• Stars
    star
    212
  • Rank 186,122 (Top 4 %)
  • Language
    Jupyter Notebook
  • Created over 7 years ago
  • Updated over 7 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

A seq2seq model that can correct spelling mistakes.

Spell-Checker

The objective of this project is to build a model that can take a sentence with spelling mistakes as input, and output the same sentence, but with the mistakes corrected. The data that we will use for this project will be twenty popular books from Project Gutenberg. Our model is designed using grid search to find the optimal architecture, and hyperparameter values. The best results, as measured by sequence loss with 15% of our data, were created using a two-layered network with a bi-direction RNN in the encoding layer and Bahdanau Attention in the decoding layer. FloydHub's GPU service was used to train the model.

All of the books that I used for training can be found in books.zip.

To view my work most easily, see the .ipynb file.

I wrote an article that explains how to create the input data (the sentences with spelling mistakes) for this model.

More Repositories

1

Text-Summarization-with-Amazon-Reviews

A seq2seq model that can generate summaries from fine food reviews on Amazon.
HTML
225
star
2

Chatbot-from-Movie-Dialogue

Built a simple chatbot from a sequence-to-sequence model with TensorFlow.
HTML
147
star
3

Movie-Reviews-Sentiment

Used two different methods to predict the sentiment (positive or negative) of movie reviews.
HTML
55
star
4

Predicting-the-Dow-Jones-with-Headlines

Used Keras to build a model (CNNs + LSTMs) to predict the opening price change of the Dow Jones.
HTML
36
star
5

500-Greatest-Albums

Exploring Rolling Stone Magazine's list of "The 500 Greatest Albums of All Time."
HTML
16
star
6

Predicting-Credit-Card-Fraud

Used TensorFlow to build a neural network that can predict fraudulent credit card transactions.
Jupyter Notebook
11
star
7

Titanic-Kaggle-Competition

My analysis for the 'Titanic: Machine Learning from Disaster' competition, hosted by Kaggle.com
HTML
8
star
8

Predicting-Similar-Questions

Used TfidfVectorizer, Doc2Vec, and deep learning to predict if pairs of questions have the same meaning.
HTML
6
star
9

Tweet-Like-Trump

A one2seq model that can generate tweets similar to those of Donald Trump.
HTML
6
star
10

Artificial-Intelligence-Nanodegree

The projects that I completed for my Artificial Intelligence Nanodegree - Udacity
Jupyter Notebook
5
star
11

Bike-Sharing-in-SF-and-Seattle

An analysis of the bike sharing services in San Francisco and Seattle.
HTML
5
star
12

NYC-Taxi-Trip-Duration

My work for Kaggle's "New York City Taxi Trip Duration" competition
Jupyter Notebook
5
star
13

neuroblastoma_gene_signature

Validate a gene signature for evaluating the survival of patients with neuroblastoma.
R
3
star
14

AirBnB-Predicting-Destination

Used TensorFlow to build a neural network that can predict which country a new AirBnB user will book their first trip to.
HTML
3
star
15

Identify-Fraud-From-Enron

Used machine learning to predict which Enron employees committed fraud; as part of my Data Analyst Nanodegree from Udacity.
Python
2
star
16

bioinformatics

Small projects related to bioinformatics
HTML
1
star
17

Language-Translation

Built a sequence-to-sequence model to translate text from English to French.
HTML
1
star
18

cooking-with-ingles

JavaScript
1
star
19

First-Neural-Network

Used Numpy to build a Neural Network to predict daily ridership of a bike sharing service.
HTML
1
star
20

Face-Generation

Built a generative adversarial network to create new faces.
HTML
1
star
21

nf-core-spatialtranscriptomicsgeomx

Jupyter Notebook
1
star
22

Comparing-Books

Used Word2Vec and Doc2Vec to compare Project Gutenberg's 20 Most Popular Books
HTML
1
star