Skip to content

kb1907/Natural-Language-Processing-Specialization

Repository files navigation

Natural Language Processing Specialization


WHAT I LEARNED


  • Use logistic regression, naïve Bayes, and word vectors to implement sentiment analysis, complete analogies & translate words.

  • Use dynamic programming, hidden Markov models, and word embeddings to implement autocorrect, autocomplete & identify part-of-speech tags for words.

  • Use recurrent neural networks, LSTMs, GRUs & Siamese networks in Trax for sentiment analysis, text generation & named entity recognition.

  • Use encoder-decoder, causal, & self-attention to machine translate complete sentences, summarize text, build chatbots & question-answering.

There are 4 Courses in this Specialization


  • In the first course of the Natural Language Processing Specialization

  • I performed sentiment analysis of tweets using logistic regression and then naïve Bayes,

  • I used vector space models to discover relationships between words and used PCA to reduce the dimensionality of the vector space and visualize those relationships, and

  • I wrote a simple English-to-French translation algorithm using pre-computed word embeddings and locality-sensitive hashing to relate words via approximate k-nearest neighbor search.

Projects


  • In the second course of the Natural Language Processing Specialization

  • I wrote a simple auto-correct algorithm using minimum edit distance and dynamic programming,

  • I applied the Viterbi Algorithm for part-of-speech (POS) tagging, which is vital for computational linguistics,

  • I wrote a better auto-complete algorithm using an N-gram language model, and

  • I wrote my own Word2Vec model that uses a neural network to compute word embeddings using a continuous bag-of-words model.

Projects


  • In the third course of the Natural Language Processing Specialization

  • I trained a neural network with GLoVe word embeddings to perform sentiment analysis of tweets,

  • I generated synthetic Shakespeare text using a Gated Recurrent Unit (GRU) language model,

  • I trained a recurrent neural network to perform named entity recognition (NER) using LSTMs with linear layers, and

  • I used so-called ‘Siamese’ LSTM models to compare questions in a corpus and identify those that are worded differently but have the same meaning.

Projects


Projects


Disclaimer


  • DeepLearning.AI makes course notes available for educational purposes.
  • Project solutions are just for educational purposes. I highly recommend trying and solving project/program assignments on your own.

All the best 🤘

About

Natural Language Processing Specialization - Course Notes & Projects

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published