Natural Language Processing (NLP) uses algorithms to understand and manipulate human language. This technology is one of the most broadly applied areas of machine learning. As AI continues to expand, so will the demand for professionals skilled at building models that analyze speech and language, uncover contextual patterns, and produce insights from text and audio. This Specialization will equip you with the state-of-the-art deep learning techniques needed to build cutting-edge NLP systems. By the end of this Specialization, you will be ready to design NLP applications that perform question-answering and sentiment analysis, create tools to translate languages and summarize text, and even build chatbots.
This is the first course of the Natural Language Processing Specialization.
Week 1: Logistic Regression for Sentiment Analysis of Tweets
- Use a simple method to classify positive or negative sentiment in tweets
Week 2: Naïve Bayes for Sentiment Analysis of Tweets
- Use a more advanced model for sentiment analysis
Week 3: Vector Space Models
- Use vector space models to discover relationships between words and use principal component analysis (PCA) to reduce the dimensionality of the vector space and visualize those relationships
Week 4: Word Embeddings and Locality Sensitive Hashing for Machine Translation
- Write a simple English-to-French translation algorithm using pre-computed word embeddings and locality sensitive hashing to relate words via approximate k-nearest neighbors search
This is the second course of the Natural Language Processing Specialization.
Week 1: Auto-correct using Minimum Edit Distance
- Create a simple auto-correct algorithm using minimum edit distance and dynamic programming
Week 2: Part-of-Speech (POS) Tagging
- Apply the Viterbi algorithm for POS tagging, which is important for computational linguistics
Week 3: N-gram Language Models
- Write a better auto-complete algorithm using an N-gram model (similar models are used for translation, determining the author of a text, and speech recognition)
Week 4: Word2Vec and Stochastic Gradient Descent
- Write your own Word2Vec model that uses a neural network to compute word embeddings using a continuous bag-of-words model
This is the third course in the Natural Language Processing Specialization.
Week 1: Sentiment with Neural Nets
- Train a neural network with GLoVe word embeddings to perform sentiment analysis of tweets
Week 2: Language Generation Models
- Generate synthetic Shakespeare text using a Gated Recurrent Unit (GRU) language model
Week 3: Named Entity Recognition (NER)
- Train a recurrent neural network to perform NER using LSTMs with linear layers
Week 4: Siamese Networks
- Use so-called ‘Siamese’ LSTM models to compare questions in a corpus and identify those that are worded differently but have the same meaning
This is the fourth course in the Natural Language Processing Specialization.
Week 1: Neural Machine Translation with Attention
- Translate complete English sentences into French using an encoder/decoder attention model
Week 2: Summarization with Transformer Models
- Build a transformer model to summarize text
Week 3: Question-Answering with Transformer Models
- Use T5 and BERT models to perform question answering
Week 4: Chatbots with a Reformer Model
- Build a chatbot using a reformer model