My Certification : https://www.coursera.org/account/accomplishments/specialization/9H3B22NBRKKJ
This first course introduces you to Tensor Flow, a popular machine learning framework. You will learn how to build a basic neural network for computer vision and use convolutions to improve your neural network.
- Introduction: A conversation with Andrew Ng
- A primer in machine learning
- The “Hello World” of neural networks
- Working through “Hello World” in TensorFlow and Python
- Week 1 - Predicting house price.ipynb
- A conversation with Andrew Ng
- An introduction to computer vision
- Writing code to load training data
- Coding a computer vision neural network
- Walk through a notebook for computer vision
- Using callbacks to control training
- Walk through a notebook with callbacks
- Week 2 - Classifying Fashion MNIST with MLP.ipynb
- A conversation with Andrew Ng
- What are convolutions and pooling?
- Implementing convolutional layers
- Implementing pooling layers
- Improving the fashion classifier with convolutions
- Walking through convolutions
- Week 3 - Classifying Fashion MNIST with CNN.ipynb
- A conversation with Andrew Ng
- Understanding ImageGenerator
- Defining a ConvNet to use complex images
- Training the ConvNet with fit_generator
- Walking through developing a ConvNet
- Walking through training the ConvNet with fit_generator
- Adding automatic validation to test accuracy
- Exploring the impact of compressing images
- Outro: Conversation with Andrew
- Week 4 - Classifying emotion with CNN.ipynb
This second course teaches you advanced techniques to improve the computer vision model you built in Course 1. You will explore how to work with real-world images in different shapes and sizes, visualize the journey of an image through convolutions to understand how a computer “sees” information, plot loss and accuracy, and explore strategies to prevent overfitting, including augmentation and dropouts. Finally, Course 2 will introduce you to transfer learning and how learned features can be extracted from models.
- Introduction: A conversation with Andrew Ng
- Training with the cats vs. dogs dataset
- Working through the notebook
- Fixing through cropping
- Looking at accuracy and loss
- Week 1 - Classifying Cats and Dogs.ipynb
- A conversation with Andrew Ng
- Introducing augmentation
- Coding augmentation with ImageDataGenerator
- Demonstrating overfitting in cats vs. dogs dataset
- Adding augmentation to cats vs. dogs dataset
- Exploring augmentation with horses vs. humans dataset
- Week 2 - Improving Cats and Dogs Classifier.ipynb
- A conversation with Andrew Ng
- Understanding transfer learning: the concepts
- Coding your own model with transferred features
- Exploring dropouts
- Exploring transfer learning with inception
- Week 3 - Transfer learning (VGG Net).ipynb
- A conversation with Andrew Ng
- Moving from binary to multi-class classification
- Exploring multi-class classification with the rock paper scissors dataset
- Training a classifier with the rock paper scissors dataset
- Testing the rock paper scissors classifier
- Week 4 - Classifying images of sign languages.ipynb
In this third course, you’ll learn how to apply neural networks to solve natural language processing problems using TensorFlow. You’ll learn how to process and represent text through tokenization so that it’s recognizable by a neural network. You’ll be introduced to new types of neural networks, including RNNs, GRUs and LSTMs, and how you can train them to understand the meaning of text. Finally, you’ll learn how to train LSTMs on existing text to create original poetry and more!
- Introduction: A conversation with Andrew Ng
- Word-based encodings
- Using APIs
- Text to sequence
- Sarcasm, really?
- Working with the Tokenizer
- Week 1.1 - Detecting sarcasm in news headlines with LSTM and CNN.ipynb
- Week 1.2 - Exploring BBC news data.ipynb
- A conversation with Andrew Ng
- The IMDB dataset
- Looking into the details
- How can we use vectors?
- More into the details
- Remember the sarcasm dataset?
- Building a classifier for the sarcasm dataset
- Let’s talk about the loss function
- Pre-tokenized datasets
- Diving into the code
- Week 2.1 - Classifying IMDB reviews data (Embedding + MLP).ipynb
- Week 2.2 - Classifying BBC news into topics (Embedding + Conv + MLP).ipynb
- A conversation with Andrew Ng
- LSTMs
- Implementing LSTMs in code
- A word from Laurence
- Accuracy and Loss
- Using a convolutional network
- Going back to the IMDB dataset
- Tips from Laurence
- Week 3.1 - Classifying IMDB reviews (Embedding + Conv1D).ipynb
- Week 3.2 - Twitter sentiment classification (GloVe).ipynb
- A conversation with Andrew Ng
- Training the data
- Finding what the next word should be
- Predicting a word
- Poetry!
- Laurence the poet
- Week 4 - Poem generation with Bi-directional LSTM.ipynb
In this fourth course, you will learn how to solve time series and forecasting problems in TensorFlow. You’ll first implement best practices to prepare data for time series learning. You’ll also explore how RNNs and ConvNets can be used for predictions. Finally, you’ll apply everything you’ve learned throughout the Specialization to build a sunspot prediction model using real-world data!
- Introduction: a conversation with Andrew Ng
- Time series examples
- Machine learning applied to time series
- Common patterns in time series
- Introduction to time series
- Train, validation, and test sets
- Metrics for evaluating performance
- Moving average and differencing
- Trailing versus centered windows
- Forecasting
- Week 1 - Create and predict synthetic data with time series decomposition.ipynb
- A conversation with Andrew Ng
- Preparing features and labels
- Feeding a windowed dataset into a neural network
- Single layer neural network
- Machine learning on time windows
- Prediction
- More on single-layer network
- Deep neural network training, tuning, and prediction
- Week 2.1 - Prepare features and labels.ipynb
- Week 2.2 - Predict synthetic data with Linear Regression.ipynb
- Week 2.3 - Predict synthetic data with MLP.ipynb
- A conversation with Andrew Ng
- Shape of the inputs to the RNN
- Outputting a sequence
- Lambda layers
- Adjusting the learning rate dynamically
- RNNs
- LSTMs
- Coding LSTMs
- More on LSTMs
- Week 3.1 - Finding an optimal learning rate for a RNN.ipynb
- Week 3.2 - LSTM.ipynb
- A conversation with Andrew Ng
- Convolutions
- Bi-directional LSTMs
- Real data – sunspots
- Train and tune the model
- Prediction
- Sunspots
- Combining our tools for analysis