The final Deep Learning Specialization course is now out

After a long wait, the final and much-anticipated course in the Coursera Deep Learning Specialization series taught by Andrew Ng, called Sequence Models, has now been released.

The first week will be about Recurrent Neural Networks, the second week will address Natural Language Processing & Word Embeddings and the final week will be about Sequence models & Attention mechanism.

 

Fast.ai 2018 has been released, and it’s truly awesome

Last year the 2017 course of fast.ai was amazing, which taught state of the art deep learning to coders. There are so many goodies in the blog post about the Fast.ai 2018 launch which is available now. This year they held the course using Pytorch instead of Keras and wrote their own library for speeding up development and were the first to add several implementations from papers to the library such as Learning Rate Finder (Smith 2015) and Stochastic Gradient Descent with Restarts (SGDR). With one line of code, you can also get the images that the classifier gets wrong.

17 of the 20 top participants in a kaggle competitors were students in the preview course.

I recommend reading the blog post and taking the course.

DeepMinds AlphaZero beat the strongest open-source chess engine

We know that computers have beat humans in chess, that was a great breakthrough and a milestone in AI.
The worlds strongest AI for chess, called Stockfish was recently dethroned by a deep reinforcement AI by Googles DeepMind called AlphaZero.

Here is a walkthrough of the third game further explanation on chess.com “How Does AlphaZero Play Chess?

Titta på AlphaZero vs Stockfish Chess Match: Game 3 från Chess på www.twitch.tv

Image recognition and AI on a Raspberry Pi 3 using MobileNets and Neural Compute Stick

If you are building a robot driven by Raspberry Pi and want to use image recognition and object detection you may want to look into Googles Mobile Nets platform which lets you do use a several mobile-first computer vision models for TensorFlow, combined with an Intel Movidius Neural Compute Stick on a Rasrberry PI 3.  The MobileNets platform is designed to be run on resource conservative devices while maintaining accuracy and the latter will give you an order of magnitude more compute power than running the detection on the raspberrys CPU.

Get a hang of reinforcement learning for games

Check this out: Introduction to AI for Video Games (Reinforcement Learning) by Siraj Raval

And this: Lecture 10: Reinforcement Learning in CS188 Artificial Intelligence, Fall 2013 (University of California, Berkley)

Also this lecture on Deep Reinforcement Learning from Stanford CS231n

And this playlist: Introduction to Reinforcement Learning by Deep Mind

If you are interested in creating Games in Unity, this is also something you shall check out https://github.com/Unity-Technologies/ml-agents

And this: Playlist (Unity Machine Learning)

This Lecture from MIT was also interesting (and the DeepTraffic assignment was fun)

DeepSchool.io

There is a new kid on the block in terms of online courses on Deep Learning.

DeepSchool.io is a set of Jupyter notebooks that teach you the basics and different concepts you need in order to get started and being productive in Depp Learning. They are also videos supporting the notebook, although not for every notebook yet.

It differs from fast.ai in that the videos are shorter and the notebooks are mostly self-explanatory.

The goal of the project is to make Deep Learning accessible to everyone, make it practical, make learning open source and fun.

These are the topics covered:

  1. Lesson 0: Introduction to regression.
  2. Lesson 1: Penalising weights to fit better (scikit learn intro)

Mathematics (optional)

  1. Lesson 2: Gradient Descent. Using basic optimisation methods.
  2. Lesson 3: Tensorflow intro: zero layer hidden networks (i.e. normal regression).
  3. Lesson 4: Tensorflow hidden layer introduction.

Deep Learning

  1. Lesson 5: Using Keras to simplify multi-layer neural nets.
  2. Lesson 6: Embeddings to deal with categorical data. (Keras)
  3. Lesson 7: Word2Vec. Embeddings to visualise words. (Tensorflow)
  4. Lesson 8: Application – Bike Sharing predictions
  5. Lesson 9: Choosing Number of Layers and more
  6. Lesson 10: XGBoost – A quick detour from Deep Learning
  7. Lesson 11: Convolutional Neural Nets (MNIST dataset)
  8. Lesson 12: CNNs and BatchNormalisation (CIFAR10 dataset)
  9. Lesson 13: Transfer Learning (Dogs vs Cats dataset)

Advanced Topics

  1. Lesson 14: LSTMs – Sentiment analysis.
  2. Lesson 15: LSTMs – Shakespeare.
  3. Lesson 16: LSTMs – Trump Tweets.
  4. Lesson 17: Trump – Stacking and Stateful LSTMs.
  5. Lesson 18: Fake News Classifier

You can read more here.

AlphaGo Zero, beats previous world champion winner AlphaGo, 100-0

You heard of DeepMinds AlphaGo that beat worlds best Go player in the game everyone said computers would still need ten years to beat humans in.

That version trained on millions of expert human gameplays and then trained on itself through reinforcement learning.

This version skips all human gameplay and learns by playing against itself through a novel reinforcement learning method. It only has the rules of the game and starts to play against itself, making adjustments and keeping the versions that improve.

Blog Post: https://deepmind.com/blog/alphago-zero-learning-scratch/
Research Page: https://deepmind.com/research/alphago/

If you would like to replicate the research, there is an open source project that is based on the paper https://github.com/gcp/leela-zero. However, in order to get the same results as AplhaGo Zero, you would need to have the same weights, and in order to achieve similar weights, you would need to have access to the same computing power as they. It would take 1700 years on commodity computers. The projects aim is to make a distributed effort to repeat the work.

Course 4 [deeplearning.ai] has been released!

The fourth course, Convolutional Neural Networks of Deeplearning.ai has now been released on coursera. People have been waiting for this one, but i think that the delay was to make the material very up to date with current research results. The four weeks of learning deals with:

  1. Foundations of Convolutional Neural Networks
  2. Deep convolutional models: case studies
  3. Object detection
  4. Special applications: Face recognition & Neural style transfer