Deep Learning Weekly Issue #138

AI-powered scooters from Lime, Google's human-like chatbot, a new deep learning framework, and more

Hey folks,

This week in deep learning we bring you personalization and recommendation at Taco Bell, Lime scooters that can predict if they are on sidewalks, a protein folding model from DeepMind, an AI that can play text-based games from Uber, and a Dataset search tool from Google.

You may also enjoy an incredible AR piano playing app, a philosophical discussion of GPT-2, a thorough review of GANs, an introduction to Bayesian neural networks, an overview of the Economics of AI, a new lightweight type-checked deep learning framework, and more.

As always, happy reading and hacking. If you have something you think should be in next week's issue, find us on Twitter: @dl_weekly.

Until next week!


Here's Why Taco Bell Added Artificial Intelligence Technology To Its Mobile App

Machine learning-based personalization and recommendation is becoming a requirement in the food industry.

Google Dataset Search comes out of beta.

After a year in public beta, you can now search through over 25 million indexed datasets.

AlphaFold: Using AI for scientific discovery

DeepMind discusses their recent paper on protein folding.

Uber’s AI plays text-based games like a human

Research out of Uber uses Seq2Seq models to generate policies agents use to play games.

Lime uses sensor data to keep scooters off sidewalks

Scooters use speedometer and accelerometer data to predict which surface they are riding on.

Mobile + Edge

On-Device Machine Learning with SwiftUI and PyTorch Mobile

Build a simple image classification iOS app that demonstrates how to integrate PyTorch Mobile with SwiftUI.

AR Pianist uses machine learning to superimpose a virtual pianist on your piano.

One of the most seamless and impressive AR demo’s I’ve seen.

Google brings cross-platform AI pipeline framework MediaPipe to the web

MediaPipe, a new optimized combined graphics and AI pipeline gets a Javascript API.


GPT-2 and the Nature of Intelligence

Gary Marcus explores the philosophical implications of large transformer models.

A Gentle Introduction to Deep Learning for Graphs

A tutorial on deep learning for graphs that has something for beginners and experts.

A Review on Generative Adversarial Networks: Algorithms, Theory, and Applications

An absurdly well referenced review of GANs.

Putting the Neural back into Networks

An introduction to spiking neurons and how they relate to artificial neural networks.

The Economics of AI Today

A great review of how economists think about AI’s impact on markets and people.

An introduction to (and puns on) Bayesian neural networks

Theory and examples behind Bayesian neural networks, implementations in JAX.

Libraries & Code

[Github] kornia/kornia

Kornia, a differentiable computer vision library for PyTorch has released a new version with GPU support for data augmentation.

Thinc: A lightweight deep learning framework

Thinc is a type-checked functional deep learning library with support for layers defined in PyTorch, TensorFlow or MXNet.

Papers & Publications

FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence

Abstract: ….In this paper, we demonstrate the power of a simple combination of two common SSL methods: consistency regularization and pseudo-labeling. Our algorithm, FixMatch, first generates pseudo-labels using the model's predictions on weakly-augmented unlabeled images. For a given image, the pseudo-label is only retained if the model produces a high-confidence prediction. The model is then trained to predict the pseudo-label when fed a strongly-augmented version of the same image. Despite its simplicity, we show that FixMatch achieves state-of-the-art performance across a variety of standard semi-supervised learning benchmarks, including 94.93% accuracy on CIFAR-10 with 250 labels and 88.61% accuracy with 40 -- just 4 labels per class....

Towards a Human-like Open-Domain Chatbot

Abstract: We present Meena, a multi-turn open-domain chatbot trained end-to-end on data mined and filtered from public domain social media conversations. This 2.6B parameter neural network is trained to minimize perplexity, an automatic metric that we compare against human judgement of multi-turn conversation quality. To capture this judgement, we propose a human evaluation metric called Sensibleness and Specificity Average (SSA), which captures key elements of good conversation. Interestingly, our experiments show strong correlation between perplexity and SSA. The fact that the best perplexity end-to-end trained Meena scores high on SSA (72% on multi-turn evaluation) suggests that a human-level SSA of 86% is potentially within reach if we can better optimize perplexity. Additionally, the full version of Meena (with a filtering mechanism and tuned decoding) scores 79% SSA, 23% higher than the next highest scoring chatbot that we evaluated.