Facebook today released the latest version of its deep learning library PyTorch with quantization and Google Cloud TPU support for speedier training of machine learning models.
Tensor processing unit support begins with the ability to train a model with a single chip and will later be extended to Cloud Pods, Facebook CTO Mike Schroepfer said today.
Also new today are PyTorch Mobile for deployment of ML on edge devices starting with Android and iOS devices; CryptTen, a tool for encrypted machine learning; and Captum, a tool for explainability of machine learning models.
The news is being announced at the PyTorch Developer Conference today at The Midway in San Francisco.
Available today, PyTorch 1.3 comes with the ability to quantize a model for inference on to either server or mobile devices. Quantization is a way to perform computation at reduced precision.
The latest version of PyTorch will support eager mode quantization at 8-bit integer with the eager mode Python API and will allow for post-training quantization in a variety of approaches like dynamic quantization and quantization-aware training.
Version 1.3 also comes with named tensors, a way to write cleaner code, PyTorch project manager Joe Spisak told VentureBeat in a phone interview.
“This is really to allow you to write cleaner code,” he said. “I’m actually able to embed those into the code instead of having a comment that says, hey, this number is the height, this number is the width etc., so it makes for more I would say readable and cleaner code and more maintainable code.”
Facebook has been working on privacy-preserving models with Google’s DeepMind research scientist Andrew Trask, who this spring launched a Udacity course to teach developers how to use things like federated learning and PySyft, an open source project with a library for encrypted deep learning with extensions of PyTorch, TensorFlow, and Keras.
Facebook also introduced two new open source frameworks: Detectron2, a new version of the Detectron object detect system, as well as speech recognition extensions typically used for translation.
PyTorch 1.1 was released this spring at the F8 developer conference with support for TensorBoard.
Over the past two years, Facebook has moved away from using its predecessor Torch or Caffe2 in an effort to make PyTorch the main tool for deep learning, CTO Mike Schroepfer said at the start of the conference. This has been essential to keeping researchers and developers on the same page.
“This means it’s now the de facto tool for doing machine learning [and] deep learning at Facebook, not only for research that happens organically, but in production, so the vast majority of our models are now trained on PyTorch,” Schroepfer said onstage. “And this is true across multiple domains both in computer vision, NLP, speech, translation — all these systems are now using PyTorch.”
Tesla, Microsoft, Uber, and other large companies have also adopted PyTorch, he said.
The PyTorch deep learning library has considerably grown in popularity among AI practitioners in the past year. Nearly 1,200 developers now contribute to the open source project, and an O’Reilly analysis released in June found that arXiv mentions of PyTorch went up 194% from January to June, and are now on par with TensorFlow mentions.
The audio problem: Learn how new cloud-based API solutions are solving imperfect, frustrating audio in video conferences. Access here