Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon

Google introduces and open-sources Lingvo, a scalable TensorFlow framework for Sequence-to-Sequence Modeling

Save for later
  • 3 min read
  • 26 Feb 2019

article-image
Google researchers announced a new TensorFlow framework, called Lingvo, last week. Lingvo offers a complete solution for collaborative deep learning research, with a particular focus towards sequence modeling tasks such as machine translation, speech recognition, and speech synthesis.  

The TensorFlow team also announced yesterday that it is open sourcing Lingvo. “To show our support of the research community and encourage reproducible research effort, we have open-sourced the framework and are starting to release the models used in our papers”, states the TensorFlow team.

https://twitter.com/GoogleAI/status/1100177047857487872

Lingvo has been designed for collaboration and has a code with a consistent interface and style that is easy to read and understand. It also consists of a flexible modular layering system that promotes code reuse. Since it involves many people using the same codebase, it is easier to employ other ideas within your models. Also, you can adapt to the existing models to new datasets with ease.

Other than that, Lingvo makes it easier to reproduce and compare the results in research. This is because Lingvo adopts a system where all the hyperparameters of a model get configured within their own dedicated sub-directory that is separate from the model logic. All the models within Lingvo are built from the same common layers which allow them to be compared with each other easily. Also, all of these models have the same overall structure from input processing to loss computation, and all the layers consist of the same interface.

google-introduces-and-open-sources-lingvo-a-scalable-tensorflow-framework-for-sequence-to-sequence-modeling-img-0

                                        Overview of the LINGVO framework

Moreover, all the hyperparameters within Lingvo get explicitly declared and their values get logged at runtime. This makes it easier to read and understand the models. Lingvo can also easily train the on production scale datasets. There’s also additional support provided for synchronous and asynchronous distributed training. In Lingvo, inference-specific graphs are built from the same shared code used for training. Also, quantization support has been built in directly into the framework.

Unlock access to the largest independent learning library in Tech for FREE!
Get unlimited access to 7500+ expert-authored eBooks and video courses covering every tech area you can think of.
Renews at €18.99/month. Cancel anytime


Lingvo initially started out for natural language processing (NLP) tasks but has become very flexible and is also applicable to models used for tasks such as image segmentation and point cloud classification. It also supports distillation, GANs, and multi-task models. Additionally, the framework does not compromise on the speed and comes with an optimized input pipeline and fast distributed training.

For more information, check out the official LINGVO research paper.

Google engineers work towards large scale federated learning

Google AI researchers introduce PlaNet, an AI agent that can learn about the world using only images

Google to acquire cloud data migration start-up ‘Alooma’