Google’s brain team is open sourcing Tensor2Tensor, a new deep learning library designed to help researchers replicate results from recent papers in the field and push the boundaries of what’s possible by trying new combinations of models, datasets and other parameters. The sheer number of variables in AI research combined with the fast pace of new developments makes it difficult for experiments run in two distinct settings to match. This is a pain for researchers and a drag on research progress.

The Tensor2Tensor library makes it easier to maintain best practices while conducting AI research. It comes equipped with key ingredients including hyperparameters, data-sets, model architectures and learning rate decay schemes.

The best part is that any of these components can be swapped in and out in a modular fashion without completely destroying everything. From a training perspective, this means that with Tensor2Tensor you can bring in new models and data sets at any time — a much simpler process than would ordinarily be possible.

Google isn’t alone in its pursuits to help make research more reproducible outside the lab. Facebook recently open sourced ParlAI, its tool to facilitate dialog research that comes prepackaged with commonly used datasets.

Similarly, Google’s Tensor2Tensor comes with models from recent Google research projects like “Attention Is All You Need” and “One Model to Learn Them All.” Everything is available now on Github so you can start training your own deep learning-powered tools.

Source :

tch

Load More By Bekim Em
Load More In News

Leave a Reply

Your email address will not be published. Required fields are marked *

Check Also

Twitter ought to hurry up and let users flag fake news

Twitter is believed to be testing a feature to allow users to report tweets containing fak…