|Facebook Shares Deep Learning Tools|
|Written by Alex Armstrong|
|Thursday, 22 January 2015|
Facebook AI Research has announced that is open sourcing the deep-learning modules that enable it to train larger neural nets in less time than those already available.
Since Yann LeCun was recruited to head Facebook's newly founded AI Group, FAIR in 2013 it has become a team of 36 people and made great strides forward.
Fortunately for the area of deep learning and convolutional nets it believes that:
Progress in science and technology accelerates when scientists share not just their results, but also their tools and methods.
Hence its decision to do just that with a set of tools that give a 23.5x speed-up over publicly available convolutional layer codes, and the ability to parallelize neural networks training over GPU cards.
The tools are being made available for Torch. an open source development environment for numerics, machine learning, and computer vision widely used at a number of academic labs as well as at Google/DeepMind, Twitter, NVIDIA, AMD, Intel, and many other companies.
The following fast nn modules for Convnets and neural networks in general are provide a a plug-in to the Torch-7 framework:
Facebook has also recently released iTorch, an interface for Torch using iPython with visualization and plotting and previously has made available fbnn, extensions to torch/nn, fbcuda, extensions to CUDA, and fblualib libraries and utilities for Lua.
Concluding the announcement on the FAIR Blog, Soumith Chintala notes:
We hope that these high-quality code releases will be a catalyst to the research community and we will continue to update them from time to time.
To be informed about new articles on I Programmer, install the I Programmer Toolbar, subscribe to the RSS feed, follow us on, Twitter, Facebook, Google+ or Linkedin, or sign up for our weekly newsletter.
or email your comment to: firstname.lastname@example.org
|Last Updated ( Thursday, 04 October 2018 )|