Wolfram Image Identification Project - A Neural Network By Any Other Name
Written by Mike James   
Thursday, 14 May 2015

Wolfram's latest project is to bring image identification to the masses with a new ImageIdentify function. How does it work? Some sort of neural network is about as far as you can get. 

If you work in the AI area, and neural networks in particular, you might be a bit upset to discover that in the latest post to his blog, Steven Wolfram seems to suggest that he has just discovered convolutional neural networks. The post is very low on awareness of anything that anyone else has done recently and presents the image classifier as if it was all Wolfram's own work.

A web site has been set up that uses the Wolfram languages ImageIdentify function to return tags for any image you care to upload. You also get the opportunity to supply the correct tag or select a more appropriate tag from the runners up. Notice that you cannot train the classifier on your own image set - it either recognises the image or it doesn't. 




The ImageIdentify function is supposed to select the appropriate machine learning method for the image. However, the only method mentioned is a neural network. You pass the image to the function and it passes you back a symbolic entity that can be further processed. Having the recognition built into Wolfram language does mean that you can use it as a component in a more complete system - you could get definitions of objects detected or compile statistics. 

This is all very reasonable, fun, and perhaps even useful. What is odd it the way that it is represented in the Wolfram blog. After a brief introduction to the function, the post proceeds to speculate on the nature of neural networks and recounts a brief history. Next neural networks are described as being "attractors", which is certainly what a Hopfield network is, but the description seems less appropriate for a standard convolution neural network. However, it gives Wolfram the connection needed to talk about his pet subject, cellular automata.

The post then continues to outline the development of the new function, but all without saying what the neural network architecture is; what the training set was; and how it was trained. As the function is just an implementation of a trained neural network, it might even be one of the many published trained networks - although this seems unlikely.



Even when a neural network gets it wrong you can see where it is coming from...


Wolfram doesn't actually claim that this is all newly invented, but if you didn't know otherwise this is the impression you get. The only reference to another type of neural network or worker in the field is to the McCulloch-Pitts Neuron which, while very interesting is hardly a modern neural network. There is no mention of Geoffrey Hinton or Yann LeCun and no mention of the Large Scale Visual Recognition Challenge or ImageNet. 

It is good to have a deep convolutional neural network in Mathematica, and it might allow non-experts and experts alike to build systems for which image recognition is just one of many steps. However, the presentation of the closed proprietary approach to neural networks without a mention of the prior state of the art leaves a nasty taste in the mouth. 



Andrew Tanenbaum Gains ACM Award

Andrew Tanenbaum has been awarded the 2023 ACM System Software Award for MINIX the operating system he created for teaching purposes and which was an important influence on Linux.

More Jetpack Compose Updates

The July release of Android Jetpack Compose has some improvements and it is faster, but is it what we want?

More News


kotlin book



or email your comment to: comments@i-programmer.info

Last Updated ( Wednesday, 20 May 2015 )