What is Deep Learning's breaking point? Is there one?

By arvind |Email | Oct 30, 2018 | 6021 Views

Artificial Intelligence is living the dream life of a celebrity, being the buzz of the town for a decade and for all the right reasons. Right as I am writing this article there is a good chance a worker has been replaced by IBM Watson or an algorithm laughing at a doctor because it was better at diagnosing cancer. There are AI's that are writing articles to (my job is in a bit of danger too; maybâ?¦.it has already been taken by us-an AI). Nah I am still good and writing, no is no shortage of a new AI start-up coming up every day, and the word on the street is some of them are good too. These AI start-ups or companies are promising to solve our personal and business related problems; all of it with the help of machine learning.
Why is that these juicer grinders, these routers all of a sudden all have a tag of powered by AI. I have AI I have AIâ?¦..there is a speaker that can order me lunch now, call play music on my command (and I love everything about it, just that it stops responding when music is played on maximum volume). 
Most of this AI hubbub is given air by reporters who have never trained a neural network and by start-ups or by those that are hoping to be acqui-hired for their engineering talent despite not having solved any real business problems. No wonder why there are so many misconceptions surrounding- what AI can and cannot do? 
Deep learning is the master key
For something that was invented in 1960 it sure took it some time to come of age, back then there were not the right tools to give it the necessary boost that big data and today's computational power could provide to turn them into something useful. Deep Learning a new practice have risen in the past few years, now we have something that can accurately apply complex neural network architectures to model patterns in data. The result speak for itself, it is impressive like nothing before. It is because of deep learning the computers are able to identify objects that are present in a photo or motion picture and transcribe speech to text better than us humans. Even Google switched Google's Translates architecture with neural networks, and now machine translation is also closing in on human performance.
As I mentioned earlier results speak for itself- better at predicting crop yield, better at diagnosing cancer and more accurately than the top doctors or physicians.
John Launchbury, a director at DARPA, describes three waves of artificial intelligence:
1. Handcrafted knowledge or expert systems like IBM's Deep Blue or Watson
2. Statistical learning, which includes machine learning and deep learning
3. Contextual adaption, which involves constructing reliable, explanatory models for real-world phenomena using sparse data, like humans do
As part of the current second wave of AI, deep learning algorithms work well because of what Launchbury calls the "manifold hypothesis" (see below). In Layman's terms, this refers to how different types of high-dimensional natural data tend to clump and be shaped differently when visualized in lower dimensions.

By mathematically manipulating and separating data clumps, deep neural networks can distinguish different data types. While neural nets can achieve nuanced classification and predication capabilities, they are essentially what Launchbury calls "spreadsheets on steroids."

With deep learning come deep problems
At the recent 'AI By The Bay' conference, Francois Chollet emphasized that deep learning is simply more powerful pattern recognition than previous statistical and machine learning methods. "The most important problem for AI today is abstraction and reasoning," explains Chollet, an AI researcher at Google and famed inventor of widely used deep learning library Keras. "Current supervised perception and reinforcement learning algorithms require lots of data, are terrible at planning, and are only doing straightforward pattern recognition." By contrast, humans "learn from very few examples, can do very long-term planning and are capable of forming abstract models of a situation and [manipulating] these models to achieve extreme generalization."
Even simple human behaviors are laborious to teach to a deep learning algorithm. Let's examine a situation such as avoiding being hit by a car as you walk down the road. If you go the supervised learning route, you'd need huge data sets of car situations with clearly labeled actions to take, such as "stop" or "move." Then you'd need to train a neural network to learn the mapping between the situation and the appropriate action.
If you go the reinforcement learning route, where you give an algorithm a goal and let it independently determine the ideal actions to take, the computer would need to die thousands of times before learning to avoid cars in different situations.
"You cannot achieve general intelligence simply by scaling up today's deep learning techniques," warns Chollet.
Humans only need to be told once to avoid cars. We're equipped with the ability to generalize from just a few examples and are capable of imagining (i.e. modeling) the dire consequences of being run over. Without losing life or limb, most of us quickly learn to avoid being overrun by motor vehicles.
While neural networks achieve statistically impressive results across large sample sizes, they are "individually unreliable" and often make mistakes humans would never make, such as classifying a toothbrush as a baseball bat.

Your results are only as good as your data. Neural networks fed inaccurate or incomplete data will simply produce the wrong results. The outcomes can be both embarrassing and damaging. In two major PR debacles, Google Images incorrectly classified African Americans as gorillas, while Microsoft's Tay learned to spew racist, misogynistic hate speech after only hours training on Twitter.
Undesirable biases are implicit in our input data. Google's massive Word2Vec embeddings are built off of 3 million words from Google News. The data set makes associations such as "father is to doctor as mother is to nurse," which reflect gender bias in our language. Researchers such as Tolga Bolukbasi of Boston University have taken to human ratings on Mechanical Turk to perform "hard de-biasing" to undo such associations.
Such tactics are essential since, according to Bolukbasi, "word embeddings not only reflect stereotypes but can also amplify them." If the term "doctor" is more associated with men than women, then an algorithm might prioritize male job applicants over female job applicants for open physician positions.
Finally, Ian Goodfellow, inventor of generative adversarial networks (GANs), showedthat neural networks can be deliberately tricked with adversarial examples. By mathematically manipulating an image in a way that is undetectable to the human eye, sophisticated attackers can trick neural networks into grossly misclassifying objects.

The dangers such adversarial attacks pose to AI systems are alarming, especially since adversarial images and original images seem identical to us. Self-driving cars could be hijacked with seemingly innocuous signage and secure systems could be compromised by data that initially appears normal.
Deep learning and beyond it
How can we overcome the limitations of deep learning and proceed toward general artificial intelligence? Chollet's initial plan of attack involves using "super-human pattern recognition, like deep learning, to augment explicit search and formal systems," starting with the field of mathematical proofs. Automated Theorem Provers (ATPs) typically use brute force search and quickly hit combinatorial explosions in practical use. In the DeepMath project, Chollet and his colleagues used deep learning to assist the proof search process, simulating a mathematician's intuitions about what lemmas (a subsidiary or intermediate theorem in an argument or proof) might be relevant.
Another approach is to develop more explainable models. In handwriting recognition, neural nets currently need to be trained on tens to hundreds of thousands of examples to perform decent classification. Instead of looking at just pixels, however, Launchbury of DARPA explains that generative models can be taught the strokes behind any given character and can use this physical construction information to disambiguate between similar numbers, such as a 9 or a 4.
Yann LeCun, inventor of convolutional neural networks (CNNs) and director of AI research at Facebook, proposes "energy-based models" as a method of overcoming limits in deep learning. Typically, a neural network is trained to produce a single output, such as an image label or sentence translation. LeCun's energy-based models instead give an entire set of possible outputs, such as the many ways a sentence could be translated, along with scores for each configuration.
Geoffrey Hinton, widely called the "father of deep learning" wants to replace neurons in neural networks with "capsules" that he believes more accurately reflect the cortical structure in the human mind. "Evolution must have found an efficient way to adapt features that are early in a sensory pathway so that they are more helpful to features that are several stages later in the pathway," Hinton explains. He hopes that capsule-based neural network architectures will be more resistant to the adversarial attacks that Goodfellow illuminated above.

Source: HOB