All artificial intelligence methods today are around machine learning modeling and use some form of sophisticated correlation or association method, which can be approximated to brute-force robot learning. It is about reverse engineering existing features/patterns and providing useful “forward engineering” solutions like:
- Self-driving cars
- Detecting diseases from X-rays/MRIs
- Robots in manufacturing
- Chatbots for customer service
- Insert your future application here.
The logic is that X event follows Y, or if X event occurs with Y in historical and simulation data, so today we can create automated models around it and use the models to predict an unknown object/variable/situation and even prescribe actions. Today, we are progressing to explain to some extent how we are predicting that way. However, today’s machine learning is about figuring the “what” in images, speech, numbers, translation and text. However, it doesn’t address the “why” question. “What” can work if the environment under which training happened also occurs, to some extent, during prediction, at least the context — right? The question so far has been, “Can we figure out all the ‘whats’ in our models?” Yes, to some extent ,if we can train on lots and lots of data within the same context. Examples of context would be playing chess or Pokemon Go, driving on streets, website browsing etc. Techniques such as deep learning/reinforcement learning with GPU hardware, big cluster farms and days/weeks of training make it possible.
The idea of machine learning or deep learning is to mimic by feature extraction, memorizing and generalizing instances of interest. Human brains work both with memorizing, generalizing problems, but also to “creatively infer causation.” This is yet to be seen in AI algorithms today and, as you can imagine, may move us toward general artificial intelligence or start a new AI winter.
Today’s AI cannot creatively infer causation on its own
Here’s are some challenges in today’s AI world:
- The models can tell that the sunset in a beach will be red or yellow. AI cannot tell why it is so on its own. It will not know that the sun’s rays scatter differently with red wavelength and also because the atmosphere has pollutants.
- The models can tell that an X-ray image shows a cancerous polyp, but it cannot tell why. It will not know that the polyp is caused because of DNA mutation, food factors and an external factor/.trigger/environment from six months ago.
- The models can tell that an umbrella in a picture is for either rain or hot sun, but cannot tell why it was designed that way in the first place. It will not know that the length and color of the umbrella were designed to reflect the rays of the sun, balancing with the average wind speed, not to fly off.
- The models can tell that a behavior is fraudulent/suspect, but cannot really explain why the fraudster is targeting this particular business and using this technique.
- The models can turn on chatbots to answer questions intelligently by learning from a large corpus of chats, text or Q&A from the past. It will, however, miss sarcasm and humor or main intent at the outset.