“Some people worry that artificial intelligence will make us feel inferior, but then, anybody in his right mind should have an inferiority complex every time he looks at a flower.”—Alan Kay
Raymond Kurzweil, the director of engineering at Google and a renowned futurist predicted in The Singularity Is Near: When Humans Transcend Biology that by 2029 computers will be cleverer than humans. While such a mindboggling prediction can be a matter of debate, there is little doubt that in the future machines will be much more sophisticated than what they are today. Indeed, these smart machines will be able to learn on their own—without human intervention—using technology!
Erik Brynjolfsson and Andrew McAfee, co-authors of The Second Machine Age: Work, Progress, and Prosperity in a Time of Brilliant Technologies say, “Technologies that used to seem like science fiction are becoming everyday reality.” Machines that can use such technologies to learn and autonomously adapt to the changes will cause serious disruption to well-established business models. Likewise, Gartner predicts that “by 2017, a significant and disruptive digital business will be launched that was conceived by a computer algorithm.”
All of these predictions are based on one common observation: Machines can learn exceptionally fast, and it is possible to build machines that can learn without any active human supervision. In fact, several of the top 10 emerging technology trends of 2015, as predicted in a recent report by World Economic Forum (WEF), are technologies for learnable machines.
Neuromorphic Technology
With advances in machine learning and rapid processing of high volumes of data, the next generation of powerful computing will be built on chips that will process information by mirroring the brain's architecture. As explained in the WEF report, such computers will be energy efficient and will have much more computing power than a conventional CPU (central processing unit).
Moreover, these computers will be able to “anticipate and learn”—instead of just responding in a pre-programmed fashion. IBM has already demonstrated this technology in a prototype form through their million-neuron TrueNorth chip in August 2014.
Gartner has predicted that by 2017, 10 percent of computers will be learning rather than processing. The report also mentions, that, “The Defense Advanced Research Project Agency (DARPA) and Ecole Polytechnique Federal de Lausanne are funding the SyNAPSE and The Human Brain projects, respectively, fostering neuromorphic computing techniques intended for pattern recognition applications, including facial recognition, object recognition, drug discovery and medical diagnostics.”
Emergent Artificial Intelligence
We are already seeing significant advancements in artificial intelligence through self-driving cars and automated flying drones. Emergent artificial intelligence takes this step further with machines that can learn automatically by using large volumes of data. Carnegie Mellon University's NELL (the Never-Ending Language Learning) computer system "not only reads facts by crawling through hundreds of millions of web pages, but attempts to improve its reading and understanding competence in the process in order to perform better in the future."
Machines with emergent artificial intelligence are not only capable of processing huge amount of information in a short time frame and arrive at an optimal solution, they also can avoid human errors and emotional biases. However, many experts are raising early warnings for potential pitfalls associated with such super-intelligent machines, which will not only replace the jobs of many human workers, but also may eventually “overcome and enslave humans.”
Programmable Material
In case of 3D printing, computer readable designs are fed to a machine that can process the design and produce a finished object using additive manufacturing technology. Medical applications of 3D printing have reached such an advanced stage that even a living tissue or a cell can be printed in this process. Other fields of application of 3D printing include automotive, aerospace, electronics as well as housing and construction.
Advertisement
“While 3D printing is purposed around building static items, 4D printing employs dynamic materials that evolve or adapt to their external environment in real-time and in direct response to changing conditions. This ability to transform is embedded within the material itself. Objects created through 4D printing are assemble and alter their structure when the material of which they are composed comes in contact with different conditions, such as moisture or humidity. The ultimate goal is to create stimuli-responsive components, materials that modify their own form or self-assemble new patterns automatically and in predictive ways,” explains a September 2014 Georgetown Journal of International Affairs article.
Advanced Robotics
Next-generation robots will learn to collaborate with humans to become their co-workers. They will no longer be used only in factory assembly lines and separated from humans by safety cages. The WEF report talks about an example: “In Japan, robots are being trialed in nursing roles: they help patients out of bed and support stroke victims in regaining control of their limbs.”
Boston Dynamics, a wholly owned subsidiary of Google, builds “advanced robots with remarkable behavior: mobility, agility, dexterity and speed.” These robots can walk, run, and climb in a variety of terrains and surfaces by learning to deal with environmental variations. An interesting Tech Times article discusses how researchers at University of Maryland are looking at whether robots can learn a skill, like cooking, by watching YouTube videos.
Apart from the five technologies mentioned above, technologies that are going to impact our lives in a profound manner are Big Data, Analytics and Internet of Things (IoT). Among these, analytics will be a de-facto technology for a learning machines.
Embedded Analytics
According to Gartner’s “Top 10 Strategic Technology Trends for 2015,” analytics will be embedded in every app so that it will continue to learn and develop insights on a continuous basis. Such technology-based learning will lead to “advanced, pervasive, and invisible analytics,” through which the apps will be able to intelligently monitor events, gather data, and learn about the context to take better decisions. By continuously analyzing data and by developing insights, these apps will create deep understanding of the business domains.
Chris Argyris, in his seminal HBR article on organizational learning, “Teaching Smart People How to Learn,” coined the terms single loop and double loop learning. Argyris explained single loop learning with an analogy of a thermostat that automatically turns on the moment the temperature of the room drops below 68 degrees. “A thermostat that could ask, ‘Why am I set at 68 degrees?’ and then explore whether or not some other temperature might more economically achieve the goal of heating the room would be engaging in double-loop learning.”
The article raised the issue of how highly skilled professionals can be very good at single loop learning but miserably fail at double loop learning. With the advancement of technology, we can now have thermostats that are intelligent enough to use double loop learning. Nest Labs, a Google company, specializes in making sensor driven self-learning thermostats and smoke detectors
Bottom line: Machines will no longer remain dependent on their masters for learning. More importantly, like humans, learning machines will be able to use technology to enhance their skills. B. F. Skinner once said, “The real problem is not whether machines think but whether men do.” The time has come to seriously ponder over the implications of this remark.
Source > td.org
No comments:
Post a Comment