In the landscape of today's technology, artificial intelligence is considered to be the prior achievement but Stephen Hawking who is the renowned physicist warned, it might also be the last until and unless we learn to avoid the risks.
Sophisticated technology continues to embed itself in modern society, from digital personal assistants and smart search algorithms in Google to experimental self-driving cars. Artificial intelligence is defined as a computer or machine being able to perform logical deduction and inference and make decisions based on past experience or insufficient or conflicting information. Current tech levels already pose frightening ethical dilemmas, such as whether military drones or other robotic systems should be designed to use lethal force against targets without direct human involvement.
Presently, programmed machines can react with greater speed and accuracy than a human operator, but are still capable of gross error without human biases and intuition to guide decision making. These issues will be intensified if AI is built into such systems, introducing independent motives that may be completely unpredictable or in direct conflict with human intentions.
Transforming the Marketplace
Incorporation of artificial intelligence will drastically shift the humankind's concept of work. Since the Industrial Age, the technology of any kind inevitably renders many jobs obsolete while creating new opportunities in emerging fields. AI offers a future in which machines do tasks in our stead, allowing people to pursue lives of leisure. Economist John Maynard Keynes went so far back in 1930 to predict a work week of 15 hours by 2030. At current population levels, all aspects of society would require a drastic reenvisioning both culturally and economically if automation created such a dramatic shift in how our days are spent.
AI: Friend or Enemy?
One significant threat of artificial intelligence is whether it chooses to use already enhanced abilities to create machines of even greater cognitive power. Advanced generations of AI could operate as high above human beings as we do above animals, and even quite possibly evolve beyond the human ability to understand it.
Protecting Against Disaster
Hawking stresses that any serious discussion of artificial intelligence must take into consideration the potential threats and how to manage them, and calls for more critical, institutional research as increasing corporate resources are devoted to realizing breakthroughs in creating artificial intelligence. Vinge postulates that AI can be tightly regulated by rules that effectively hardwire benevolent behavior into autonomous robots, as imagined by science fiction writer Isaac Asimov's Three Laws of Robotics. Vinge also warns that natural human competition will likely lead to the development of unrestricted models of AI, so even such safeguards may not be enough to control the Singularity if it occurs.