AI News, Allen Institute for Artificial Intelligence Reviews
The Center investigates problems in data mining, natural language processing, the Semantic Web and other web search topics.
Etzioni was the first student to major in computer science at Harvard University, where he earned a bachelor's degree in 1986.
Etzioni's research is focused on basic problems in the study of intelligence, machine reading, machine learning and web search.
Past projects include Internet Softbots---the study of intelligent agents in the context of real-world software testbeds.
Etzioni is an entrepreneur who has founded or co-founded several business ventures, including MetaCrawler (bought by Infospace), Netbot (bought by Excite), and ClearForest (bought by Reuters).
He co-founded Decide, a company whose website Decide.com helped consumers make buying decisions using previous price history and recommendations from other users.
Ethics of artificial intelligence
The ethics of artificial intelligence is the part of the ethics of technology specific to robots and other artificially intelligent beings.
divided into roboethics, a concern with the moral behavior of humans as they design, construct, use and treat artificially intelligent beings, and machine ethics, which is concerned with the moral behavior of artificial moral agents (AMAs).
It has been suggested that robot rights, such as a right to exist and perform its own mission, could be linked to robot duty to serve human, by analogy with linking human rights to human duties before society.
Pamela McCorduck counters that, speaking for women and minorities 'I'd rather take my chances with an impartial computer,' pointing out that there are conditions where we would prefer to have automated judges and police that have no personal agenda at all.
However, Kaplan and Haenlein stress that AI systems are only as smart as the data used to train them since they are, in their essence, nothing more than fancy curve-fitting machines: Using AI to support a court ruling can be highly problematic if past rulings show bias toward certain groups since those biases get formalized and engrained, which makes them even more difficult to spot and fight against.
'If any major military power pushes ahead with the AI weapon development, a global arms race is virtually inevitable, and the endpoint of this technological trajectory is obvious: autonomous weapons will become the Kalashnikovs of tomorrow', says the petition, which includes Skype co-founder Jaan Tallinn and MIT professor of linguistics Noam Chomsky as additional supporters against AI weaponry.
Regarding the potential for smarter-than-human systems to be employed militarily, the Open Philanthropy Project writes that these scenarios 'seem potentially as important as the risks related to loss of control', but that research organizations investigating AI's long-run social impact have spent relatively little time on this concern: 'this class of scenarios has not been a major focus for the organizations that have been most active in this space, such as the Machine Intelligence Research Institute (MIRI) and the Future of Humanity Institute (FHI), and there seems to have been less analysis and debate regarding them'.
To account for the nature of these agents, it has been suggested to consider certain philosophical ideas, like the standard characterizations of agency, rational agency, moral agency, and artificial agency, which are related to the concept of AMAs.
In 2009, during an experiment at the Laboratory of Intelligent Systems in the Ecole Polytechnique Fédérale of Lausanne in Switzerland, robots that were programmed to cooperate with each other (in searching out a beneficial resource and avoiding a poisonous one) eventually learned to lie to each other in an attempt to hoard the beneficial resource.
In 2009, academics and technical experts attended a conference organized by the Association for the Advancement of Artificial Intelligence to discuss the potential impact of robots and computers and the impact of the hypothetical possibility that they could become self-sufficient and able to make their own decisions.
They noted that some machines have acquired various forms of semi-autonomy, including being able to find power sources on their own and being able to independently choose targets to attack with weapons.
In a paper on the acquisition of moral values by robots, Nayef Al-Rodhan mentions the case of neuromorphic chips, which aim to process information similarly to humans, nonlinearly and with millions of interconnected artificial neurons.
Inevitably, this raises the question of the environment in which such robots would learn about the world and whose morality they would inherit - or if they end up developing human 'weaknesses' as well: selfishness, a pro-survival attitude, hesitation etc.
Wendell Wallach and Colin Allen conclude that attempts to teach robots right from wrong will likely advance understanding of human ethics by motivating humans to address gaps in modern normative theory and by providing a platform for experimental investigation.
Nick Bostrom and Eliezer Yudkowsky have argued for decision trees (such as ID3) over neural networks and genetic algorithms on the grounds that decision trees obey modern social norms of transparency and predictability (e.g.
while Chris Santos-Lang argued in the opposite direction on the grounds that the norms of any age must be allowed to change and that natural failure to fully satisfy these particular norms has been essential in making humans less vulnerable to criminal 'hackers'.
Many researchers have argued that, by way of an 'intelligence explosion' sometime in the 21st century, a self-improving AI could become so vastly more powerful than humans that we would not be able to stop it from achieving its goals. In
However, instead of overwhelming the human race and leading to our destruction, Bostrom has also asserted that super-intelligence can help us solve many difficult problems such as disease, poverty, and environmental destruction, and could help us to “enhance” ourselves.
Unless moral philosophy provides us with a flawless ethical theory, an AI's utility function could allow for many potentially harmful scenarios that conform with a given ethical framework but not 'common sense'.
Amazon, Google, Facebook, IBM, and Microsoft have established a non-profit partnership to formulate best practices on artificial intelligence technologies, advance the public's understanding, and to serve as a platform about artificial intelligence.
They stated: 'This partnership on AI will conduct research, organize discussions, provide thought leadership, consult with relevant third parties, respond to questions from the public and media, and create educational material that advance the understanding of AI technologies including machine perception, learning, and automated reasoning.'
The same idea can be found in the Emergency Medical Hologram of Starship Voyager, which is an apparently sentient copy of a reduced subset of the consciousness of its creator, Dr. Zimmerman, who, for the best motives, has created the system to give medical assistance in case of emergencies.
This event caused an ethical schism between those who felt bestowing organic rights upon the newly sentient Geth was appropriate and those who continued to see them as disposable machinery and fought to destroy them.
Smart Robots are Still Far From Reach
“Passing even a fourth-grade science test isn’t a single task.
The exams require specific and general knowledge to pass, and Clark can easily check his research’s progress by grading the computer’s performance.
Eventually, rather than eighth-grade science-test diagrams, Etzioni’s team will design algorithms that interpret images, diagrams and text in advanced scientific papers to make new connections and insights, based on its knowledge.
The robot apocalypse presaged in The Terminator might not (and almost certainly won’t) come to pass, but smarter machines aren’t exactly risk-free.
A block away, a man without teeth yells incoherently at four police officers imploring him to stand and put on his shoes.
Is this a glimpse of the future, where more and more people are left behind, replaced by machines that think better and act faster than humanly possible?
A World Economic Forum analysis last year estimated that by 2020, automation and robots will eliminate roughly 5 million jobs in 15 of the world’s developed and emerging economies.
- On 23. januar 2021
Amir Husain: "The Sentient Machine: The Coming Age of Artificial Intelligence" | Talks at Google
The Sentient Machine addresses broad existential questions surrounding the coming of AI: Why are we valuable? What can we create in this world? How are we ...
Artificial Intelligence—Genuine Emotions. How Robots Shape Our Lives
MAK DESIGN SUMMIT Closing event of the exhibition Hello, Robot. Design between Human and Machine in context of the VIENNA BIENNALE 2017: Robots.
Mohit Iyyer: Using Deep Learning to Understand Creative Language
Title: Using Deep Learning to Understand Creative Language Abstract: Creative language—the sort found in novels, film, and comics—contains a wide range of ...
Matt Gardner: Feature Generation from Knowledge Graphs
Matt Gardner: Feature Generation from Knowledge Graphs Abstract: A lot of attention has recently been given to the creation of large knowledge bases that ...
Jay Pujara: Better Knowledge Graphs Through Probabilistic Graphical Models
Jay Pujara Title: Better Knowledge Graphs Through Probabilistic Graphical Models Abstract: Automated question answering, knowledgeable digital assistants, ...
Researchers teach AI to think like a dog and find out what they know about the world
Subscribe to get more product reviews video: -- Researchers teach AI to think like a dog and find out what they know about the world What ..
Kenneth Stanley: Why Greatness Cannot Be Planned: The Myth of the Objective
Presented at TTI/Vanguard's Collaboration and the Workplace of the Future. September 30, 2015 | Washington, DC Why Greatness Cannot Be Planned: The ...
Deliver Business Results by Combining Optimization and Machine Learning
Explore the benefits and use cases of combining machine learning techniques with decision optimization to deliver business results. Learn more: ...
Marcus Olsson | Augmented, Virtual, & Mixed Reality | SingularityU Nordic Summit 2018
Marcus is a Product Expert and a VR-pioneer. He is the CEO and Co-founder of the VR-company SceneThere, and the founder of the strategic advisory firm ...
How can we overcome our fears? How do we reconnect with our intuition? What is the power and magic of using your intention? -To what extent can we use it to ...