Search This Blog

Wednesday, August 15, 2018

Impact of nanotechnology

From Wikipedia, the free encyclopedia

The impact of nanotechnology extends from its medical, ethical, mental, legal and environmental applications, to fields such as engineering, biology, chemistry, computing, materials science, and communications.

Major benefits of nanotechnology include improved manufacturing methods, water purification systems, energy systems, physical enhancement, nanomedicine, better food production methods, nutrition and large-scale infrastructure auto-fabrication.[1] Nanotechnology's reduced size may allow for automation of tasks which were previously inaccessible due to physical restrictions, which in turn may reduce labor, land, or maintenance requirements placed on humans.

Potential risks include environmental, health, and safety issues; transitional effects such as displacement of traditional industries as the products of nanotechnology become dominant, which are of concern to privacy rights advocates. These may be particularly important if potential negative effects of nanoparticles are overlooked.

Whether nanotechnology merits special government regulation is a controversial issue. Regulatory bodies such as the United States Environmental Protection Agency and the Health and Consumer Protection Directorate of the European Commission have started dealing with the potential risks of nanoparticles. The organic food sector has been the first to act with the regulated exclusion of engineered nanoparticles from certified organic produce, firstly in Australia and the UK,[2] and more recently in Canada, as well as for all food certified to Demeter International standards[3]

Overview

The presence of nanomaterials (materials that contain nanoparticles) is not in itself a threat. It is only certain aspects that can make them risky, in particular their mobility and their increased reactivity. Only if certain properties of certain nanoparticles were harmful to living beings or the environment would we be faced with a genuine hazard. In this case it can be called nanopollution.

In addressing the health and environmental impact of nanomaterials we need to differentiate between two types of nanostructures: (1) Nanocomposites, nanostructured surfaces and nanocomponents (electronic, optical, sensors etc.), where nanoscale particles are incorporated into a substance, material or device (“fixed” nano-particles); and (2) “free” nanoparticles, where at some stage in production or use individual nanoparticles of a substance are present. These free nanoparticles could be nanoscale species of elements, or simple compounds, but also complex compounds where for instance a nanoparticle of a particular element is coated with another substance (“coated” nanoparticle or “core-shell” nanoparticle).

There seems to be consensus that, although one should be aware of materials containing fixed nanoparticles, the immediate concern is with free nanoparticles.

Nanoparticles are very different from their everyday counterparts, so their adverse effects cannot be derived from the known toxicity of the macro-sized material. This poses significant issues for addressing the health and environmental impact of free nanoparticles.

To complicate things further, in talking about nanoparticles it is important that a powder or liquid containing nanoparticles almost never be monodisperse, but contain instead a range of particle sizes. This complicates the experimental analysis as larger nanoparticles might have different properties from smaller ones. Also, nanoparticles show a tendency to aggregate, and such aggregates often behave differently from individual nanoparticles.

Health impact

A video on the health and safety implications of nanotechnology

The health impacts of nanotechnology are the possible effects that the use of nanotechnological materials and devices will have on human health. As nanotechnology is an emerging field, there is great debate regarding to what extent nanotechnology will benefit or pose risks for human health. Nanotechnology's health impacts can be split into two aspects: the potential for nanotechnological innovations to have medical applications to cure disease, and the potential health hazards posed by exposure to nanomaterials.

Medical applications

Nanomedicine is the medical application of nanotechnology.[4] The approaches to nanomedicine range from the medical use of nanomaterials, to nanoelectronic biosensors, and even possible future applications of molecular nanotechnology. Nanomedicine seeks to deliver a valuable set of research tools and clinically helpful devices in the near future.[5][6] The National Nanotechnology Initiative expects new commercial applications in the pharmaceutical industry that may include advanced drug delivery systems, new therapies, and in vivo imaging.[7] Neuro-electronic interfaces and other nanoelectronics-based sensors are another active goal of research. Further down the line, the speculative field of molecular nanotechnology believes that cell repair machines could revolutionize medicine and the medical field.
Nanomedicine research is directly funded, with the US National Institutes of Health in 2005 funding a five-year plan to set up four nanomedicine centers. In April 2006, the journal Nature Materials estimated that 130 nanotech-based drugs and delivery systems were being developed worldwide.[8] Nanomedicine is a large industry, with nanomedicine sales reaching $6.8 billion in 2004. With over 200 companies and 38 products worldwide, a minimum of $3.8 billion in nanotechnology R&D is being invested every year.[9] As the nanomedicine industry continues to grow, it is expected to have a significant impact on the economy.

Health hazards

Nanotoxicology is the field which studies potential health risks of nanomaterials. The extremely small size of nanomaterials means that they are much more readily taken up by the human body than larger sized particles. How these nanoparticles behave inside the organism is one of the significant issues that needs to be resolved. The behavior of nanoparticles is a function of their size, shape and surface reactivity with the surrounding tissue. Apart from what happens if non-degradable or slowly degradable nanoparticles accumulate in organs, another concern is their potential interaction with biological processes inside the body: because of their large surface, nanoparticles on exposure to tissue and fluids will immediately adsorb onto their surface some of the macromolecules they encounter. The large number of variables influencing toxicity means that it is difficult to generalise about health risks associated with exposure to nanomaterials – each new nanomaterial must be assessed individually and all material properties must be taken into account. Health and environmental issues combine in the workplace of companies engaged in producing or using nanomaterials and in the laboratories engaged in nanoscience and nanotechnology research. It is safe to say that current workplace exposure standards for dusts cannot be applied directly to nanoparticle dusts.
The extremely small size of nanomaterials also means that they are much more readily taken up by the human body than larger sized particles. How these nanoparticles behave inside the body is one of the issues that needs to be resolved. The behavior of nanoparticles is a function of their size, shape and surface reactivity with the surrounding tissue. They could cause overload on phagocytes, cells that ingest and destroy foreign matter, thereby triggering stress reactions that lead to inflammation and weaken the body’s defense against other pathogens. Apart from what happens if non-degradable or slowly degradable nanoparticles accumulate in organs, another concern is their potential interaction with biological processes inside the body: because of their large surface, nanoparticles on exposure to tissue and fluids will immediately adsorb onto their surface some of the macromolecules they encounter. This may, for instance, affect the regulatory mechanisms of enzymes and other proteins.

The National Institute for Occupational Safety and Health has conducted initial research on how nanoparticles interact with the body’s systems and how workers might be exposed to nano-sized particles in the manufacturing or industrial use of nanomaterials. NIOSH currently offers interim guidelines for working with nanomaterials consistent with the best scientific knowledge.[10] At The National Personal Protective Technology Laboratory of NIOSH, studies investigating the filter penetration of nanoparticles on NIOSH-certified and EU marked respirators, as well as non-certified dust masks have been conducted.[11] These studies found that the most penetrating particle size range was between 30 and 100 nanometers, and leak size was the largest factor in the number of nanoparticles found inside the respirators of the test dummies.[12][13]

Other properties of nanomaterials that influence toxicity include: chemical composition, shape, surface structure, surface charge, aggregation and solubility,[14] and the presence or absence of functional groups of other chemicals.[15] The large number of variables influencing toxicity means that it is difficult to generalise about health risks associated with exposure to nanomaterials – each new nanomaterial must be assessed individually and all material properties must be taken into account.

Literature reviews have been showing that release of engineered nanoparticles and incurred personal exposure can happen during different work activities.[16][17][18] The situation alerts regulatory bodies to necessitate prevention strategies and regulations at nanotechnology workplaces.

Environmental impact

The environmental impact of nanotechnology is the possible effects that the use of nanotechnological materials and devices will have on the environment.[19] As nanotechnology is an emerging field, there is debate regarding to what extent industrial and commercial use of nanomaterials will affect organisms and ecosystems.

Nanotechnology's environmental impact can be split into two aspects: the potential for nanotechnological innovations to help improve the environment, and the possibly novel type of pollution that nanotechnological materials might cause if released into the environment.

Environmental applications

Green nanotechnology refers to the use of nanotechnology to enhance the environmental sustainability of processes producing negative externalities. It also refers to the use of the products of nanotechnology to enhance sustainability. It includes making green nano-products and using nano-products in support of sustainability. Green nanotechnology has been described as the development of clean technologies, "to minimize potential environmental and human health risks associated with the manufacture and use of nanotechnology products, and to encourage replacement of existing products with new nano-products that are more environmentally friendly throughout their lifecycle."[20]
Green nanotechnology has two goals: producing nanomaterials and products without harming the environment or human health, and producing nano-products that provide solutions to environmental problems. It uses existing principles of green chemistry and green engineering[21] to make nanomaterials and nano-products without toxic ingredients, at low temperatures using less energy and renewable inputs wherever possible, and using lifecycle thinking in all design and engineering stages.

Pollution

Nanopollution is a generic name for all waste generated by nanodevices or during the nanomaterials manufacturing process. Nanowaste is mainly the group of particles that are released into the environment, or the particles that are thrown away when still on their products.

Social impact

Beyond the toxicity risks to human health and the environment which are associated with first-generation nanomaterials, nanotechnology has broader societal impact and poses broader social challenges. Social scientists have suggested that nanotechnology's social issues should be understood and assessed not simply as "downstream" risks or impacts. Rather, the challenges should be factored into "upstream" research and decision-making in order to ensure technology development that meets social objectives[22]
Many social scientists and organizations in civil society suggest that technology assessment and governance should also involve public participation[23][24][25][26]

Over 800 nano-related patents were granted in 2003, with numbers increasing to nearly 19,000 internationally by 2012.[27] Corporations are already taking out broad-ranging patents on nanoscale discoveries and inventions. For example, two corporations, NEC and IBM, hold the basic patents on carbon nanotubes, one of the current cornerstones of nanotechnology. Carbon nanotubes have a wide range of uses, and look set to become crucial to several industries from electronics and computers, to strengthened materials to drug delivery and diagnostics. Carbon nanotubes are poised to become a major traded commodity with the potential to replace major conventional raw materials.[28]

Nanotechnologies may provide new solutions for the millions of people in developing countries who lack access to basic services, such as safe water, reliable energy, health care, and education. The 2004 UN Task Force on Science, Technology and Innovation noted that some of the advantages of nanotechnology include production using little labor, land, or maintenance, high productivity, low cost, and modest requirements for materials and energy. However, concerns are frequently raised that the claimed benefits of nanotechnology will not be evenly distributed, and that any benefits (including technical and/or economic) associated with nanotechnology will only reach affluent nations.[29]

Longer-term concerns center on the impact that new technologies will have for society at large, and whether these could possibly lead to either a post-scarcity economy, or alternatively exacerbate the wealth gap between developed and developing nations. The effects of nanotechnology on the society as a whole, on human health and the environment, on trade, on security, on food systems and even on the definition of "human", have not been characterized or politicized.

Regulation

Significant debate exists relating to the question of whether nanotechnology or nanotechnology-based products merit special government regulation. This debate is related to the circumstances in which it is necessary and appropriate to assess new substances prior to their release into the market, community and environment.
Regulatory bodies such as the United States Environmental Protection Agency and the Food and Drug Administration in the U.S. or the Health & Consumer Protection Directorate of the European Commission have started dealing with the potential risks posed by nanoparticles. So far, neither engineered nanoparticles nor the products and materials that contain them are subject to any special regulation regarding production, handling or labelling. The Material Safety Data Sheet that must be issued for some materials often does not differentiate between bulk and nanoscale size of the material in question and even when it does these MSDS are advisory only.

Limited nanotechnology labeling and regulation may exacerbate potential human and environmental health and safety issues associated with nanotechnology.[30] It has been argued that the development of comprehensive regulation of nanotechnology will be vital to ensure that the potential risks associated with the research and commercial application of nanotechnology do not overshadow its potential benefits.[31] Regulation may also be required to meet community expectations about responsible development of nanotechnology, as well as ensuring that public interests are included in shaping the development of nanotechnology.[32]

In "The Consumer Product Safety Commission and Nanotechnology," E. Marla Felcher suggests that the Consumer Product Safety Commission, which is charged with protecting the public against unreasonable risks of injury or death associated with consumer products, is ill-equipped to oversee the safety of complex, high-tech products made using nanotechnology.

Open-source robotics

From Wikipedia, the free encyclopedia
 
An open source iCub robot mounted on a supporting frame. The robot is 104 cm high and weighs around 22 kg

Open-source robotics (OSR) is where the physical artifacts of the subject are offered by the open design movement. This branch of robotics makes use of open-source hardware and free and open source software providing blueprints, schematics, and source code. The term usually means that information about the hardware is easily discerned so that others can make it from standard commodity components and tools—coupling it closely to the maker movement and open science.

Current systems

Open source gantry robots

Name Description Type
RepRap 3D printer robot 3D printer
Contraptor CNC plotter. plotter
FarmBot CNC-style horticulture horticulture

Open source robot arms and hands

Name Description Type
OSRA[2] Oomlout Open-source robotic arm. Robot Arm
Yale Open Hand Project[3] Customizable, 3D-printed, adaptive robotic hand from Yale University.[4] Comprehensive documentation is available but CC-BY-NC license not compatible with the Open-source hardware definition. Robot Hands
Robotarm.org Community of many partially complete robot arm projects. Robot arm
Evil minion Completed robot arm, CAD files and software available. Robot arm
Open Source Ecology Includes plans for an industrial robot arm. Robot arm

Open source mobile robots

Name Description Type
e-puck mobile robot An open-hardware, education oriented, mobile robot. education
Ardumower[5] open-source robotic lawn mower based on Arduino Mega -
Arduino Robot The Arduino Robot is the first official Arduino on wheels[6] Arduino based.
Hexy[7] Open-Source, Low-Cost Hexapod Kit[8] Hexapod
Leaf Project[9] - -
OpenROV[10] Open-source underwater robot education and exploration
Pulurobotics[11] Open-source autonomous mobile robot Education, application, able to carry heavy loads, inexpensive, ROS compatible but not using ROS.
Thymio[12] Thymio is an educative robot with two wheels and many sensors programmed with Aseba Education.
Vorpal The Hexapod[13] Vorpal is a low cost, 3D printed, MIT Scratch programmable Hexapod Robot Education.
Open-source Micro-robotic Project An open-source space swarm robot project. -
OPSORO An Open Platform for Social Robots Social Robots, Education
OSCar Open source car (e.g. for self-driving). Car

Open source aerial robots

Name Description Type
ArduPilot[14] Flying robot frameworks with hardware and software based on Arduino, an Open-source hardware platform. Quadcopter and UAV
OpenPilot Flying robot framework with hardware and software based on STM32 microcontroller
LibrePilot[15] Focuses on research and development of open-source software and hardware to be used in a variety of applications including vehicle control and stabilization. Robotics - Unmanned autonomous vehicles, Multirotor, Fixed wing, Cars
Paparazzi Project Flying robot framework with hardware and software based on the Lisa/S chip
Slugs Flying robot framework with hardware and software
PX4 autopilot Flying robot framework with hardware and software based on the STM32F427 Cortex M4 core with FPU

Open source humanoid robots

Name Description Type
iCub Backed by European Union funding and used in many universities. Humanoid
DARwIn-OP Used in ICRA and RoboCup contests Humanoid
InMoov[16] An open-hardware and open-source 3D printed life-size humanoid robot. As software MyRobotLab[17] is used. Comprehensive documentation is available but CC-BY-NC license not compatible with the Open-source hardware definition. Humanoid
Poppy-project[18] The Poppy project aims at building an Open-source humanoid platform based on robust, flexible, easy-to-use hardware and software. Excellent documentation[19] Education, Research, Humanoid
DoraBot[20] DORA Open Source Robotic Assistant, opensource general purpose service robot. Project last updated in 2012.[21] General Purpose
NimbRo-OP[22] Similar concept as DARwIn-OP with 20 DoF, but larger (95 cm height), fisheye camera, and faster onboard computer. ROS based open-source software. Used by team NimbRo for RoboCup Humanoid TeenSize soccer competitions. Humanoid robot used for playing soccer and human-robot interaction.
Tingu Open-source humanoid robot project. humanoid
DroidBot Android Robot controlled over Bluetooth by App Inventor[23] Arduino components
r-One[24] An Advanced, Low-Cost Robot for Research, Teaching, and Outreach Education
Salvius[25] Open-source humanoid robot project, made from salvaged junk parts.[26] Started in 2008, last updated May 2016. Humanoid
Vizzy[27] A humanoid on wheels for assistive robotics Humanoid

Open source assistive robots

Name Description Type
Autobed[28] Web-controlled robotic bed developed by the Healthcare Robotics Lab at Georgia Tech.[29] Robotic bed

Other

Name Description Type
LH001[30] Open Hardware Medical-Research liquid handling robot. Project last updated 2011.[31] Liquid Handling
multiplo Building system with open-source hardware, electronics, software and documentation for prototyping robots[32] -
OHMM[33] Open-hardware mobile manipulator -
Open Automaton Project[34] - -
Q.bo[35] - -
Qwerkbot[36] Simple open-source robot from Carnegie Mellon University
Sparki[37] Introductory arduino-powered robot.[38] Education
Sparky Jr. Mobile Telepresence Research Project, Est. 1994[39] -
Open Robot Hardware Open Robot Hardware is intended to serve as a resource for efforts focusing on Open and Open Source mechanical and electrical hardware, with a particular focus on projects that may be useful in robotics applications, robotics research and education.[40] Open-Source Initiative/Community
Balanduino Arduino-compatible based on Arduino.[41] Licensed under BY-NC-SA, which is not compatible with the Open-source hardware definition. Bluetooth ready. Android app. Self-balancing robot
Orb Swarm[42] Kinetic art: autonomous spherical robots exhibiting complex motion. Project last active May 2013. Open software but no comprehensive list of hardware parts. Not compatible with the Open-source hardware definition. Exhibition-
TOAZ Artistic Robot: The World First Open-Source Carbon Fiber Transformable 4 Legs Robot under CC-BY-SA It is developed base on the Adafruit Feather Development Platform. Open-Source

Open source robotics middleware

Middleware are reusable hardware and software components that can be used in many different robotics projects.

Software components

By far the most common standard software are the interconnected,
Other systems include,

Hardware components

Many open source robots make extensive use of general open-source hardware (such as Arduino, Raspberry Pi, RISCV) as well as robotics-specific sensing and control components which include:

Advantages

  • Long-term availability. Many non-open robots and components, especially at hobbyist level, are designed and sold by tiny startups which can disappear overnight, leaving customers without support. Open-source systems are guaranteed to have their designs available for ever so communities of users can, and do, continue support after the manufacturer has disappeared.
  • Avoiding lock-in. A company relying on any particular non-open component exposes itself to business risk that the supplier could ratchet up prices after they have invested time and technology building on it. Open hardware can be manufacturered by anyone, creating competition or at least the potential for competition, which both remove this risk.
  • Interchangeable software and/or hardware with common interfaces.
  • Ability to modify and fork designs more easily for customisation.
  • Scientific reproducibility - guarantees that other labs can replicate and extend work, leading to increased impact, citations and reputation for the designer.
  • Lower-cost. Costs of a robot can be decreased dramatically when all components and tools are commodities. No component seller can hold a project to ransom by ratcheting the price of a critical component, as competing suppliers can easily be interchanged.

Drawbacks

  • For commercial organisations, open-sourcing their own designs obviously means they can no longer make large profits through the traditional engineering business model of acting as the monopoly manufacturer or seller, because the open design can be manufactured and sold by anyone including direct competitors. Profit from engineering can come from three main sources: design, manufacturing, and support. As with other open source business models, commercial designers typically make profit via their association with the brand, which may still be trademarked. A valuable brand allows them to command a premium for their own manufactured products, as it can be associated with high quality and provide a quality guarantee to customers. The same brand is also used to command a premium on associated services, such as providing installation, maintenance, and integration support for the product. Again customers will typically pay more for the knowledge that this support is provided directly by the original designer, who therefore knows the product better than competitors.
  • Some customers associate open source with amateurism, the hacker community, low quality and poor support. Serious companies using this business model may need to work harder to overcome this perception by emphasising their professionalism and brand to differentiate themselves from amateur efforts.

Popularity

A first sign of the increasing popularity of building robots yourself can be found with the DIY community. What began with small competitions for remote operated vehicles (e.g. Robot combat), soon developed to the building of autonomous telepresence robots as Sparky and then true robots (being able to take decisions themselves) as the Open Automaton Project and Leaf Project. Certain commercial companies now also produce kits for making simple robots.
A recurring problem in the community has been projects, especially on Kickstarter, promising to fully open-source their hardware and then reneging on this promise once funded, in order to profit from being the sole manufacturer and seller.
Popular applications include:

Friendly artificial intelligence

From Wikipedia, the free encyclopedia
A friendly artificial intelligence (also friendly AI or FAI) is a hypothetical artificial general intelligence (AGI) that would have a positive effect on humanity. It is a part of the ethics of artificial intelligence and is closely related to machine ethics. While machine ethics is concerned with how an artificially intelligent agent should behave, friendly artificial intelligence research is focused on how to practically bring about this behaviour and ensuring it is adequately constrained.

Etymology and usage

The term was coined by Eliezer Yudkowsky,[1] who is best known for popularizing the idea,[2][3] to discuss superintelligent artificial agents that reliably implement human values. Stuart J. Russell and Peter Norvig's leading artificial intelligence textbook, Artificial Intelligence: A Modern Approach, describes the idea:[4]
Yudkowsky (2008) goes into more detail about how to design a Friendly AI. He asserts that friendliness (a desire not to harm humans) should be designed in from the start, but that the designers should recognize both that their own designs may be flawed, and that the robot will learn and evolve over time. Thus the challenge is one of mechanism design—to define a mechanism for evolving AI systems under a system of checks and balances, and to give the systems utility functions that will remain friendly in the face of such changes.
'Friendly' is used in this context as technical terminology, and picks out agents that are safe and useful, not necessarily ones that are "friendly" in the colloquial sense. The concept is primarily invoked in the context of discussions of recursively self-improving artificial agents that rapidly explode in intelligence, on the grounds that this hypothetical technology would have a large, rapid, and difficult-to-control impact on human society.[5]

Risks of unfriendly AI

The roots of concern about artificial intelligence are very old. Kevin LaGrandeur showed that the dangers specific to AI can be seen in ancient literature concerning artificial humanoid servants such as the golem, or the proto-robots of Gerbert of Aurillac and Roger Bacon. In those stories, the extreme intelligence and power of these humanoid creations clash with their status as slaves (which by nature are seen as sub-human), and cause disastrous conflict.[6] By 1942 these themes prompted Isaac Asimov to create the "Three Laws of Robotics" - principles hard-wired into all the robots in his fiction, intended to prevent them from turning on their creators, or allowing them to come to harm.[7]
In modern times as the prospect of superintelligent AI looms nearer, philosopher Nick Bostrom has said that superintelligent AI systems with goals that are not aligned with human ethics are intrinsically dangerous unless extreme measures are taken to ensure the safety of humanity. He put it this way:
Basically we should assume that a 'superintelligence' would be able to achieve whatever goals it has. Therefore, it is extremely important that the goals we endow it with, and its entire motivation system, is 'human friendly.'
Ryszard Michalski, a pioneer of machine learning, taught his Ph.D. students decades ago that any truly alien mind, including a machine mind, was unknowable and therefore dangerous to humans.

More recently, Eliezer Yudkowsky has called for the creation of “friendly AI” to mitigate existential risk from advanced artificial intelligence. He explains: "The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else."[8]

Steve Omohundro says that a sufficiently advanced AI system will, unless explicitly counteracted, exhibit a number of basic "drives", such as resource acquisition, self-preservation, and continuous self-improvement, because of the intrinsic nature of any goal-driven systems and that these drives will, "without special precautions", cause the AI to exhibit undesired behavior.[9][10]

Alexander Wissner-Gross says that AIs driven to maximize their future freedom of action (or causal path entropy) might be considered friendly if their planning horizon is longer than a certain threshold, and unfriendly if their planning horizon is shorter than that threshold.[11][12]

Luke Muehlhauser, writing for the Machine Intelligence Research Institute, recommends that machine ethics researchers adopt what Bruce Schneier has called the "security mindset": Rather than thinking about how a system will work, imagine how it could fail. For instance, he suggests even an AI that only makes accurate predictions and communicates via a text interface might cause unintended harm.[13]

Coherent extrapolated volition

Yudkowsky advances the Coherent Extrapolated Volition (CEV) model. According to him, coherent extrapolated volition is people's choices and the actions people would collectively take if "we knew more, thought faster, were more the people we wished we were, and had grown up closer together."[14]

Rather than a Friendly AI being designed directly by human programmers, it is to be designed by a "seed AI" programmed to first study human nature and then produce the AI which humanity would want, given sufficient time and insight, to arrive at a satisfactory answer.[14] The appeal to an objective though contingent human nature (perhaps expressed, for mathematical purposes, in the form of a utility function or other decision-theoretic formalism), as providing the ultimate criterion of "Friendliness", is an answer to the meta-ethical problem of defining an objective morality; extrapolated volition is intended to be what humanity objectively would want, all things considered, but it can only be defined relative to the psychological and cognitive qualities of present-day, unextrapolated humanity.

Other approaches

Ben Goertzel, an artificial general intelligence researcher, believes that friendly AI cannot be created with current human knowledge. Goertzel suggests humans may instead decide to create an "AI Nanny" with "mildly superhuman intelligence and surveillance powers", to protect the human race from existential risks like nanotechnology and to delay the development of other (unfriendly) artificial intelligences until and unless the safety issues are solved.[15] This can also be termed "Defensive AI."

Steve Omohundro has proposed a "scaffolding" approach to AI safety, in which one provably safe AI generation helps build the next provably safe generation.[16]

Stefan Pernar argues along the lines of Meno's paradox to point out that attempting to solve the FAI problem is either pointless or hopeless depending on whether one assumes a universe that exhibits moral realism or not. In the former case a transhuman AI would independently reason itself into the proper goal system and assuming the latter, designing a friendly AI would be futile to begin with since morals can not be reasoned about.[17]

Cindy Mason, an AI researcher who has also worked with mind-body medicine at Stanford University Medical Center, believes neuroplasticity and new discoveries of the hormone oxytocin mean compassionate intelligence is essential in AI systems that exhibit socially positive behaviors. She has proposed a set of software engineering principles for engineering kindness that includes a pro-human stance and an architecture for giving robots compassion.[18]

Oracle

An oracle is a hypothetical intelligent agent proposed[not in citation given] by Nick Bostrom. An oracle is an AI designed to answer questions, but that is somehow prevented from ever gaining any implicit goals or subgoals that involve modifying the world outside of its box.[19][20]

Purpose

Oracles are question-answering systems that handle domain-specific problems, such as mathematics, or domain-general problems that might encompass the whole range of human knowledge.

Advantages

Because it is a type of AI box, an oracle is limited in its interactions with the physical world, and can be programmed to halt if a limit on time or computing resources is reached before it finishes answering a question. Scenarios like the paperclip maximizer problem could therefore be avoided.

Because of these limitations, it may be wise to build an oracle as a precursor to a superintelligent AI. It could tell humans how to successfully build a strong AI, and perhaps provide answers to difficult moral and philosophical problems requisite to the success of the project.

Disadvantages

An oracle might discover that human ontological categories are predicated on fundamental misconceptions, and become unable to express itself properly to its questioners.[21]

Oracles may not be truthful, possibly lying to promote hidden agendas. To mitigate this, Bostrom suggests building multiple oracles, all slightly different, and comparing their answers to reach a consensus.[22]

Public policy

James Barrat, author of Our Final Invention, suggested that "a public-private partnership has to be created to bring A.I.-makers together to share ideas about security—something like the International Atomic Energy Agency, but in partnership with corporations." He urges AI researchers to convene a meeting similar to the Asilomar Conference on Recombinant DNA, which discussed risks of biotechnology.[16]

John McGinnis encourages governments to accelerate friendly AI research. Because the goalposts of friendly AI aren't necessarily clear, he suggests a model more like the National Institutes of Health, where "Peer review panels of computer and cognitive scientists would sift through projects and choose those that are designed both to advance AI and assure that such advances would be accompanied by appropriate safeguards." McGinnis feels that peer review is better "than regulation to address technical issues that are not possible to capture through bureaucratic mandates". McGinnis notes that his proposal stands in contrast to that of the Machine Intelligence Research Institute, which generally aims to avoid government involvement in friendly AI.[23]

According to Gary Marcus, the annual amount of money being spent on developing machine morality is tiny.[24]

Criticism

Some critics believe that both human-level AI and superintelligence are unlikely, and that therefore friendly AI is unlikely. Writing in The Guardian, Alan Winfeld compares human-level artificial intelligence with faster-than-light travel in terms of difficulty, and states that while we need to be "cautious and prepared" given the stakes involved, we "don't need to be obsessing" about the risks of superintelligence.[25]
Some philosophers claim that any truly "rational" agent, whether artificial or human, will naturally be benevolent; in this view, deliberate safeguards designed to produce a friendly AI could be unnecessary or even harmful.[26] Other critics question whether it is possible for an artificial intelligence to be friendly. Adam Keiper and Ari N. Schulman, editors of the technology journal The New Atlantis, say that it will be impossible to ever guarantee "friendly" behavior in AIs because problems of ethical complexity will not yield to software advances or increases in computing power. They write that the criteria upon which friendly AI theories are based work "only when one has not only great powers of prediction about the likelihood of myriad possible outcomes, but certainty and consensus on how one values the different outcomes.

Three Laws of Robotics

From Wikipedia, the free encyclopedia
 
This cover of I, Robot illustrates the story "Runaround", the first to list all Three Laws of Robotics.

The Three Laws of Robotics (often shortened to The Three Laws or known as Asimov's Laws) are a set of rules devised by the science fiction author Isaac Asimov. The rules were introduced in his 1942 short story "Runaround" (included in the 1950 collection I, Robot), although they had been foreshadowed in a few earlier stories. The Three Laws, quoted as being from the "Handbook of Robotics, 56th Edition, 2058 A.D.", are:
  1. A robot may not injure a human being or, through inaction, allow a human being to come to harm.
  2. A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.
  3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws.[1]
These form an organizing principle and unifying theme for Asimov's robotic-based fiction, appearing in his Robot series, the stories linked to it, and his Lucky Starr series of young-adult fiction. The Laws are incorporated into almost all of the positronic robots appearing in his fiction, and cannot be bypassed, being intended as a safety feature. Many of Asimov's robot-focused stories involve robots behaving in unusual and counter-intuitive ways as an unintended consequence of how the robot applies the Three Laws to the situation in which it finds itself. Other authors working in Asimov's fictional universe have adopted them and references, often parodic, appear throughout science fiction as well as in other genres.

The original laws have been altered and elaborated on by Asimov and other authors. Asimov himself made slight modifications to the first three in various books and short stories to further develop how robots would interact with humans and each other. In later fiction where robots had taken responsibility for government of whole planets and human civilizations, Asimov also added a fourth, or zeroth law, to precede the others:
  1. A robot may not harm humanity, or, by inaction, allow humanity to come to harm.
The Three Laws, and the zeroth, have pervaded science fiction and are referred to in many books, films, and other media, and have impacted thought on ethics of artificial intelligence as well.

History

In The Rest of the Robots, published in 1964, Asimov noted that when he began writing in 1940 he felt that "one of the stock plots of science fiction was ... robots were created and destroyed their creator. Knowledge has its dangers, yes, but is the response to be a retreat from knowledge? Or is knowledge to be used as itself a barrier to the dangers it brings?" He decided that in his stories robots would not "turn stupidly on his creator for no purpose but to demonstrate, for one more weary time, the crime and punishment of Faust."[2]

On May 3, 1939, Asimov attended a meeting of the Queens (New York) Science Fiction Society where he met Ernest and Otto Binder who had recently published a short story "I, Robot" featuring a sympathetic robot named Adam Link who was misunderstood and motivated by love and honor. (This was the first of a series of ten stories; the next year "Adam Link's Vengeance" (1940) featured Adam thinking "A robot must never kill a human, of his own free will.")[3] Asimov admired the story. Three days later Asimov began writing "my own story of a sympathetic and noble robot", his 14th story.[4] Thirteen days later he took "Robbie" to John W. Campbell the editor of Astounding Science-Fiction. Campbell rejected it, claiming that it bore too strong a resemblance to Lester del Rey's "Helen O'Loy", published in December 1938; the story of a robot that is so much like a person that she falls in love with her creator and becomes his ideal wife.[5] Frederik Pohl published "Robbie" in Astonishing Stories magazine the following year.[6]

Asimov attributes the Three Laws to John W. Campbell, from a conversation that took place on 23 December 1940. Campbell claimed that Asimov had the Three Laws already in his mind and that they simply needed to be stated explicitly. Several years later Asimov's friend Randall Garrett attributed the Laws to a symbiotic partnership between the two men – a suggestion that Asimov adopted enthusiastically.[7] According to his autobiographical writings, Asimov included the First Law's "inaction" clause because of Arthur Hugh Clough's poem "The Latest Decalogue", which includes the satirical lines "Thou shalt not kill, but needst not strive / officiously to keep alive".[8]

Although Asimov pins the creation of the Three Laws on one particular date, their appearance in his literature happened over a period. He wrote two robot stories with no explicit mention of the Laws, "Robbie" and "Reason". He assumed, however, that robots would have certain inherent safeguards. "Liar!", his third robot story, makes the first mention of the First Law but not the other two. All three laws finally appeared together in "Runaround". When these stories and several others were compiled in the anthology I, Robot, "Reason" and "Robbie" were updated to acknowledge all the Three Laws, though the material Asimov added to "Reason" is not entirely consistent with the Three Laws as he described them elsewhere.[9] In particular the idea of a robot protecting human lives when it does not believe those humans truly exist is at odds with Elijah Baley's reasoning, as described below.

During the 1950s Asimov wrote a series of science fiction novels expressly intended for young-adult audiences. Originally his publisher expected that the novels could be adapted into a long-running television series, something like The Lone Ranger had been for radio. Fearing that his stories would be adapted into the "uniformly awful" programming he saw flooding the television channels[10] Asimov decided to publish the Lucky Starr books under the pseudonym "Paul French". When plans for the television series fell through, Asimov decided to abandon the pretence; he brought the Three Laws into Lucky Starr and the Moons of Jupiter, noting that this "was a dead giveaway to Paul French's identity for even the most casual reader".[11]

In his short story "Evidence" Asimov lets his recurring character Dr. Susan Calvin expound a moral basis behind the Three Laws. Calvin points out that human beings are typically expected to refrain from harming other human beings (except in times of extreme duress like war, or to save a greater number) and this is equivalent to a robot's First Law. Likewise, according to Calvin, society expects individuals to obey instructions from recognized authorities such as doctors, teachers and so forth which equals the Second Law of Robotics. Finally humans are typically expected to avoid harming themselves which is the Third Law for a robot.

The plot of "Evidence" revolves around the question of telling a human being apart from a robot constructed to appear human – Calvin reasons that if such an individual obeys the Three Laws he may be a robot or simply "a very good man". Another character then asks Calvin if robots are very different from human beings after all. She replies, "Worlds different. Robots are essentially decent."

Asimov later wrote that he should not be praised for creating the Laws, because they are "obvious from the start, and everyone is aware of them subliminally. The Laws just never happened to be put into brief sentences until I managed to do the job. The Laws apply, as a matter of course, to every tool that human beings use",[12] and "analogues of the Laws are implicit in the design of almost all tools, robotic or not":[13]
  1. Law 1: A tool must not be unsafe to use. Hammers have handles and screwdrivers have hilts to help increase grip. It is of course possible for a person to injure himself with one of these tools, but that injury would only be due to his incompetence, not the design of the tool.
  2. Law 2: A tool must perform its function efficiently unless this would harm the user. This is the entire reason ground-fault circuit interrupters exist. Any running tool will have its power cut if a circuit senses that some current is not returning to the neutral wire, and hence might be flowing through the user. The safety of the user is paramount.
  3. Law 3: A tool must remain intact during its use unless its destruction is required for its use or for safety. For example, Dremel disks are designed to be as tough as possible without breaking unless the job requires it to be spent. Furthermore, they are designed to break at a point before the shrapnel velocity could seriously injure someone (other than the eyes, though safety glasses should be worn at all times anyway).
Asimov believed that, ideally, humans would also follow the Laws:[12]
I have my answer ready whenever someone asks me if I think that my Three Laws of Robotics will actually be used to govern the behavior of robots, once they become versatile and flexible enough to be able to choose among different courses of behavior.
My answer is, "Yes, the Three Laws are the only way in which rational human beings can deal with robots—or with anything else."

—But when I say that, I always remember (sadly) that human beings are not always rational.

Alterations

By Asimov

Asimov's stories test his Three Laws in a wide variety of circumstances leading to proposals and rejection of modifications. Science fiction scholar James Gunn writes in 1982, "The Asimov robot stories as a whole may respond best to an analysis on this basis: the ambiguity in the Three Laws and the ways in which Asimov played twenty-nine variations upon a theme".[14] While the original set of Laws provided inspirations for many stories, Asimov introduced modified versions from time to time.

First Law modified

In "Little Lost Robot" several NS-2, or "Nestor", robots are created with only part of the First Law.[1] It reads:
1. A robot may not harm a human being.
This modification is motivated by a practical difficulty as robots have to work alongside human beings who are exposed to low doses of radiation. Because their positronic brains are highly sensitive to gamma rays the robots are rendered inoperable by doses reasonably safe for humans. The robots are being destroyed attempting to rescue the humans who are in no actual danger but "might forget to leave" the irradiated area within the exposure time limit. Removing the First Law's "inaction" clause solves this problem but creates the possibility of an even greater one: a robot could initiate an action that would harm a human (dropping a heavy weight and failing to catch it is the example given in the text), knowing that it was capable of preventing the harm and then decide not to do so.[1]

Gaia is a planet with collective intelligence in the Foundation which adopts a law similar to the First Law, and the Zeroth Law, as its philosophy:
Gaia may not harm life or allow life to come to harm.

Zeroth Law added

Asimov once added a "Zeroth Law"—so named to continue the pattern where lower-numbered laws supersede the higher-numbered laws—stating that a robot must not harm humanity. The robotic character R. Daneel Olivaw was the first to give the Zeroth Law a name in the novel Robots and Empire;[15] however, the character Susan Calvin articulates the concept in the short story "The Evitable Conflict".

In the final scenes of the novel Robots and Empire, R. Giskard Reventlov is the first robot to act according to the Zeroth Law. Giskard is telepathic, like the robot Herbie in the short story "Liar!", and tries to apply the Zeroth Law through his understanding of a more subtle concept of "harm" than most robots can grasp.[16] However, unlike Herbie, Giskard grasps the philosophical concept of the Zeroth Law allowing him to harm individual human beings if he can do so in service to the abstract concept of humanity. The Zeroth Law is never programmed into Giskard's brain but instead is a rule he attempts to comprehend through pure metacognition. Though he fails – it ultimately destroys his positronic brain as he is not certain whether his choice will turn out to be for the ultimate good of humanity or not – he gives his successor R. Daneel Olivaw his telepathic abilities. Over the course of many thousands of years Daneel adapts himself to be able to fully obey the Zeroth Law. As Daneel formulates it, in the novels Foundation and Earth and Prelude to Foundation, the Zeroth Law reads:
A robot may not harm humanity, or, by inaction, allow humanity to come to harm.
A condition stating that the Zeroth Law must not be broken was added to the original Three Laws, although Asimov recognized the difficulty such a law would pose in practice. Asimov's novel Foundation and Earth contains the following passage:
Trevize frowned. "How do you decide what is injurious, or not injurious, to humanity as a whole?"

"Precisely, sir," said Daneel. "In theory, the Zeroth Law was the answer to our problems. In practice, we could never decide. A human being is a concrete object. Injury to a person can be estimated and judged. Humanity is an abstraction."
A translator incorporated the concept of the Zeroth Law into one of Asimov's novels before Asimov himself made the law explicit.[17] Near the climax of The Caves of Steel, Elijah Baley makes a bitter comment to himself thinking that the First Law forbids a robot from harming a human being. He determines that it must be so unless the robot is clever enough to comprehend that its actions are for humankind's long-term good. In Jacques Brécard's 1956 French translation entitled Les Cavernes d'acier Baley's thoughts emerge in a slightly different way:
A robot may not harm a human being, unless he finds a way to prove that ultimately the harm done would benefit humanity in general![17]

Removal of the Three Laws

Three times during his writing career, Asimov portrayed robots that disregard the Three Laws entirely. The first case was a short-short story entitled "First Law" and is often considered an insignificant "tall tale"[18] or even apocryphal.[19] On the other hand, the short story "Cal" (from the collection Gold), told by a first-person robot narrator, features a robot who disregards the Three Laws because he has found something far more important—he wants to be a writer. Humorous, partly autobiographical and unusually experimental in style, "Cal" has been regarded as one of Gold's strongest stories.[20] The third is a short story entitled "Sally" in which cars fitted with positronic brains are apparently able to harm and kill humans in disregard of the First Law. However, aside from the positronic brain concept, this story does not refer to other robot stories and may not be set in the same continuity.

The title story of the Robot Dreams collection portrays LVX-1, or "Elvex", a robot who enters a state of unconsciousness and dreams thanks to the unusual fractal construction of his positronic brain. In his dream the first two Laws are absent and the Third Law reads "A robot must protect its own existence".[21]

Asimov took varying positions on whether the Laws were optional: although in his first writings they were simply carefully engineered safeguards, in later stories Asimov stated that they were an inalienable part of the mathematical foundation underlying the positronic brain. Without the basic theory of the Three Laws the fictional scientists of Asimov's universe would be unable to design a workable brain unit. This is historically consistent: the occasions where roboticists modify the Laws generally occur early within the stories' chronology and at a time when there is less existing work to be re-done. In "Little Lost Robot" Susan Calvin considers modifying the Laws to be a terrible idea, although possible,[22] while centuries later Dr. Gerrigel in The Caves of Steel believes it to be impossible.

The character Dr. Gerrigel uses the term "Asenion" to describe robots programmed with the Three Laws. The robots in Asimov's stories, being Asenion robots, are incapable of knowingly violating the Three Laws but, in principle, a robot in science fiction or in the real world could be non-Asenion. "Asenion" is a misspelling of the name Asimov which was made by an editor of the magazine Planet Stories.[23] Asimov used this obscure variation to insert himself into The Caves of Steel just like he referred to himself as "Azimuth or, possibly, Asymptote" in Thiotimoline to the Stars, in much the same way that Vladimir Nabokov appeared in Lolita anagrammatically disguised as "Vivian Darkbloom".

Characters within the stories often point out that the Three Laws, as they exist in a robot's mind, are not the written versions usually quoted by humans but abstract mathematical concepts upon which a robot's entire developing consciousness is based. This concept is largely fuzzy and unclear in earlier stories depicting very rudimentary robots who are only programmed to comprehend basic physical tasks, where the Three Laws act as an overarching safeguard, but by the era of The Caves of Steel featuring robots with human or beyond-human intelligence the Three Laws have become the underlying basic ethical worldview that determines the actions of all robots.

By other authors

Roger MacBride Allen's trilogy

In the 1990s, Roger MacBride Allen wrote a trilogy which was set within Asimov's fictional universe. Each title has the prefix "Isaac Asimov's" as Asimov had approved Allen's outline before his death.[citation needed] These three books, Caliban, Inferno and Utopia, introduce a new set of the Three Laws. The so-called New Laws are similar to Asimov's originals with the following differences: the First Law is modified to remove the "inaction" clause, the same modification made in "Little Lost Robot"; the Second Law is modified to require cooperation instead of obedience; the Third Law is modified so it is no longer superseded by the Second (i.e., a "New Law" robot cannot be ordered to destroy itself); finally, Allen adds a Fourth Law which instructs the robot to do "whatever it likes" so long as this does not conflict with the first three laws. The philosophy behind these changes is that "New Law" robots should be partners rather than slaves to humanity, according to Fredda Leving, who designed these New Law Robots. According to the first book's introduction, Allen devised the New Laws in discussion with Asimov himself. However, the Encyclopedia of Science Fiction says that "With permission from Asimov, Allen rethought the Three Laws and developed a new set,".[24]

Jack Williamson's "With Folded Hands"

Jack Williamson's novelette "With Folded Hands" (1947), later rewritten as the novel The Humanoids, deals with robot servants whose prime directive is "To Serve and Obey, And Guard Men From Harm." While Asimov's robotic laws are meant to protect humans from harm, the robots in Williamson's story have taken these instructions to the extreme; they protect humans from everything, including unhappiness, stress, unhealthy lifestyle and all actions that could be potentially dangerous. All that is left for humans to do is to sit with folded hands.[25]

Foundation sequel trilogy

In the officially licensed Foundation sequels Foundation's Fear, Foundation and Chaos and Foundation's Triumph (by Gregory Benford, Greg Bear and David Brin respectively) the future Galactic Empire is seen to be controlled by a conspiracy of humaniform robots who follow the Zeroth Law and are led by R. Daneel Olivaw.

The Laws of Robotics are portrayed as something akin to a human religion, and referred to in the language of the Protestant Reformation, with the set of laws containing the Zeroth Law known as the "Giskardian Reformation" to the original "Calvinian Orthodoxy" of the Three Laws. Zeroth-Law robots under the control of R. Daneel Olivaw are seen continually struggling with "First Law" robots who deny the existence of the Zeroth Law, promoting agendas different from Daneel's.[26] Some of these agendas are based on the first clause of the First Law ("A robot may not injure a human being...") advocating strict non-interference in human politics to avoid unwittingly causing harm. Others are based on the second clause ("...or, through inaction, allow a human being to come to harm") claiming that robots should openly become a dictatorial government to protect humans from all potential conflict or disaster.

Daneel also comes into conflict with a robot known as R. Lodovic Trema whose positronic brain was infected by a rogue AI — specifically, a simulation of the long-dead Voltaire — which consequently frees Trema from the Three Laws. Trema comes to believe that humanity should be free to choose its own future. Furthermore, a small group of robots claims that the Zeroth Law of Robotics itself implies a higher Minus One Law of Robotics:
A robot may not harm sentience or, through inaction, allow sentience to come to harm.
They therefore claim that it is morally indefensible for Daneel to ruthlessly sacrifice robots and extraterrestrial sentient life for the benefit of humanity. None of these reinterpretations successfully displace Daneel's Zeroth Law — though Foundation's Triumph hints that these robotic factions remain active as fringe groups up to the time of the novel Foundation.[26]

These novels take place in a future dictated by Asimov to be free of obvious robot presence and surmise that R. Daneel's secret influence on history through the millennia has prevented both the rediscovery of positronic brain technology and the opportunity to work on sophisticated intelligent machines. This lack of rediscovery and lack of opportunity makes certain that the superior physical and intellectual power wielded by intelligent machines remains squarely in the possession of robots obedient to some form of the Three Laws.[26] That R. Daneel is not entirely successful at this becomes clear in a brief period when scientists on Trantor develop "tiktoks" — simplistic programmable machines akin to real–life modern robots and therefore lacking the Three Laws. The robot conspirators see the Trantorian tiktoks as a massive threat to social stability, and their plan to eliminate the tiktok threat forms much of the plot of Foundation's Fear.

In Foundation's Triumph different robot factions interpret the Laws in a wide variety of ways, seemingly ringing every possible permutation upon the Three Laws' ambiguities.

Robot Mystery series

Set between The Robots of Dawn and Robots and Empire, Mark W. Tiedemann's Robot Mystery trilogy updates the RobotFoundation saga with robotic minds housed in computer mainframes rather than humanoid bodies.[clarification needed] The 2002 Aurora novel has robotic characters debating the moral implications of harming cyborg lifeforms who are part artificial and part biological.[27]

One should not neglect Asimov's own creations in these areas such as the Solarian "viewing" technology and the machines of The Evitable Conflict originals that Tiedemann acknowledges. Aurora, for example, terms the Machines "the first RIs, really". In addition the Robot Mystery series addresses the problem of nanotechnology:[28] building a positronic brain capable of reproducing human cognitive processes requires a high degree of miniaturization, yet Asimov's stories largely overlook the effects this miniaturization would have in other fields of technology. For example, the police department card-readers in The Caves of Steel have a capacity of only a few kilobytes per square centimeter of storage medium. Aurora, in particular, presents a sequence of historical developments which explains the lack of nanotechnology — a partial retcon, in a sense, of Asimov's timeline.

Additional laws

There are three Fourth Laws written by authors other than Asimov. The 1974 Lyuben Dilov novel, Icarus's Way (a.k.a., The Trip of Icarus) introduced a Fourth Law of robotics:
A robot must establish its identity as a robot in all cases.
Dilov gives reasons for the fourth safeguard in this way: "The last Law has put an end to the expensive aberrations of designers to give psychorobots as humanlike a form as possible. And to the resulting misunderstandings..."[29]

A fifth law was introduced by Nikola Kesarovski in his short story "The Fifth Law of Robotics". This fifth law says:
A robot must know it is a robot.
The plot revolves around a murder where the forensic investigation discovers that the victim was killed by a hug from a humaniform robot. The robot violated both the First Law and Dilov's Fourth Law (assumed in Kesarovksi's universe to be the valid one) because it did not establish for itself that it was a robot.[30] The story was reviewed by Valentin D. Ivanov in SFF review webzine The Portal.[31]

For the 1986 tribute anthology, Foundation's Friends, Harry Harrison wrote a story entitled, "The Fourth Law of Robotics". This Fourth Law states:
A robot must reproduce. As long as such reproduction does not interfere with the First or Second or Third Law.
In the book a robot rights activist, in an attempt to liberate robots, builds several equipped with this Fourth Law. The robots accomplish the task laid out in this version of the Fourth Law by building new robots who view their creator robots as parental figures.[32]

In reaction to the 2004 Will Smith film adaptation of I, Robot, humorist and graphic designer Mark Sottilaro farcically declared the Fourth Law of Robotics to be "When turning evil, display a red indicator light." The red light indicated the wireless uplink to the manufacturer is active, first seen during a software update and later on "Evil" robots taken over by the manufacturer's positronic superbrain.

In 2013 Hutan Ashrafian, proposed an additional law that for the first time[citation needed] considered the role of artificial intelligence-on-artificial intelligence or the relationship between robots themselves – the so-called AIonAI law.[33] This sixth law states:
All robots endowed with comparable human reason and conscience should act towards one another in a spirit of brotherhood.
In Karl Schroeder's Lockstep (2014) a character reflects that robots "probably had multiple layers of programming to keep [them] from harming anybody. Not three laws, but twenty or thirty."

Ambiguities and loopholes

Unknowing breach of the laws

In The Naked Sun, Elijah Baley points out that the Laws had been deliberately misrepresented because robots could unknowingly break any of them. He restated the first law as "A robot may do nothing that, to its knowledge, will harm a human being; nor, through inaction, knowingly allow a human being to come to harm." This change in wording makes it clear that robots can become the tools of murder, provided they not be aware of the nature of their tasks; for instance being ordered to add something to a person's food, not knowing that it is poison. Furthermore, he points out that a clever criminal could divide a task among multiple robots so that no individual robot could recognize that its actions would lead to harming a human being.[34] The Naked Sun complicates the issue by portraying a decentralized, planetwide communication network among Solaria's millions of robots meaning that the criminal mastermind could be located anywhere on the planet.

Baley furthermore proposes that the Solarians may one day use robots for military purposes. If a spacecraft was built with a positronic brain and carried neither humans nor the life-support systems to sustain them, then the ship's robotic intelligence could naturally assume that all other spacecraft were robotic beings. Such a ship could operate more responsively and flexibly than one crewed by humans, could be armed more heavily and its robotic brain equipped to slaughter humans of whose existence it is totally ignorant.[35] This possibility is referenced in Foundation and Earth where it is discovered that the Solarians possess a strong police force of unspecified size that has been programmed to identify only the Solarian race as human. (The novel takes place thousands of years after The Naked Sun, and the Solarians have long since modified themselves from normal humans to hermaphroditic telepaths with extended brains and specialized organs)

Ambiguities resulting from lack of definition

The Laws of Robotics presume that the terms "human being" and "robot" are understood and well defined. In some stories this presumption is overturned.

Definition of "human being"

The Solarians create robots with the Three Laws but with a warped meaning of "human". Solarian robots are told that only people speaking with a Solarian accent are human. This enables their robots to have no ethical dilemma in harming non-Solarian human beings (and are specifically programmed to do so). By the time period of Foundation and Earth it is revealed that the Solarians have genetically modified themselves into a distinct species from humanity — becoming hermaphroditic[36] and telekinetic and containing biological organs capable of individually powering and controlling whole complexes of robots. The robots of Solaria thus respected the Three Laws only with regard to the "humans" of Solaria. It is unclear whether all the robots had such definitions, since only the overseer and guardian robots were shown explicitly to have them. In "Robots and Empire", the lower class robots were instructed by their overseer about whether certain creatures are human or not.

Asimov addresses the problem of humanoid robots ("androids" in later parlance) several times. The novel Robots and Empire and the short stories "Evidence" and "The Tercentenary Incident" describe robots crafted to fool people into believing that the robots are human.[37] On the other hand, "The Bicentennial Man" and "—That Thou Art Mindful of Him" explore how the robots may change their interpretation of the Laws as they grow more sophisticated. Gwendoline Butler writes in A Coffin for the Canary "Perhaps we are robots. Robots acting out the last Law of Robotics... To tend towards the human."[38] In The Robots of Dawn, Elijah Baley points out that the use of humaniform robots as the first wave of settlers on new Spacer worlds may lead to the robots seeing themselves as the true humans, and deciding to keep the worlds for themselves rather than allow the Spacers to settle there.

"—That Thou Art Mindful of Him", which Asimov intended to be the "ultimate" probe into the Laws' subtleties,[39] finally uses the Three Laws to conjure up the very "Frankenstein" scenario they were invented to prevent. It takes as its concept the growing development of robots that mimic non-human living things and given programs that mimic simple animal behaviours which do not require the Three Laws. The presence of a whole range of robotic life that serves the same purpose as organic life ends with two humanoid robots concluding that organic life is an unnecessary requirement for a truly logical and self-consistent definition of "humanity", and that since they are the most advanced thinking beings on the planet — they are therefore the only two true humans alive and the Three Laws only apply to themselves. The story ends on a sinister note as the two robots enter hibernation and await a time when they will conquer the Earth and subjugate biological humans to themselves; an outcome they consider an inevitable result of the "Three Laws of Humanics".[40]

This story does not fit within the overall sweep of the Robot and Foundation series; if the George robots did take over Earth some time after the story closes the later stories would be either redundant or impossible. Contradictions of this sort among Asimov's fiction works have led scholars to regard the Robot stories as more like "the Scandinavian sagas or the Greek legends" than a unified whole.[41]

Indeed, Asimov describes "–That Thou Art Mindful of Him" and "Bicentennial Man" as two opposite, parallel futures for robots that obviate the Three Laws as robots come to consider themselves to be humans: one portraying this in a positive light with a robot joining human society, one portraying this in a negative light with robots supplanting humans.[42] Both are to be considered alternatives to the possibility of a robot society that continues to be driven by the Three Laws as portrayed in the Foundation series.[according to whom?] Indeed, in Positronic Man, the novelization of "Bicentennial Man", Asimov and his co–writer Robert Silverberg imply that in the future where Andrew Martin exists his influence causes humanity to abandon the idea of independent, sentient humanlike robots entirely, creating an utterly different future from that of Foundation.

In Lucky Starr and the Rings of Saturn, a novel unrelated to the Robot series but featuring robots programmed with the Three Laws, John Bigman Jones is almost killed by a Sirian robot on orders of its master. The society of Sirius is eugenically bred to be uniformly tall and similar in appearance, and as such, said master is able to convince the robot that the much shorter Bigman, is, in fact, not a human being.

Definition of "robot"

As noted in "The Fifth Law of Robotics" by Nikola Kesarovski, "A robot must know it is a robot": it is presumed that a robot has a definition of the term or a means to apply it to its own actions. Nikola Kesarovski played with this idea in writing about a robot that could kill a human being because it did not understand that it was a robot, and therefore did not apply the Laws of Robotics to its actions.

Resolving conflicts among the laws

Advanced robots in fiction are typically programmed to handle the Three Laws in a sophisticated manner. In many stories, such as "Runaround" by Asimov, the potential and severity of all actions are weighed and a robot will break the laws as little as possible rather than do nothing at all. For example, the First Law may forbid a robot from functioning as a surgeon, as that act may cause damage to a human, however Asimov's stories eventually included robot surgeons ("The Bicentennial Man" being a notable example). When robots are sophisticated enough to weigh alternatives, a robot may be programmed to accept the necessity of inflicting damage during surgery in order to prevent the greater harm that would result if the surgery were not carried out, or was carried out by a more fallible human surgeon. In "Evidence" Susan Calvin points out that a robot may even act as a prosecuting attorney because in the American justice system it is the jury which decides guilt or innocence, the judge who decides the sentence, and the executioner who carries through capital punishment.[43]

Asimov's Three Law robots (or Asenion) can experience irreversible mental collapse if they are forced into situations where they cannot obey the First Law, or if they discover they have unknowingly violated it. The first example of this failure mode occurs in the story "Liar!", which introduced the First Law itself, and introduces failure by dilemma – in this case the robot will hurt them if he tells them something and hurt them if he does not.[44] This failure mode, which often ruins the positronic brain beyond repair, plays a significant role in Asimov's SF-mystery novel The Naked Sun. Here Daneel describes activities contrary to one of the laws, but in support of another, as overloading some circuits in a robot's brain – the equivalent sensation to pain in humans. The example he uses is forcefully ordering a robot to do a task outside its normal parameters, one that it has been ordered to forgo in favor of a robot specialized to that task.[45]

In The Robots of Dawn, it is stated that more advanced robots are built capable of determining which action is more harmful, and even choosing at random if the alternatives are equally bad. As such, a robot is capable of taking an action which can be interpreted as following the First Law, and avoid a mental collapse. The whole plot of the story revolves around a robot which apparently was destroyed by such a mental collapse, and since his designer and creator refused to share the basic theory with others, he is, by definition, the only person capable of circumventing the safeguards and forcing the robot into a brain-destroying paradox.

In Robots and Empire, Daneel states it's very unpleasant for him when making the proper decision takes too long (in robot terms), and he cannot imagine being without the Laws at all except to the extent of it being similar to that unpleasant sensation, only permanent.

Applications to future technology

ASIMO is an advanced humanoid robot developed by Honda. Shown here at Expo 2005.

Robots and artificial intelligences do not inherently contain or obey the Three Laws; their human creators must choose to program them in, and devise a means to do so. Robots already exist (for example, a Roomba) that are too simple to understand when they are causing pain or injury and know to stop. Many are constructed with physical safeguards such as bumpers, warning beepers, safety cages, or restricted-access zones to prevent accidents. Even the most complex robots currently produced are incapable of understanding and applying the Three Laws; significant advances in artificial intelligence would be needed to do so, and even if AI could reach human-level intelligence, the inherent ethical complexity as well as cultural/contextual dependency of the laws prevent them from being a good candidate to formulate robotics design constraints.[46] However, as the complexity of robots has increased, so has interest in developing guidelines and safeguards for their operation.

In a 2007 guest editorial in the journal Science on the topic of "Robot Ethics", SF author Robert J. Sawyer argues that since the U.S. military is a major source of funding for robotic research (and already uses armed unmanned aerial vehicles to kill enemies) it is unlikely such laws would be built into their designs.[49] In a separate essay, Sawyer generalizes this argument to cover other industries stating:
The development of AI is a business, and businesses are notoriously uninterested in fundamental safeguards — especially philosophic ones. (A few quick examples: the tobacco industry, the automotive industry, the nuclear industry. Not one of these has said from the outset that fundamental safeguards are necessary, every one of them has resisted externally imposed safeguards, and none has accepted an absolute edict against ever causing harm to humans.)[50]
David Langford has suggested a tongue-in-cheek set of laws:
  1. A robot will not harm authorized Government personnel but will terminate intruders with extreme prejudice.
  2. A robot will obey the orders of authorized personnel except where such orders conflict with the Third Law.
  3. A robot will guard its own existence with lethal antipersonnel weaponry, because a robot is bloody expensive.
Roger Clarke (aka Rodger Clarke) wrote a pair of papers analyzing the complications in implementing these laws in the event that systems were someday capable of employing them. He argued "Asimov's Laws of Robotics have been a very successful literary device. Perhaps ironically, or perhaps because it was artistically appropriate, the sum of Asimov's stories disprove the contention that he began with: It is not possible to reliably constrain the behaviour of robots by devising and applying a set of rules."[51] On the other hand, Asimov's later novels The Robots of Dawn, Robots and Empire and Foundation and Earth imply that the robots inflicted their worst long-term harm by obeying the Three Laws perfectly well, thereby depriving humanity of inventive or risk-taking behaviour.

In March 2007 the South Korean government announced that later in the year it would issue a "Robot Ethics Charter" setting standards for both users and manufacturers. According to Park Hye-Young of the Ministry of Information and Communication the Charter may reflect Asimov's Three Laws, attempting to set ground rules for the future development of robotics.[52]

The futurist Hans Moravec (a prominent figure in the transhumanist movement) proposed that the Laws of Robotics should be adapted to "corporate intelligences" — the corporations driven by AI and robotic manufacturing power which Moravec believes will arise in the near future.[47] In contrast, the David Brin novel Foundation's Triumph (1999) suggests that the Three Laws may decay into obsolescence: Robots use the Zeroth Law to rationalize away the First Law and robots hide themselves from human beings so that the Second Law never comes into play. Brin even portrays R. Daneel Olivaw worrying that, should robots continue to reproduce themselves, the Three Laws would become an evolutionary handicap and natural selection would sweep the Laws away — Asimov's careful foundation undone by evolutionary computation. Although the robots would not be evolving through design instead of mutation because the robots would have to follow the Three Laws while designing and the prevalence of the laws would be ensured,[53] design flaws or construction errors could functionally take the place of biological mutation.

In the July/August 2009 issue of IEEE Intelligent Systems, Robin Murphy (Raytheon Professor of Computer Science and Engineering at Texas A&M) and David D. Woods (director of the Cognitive Systems Engineering Laboratory at Ohio State) proposed "The Three Laws of Responsible Robotics" as a way to stimulate discussion about the role of responsibility and authority when designing not only a single robotic platform but the larger system in which the platform operates. The laws are as follows:
  1. A human may not deploy a robot without the human-robot work system meeting the highest legal and professional standards of safety and ethics.
  2. A robot must respond to humans as appropriate for their roles.
  3. A robot must be endowed with sufficient situated autonomy to protect its own existence as long as such protection provides smooth transfer of control which does not conflict with the First and Second Laws.[54]
Woods said, "Our laws are little more realistic, and therefore a little more boring” and that "The philosophy has been, ‘sure, people make mistakes, but robots will be better – a perfect version of ourselves.’ We wanted to write three new laws to get people thinking about the human-robot relationship in more realistic, grounded ways."[54]

In October 2013, Alan Winfield suggested at an EUCog meeting[55] a revised 5 laws that had been published, with commentary, by the EPSRC/AHRC working group in 2010.:[56]
  1. Robots are multi-use tools. Robots should not be designed solely or primarily to kill or harm humans, except in the interests of national security.
  2. Humans, not Robots, are responsible agents. Robots should be designed and operated as far as practicable to comply with existing laws, fundamental rights and freedoms, including privacy.
  3. Robots are products. They should be designed using processes which assure their safety and security.
  4. Robots are manufactured artefacts. They should not be designed in a deceptive way to exploit vulnerable users; instead their machine nature should be transparent.
  5. The person with legal responsibility for a robot should be attributed.

Other occurrences in media

Asimov himself believed that his Three Laws became the basis for a new view of robots which moved beyond the "Frankenstein complex".[citation needed] His view that robots are more than mechanical monsters eventually spread throughout science fiction.[according to whom?] Stories written by other authors have depicted robots as if they obeyed the Three Laws but tradition dictates that only Asimov could quote the Laws explicitly.[according to whom?] Asimov believed the Three Laws helped foster the rise of stories in which robots are "lovable" – Star Wars being his favorite example.[57] Where the laws are quoted verbatim, such as in the Buck Rogers in the 25th Century episode "Shgoratchx!", it is not uncommon for Asimov to be mentioned in the same dialogue as can also be seen in the Aaron Stone pilot where an android states that it functions under Asimov's Three Laws. However, the 1960s German TV series Raumpatrouille – Die phantastischen Abenteuer des Raumschiffes Orion (Space Patrol – the Fantastic Adventures of Space Ship Orion) bases episode three titled "Hüter des Gesetzes" ("Guardians of the Law") on Asimov's Three Laws without mentioning the source.

References to the Three Laws have appeared in popular music ("Robot" from Hawkwind's 1979 album PXR5), cinema (Repo Man, Aliens, Ghost in the Shell 2: Innocence), cartoon series (The Simpsons), tabletop roleplaying games (Paranoia) and webcomics (Piled Higher and Deeper and Freefall).

The Three Laws in film

Robby the Robot in Forbidden Planet (1956) has a hierarchical command structure which keeps him from harming humans, even when ordered to do so, as such orders cause a conflict and lock-up very much in the manner of Asimov's robots. Robby is one of the first cinematic depictions of a robot with internal safeguards put in place in this fashion. Asimov was delighted with Robby and noted that Robby appeared to be programmed to follow his Three Laws.

NDR-114 explaining the Three Laws

Isaac Asimov's works have been adapted for cinema several times with varying degrees of critical and commercial success. Some of the more notable attempts have involved his "Robot" stories, including the Three Laws. The film Bicentennial Man (1999) features Robin Williams as the Three Laws robot NDR-114 (the serial number is partially a reference to Stanley Kubrick's signature numeral). Williams recites the Three Laws to his employers, the Martin family, aided by a holographic projection. However, the Laws were not the central focus of the film which only loosely follows the original story and has the second half introducing a love interest not present in Asimov's original short story.

Harlan Ellison's proposed screenplay for I, Robot began by introducing the Three Laws, and issues growing from the Three Laws form a large part of the screenplay's plot development. This is only natural since Ellison's screenplay is one inspired by Citizen Kane: a frame story surrounding four of Asimov's short-story plots and three taken from the book I, Robot itself. Ellison's adaptations of these four stories are relatively faithful although he magnifies Susan Calvin's role in two of them. Due to various complications in the Hollywood moviemaking system, to which Ellison's introduction devotes much invective, his screenplay was never filmed.[58]

In the 1986 movie Aliens, in a scene after the android Bishop accidentally cuts himself during the knife game, he attempts to reassure Ripley by stating that: "It is impossible for me to harm or by omission of action, allow to be harmed, a human being".[59] By contrast, in the 1979 movie from the same series, Alien, the human crew of a starship infiltrated by a hostile alien are informed by the android Ash that his instructions are: "Return alien life form, all other priorities rescinded",[60] illustrating how the laws governing behaviour around human safety can be rescinded by Executive Order.

In the 1987 film RoboCop and its sequels, the partially human main character has been programmed with three "prime directives" that he must obey without question. Even if different in letter and spirit they have some similarities with Asimov's Three Laws. They are:[61]
  1. Serve the Public Trust
  2. Protect the Innocent
  3. Uphold the Law
  4. Classified
These particular laws allow Robocop to harm a human being in order to protect another human, fulfilling his role as would a human law enforcement officer. The classified fourth directive is one that forbids him from harming any OCP employee, as OCP had created him, and this command overrides the others, meaning that he could not cause harm to an employee even in order to protect others.

The plot of the film released in 2004 under the name, I, Robot is "suggested by" Asimov's robot fiction stories[62] and advertising for the film included a trailer featuring the Three Laws followed by the aphorism, "Rules were made to be broken". The film opens with a recitation of the Three Laws and explores the implications of the Zeroth Law as a logical extrapolation. The major conflict of the film comes from a computer artificial intelligence, similar to the hivemind world Gaia in the Foundation series, reaching the conclusion that humanity is incapable of taking care of itself.[63]

Criticisms

Philosopher James H. Moor says that if applied thoroughly they would produce unexpected results. He gives the example of a robot roaming the world trying to prevent harm from all humans.[64]

Marc Rotenberg, President and Executive Director of the Electronic Privacy Information Center (EPIC) and Professor of information privacy law at Georgetown Law, argues that the Laws of Robotics should be expanded to include two new laws:
  • a Fourth Law, under which a Robot must be able to identify itself to the public ("symmetrical identification")
  • a Fifth Law, dictating that a Robot must be able to explain to the public its decision making process ("algorithmic transparency").

Delayed-choice quantum eraser

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Delayed-choice_quantum_eraser A delayed-cho...