Short Communication Volume 2 Issue 2
PHD Freelance Philosopher, USA
Correspondence: Rongqing Dai, PHD Freelance Philosopher, USA, Tel 1-732-395-3163
Received: October 15, 2016 | Published: April 10, 2017
Citation: Dai R. How could robots challenge humans? Int Rob Auto J. 2017;2(2):41-43. DOI: 10.15406/iratj.2017.02.00014
For the issue whether robots might challenge humans in the future, people have so far focused on the possibility of creating a robot that could be considered as a human in the sense that it could indeed think like a human instead of being solely a tool of humans operated with programmed instructions. Based upon this line of thought, it seems that we do not need to worry about the threat of robots since nobody could yet provide any plausible evidence that it is possible to produce this type of robots even after Google’s Alpha Go defeated human Go Masters. However, this way of thinking is philosophically wrong since the fundamental factor to enable robots to challenge humans would be their sociability once they are infused certain level of intelligence. This writing would propose the criteria for robots to be considered as sociable and discuss how the robots would pose potential threats to humans.
Keywords: robot, sociability, community, human, philosophy
The debate over “if robots would overtake humans” has been heated up in recent years by warnings against the potential threat of unregulated development of robots from some academic or industrial superstars. However, what is obviously missing in those warnings is a clear description of any realistic scenario by which robots could assuredly challenge humans as a whole, not as puppets programmed and controlled by humans, but as autonomous powers acting on their own "will". If this type of scenarios would never be realistic then we might still not need to worry too much about the so-called demonic threat of robots as warned by some experts. This is because even if we might possibly see robots be used as ruthless killing machines in near future by terrorists, dictators and warlords as warned by the elite scientists and experts1, it is just another form of human threat in the end. However, if robots could act on their own "will", not as puppets programmed and controlled by humans, then humans do need to start worrying about how to prevent the peril from happening instead of how to win debates over imaginary dangers. The reason that people on both sides of the debate could not see or show a very clear scenario that robots could indeed challenge humans in a very realistic way is truly a philosophical issue. So far all discussions on the issues like whether robots would threaten humans in the future have focused on the possibility of creating a robot that could be considered as a human in the sense that it could indeed think as a human instead of being solely a tool of humans operated with programmed instructions. According to this line of thought it seems that we still do not need to worry about the threat of robots to our human species as a whole since nobody could yet provide any plausible reason that it is possible to produce this type of robots even after Google’s Alpha Go defeated human Go Masters.
Unfortunately this way of thinking is philosophically incorrect because people who are thinking in this way are missing a fundamental point about our own human nature: human beings are social creatures. An important reason that we could survive as what we are now and could do what we are doing now is because we are living and acting as a societal community. Similarly, when we estimate the potential of robots we should not solely focus our attention on their individual intelligence (which of course is so far infused by humans), but should also take into consideration their sociability (which of course would be initially created by humans).
This would further lead to another philosophical question: what would fundamentally determine the sociability of robots? There might be a wide range of arguments on this question. But in term of being able to challenge humans I would argue that the fundamental criteria of sociability for robots could be defined as follows:
Once robots could possess the above capacities and start to “live” together as a mutually dependent multitude, we should reasonably view them as sociable beings. Sociable robots could form community of robots. Once robots could function as defined above and form a community they would no longer need to live as slaves of their human masters. Once that happens, it would be the beginning of a history that robots could possibly challenge humans or start their cause of taking over humans. Since not all the capacities mentioned above exist (at least publically) in this world today, to avoid any unnecessary argument, it would be wise to make our judgment based upon whether any known scientific principle would be violated in any practical attempt to realize any particular capacity among those mentioned above. Communication with other machines, moving objects, operating and repairing machine systems and exploring natural resources are all among nowadays common practices with programmed machineries. Therefore, even though we might not have a single robot or a group of single robots possess all the capacities mentioned above, there is no fundamental reason for any of the capacities mentioned above to be considered as not producible according to any known scientific principle, the only thing left to do would be to integrate those capacities together onto a single whole robot (and thus a group of single robots). Since we don’t see any known scientific principle that would prevent any of those capacities from being realized, we should reasonably expect that with money to be invested and with time to be spent the creation of sociable robots as defined earlier could foreseeable become real unless some special efforts to be made by humans in this world to prevent that from happening.
Although sociability would be a critical precondition for robots to challenge humans, it might still not be sufficient for robots to pose any threat to humans yet. In order for robots to become a real threat to humans, they need to possess the ability to fight or combat. Unfortunate for humans, fighting ability of robots might be more real than their sociability. It is reasonable to expect that human manufacturers of robots would make great efforts to integrate as much the most advanced technology available as possible into the design and production of robots. Therefore, based upon some common knowledge about nowadays technology and what we have already witnessed about what robots could do, we might very moderately expect that an army of robots would be capable of doing the following:
Besides, according to the definition of sociability of robots given above, robots in their own community would be able to 1) help each other to recover from damage or shutdown, and thus it would not be an issue for robots to replace their existing operating system or application programs if needed, and the same would be true for the replacement or addition of required new hardware parts; 2) manufacture new parts for producing new robots, and thus as long as there are designs for new software or hardware, they could produce the final products based upon the design.
The above two points are what robots could be practically made to do even today. However, in order for robots to become a radical threat to humans in the sense of being able to win a full-scale war against humans, they need to be able to perform complicated logical reasoning when facing various unfamiliar situations. Because human knowledge system is limited, when humans creating robots, the complete set of logic infused into the computer system of any robot would be restrained within certain framework due to the limit of human knowledge. But the natural and social reality is wide open, and thus the logic system of robots might not be good enough to handle some complicated situations when they do not look familiar to robots. On the other hand, as we have been aware that being able to think like humans and being able to challenge humans are two different things. Since in this discussion we are more interested in how robots could challenge or threaten humans, we don’t need to look into the popular topic of the general learning ability of robots. We only need to examine the possibility for robots to perform complicated logical reasoning to challenge humans, which might be a more difficult goal than any capability or functionality so far mentioned in this writing. There could be two different ways to achieve the goal of making robots smart enough in terms of intelligence to challenge humans.
We might call the first way as Nurturing way, by which humans continue to improve the logical reasoning ability of robots through AI programming development even after the robots have formed a community. Humans keep nurturing the community of robots in this way until at one point they are good enough to win the full-scale war against humans and then set them off to fight against humans. (To people without a technical background, this might sound like a wishful thinking without assured certainty; but people with some basic programming background would be able to see as long as time and money are invested in creating a society of robots that could challenge humans, this is hundred percent doable). The second way would be an Evolution way, by which from the very beginning humans create a community of robots that could make their own evolution through software and hardware upgrading. The main challenge for robots to be able to evolve would be how they could evolve through design for upgrading their own software and hardware. The task to make robots able to evolve by themselves could then be reduced to two simpler tasks: 1) to enable robots to identify needs, 2) to enable robots to make software and hardware designs based upon needs. The first
goal of identifying needs could be achieved by recording the history of failure to accomplish a previous mission, which could in turn be achieved by examining (through some fuzzy logic type programming) how a previous mission was accomplished. The second goal of designing based upon needs might be a bit more complicated in principle, but still possible to be fulfilled. This second approach (i.e. the Evolution way) would be a bigger challenge than the Nurturing way mentioned above and so far we still cannot see a hundred percent certainty for this to happen in the future even if money and time would be invested. However, even if humans failed to create the evolutionary community of robots, they still could help robots to be intelligent enough to fight a full scale war against humans through the Nurturing way mentioned above.
There is still one critical question left for this writing to answer which is why any reasonable humans would create a community of robots that is socially independent of humans with lethal power and help them to fight against humans instead of making them tools or slaves of humans?
We need to look at this question from two different levels. First, whether someone who is able to mobilize and organize resource to create a community of sociable robots would indeed have the intention to do so is a social issue, which is not under any hard restriction provided by natural laws. As long as something is possible to happen according to natural laws, we could not exclude the possibility solely based upon our own wishful thinking about the intentions of all humans. Second, human civilization contains some suicidal gene in itself. The competition of human society would provide enough motives for people who are able to do something to enhance their own competing power to push their creativity and productivity to the maximal edge. Furthermore, history has proven that humans are vulnerable to ignorance of many potential risks when they are going to extremes for their own benefits. Especially, once some groups of humans are capable of doing something with potentially dangerous risks for others and themselves, a very few decision makers or even one single person could make the difference of whether they would actually do it or not. Since there is no natural law to prevent the community of sociable robots with lethal power from being created, without social efforts of regulations, we could potentially come to a point when we need to count on the psychological stability of very few or even a single person to determine whether humans would be threatened by robots or not.
Even though we might agree that in order to gain advantages in competition human enterprises might go for various extremes including creating socialized robots, we still need to see what kind of advantages might be gained by creating socialized robots in order to estimate what might be its consequence. Actually, the horrifying potential of robots’ rebellion against their human masters as shown in fiction movies might be the only critical risk for making robots capable of getting socialized among themselves. A socialized community of robots would be definitely more productive when accomplishing a task than a group of robots that are remotely controlled by humans. For interstellar missions, the restriction of slower than light speed communication would cause remote control very difficult; and even for missions on earth, for the same reason that we would want a group of humans to cooperate with each other efficiently, we might hope robots to cooperate with each other effectively in civil or military actions so that they might successfully accomplish their tasks. Therefore, as long as it is possible to have robots to act in socialized community, the benefit would be clear and thus the temptations of making it happen could be very high.
Besides, the risk that socialized robots might rebel their human masters might look very minor at the beginning. This is because whoever creates the socialized community of robots could encode their loyalty rules and thus they would never rebel their own human masters. However, in the long run, due to the selfish nature of human beings, any loyalty code given to robots could be a hazard to the human society if it is not properly created. This is because human masters would naturally want their robot armies only loyal to themselves for a loyal army of robots might be even more reliable than any human squad. However, if not properly programmed, after the masters die, those armies of robots with the autonomous capacity of survival will continue to function without loyalty to anyone.
Even though socialized robots might possibly become a threat to human civilization in the future, it might be very hard to stop it from happening in the global range since that, as discussed in this writing, technically it is possible to make it happen and functionally socialized robots might benefit their human masters. But on the other hand, if we could plan in advance and proceed rationally in a civilized way, then we might even make the society of robots as part of our earth civilization and have them partner with us as good citizens. When that happens robots might contribute to our civilization positively with their superior advantages mentioned earlier in this writing.
None.
The author declares no conflict of interest.
©2017 Dai. This is an open access article distributed under the terms of the, which permits unrestricted use, distribution, and build upon your work non-commercially.