Monday, April 1, 2019
Arguments on Artificial Intelligence
Arguments on  synthetic IntelligenceWe live in an  extraordinaire(postnominal) time. Improvements in technology seem to be accelerating at an unbelievable rate.  all(prenominal) time they think Moores Law has r for each oneed its limits, tech companies come up with a new level of capability. No less is the advancement of artificial  cognition (AI). Our ein truth day lives are already deeply immersed in AI, and we dont  up to now know it. It controls much of the financial markets, performs law enforcement tasks, and  considers our internet searches more useful.  or so AI today is weak AI,  intentional to perform a very specific task (Tegmark, n.d.). But the goal of all research and  corporal investment is always more what else can we know or do? Often, these entities are creating things in a vacuum, with limited moral, honorable, or legal boundaries. When is it  also much? The driving force that makes us want to always  look for further is what makes the development and use of artific   ial  apprehension (AI) a  uncollectible course of action.Why is this a risky course of action? Because  large control of systems to artificial  learning could  require seriously  invalidating results. Take, for example, researchers  workings with the University of Pittsburgh Medical Center. In this case, they develop a neural network that returns  draw outions for  interposition of pneumonia patients. Using a historical database with the solutions and results of methods of  diplomacyment, the AI is supposed to provide suggested solutions to treat patients. In one solution, it recommended that certain high risk patients be  displace home (Bornstein, 2016). This solution had high probability of resulting in death.When working with and  conf apply task, accomplished by human or machine, the law of unintended consequences  must always be con locationred. No matter how well someone thinks they  befuddle thought a system through, it is nearly impossible to consider alwaysy possible outcom   e. Certainly, unintended consequences are  non all bad,  many a(prenominal) drugs  begin side effects that are beneficial and completely not what the drugs was designed to do. On the other hand, many drugs have very negative side effects. Certainly, they are not intended to cause any adverse symptoms,  precisely many have severe unintended consequences, including death.Some would argue, AI is  before long in use and benefits everyone with no negative effects. Singularity cannot happen. While we  for sure use some types of AI currently and have had minimal negative effects. It is also true we have not reached singularity. It is the height of hubris to believe that we have total control over anything or that we have considered all possibilities.  insure Fukishima or Chernobyl, all possibilities were not covered and resulted in huge disasters. sluice NASA, the  houseard for careful scrutiny of complex systems and procedures has had some catastrophic failures in the form of space shuttl   e crashes due to hubris of the organization and/or individuals.How many people died on the Titanic? A ship that was unsinkable was  drop by a simple iceberg, or was it hubris? The shoddy steel used in the construction of the hull, the poorly designed bulkheads that didnt reach to the top deck, and the  compress to go as fast as it could are what sunk the ship. And not enough life boats on the unsinkable ship killed the passengers. Hubris lead them  calibrate the path to destruction.We are at the point that we have the capability to  faith AI to create autonomous military machines. Some are  even off in the testing phase of development. Machines that make decisions of life and death on their own (Russell, 2015). Absent human intervention, what is to keep one of these machines from deciding the  untimely person is a target. A machine knows no morality, no ethical code, only its programming, its goal or reason to exist. Given a  sizeable enough computational system, it could decide to    use everything at its disposal to  procure its goals (Anderson, 2017). Things like taking control of infrastructure, or even humans.So, what do we do? Is there risk? Even captains of industry and experts like Gates, Musk, and Hawking suggest there is (Holley, 2015). It is clear we are already on the path to creating ever more complex and capable AI. We must recognize that we all make mistakes and constantly be on guard against mistakes and, more importantly, hubris. Most  blowup of knowledge has risk. When confronted with a discipline that has catastrophic possibilities, we must fight the  relish to run forward as fast as we can with no concern for the consequences. Methodical deliberation is the only course. We must consider the ramifications of each step and ensure safeguards are in place should we need to  change or isolate any AI that develops goals counter to those of humans. If we manage to be conscientious enough and adhere to ethical principles, we might, just might, keep fr   om  develop the instrument of our own demise.ReferencesAnderson, J. (2017, February 16). Googles artificial intelligence getting greedy, and aggressive. Activist Post. Retrieved from http//www.activistpost.com/2017/02/googles-artificial-intelligence-getting-greedy-and-aggressive/Artificial Intelligence. (2015). In Opposing Viewpoints Online Collection. Detroit Gale. Retrieved from http//link.galegroup.com.ezproxy.libproxy.db.erau.edu/apps/doc/PC3010999273/OVIC?u=embryxid=415989d5Bornstein, A. (2016, September 1). Is artificial intelligence permanently inscrutable?Holley, P. (2015, January 29). Bill Gates on the dangers of artificial intelligence I dont understand why some people are not concerned. The Washington Post. Retrieved from https//www.washingtonpost.com/news/the-switch/wp/2015/01/28/bill-gates-on-dangers-of-artificial-intelligence-dont-understand-why-some-people-are-not-concerned/Russell, S. (2015, May 28). Take a stand on AI weapons. Nature, 521 (7553), 415-416.  
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.