Группа авторов

Cyberphysical Smart Cities Infrastructures


Скачать книгу

point of trust, though, does bring up an obvious issue. For many, practicing medicine is a lifelong journey of continual education, and as things are now, most doctors do not know how to code at the level of creating AI. This means that HCPs and programmers must work together to identify trends, patterns, and known symptomatic association that the AI would use. This limitation is likely why AI is kept in the passenger seat and used as a secondary diagnosis, rather than replacing HCPs as the primary diagnostician.

      Earlier, it was brought up how HCPs are often overworked and understaffed. In some cases, AI has moved from beyond the virtual aspect of diagnostics and made its way into the physical world. In these cases, AI is often represented by carebots. These carebots can perform a variety of functions including providing company for the elderly or delivering medication to patients [3]. Although AI and robotics are making large steps forward, it will likely still be a long time before AI carebots are able to perform more than simple basic tasks or are widespread across healthcare. Carebots are not the only physical presence of AI in healthcare. AI is also assisting HCPs in surgical matters via robots and in some cases is even performing surgical procedures by themselves [3]. These solo surgical procedures are small, simple, basic surgeries, but the AI is still performing the process without the control of an HCP.

      Ethics and morality have a large impact on AI in healthcare. Whereas science, technology, and innovation limit what is possible, ethics and morality limit what should be done versus not. Just because the ability to do something exists does not mean that it should be done. HCPs and AI programmers are making attempts to incorporate ethics and morality into ML and final AI products by implementing standards [1]. According to Bryson, these standards are set via consensus of the masses. In other words, if the majority of people feel that it is the ethical or moral thing to do, then that becomes the standard. This obviously is highly flawed as that standard can change based off cultural or societal needs and wants. If the standards are based on an extremist's point of view, for instance, the Taliban, does that mean that these AI carebots should not assist someone if they identify them as a woman, or a person of color, or a foreign ethnic group?

      Human beings are flawed creatures to begin with; just because there is a consensus for how to go about doing something does not necessarily mean that it is right. Look at the Salem witch trials. At the time, it was the consensus of the majority that women who were deemed to be partaking in witchcraft be burned at the stake to protect the rest in the village. If the goal of AI is to have autonomous robots that function completely independently of a user that take new information and learn from it, how do we instill ethics into it?

      The Institute of Electrical and Electronic Engineers (IEEEs) has attempted to overcome this barrier by listening to as many voices in AI as possible and unifying them. The IEEE's Initiative for Ethical Considerations in Artificial Intelligence and Autonomous Systems considers various voices, communities, and school of thoughts and then translates this into education and training. IEEE strives to “prioritize ethical considerations in the design and development of autonomous and intelligent system” [1]. However, is that enough? Rigby does not believe so. In fact, he believes that “current policy and ethical guidelines for AI technology are lagging behind the progress AI has made in the healthcare field” [2]. When we use Moore's law to view the pace of technological innovation, we see that every 18–24 months the ability of technology basically doubles. This pace cannot be matched by those who implement public policy or guidelines. The result is that new technology and innovation is being implemented ethically agnostic or ethically askew, or somewhere in between, there is no standard.

      Another ethical dilemma that should be considered involves the use of ML. If ML's goal is to identify and capture patterns, then what happens when it identifies a pattern within its own coding as being unethical or immoral? This goes back to the idea that humans are flawed, and depending on who, where, and when the ethical creation came from, it could still be flawed. At that intersection, the AI is at a crossroads. Does it defy the limitations put on it by the ethics imposed, or does it continue doing something that it perceives to be ethically wrong? With the rate of technological innovation, these questions must be addressed and answered sooner rather than later.

      There is no doubt that in this era of technological advance and innovation, data is the currency that powers all. Data can be in any number of forms, but it is the driving force behind closed doors as to what decisions are being made. Even in the case of AI, this holds true. Gibian discusses in his article how having more data allows AI to be improved and be more accurate, even surpassing that of humans [5]. We see how data drives ML to make better informed decisions and produce better results. However, what happens when that data is corrupted or unavailable? How do we even know whether it is reliable? The answer to these questions centers around cybersecurity, securing our devices, and protecting our data.