Toward an Ethics of AI Assistants: an Initial Framework

Philosophy & Technology - Tập 31 - Trang 629-653 - 2018
John Danaher1
1School of Law, NUI Galway, Galway, Ireland

Tóm tắt

Personal AI assistants are now nearly ubiquitous. Every leading smartphone operating system comes with a personal AI assistant that promises to help you with basic cognitive tasks: searching, planning, messaging, scheduling and so on. Usage of such devices is effectively a form of algorithmic outsourcing: getting a smart algorithm to do something on your behalf. Many have expressed concerns about this algorithmic outsourcing. They claim that it is dehumanising, leads to cognitive degeneration, and robs us of our freedom and autonomy. Some people have a more subtle view, arguing that it is problematic in those cases where its use may degrade important interpersonal virtues. In this article, I assess these objections to the use of AI assistants. I will argue that the ethics of their use is complex. There are no quick fixes or knockdown objections to the practice, but there are some legitimate concerns. By carefully analysing and evaluating the objections that have been lodged to date, we can begin to articulate an ethics of personal AI use that navigates those concerns. In the process, we can locate some paradoxes in our thinking about outsourcing and technological dependence, and we can think more clearly about what it means to live a good life in the age of smart machines.

Tài liệu tham khảo

Burgos, D, Van Nimwegen, C, Van Oostendorp, H. and Koper, R. (2007). Game-based learning and immediate feedback. The case study of the Planning Educational Task. International Journal of Advanced Technology in Learning Available at http://hdl.handle.net/1820/945 (accessed 29/11/2016). Burrell, J. (2016). How the machine thinks: Understanding opacity in machine learning systems. Big Data and Society. https://doi.org/10.1177/2053951715622512. Carr, N. (2014). The glass cage: Where automation is taking us. London: The Bodley Head. Crawford, M. (2015). The world beyond your head. New York: Farrar, Strauss and Giroux. Danaher, J. (2016a). The threat of algocracy: Reality, resistance and accommodation. Philosophy and Technology, 29(3), 245–268. Danaher, J. (2016b). Why internal moral enhancement might be politically better than external moral enhancement. Neuroethics. https://doi.org/10.1007/s12152-016-9273-8 Dworkin, G. (1988). The theory and practice of autonomy. Cambridge: CUP. Frankfurt, H. (1971). Freedom of the will and the concept of a person. Journal of Philosophy, 68, 5–20. Frischmann, B. (2014). Human-focused Turing tests: A framework for judging nudging and the techno-social engineering of humans. Cardozo Legal Studies Research Paper No. 441 - available at https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2499760 (accessed 29/11/2016). Giublini, A., & Savulescu, J. (2018). The Artificial Moral Advisor. The 'Ideal Observer' meets Artificial Intelligence. Philosophy and Technology, 31(2):169–188. Hare, S., & Vincent, N. (2016). Happiness, cerebroscopes and incorrigibility: Prospects for Neuroeudaimonia. Neuroethics, 9(1), 69–84. Heersmink, R. (2015). Extended mind and cognitive enhancement: Moral aspects of extended cognition. Phenomenal Cognitive Science. https://doi.org/10.1007/s11097-015-9448-5. Heersmink, R. (2013). A taxonomy of cognitive artifacts: Function, information and categories. Review of Philosophical Psychology, 4(3), 465–481. Kelly, S., & Dreyfus, H. (2011). All things shining. New York: Free Press. Kirsh, D. (2010). Thinking with external representations. AI and Society, 25, 441–454. Kirsh, D. (1995). The intelligent use of space. Artificial Intelligence, 73, 31–68. Krakauer, D. (2016). Will AI harm us? Better to ask how we’ll reckon with our hybrid nature. Nautilus 6 September 2016 - available at http://nautil.us/blog/will-ai-harm-us-better-to-ask-how-well-reckon-with-our-hybrid-nature (accessed 29/11/2016). Luper, S. (2014). Life’s meaning. In Luper (Ed.), The Cambridge Companion to Lie and Death. Cambridge: Cambridge University Press. Mittelstadt, B., Allo, P., Taddeo, M., Wachter, S., & Floridi, L. (2016). The ethics of algorithms: Mapping the debate. Big Data and Society. https://doi.org/10.1177/2053951716679679. Morozov, E. (2013). The real privacy problem. MIT Technology Review. Available at http://www.technologyreview.com/featuredstory/520426/the-real-privacy-problem/ (accessed 29/11/16). Mullainathan, S. and Shafir, E. (2014) Freeing up intelligence. Scientific American Mind Jan/Feb: 58–63. Mullainathan, S., & Shafir, E. (2012). Scarcity: The true cost of not having enough. London: Penguin. Nagel, S. (2010). Too much of a good thing? Enhancement and the burden of self-determination. Neuroethics, 3, 109–119. Nass, C. and Flatow, I. (2013) The myth of multitasking. NPR: Talk of the Nation 10 May 2013 - available at http://www.npr.org/2013/05/10/182861382/the-myth-of-multitasking (accessed 29/11/2016). van Nimwegen, C., Burgos, D., Oostendorp, H and Schijf, H. (2006). The paradox of the assisted user: Guidance can be counterproductive. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems 917–926. Newport, C. (2016). Deep Work. New York: Grand Central Publishing. Norman, D. (1991). Cognitive artifacts. In J. M. Carroll (Ed.), Designing interaction: Psychology at the human-computer interface. Cambridge: Cambridge University Press. Ophir, E., Nass, C., & Wagner, A. (2009). Cognitive control in media multitaskers. PNAS, 107(37), 15583–15587. Pinker, S. (2010). The cognitive niche: Coevolution of intelligence, sociality, and language. PNAS, 107(Suppl 2), 8993–8999. Plato. The Phaedrus. From Plato in Twelve Volumes, Vol. 9, translated by Harold N. Fowler. Cambridge, MA, Harvard University Press; London, William Heinemann Ltd. 1925. Available at http://www.english.illinois.edu/-people-/faculty/debaron/482/482readings/phaedrus.html (accessed 29/11/2016). Raz, J. (1986). The morality of freedom. Oxford: OUP. Russell, S. and Norvig, P. (2016) Artificial intelligence: A modern approach (Global 3rd edition). Essex: Pearson. Sandel, M. (2012). What money can’t buy: The moral limits of markets. London: Penguin. Scheibehenne, B., Greifeneder, R., & Todd, P. M. (2010). Can there ever be too many options? A meta-analytic review of choice overload. Journal of Consumer Research, 37, 409–425. Scherer, M. (2016). Regulating artificial intelligence systems: Challenges, competencies and strategies. Harvard Journal of Law and Technology, 29(2), 354–400. Schwartz, B. (2004). The paradox of choice: Why less is more. New York, NY: Harper Collins. Selinger, E. and Frischmann, B. (2016). The dangers of Smart Communication Technology. The Arc Mag 13 September 2016 - available at https://thearcmag.com/the-danger-of-smart-communication-technology-c5d7d9dd0f3e#.3yuhicpw8 (accessed 29/11/2016). Selinger, E. (2014a). Today’s Apps are Turning us Into Sociopaths. WIRED 26 February 2014 - available at https://www.wired.com/2014/02/outsourcing-humanity-apps/ (accessed 29/11/2016). Selinger, E. (2014b). Don’t outsource your dating Life. CNN: Edition 2 May 2014 - available at http://edition.cnn.com/2014/05/01/opinion/selinger-outsourcing-activities/index.html (accessed 29/11/2016). Selinger, E. (2014c). Outsourcing Your Mind and Intelligence to Computer/Phone Apps. Institute for Ethics and Emerging Technologies 8 April 2014 - available at http://ieet.org/index.php/IEET/more/selinger20140408 (accessed 29/11/2014). Shah, A. K., Mullainathan, S., & Shafir, E. (2012). Some consequences of having too little. Science, 338, 682–685. Slamecka, N., & Graf, P. (1978). The generation effect: The delineation of a phenomenon. Journal of Experimental Psychology: Human Learning and Memory., 4(6), 592–604. Smuts, A. (2013). The good cause account of the meaning of life. Southern Philosophy Journal, 51(4), 536–562. Sunstein, C. (2016). The ethics of influence. Cambridge, UK: Cambridge University Press. Sunstein, C. (2017). # Republic: Divided democracy in an age of social media. Princeton, NJ: Princeton University Press. Thaler, R., & Sunstein, C. (2009). Nudge: Improving decisions about health, wealth and happiness. London: Penguin. Wertheimer, A. (1987). Coercion. Princeton, NJ: Princeton University Press. Whitehead, A. N. (1911). An introduction to mathematics. London: Williams and Norgate. Wu, T. (2017). The Attention Merchants. New York: Atlantica. Yeung, K. (2017). ‘Hypernudge’: Big data as a mode of regulation by design. Information, Communication and Society, 20(1), 118–136.