Can AI be trustworthy?
DOI:
https://doi.org/10.18012/arf.v11iEspecial.69910Palavras-chave:
Artificial Intelligence, Distrust, Human replacement, Ethics of AI, ExplicabilityResumo
In this paper, I address why some people distrust of Artificial Intelligence, and how this discipline can be trustworthy. Specifically, I describe the origins of such distrust, make a prognosis of the current state of uncertainty it provokes, and offer a key for a trustworthy AI. In the first section, I deal with why machines have not been considered trustworthy, and, more importantly, with the core of distrust of AI: as I contend here, there has been a shift from an academic debate about AI minds, due to skepticism about other minds, to fears on day-to-day life, mainly due to concerns of human replacement. In the second section, I make a prognosis about the current state of distrust AI provokes. Finally, I offer a key for a trustworthy AI based on explicability. As I argue, explicability via experts who have been certified by institutions are facilitators of a trustworthy AI.
Downloads
Referências
ANDERSON, C.; SINGER, M. The sensitive left and the impervious right: multilevel models. Comparative Political Studies, v. 41, p. 564-599, 2008.
BABBAGE, C. Excerpt from ‘Passages from the Life of a Philosopher’. In: BABBAGE, C. Babbage’s calculating engines: being a collection of papers relating to them; their history, and construction. Edited by Henry P. Babbage.Cambridge: Cambridge University Press, 2010. p. 83-288.
BAIER, A. Trust: the Tanner lectures of human values, 1991. <https://tannerlectures.utah.edu/_documents/a-to-z/b/baier92.pdf>.
BAROTTA, P.; GRONDA, R. Epistemic inequality and the grounds of trust in scientific experts. In: FABRIS, A. (Ed.) Trust: a philosophical approach. Cham: Springer, 2020. p. 81-94.
BEAUCHAMP, T.; CHILDRESS J. Principles of biomedical ethics. Oxford: Oxford University Press, 2021.
BIRD, A.; EMMA, T. Natural kinds._In: ZALTA, E.; NODELMAN, U. (Ed.) The Stanford Encyclopedia of Philosophy, 2023. <https://plato.stanford.edu/archives/spr2023/entries/natural-kinds>.
BROWN, M.; ROBINSON, L.; CAMPIONE, G. C.; WUENSCH, K.; HILDEBRANDT, T.; MICALI, N. Intolerance of uncertainty in eating disorders: a systematic review and metaanalysis. European Eating Disorders Review, v. 25, p. 329-343, 2017.
CARLETON, R. N. Fear of the unknown: one fear to rule them all? Journal of Anxiety Disorders,v. 41, p. 5-21, 2016.
COPELAND, J. Artificial intelligence: a philosophical introduction. Malden, MA: Blackwell, 2001.
COWLS, J.; FLORIDI, L.: TADEO, M. The challenges and opportunities of ethical AI. Artificially Intelligent, 2018. <https://digitransglasgow.github.io/ArtificiallyIntelligent/contributions/04_Alan_Turing_Institute.html>
DESCARTES, R. Oeuvres. Edited by Charles Adam and Paul Tannery, new edn, edited by the CNRS, 11 vols. Paris: Vrin, 1974-1976. [all Descartes’ works are cited as AT in this paper]
FLORIDI, L. Introduction: The importance of an ethics-first approach to the development of AI. In: FLORIDI, L. Ethics, governance, and policies in artificial intelligence. Cham: Springer, 2021a. p. 1-4.
FLORIDI, L. A unified framework of five principles for AI in society. In: FLORIDI, L. Ethics, governance, and policies in Artificial Intelligence. Cham: Springer, 2021b. p. 5-17.
FREDERIKSEN, M.; LARSEN, C; H. LOLLE. Education and trust: exploring the association across social relationships and nations. Acta Sociologica v. 59, p. 293-308, 2016.KLEIN, N. AI machines aren’t ‘hallucinating’. But their makers are. The Guardian, May 8, 2023. available at: <https://www.theguardian.com/commentisfree/2023/may/08/ai-machines-hallucinating-naomi-klein>.
HAKHVERDIAN, A.; MAYNE SOURCE, Q. Institutional trust, education, and corruption: a micro-macro interactive approach. The Journal of Politics v. 74, p. 739-750, 2012.
HARDIN, R. Trust and trustworthiness. New York: Russel Sage Foundation, 2002.
HARDWIG, R. The role of trust in knowledge. The Journal of Philosophy, v. 88, p. 693-708, 1991.
HENDRIKS, F., KIENHUES, D.: R. BROMME. Trust in science and the science of trust. In: BLÖBAUM, B. (Ed.) Trust and communication in a digitalized world. Cham: Springer, 2016. p. 143-160.
MCLEOD, C. Trust. In: ZALTA, E.; NODELMAN, U. (Ed.) The Stanford Encyclopedia of Philosophy, 2021. <https://plato.stanford.edu/archives/fall2021/entries/trust/>.
MURRAY, S. The frame problem. In: ZALTA, E. (Ed.) The Stanford Encyclopedia of Philosophy, 2016. <https://plato.stanford.edu/archives/spr2016/entries/frame-problem/>.
ORIGGI, G. Is trust an epistemological notion? Episteme v. 1, p. 61-72, 2004.
PUTNAM, H. The meaning of ‘meaning’._Minnesota Studies in the Philosophy of Science, v. 7, p. 215-271, 1975.
SAYGIN, A. P.; CICKELI, I.; AKMAN V. Turing Test: 50 years later. In: MOOR, J. (Ed.) The Turing test: the elusive standard of artificial intelligence. Dordrecht: Kluwer Academic Publishers, 2003. p. 23-78.
SEARLE, J. Making the social world: the structure of human civilization. Oxford: Oxford University Press, 2010.
SHANAHAN, M. The frame problem. In: ZALTA, E. (Ed.) The Stanford Encyclopedia of Philosophy, 2016. <https://plato.stanford.edu/archives/spr2016/entries/frame-problem>.
SHIHATA, S.; MCEVOY, P; MULLAN, B.; CARLETON, R. Intolerance of uncertainty in emotional disorders: What uncertainties remain? Journal of Anxiety Disorders, v. 41, p. 115-124, 2016.
SMITH, B. The promise of Artificial Intelligence: reckoning and judgement. Cambridge, MA: The MIT Press, 2019.
SMITH, B.; LODDO, O. G.; LORINI, G. On credentials. Journal of Social Ontology, v. 6, p. 47-67, 2020. SWADE, D. The difference engine: Charles Babbage and the quest to build the first computer. London: Penguin Books, 2002.
TANOVIC, E., GEE, D. G.; JOORMANN, J. Intolerance of uncertainty: neural and psychophysiological correlates of the perception of uncertainty as threatening. Clinical Psychology Review, n. 60, p. 87-99, 2018.
TENNIE, C.; CALL, J.; TOMASELLO, M. Ratchering up the ratchet: on the evolution of cumulative culture. Philosophical Transactions of the Royal Society B: Biological Sciences, v.364, p.2405-2415, 2009. <http://doi.org/10.1098/rstb.2009.0052>.
TOMASELLO, M. Why we cooperate. Cambridge, MA: The MIT Press, 2009.
TOMASELLO, M. Being human: a theory of ontogeny. Cambridge, MA: Harvard University Press, 2019.
TURING, A. Computing intelligence and machinery. Mind,n. 59, p. 433-460, 1950.
TURING, A. Can digital computers think? In: SHIEBER, S. (Ed.) The Turing test: verbal behavior as the hallmark of intelligence. Cambridge, MA: The MIT Press, 1951. p. 111-116.
TURING, A.; BRAITHWAITE, R.; JEFFERSON, G.; M. NEWMAN. Can automatic calculating machines be said to think? In: COPELAND, J. (Ed.) The essential Turing: seminal writings in computing, logic, philosophy, artificial intelligence, and artificial life plus the secrets of enigma. Oxford: Oxford University Press, 1952. p. 487-506.
WEIZENBAUM, J.. Computer power and human reason: from judgement to calculation. San Francisco: W. H. Freeman & Company, 1976.
WILLIAMS, D. Yoval Noah Harari argues that AI has hacked the operating system of human civilization, 2023. <https://www.economist.com/by-invitation/2023/04/28/yuval-noah-harari-argues-that-ai-has-hacked-the-operating-system-of-human-civilisation>.
Arquivos adicionais
Publicado
Como Citar
Edição
Seção
Licença
Este trabalho está licenciado sob uma licença Creative Commons Attribution 4.0 International License.
Política de Direito Autoral para os itens publicados pela Revista:
1.Esta revista é regida por uma Licença da Creative Commons aplicada a revistas eletrônicas. Esta licença pode ser lida no link a seguir: Creative Commons Attribution 4.0 International (CC BY 4.0).
2.Consonante a essa politica, a revista declara que os autores são os detentores do copyright de seus artigos sem restrição, e podem depositar o pós-print de seus artigos em qualquer repositório ou site.
Política de Direito de Uso dos Metadados para informações contidas nos itens do repositório
1. Qualquer pessoa e/ou empresa pode acessar os metadados dos itens publicados gratuitamente e a qulquer tempo.
2.Os metadados podem ser usados sem licença prévia em qualquer meio, mesmo comercialmente, desde que seja oferecido um link para o OAI Identifier ou para o artigo que ele desceve, sob os termos da licença CC BY aplicada à revista.
Os autores que têm seus trabalhos publicados concordam que com todas as declarações e normas da Revista e assumem inteira responsabilidade pelas informações prestadas e ideias veiculadas em seus artigos, em conformidade com a Política de Boas Práticas da Revista.