Academia.eduAcademia.edu

Outline

The State Space of Artificial Intelligence

2020, Minds and Machines

https://doi.org/10.1007/S11023-020-09538-3

Abstract

The goal of the paper is to develop and propose a general model of the state space of AI. Given the breathtaking progress in AI research and technologies in recent years, such conceptual work is of substantial theoretical interest. The present AI hype is mainly driven by the triumph of deep learning neural networks. As the distinguishing feature of such networks is the ability to self-learn, self-learning is identified as one important dimension of the AI state space. Another dimension is recognized as generalization, the possibility to go over from specific to more general types of problems. A third dimension is semantic grounding. Our overall analysis connects to a number of known foundational issues in the philosophy of mind and cognition: the blockhead objection, the Turing test, the symbol grounding problem, the Chinese room argument, and use theories of meaning. It shall finally be argued that the dimension of grounding decomposes into three sub-dimensions. And the dimension o...

References (48)

  1. Bengio, Yoshua, Dong-Hyun Lee, Jörg Bornschein, Thomas Mesnard & Zhouhan Lin (2016): Towards Biologically Plausible Deep Learning. arXiv :1502.04156 v3.
  2. Block, Ned (1981). Psychologism and Behaviorism. Philosophical Review, 90(1), 5-43.
  3. Block, Ned (1998): Semantics, conceptual role. In The Routledge Encylopedia of Philosophy, ed. E. Craig. London: Routledge.
  4. Bostrom, Nick (2013). Superintelligence. Paths: Oxford University Press.
  5. Botvinick, Matthew M., Ritter, Sam, Wang, Jane X., Kurth-Nelson, Zeb, & Hassabis, Demis (2019). Reinforcement Learning, Fast and Slow. Trends Cognitive Sci, 23(5), 408-422.
  6. Brockman, John, editor (2019): Possible Minds. 25 Ways of Looking at AI. Penguin Press.
  7. Buckner, Cameron (2018). Empiricism without Magic: Transformational Abstraction in Deep Convolu- tional Neural Networks. Synthese, 195, 5339-5372.
  8. Buckner, Cameron (2019). Deep learning: a philosophical introduction. Philosophy Compass, 2019, e12625.
  9. Chomsky, Noam (1980). Rules and Representations. Behavioral and Brain Sciences, 3(127), 1-61.
  10. Cummins, Robert C. (1996). Representations, Targets, and Attitudes. Cambridge: MIT Press.
  11. Fodor, Jerry (1987): Psychosemantics. MIT Press.
  12. Ford, Martin, editor (2018): Architects of Intelligence: The truth about AI from the people building it. Packt Publishing.
  13. Goodfellow, Ian, Yoshua Bengio & Aaron Courville (2016): Deep Learning. MIT Press.
  14. Harnad, Stevan (1989). Minds, Machines and Searle. J Theoretical Exp Artifi Intell, 1, 5-25.
  15. Harnad, Stevan (1990). The Symbol Grounding Problem. Physica D: Nonlinear Phenomena, 42, 335-346.
  16. Harnad, Stevan (2001): What's Wrong and Right About Searle's Chinese Room Argument? In M. Bishop & J. Preston (eds.): Essays on Searle's Chinese Room Argument. Oxford University Press.
  17. Hassabis, Demis, Kumaran, Dharshan, Summerfield, Christopher, & Botvinick, Matthew (2017). Neuro- science-Inspired Artificial Intelligence. Neuron, 95, 245-258.
  18. Hinton, Geoffrey E., & Salakhutdinov, Ruslan R. (2006). Reducing the dimensionality of data with neural networks. Science, 313, 504-507.
  19. Hsu, Feng-hsiung (2002): Behind Deep Blue: Building the Computer that Defeated the World Chess Champion. Princeton University Press.
  20. Kripke, Saul A. (1982): Wittgenstein on Rules and Private Language. Harvard University Press.
  21. Krizhevsky, Alex, Ilya Sutskever & Geoffrey E. Hinton (2012): ImageNet classification with deep convo- lutional neural networks. Advances in Neural Information Processing Systems 25 (NIPS 2012), Vol. 1: 1097-1105.
  22. LeCun, Yann, Bengio, Yoshua, & Hinton, Geoffrey E. (2015). Deep learning. Nature, 521, 436-444.
  23. López-Rubio, Ezequiel. (2018). Computational functionalism for the deep learning era. Minds Machines, 28, 667-688.
  24. Lyre, Holger (2016). Active content externalism. Rev Philos Psychol, 7(1), 17-33.
  25. Lyre, Holger (2010): Humean Perspectives on Structural Realism. In: F. Stadler (ed.): The Present Situa- tion in the Philosophy of Science. Springer, p. 381-397.
  26. Millikan, Ruth (1984). Language, Thought and Other Biological Categories. Cambridge: MIT Press.
  27. Mnih, Volodymyr, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra & Martin Riedmiller (2013): Playing Atari with Deep Reinforcement Learning. arXiv :1312.5602.
  28. Müller-Schloer, Christian & Sven Tomforde (2017): Organic Computing-Technical Systems for Survival in the Real World. Birkhäuser.
  29. Páez, Andrés (2019). The Pragmatic Turn in Explainable Artificial Intelligence (XAI). Minds Mach, 29, 441-459.
  30. Ramsey, William (2007). Representation Reconsidered. Cambridge: Cambridge University Press.
  31. Robbins, Philip & Murat Aydede, editors (2009): The Cambridge Handbook of Situated Cognition. Cam- bridge University Press.
  32. Schaul, Tom, & Schmidhuber, Jürgen (2010). Metalearning. Scholarpedia, 5(6), 4650.
  33. Schmidhuber, Jürgen (2015a). Deep Learning in Neural Networks: an Overview. Neural Networks, 61, 85-117.
  34. Schmidhuber, Jürgen (2015b). Deep Learning. Scholarpedia, 10(11), 32832. The State Space of Artificial Intelligence Schubbach, Arno (2019): Judging Machines. Philosophical Aspects of Deep Learning. Synthese. https :// doi.org/10.1007/s1122 9-019-02167 -z.
  35. Searle, John R. (1980). Minds, brains and programs. Behavioral Brain Sci, 3, 417-457.
  36. Searle, John R. (1990). Is the Brain's Mind a Computer Program? Sci Am, 1, 26-31.
  37. Sejnowski, Terrence J (2018): The Deep Learning Revolution. MIT Press.
  38. Shea, Nicolas (2018). Representation in Cognitive Science. Oxford: Oxford University Press.
  39. Siegelmann, H. T., & Sontag, E. D. (1995). On the computational power of neural nets. J Comput Syst Sci, 50(1), 132-150.
  40. Silver, David, et al. (2016). Mastering the game of Go with deep neural networks and tree search. Nature, 529, 484-489.
  41. Silver, David, et al. (2017). Mastering the game of Go without human knowledge. Nature, 550, 354-359.
  42. Sutton, Richard & Andrew Barto (2018): Reinforcement Learning: An Introduction. 2nd edition. MIT Press.
  43. Taddeo, Mariarosaria, & Floridi, Luciano (2005). Solving the symbol grounding problem: a critical review of fifteen years of research. J Experimental Theoretical Artifi Intell, 17(4), 419-445.
  44. Tegmark, Max (2017): Life 3.0: Being Human in the Age of Artificial Intelligence. Allen Lane. Turing, Alan (1950). Computing machinery and intelligence. Mind, 49, 433-460.
  45. Ullman, Shimon (2019). Using neuroscience to develop artificial intelligence. Science, 363(6428), 692-693.
  46. Weizenbaum, Joseph (1976). Computer Power and Human Reason. From Judgement to Calculation. W. H: Freeman.
  47. Wittgenstein, Ludwig (1953): Philosophical investigations. Macmillan Publishing Company. Wittgenstein, Ludwig (1956): Remarks on the foundations of mathematics. Blackwell.
  48. Würtz, Rolf P., editor (2008): Organic Computing (Understanding Complex Systems). Springer. Zednik, Carlos (2019). Solving the black box problem: a normative framework for explainable artificial intelligence. Philos Technol. https ://doi.org/10.1007/s1334 7-019-00382 -7.