Academia.eduAcademia.edu

Outline

A Note on Artificial Intelligence

Abstract

Note 4 A Note on Artificial Intelligence and the critical recursive implementation: The lagging problem of 'background knowledge' 1.

Key takeaways
sparkles

AI

  1. The text critiques AI's reliance on recurrent models and advocates for recursive processing to mimic human cognition.
  2. Cognitive scientists are divided into AI-soft, favoring data-driven models, and AI-hard, emphasizing inner theories of mind.
  3. Judea Pearl's theory distinguishes between probabilistic and causal reasoning, highlighting the need for asymmetrical language.
  4. The dual mechanism model proposes that true AI requires both associative learning and recursive structures for effective processing.
  5. AI systems face challenges like blending and catastrophic interference, necessitating advanced architectures for human-like understanding.

References (42)

  1. Q)uestion put to our AI-OS : Who is the first president of the United States? So, there are two boxes: box-1 = three (P)eople (P1,2,3) (box-1 = time-step-1 of input for box-2) 19 box-2 = AI (with DME 2 -recursion
  2. Box-1: P1 and P2 think they know the answer to Q, but P3 is unsure. As each of the three people in box-1 (outside the computer, say a room) enters into box-2 (the computer) to deliver his input/answer, a scenario unfolds: the third person (P3) is unsure of the answer and so waits until P2 exits box-2 and returns to box-1 (imagine P2 and P3 bump into one another, and P3 whispers to ask what the (R)esponse is to Q: P2 says incorrectly (as P2's grammar is erroneously operating under an 'alternating response' procedural grammar) that R=John Adams (erroneously generated from an mistaken A n B m grammar):
  3. A [(George Washington), [John Adams 1 [George Washington , John Adams]]] B] P3 now enters into Box-2 to deliver his R-input 'John Adams'. At this time- step, the AI (box-2) finds the net-value R to Q (2 over 1 out-weight responses) and moves on. So when P3 returns to ask google (AI) a follow-up Q 'Who the first president of the United States?' the Response from a 2-1 weighted input-to-output system is 'John Adams' and so P3 feels reassured of the 'correct' response and assumed 'knowledge' of Q.
  4. Recall, that for box-1, this constitutes (G)eneration-1 of input-virgin input which would be used for the first time to deliver an approximate result. Of course, over many thousands of generations, any input/answer may change over time until it reaches stabilization. Usually, such G-1 (as a starting point) is either soft-wared/ programmed into the OS (top- down) as initial default settings, or that some predesigned architectural template is built-in (innately) to capture tailored types of weights and distributions of the specified incoming- data stream. [12.1] Correct: Let's consider how a "center-embedded" structure would look like for the R given by P1: [A: who is the first US Pres? [AB list of all US presidents in order] B: George Washington] [A: Question [AB n = George Washington, John Adams, Thomas Jefferson…..etc.] B: Response] (where variable {n} = repeat list as you know it) [12.2] Incorrect: Let's consider how a "center-embedded" structure would look like for the R given by P2: [A: who is the first US Pres? [AB list of all US presidents in order] B: John Adams] [A: Question [AB m= John Adams, George Washington, __Thomas Jefferson.etc.] B: Response] (where variable {m} = alternate list as you know it)
  5. Where AB m = [John Adams, George Washington, John Adams, Thomas Jefferson…..etc.]
  6. Recall, what we want from a self-learning AI-OS ability is two-fold:
  7. To identify that the nature of the error is actually encoded in the center- embedded structure-to realize that such encoding can actually be examined by the periphery [a,b,c] in determining the nature of P2's error (that [AB m ]-procedural grammar was wrongly put in operation rather than the correct [AB n ]-grammar).
  8. Abler, W.L. (1989). On the particulate principle of self-diversifying systems. J. of Social and Biological Structures, 12, 1-13.
  9. Baumgartner, P. & Sabine Payr (Eds. 1995). Speaking Minds: Interviews with twenty eminent cognitive scientists. Princeton University Press.
  10. Bickerton, D. (2010). 'On two incompatible theories of language evolution' (chapter 14) In Larson, R., V. Déprez., & H. Yamalido (eds). The Evolution of Human Language: A Biolinguistic Perspective. Cambridge University Press.
  11. Chomsky, N. (1956). Three models for the description of language. IRE Transactions on Information Theory. (MIT). (1957). Syntactic Structures. Mouton & co. The Hague. (1958). Review of B.F. Skinner, 'Verbal Behavior'. Language, Vol 35. pp.26-58. (1966). Cartesian Linguistics: A chapter in the history of rationalist thought. University Press of Amercia. (2002). On Nature and Langauge. Cambridge University Press.
  12. Clahsen, H. (1999). Lexical entries and rules of language. A multidisciplinary study of German inflection. Behavioral and Brain Science, 22. 991-1060. (Target article).
  13. Clahsen, H., G. Marcus., S. Bartke & R. Wiese (1995). Compounding and inflection in German child language. In G. Booij and J. van Marle (eds) Yearbook of Morphology. Kluwer.
  14. Fitch, W. (2010). 'Three meanings of "recursion": Key distinctions for biolinguists (Chapter 4) In Larson, R., V. Déprez, H. Yamakido (eds). The evolution of Human Language.
  15. Dreyfus, H & S. Dreyfus (1986). Mind over Machine. The Free Press: NY.
  16. Galasso, J. (2016). From Merge to Move: A Minimalist Perspective on the design of language and its role in early child syntax. LINGCOM Studies in Theoretical Linguistics, 59.
  17. Goldschmidt, R. (1940/1982). The material Basis of Evolution. Yale University Press.
  18. Gordon, P, (1985). Level-ordering in lexical development. Cognition, 21. 73-98. (See 'Rat-eater' experiment).
  19. Gould, S.J. (2007). Punctuated Equilibrium. Harvard University Press.
  20. Hadley, R.F. (2000). Cognition and the computational power of connectionist networks. Connection Science, 12.
  21. Hebb, D. (1949). Organization of Behavior. New York: Wiley.
  22. Marcus, G. (2001). The Algebraic Mind: Integrating Connectionism and Cognitive Science. MIT Press. (2017). Artificial Intelligence is Stuck. Here's how to move if forward. New York Times, July 29. (2018). The deepest problem with deep learning. Article reprinted on TheAtlantic.com. https://medium.com/@GaryMarcus/the-deepest-problem-with-deep-learning- 91c5991f5695 For summary of Marcus v Elman debates, see http://psych.nyu.edu/marcus/TAM/author_response.html
  23. Marcus, G., S. Pinker, M. Ullman, J. Hollander, T. Rosen & F. Xu (1992). Over-regularization in language acquisition. Monographs of the Society for Research in Child Development. 57.
  24. Marcus, G., U, Brinkmann, H. Clahsen, R. Wiese & S. Pinker (1995). German Inflection: The exception to the rule. Cognitive Psychology, 29. 189-256.
  25. Miller, G. (1955). The magic number seven, plus or minus two: Some limits on our capacity for processing information. Published in 1956, Psychological Review 63: 81-97. (1956). Human memory and the storage of information. IRE Transactions on Information Theory. (MIT).
  26. Minsky, M. & S. Papert (1969). Perceptrons: An Introduction to Computational Geometry. MIT Press.
  27. Newell, A. & H. Simon (1956). The logic theory machine: a complex information processing system. IRE Transactions on Information Theory. (MIT).
  28. 'Parallel Distributional Processing' Research Group. UC San Diego.
  29. Pearl, Judae (2018). Theoretical Impediments to Machine Learning with Seven Sparks from the Causal Revolution. Technical Report R-475, July 2018. (2018). The Book of Why: The new science of cause and effect. Basic books.
  30. Pinker, S. (1999). Words and Rules. Basic Books.
  31. Rosenblatt, F. (1959). Two theorems of statistical separability in the perceptron. Ms. Proceedings of a symposium on the mechanism of thought processes. London.
  32. Tomasello, M. & J. Call (1997). Primate Cognition. Oxford Press.
  33. Toulmin, S. (1961). Forecast and Understanding. University Press, Indiana.
  34. Wexler, K. (2003). 'Lenneberg's Dream' (Chapter 1, pp 11-61) In Levy, Y. & J. Schaeffer (eds). Language Competence across Populations. Mahwah, Erlbaum.
  35. Newell, A. & H. Simon (1956). The logic theory machine: a complex information processing system. IRE Transactions on Information Theory. (MIT).
  36. 'Parallel Distributional Processing' Research Group. UC San Diego.
  37. Pearl, Judae (2018). Theoretical Impediments to Machine Learning with Seven Sparks from the Causal Revolution. Technical Report R-475, July 2018. (2018). The Book of Why: The new science of cause and effect. Basic books.
  38. Pinker, S. (1999). Words and Rules. Basic Books.
  39. Rosenblatt, F. (1959). Two theorems of statistical separability in the perceptron. Ms. Proceedings of a symposium on the mechanism of thought processes. London.
  40. Tomasello, M. & J. Call (1997). Primate Cognition. Oxford Press.
  41. Toulmin, S. (1961). Forecast and Understanding. University Press, Indiana.
  42. Wexler, K. (2003). 'Lenneberg's Dream' (Chapter 1, pp 11-61) In Levy, Y. & J. Schaeffer (eds). Language Competence across Populations. Mahwah, Erlbaum.