Academia.eduAcademia.edu

Logic, language and the brain Michiel van Lambalgen

Logic, language and the brain Michiel van Lambalgen Cognitive Science Center Amsterdam https://staff.science.uva.nl/˜michiell Aim and program • aim: explain the use of computational logic in cognitive science – the domain is language comprehension and production – show how logical modelling leads to testable predictions, both for behaviour and brain imaging – show how logical modelling connects to biological issues, e.g. neural substrate of linguistic processing, and evolutionary considerations • lecture 1: the big picture – time, tense and cognition • lecture 2: the event calculus • lecture 3: verb tense and closed world reasoning • lecture 4: predictions for EEG • lecture 5: executive function and behavioural predictions for autism and ADHD; neural network architecture Literature 1. series of articles by Baggio, vL & Hagoort (most recently in Journal of Memory and Language 2008) 2. (with Fritz Hamm) The proper treatment of events, Blackwell 2004 (linguistic details); cited as PToE 3. (with Keith Stenning), Human reasoning and cognitive science, MIT Press 2008 (evolutionary and methodological considerations); cited as HRCS The big picture – time, tense and cognition • methodology: a role for logic in cognitive science • an ancestral precursor to language: planning • planning, causality, time and tense Why logic is unpopular in cognitive science: 5 misconceptions • experiments with reasoning tasks, such as the famous Wason selection task show that logical form is not a determinant of reasoning • logic cannot deal with vagueness and graded concepts • logic cannot deal with uncertainty and must be replaced by probability theory, which is after all the calculus of uncertainty • what we know about the neocortex suggests that the computations executed by the brain are very different from logical computations • the computational complexity of logic is too high for logic to be a realistic model of cognition A novel variation: the two-rule task (S & vL) Below is depicted a set of four cards, of which you can see only the exposed face but not the hidden back. On each card, there is a 4 or 7 on one of its sides and a A or K on the other. Also below there are two rule which apply only to the four cards. It is given that exactly one rule is true. Your task is to decide which if any of these four cards you must turn in order to decide which rule is true. Don’t turn unnecessary cards. Tick the cards you want to turn. 1. if there is a A on one side, then there is an 4 on the other side 2. if there is an K on one side, then there is an 4 on the other side Cards: A K 4 7 A role for logic in cognitive science • David Marr’s hierarchy of explanation in cognitive science: a cognitive capacity can be studied at (at least) the following levels 1. abstract specification of the capacity, including input–output behaviour, in a formal language [informational level] 2. algorithm computing the input–output behaviour [algorithmic level] 3. neural implementation of the algorithm [neural level] • example: informational level description of language comprehension could run as follows • input: discourse is transformed into output: discourse model (called situation model in pyscholinguistics) • DRT is an example, but we will not adopt it because (i) it is monotonic (this conflicts with a processing principle called immediacy – see lecture 4), and (ii) the algorithm is not easily related to neural implementation, or to what is known about cognition Logic as high-level description of cognitive functions • another example: ’executive function’ • ‘executive function’ is an umbrella term for processes responsible for higher-level action control that are necessary for maintaining a goal and achieving it in possibly adverse circumstances • we may take executive function to be composed of planning, initiation, inhibition, coordination and control of action sequences, leading toward a goal held in working memory • several of these components have logical specifications • this will be illustrated by showing how executive function is involved in discourse production and comprehension Discourse • discourse is not just bunch of sentences: what is special about discourse is that it is always interpreted relative to the verbal and nonverbal context – e.g. world knowledge • experimental evidence for the reality of temporally organised discourse models – memory for narrated events not in order of narration, but in order of occurrence (‘Max fell. John pushed him.’) – events occurring at temporally contiguous times are related more closely in memory • experimental clues to the constitution of discourse models – causal information about the events of interest is represented, including default expectations: sentences of the type ‘John was beaming. A moment later . . . ’ have shorter reading times then sentences of type ‘John was beaming. An hour later . . . ’ – this is interpreted as indicating that the default continuation of the discourse model that supports the first sentence doesn’t support the second Discourse comprehension in a wider context • the construction of discourse models is not specific to language! • a discourse model consists of entities (e.g. introduced by indefinite NPs), relations between these entities, events (e.g. introduced by VPs), temporal relations between events • structures consisting of events and relations between them occur elsewhere in cognition as well • a plan is a sequence of actions (together with the times at which these actions have to be executed) which achieves a given goal • (the assumption of sequentiality can be relaxed: often actions have to be executed in parallel) • a discourse model is analogous to a plan and it may be the case that the same construction algorithm is applied Why is planning possibly important? • amplified definition: planning consists in setting a goal and proposing a sequence of actions which provably leads to the achievement of that goal, taking into account properties of, and events in, the world • (hierarchical) planning occurs in humans and in nonhuman primates • evolution of language: planning has been hypothesized to be co-opted for syntax because of ‘recursive’ structure (Lashley 1951; Greenfield 1991; Steedman 2002) • Broca’s area (involved in language) immediately adjacent to areas for motor planning (Greenfield et al. 1972) • possibility for discourse may distinguish ape language/human language • planning can organize (e.g. temporal features of) discourse so as to determine the event structure described by the discourse analogous to a structured set of actions leading to a goal • here, events are the denotata of untensed sentences and verb phrases Intermezzo: planning and time ‘Using goal-plan knowledge to merge the past with the present and the future in narrating events on line’ (Trabasso and Stein 1993) • uses Berman and Slobin’s ‘Frog where are you?’ paradigm • plans act like glue: ‘the plan unites the past (a desired state) with the present (an attempt) and the future (the attainment of that state) . . . ‘causality and planning provide the medium through which the past is glued to the present and the future’ • basic narrative unit: GAO – ‘goal-action-outcome sequence’ • the capacity to structure a narrative in GAOs develops gradually: – absent at 3 (almost no GAOs; story is basically a sequence of stills, none of the actions described relevant to the boy’s goal) – sharp increase GAOs starting age 4 until age 9 – maximum reached at age 21 The frog story: opening scenes #1 #2 #3 The frog story: a typical mishap Planning: what is the informational level description? • amplified definition: planning consists in setting a goal and proposing an ordering of actions which provably leads to the achievement of that goal, taking into account properties of, and events in, the world • one needs – facts and causal laws holding in the world, including laws about consequences and preconditions of actions – a goal – a reasoning mechanism that outputs a plan – the reasoning mechanism should not suffer from the ‘frame problem’, i.e. should not take into account all logical possibilities – this makes it non-monotonic: goal G can be expected to be attainable in circumstances C, but not anymore in C ∪ D Logical characterisation of planning • computing a plan is closed world reasoning (CWR) in a formal system for representing and reasoning with causes, actions and consequences • CWR: ’if you have no reason to expect event e, assume e will not occur’ • this principle can be viewed as the construction of a very special kind of model • toy example – a:= turn light on (‘action’ or ‘event’), r:= light is on (‘state’) – predicates: holds(x , n), do(e, n), ab(n, k ) – causal axiom: ∀n∀k(¬holds(r , n)∧do(a, n)∧k > n∧¬ab(n, k ) → holds(r , k )) – goal: holds(r , 10 ) – boundary condition: ¬holds(r , 1 ); time is interval [1, 10] The output of planning is a model with events and states • the obvious plan is do(a, 9 ) • (this works only if ¬holds(r , 9 ) and ¬ab(9 , 10 )) • the obvious plan together with the conditions for its execution determine a model defined by holds(r, n) ↔ n = 10, do(a, n) ↔ n = 9, ab(n, k) ↔ n 6= n. • this model is minimal in the sense that the extensions of the predicates are as small as possible (‘closed world model’) • there are other models, e.g. holds(r, n) ↔ n = 2, 4, . . . , 10, do(a, n) ↔ n = 1, 3, . . . , 9, ab(n, k) ↔ (n, k) = (1, 3), (3, 5), (5, 7), (7, 9). • here extensions of predicates are larger – and ‘unexplained’ Computational features of the ‘closed world model’ • CWR comes in 3 forms – CWR for rules: if one knows that q and the only rule with q as consequent is p → q, then assume q; in effect one thus replaces p → q by p ↔ q – CWR for facts: if there is no rule with q as consequent (in particular, if q is not a fact), one may assume ¬q – CWR for goals: if the given goal is to make A(c) true, then one may assume A(x) can be made false for x 6= c • CWRr : ∃n(¬holds(r , n) ∧ do(a, n) ∧ k > n ∧ ¬ab(n, k )) ↔ holds(r , k ) • CWRf : ∀k∀n¬ab(n, k) • CWRg : try to ensure that the goal is satisfied minimally, i.e. holds(r, n) ↔ n = 10 The computation • CWRr : ∃n(¬holds(r , n) ∧ do(a, n) ∧ k > n ∧ ¬ab(n, k )) ↔ holds(r , k ) • CWRf : ∀k∀n¬ab(n, k) • CWRg : try to ensure that holds(r, n) ↔ n = 10 CWRg gives ∃n(¬holds(r , n) ∧ do(a, n) ∧ k > n ∧ ¬ab(n, k )) ↔ k = 10 CWRf gets rid of ab: ∃n(¬holds(r , n) ∧ do(a, n) ∧ k > n) ↔ k = 10 and we have ∃n < 10(¬holds(r , n) ∧ do(a, n)) ∀n < 9(¬holds(r , n) → ¬do(a, n)) Again CWRg gives ∀n < 9¬do(a, n), whence do(a, 9) What we can learn from this • we specified a planning algorithm through closed world reasoning • the output is a time-based model containing actions/events and states (together with their times of occurrence) which is minimal in two senses: – the least number of actions to be performed – the least interference from outside events • such models are quite like discourse models • there is also a connection with tense/aspect What we can learn from this • consider the sentence in present perfect ‘Bill has turned on the light’ • identify time point 10 with now (utterance time) • the pres. perf. is formulated in planning terms as the instruction to satisfy the goal ‘light on now ’ (formally holds(r, 10)) • this captures ‘present relevance’ of the present perfect: reference time (given as now ) anchors the state consequent upon the event (namely r); event time itself is inferred (as do(a, 9)) • generally, tenses will be viewed as goals to be satisfied in the current discourse model • this can also be seen as an update process: first the (empty) discourse model is updated with holds(r, 10), and then, after CWR, with do(a, 9) • the discourse model for ‘Bill has turned on the light’ has elements {bill, a, r} and relations holds(r, 10), do(a, 9) A different computation • what we’ve just seen is, logically speaking, the computation of the minimal fixpoint of a logic program • the syntactic counterpart of this computation is backtracking from a given goal • ?holds(r, 10), ¬holds(r, 1), . . . , ¬holds(r, 9) reduces to • ?n < 10, ¬holds(r, n), do(a, n), ¬ab(n, 10), ¬holds(r, 1), . . . , ¬holds(r • the only unification that leads to success with this goal is n = 9; e.g. n = 1 leads to • ?do(a, 1), ¬ab(1, 10), ¬holds(r, 1), . . . , ¬holds(r, 9) • negation as failure reduces this goal to • ?do(a, 1), ¬holds(r, 1), . . . , ¬holds(r, 9) • but one cannot make ¬holds(r, 2) succeed, i.e. one cannot show that the derivation from the goal ?holds(r, 2) fails – indeed this derivation succeeds • the same argument works for all n ≤ 8 Language as planned action: the bigger picture • preceding considerations will lead to detailed computational model of tense/aspect processing, with predictions for brain imaging (see lecture 4) • the claim is that the computations identified are what is ‘really’ going on – at least ‘really’ in the sense of Marr’s informational and algorithmic level • there is also a suggestion for a biologically plausible neural implementation: discourse models as stable states (lectures 4, 5) • language is (externalised!) planned action in the sense that (i) discourse production results in a series of goals to be satisfied by a plan, and (ii) discourse comprehension involves the computation of a plan, which can be identified with a discourse model • importance of flexibility of planning in organising discourse – clue to origin (evodevo; HCRS)