• No results found

PP 2019 26: Towards a logical formalisation of Theory of Mind: a study on False Belief Tasks

N/A
N/A
Protected

Academic year: 2020

Share "PP 2019 26: Towards a logical formalisation of Theory of Mind: a study on False Belief Tasks"

Copied!
14
0
0

Loading.... (view fulltext now)

Full text

(1)

Towards a logical formalisation of Theory of Mind:

1

a study on False Belief Tasks

2

Anthia Solaki and Fernando R. Vel´azquez-Quesada 3

Institute for Logic, Language and Computation, Universiteit van Amsterdam.

4

a.solaki2@uva.nl,F.R.VelazquezQuesada@uva.nl

5

Abstract. Theory of Mind, the cognitive capacity to attribute internal

men-6

tal states to oneself and others, is a crucial component of social skills. Its

7

formal study has become important, witness recent research on reasoning

8

and information update by intelligent agents, and some proposals for its

9

formal modelling have put forward settings based on Epistemic Logic

10

(EL). Still, due to intrinsic idealisations, it is questionable whetherELcan

11

be used to model the high-order cognition of ‘real’ agents. This manuscript

12

proposes a mental attribution modelling logical framework that is more

13

in-line with findings in cognitive science. We introduce the setting and

14

some of its technical features, and argue why it does justice to empirical

15

observations, using it for modelling well-known False-Belief Tasks.

16

Keywords: theory of mind, mental state attribution, false belief tasks, temporal 17

model, dynamic epistemic logic 18

1

Introduction

19

An important feature of how people function in social scenarios is that ofTheory 20

of Mind(ToM), the cognitive capacity to attribute internal mental states, such as 21

knowledge and beliefs, to oneself and others [1].1Theory of Mind is a crucial

22

component of social skills: someone who understands that others might have 23

mental states different from hers, and can reason about those states, is much 24

better suited to understand their behaviour, and thus act and react appropriately. 25

Theory of Mind is slowly developed in the course of our lives [3,4] (and 26

at different speed for different types of persons [5,6]), starting with the ability 27

to make first-order attributions (e.g., someone knowing/believing that “Mary 28

believes that the ball is in the bag”) and progressing through attributions of second-29

ordermental states (e.g., someone knowing/believing that “Mary believes that 30

John believes that the ball is in the closet”). When testing one’s ToM, an extensively 31

used experiment is theSally-AnneFalse-Belief Task. 32

Example1 (TheSally-Anne(SA)task) The following is adapted from [3].

33

1There has been a debate on how this understanding of others’ mental states is achieved

(2)

Sally and Anne are in a room in which there are a basket and a box. Sally is holding a marble. Then, after putting the marble into the basket, Sally leaves the room. While Sally is away, Anne transfers the marble to the box. Then Sally comes back. 34

To pass the test, the subject should answer correctly the question“where does 35

Sally believe the marble to be?”. This requires for the subject to distinguish between 36

her own true belief (“the marble is in the box”) and Sally’sfalsebelief (“the marble 37

is in the basket”). Experiments have shown that, while children older than 4 38

years old tend to answer correctly, younger children (or children on the autism 39

spectrum) tend to fail the test, reporting their own belief [3]. (But see [7].) J 40

In the enterprise of studying and understanding ToM, there has been a 41

growing interest on the use of formal frameworks. A seemingly natural choice 42

is Epistemic Logic (EL) [8,9], as it provides tools for representing not only the 43

knowledge/beliefs agents have about ontic facts, but also the knowledge/beliefs 44

they have about their own and others’ knowledge/beliefs. However, usingEL 45

has some drawbacks. First, withinEL’s standard relational ‘Kripke’ semantics, 46

knowledge/beliefs are closed under logical consequence (thelogical omniscience 47

problem; [10]). Moreover, the extra relational requirements for ‘faithful’ repre-48

sentations of knowledge and beliefs turn them into S5 and KD45 modal logics, 49

respectively, thus yielding fully (positive and negative) introspective agents. 50

There is an even more fundamental reason whyELmight not be well-suited 51

for representing realistic high-order attributions. Semantically, both knowledge 52

and beliefs correspond to a universal quantification (φis known/believed iffit is 53

the case inallthe alternatives the agent considers possible); still, for real agents, 54

these notions involve more elaborate considerations (e.g., observation, commu-55

nication, reasoning). This ‘simple’ universal quantification works becauseEL 56

uses a loaded model, which contains not only the (maximally consistent) alter-57

natives the agent considers possible, but also every other alternativeevery other 58

agentconsiders possible.2In a few words, the semantic interpretation of

(high-59

order) knowledge/beliefs formulas is simple because the model is complex. Real 60

agents might not be able to have such a loaded structure ‘in their mind’, and 61

thus it is questionable whether the use of traditionalELcan provide a proper 62

picture of the way real agents deal with mental attribution scenarios. 63

In light of these issues, one could even wonder whether it makes sense to 64

use logical tools for dealing with results of empirical research. Indeed, it has 65

been argued that psychological experiments and logic are essentially different3,

66

understanding the former as the study of empirical findings on the behaviour 67

of real ‘fallible’ agents, and the latter as a normative discipline studying what 68

‘rational’ agentsshoulddo. However, other authors (e.g., [14,15]) have justified 69

why bridging these two views is a worthwhile endeavour that also has promis-70

ing applications (especially on reasoning and information update by intelligent 71

2Frameworks for representing acts of private communication [11] make this clear. Their

additional structures,action models, have one ‘event’ for each different perspective the agents might have about the communication, and the model after the communication contains roughly one copy of the original model for each one of these perspectives.

(3)

agents). Indeed, empirical research benefits from using formal tools to explain 72

their discoveries and understand their consequences, and logical frameworks 73

become richer and more ‘useful’ when they capture human limitations and 74

prescribe behaviour attainable by real agents. 75

This work seeks a ToM’s logical setting that is more in-line with the findings 76

in cognitive science, with non-trivial and competent agents whose underlying 77

reasoning is reflected in the syntax and semantics.4 To that end, we aim at the

78

converse direction to that ofEL. Our structures are simple, encoding only basic 79

facts, and thus resembling the ‘frugal’ way real agents keep information stored. 80

However, interpretations of mental state attributions show that agents engage 81

in the, oftentimes strenuous, process of recalling these facts and deriving further 82

information on their basis. 83

Outline The text is organised as follows.Section 2introduces thetemporal visi-84

bilityframework, presenting its model and formal language, and also discussing 85

some of its technical aspects. Then,Section 3relates the features of the setting 86

with findings in the cognitive science literature, using it to model well-known 87

mental attribution tasks in detail, and comparing it with other related formal 88

settings.Section 4closes, recapitulating the highlights, discussing ways in which 89

the framework can be extended, and suggesting lines for further research. 90

2

Visibility in a temporal setting

91

In most mental attribution tasks, beliefs5are, at their lower (ontic) order, about

92

the location of certain objects (e.g., the marble’s location in the Sally-Anne Task). 93

We do take objects as the main entities about which agents have mental attitudes; 94

still, for simplicity, we will work with these objects’colours. LetA,∅be the set 95

of agents (a,b, . . .), andO ,∅be the set of objects (o,p,q, . . .). For eacho ∈ O, 96

the setRocontains the colours the object might have; defineRO:=So∈ORo. The 97

model is a temporal structure, with each stage (state) fully described by both the 98

colour of each object and the objects and agents each agent sees. 99

Definition2.1 (Temporal visibility model) Atemporal visibility (TV) modelis 100

a tuplehn,S, τ, κ, νiwith(i)nNthe index of the ‘most recent’ (current) stage;

101

(ii)S a finite set of states with|S| = n;(iii)τ : S → {1..n}the temporal index

102

(bijective) function, indicating the temporal indexτ(s)∈ {1..n}of each statesS;

103

(iv)κ:S→(ORO) thecolouringfunction, withκ(s,o) (abbreviated asκs(o))

104

the colour object ohas at state s;6 (v) ν : S (A ℘(AO)) the visibility

105

function, withν(s,a) (abbreviated asνs(a)) theentities(agents and objects) agent 106

asees at states.7Given aTVmodel, lets

last ∈Sbe its (unique) state satisfying 107

τ(slast)=n. J

108

4In particular, one goal is to find a system that provides a plausible answer on why

people find mental attribution tasks increasingly difficult as their order increases.

5Following the common parlance in the literature describing the tasks we later model,

the termbelief will be used for referring to an agent’s mental state.

6Each object has a proper colour:κ

s(o)∈Roholds for alls∈Sando∈O. 7Every agent can see herself in every state:aν

(4)

Example2 Take theSally-Anne Task, with Sally (Sa), Anne (An) and the marble 109

(mar). Consider a two-state model M with (i) s1 the initial state, where both

110

agents see all agents and objects (νs1(Sa)=νs1(An)={Sa,An,mar}) and the object

111

is black (κs1(mar)=black, read as ‘the marble is in Sally’s hands’), and(ii)s2the

112

‘next’ state, where both agents still see everything, but now the object is white 113

(κs2(mar)=white, read as ‘the marble is in the basket’). The model is depicted as

114

s1

ν(Sa)=ν(An)={Sa,An,mar} mar

l

Sa

m

An

s2

ν(Sa)=ν(An)={Sa,An,mar} mar

l

Sa

m

An

J 115

Representing actions ATVmodel contains not only a state representing the 116

current situation (the stateτ−1(n)) but also states indicating how the situation

117

was in the past (up to the initialτ−1(1)). One can provide operations thatextend

118

the current model with a state depicting the outcome of a certain activity (the 119

way the situationwillbe). In theSally-Anne Task, some acts modify the colour 120

of objects (Sally puts the marble into the basket) and some others modify the 121

agents’ visibility (Sally leaves the room). Here are operations for them. 122

Definition2.2 (Colour change) Let M = hn,S, τ, κ, νi be aTV model, with snew <S; take a set of objects{p1, . . . ,pk} ⊆ O, withci ∈ Rpi a proper colour for

eachpi. The colour assignment [p1:=c1, . . . ,pk:=ck] produces theTVmodel

M[p1:=c1,...,pk:=ck]=hn+1,S∪ {snew}, τ

0, κ0, ν0i

in which(i)τ0

preserves the temporal position of states inS, makingsnew the 123

most recent (soτ0

(s) :=τ(s) fors∈ S, andτ0

(snew) := n+1);(ii)κ0is exactly as 124

κfor states inS, with the newsnew taking the colouring ofslastfor objects not 125

mentioned by the assignment, and following the assignment for the colour of 126

the objects it mentions (so, for anyo∈ O, defineκ0

s(o) := κs(o) fors∈ S, with 127

κ0

snew(o) := κslast(o) wheno < {p1, . . . ,pk}, andκ

0

snew(pj) := cj wheno= pj);(iii)ν

0 128

preserves the visibility assignment for states inS, with visibility insnewexactly 129

as inslast(so, for anya∈A, defineν0s(a) :=νs(a) fors∈S, andν 0

snew(a) :=νslast(a)).J

130

Definition2.3 (Visibility change) LetM=hn,S, τ, κ, νibe aTVmodel, with snew < S; take a set of agents {b1, . . . ,bk} ⊆ A, and let Xi ⊆ A∪Obe a set of agents and objects for every bi, satisfying bi ∈ Xi. The visibility assignment [b1←X1, . . . ,bk←Xk] produces theTVmodel

M[b1←X1,...,bk←Xk] =hn+1,S∪ {snew}, τ

0, κ0, ν0i

in which(i)τ0

preserves the temporal position of states inS, makingsnew the 131

most recent (soτ0

(s) := τ(s) for s ∈ S, and τ0

(snew) := n+1);(ii)κ0

preserves 132

the colouring assignment for states inS, with the colouring insnew exactly as 133

inslast(so, for anyo∈ O, defineκ0s(o) :=κs(o) fors ∈S, andκ 0

snew(o) := κslast(o));

134

(iii)ν0

is exactly asνfor states inS, with the newsnewtaking the visibility ofslast 135

for agents not mentioned by the assignment, and following the assignment for 136

those agents it mentions (so, for anya∈ A, defineν0

s(a) :=νs(a) fors∈ S, with 137

ν0

snew(a) :=νslast(a) whena<{b1, . . . ,bk}, andν

0

snew(bj) :=Xjwhena=bj). J

(5)

The operations describe a change in the current situation; in this sense, they 139

are analogous to model operations inDynamic Epistemic Logic(DEL; [16,17]). 140

Still, there is an important difference. Typically,DELmodels describe only the 141

current situation, so model operations return a structure representing also a 142

single situation (the ‘next’ one). In contrast, while aTV model describes how 143

the situation is at the current stage (the stateτ−1(n)), it might also describe how 144

the situation was in the past (the other states). Thus, while the operations add a 145

state describing the situation the action produces, they also retain the states of the 146

original model, hence keeping track of the past. In this sense, theTVsetting can 147

be understood as a ‘dynamic temporal’: an underlying temporal structure that 148

can beextendedby dynamic ‘model change’ operations. Other proposals using 149

similar ideas include [18] (cf. [19,20]), which redefines the operation representing 150

acts of (public and) private communication [11] to preserve previous stages, and 151

[21], whose models ‘remember’ the initial epistemic situation. 152

A formal language The languageL, for describingTVmodels, contains basic

153

formulas expressing the (high-order) beliefs agents have about the colour of an 154

object, and it is closed under both the standard Boolean operators as well as 155

modalities for describing what will be the case after an action takes place. 156

Definition2.4 (LanguageL) GivenA,Oand{Ro}oO, formulasφof the lan-157

guageLare given by

158

φ::=Ba1· · ·Bak(oCc)|¬φ|φ∧φ|[α]φ fork>1,{a1, . . . ,ak} ⊆A,o∈O,c∈Ro

α::=p1:=c1, . . . ,pi:=ci|b1←X1, . . . ,bj←Xj fori>1,{p1, . . . ,pi} ⊆O,ci∈Rpi,

j>1,{b1, . . . ,bj} ⊆A,XiAPwithbiXi

159

Formulas of the formBa1· · ·Bak(oCc), calledmental attribution formulas, are read

160

as“agent a1believes that . . . that agent akbelieves that o has colour c”. Other Boolean 161

connectives (∨,→,↔) are defined in the standard way. J

162

Formulas inLare evaluated in aTVmodel with respect its last stateslast, the

163

fullest representation of the scenario available up that point. Nevertheless, as 164

the definition shows, the truth-value of formulas is influenced by earlier states. 165

Definition2.5 (Semantic interpretation) LetM =hn,S, τ, κ, νibe a tempo-166

ral visibility model. The following definitions will be useful. 167

Takeχ:=Ba

1· · ·Bak(oCc). Itsvisibility conditionons∈S, denoted by visχ(s),

and listing the requirements forχto be evaluated ats(agenta1can see agent

a2, . . . , agentak−1can see agentak, agentakcan see objecto), is given by

visχ(s) iffdef a2∈νs(a1) & . . . &ak∈νs(ak−1) &o∈νs(ak).

TakesSandt6τ(s). Thet-predecessorofs, denoted by [s]−t, is the (unique) state appearing exactlytstages befores,8and it is formally defined as

[s]−t:=τ−1(τ(s)−t)

8In particular, [s]

(6)

For evaluatingχ :=Ba1· · ·Bak(oCc), the process starts fromslast, going ‘back in

168

time’ one step at the time, looking for a state satisfyingχ’s visibility condition. 169

If suchs0

is reached,χ’s truth-value depends only on whetherohas colourcat 170

s0

; otherwise,χis false. Formally, and by using “&” for a natural-language dis-171

junction (just as “&” stands for a natural-language conjunction), the satisfaction 172

relationbetween aTVmodel and a mental attribution formula is given by 173

MBa1· · ·Bak(oCc) iffdef

τ(

&

slast)−1

i=0

                              

no−latter−vis

z }| { i

&

j=1 not visBa1···Bak(oCc)([slast]−(j−1))

& visBa1···Bak(oCc)([slast]−i))

| {z } vis

& κ[slast]−i(o)=c

| {z } col                                174

Thus,Ba1· · ·Bak(oCc) holds atMwhen there is a state (the quantification indicated

175

by the main disjunction) in which the visibility condition is satisfied (the vis 176

part), the object has the indicated colour (thecolpart), and there is no ‘more 177

recent’ state satisfying the visibility condition (theno−lattervispart).

178

Boolean operators are interpreted as usual. For ‘action’ modalities, 179

M[α]φ iffdef M[α]φ J

180

Before an example of the framework at work, there are four points worth-181

while to emphasise.(i)The semantic interpretation ofχ:=Ba1· · ·Bak(oCc)

cap-182

tures the discussed intuitive idea. On the one hand, if the visibility condition 183

fails at every state, the formula is false (every disjunct fails in itsvispart). On the 184

other hand, if some states satisfy the visibility condition, lets0

be the time-wise 185

latest (i.e.,s0 :=τ−1(max{τ(s) | vis

χ(s)})); then,M χiffκs0(o) =c.(ii)For the

186

sake of simplicity, we assume that, when an agentasees an agentb, andbsees 187

an objecto, thenain fact seesb seeing o, as it should be intuitively the case in 188

order for a formula likeBaBb(oCc) to be evaluated.9(iii)The term ‘belief’ here 189

does not have the strongELreading; it is rather understood as“truth according 190

to the agent’s current information about what has happened so far”(a form ofdefault 191

reasoning[24,25]: the agent assumes that things remain the way she saw them 192

last).(iv)Attributions to oneself boil down to thecolpart of the interpretation, 193

given the properties ofν, thus giving any agent full positive introspection. 194

Example3 Recall theSally-Anne Task, with its first two stages represented by

195

the modelMinExample 2. The story continues with Sally leaving the room, after 196

which she can see neither Anne nor the marble anymore, and Anne can only 197

9Notice that visibility of each agent is not ‘common knowledge’: knowledge relies

(7)

see the marble. This is represented by an operation extending the model with a 198

new state (s3) in which bothSa’s andAn’s visibility have changed, yielding the

199

modelM[Sa←{Sa},An←{An,mar}] =M0below.

200

s1

ν(Sa)=ν(An)={Sa,An,mar} mar

l

Sa

m

An

s2

ν(Sa)=ν(An)={Sa,An,mar} mar

l

Sa

m

An

s3

ν(Sa)={Sa};ν(An)={An,mar}

m

An mar

l

Sa 201

Does Anne believe that the marble is white? Intuitively, the answer should be

202

“yes”, and the system agrees:M0

BAn(marCwhite) holds, as atslastAnne sees 203

the marble (mar∈νs

3(An)), and the marble is indeed white (κs3(mar)=white).

204

Does Sally believe that the marble is white? The answer is“yes”, but for a

205

different reason:M0

BSa(marCwhite) holds because(i)althoughSacannot 206

seemarnow (ats3),(ii)the last time she saw it (s2),marwas white.

207

Does Anne believe that Sally believes that the marble is white? The relevant

208

state is the last time Anne saw Sally looking at the marble, i.e.,s2. Sincemar

209

is white ats2, indeedM0BAnBSa(marCwhite). 210

Finally, does Sally believe that Anne believes that the marble is white? As

211

before, we can verify thatM0BSaBAn(marCwhite). J 212

TV models from a modal perspective Readers familiar with modal logic [26] 213

will have noticed that aTVmodel is just a domain with a predecessor relation 214

(more precisely, a finite linear temporal structure); thus, it can also be described 215

by more standard modal languages. This will be made precise now, in order to 216

make explicit what the semantic evaluation of mental attribution formulas boils 217

down to. For simplicity, the focus will beL0

: the fragment ofLthat does not

218

include the dynamic modalities [p1:=c1, . . . ,pi:=ci] and [b1←X1, . . . ,bj←Xj]. 219

A modal language for describing a TV model requires special atoms for 220

agents’ visibility and objects’ colour. For the modalities, evaluating mental at-221

tribution formulas might require visiting previous states, so temporal operators 222

are needed. A suitable one for expressing what mental attribution formulas 223

encode is thesinceoperator S(φ, ψ) [27] (more precisely, itsstrictversion, found 224

also in, e.g., [28]), read as“sinceφwas true,ψhas been the case”.10Given a linear

225

structureM=hW,≺,ViandwW, the formula is interpreted as follows.

226

(M,w)S(φ, ψ) iffdef there isu ∈ W with(i) uw,(ii)(M,u) φ, and (iii)(M,v)ψfor everyv∈Wsuch thatuvw.11

227

10Note: a single ‘predecessor’ modality is insufficient, as the number of back steps the

recursive exploration requires isa prioriunknown. A modality for its reflexive and transitive closure is still not enough: it takes care of the recursive search for a state satisfying the visibility condition, but on its own cannot indicate that every state up to that point shouldnotsatisfy it. More on the adequacy ofsincecan be found in [27].

11Within propositional dynamic logic [29], and in the presence of the converse, thesince

modality can be defined as S(φ, ψ) :=h(; (?φ∪?(¬φ∧ψ)))+iφ, with “?” indicating

relational test, “;” indicating sequential composition, “∪” indicating non-deterministic

(8)

Thus, letLSbe the modal language whose formulas are given by

φ::=^ab|^ao|oCc| ¬φ|φ∧φ|S(φ, φ)

fora,b∈A,oOandcRo. The semantic interpretation of the atoms^ab,^ao

228

andoCcover aTV‘pointed’ model (M,s) is the natural one (look ats’s contents, 229

given byνs andκs); the semantic interpretation of S(φ, ψ) is as above, with≺ 230

taken to be the ‘‘strictly earlier than”relation over states inS, defined ass≺s0 iffdef 231

τ(s)< τ(s0). Then, by using the abbreviation visa1···ano:=^a1a2∧ · · · ∧^ak−1ak∧^ako

232

(so visa1···ano∈ LSexpresses the visibility condition of the formulaBa1· · ·Bak(oCc)),

233

the translationtr:L0→ LSis defined as

234

tr(Ba1· · ·Bak(oCc)):=(visa1···ano∧oCc)∨(¬visa1···ano∧S(visa1···ano∧oCc,¬visa1···ano)),

tr(¬φ):tr(φ), tr(φ∧ψ):=tr(φ)∧tr(ψ).

235

Then, M φiff(M,slast) tr(φ) holds for anyTV modelMand anyφ ∈ L0 . 236

The crucial case, for mental attribution formulas, is apparent:tr(Ba1· · ·Bak(oCc))

237

holds atslast in Mif and only if either the visibility condition holds and the 238

object has the indicated colour (visa1···ano∧oCc), or else the visibility condition

239

fails (¬visa

1···ano) and there is a state in the past where both visibility and colour

240

were satisfied, and since then visibility has failed (S(visa1···ano∧oCc,¬visa1···ano)).

241

This is exactly what the semantic interpretation ofBa1· · ·Bak(oCc) inMrequires.

242

Bisimulation The translationtrprovides an insight on the semantic clause for 243

mental attribution formulas. Equally illuminating is a bisimulation forL0 . 244

Definition2.6 (TV-bisimulation) TwoTVmodelsM=hn,S, τ, κ, νiandM0= 245

hm,S0, τ0, κ0, ν0i

(withslastands0lasttheir respective ‘last’ states) are said to be TV-246

bisimilar (notation:M↔

––M0) if and only if, for any mental attribution formula 247

χ:=Ba1· · ·Bak(oCc),(I) Forth:if there ist∈Ssuch that(i)visχ(t) holds,(ii)visχ(r)

248

fails for every r ∈ S with τ(t) < τ(r) 6 τ(slast), and(iii) κt(o) = c, then there

249

is t0 S0

such that (i) visχ(t0) holds, (ii) visχ(r0) fails for every r0 ∈ S with 250

τ0 (t0

)< τ0 (r0

)6τ0 (s0

last), and(iii)κt0(o)=c.(II) Back:vice versa. J 251

It can be proved that, wheneverMandM0

are TV-bisimilar, both models 252

satisfy the sameL0

-formulas.12The colour of an object is relevant only if some

253

agent can see it (so, no ‘atom’ clause is needed). Note also how twoTVmodels 254

satisfying the sameL0-formulas might differ in their cardinality, and also make 255

the same formula true in different ways (e.g.,¬Ba(oCc) holds inMbecause, at

256

slast, agentaseesohaving a colour other thanc, but it holds inM0

because, as 257

far asM0

is concerned, agentahas never seeno). Finally, notice how, although 258

TV-bisimulation impliesL0

-equivalence, it does not implyL-equivalence. Take

259

A={a}andO={o}, withs1a state in whichaseesobeing white, ands2one in

260

whichadoes not seeo. TakeMto be the model with onlys1, andM0to be the

261

model with boths1 ands2. The models areTV-bisimilar, henceL0-equivalent.

262

12SinceL0

(9)

Yet, they can be distinguished by the formula [o:=black]Ba(oCblack) (true inM, 263

false inM0): the different reasons whyL0

-formulas are made true in bisimilar 264

models become salient when actions enter the picture. For a bisimulation for 265

LS, it is enough to consider the mutual satisfaction of atoms in bisimilar points,

266

and suitableSinceconditions, as the ones discussed in [30, p.413]. 267

3

On modelling mental attribution scenarios

268

The TV framework aims to model belief attributions in a more cognitively 269

plausible way (compared with EL), revealing features thought of as crucial 270

ingredients of social cognition. Let’s justify these claims. 271

Informational economy On the one hand, a state in aTV model contains a 272

bare informational ‘minimum’: only basic facts regarding objects and agents’ 273

visibility. The operations on the model also induce ‘minimal’ changes, in ac-274

cordance to the criterion of informational economy in belief revision [31]. On 275

the other hand, the non-standard semantic clause for belief is complex, as the 276

state representing the current situation might not have all information neces-277

sary to evaluate a complex belief attribution, and thus the information at other 278

(previous) stages might be needed. A ‘backtracking’ process might be difficult 279

and time-consuming, depending on how many different states an agent needs 280

to ‘remember’, and our clause is sensitive to this observation, unlike the usual 281

modal interpretations. The level of complexity that one finds on theTV frame-282

work for both representing a situation (low) and evaluating mental attributions 283

(high) can be contrasted with whatELdoes, as discussed inSection 1. 284

Perspective shifting Another important feature, identified in analyses of ToM 285

and formalisations of False-Belief Tasks (FBTs), isperspective shifting[32]. Suc-286

cessful performance in the tasks (i.e., making correct attributions) requires a 287

perspective shift: stepping into the shoes of another agent.13Asking for the vis-288

ibility condition ensures precisely that agents change perspectives, even if that 289

means having to recall earlier stages. Making multiple shifts, e.g. in complex 290

high-order attributions, may be difficult compared to plainly attributing one’s 291

own belief to others, capturing why agents might fail in the tasks. 292

Principle of inertia A further crucial notion is theprinciple of inertia[6,33,34]: 293

an agent’s beliefs are preserved unless there is reason to the contrary. In our 294

case, reason to the contrary amounts to the satisfaction of visibility; if this is not 295

satisfied in the state of evaluation, then, essentially, the agent maintains beliefs 296

formed in earlier stages, where necessary information was available. 297

Dual process theories of reasoning Besides ToM, the TVsetting is in agree-298

ment with the literature supporting thedual process theories of reasoning[35,36,37]. 299

According to them, there are two systems underlying human reasoning. Sys-300

tem 1 (thefastmode) is quick, unconscious and automatic, often governed by 301

habit, biases and heuristics developed in the course of evolution. System 2 (the 302

slowmode) is gradual, deliberate and rule-based, and requires cognitive effort. 303

13In fact, unsuccessful performance, e.g. of autistic children, is often connected with a

(10)

System 1 is at play most of the time, constructing our idea of the world with el-304

ementary cues and avoiding cognitive overload. When rule-based calculations 305

become necessary, e.g. in face of a demanding task, System 2 takes over, building 306

on inputs of System 1 to slowly produce an output in a step-wise fashion. 307

We argue that agents’ higher order reasoning roughly follows this pattern. 308

System 1 keeps track only of a bare-minimum of information (basic facts), with-309

out overloading memory with information that can be later inferred. Whenever 310

a task requires more than what is stored (as higher-order attributions), System 311

2 takes over, using the inputs of System 1. This is precisely the pattern of our 312

semantics, with our models and updates encoding only basic facts. Whenever 313

a demanding task appears, such as the evaluation of a mental attribution, our 314

agents follow the cognitively hard calculations of our semantic clause.14On the

315

basis of elementary facts regarding whom/what they observed, they test certain 316

conditions and trace back earlier states. It is only after this slow and effortful 317

process that they can determine whether a higher-order attribution holds. 318

3.1 Detailed examples 319

False-Belief Tasks use stories to test the ability to attribute mental states to others. 320

In what follows, we provide formal representations of some of these storylines, 321

to the level of abstraction allowed by our framework’s constructions. 322

Example4 (First-orderFBT:theSally-Anne(SA)task) The full storyline (

Ex-323

ample 1) can be represented within theTVframework, modulo minor changes,

324

as already hinted at.(1)Sally and Anne are in a room, with Sally holding the 325

marble (the model with only states1inExample 2).(2)Sally puts the marble into

326

the basket (the full model inExample 2).(3)Sally leaves the room (the model in 327

Example 3).(4) Anne transfers the marble to the box (the model inFigure 1).

328

The task’s last step, Sally coming back to the room, prepares the audience for 329

the crucial question:“where does Sally believe the marble is?”. The action changes 330

Sally’s visibility (she can see Anne now), but it does not change the crucial fact 331

that she cannot see the marble. Thus, it is not relevant for our purposes. 332

So, which are Anne’s and Sally’s final high-order beliefs? According to 333

the framework, with M the model in Figure 1(top): M BSa(marCwhite)∧ 334

BAn(marCgreen), andMBSaBAn(marCwhite)∧BAnBSa(marCwhite). J 335

Example5 (Second-orderFBT:thechocolate(C)task) Adapted from [39], the

336

task is as follows. (1) Mary and John are in a room, with a chocolate bar in 337

the room’s table.(2)John puts the chocolate into the drawer, then(3)leaving the 338

room.(4)Mary transfers the chocolate to the box.(5)John peeks into the room, 339

without Mary noticing, and sees the chocolate in the box. 340

TheTVmodelling works stepwise, with the initial situation represented bys1

341

(blackindicates the chocolate is on the table), and each subsequent action adding 342

14Although it is always possible to evaluate attributions of any length (like in

possible-worlds semantics), our semantic clause offers a mechanism to account for human

(11)

s1

ν(Sa)=ν(An)={Sa,An,mar} mar

l

Sa

m

An

s2

ν(Sa)=ν(An)={Sa,An,mar} mar

l

Sa

m

An

s3

ν(Sa)={Sa};ν(An)={An,mar}

m

An mar

l

Sa

s4

ν(Sa)={Sa};ν(An)={An,mar}

m

An mar

l

Sa

s1

ν(Jo)=ν(Ma)={Jo,Ma,cho} cho

l

Jo

m

Ma

s2

ν(Jo)=ν(Ma)={Jo,Ma,cho} cho

l

Jo

m

Ma

s3

ν(Jo)={Jo};ν(Ma)={Ma,cho}

m

Ma cho

l

Jo

s4

ν(Jo)={Jo};ν(Ma)={Ma,cho}

m

Ma cho

l

Jo

s5

ν(Jo)={Jo,Ma,cho};ν(Ma)={Ma,cho}

m

[image:11.612.141.476.115.284.2]

Ma cho

m

Jo

Fig. 1:TVrepresentations ofSally-Anne Task(top) andChocolate Task(bottom).

a state. By putting the chocolate into the drawer (white), John producess2, and by

343

leaving the room he producess3. Mary createss4when she moves the chocolate

344

to the box (green), and finallys5 emerges when John peeks into the room. In

345

the final model, displayed inFigure 1(bottom), we have the following:(i)M 346

BMa(choCgreen)∧BJo(choCgreen),(ii)MBMaBJo(choCwhite)BJoBMa(choCgreen),

347

and(iii)MBMaBJoBMa(choCwhite)∧BJoBMaBJo(choCwhite). J 348

OtherFBTs (theIce Cream Task[40], thePuppy Task[41] and theBake-saletask 349

[42]) can be also represented in theTVframework, their crucial ToM features still 350

preserved. Still, some sources of change in zero- or higher- order information in 351

such dynamic scenarios might not be captured by our operations. While concep-352

tually similar examples can fit into our setting, up to some level of abstraction, 353

different operations might be required for other scenarios (Section 4). 354

3.2 Comparison with other proposals for mental attributions 355

Through a relational ‘preference’ framework for modelling different degrees of 356

belief, [43] studies three kinds of agents (including agents on the autism spec-357

trum), each endowed with specific “properties” as higher-order reasoners. Our 358

attempt does not focus on agents with specific strategies when evaluating belief 359

attributions, working instead onanyagent’s reasoning behind such process. 360

In [6], the authors provide a non-monotonic, closed-world reasoning for-361

malization of first-orderFBTs, implemented within logical programming. They 362

useevent calculus, with belief treated as a predicate, and rely on the principle of 363

inertia. While we design a different formalism, we still account for these features 364

without restricting ourselves to specific types of agents or orders of beliefs. 365

Another interesting logical formalization ofFBTs is given in [32,33,34]. These 366

papers use a proof-theoretic Hybrid Logic system for identifying perspective 367

shifts, while using inertia. The straightforward difference is that our approach 368

is rather semantic, with models keeping track of the actions involved, and in 369

(12)

The framework of [44] uses EL-beliefs plus special atoms indicating the 371

location of objects and the agents’ visibility, then representing changes in the 372

situation as action-model-based acts of (private) communication that rely on 373

agents’ visibility.15 The differences between our proposal and [44] have been

374

discussed: the contrast between complex models that simplify answering mental 375

attribution questions (EL) and simple states that require a complex process 376

for deciding high-order belief issues (here). The representation of actions also 377

differs: while [44] uses (a variation of) the heavy action models machinery (for 378

private communication), the actions of visibility and colour change presented 379

here simply modify atomic information. Finally, [44] also proposes two criteria 380

of success in formalizingFBTs:(i)robustness(being able to deal with as many 381

FBTs as possible, with no strict limit on the order of belief attribution), and 382

(ii)faithfulness(each action of the story should correspond to an action in the 383

formalism in a natural way). TheTVframework fulfils these requirements: it is 384

robust enough to deal with differentFBTs (seeSubsection 3.1and the discussion 385

therein), and the actions in the stories have a straightforward representation. 386

4

Summary and ongoing

/

future work

387

This paper has introduced a temporal framework suitable for capturing ‘real’ 388

agents’ mental state attributions. Its most important feature is the contrast be-389

tween a ‘simple’ semantic model (encoding only objects’ colours and agents’ 390

visibility) and a ‘complex’ clause for interpreting mental state attributions (es-391

sentially a temporal“since”operator). We have argued for its adequacy towards 392

representing important features of social cognition, as informational economy, 393

perspective shifting, inertia, and connections with dual process theories, with 394

these points exemplified through the modelling of commonFBTs. 395

This project presents several lines for further research. On the technical side, 396

there are still aspects of the logical setting to be investigated (e.g., axiomatisa-397

tion). Equally interesting is the exploration of extensions for modelling more 398

empirical findings. The main points made above on the adequacy of the frame-399

work make for a suitable basis for such extensions. Here are two possibilities. 400

A perspective function The setting can be fine-tuned to capture special types 401

of high-order reasoning (see case-studies of [16]). For example, autistic children 402

tend to fail theFBTs because they attribute their own beliefs to others [5]. This 403

and other similar situations can be accommodated through the introduction of 404

a perspectivefunction π :A → (AA) (with πa(b) = cunderstood as“agent

405

a considers agent b to have the perspective of agent c”), which then can be used to 406

define an appropriate variation of the visibility condition. In this way, an autistic 407

agentawould be one for whichπa(x)=afor anyx∈A, essentially relying only

408

on her own information, and thus attributing her own belief to others. 409

Different states for different agents at the same stage Another extension is 410

towards capturing scenarios involving communicative actions, including lying 411

15For example, the act through which, in the absence of Sally, Anne moves the marble

(13)

and spread of misinformation (e.g., the Puppy Task, the Bake Sale Task) and 412

other manifestations of social cognition (e.g., negotiations, games). With them, 413

it makes sense to include different states for different agents at the same stage, 414

each one of them representing the (potentially different) information different 415

agents might have about the situation at the same stage. 416

References

417

1. Premack, D., Woodruff, G.: Does the chimpanzee have a theory of mind? Behavioral

418

and Brain Sciences1(4) (1978) 515–526

419

2. Carruthers, P., Smith, P.K.: Theories of Theories of Mind. CUP (1996)

420

3. Wimmer, H., Perner, J.: Beliefs about beliefs: Representation and constraining

func-421

tion of wrong beliefs in young children’s understanding of deception. Cognition

422

13(1) (1983) 103 – 128

423

4. Wellman, H.M.: From desires to beliefs: Acquisition of a theory of mind. In: Natural

424

theories of mind: Evolution, development and simulation of everyday mindreading.

425

Basil Blackwell, Cambridge, MA, US (1991) 19–38

426

5. Baron-Cohen, S., Leslie, A.M., Frith, U.: Does the autistic child have a “theory of

427

mind”? Cognition21(1) (1985) 37–46

428

6. Stenning, K., van Lambalgen, M.: Human Reasoning and Cognitive Science. MIT

429

Press (2008)

430

7. Setoh, P., Scott, R.M., Baillargeon, R.: Two-and-a-half-year-olds succeed at a

tra-431

ditional false-belief task with reduced processing demands. Proceedings of the

432

National Academy of Sciences113(47) (2016) 13360–13365

433

8. Hintikka, J.: Knowledge and Belief. Cornell University Press, Ithaca, N.Y. (1962)

434

9. Fagin, R., Halpern, J.Y., Moses, Y., Vardi, M.Y.: Reasoning about knowledge. The

435

MIT Press, Cambridge, Mass. (1995)

436

10. Stalnaker, R.: The problem of logical omniscience, I. Synthese89(3) (1991) 425–440

437

11. Baltag, A., Moss, L.S., Solecki, S.: The logic of public announcements and common

438

knowledge and private suspicions. In Gilboa, I., ed.: Proceedings of the 7th

Confer-439

ence on Theoretical Aspects of Rationality and Knowledge (TARK-98), Evanston, IL,

440

USA, July 22-24, 1998, Morgan Kaufmann (1998) 43–56

441

12. Frege, G.: Grundlagen der Arithmetik. Breslau: Wilhelm Koebner (1884)

442

13. Pelletier, F.J., Elio, R., Hanson, P.: Is logic all in our heads? From naturalism to

443

psychologism. Studia Logica88(1) (2008) 3–66

444

14. Verbrugge, R.: Logic and social cognition. Journal of Philosophical Logic38(6) (2009)

445

649–680

446

15. van Benthem, J.: Logic and reasoning: Do the facts matter? Studia Logica88(1) (2008)

447

67–84

448

16. van Ditmarsch, H., van der Hoek, W., Kooi, B.: Dynamic Epistemic Logic. Volume

449

337 of Synthese Library Series. Springer, Dordrecht, The Netherlands (2008)

450

17. van Benthem, J.: Logical Dynamics of Information and Interaction. CUP (2011)

451

18. Yap, A.: Dynamic epistemic logic and temporal modality. In Girard, P., Roy, O.,

452

Marion, M., eds.: Dynamic Formal Epistemology. Springer (2011) 33–50

453

19. Sack, J.: Temporal languages for epistemic programs. Journal of Logic, Language

454

and Information17(2) (2008) 183–216

455

20. Renne, B., Sack, J., Yap, A.: Logics of temporal-epistemic actions. Synthese193(3)

456

(2016) 813–849

457

21. Baltag, A., ¨Ozg ¨un, A., Sandoval, A.L.V.: APAL with memory is better. In Moss, L.S.,

458

de Queiroz, R.J.G.B., Mart´ınez, M., eds.: 25th International Workshop WoLLIC 2018.

459

Volume 10944 of Lecture Notes in Computer Science., Springer (2018) 106–129

(14)

22. Gasquet, O., Goranko, V., Schwarzentruber, F.: Big brother logic: visual-epistemic

461

reasoning in stationary multi-agent systems. Autonomous Agents and Multi-Agent

462

Systems30(5) (2016) 793–825

463

23. Charrier, T., Herzig, A., Lorini, E., Maffre, F., Schwarzentruber, F.: Building epistemic

464

logic from observations and public announcements. In: Proceedings of the Fifteenth

465

International Conference on Principles of Knowledge Representation and Reasoning.

466

KR’16, AAAI Press (2016) 268–277

467

24. Reiter, R.: A logic for default reasoning. Artificial Intelligence13(1-2) (1980) 81–132

468

25. Ben-David, S., Ben-Eliyahu-Zohary, R.: A modal logic for subjective default

reason-469

ing. Artificial Intelligence116(1-2) (2000) 217–236

470

26. Blackburn, P., de Rijke, M., Venema, Y.: Modal logic. CUP, Cambridge, UK (2001)

471

27. Kamp, H.: Tense Logic and the Theory of Linear Order. PhD thesis, University of

472

California (1968)

473

28. Burgess, J.P.: Axioms for tense logic. I. “since” and “until”. Notre Dame Journal of

474

Formal Logic23(4) (1982) 367–374

475

29. Harel, D., Kozen, D., Tiuryn, J.: Dynamic Logic. MIT Press, Cambridge, USA (2000)

476

30. Kurtonina, N., De Rijke, M.: Bisimulations for temporal logic. Journal of Logic,

477

Language and Information6(4) (1997) 403–425

478

31. G¨ardenfors, P.: Knowledge in Flux. Modelling the Dynamics of Epistemic States.

479

MIT Press (1988)

480

32. Bra ¨uner, T.: Hybrid-logical reasoning in the smarties and Sally-Anne tasks. Journal

481

of Logic, Language and Information23(4) (2014) 415–439

482

33. Bra ¨uner, T.: Hybrid-logical reasoning in the smarties and Sally-Anne tasks: What

483

goes wrong when incorrect responses are given? In: Proceedings of the 37th

An-484

nual Meeting of the Cognitive Science Society, Pasadena, California, USA, Cognitive

485

Science Society (2015) 273–278

486

34. Bra ¨uner, T., Blackburn, P., Polyanskaya, I.: Second-order false-belief tasks: Analysis

487

and formalization. In: 23rd International Workshop, WoLLIC 2016, Springer (2016)

488

125–144

489

35. Kahneman, D.: Thinking, fast and slow. Farrar, Straus and Giroux, New York (2011)

490

36. Evans, J.: Dual process theories. In Ball, L., Thompson, V., eds.: The Routledge

491

International Handbook of Thinking and Reasoning. Routledge (2018) 151–64

492

37. Stanovich, K.E., West, R.F.: Individual differences in reasoning: Implications for the

493

rationality debate? Behavioral and Brain Sciences23(5) (2000) 645–665

494

38. Cowan, N.: The magical number 4 in short-term memory: A reconsideration of

495

mental storage capacity. The Behavioral and Brain Sciences24(2001) 87–114

496

39. Flobbe, L., Verbrugge, R., Hendriks, P., Kr¨amer, I.: Children’s application of theory

497

of mind in reasoning and language. Journal of Logic, Language and Information

498

17(4) (2008) 417–442

499

40. Perner, J., Wimmer, H.: “JohnThinksthat MaryThinksthat . . . ” attribution of

second-500

order beliefs by 5- to 10-year-old children. Journal of Experimental Child Psychology

501

39(3) (1985) 437–471

502

41. Sullivan, K., Zaitchik, D., Tager-Flusberg, H.: Preschoolers can attribute second-order

503

beliefs. Developmental Psychology30(1994) 395–402

504

42. Hollebrandse, B., van Hout, A., Hendriks, P.: Children’s first and second-order

false-505

belief reasoning in a verbal and a low-verbal task. Synthese191(3) (2014) 321–333

506

43. Ditmarsch, H.V., Labuschagne, W.: My beliefs about your beliefs: A case study in

507

theory of mind and epistemic logic. Synthese155(2) (2007) 191–209

508

44. Bolander, T.: Seeing is believing: Formalising false-belief tasks in dynamic epistemic

509

logic. In van Ditmarsch, H., Sandu, G., eds.: Jaakko Hintikka. Springer (2018) 207–236

Figure

Fig. 1: TV representations of Sally-Anne Task (top) and Chocolate Task (bottom).

References

Related documents

Another vital conjoin between these two visions of a smart city is the prioritisation of data capture and analysis as a means for underpinning evidence- informed policy

The concept of resilience can be illustrated in different ways. the Futurenet target year 2050) with capacity reductions determined by physical process models responding to

The UK Chandler Hospital Trauma Center is committed to providing timely, high-quality, cost-effective care for patients with complex, multisystem trauma or single-system trauma

London (Southampton, England), Edinburgh (Scotland), Kirkwall (Orkney Islands, Scotland), Lerwick (Shetland Islands, Scotland), Torshavn (Faroe Islands, Denmark), Akureyri

antigens and acute graft versus host disease after bone marrow transplantation.. Balduzzi A, Gooley T, Anasetti C

P101 If medical advice is needed, have product container or label at hand.. P310 Immediately call a POISON CENTER

Figure 1(ii) (a-e) shows typical ground surfaces with grits ranging from 120 to 600 grits showing a reduction in the intensity and size of the resulting grit lines. The reductions

Bad radiometric data in bands used in the production of the MODIS cloud mask will mean less tests applied in a given processing path, thus lessening the confidence in the quality of