12
1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050 051 052 053 054 055 056 057 058 059 060 061 062 063 064 065 066 067 068 069 070 071 072 073 074 075 076 077 078 079 080 081 082 083 084 085 086 087 088 089 090 091 092 093 094 095 096 097 098 099 ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE. Multi-Relational Question Answering from Narratives: Machine Reading and Reasoning in Simulated Worlds Anonymous ACL submission Abstract Question Answering (QA), as a research field, has primarily focused on either knowledge bases (KBs) or free text as a source of knowledge. These two sources have historically shaped the kinds of ques- tions that are asked over these sources, and the methods developed to answer them. In this work, we look towards a practi- cal use-case of QA over user-instructed knowledge that uniquely combines ele- ments of both structured QA over knowl- edge bases, and unstructured QA over narrative, introducing the task of multi- relational QA over personal narrative. As a first step towards this goal, we make three key contributions: (i) we generate and release TEXTWORLDSQA, a set of five diverse datasets, where each dataset contains dynamic narrative that describes entities and relations in a simulated world, paired with variably compositional ques- tions over that knowledge, (ii) we perform a thorough evaluation and analysis of sev- eral state-of-the-art QA models and their variants at this task, and (iii) we release a lightweight Python-based framework we call TEXTWORLDS for easily generating arbitrary additional worlds and narrative, with the goal of allowing the community to create and share a growing collection of diverse worlds as a test-bed for this task. 1 Introduction Personal devices that interact with users via nat- ural language conversation are becoming ubiqui- tous (e.g., Siri, Alexa), however, very little of that conversation today allows the user to teach, and then query, new knowledge. Most of the focus in She and Amy are both TAs for this course Which phd students are advised by the department head? John is now the department head User queries taught knowledge User teaches new knowledge Figure 1: Illustration of our task: relational question answering from dynamic knowledge ex- pressed via personal narrative these personal devices has been on Question An- swering (QA) over general world-knowledge (e.g., “who was the president in 1980” or “how many ounces are in a cup”). These devices open a new and exciting possibility of enabling end-users to teach machines in natural language, e.g., by ex- pressing the state of their personal world to its vir- tual assistant (e.g., via narrative about people and events in that user’s life) and enabling the user to ask questions over that personal knowledge (e.g., “which engineers in the QC team were involved in the last meeting with the director?”). This type of questions highlight a unique blend of two conventional streams of research in Question Answering (QA) – QA over struc- tured sources such as knowledge bases (KBs), and QA over unstructured sources such as free text. This blend is a natural consequence of our problem setting: (i) users may choose to express rich relational knowledge about their world, in turn enabling them to pose complex composi- tional queries (e.g., “all CS undergrads who took my class last semester”), while at the same time (ii) personal knowledge generally evolves through

Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

  • Upload
    others

  • View
    11

  • Download
    1

Embed Size (px)

Citation preview

Page 1: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

1

000

001

002

003

004

005

006

007

008

009

010

011

012

013

014

015

016

017

018

019

020

021

022

023

024

025

026

027

028

029

030

031

032

033

034

035

036

037

038

039

040

041

042

043

044

045

046

047

048

049

050

051

052

053

054

055

056

057

058

059

060

061

062

063

064

065

066

067

068

069

070

071

072

073

074

075

076

077

078

079

080

081

082

083

084

085

086

087

088

089

090

091

092

093

094

095

096

097

098

099

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Multi-Relational Question Answering from Narratives:Machine Reading and Reasoning in Simulated Worlds

Anonymous ACL submission

Abstract

Question Answering (QA), as a researchfield, has primarily focused on eitherknowledge bases (KBs) or free text as asource of knowledge. These two sourceshave historically shaped the kinds of ques-tions that are asked over these sources, andthe methods developed to answer them.In this work, we look towards a practi-cal use-case of QA over user-instructedknowledge that uniquely combines ele-ments of both structured QA over knowl-edge bases, and unstructured QA overnarrative, introducing the task of multi-relational QA over personal narrative. Asa first step towards this goal, we makethree key contributions: (i) we generateand release TEXTWORLDSQA, a set offive diverse datasets, where each datasetcontains dynamic narrative that describesentities and relations in a simulated world,paired with variably compositional ques-tions over that knowledge, (ii) we performa thorough evaluation and analysis of sev-eral state-of-the-art QA models and theirvariants at this task, and (iii) we releasea lightweight Python-based framework wecall TEXTWORLDS for easily generatingarbitrary additional worlds and narrative,with the goal of allowing the communityto create and share a growing collection ofdiverse worlds as a test-bed for this task.

1 Introduction

Personal devices that interact with users via nat-ural language conversation are becoming ubiqui-tous (e.g., Siri, Alexa), however, very little of thatconversation today allows the user to teach, andthen query, new knowledge. Most of the focus in

She and Amy are both TAs for this course

Which phdstudents are advised by the department head?

John is now the department head

Userqueriestaught

knowledge

Userteachesnewknowledge

Figure 1: Illustration of our task: relationalquestion answering from dynamic knowledge ex-pressed via personal narrative

these personal devices has been on Question An-swering (QA) over general world-knowledge (e.g.,“who was the president in 1980” or “how manyounces are in a cup”). These devices open a newand exciting possibility of enabling end-users toteach machines in natural language, e.g., by ex-pressing the state of their personal world to its vir-tual assistant (e.g., via narrative about people andevents in that user’s life) and enabling the user toask questions over that personal knowledge (e.g.,“which engineers in the QC team were involved inthe last meeting with the director?”).

This type of questions highlight a uniqueblend of two conventional streams of researchin Question Answering (QA) – QA over struc-tured sources such as knowledge bases (KBs),and QA over unstructured sources such as freetext. This blend is a natural consequence of ourproblem setting: (i) users may choose to expressrich relational knowledge about their world, inturn enabling them to pose complex composi-tional queries (e.g., “all CS undergrads who tookmy class last semester”), while at the same time(ii) personal knowledge generally evolves through

Page 2: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

2

100

101

102

103

104

105

106

107

108

109

110

111

112

113

114

115

116

117

118

119

120

121

122

123

124

125

126

127

128

129

130

131

132

133

134

135

136

137

138

139

140

141

142

143

144

145

146

147

148

149

150

151

152

153

154

155

156

157

158

159

160

161

162

163

164

165

166

167

168

169

170

171

172

173

174

175

176

177

178

179

180

181

182

183

184

185

186

187

188

189

190

191

192

193

194

195

196

197

198

199

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

1. There is an associate professor named Andy2. He returned from a sabbatical3. This professor currently has funding4. There is a masters level course called G3015. That course is taught by him6. That class is part of the mechanical

engineering department7. Roslyn is a student in this course8. U203 is a undergraduate level course9. Peggy and that student are TAs for this

course

…What students are advised by a professor with funding?[Albertha, Roslyn, Peggy, Lucy, Racquel]

What assistant professors advise students who passed their thesis proposal? [Andy]

Which courses have masters student TAs? [G301, U101 ]

Who are the professors working on unsupervised machine learning?[Andy, Hanna]

1. There is a new important mobile project2. That project is in the implementation stage3. Hiram is a tester on mobile project4. Mobile project has moved to the deployment

stage5. Andrew created a new issue for mobile

project: fails with apache stack6. Andrew is no longer assigned to that project7. That developer resolved the changelog needs

to be added issue

Are there any developers assigned to projects in the evaluation stage? [Tawnya, Charlott, Hiram]

Who is the null pointer exception during parsing issue assigned to? Hiram

Are there any issues that are resolved for experimental projects?[saving data throws exception,wrong pos tag on consecutive words]

AcademicDepartmentWorld SoftwareEngineeringWorld

Figure 2: Illustrative snippets from two sample worlds. We aim to generate natural-sounding first-personnarratives from five diverse worlds, covering a range of different events, entities and relations.

time and has an open and growing set of relations,making natural language the only practical inter-face for creating and maintaining that knowledgeby non-expert users. In short, the task that we ad-dress in this work is: multi-relational questionanswering from dynamic knowledge expressedvia narrative.

Although we hypothesize that question-answering over personal knowledge of this sortis ubiquitous (e.g., between a professor and theiradministrative assistant, or even if just in theuser’s head), such interactions are rarely recorded,presenting a significant practical challenge tocollecting a sufficiently large real-world dataset ofthis type. At the same time, we hypothesize thatthe technical challenges involved in developingmodels for relational question answering fromnarrative would not be fundamentally impactedif addressed via sufficiently rich, but controlledsimulated narratives. Such simulations alsooffer the advantage of enabling us to directlyexperiment with stories and queries of differentcomplexity, potentially offering additional insightinto the fundamental challenges of this task.

While our problem setting blends the problemsof relational question answering over knowledgebases and question answering over text, our hy-pothesis is that end-to-end QA models may learn

to answer such multisentential relational queries,without relying on an intermediate knowledgebase representation. In this work, we conductan extensive evaluation of a set of state-of-the-artend-to-end QA models on our task and analyzetheir results.

2 Related Work

Question answering has been mainly studied intwo different settings: KB-based and text-based.KB-based QA mostly focuses on parsing ques-tions to logical forms (Zelle and Mooney, 1996;Zettlemoyer and Collins, 2012; Berant et al., 2013;Kwiatkowski et al., 2013; Yih et al., 2015) in orderto better retrieve answer candidates from a knowl-edge base. Text-based QA aims to directly an-swer questions from the input text. This includesworks on early information retrieval-based meth-ods (Banko et al., 2002; Ahn et al., 2004) andmethods that build on extracted structured repre-sentations from both the question and the inputtext (Sachan et al., 2015; Sachan and Xing, 2016;Khot et al., 2017; Khashabi et al., 2018b). Al-though these structured presentations make rea-soning more effective, they rely on sophisticatedNLP pipelines and suffer from error propaga-tion. More recently, end-to-end neural archi-tectures have been successfully applied to text-

Page 3: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

3

200

201

202

203

204

205

206

207

208

209

210

211

212

213

214

215

216

217

218

219

220

221

222

223

224

225

226

227

228

229

230

231

232

233

234

235

236

237

238

239

240

241

242

243

244

245

246

247

248

249

250

251

252

253

254

255

256

257

258

259

260

261

262

263

264

265

266

267

268

269

270

271

272

273

274

275

276

277

278

279

280

281

282

283

284

285

286

287

288

289

290

291

292

293

294

295

296

297

298

299

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

based QA, including Memory-augmented neuralnetworks (Sukhbaatar et al., 2015; Miller et al.,2016; Kumar et al., 2016) and attention-based neu-ral networks (Hermann et al., 2015; Chen et al.,2016; Kadlec et al., 2016; Dhingra et al., 2017;Xiong et al., 2017; Seo et al., 2017; Chen et al.,2017). In this work, we focus on QA over text(where the text is generated from a supportingKB) and evaluate several state-of-the-art memory-augmented and attention-based neural architec-tures on our QA task. In addition, we consider asequence-to-sequence model baseline (Bahdanauet al., 2015), which has been widely used in dia-log (Vinyals and Le, 2015; Ghazvininejad et al.,2017) and recently been applied to generating an-swer values from Wikidata (Hewlett et al., 2016).

There are numerous datasets available for evalu-ating the capabilities of QA systems. For example,MCTest (Richardson et al., 2013) contains com-prehension questions for fictional stories. AllenAI Science Challenge (Clark, 2015) contains sci-ence questions that can be answered with knowl-edge from text books. RACE (Lai et al., 2017)is an English exam dataset for middle and highschool Chinese students. MULTIRC (Khashabiet al., 2018a) is a dataset that focuses on evaluatingmulti-sentence reasoning skills. These datasetsall require humans to carefully design multiple-choice questions and answers, so that certain as-pects of the comprehension and reasoning capa-bilities are properly evaluated. As a result, it isdifficult to collect them at scale. Furthermore, asthe knowledge required for answering each ques-tion is not clearly specified in these datasets, it canbe hard to identify the limitations of QA systemsand propose improvements.

Weston et al. (2015) proposes to use syntheticQA tasks (the BABI dataset) to better under-stand the limitations of QA systems. BABI buildson a simulated physical world similar to interac-tive fiction (Montfort, 2005) with simple objectsand relations and includes 20 different reasoningtasks. Various types of end-to-end neural net-works (Sukhbaatar et al., 2015; Lee et al., 2015;Peng et al., 2015) have demonstrated promisingaccuracies on this dataset. However, the per-formance can hardly translate to real-world QAdatasets, as BABI uses a small vocabulary (150words) and short sentences with limited languagevariations (e.g., nesting sentences, coreference).A more sophisticated QA dataset with a support-

ing KB is WIKIMOVIES (Miller et al., 2016),which contains 100k questions about movies,each of them is answerable by using either a KBor a Wikipedia article. However, WIKIMOVIES ishighly domain-specific, and similar to BABI, thequestions are designed to be in simple forms withlittle compositionality and hence limit the diffi-culty level of the tasks.

Our dataset differs in the above datasets in that(i) it contains five different realistic domains per-mitting cross-domain evaluation to test the abil-ity of models to generalize beyond a fixed set ofKB relations, (ii) it exhibits rich referring expres-sions and linguistic variations (vocabulary muchlarger than the BABI dataset), (iii) questions in ourdataset are designed to be deeply compositionaland can cover multiple relations mentioned acrossmultiple sentences.

Other large-scale QA datasets include Cloze-style datasets such as CNN/Daily Mail (Her-mann et al., 2015), Children’s Book Test (Hillet al., 2015), and Who Did What (Onishiet al., 2016); datasets with answers beingspans in the document, such as SQuAD (Ra-jpurkar et al., 2016), NewsQA (Trischler et al.,2016), and TriviaQA (Joshi et al., 2017); anddatasets with human generated answers, for in-stance, MS MARCO (Nguyen et al., 2016) andSearchQA (Dunn et al., 2017). One commondrawback of these datasets is the difficulty in ac-cessing a system’s capability of integrating infor-mation across a document context. Kocisky et al.(2017) recently emphasized this issue and pro-posed NarrativeQA, a dataset of fictional storieswith questions that reflect the complexity of nar-ratives: characters, events, and evolving relations.Our dataset contains similar narrative elements,but it is created with a supporting KB and hence itis easier to analyze and interpret results in a con-trolled setting.

3 TEXTWORLDS: Simulated Worlds forMulti-Relational QA from Narratives

In this work, we synthesize narratives in five di-verse worlds, each containing a thousand narra-tives and where each narrative describes the evolu-tion of a simulated user’s world from a first-personperspective. In each narrative, the simulated usermay introduce new facts, update existing facts orexpress a state change (e.g., “Homework 3 is nowdue on Friday” or “Samantha passed her the-

Page 4: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

4

300

301

302

303

304

305

306

307

308

309

310

311

312

313

314

315

316

317

318

319

320

321

322

323

324

325

326

327

328

329

330

331

332

333

334

335

336

337

338

339

340

341

342

343

344

345

346

347

348

349

350

351

352

353

354

355

356

357

358

359

360

361

362

363

364

365

366

367

368

369

370

371

372

373

374

375

376

377

378

379

380

381

382

383

384

385

386

387

388

389

390

391

392

393

394

395

396

397

398

399

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Statistics Value

# of total stories 5,000# of total questions 1,207,022Avg. # of entity mentions (per story) 217.4Avg. # of correct answers (per question) 8.7Avg. # of facts in stories 100Avg. # of tok. in stories 837.5Avg. # of tok. in questions 8.9Avg. # of tok. in answers 1.5Vocabulary size (tok.) 1,994Vocabulary size (entity) 10,793

Table 1: TEXTWORLDSQA dataset statistics

sis defense”). Each narrative is interleaved withquestions about the current state of the world, andquestions range in complexity depending on theamount of knowledge that needs to be integrated toanswer them. This allows us to benchmark a rangeof QA models at their ability to answer questionsthat require different extents of relational reason-ing to be answered.

The set of worlds that we simulate as part of thiswork are as follows:

1. MEETING WORLD: This world describes sit-uations related to professional meetings, e.g.,meetings being set/cancelled, people attendingmeetings, topics of meetings.

2. HOMEWORK WORLD: This world describessituations from the first-person perspective ofa student, e.g., courses taken, assignments indifferent courses, deadlines of assignments.

3. SOFTWARE ENGINEERING WORLD: Thisworld describes situations from the first-personperspective of a software development man-ager, e.g., task assignment to different projectteam members, stages of software develop-ment, bug tickets.

4. ACADEMIC DEPARTMENT WORLD: Thisworld describes situations from the first-personperspective of a professor, e.g., teaching assign-ments, faculty going/returning from sabbati-cals, students from different departments tak-ing/dropping courses.

5. SHOPPING WORLD: This world describes sit-uations about a person shopping for variousoccasions, e.g., adding items to a shoppinglist, purchasing items at different stores, notingwhere items are on sale.

3.1 Narrative

Each world is represented by a set of entities E anda set of unary, binary or ternary relations R. For-mally, a single step in one simulation of a worldinvolves a combination of instantiating new enti-ties and defining new (or mutating existing) re-lations between entities. Practically, we imple-ment each world as a collection of classes andmethods, with each step of the simulation creat-ing or mutating class instances by sampling en-tities and methods on those entities. By design,these classes and methods are easy to extend, toeither enrich existing worlds or create new ones.Each simulation step is then expressed as a naturallanguage statement, which is added to the narra-tive. In the process of generating a natural lan-guage expression, we employ a rich mechanismfor generating anaphora, such as “meeting withJohn about the performance review” and “meetingthat I last added”, in addition to simple pronounreferences. This allows us to generate more nat-ural and flowing narratives. These references aregenerated and composed automatically by the un-derlying TEXTWORLDS framework, significantlyreducing the effort needed to build new worlds.Furthermore, all generated stories also provide ad-ditional annotation that maps all entities to under-lying gold-standard KB ids, allowing to performexperiments that provide models with different de-grees of access to the “simulation oracle”.

We generate 1,000 narratives within each world,where each narrative consists of 100 sentences,plus up to 300 questions interleaved randomlywithin the narrative. See Figure 1 for two examplenarratives. Each story in a given world samples itsentities from a large general pool of entity namescollected from the web (e.g., people names, uni-versity names). Although some entities do overlapbetween stories, each story in a given world con-tains a unique flow of events and entities involvedin those events. See Table 1 for the data statistics.

3.2 Questions

Formally, questions are queries over theknowledge-base in the state defined up tothe point when the question is asked in the narra-tive. In the narrative, the questions are expressedin natural language, employing the same anaphoramechanism used in generating the narrative (e.g.,“who is attending the last meeting I added?”).

We categorize each question according to their

Page 5: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

5

400

401

402

403

404

405

406

407

408

409

410

411

412

413

414

415

416

417

418

419

420

421

422

423

424

425

426

427

428

429

430

431

432

433

434

435

436

437

438

439

440

441

442

443

444

445

446

447

448

449

450

451

452

453

454

455

456

457

458

459

460

461

462

463

464

465

466

467

468

469

470

471

472

473

474

475

476

477

478

479

480

481

482

483

484

485

486

487

488

489

490

491

492

493

494

495

496

497

498

499

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Dataset QuestionsSingle Entity/Relation Multiple entities

Single relation Two relations Three relationsMEETING 57,590 (41.16%) 46,373 (33.14%) 30,391 (21.72%) 5,569 (3.98%)HOMEWORK 45,523 (24.10%) 17,964 (9.51%) 93,669 (49.59%) 31,743 (16.80%)SOFTWARE 47,565 (20.59%) 51,302 (22.20%) 66,026 (28.58%) 66,150 (28.63%)ACADEMIC 46,965 (24.81%) 54,581 (28.83%) 57,814 (30.53%) 29,982 (15.83%)SHOPPING 111,522 (26.25%) 119,890 (28.22%) 107,418 (25.29%) 85,982 (20.24%)All 309,165 (26.33%) 290,110 (24.71%) 355,318 (30.27%) 219,426 (18.69%)

Table 2: Dataset statistics by question type.

compositionality, broadly, as a proxy to theamount of information that needs to be integratedacross the whole narrative in order to answer it.We categorize each question in our dataset intoone of the following four categories:

Single Entity/Single Relation Answers to thesequestions are a single entity, e.g. “what is John’semail address?”, or expressed in lambda-calculusnotation:

λx.EmailAddress(John, x)

The answers to these questions are found in a sin-gle sentence in the narrative, although it is pos-sible that the answer may change through thecourse of the narrative (e.g., “John’s new office isGHC122”).

Multi-Entity/Single Relation Answers to thesequestions can be multiple entities but involve asingle relation, e.g., “Who is enrolled in the Mathclass?”, or expressed in lambda calculus notation:

λx.TakingClass(x, Math)

Unlike the previous category, answers to thesequestions can be sets of entities.

Multi-Entity/Two Relations Answers to thesequestions can be multiple entities and involve tworelations, e.g., “Who is enrolled in courses that Iam teaching?”, or expressed in lambda calculus:

λx.∃y.EnrolledInClass(x, y)∧ CourseTaughtByMe(y)

Multi-Entity/Three Relations Answers to thesequestions can be multiple entities and involvethree relations, e.g., “Which undergraduates areenrolled in courses that I am teaching?”, or ex-pressed in lambda calculus notation:

λx.∃y.EnrolledInClass(x, y)∧ CourseTaughtByMe(y)

∧ Undergrad(x)

In the data that we generate, answers to questionsare always sets of spans in the narrative (the rea-son for this constraint is for easier evaluation ofseveral existing machine-reading models; this as-sumption can easily be relaxed in the simulation).In all of our evaluations, we will partition our re-sults by one of the four question categories listedabove, which we hypothesize correlates with thedifficulty of a question.

4 Methods

We develop several baselines for our QA task, in-cluding a logistic regression model and four differ-ent neural network models: Seq2Seq (Bahdanauet al., 2015), MemN2N (Sukhbaatar et al., 2015),BiDAF (Seo et al., 2017), and DrQA (Chen et al.,2017). These models generate answers in differ-ent ways, e.g., predicting a single entity, predict-ing spans of text, or generating answer sequences.Therefore, we implement two experimental set-tings: ENTITY and RAW. In the ENTITY setting,given a question and a story, we treat all the en-tity spans in the story as candidate answers, andthe prediction task becomes a classification prob-lem. In the RAW setting, a model needs to pre-dict the answer spans. For logistic regression andMemN2N, we adopt the ENTITY setting as theyare naturally classification models. This ideallyprovides an upper bound on the performance whenconsidering answer candidate generation. For allthe other models, we can apply the RAW setting.

4.1 Logistic Regression

The logistic regression baseline predicts the likeli-hood of an answer candidate being a true answer.For each answer candidate e and a given ques-tion, we extract the following features: (1) The fre-quency of e in the story; (2) The number of wordswithin e; (3) Unigrams and bigrams within e; (4)Each non-stop question word combined with eachnon-stop word within e; (5) The average minimum

Page 6: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

6

500

501

502

503

504

505

506

507

508

509

510

511

512

513

514

515

516

517

518

519

520

521

522

523

524

525

526

527

528

529

530

531

532

533

534

535

536

537

538

539

540

541

542

543

544

545

546

547

548

549

550

551

552

553

554

555

556

557

558

559

560

561

562

563

564

565

566

567

568

569

570

571

572

573

574

575

576

577

578

579

580

581

582

583

584

585

586

587

588

589

590

591

592

593

594

595

596

597

598

599

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

distance between each non-stop question word ande in the story; (6) The common words (excludingstop words) between the question and the text sur-rounding of e (within a window of 10 words); (7)Sum of the frequencies of the common words tothe left of e, to the right e, and both. These featuresare designed to help the model pick the correct an-swer spans. They have shown to be effective foranswer prediction in previous work (Chen et al.,2016; Rajpurkar et al., 2016).

We associate each answer candidate with a bi-nary label indicating whether it is a true answer.We train a logistic regression classifier to pro-duce a probability score for each answer candi-date. During test, we search for an optimal thresh-old that maximizes the F1 performance on the val-idation data. During training, we optimize thecross-entropy loss using Adam (Kingma and Ba,2014) with an initial learning rate of 0.01. We usea batch size of 10, 000 and train with 5 epochs.Training takes roughly 10 minutes for each do-main on a Titan X GPU.

4.2 Seq2SeqThe seq2seq model is based on the sequence tosequence model presented in (Bahdanau et al.,2015), which includes an attention model. Bah-danau et al. (Bahdanau et al., 2015) have used thismodel to build a neural based machine translationperforming at the state-of-the-art. We adopt thismodel to fit our own domain by including a pre-processing step in which all facts are concatenatedwith a dedicated token, while eliminating all pre-viously asked questions, and the current questionis added at the end of the list of facts. The answersare treated as a sequence of words. We use wordembeddings (Zou et al., 2013), as it was shownto improve accuracy. We use 3 GRU (Cho et al.,2014) connected layers, each with a capacity of256. Our batch size was set to 16. We use gradientdescent with an initial learning rate of 0.5 and a de-cay factor of 0.99. We iterated on the data for a to-tal of 50, 000 steps (roughly 5 epochs). The train-ing process for each domain took approximately48 hours on a Titan X GPU.

4.3 MemN2NEnd-To-End Memory Network (MemN2N) is aneural architecture that encodes both long-termand short-term context into a memory and it-eratively reads from the memory (i.e., multi-ple hops) relevant information to answer a ques-

tion (Sukhbaatar et al., 2015). It has been shownto be effective for a variety of question answeringtasks (Weston et al., 2015; Sukhbaatar et al., 2015;Hill et al., 2015).

In this work, we directly apply MemN2N toour task with a small modification. Originally,MemN2N was designed to produce a single an-swer for a question, so at the prediction layer, ituses softmax to select the best answer from theanswer candidates. In order to account for multi-ple answers for a given question, we modify theprediction layer to apply the logistic function andoptimize the cross entropy loss instead. For train-ing, we use the implementation setting in a pub-licly available MemN2N package 1. We train themodel for 100 epochs and it takes about 2 hoursfor each domain on a Titan X GPU.

4.4 BiDAF-M

BiDAF (Bidirectional Attention Flow Net-works) (Seo et al., 2017) is one of the top-performing models on the span-based questionanswering dataset SQuAD (Rajpurkar et al.,2016). We reimplement BiDAF with simplifiedparameterizations and change the prediction layerso that it can predict multiple answer spans.

Specifically, we encode the input story{x1, ..., xT } and a given question {q1, ..., qJ} atthe character level and the word level, where thecharacter level uses CNNs and the word level usespre-trained word vectors. The concatenation ofthe character and word embeddings are passedto a bidirectional LSTM to produce a contextualembedding for each word in the story contextand in the question. Then, we apply the samebidirectional attention flow layer to model theinteractions between the context and questionembeddings, producing question-aware featurevectors for each word in the context, denotedas G ∈ Rdg×T . G is then fed into a bidirec-tional LSTM layer to obtain a feature matrixM1 ∈ Rd1×T for predicting the start offset ofthe answer span, and M1 is then passed intoanother bidirectional LSTM layer to obtain afeature matrix M2 ∈ Rd2×T for predicting theend offset of the answer span. We then computetwo probability scores for each word i in thenarrative: pstart = sigmoid(wT

1 [G;M1]) andpend = sigmoid(wT

2 [G;M1;M2]), where w1

and w2 are trainable weights. The training objec-

1https://github.com/domluna/memn2n

Page 7: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

7

600

601

602

603

604

605

606

607

608

609

610

611

612

613

614

615

616

617

618

619

620

621

622

623

624

625

626

627

628

629

630

631

632

633

634

635

636

637

638

639

640

641

642

643

644

645

646

647

648

649

650

651

652

653

654

655

656

657

658

659

660

661

662

663

664

665

666

667

668

669

670

671

672

673

674

675

676

677

678

679

680

681

682

683

684

685

686

687

688

689

690

691

692

693

694

695

696

697

698

699

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Within-World MEETING HOMEWORK SOFTWARE DEPARTMENT SHOPPING Avg. F1

Logistic Regression 50.1 55.7 60.9 55.9 61.1 56.7Seq2Seq 22.5 32.6 16.7 39.1 31.5 28.5MemN2N 55.4 46.6 69.5 67.3 46.3 57.0BiDAF-M 81.8 76.9 68.4 68.2 68.7 72.8DrQA-M 81.2 83.6 79.1 76.4 76.5 79.4

Cross-World MEETING HOMEWORK SOFTWARE DEPARTMENT SHOPPING Avg. F1

Logistic Regression 9.0 9.1 11.1 9.9 7.2 9.3Seq2Seq 8.8 3.5 1.9 5.4 2.6 4.5MemN2N 23.6 2.9 4.7 14.6 0.07 9.2BiDAF-M 34.0 6.9 16.1 22.2 3.9 16.6DrQA-M 46.5 12.2 23.1 28.5 9.3 23.9

Table 3: F1 scores for different baselines evaluated on both within-world and across-world settings.

tive is simply the sum of cross-entropy losses forpredicting the start and end indices.

We use 50 1D filters for CNN character embed-ding, each with a width of 5. The word embeddingsize is 300 and the hidden dimension for LSTMs is128. For optimization, we use Adam (Kingma andBa, 2014) with an initial learning rate of 0.001,and use a minibatch size of 32 for 15 epochs. Thetraining process takes roughly 20 hours for eachdomain on a Titan X GPU.

4.5 DrQA-M

DrQA (Chen et al., 2017) is an open-domain QAsystem that has demonstrated strong performanceon multiple QA datasets. We modify the Doc-ument Reader component of DrQA and imple-ment it in a similar framework as BiDAF-M forfair comparisons. First, we employ the samecharacter-level and word-level encoding layers toboth the input story and a given question. We thenuse the concatenation of the character and wordembeddings as the final embeddings for words inthe story and in the question. We compute thealigned question embedding (Chen et al., 2017) asa feature vector for each word in the story and con-catenate it with the story word embedding and passit into a bidirectional LSTM to obtain the contex-tual embeddings E ∈ Rd×T for words in the story.Another bidirectional LSTM is used to obtain thecontextual embeddings for the question, and self-attention is used to compress them into one singlevector q ∈ Rd. The final prediction layer uses abilinear term to compute scores for predicting thestart offset: pstart = sigmoid(qTW1E) and an-other bilinear term for predicting the end offset:pend = sigmoid(qTW2E), where W1 and W2

are trainable weights. The training loss is the same

as in BiDAF-M, and we use the same parametersetting. Training takes roughly 10 hours for eachdomain on a Titan X GPU.

5 Experiments

We use two evaluation settings for measuring per-formance at this task: within-world and across-world. In the within-world evaluation setting, wetest on the same world that the model was trainedon. We then compute the precision, recall and F1

for each question and report the macro-average F1score for questions in each world. In the across-world evaluation setting, the model is trained onfour out of the five worlds, and tested on the re-maining world. The across-world regime is obvi-ously more challenging, as it requires the model tobe able to learn to generalize to unseen relationsand vocabulary. We consider the across-worldevaluation setting to be the main evaluation crite-ria for any future models used on this dataset, as itmimics the practical requirement of any QA sys-tem used in personal assistants: it has to be ableto answer questions on any new domain the userintroduces to the system.

5.1 Results

We draw several important observations from ourresults. First, we observe that more compositionalquestions (i.e., those that integrate multiple rela-tions) are more challenging for most models - asall models (except Seq2seq) decrease in perfor-mance with the number of relations composed ina question (Figure 5.1). This can be in part ex-plained by the fact that more composition ques-tions are typically longer, and also require themodel to integrate more sources of information inthe narrative in order to answer them. One surpris-

Page 8: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

8

700

701

702

703

704

705

706

707

708

709

710

711

712

713

714

715

716

717

718

719

720

721

722

723

724

725

726

727

728

729

730

731

732

733

734

735

736

737

738

739

740

741

742

743

744

745

746

747

748

749

750

751

752

753

754

755

756

757

758

759

760

761

762

763

764

765

766

767

768

769

770

771

772

773

774

775

776

777

778

779

780

781

782

783

784

785

786

787

788

789

790

791

792

793

794

795

796

797

798

799

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

1 2 3

Average number of relations in query

0

10

20

30

40

50

60

70

80F

1

Within-world

1 2 3

Average number of relations in query

0

10

20

30

40

50

60

70

80

Cross-world

DrQA-M

Bidaf-M

Logistic regression

Memory networks

Seq2seq

Figure 3: F1 score breakdown based on the num-ber of relations involved in the questions.

ing observation from our results is that the perfor-mance on questions that ask about a single relationand have only a single answer is lower than ques-tions that ask about a single relation but that canhave multiple answers (see detailed results in theAppendix). This is in part because questions thatcan have multiple answers typically have canoni-cal entities as answers (e.g., person’s name), andthese entities generally repeat in the text, makingit easier for the model to find the correct answer.

Table 3 reports the overall (macro-average) F1scores for different baselines. We can see thatBiDAF-M and DrQA-M perform surprisingly wellin the within-world evaluation even though theydo not use any entity span information. In partic-ular, DrQA-M outperforms BiDAF-M which sug-gests that modeling question-context interactionsusing simple bilinear terms have advantages overusing more complex bidirectional attention flows.The lower performance of MemN2N suggests thatits effectiveness on the BABI dataset does not di-rectly transfer to our dataset. Note that the orig-inal MemN2N architecture uses simple bag-of-words and position encoding for sentences. Thismay work well on dataset with a simple vocabu-lary, for example, MemN2N performs better in theSOFTWARE world compared to other worlds as theSOFTWARE world has a much smaller vocabulary.In general, we believe that better text representa-tions for questions and narratives can lead to im-proved performance. Seq2Seq model also did notperform as well. This is due to the inherent diffi-culty of generation and encoding long sequences.We found that it performs better when training and

testing on shorter stories (limited to 30 facts). In-terestingly, the logistic regression baseline outper-forms Seq2Seq and MemN2N, but there is still alarge performance gap to BiDAF-M and DrQA-M,and the gap is greater for questions that composemultiple relations.

In the across-world setting, the performance ofall methods dramatically decreases.2 This sug-gests the limitations of these methods in gener-alizing to unseen relations and vocabulary. Thespan-based models BiDAF-M and DrQA-M havean advantage in this setting as they can learnto answer questions based on the alignment be-tween the question and the narrative. However, thelow performance still suggests their limitations intransferring question answering capabilities.

6 Conclusion

In this work, we have taken the first steps towardsthe task of multi-relational question answering ex-pressed through personal narrative. Our hypoth-esis is that this task will become increasingly im-portant as users begin to teach personal knowledgeabout their world to the personal assistants em-bedded in their devices. This task naturally syn-thesizes two main branches of question answer-ing research: QA over KBs and QA over freetext. One of our main contributions is a collec-tion of diverse datasets that feature rich composi-tional questions over a dynamic knowledge graphexpressed through simulated narrative. Anothercontribution of our work is a thorough set of ex-periments and analysis of different types of end-to-end architectures for QA at their ability to an-swer multi-relational questions of varying degreesof compositionality. Our long-term goal is thatboth the data and the simulation code we releasewill inspire and motivate the community to looktowards the vision of letting end-users teach ourpersonal assistants about the world around us.

ReferencesDavid Ahn, Valentin Jijkoun, Gilad Mishne, Karin

Müller, Maarten de Rijke, Stefan Schlobach,M Voorhees, and L Buckland. 2004. Usingwikipedia at the trec qa track. In TREC. Citeseer.

2In order to allow generalization across different domainsfor the Seq2Seq model, we replace entities appearing in eachstory with an id that correlates to their appearance order. Af-ter the model outputs its prediction, the entity ids are con-verted back to the entity phrase.

Page 9: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

9

800

801

802

803

804

805

806

807

808

809

810

811

812

813

814

815

816

817

818

819

820

821

822

823

824

825

826

827

828

829

830

831

832

833

834

835

836

837

838

839

840

841

842

843

844

845

846

847

848

849

850

851

852

853

854

855

856

857

858

859

860

861

862

863

864

865

866

867

868

869

870

871

872

873

874

875

876

877

878

879

880

881

882

883

884

885

886

887

888

889

890

891

892

893

894

895

896

897

898

899

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Ben-gio. 2015. Neural machine translation by jointlylearning to align and translate. In ICLR.

Michele Banko, Eric Brill, Susan Dumais, and JimmyLin. 2002. Askmsr: Question answering usingthe worldwide web. In Proceedings of 2002 AAAISpring Symposium on Mining Answers from Textsand Knowledge Bases, pages 7–9.

Jonathan Berant, Andrew Chou, Roy Frostig, and PercyLiang. 2013. Semantic parsing on freebase fromquestion-answer pairs. In Proceedings of the 2013Conference on Empirical Methods in Natural Lan-guage Processing, pages 1533–1544.

Danqi Chen, Jason Bolton, and Christopher D Man-ning. 2016. A thorough examination of thecnn/daily mail reading comprehension task. In ACL.

Danqi Chen, Adam Fisch, Jason Weston, and AntoineBordes. 2017. Reading wikipedia to answer open-domain questions. In ACL.

Kyunghyun Cho, Bart Van Merriënboer, Caglar Gul-cehre, Dzmitry Bahdanau, Fethi Bougares, HolgerSchwenk, and Yoshua Bengio. 2014. Learningphrase representations using rnn encoder-decoderfor statistical machine translation. In EMNLP.

Peter Clark. 2015. Elementary school science and mathtests as a driver for ai: take the aristo challenge! InAAAI, pages 4019–4021.

Bhuwan Dhingra, Hanxiao Liu, Zhilin Yang,William W Cohen, and Ruslan Salakhutdinov.2017. Gated-attention readers for text comprehen-sion. In ACL.

Matthew Dunn, Levent Sagun, Mike Higgins, UgurGuney, Volkan Cirik, and Kyunghyun Cho. 2017.Searchqa: A new q&a dataset augmented withcontext from a search engine. arXiv preprintarXiv:1704.05179.

Marjan Ghazvininejad, Chris Brockett, Ming-WeiChang, Bill Dolan, Jianfeng Gao, Wen-tau Yih, andMichel Galley. 2017. A knowledge-grounded neuralconversation model. In AAAI.

Karl Moritz Hermann, Tomas Kocisky, EdwardGrefenstette, Lasse Espeholt, Will Kay, Mustafa Su-leyman, and Phil Blunsom. 2015. Teaching ma-chines to read and comprehend. In Advances in Neu-ral Information Processing Systems, pages 1693–1701.

Daniel Hewlett, Alexandre Lacoste, Llion Jones, IlliaPolosukhin, Andrew Fandrianto, Jay Han, MatthewKelcey, and David Berthelot. 2016. Wikireading: Anovel large-scale language understanding task overwikipedia. In ACL.

Felix Hill, Antoine Bordes, Sumit Chopra, and JasonWeston. 2015. The goldilocks principle: Readingchildren’s books with explicit memory representa-tions. arXiv preprint arXiv:1511.02301.

Mandar Joshi, Eunsol Choi, Daniel S Weld, and LukeZettlemoyer. 2017. Triviaqa: A large scale distantlysupervised challenge dataset for reading comprehen-sion. In ACL.

Rudolf Kadlec, Martin Schmid, Ondrej Bajgar, and JanKleindienst. 2016. Text understanding with the at-tention sum reader network. In ACL.

Daniel Khashabi, Snigdha Chaturvedi, Michael Roth,Shyam Upadhyay, and Dan Roth. 2018a. Lookingbeyond the surface: A challenge set for reading com-prehension over multiple sentences. In NAACL.

Daniel Khashabi, Tushar Khot Ashish Sabharwal, andDan Roth. 2018b. Question answering as global rea-soning over semantic abstractions. In AAAI.

Tushar Khot, Ashish Sabharwal, and Peter Clark. 2017.Answering complex questions using open informa-tion extraction. In ACL.

Diederik P Kingma and Jimmy Ba. 2014. Adam: Amethod for stochastic optimization. arXiv preprintarXiv:1412.6980.

Tomáš Kocisky, Jonathan Schwarz, Phil Blunsom,Chris Dyer, Karl Moritz Hermann, Gábor Melis,and Edward Grefenstette. 2017. The narrativeqareading comprehension challenge. arXiv preprintarXiv:1712.07040.

Ankit Kumar, Ozan Irsoy, Peter Ondruska, MohitIyyer, James Bradbury, Ishaan Gulrajani, VictorZhong, Romain Paulus, and Richard Socher. 2016.Ask me anything: Dynamic memory networks fornatural language processing. In International Con-ference on Machine Learning, pages 1378–1387.

Tom Kwiatkowski, Eunsol Choi, Yoav Artzi, and LukeZettlemoyer. 2013. Scaling semantic parsers withon-the-fly ontology matching. In Proceedings of the2013 Conference on Empirical Methods in NaturalLanguage Processing, pages 1545–1556.

Guokun Lai, Qizhe Xie, Hanxiao Liu, Yiming Yang,and Eduard Hovy. 2017. Race: Large-scale read-ing comprehension dataset from examinations. InEMNLP.

Moontae Lee, Xiaodong He, Wen-tau Yih, JianfengGao, Li Deng, and Paul Smolensky. 2015. Reason-ing in vector space: An exploratory study of ques-tion answering. arXiv preprint arXiv:1511.06426.

Alexander Miller, Adam Fisch, Jesse Dodge, Amir-Hossein Karimi, Antoine Bordes, and Jason We-ston. 2016. Key-value memory networks fordirectly reading documents. arXiv preprintarXiv:1606.03126.

Nick Montfort. 2005. Twisty Little Passages: an ap-proach to interactive fiction. Mit Press.

Page 10: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

10

900

901

902

903

904

905

906

907

908

909

910

911

912

913

914

915

916

917

918

919

920

921

922

923

924

925

926

927

928

929

930

931

932

933

934

935

936

937

938

939

940

941

942

943

944

945

946

947

948

949

950

951

952

953

954

955

956

957

958

959

960

961

962

963

964

965

966

967

968

969

970

971

972

973

974

975

976

977

978

979

980

981

982

983

984

985

986

987

988

989

990

991

992

993

994

995

996

997

998

999

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao,Saurabh Tiwary, Rangan Majumder, and Li Deng.2016. Ms marco: A human generated machinereading comprehension dataset. arXiv preprintarXiv:1611.09268.

Takeshi Onishi, Hai Wang, Mohit Bansal, Kevin Gim-pel, and David McAllester. 2016. Who did what:A large-scale person-centered cloze dataset. arXivpreprint arXiv:1608.05457.

Baolin Peng, Zhengdong Lu, Hang Li, and Kam-FaiWong. 2015. Towards neural network-based reason-ing. arXiv preprint arXiv:1508.05508.

Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, andPercy Liang. 2016. Squad: 100,000+ questions formachine comprehension of text. In EMNLP.

Matthew Richardson, Christopher JC Burges, and ErinRenshaw. 2013. Mctest: A challenge dataset forthe open-domain machine comprehension of text.In Proceedings of the 2013 Conference on Empiri-cal Methods in Natural Language Processing, pages193–203.

Mrinmaya Sachan, Kumar Dubey, Eric Xing, andMatthew Richardson. 2015. Learning answer-entailing structures for machine comprehension. InProceedings of the 53rd Annual Meeting of theAssociation for Computational Linguistics and the7th International Joint Conference on Natural Lan-guage Processing (Volume 1: Long Papers), vol-ume 1, pages 239–249.

Mrinmaya Sachan and Eric Xing. 2016. Machine com-prehension using rich semantic representations. InProceedings of the 54th Annual Meeting of the As-sociation for Computational Linguistics (Volume 2:Short Papers), volume 2, pages 486–492.

Minjoon Seo, Aniruddha Kembhavi, Ali Farhadi, andHannaneh Hajishirzi. 2017. Bidirectional attentionflow for machine comprehension. In ICLR.

Sainbayar Sukhbaatar, Jason Weston, Rob Fergus, et al.2015. End-to-end memory networks. In Advancesin neural information processing systems, pages2440–2448.

Adam Trischler, Tong Wang, Xingdi Yuan, Justin Har-ris, Alessandro Sordoni, Philip Bachman, and Ka-heer Suleman. 2016. Newsqa: A machine compre-hension dataset. arXiv preprint arXiv:1611.09830.

Oriol Vinyals and Quoc Le. 2015. A neural conversa-tional model. arXiv preprint arXiv:1506.05869.

Jason Weston, Antoine Bordes, Sumit Chopra, Alexan-der M Rush, Bart van Merriënboer, Armand Joulin,and Tomas Mikolov. 2015. Towards ai-completequestion answering: A set of prerequisite toy tasks.arXiv preprint arXiv:1502.05698.

Caiming Xiong, Victor Zhong, and Richard Socher.2017. Dynamic coattention networks for questionanswering. In ICLR.

Scott Wen-tau Yih, Ming-Wei Chang, Xiaodong He,and Jianfeng Gao. 2015. Semantic parsing viastaged query graph generation: Question answeringwith knowledge base. In ACL.

John M Zelle and Raymond J Mooney. 1996. Learn-ing to parse database queries using inductive logicprogramming. In Proceedings of the national con-ference on artificial intelligence, pages 1050–1055.

Luke S Zettlemoyer and Michael Collins. 2012. Learn-ing to map sentences to logical form: Structuredclassification with probabilistic categorial gram-mars. arXiv preprint arXiv:1207.1420.

Will Y Zou, Richard Socher, Daniel Cer, and Christo-pher D Manning. 2013. Bilingual word embeddingsfor phrase-based machine translation. In Proceed-ings of the 2013 Conference on Empirical Methodsin Natural Language Processing, pages 1393–1398.

Page 11: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

11

1000

1001

1002

1003

1004

1005

1006

1007

1008

1009

1010

1011

1012

1013

1014

1015

1016

1017

1018

1019

1020

1021

1022

1023

1024

1025

1026

1027

1028

1029

1030

1031

1032

1033

1034

1035

1036

1037

1038

1039

1040

1041

1042

1043

1044

1045

1046

1047

1048

1049

1050

1051

1052

1053

1054

1055

1056

1057

1058

1059

1060

1061

1062

1063

1064

1065

1066

1067

1068

1069

1070

1071

1072

1073

1074

1075

1076

1077

1078

1079

1080

1081

1082

1083

1084

1085

1086

1087

1088

1089

1090

1091

1092

1093

1094

1095

1096

1097

1098

1099

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Dataset QuestionsSingle Entity/Relation Multiple Entities

Single Relation Two Relations Three RelationsP R F1 P R F1 P R F1 P R F1

Logistic RegressionMEETING 42.0 78.1 51.0 50.6 74.6 56.6 33.3 66.3 41.1 31.8 57.6 38.0HOMEWORK 39.7 57.8 44.2 98.6 99.1 98.8 57.4 78.7 62.2 25.4 42.0 28.0SOFTWARE 55.0 73.3 59.0 54.3 98.2 66.5 58.2 76.0 62.3 46.3 84.6 56.4DEPARTMENT 42.6 65.9 48.0 59.0 82.5 65.1 38.8 52.7 41.2 42.5 64.6 46.9SHOPPING 53.1 70.2 56.2 79.6 83.4 79.0 53.1 60.5 52.3 53.4 67.9 56.0Average 46.5 69.1 51.7 68.4 87.6 73.2 48.2 66.8 51.8 39.9 63.3 45.1

Sequence-to-SequenceMEETING 27.9 18.3 22.1 48.1 12.1 19.3 42.1 15.0 22.1 33.7 19.7 24.8HOMEWORK 16.3 9.0 11.6 71.9 9.3 16.4 75.3 35.9 48.6 32.9 15.6 21.1SOFTWARE 42.5 21.5 28.5 44.8 8.5 14.2 50.0 6.3 11.2 45.5 7.4 12.7DEPARTMENT 49.9 35.6 41.5 54.1 20.3 29.6 57.2 38.0 45.7 43.9 39.7 41.7SHOPPING 25.8 16.0 19.8 71.3 28.2 40.5 33.3 19.3 24.4 46.9 31.4 37.6Average 32.5 20.1 24.7 58.0 15.7 24.0 51.6 22.9 30.4 40.6 22.7 27.6

MemN2NMEETING 56.9 56.0 54.7 66.8 58.4 58.6 57.0 57.5 54.8 38.7 40.7 38.8HOMEWORK 42.6 41.2 41.3 97.9 63.7 73.9 60.4 47.9 49.4 36.5 29.0 30.1SOFTWARE 68.5 71.6 68.5 72.9 73.2 70.9 69.7 67.3 66.1 75.0 74.8 72.6DEPARTMENT 56.3 74.3 61.3 78.5 87.0 80.2 59.4 76.6 63.2 57.8 74.2 61.6SHOPPING 51.3 45.4 45.5 74.9 54.1 59.0 45.6 40.6 40.2 44.3 37.6 37.9Average 55.1 57.7 54.3 78.2 67.3 68.5 58.4 58.0 54.8 50.4 51.3 48.2

BIDAF-MMEETING 87.6 92.4 88.2 78.6 86.1 79.2 68.9 89.6 74.6 73.9 94.4 80.0HOMEWORK 79.9 97.4 84.5 86.8 81.0 82.4 76.4 90.0 78.9 47.0 78.5 55.5SOFTWARE 48.0 89.4 57.4 68.5 93.6 75.8 62.4 86.1 67.5 62.7 90.9 71.3DEPARTMENT 57.0 64.6 58.1 73.6 85.9 76.6 67.0 83.2 70.8 63.1 71.4 64.0SHOPPING 60.5 87.1 66.9 76.7 90.9 79.8 57.1 89.0 65.8 53.2 88.5 62.0Average 66.6 86.2 71.0 76.8 87.5 78.8 66.4 87.6 71.5 60.0 84.7 66.6

DrQA-MMEETING 77.1 94.2 81.0 80.6 95.8 85.1 68.6 95.7 76.8 64.1 97.9 74.3HOMEWORK 88.8 97.9 91.4 85.2 80.2 81.4 85.0 94.7 87.9 51.6 85.8 60.2SOFTWARE 72.7 96.0 78.9 78.6 93.3 82.7 79.4 89.4 80.9 66.3 93.2 74.5DEPARTMENT 67.1 97.9 76.1 80.3 95.0 84.1 67.1 94.4 74.8 55.8 95.2 66.9SHOPPING 71.5 93.9 77.7 86.4 94.8 88.7 62.8 91.1 71.4 62.4 90.7 69.7Average 75.4 96.0 81.0 82.2 91.8 84.4 72.6 93.1 78.4 60.0 92.6 69.1

Table 4: Test performance at the task of question answering by question type using the within-worldevaluation.

Page 12: Multi-Relational Question Answering from Narratives ...azariaa.com/Content/Publications/PersonalNarrativeQA.pdf · 1 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015

12

1100

1101

1102

1103

1104

1105

1106

1107

1108

1109

1110

1111

1112

1113

1114

1115

1116

1117

1118

1119

1120

1121

1122

1123

1124

1125

1126

1127

1128

1129

1130

1131

1132

1133

1134

1135

1136

1137

1138

1139

1140

1141

1142

1143

1144

1145

1146

1147

1148

1149

1150

1151

1152

1153

1154

1155

1156

1157

1158

1159

1160

1161

1162

1163

1164

1165

1166

1167

1168

1169

1170

1171

1172

1173

1174

1175

1176

1177

1178

1179

1180

1181

1182

1183

1184

1185

1186

1187

1188

1189

1190

1191

1192

1193

1194

1195

1196

1197

1198

1199

ACL 2018 Submission ***. Confidential Review Copy. DO NOT DISTRIBUTE.

Dataset QuestionsSingle Entity/Relation Across Entities

Single Relation Two Relations Three RelationsLogistic Regression

MEETING 8.8 10.9 7.2 5.6HOMEWORK 7.5 20.2 8.5 6.7SOFTWARE 8.2 12.0 12.9 10.6DEPARTMENT 7.4 14.4 9.7 6.1SHOPPING 8.2 9.0 5.9 6.6Average 8.0 13.3 8.8 7.1

Sequence-to-SequenceMEETING 7.4 8.1 10.0 14.0HOMEWORK 4.2 2.9 3.1 2.3SOFTWARE 5.0 0.6 0.9 1.1DEPARTMENT 5.5 4.0 5.6 5.6SHOPPING 2.5 2.6 2.3 2.8Average 4.9 3.6 4.4 5.2

MemN2NMEETING 9.0 34.2 33.0 27.4HOMEWORK 3.3 12.4 1.0 2.5SOFTWARE 13.4 0.8 3.2 2.9DEPARTMENT 12.9 20.8 13.0 9.4SHOPPING 0.1 0.07 0.05 0.03Average 7.8 13.7 10.1 8.4

BIDAF-MMEETING 31.1 40.2 30.4 30.0HOMEWORK 10.4 20.3 2.3 7.8SOFTWARE 19.2 13.4 22.7 9.1DEPARTMENT 23.3 30.5 19.0 13.5SHOPPING 5.6 3.2 2.6 3.4Average 17.9 21.5 15.4 12.8

DrQA-MMEETING 44.5 58.8 33.3 37.1HOMEWORK 19.8 30.1 5.9 9.4SOFTWARE 26.4 23.4 24.0 19.4DEPARTMENT 31.0 38.8 24.4 15.7SHOPPING 19.3 2.3 6.7 7.1Average 28.2 30.7 18.9 17.7

Table 5: Test performance (F1 score) at the task of question answering by question type using theacross-world evaluation.