How to Train Your AI PA: A Novel Approach to Timeline Evaluation and Inference
We live in a time where information is more readily available in greater quantities than ever before. The question is how do me make the best use of it? One recent method for summarising massive amounts of data and presenting it in an accessible way is timeline generation.
Timeline generation (TLG) is a way of representing a large amount of temporally dependent information concisely. It is query driven; we retrieve a corpus of text linked to some entity, event or other term. We then select a number of the constituent sentences, timestamp them and return them as output (Fig 1). The canonical TLG model makes this selection by fitting a topic model over the corpus. This is used to cluster these articles into stories. The most relevant of these stories are selected and summarised through some flavour of sentence-selection. It can be seen as a generalisation of the multi-document summarisation task, where we have introduced temporal dependency and structure.
In this paper, we first outline the canonical timeline generation model. We look at several domains where it has been applied, as well as its statistical foundation. Through surveying the current body of work for this model, we define two fundamental issues with current implementations: namely the process by which timelines are evaluated a