A Neural Network for Factoid Question Answering over Paragraphs
Focused on Quiz Bowl domain, where the task is to match multi sentences(paragraph-length text) to entities. Our model improves upon the existing dt-rnn model by jointly learning answer and question representations
in the same vector space rather than learning them separately.(IBM paper showed a similar result, use the same filter to both question and answer get the best result. Their embeddings are pretrained via word2vec, both question and answer share the same embedding matrix. They did not conduct experiment on using different embedding matrix for questions and ansers.).
这个里面的loss是不是有问题?理论上讲,如果理论上讲,正确答案离 更近的话, loss应该是0,但是目前这个不是。
Deep Learning for Answer Sentence Selection [It’s almost a survey]
Answer sentence selection is the task of identifying sentences that contain the
answer to a given question. We propose a novel approach to solving this task via means of distributed representations, and learn to match questions with answers by considering their semantic encoding. This contrasts prior work on this task, which typically relies on classifiers with large numbers of hand-crafted syntactic and semantic features and various external resources..
Question answering can broadly be divided into two categories. One approach focuses on semantic parsing, where answers are retrieved by turning a question into a database query and subsequently applying that query to an existing knowledge base. The other category is open domain question answering, which is more closely related to the field of information retrieval.
In this paper, we focus on answer sentence selection, the task that selects
the correct sentences answering a factual question from a set of candidate sentences.
Another line of work—closely related to the model presented here—is the application of recursive neural networks to factoid question answering over paragraphs. A key difference to our approach is that this model, given a question, selects answers from a relatively small fixed set of candidates encountered during training. On the other hand, the task of answer sentence selection that we address here, requires picking an answer from among a set of candidate sentences not encountered during training. In addition, each question has different numbers of candidate sentences.
Then, what does our data look like?.
Answer sentence selection requires both semantic and syntactic information in order to establish both what information the question seeks to answer, as well as whether a given candidate contains the required information, with current state-of-the-art approaches mostly focusing on syntactic matching (a number of paper has been published based on tree-edit feature) between questions and answers.
Our solution to this problem assumes that correct answers have high semantic similarity to questions. Unlike previous work, which measured the similarity mainly using syntactic information and handcrafted semantic resources, we model questions and answers as vectors, and evaluate the relatedness of each QA pair in a shared vector space.
Answer candidates were chosen using a combination of overlapping non-stop word counts and pattern matching.
The task is to rank the candidate answers based on their relatedness to the question, and is thus measured in Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR), which are standard metrics in Information Retrieval and Question Answering. Whereas MRR measures the rank of any correct answer, MAP examines the ranks of all the correct answers. In general, MRR is slightly
higher than MAP on the same list of ranked outputs, except that they are the same in the case where each question has exactly one correct answer.
One weakness of the distributed approach is that—unlike symbolic approaches—distributed representations are not very well equipped for dealing with cardinal numbers and proper nouns, especially considering the small dataset. They solve the by adding a words co-occurring count(boolean count and tf*idf count).