Babi Models
Collection
3 items
•
Updated
Fine tune and evaluate transformer model on facebook's bAbi tasks.
Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks
Training Code: p208p2002/bAbi-tasks-with-transformer-model
task_no | task_name | score |
---|---|---|
qa1 | single-supporting-fact | 100 |
qa2 | two-supporting-facts | 99.4 |
qa3 | three-supporting-facts | 62.0 |
qa4 | two-arg-relations | 100 |
qa5 | three-arg-relations | 96.0 |
qa6 | yes-no-questions | 100 |
qa7 | counting | 100 |
qa8 | lists-sets | 95.6 |
qa9 | simple-negation | 100 |
qa10 | indefinite-knowledge | 100 |
qa11 | basic-coreference | 100 |
qa12 | conjunction | 100 |
qa13 | compound-coreference | 100 |
qa14 | time-reasoning | 100 |
qa15 | basic-deduction | 100 |
qa16 | basic-induction | 100 |
qa17 | positional-reasoning | 100 |
qa18 | size-reasoning | 100 |
qa19 | path-finding | 100 |
qa20 | agents-motivations | 100 |
# Please use with the follow template
INPUT_TEMPLATE = """
Context:
{context}
Question:
{question}
Answer:
{answer}
"""
input_text = INPUT_TEMPLATE.format_map({
"context":context,
"question":question,
"answer":answer
}).strip()