Git Product home page Git Product logo

kilt's Introduction

KILT logo

A Benchmark for Knowledge Intensive Language Tasks

http://kiltbenchmark.com/

The KILT benchmark is described in the following paper:

@inproceedings{petroni-etal-2021-kilt,
    title = "{KILT}: a Benchmark for Knowledge Intensive Language Tasks",
    author = {Petroni, Fabio  and Piktus, Aleksandra  and
      Fan, Angela  and Lewis, Patrick  and
      Yazdani, Majid  and De Cao, Nicola  and
      Thorne, James  and Jernite, Yacine  and
      Karpukhin, Vladimir  and Maillard, Jean  and
      Plachouras, Vassilis  and Rockt{\"a}schel, Tim  and
      Riedel, Sebastian},
    booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association 
                 for Computational Linguistics: Human Language Technologies",
    month = jun,
    year = "2021",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2021.naacl-main.200",
    doi = "10.18653/v1/2021.naacl-main.200",
    pages = "2523--2544",
}

https://arxiv.org/abs/2009.02252

Setup the env

conda create -n kilt37 -y python=3.7 && conda activate kilt37
pip install -e .

KILT knowledge source

The KILT knowledge source can be downloaded here: kilt_knowledgesource.json (34.76GiB).
It is based on the 2019/08/01 Wikipedia dump.
We use mongoDB to index the knowledge base (but you can use any json-based db).
To import the knowledge source in mongoDB run:

wget http://dl.fbaipublicfiles.com/KILT/kilt_knowledgesource.json
mongoimport --db kilt --collection knowledgesource --file kilt_knowledgesource.json

Structure of each record

{
 'wikipedia_title': 'Email marketing',
 'wikipedia_id': 1101759, 
 'text': ['p1', 'p2',...., 'pn'], # list of paragraph text
 'anchors': [{"text":,"href":,"paragraph_id":,"start":,"end":} ]  , 
 'categories': 'comma separated list of categories'
 'history': # some info from wikipedia, including original url
 'wikidata_info': # wikidata info
 }

Query the knowledge source

from kilt.knowledge_source import KnowledgeSource

# get the knowledge souce
ks = KnowledgeSource()

# count entries - 5903530
ks.get_num_pages()

# get page by id
page = ks.get_page_by_id(27097632)

# get pages by title
page = ks.get_page_by_title("Michael Jordan")

KILT data

Examples: KILT example

download the data

mkdir data
python scripts/download_all_kilt_data.py
python scripts/get_triviaqa_input.py

You can also download and use the KILT data through the HuggingFace's nlp library.

Note that we release only the input for the test sets, without answers. Test answers are used for the KILT challenge on EvalAI where participants can upload their models’ predictions and be listed on the public leaderboard (there are strict submission limits to discourage overfitting on test data).

KILT data format

{'id': # original data point id if available otherwise unique id
 'input': # question / claim / sentence / etc
 'output': [ # each element might contain an answer, a provenance or both
    {
    'answer': # answer in textual form
    'provenance': [
        # evidence set for the answer from the KILT ks
        {
            'wikipedia_id':  # *mandatory* 
            'title': 
            'section': 
            'start_paragraph_id': 
            'start_character': 
            'end_paragraph_id':
            'end_character': 
            'bleu_score': # wrt original evidence
            'meta': # dataset/task specific
        }
        ] 
      }
    ]
 'meta': # dataset/task specific
 }

KILT data catalogue

dataset task train dev test
FEVER Fact Checking fever-train-kilt.jsonl
(104,966 lines, 38.9MiB)
fever-dev-kilt.jsonl
(10,444 lines, 6.17MiB)
fever-test_without_answers-kilt.jsonl
(10,100 lines, 839kiB)
AIDA CoNLL-YAGO Entity Linking aidayago2-train-kilt.jsonl
(18,395 lines, 70.1MiB)
aidayago2-dev-kilt.jsonl
(4,784 lines, 21.1MiB)
aidayago2-test_without_answers-kilt.jsonl
(4,463 lines, 14.4MiB)
WNED-WIKI Entity Linking - wned-dev-kilt.jsonl
(3,396 lines, 12.9MiB)
wned-test_without_answers-kilt.jsonl
(3,376 lines, 13.3MiB)
WNED-CWEB Entity Linking - cweb-dev-kilt.jsonl
(5,599 lines, 90.2MiB)
cweb-test_without_answers-kilt.jsonl
(5,543 lines, 100MiB)
T-REx Slot Filling trex-train-kilt.jsonl
(2,284,168 lines, 1.75GiB)
trex-dev-kilt.jsonl
(5,000 lines, 3.80MiB)
trex-test_without_answers-kilt.jsonl
(5,000 lines, 896kiB)
Zero-Shot RE Slot Filling structured_zeroshot-train-kilt.jsonl
(147,909 lines, 71.4MiB)
structured_zeroshot-dev-kilt.jsonl
(3,724 lines, 2.27MiB)
structured_zeroshot-test_without_answers-kilt.jsonl
(4,966 lines, 1.22MiB)
Natural Questions Open Domain QA nq-train-kilt.jsonl
(87,372 lines, 51.9MiB)
nq-dev-kilt.jsonl
(2,837 lines, 7.94MiB)
nq-test_without_answers-kilt.jsonl
(1,444 lines, 334kiB)
HotpotQA Open Domain QA hotpotqa-train-kilt.jsonl
(88,869 lines, 52.8MiB)
hotpotqa-dev-kilt.jsonl
(5,600 lines, 3.97MiB)
hotpotqa-test_without_answers-kilt.jsonl
(5,569 lines, 778kiB)
TriviaQA Open Domain QA triviaqa-train_id-kilt.jsonl*
(61,844 lines, 102MiB)
triviaqa-dev_id-kilt.jsonl*
(5,359 lines, 9.81MiB)
triviaqa-test_id_without_answers-kilt.jsonl*
(6,586 lines, 123kiB)
ELI5 Open Domain QA eli5-train-kilt.jsonl
(272,634 lines, 548MiB)
eli5-dev-kilt.jsonl
(1,507 lines, 14.1MiB)
eli5-test_without_answers-kilt.jsonl
(600 lines, 99kiB)
Wizard of Wikipedia Dialogue wow-train-kilt.jsonl
(63,734 lines, 48.9MiB)
wow-dev-kilt.jsonl
(3,054 lines, 2.42MiB)
wow-test_without_answers-kilt.jsonl
(2,944 lines, 1.29MiB)

* run python scripts/get_triviaqa_input.py to get the question associated with each id

Additional data

For Entity Linking, in addition to the AIDA CoNLL-YAGO train set, the whole knowledge source can be used as training data by exploiting hyperlinks. To facilitate experimentation, we release such data in KILT format following the splits of BLINK:

We also provide a script to map the TAC-KBP 2010 dataset to the knowledge source and format of KILT.

Run the retrieval evaluation

Please follow this README.

Mapping scripts

Mapping scripts are located in kilt/datasets/. See scripts/map_datasets.py for an example.

Troubleshooting

If the module cannot be found, preface the python command with PYTHONPATH=.

If the experiments fail on GPU memory allocation, try reducing batch size.

License

KILT is MIT licensed. See the LICENSE file for details.

kilt's People

Contributors

anthonywchen avatar fabiopetroni avatar gizacard avatar jeanm avatar nicola-decao avatar ola13 avatar pminervini avatar sashank06 avatar yuwfan avatar zouharvi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

kilt's Issues

Script for generating kilt_knowledgesource.json file

Hi!
Would it be possible for You to include script used for converting Wikipedia dump into a kilt_knowledgesource.json file?
I am planning on generating the file for languages different than English.
Thanks!

Missing file for BLINK

BLINK_connector.py file has a hardcoded file path:
WIKIPEDIA_TITLE2ID = "/checkpoint/fabiopetroni/KILT/Wikipedia_title2id.p"
Can this file be made available?

Error while running bm25 predictions

FileNotFoundError: [Errno 2] No such file or directory: '\\kilt\\kilt\\configs\\retriever\\default_bm25.json'

In the retrieval configs there is no default file for bm25. Can you guys add that? Thanks

Serious inconsistencies in Wizard of Wikipedia data

If we look at the validation data, in the original data there are 8840 wizard turns all unique (including both the validation sets).
Out of the 3058 val instances only 1069 "answers" (or wizard utterances) are unique. That is, the same wizard utterance is shared by multiple conversational histories.

I also checked the latest version, and the issue still seems to be there. If you search for "answer": "Brown hair is the second in the file - http://dl.fbaipublicfiles.com/KILT/wow-dev-kilt.jsonl, you'll find the answer being used in 38 different conversations. That's not the case in the original data.

The same can be said about the train set. Out of 94577 training instances, there are only 63733 unique inputs and 20427 unique answers. I didn't check the current online version of the file though, because I'm assuming the same bug affects both validation and training files.

Here are the most 10 frequent validation answers (with frequencies)

[("I think I'm going to have mine done by a professional hairdresser,", 45),
 ('Jazz originated in the late 19th century', 44),
 ('Red is the colour at the end of the visible spectrum of light', 38),
 ('Brown hair is the second most common human hair color, after black hair, my hair is also brown but on the lighter side. some parts of my hair change to blonde in the summer.', 38),
 ('I am not 100% sure on that however, I do know that it was founded by Enzo Ferrari and the company built its first car in 1940.', 37),
 ("I probably wouldn't. I'm happy with black hair. Although, hair coloring is definitely on the rise, as 75% of women and 18% of men in Copenhagen, for example, have reported dying their hair, if that gives you any indication.", 37),
 ('Not 100% sure on that, either but Brand Finance rated the car the worlds most powerful car in 2014. That is awesome and I think I need a Ferrari. lol', 29),
 ('It has a wavelenght that starts at 625 nanometres.', 29),
 ('Hello, have you colored your hair before? It is practice of changing the hair color', 28),
 ("I've herd something crazy like 75% of women and 18% of men use hair dye.", 28)]

10 most frequent training answers (with frequencies)

[('It originated from Italy.', 343),
 ('I have one dog! I love selectively bred dogs.', 259),
 ('The first mention of it was in the 10th century, but nobody knows for sure who invented it.', 237),
 ("It's different. Our pizza was invented in Naples, and that's been popular around the world.", 210),
 ('Not right now, but I wish I did they are great for companionship and they can hunt vermin...lol', 190),
 ('So do I! it is one of the three primary colours??', 182),
 ('Yep, blue mixed with green and violet to make turquoise is great as well', 180),
 ("Yes, I see where you're coming from, but theres also potential for dogs proficient in hunting and herding, pulling loads,", 173),
 ('I think veganism is a bit narcissistic. The philosphy behind it I think elevates animals status illogically.', 171),
 ('yea it was founded by richard and maurice mcdonald in san bernardino, california', 166)]

I haven't checked any other datasets, so can't speak for them.

Pip installer?

Hi All, are you planning to make this pip installable?

Kilt with latest Wikipedia dump

Hi

Is there any script to generate KILT format dataset from the latest wikipedia dump? most specifically to train BLINK encoder models with new dataset.

Thanks.

Additional KB fields

I would like to request an extension to the KB structure, closer to the one released as part of the SLING Wikiflow or TREC CAR and/or data pipelines:
https://github.com/google/sling/blob/master/doc/guide/wikiflow.md
http://trec-car.cs.unh.edu/datareleases/v2.4-release.html

In particular, access to precomputed inlinks and incoming anchor text for a page, the occurrence of the page coming from disambiguation pages as well as redirects. These all provide rich name variety and signals commonly used in entity retrieval and linking tasks.

It seems like they could be taken from the output of WikiFlow or the CAR pipeline if the same wiki dump is used.

Are there plans to add/support these rich fields?

mapping wikidata to wikipedia

the issue

Hi,

Several thousands (11020 exactly) of wikipedia articles in the KILT knowledge source point to the same Wikidata item (often one of the articles is a disambiguation page).
However, in all of the examples I’ve tried, both articles have distinct Wikidata items (when following the 'Wikidata item' link on their page)
For example https://en.wikipedia.org/wiki/Ambale,_Chamarajanagar links to https://www.wikidata.org/wiki/Q4740999 and https://en.wikipedia.org/wiki/Ambale links to https://www.wikidata.org/wiki/Q48441930 (but in KILT both link to Q4740999)
Perhaps it is just a coincidence and these articles have been updated after you collected the data but I wondered how did you do the mapping between wikipedia and wikidata ?

Follow-up question: there is no indication in KILT that the wikipedia article is a disambiguation page or not, is it ?

disclaimer

I use HF version of KILT but I seriously doubt that it is the cause of the issue

more examples of 1-many mappings in KILT

formatted like QID and linked wikipedia articles in KILT

Q5584686 ['Gordi', 'Gordi (band)']      
Q2053115 ['Pat McGrath', 'Pat McGrath (make-up artist)']
Q48814857 ['Violence (Editors album)', 'Magazine (Editors song)', 'Hallelujah (So Low)']
Q7601873 ['Stargaze', 'StarGaze']
Q7976034 ['Wayne Bell', 'Wayne Bell (disambiguation)']
Q7345802 ['Robert Ironside (footballer)', 'Robert Ironside']
Q1227528 ['Directorate of Military Intelligence', 'Directorate of Military Intelligence (United Kingdom)']
Q195154 ['Rachel Corrie', 'Images of Rachel Corrie']
Q4409967 ['A modern fusion splicer', 'Fusion splicing']
Q2135463 ['Puhar, Nagapattinam', 'Puhar']
Q7207804 ['Pohádka', 'Pohádka (disambiguation)']
Q1660897 ['Inan', 'İnan']
Q7333686 ['John Boden', 'John Boden (cricketer)']
Q5552927 ['Gerry Mullan (footballer)', 'Gerry Mullan']
Q1335355 ['Brotula', 'Viviparous brotula']
Q12956809 ['Wild League (water polo)', 'Wild League']
Q19083 ['Kingdom of Iberia (antiquity)', 'Kingdom of Iberia']
Q933263 ['Ruby laser', 'A ruby laser']
Q7614510 ['Steven Bradbury (disambiguation)', 'Steven Bradbury']
Q5672371 ['Harry Simmons', 'Harry Simmons (baseball)']
Q6265672 ['John Westbury (MP)', 'John de Westbury']
Q7807635 ['Timpanogos', 'Timpanogos (disambiguation)']
Q42308 ['Occupation of Kharkiv', 'Kharkiv']
Q4740999 ['Ambale', 'Ambale, Chamarajanagar']
Q4770391 ['Another Weekend', 'Another Weekend (Five Star song)']
Q1890696 ['Mannerheim (family)', 'Mannerheim (disambiguation)']
Q2669498 ['Nacajuca', 'Nacajuca Municipality']
Q7680683 ['Tamanduateí (São Paulo Metro)', 'Tamanduateí (CPTM)']
Q1677353 ['Jackson Lake', 'Jackson Lake State Park']
Q26361 ['Podkamennaya Tunguska', 'Podkamennaya Tunguska River']
Q7286397 ['Rajpuri, Raigad', 'Rajpuri']
Q6575542 ['Gallery of United States Supreme Court composition templates', 'List of Justices of the Supreme Court of the United States by court composition']
Q3459340 ['Kaimri, Estonia', 'Kaimri']
Q7098625 ['Opposition (Malaysia)', 'Leader of the Opposition (Malaysia)']
Q1798855 ['La Hague', 'Cap de la Hague']
Q1361367 ['Erotikon', 'Erotikon (1920 film)']
Q16797896 ['San Carlos Bay (disambiguation)', 'San Carlos Bay']
Q18340038 ['Ishige (alga)', 'Ishige']
Q5237295 ['David McClure (footballer)', 'David McClure']
Q5537920 ['George Clancy (politician)', 'George Clancy']
Q5880780 ["Holland's Leaguer (play)", "Holland's Leaguer"]
Q2829213 ['Al-Fath ibn Khaqan (al-Andalus)', 'Al-Fath ibn Khaqan']
Q5432318 ['Falling Forward (Sandi Patty album)', 'Falling Forward']
Q5250281 ['Deep Roots (radio program)', 'Deep Roots']
Q5301815 ['Douglas Miller', 'Douglas Miller (Alberta politician)']
Q5503475 ['Friday (comics)', 'Friday (2000 AD)']
Q16931046 ['Red Oak Creek', 'Red Oak Creek (Trinity River)']
Q6833259 ['Michael Osborne (footballer)', 'Michael Osborne']
Q6396067 ['Kevin Corby', 'Kevin Corby (cricketer)']
Q2213258 ['Lode Wyns (athlete)', 'Lode Wyns']
Q7354485 ['Rock Is Dead', 'Rock Is Dead (The Doors song)']
Q2946994 ['Cesare Benedetti (disambiguation)', 'Cesare Benedetti']
Q6830341 ['Michael Flynn (disambiguation)', 'Michael Flynn']
Q7791556 ['Thomas Kirkpatrick', 'Thomas Kirkpatrick (Canadian politician)']
Q1459271 ['Similarity Matrix of Proteins', 'SIMAP']

Fail to reproduce 22,220,793 passages for DPR

Hi, according to the paper, there should be 22,220,793 passages in the KILT knowledge source. However, my reproduction results in 24,853,658 passages.

I count the passage number using the code below

passage_num = 0
for sample in kilt_wiki["full"]:
    text = [x.strip() for x in sample["text"]["paragraph"] if "BULLET::::" not in x]
    word_num = len(" ".join(text).split())
    passage_num += word_num//100 + int(bool(word_num%100))

Could you share the 22,220,793 passages or give more detail for reproduction?

Issue with eval_retrieval.py

Hello! I believe that there is an issue with kilt/eval_retrieval.py. In the code, when specifying the k value, the code grabs the k least similar passages instead of the k most similar passages.

For example, on NQ validation set, the R-Precision value using the current evaluation script is 10.43.

When I flip the list of retrieved passages in line 30
for provenance in output["provenance"][::-1]:

I get a R-Precision value of 53.68, which aligns with paper results.

unicode error when running get_triviaqa_input.py

Running into this issue when running get_triviaqa_input.py

$ python get_triviaqa_input.py

  1. download TriviaQA original tar.gz file
    100%|##########| 2.67G/2.67G [06:28<00:00, 6.87MiB/s]
  2. extract tar.gz file
    Extracting qa/wikipedia-train.json: 33%|###3 | 1/3 [02:35<05:10, 155.24s/iExtracting qa/wikipedia-dev.json: 33%|###3 | 1/3 [03:34<05:10, 155.24s/it]Extracting qa/wikipedia-test-without-answers.json: 67%|######6 | 2/3 [03:34<0Extracting qa/wikipedia-test-without-answers.json: 100%|##########| 3/3 [03:34<0Extracting qa/wikipedia-test-without-answers.json: 100%|##########| 3/3 [03:34<00:00, 71.58s/it]
  3. remove tar.gz file
  4. getting original questions
    qa/wikipedia-train.json
    Traceback (most recent call last):
    File "...KILT\scripts\get_triviaqa_input.py", line 88, in
    data = json.load(fin)
    File "...\miniconda3\lib\json_init_.py", line 293, in load
    return loads(fp.read(),
    File "...\miniconda3\lib\encodings\cp1252.py", line 23, in decode
    return codecs.charmap_decode(input,self.errors,decoding_table)[0]
    UnicodeDecodeError: 'charmap' codec can't decode byte 0x81 in position 2904: character maps to

Integration with DPR Repository

Hello! Thank you for the great repository/paper. I noticed some issues with how some of the DPR_connector code interacts with the DPR repository.

  1. This line leads to an error because DenseRetriever class in the DPR repo doesn't have that function. I believe that this line should be changed to: self.retriever.index.index_data(input_paths)

  2. This is really minor: In the Retriever README there is this instruction:

change line 185 of src/dpr/dense_retriever.py to

try:
    db_id, doc_vector = doc
except:
    title, db_id, doc_vector = doc

except that this isn't line 185 anymore.

Thanks again for the great work!

Best,
Anthony

ELI5 KILT annotation

Hi, thanks for creating the dataset! I have two questions regarding to the annotated ELI5 data:

  • According to the KILT data format in the main README, the field output/provenance/meta contains task-specific data. What does this meta field contain for the ELI5 dataset? Are these the manual annotation for instances with low overlap between passages and answers mentioned in section 4? If so, could you share a bit more on how is this manual annotation performed?
  • I noticed that for some annotation in the field output/provenance/meta/evidence_span contain a string "highlight sentence(s) containing evidence, not only the answer" at the end, (for instance, data in ELI5 dev set with id "3atjp2", as well as "49wqfo"). What are the meanings of such string?

Thank you!

Error connecting to db server

Hi,

after running
mongoimport --db kilt --collection knowledgesource --file kilt_knowledgesource.json

I constantly receive

2020-10-07T13:54:09.824+0200 [........................] kilt.knowledgesource 0B/34.8GB (0.0%)
2020-10-07T13:54:10.331+0200 [........................] kilt.knowledgesource 0B/34.8GB (0.0%)
2020-10-07T13:54:10.331+0200 Failed: error connecting to db server: no reachable servers
2020-10-07T13:54:10.331+0200 imported 0 documents

I everything OK with the server?

Question about submit

I want to reproduce a model on KILT(for example triviaQA)
image
and submit it on evalai.
image
is that something wrong(or I should wait for several hours to check the result)? I can't see the result

How can one output be used for recall@k?

Hi.

I'm using the evaluation script for recall@k on T-REx dataset.
I thought it needed the top-k retrieved pages as the outputs of predictions.

However, this line says "guess should provide exactly one output".

), f"guess should provide exactly one output for {guess_item['id']}"

I'm curious about how just one output can be used for recall@k?
Could you explain what I am misunderstanding regarding how recall@k is computed?

Missing attribution data in KILT NQ training

My understanding is a lot of the value from KILT comes from the gold attribution data. In the KILT data format wikipedia_id is listed as mandatory, but when I started working with nq-train-kilt.jsonl I quickly found that while ~77k examples have attribution ~10k examples don't.

Is this expected or a bug?

I checked nq-dev-kilt.jsonl and found every example there has attribution data.

Custom Dataset

Hi Team ,

Firstly thanks a lot for releasing KILT , its a very useful library and appreciate your hard work made in developing it , I wish to ask 2 questions WRT tool .

  1. How to load the latest Wikipedia dump any scripts or latest dataset you have or you plan to release anytime soon ?

  2. How to load our own custom dataset in KILT for all Tasks mentioned in the KILT library?

Looking forward for your response

Thanks & Regards

Ajay Sahu

How to distinguish the passage id in the text field

Hi, thanks for constructing and sharing the KILT corpus. I am using python to do some pre-processing work and I can't align the passage_id from the anchor field to the elements in the text field. In other words, the passage with passage_id i usually is not the i-th element in the text field.

I wonder if there are any rules to distinguish the passage id (presented in anchor) in the text field?

Thanks.

Numeric conversion markup resulting in missing number data

It appears that there is an issue with the handling of {{convert}} tags and some other numeric markup tags have issues being parsed for the text.

One example:
https://en.wikipedia.org/w/index.php?title=Tesla%20Roadster%20(2008)&amp;oldid=908332252

Rendered Wiki Text:
Tesla said the top speed is electronically limited to 201 km/h (125 mph). Tesla claims it has a weight of 1,305 kg (2,877 lb), a drag coefficient of Cd=0.35–0.36[110] and a rolling resistance of Crr=0.011.

Kilt text:
Tesla said the top speed is electronically limited to . Tesla claims it has a weight of , a and a .

The raw wikimarkup:
Tesla said the top speed is electronically limited to {{convert|125|mph|abbr=on|order=flip}}. Tesla claims it has a weight of {{convert|2877|lb||abbr=on|order=flip}}, a {{Cd|0.35–0.36|link=car|long=yes}}

This may impact QA or retrieval systems.

eli5 implicit knowledge model reproduce scores

Hi, I saw the scores for t5-base and bart models using implicit knowledge get around 19-20 rougeL, and was having issues matching that score. Currently, I am getting around 11 rougeL. I was wondering what hyperparameters and scheduler were used? Thanks

BART-DPR and RAG baselines

Hi, thanks for your excellent work. I am confused about BART-DPR and RAG baselines; the paper reports both downstream performance and page-level R-Precision. But it seems that their's retrieval is at the passage level. How can we use page-level retrieval for downstream tasks?

Missing original answers in TriviaQA

Hi,

In https://github.com/facebookresearch/KILT/blob/master/kilt/datasets/triviaqa.py#L58 you discard the original answer, keeping only the aliases
However this might be problematic to train a system since some aliases are not accurate (e.g. 'Bring On The Empty Horses' is a valid alias for 'David Niven' because https://en.wikipedia.org/wiki/Bring_On_The_Empty_Horses redirects to https://en.wikipedia.org/wiki/David_Niven#Writing)
Of course this also affects the Hugging Face release of KILT

Btw how did you train the models to obtain the results in the paper, what answer do you keep from the aliases?

Bests,

Can Multi-task DPR model parameter file be provided?

Thanks for your great work!
I want to do some experiments on the wizard of wikipedia benchmark. I use the DPR query encoder from http://dl.fbaipublicfiles.com/KILT/dpr_multi_set_hf_bert.0 and passage indexs from http://dl.fbaipublicfiles.com/KILT/kilt_passages_2048_0.pkl and get page-level recall-1 = 26.62, recall-5 = 48.82 in the dev set which is similar in your paper KILT: a Benchmark for Knowledge Intensive Language Tasks.

I notice that there is a Multi-task DPR which can get about recall-1 = 41.07 reported in your paper. Can you provide it's model parameter file (query-encoder file and passage-encoder file) which can help me a lot because retrain a DPR across several tasks is too time-consuming. Thanks a lot!

Aligning new datasets into KILT format

Congrats on the great work. I was thinking on aligning a new dataset into the KILT format to facilitate adoption, can you comment if:

  1. You would consider additions to the benchmark if it fits the task spirit.
  2. The code that does the transform and alignment (for example for Fever or NQ) will be released, after a quick search I couldn't find it.

Thanks!

Tac-kbp 2010 dataset format

Hi, could you tell me what's the format of tackbp 2010 dataset you used for this script? I remeber the raw tackbp dataset I got is xml file. Do I need to process the raw tackbp 2010 dataset so that I can use your script? Thanks!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.