Loading examples/evaluate_ict_zeroshot_nq.sh→examples/evaluate_retriever_nq.sh +4 −4 Original line number Diff line number Diff line #!/bin/bash # Evaluate natural question test data given Wikipedia embeddings and pretrained # ICT model # ICT model or a finetuned model for Natural Question task # Datasets can be downloaded from the following link: # https://github.com/facebookresearch/DPR/blob/master/data/download_data.py EVIDENCE_DATA_DIR=<Specify path of Wikipedia dataset> EMBEDDING_PATH=<Specify path of the embeddings> CHECKPOINT_PATH=<Specify path of pretrained ICT model> CHECKPOINT_PATH=<Specify path of pretrained ICT model or finetuned model> QA_FILE=<Path of the natural question test dataset> QA_FILE=<Path of the natural question dev or test dataset> python tasks/main.py \ --task ICT-ZEROSHOT-NQ \ --task RETRIEVER-EVAL \ --tokenizer-type BertWordPieceLowerCase \ --num-layers 12 \ --hidden-size 768 \ Loading Loading
examples/evaluate_ict_zeroshot_nq.sh→examples/evaluate_retriever_nq.sh +4 −4 Original line number Diff line number Diff line #!/bin/bash # Evaluate natural question test data given Wikipedia embeddings and pretrained # ICT model # ICT model or a finetuned model for Natural Question task # Datasets can be downloaded from the following link: # https://github.com/facebookresearch/DPR/blob/master/data/download_data.py EVIDENCE_DATA_DIR=<Specify path of Wikipedia dataset> EMBEDDING_PATH=<Specify path of the embeddings> CHECKPOINT_PATH=<Specify path of pretrained ICT model> CHECKPOINT_PATH=<Specify path of pretrained ICT model or finetuned model> QA_FILE=<Path of the natural question test dataset> QA_FILE=<Path of the natural question dev or test dataset> python tasks/main.py \ --task ICT-ZEROSHOT-NQ \ --task RETRIEVER-EVAL \ --tokenizer-type BertWordPieceLowerCase \ --num-layers 12 \ --hidden-size 768 \ Loading