Pyserini Reproductions (original) (raw)

The main results table provides commands for reproducing runs using the following models:

The symbol shown in parentheses (e.g., bm25-flat) refers the parameter used in the --condition option for programmatic execution (see below). In addition, we provide the following models whose runs can be reproduced, but are not shown in the results table:

BM25 Flat BM25 MF SPLADE Contriever BGE Cohere
nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100
trec-covid 0.595 0.109 0.656 0.114 0.727 0.128 0.596 0.091 0.781 0.141 0.818 0.159
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-trec-covid.flat \ --topics beir-v1.0.0-trec-covid-test \ --output run.beir.bm25-flat.trec-covid.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-covid-test \ run.beir.bm25-flat.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-covid-test \ run.beir.bm25-flat.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-covid-test \ run.beir.bm25-flat.trec-covid.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-trec-covid.multifield \ --topics beir-v1.0.0-trec-covid-test \ --output run.beir.bm25-multifield.trec-covid.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-covid-test \ run.beir.bm25-multifield.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-covid-test \ run.beir.bm25-multifield.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-covid-test \ run.beir.bm25-multifield.trec-covid.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-trec-covid.splade-pp-ed \ --topics beir-v1.0.0-trec-covid.test.splade-pp-ed \ --output run.beir.splade-pp-ed.trec-covid.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-covid-test \ run.beir.splade-pp-ed.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-covid-test \ run.beir.splade-pp-ed.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-covid-test \ run.beir.splade-pp-ed.trec-covid.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-trec-covid.contriever-msmarco \ --topics beir-v1.0.0-trec-covid-test \ --output run.beir.contriever-msmarco.trec-covid.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-covid-test \ run.beir.contriever-msmarco.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-covid-test \ run.beir.contriever-msmarco.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-covid-test \ run.beir.contriever-msmarco.trec-covid.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-trec-covid.bge-base-en-v1.5 \ --topics beir-v1.0.0-trec-covid-test \ --output run.beir.bge-base-en-v1.5.faiss.trec-covid.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-covid-test \ run.beir.bge-base-en-v1.5.faiss.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-covid-test \ run.beir.bge-base-en-v1.5.faiss.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-covid-test \ run.beir.bge-base-en-v1.5.faiss.trec-covid.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-trec-covid.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-trec-covid-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-trec-covid-test \ --output run.beir.cohere-embed-english-v3.0.trec-covid.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-covid-test \ run.beir.cohere-embed-english-v3.0.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-covid-test \ run.beir.cohere-embed-english-v3.0.trec-covid.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-covid-test \ run.beir.cohere-embed-english-v3.0.trec-covid.txt
bioasq 0.522 0.769 0.465 0.715 0.498 0.739 0.383 0.607 0.415 0.632 0.457 0.679
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-bioasq.flat \ --topics beir-v1.0.0-bioasq-test \ --output run.beir.bm25-flat.bioasq.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-bioasq-test \ run.beir.bm25-flat.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-bioasq-test \ run.beir.bm25-flat.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-bioasq-test \ run.beir.bm25-flat.bioasq.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-bioasq.multifield \ --topics beir-v1.0.0-bioasq-test \ --output run.beir.bm25-multifield.bioasq.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-bioasq-test \ run.beir.bm25-multifield.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-bioasq-test \ run.beir.bm25-multifield.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-bioasq-test \ run.beir.bm25-multifield.bioasq.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-bioasq.splade-pp-ed \ --topics beir-v1.0.0-bioasq.test.splade-pp-ed \ --output run.beir.splade-pp-ed.bioasq.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-bioasq-test \ run.beir.splade-pp-ed.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-bioasq-test \ run.beir.splade-pp-ed.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-bioasq-test \ run.beir.splade-pp-ed.bioasq.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-bioasq.contriever-msmarco \ --topics beir-v1.0.0-bioasq-test \ --output run.beir.contriever-msmarco.bioasq.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-bioasq-test \ run.beir.contriever-msmarco.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-bioasq-test \ run.beir.contriever-msmarco.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-bioasq-test \ run.beir.contriever-msmarco.bioasq.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-bioasq.bge-base-en-v1.5 \ --topics beir-v1.0.0-bioasq-test \ --output run.beir.bge-base-en-v1.5.faiss.bioasq.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-bioasq-test \ run.beir.bge-base-en-v1.5.faiss.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-bioasq-test \ run.beir.bge-base-en-v1.5.faiss.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-bioasq-test \ run.beir.bge-base-en-v1.5.faiss.bioasq.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-bioasq.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-bioasq-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-bioasq-test \ --output run.beir.cohere-embed-english-v3.0.bioasq.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-bioasq-test \ run.beir.cohere-embed-english-v3.0.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-bioasq-test \ run.beir.cohere-embed-english-v3.0.bioasq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-bioasq-test \ run.beir.cohere-embed-english-v3.0.bioasq.txt
nfcorpus 0.322 0.246 0.325 0.250 0.347 0.284 0.328 0.301 0.373 0.337 0.386 0.351
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-nfcorpus.flat \ --topics beir-v1.0.0-nfcorpus-test \ --output run.beir.bm25-flat.nfcorpus.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nfcorpus-test \ run.beir.bm25-flat.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nfcorpus-test \ run.beir.bm25-flat.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nfcorpus-test \ run.beir.bm25-flat.nfcorpus.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-nfcorpus.multifield \ --topics beir-v1.0.0-nfcorpus-test \ --output run.beir.bm25-multifield.nfcorpus.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nfcorpus-test \ run.beir.bm25-multifield.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nfcorpus-test \ run.beir.bm25-multifield.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nfcorpus-test \ run.beir.bm25-multifield.nfcorpus.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-nfcorpus.splade-pp-ed \ --topics beir-v1.0.0-nfcorpus.test.splade-pp-ed \ --output run.beir.splade-pp-ed.nfcorpus.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nfcorpus-test \ run.beir.splade-pp-ed.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nfcorpus-test \ run.beir.splade-pp-ed.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nfcorpus-test \ run.beir.splade-pp-ed.nfcorpus.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-nfcorpus.contriever-msmarco \ --topics beir-v1.0.0-nfcorpus-test \ --output run.beir.contriever-msmarco.nfcorpus.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nfcorpus-test \ run.beir.contriever-msmarco.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nfcorpus-test \ run.beir.contriever-msmarco.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nfcorpus-test \ run.beir.contriever-msmarco.nfcorpus.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-nfcorpus.bge-base-en-v1.5 \ --topics beir-v1.0.0-nfcorpus-test \ --output run.beir.bge-base-en-v1.5.faiss.nfcorpus.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nfcorpus-test \ run.beir.bge-base-en-v1.5.faiss.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nfcorpus-test \ run.beir.bge-base-en-v1.5.faiss.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nfcorpus-test \ run.beir.bge-base-en-v1.5.faiss.nfcorpus.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-nfcorpus.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-nfcorpus-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-nfcorpus-test \ --output run.beir.cohere-embed-english-v3.0.nfcorpus.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nfcorpus-test \ run.beir.cohere-embed-english-v3.0.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nfcorpus-test \ run.beir.cohere-embed-english-v3.0.nfcorpus.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nfcorpus-test \ run.beir.cohere-embed-english-v3.0.nfcorpus.txt
nq 0.305 0.751 0.329 0.760 0.538 0.930 0.498 0.925 0.541 0.942 0.616 0.956
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-nq.flat \ --topics beir-v1.0.0-nq-test \ --output run.beir.bm25-flat.nq.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nq-test \ run.beir.bm25-flat.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nq-test \ run.beir.bm25-flat.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nq-test \ run.beir.bm25-flat.nq.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-nq.multifield \ --topics beir-v1.0.0-nq-test \ --output run.beir.bm25-multifield.nq.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nq-test \ run.beir.bm25-multifield.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nq-test \ run.beir.bm25-multifield.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nq-test \ run.beir.bm25-multifield.nq.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-nq.splade-pp-ed \ --topics beir-v1.0.0-nq.test.splade-pp-ed \ --output run.beir.splade-pp-ed.nq.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nq-test \ run.beir.splade-pp-ed.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nq-test \ run.beir.splade-pp-ed.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nq-test \ run.beir.splade-pp-ed.nq.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-nq.contriever-msmarco \ --topics beir-v1.0.0-nq-test \ --output run.beir.contriever-msmarco.nq.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nq-test \ run.beir.contriever-msmarco.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nq-test \ run.beir.contriever-msmarco.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nq-test \ run.beir.contriever-msmarco.nq.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-nq.bge-base-en-v1.5 \ --topics beir-v1.0.0-nq-test \ --output run.beir.bge-base-en-v1.5.faiss.nq.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nq-test \ run.beir.bge-base-en-v1.5.faiss.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nq-test \ run.beir.bge-base-en-v1.5.faiss.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nq-test \ run.beir.bge-base-en-v1.5.faiss.nq.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-nq.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-nq-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-nq-test \ --output run.beir.cohere-embed-english-v3.0.nq.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-nq-test \ run.beir.cohere-embed-english-v3.0.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-nq-test \ run.beir.cohere-embed-english-v3.0.nq.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-nq-test \ run.beir.cohere-embed-english-v3.0.nq.txt
hotpotqa 0.633 0.796 0.603 0.740 0.687 0.818 0.638 0.777 0.726 0.873 0.707 0.823
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-hotpotqa.flat \ --topics beir-v1.0.0-hotpotqa-test \ --output run.beir.bm25-flat.hotpotqa.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-hotpotqa-test \ run.beir.bm25-flat.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-hotpotqa-test \ run.beir.bm25-flat.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-hotpotqa-test \ run.beir.bm25-flat.hotpotqa.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-hotpotqa.multifield \ --topics beir-v1.0.0-hotpotqa-test \ --output run.beir.bm25-multifield.hotpotqa.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-hotpotqa-test \ run.beir.bm25-multifield.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-hotpotqa-test \ run.beir.bm25-multifield.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-hotpotqa-test \ run.beir.bm25-multifield.hotpotqa.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-hotpotqa.splade-pp-ed \ --topics beir-v1.0.0-hotpotqa.test.splade-pp-ed \ --output run.beir.splade-pp-ed.hotpotqa.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-hotpotqa-test \ run.beir.splade-pp-ed.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-hotpotqa-test \ run.beir.splade-pp-ed.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-hotpotqa-test \ run.beir.splade-pp-ed.hotpotqa.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-hotpotqa.contriever-msmarco \ --topics beir-v1.0.0-hotpotqa-test \ --output run.beir.contriever-msmarco.hotpotqa.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-hotpotqa-test \ run.beir.contriever-msmarco.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-hotpotqa-test \ run.beir.contriever-msmarco.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-hotpotqa-test \ run.beir.contriever-msmarco.hotpotqa.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-hotpotqa.bge-base-en-v1.5 \ --topics beir-v1.0.0-hotpotqa-test \ --output run.beir.bge-base-en-v1.5.faiss.hotpotqa.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-hotpotqa-test \ run.beir.bge-base-en-v1.5.faiss.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-hotpotqa-test \ run.beir.bge-base-en-v1.5.faiss.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-hotpotqa-test \ run.beir.bge-base-en-v1.5.faiss.hotpotqa.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-hotpotqa.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-hotpotqa-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-hotpotqa-test \ --output run.beir.cohere-embed-english-v3.0.hotpotqa.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-hotpotqa-test \ run.beir.cohere-embed-english-v3.0.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-hotpotqa-test \ run.beir.cohere-embed-english-v3.0.hotpotqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-hotpotqa-test \ run.beir.cohere-embed-english-v3.0.hotpotqa.txt
fiqa 0.236 0.539 0.236 0.539 0.347 0.631 0.329 0.656 0.406 0.742 0.421 0.736
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-fiqa.flat \ --topics beir-v1.0.0-fiqa-test \ --output run.beir.bm25-flat.fiqa.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fiqa-test \ run.beir.bm25-flat.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fiqa-test \ run.beir.bm25-flat.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fiqa-test \ run.beir.bm25-flat.fiqa.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-fiqa.multifield \ --topics beir-v1.0.0-fiqa-test \ --output run.beir.bm25-multifield.fiqa.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fiqa-test \ run.beir.bm25-multifield.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fiqa-test \ run.beir.bm25-multifield.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fiqa-test \ run.beir.bm25-multifield.fiqa.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-fiqa.splade-pp-ed \ --topics beir-v1.0.0-fiqa.test.splade-pp-ed \ --output run.beir.splade-pp-ed.fiqa.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fiqa-test \ run.beir.splade-pp-ed.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fiqa-test \ run.beir.splade-pp-ed.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fiqa-test \ run.beir.splade-pp-ed.fiqa.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-fiqa.contriever-msmarco \ --topics beir-v1.0.0-fiqa-test \ --output run.beir.contriever-msmarco.fiqa.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fiqa-test \ run.beir.contriever-msmarco.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fiqa-test \ run.beir.contriever-msmarco.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fiqa-test \ run.beir.contriever-msmarco.fiqa.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-fiqa.bge-base-en-v1.5 \ --topics beir-v1.0.0-fiqa-test \ --output run.beir.bge-base-en-v1.5.faiss.fiqa.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fiqa-test \ run.beir.bge-base-en-v1.5.faiss.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fiqa-test \ run.beir.bge-base-en-v1.5.faiss.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fiqa-test \ run.beir.bge-base-en-v1.5.faiss.fiqa.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-fiqa.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-fiqa-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-fiqa-test \ --output run.beir.cohere-embed-english-v3.0.fiqa.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fiqa-test \ run.beir.cohere-embed-english-v3.0.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fiqa-test \ run.beir.cohere-embed-english-v3.0.fiqa.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fiqa-test \ run.beir.cohere-embed-english-v3.0.fiqa.txt
signal1m 0.330 0.370 0.330 0.370 0.301 0.340 0.278 0.322 0.289 0.311 0.263 0.283
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-signal1m.flat \ --topics beir-v1.0.0-signal1m-test \ --output run.beir.bm25-flat.signal1m.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-signal1m-test \ run.beir.bm25-flat.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-signal1m-test \ run.beir.bm25-flat.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-signal1m-test \ run.beir.bm25-flat.signal1m.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-signal1m.multifield \ --topics beir-v1.0.0-signal1m-test \ --output run.beir.bm25-multifield.signal1m.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-signal1m-test \ run.beir.bm25-multifield.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-signal1m-test \ run.beir.bm25-multifield.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-signal1m-test \ run.beir.bm25-multifield.signal1m.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-signal1m.splade-pp-ed \ --topics beir-v1.0.0-signal1m.test.splade-pp-ed \ --output run.beir.splade-pp-ed.signal1m.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-signal1m-test \ run.beir.splade-pp-ed.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-signal1m-test \ run.beir.splade-pp-ed.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-signal1m-test \ run.beir.splade-pp-ed.signal1m.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-signal1m.contriever-msmarco \ --topics beir-v1.0.0-signal1m-test \ --output run.beir.contriever-msmarco.signal1m.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-signal1m-test \ run.beir.contriever-msmarco.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-signal1m-test \ run.beir.contriever-msmarco.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-signal1m-test \ run.beir.contriever-msmarco.signal1m.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-signal1m.bge-base-en-v1.5 \ --topics beir-v1.0.0-signal1m-test \ --output run.beir.bge-base-en-v1.5.faiss.signal1m.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-signal1m-test \ run.beir.bge-base-en-v1.5.faiss.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-signal1m-test \ run.beir.bge-base-en-v1.5.faiss.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-signal1m-test \ run.beir.bge-base-en-v1.5.faiss.signal1m.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-signal1m.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-signal1m-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-signal1m-test \ --output run.beir.cohere-embed-english-v3.0.signal1m.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-signal1m-test \ run.beir.cohere-embed-english-v3.0.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-signal1m-test \ run.beir.cohere-embed-english-v3.0.signal1m.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-signal1m-test \ run.beir.cohere-embed-english-v3.0.signal1m.txt
trec-news 0.395 0.447 0.398 0.422 0.415 0.441 0.428 0.492 0.442 0.499 0.504 0.543
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-trec-news.flat \ --topics beir-v1.0.0-trec-news-test \ --output run.beir.bm25-flat.trec-news.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-news-test \ run.beir.bm25-flat.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-news-test \ run.beir.bm25-flat.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-news-test \ run.beir.bm25-flat.trec-news.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-trec-news.multifield \ --topics beir-v1.0.0-trec-news-test \ --output run.beir.bm25-multifield.trec-news.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-news-test \ run.beir.bm25-multifield.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-news-test \ run.beir.bm25-multifield.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-news-test \ run.beir.bm25-multifield.trec-news.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-trec-news.splade-pp-ed \ --topics beir-v1.0.0-trec-news.test.splade-pp-ed \ --output run.beir.splade-pp-ed.trec-news.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-news-test \ run.beir.splade-pp-ed.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-news-test \ run.beir.splade-pp-ed.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-news-test \ run.beir.splade-pp-ed.trec-news.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-trec-news.contriever-msmarco \ --topics beir-v1.0.0-trec-news-test \ --output run.beir.contriever-msmarco.trec-news.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-news-test \ run.beir.contriever-msmarco.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-news-test \ run.beir.contriever-msmarco.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-news-test \ run.beir.contriever-msmarco.trec-news.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-trec-news.bge-base-en-v1.5 \ --topics beir-v1.0.0-trec-news-test \ --output run.beir.bge-base-en-v1.5.faiss.trec-news.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-news-test \ run.beir.bge-base-en-v1.5.faiss.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-news-test \ run.beir.bge-base-en-v1.5.faiss.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-news-test \ run.beir.bge-base-en-v1.5.faiss.trec-news.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-trec-news.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-trec-news-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-trec-news-test \ --output run.beir.cohere-embed-english-v3.0.trec-news.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-trec-news-test \ run.beir.cohere-embed-english-v3.0.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-trec-news-test \ run.beir.cohere-embed-english-v3.0.trec-news.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-trec-news-test \ run.beir.cohere-embed-english-v3.0.trec-news.txt
robust04 0.407 0.375 0.407 0.375 0.468 0.385 0.473 0.392 0.444 0.351 0.541 0.417
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-robust04.flat \ --topics beir-v1.0.0-robust04-test \ --output run.beir.bm25-flat.robust04.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-robust04-test \ run.beir.bm25-flat.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-robust04-test \ run.beir.bm25-flat.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-robust04-test \ run.beir.bm25-flat.robust04.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-robust04.multifield \ --topics beir-v1.0.0-robust04-test \ --output run.beir.bm25-multifield.robust04.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-robust04-test \ run.beir.bm25-multifield.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-robust04-test \ run.beir.bm25-multifield.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-robust04-test \ run.beir.bm25-multifield.robust04.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-robust04.splade-pp-ed \ --topics beir-v1.0.0-robust04.test.splade-pp-ed \ --output run.beir.splade-pp-ed.robust04.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-robust04-test \ run.beir.splade-pp-ed.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-robust04-test \ run.beir.splade-pp-ed.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-robust04-test \ run.beir.splade-pp-ed.robust04.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-robust04.contriever-msmarco \ --topics beir-v1.0.0-robust04-test \ --output run.beir.contriever-msmarco.robust04.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-robust04-test \ run.beir.contriever-msmarco.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-robust04-test \ run.beir.contriever-msmarco.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-robust04-test \ run.beir.contriever-msmarco.robust04.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-robust04.bge-base-en-v1.5 \ --topics beir-v1.0.0-robust04-test \ --output run.beir.bge-base-en-v1.5.faiss.robust04.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-robust04-test \ run.beir.bge-base-en-v1.5.faiss.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-robust04-test \ run.beir.bge-base-en-v1.5.faiss.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-robust04-test \ run.beir.bge-base-en-v1.5.faiss.robust04.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-robust04.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-robust04-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-robust04-test \ --output run.beir.cohere-embed-english-v3.0.robust04.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-robust04-test \ run.beir.cohere-embed-english-v3.0.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-robust04-test \ run.beir.cohere-embed-english-v3.0.robust04.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-robust04-test \ run.beir.cohere-embed-english-v3.0.robust04.txt
arguana 0.397 0.932 0.414 0.943 0.520 0.974 0.446 0.977 0.636 0.992 0.540 0.982
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-arguana.flat \ --topics beir-v1.0.0-arguana-test \ --output run.beir.bm25-flat.arguana.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-arguana-test \ run.beir.bm25-flat.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-arguana-test \ run.beir.bm25-flat.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-arguana-test \ run.beir.bm25-flat.arguana.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-arguana.multifield \ --topics beir-v1.0.0-arguana-test \ --output run.beir.bm25-multifield.arguana.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-arguana-test \ run.beir.bm25-multifield.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-arguana-test \ run.beir.bm25-multifield.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-arguana-test \ run.beir.bm25-multifield.arguana.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-arguana.splade-pp-ed \ --topics beir-v1.0.0-arguana.test.splade-pp-ed \ --output run.beir.splade-pp-ed.arguana.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-arguana-test \ run.beir.splade-pp-ed.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-arguana-test \ run.beir.splade-pp-ed.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-arguana-test \ run.beir.splade-pp-ed.arguana.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-arguana.contriever-msmarco \ --topics beir-v1.0.0-arguana-test \ --output run.beir.contriever-msmarco.arguana.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-arguana-test \ run.beir.contriever-msmarco.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-arguana-test \ run.beir.contriever-msmarco.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-arguana-test \ run.beir.contriever-msmarco.arguana.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "" \ --index beir-v1.0.0-arguana.bge-base-en-v1.5 \ --topics beir-v1.0.0-arguana-test \ --output run.beir.bge-base-en-v1.5.faiss.arguana.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-arguana-test \ run.beir.bge-base-en-v1.5.faiss.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-arguana-test \ run.beir.bge-base-en-v1.5.faiss.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-arguana-test \ run.beir.bge-base-en-v1.5.faiss.arguana.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-arguana.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-arguana-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-arguana-test \ --output run.beir.cohere-embed-english-v3.0.arguana.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-arguana-test \ run.beir.cohere-embed-english-v3.0.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-arguana-test \ run.beir.cohere-embed-english-v3.0.arguana.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-arguana-test \ run.beir.cohere-embed-english-v3.0.arguana.txt
webis-touche2020 0.442 0.582 0.367 0.538 0.247 0.471 0.204 0.442 0.257 0.487 0.326 0.516
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-webis-touche2020.flat \ --topics beir-v1.0.0-webis-touche2020-test \ --output run.beir.bm25-flat.webis-touche2020.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-webis-touche2020-test \ run.beir.bm25-flat.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-webis-touche2020-test \ run.beir.bm25-flat.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-webis-touche2020-test \ run.beir.bm25-flat.webis-touche2020.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-webis-touche2020.multifield \ --topics beir-v1.0.0-webis-touche2020-test \ --output run.beir.bm25-multifield.webis-touche2020.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-webis-touche2020-test \ run.beir.bm25-multifield.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-webis-touche2020-test \ run.beir.bm25-multifield.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-webis-touche2020-test \ run.beir.bm25-multifield.webis-touche2020.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-webis-touche2020.splade-pp-ed \ --topics beir-v1.0.0-webis-touche2020.test.splade-pp-ed \ --output run.beir.splade-pp-ed.webis-touche2020.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-webis-touche2020-test \ run.beir.splade-pp-ed.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-webis-touche2020-test \ run.beir.splade-pp-ed.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-webis-touche2020-test \ run.beir.splade-pp-ed.webis-touche2020.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-webis-touche2020.contriever-msmarco \ --topics beir-v1.0.0-webis-touche2020-test \ --output run.beir.contriever-msmarco.webis-touche2020.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-webis-touche2020-test \ run.beir.contriever-msmarco.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-webis-touche2020-test \ run.beir.contriever-msmarco.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-webis-touche2020-test \ run.beir.contriever-msmarco.webis-touche2020.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-webis-touche2020.bge-base-en-v1.5 \ --topics beir-v1.0.0-webis-touche2020-test \ --output run.beir.bge-base-en-v1.5.faiss.webis-touche2020.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-webis-touche2020-test \ run.beir.bge-base-en-v1.5.faiss.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-webis-touche2020-test \ run.beir.bge-base-en-v1.5.faiss.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-webis-touche2020-test \ run.beir.bge-base-en-v1.5.faiss.webis-touche2020.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-webis-touche2020.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-webis-touche2020-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-webis-touche2020-test \ --output run.beir.cohere-embed-english-v3.0.webis-touche2020.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-webis-touche2020-test \ run.beir.cohere-embed-english-v3.0.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-webis-touche2020-test \ run.beir.cohere-embed-english-v3.0.webis-touche2020.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-webis-touche2020-test \ run.beir.cohere-embed-english-v3.0.webis-touche2020.txt
cqadupstack (average) 0.302 0.580 0.299 0.606 0.334 0.650 0.345 0.663 0.424 0.762 0.415 0.745
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-android.flat \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.bm25-flat.cqadupstack-android.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-english.flat \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.bm25-flat.cqadupstack-english.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gaming.flat \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.bm25-flat.cqadupstack-gaming.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gis.flat \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.bm25-flat.cqadupstack-gis.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-mathematica.flat \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.bm25-flat.cqadupstack-mathematica.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-physics.flat \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.bm25-flat.cqadupstack-physics.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-programmers.flat \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.bm25-flat.cqadupstack-programmers.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-stats.flat \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.bm25-flat.cqadupstack-stats.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-tex.flat \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.bm25-flat.cqadupstack-tex.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-unix.flat \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.bm25-flat.cqadupstack-unix.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-webmasters.flat \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.bm25-flat.cqadupstack-webmasters.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-wordpress.flat \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.bm25-flat.cqadupstack-wordpress.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-android.multifield \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.bm25-multifield.cqadupstack-android.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-english.multifield \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.bm25-multifield.cqadupstack-english.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gaming.multifield \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.bm25-multifield.cqadupstack-gaming.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gis.multifield \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.bm25-multifield.cqadupstack-gis.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-mathematica.multifield \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.bm25-multifield.cqadupstack-mathematica.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-physics.multifield \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.bm25-multifield.cqadupstack-physics.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-programmers.multifield \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.bm25-multifield.cqadupstack-programmers.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-stats.multifield \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.bm25-multifield.cqadupstack-stats.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-tex.multifield \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.bm25-multifield.cqadupstack-tex.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-unix.multifield \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.bm25-multifield.cqadupstack-unix.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-webmasters.multifield \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.bm25-multifield.cqadupstack-webmasters.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-wordpress.multifield \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.bm25-multifield.cqadupstack-wordpress.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-android.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-android.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-android.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-english.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-english.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-english.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gaming.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-gaming.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-gaming.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gis.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-gis.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-gis.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-mathematica.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-mathematica.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-mathematica.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-physics.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-physics.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-physics.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-programmers.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-programmers.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-programmers.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-stats.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-stats.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-stats.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-tex.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-tex.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-tex.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-unix.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-unix.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-unix.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-webmasters.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-webmasters.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-webmasters.txt \ --output-format trec \ --hits 1000 --impact --remove-query python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-wordpress.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-wordpress.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-wordpress.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-android.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.contriever-msmarco.cqadupstack-android.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-english.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.contriever-msmarco.cqadupstack-english.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-gaming.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.contriever-msmarco.cqadupstack-gaming.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-gis.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.contriever-msmarco.cqadupstack-gis.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-mathematica.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.contriever-msmarco.cqadupstack-mathematica.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-physics.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.contriever-msmarco.cqadupstack-physics.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-programmers.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.contriever-msmarco.cqadupstack-programmers.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-stats.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.contriever-msmarco.cqadupstack-stats.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-tex.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.contriever-msmarco.cqadupstack-tex.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-unix.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.contriever-msmarco.cqadupstack-unix.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-webmasters.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.contriever-msmarco.cqadupstack-webmasters.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-wordpress.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.contriever-msmarco.cqadupstack-wordpress.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-android.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-english.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-gaming.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-gis.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-mathematica.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-physics.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-programmers.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-stats.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-tex.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-unix.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-webmasters.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-wordpress.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-android.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-android-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-android-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-english.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-english-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-english-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-gaming.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-gis.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-gis-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-mathematica.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-physics.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-physics-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-programmers.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-stats.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-stats-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-tex.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-tex-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-unix.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-unix-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-webmasters.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt \ --hits 1000 --remove-query python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-wordpress.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt
quora 0.789 0.973 0.789 0.973 0.834 0.986 0.865 0.994 0.889 0.997 0.887 0.996
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-quora.flat \ --topics beir-v1.0.0-quora-test \ --output run.beir.bm25-flat.quora.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-quora-test \ run.beir.bm25-flat.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-quora-test \ run.beir.bm25-flat.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-quora-test \ run.beir.bm25-flat.quora.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-quora.multifield \ --topics beir-v1.0.0-quora-test \ --output run.beir.bm25-multifield.quora.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-quora-test \ run.beir.bm25-multifield.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-quora-test \ run.beir.bm25-multifield.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-quora-test \ run.beir.bm25-multifield.quora.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-quora.splade-pp-ed \ --topics beir-v1.0.0-quora.test.splade-pp-ed \ --output run.beir.splade-pp-ed.quora.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-quora-test \ run.beir.splade-pp-ed.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-quora-test \ run.beir.splade-pp-ed.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-quora-test \ run.beir.splade-pp-ed.quora.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-quora.contriever-msmarco \ --topics beir-v1.0.0-quora-test \ --output run.beir.contriever-msmarco.quora.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-quora-test \ run.beir.contriever-msmarco.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-quora-test \ run.beir.contriever-msmarco.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-quora-test \ run.beir.contriever-msmarco.quora.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "" \ --index beir-v1.0.0-quora.bge-base-en-v1.5 \ --topics beir-v1.0.0-quora-test \ --output run.beir.bge-base-en-v1.5.faiss.quora.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-quora-test \ run.beir.bge-base-en-v1.5.faiss.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-quora-test \ run.beir.bge-base-en-v1.5.faiss.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-quora-test \ run.beir.bge-base-en-v1.5.faiss.quora.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-quora.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-quora-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-quora-test \ --output run.beir.cohere-embed-english-v3.0.quora.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-quora-test \ run.beir.cohere-embed-english-v3.0.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-quora-test \ run.beir.cohere-embed-english-v3.0.quora.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-quora-test \ run.beir.cohere-embed-english-v3.0.quora.txt
dbpedia-entity 0.318 0.468 0.313 0.398 0.437 0.562 0.413 0.541 0.407 0.530 0.434 0.536
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-dbpedia-entity.flat \ --topics beir-v1.0.0-dbpedia-entity-test \ --output run.beir.bm25-flat.dbpedia-entity.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-dbpedia-entity-test \ run.beir.bm25-flat.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-dbpedia-entity-test \ run.beir.bm25-flat.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-dbpedia-entity-test \ run.beir.bm25-flat.dbpedia-entity.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-dbpedia-entity.multifield \ --topics beir-v1.0.0-dbpedia-entity-test \ --output run.beir.bm25-multifield.dbpedia-entity.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-dbpedia-entity-test \ run.beir.bm25-multifield.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-dbpedia-entity-test \ run.beir.bm25-multifield.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-dbpedia-entity-test \ run.beir.bm25-multifield.dbpedia-entity.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-dbpedia-entity.splade-pp-ed \ --topics beir-v1.0.0-dbpedia-entity.test.splade-pp-ed \ --output run.beir.splade-pp-ed.dbpedia-entity.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-dbpedia-entity-test \ run.beir.splade-pp-ed.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-dbpedia-entity-test \ run.beir.splade-pp-ed.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-dbpedia-entity-test \ run.beir.splade-pp-ed.dbpedia-entity.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-dbpedia-entity.contriever-msmarco \ --topics beir-v1.0.0-dbpedia-entity-test \ --output run.beir.contriever-msmarco.dbpedia-entity.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-dbpedia-entity-test \ run.beir.contriever-msmarco.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-dbpedia-entity-test \ run.beir.contriever-msmarco.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-dbpedia-entity-test \ run.beir.contriever-msmarco.dbpedia-entity.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-dbpedia-entity.bge-base-en-v1.5 \ --topics beir-v1.0.0-dbpedia-entity-test \ --output run.beir.bge-base-en-v1.5.faiss.dbpedia-entity.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-dbpedia-entity-test \ run.beir.bge-base-en-v1.5.faiss.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-dbpedia-entity-test \ run.beir.bge-base-en-v1.5.faiss.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-dbpedia-entity-test \ run.beir.bge-base-en-v1.5.faiss.dbpedia-entity.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-dbpedia-entity.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-dbpedia-entity-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-dbpedia-entity-test \ --output run.beir.cohere-embed-english-v3.0.dbpedia-entity.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-dbpedia-entity-test \ run.beir.cohere-embed-english-v3.0.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-dbpedia-entity-test \ run.beir.cohere-embed-english-v3.0.dbpedia-entity.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-dbpedia-entity-test \ run.beir.cohere-embed-english-v3.0.dbpedia-entity.txt
scidocs 0.149 0.348 0.158 0.356 0.159 0.373 0.165 0.378 0.217 0.496 0.203 0.451
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-scidocs.flat \ --topics beir-v1.0.0-scidocs-test \ --output run.beir.bm25-flat.scidocs.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scidocs-test \ run.beir.bm25-flat.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scidocs-test \ run.beir.bm25-flat.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scidocs-test \ run.beir.bm25-flat.scidocs.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-scidocs.multifield \ --topics beir-v1.0.0-scidocs-test \ --output run.beir.bm25-multifield.scidocs.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scidocs-test \ run.beir.bm25-multifield.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scidocs-test \ run.beir.bm25-multifield.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scidocs-test \ run.beir.bm25-multifield.scidocs.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-scidocs.splade-pp-ed \ --topics beir-v1.0.0-scidocs.test.splade-pp-ed \ --output run.beir.splade-pp-ed.scidocs.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scidocs-test \ run.beir.splade-pp-ed.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scidocs-test \ run.beir.splade-pp-ed.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scidocs-test \ run.beir.splade-pp-ed.scidocs.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-scidocs.contriever-msmarco \ --topics beir-v1.0.0-scidocs-test \ --output run.beir.contriever-msmarco.scidocs.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scidocs-test \ run.beir.contriever-msmarco.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scidocs-test \ run.beir.contriever-msmarco.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scidocs-test \ run.beir.contriever-msmarco.scidocs.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-scidocs.bge-base-en-v1.5 \ --topics beir-v1.0.0-scidocs-test \ --output run.beir.bge-base-en-v1.5.faiss.scidocs.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scidocs-test \ run.beir.bge-base-en-v1.5.faiss.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scidocs-test \ run.beir.bge-base-en-v1.5.faiss.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scidocs-test \ run.beir.bge-base-en-v1.5.faiss.scidocs.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-scidocs.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-scidocs-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-scidocs-test \ --output run.beir.cohere-embed-english-v3.0.scidocs.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scidocs-test \ run.beir.cohere-embed-english-v3.0.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scidocs-test \ run.beir.cohere-embed-english-v3.0.scidocs.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scidocs-test \ run.beir.cohere-embed-english-v3.0.scidocs.txt
fever 0.651 0.918 0.753 0.931 0.788 0.946 0.758 0.949 0.863 0.972 0.890 0.965
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-fever.flat \ --topics beir-v1.0.0-fever-test \ --output run.beir.bm25-flat.fever.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fever-test \ run.beir.bm25-flat.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fever-test \ run.beir.bm25-flat.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fever-test \ run.beir.bm25-flat.fever.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-fever.multifield \ --topics beir-v1.0.0-fever-test \ --output run.beir.bm25-multifield.fever.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fever-test \ run.beir.bm25-multifield.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fever-test \ run.beir.bm25-multifield.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fever-test \ run.beir.bm25-multifield.fever.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-fever.splade-pp-ed \ --topics beir-v1.0.0-fever.test.splade-pp-ed \ --output run.beir.splade-pp-ed.fever.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fever-test \ run.beir.splade-pp-ed.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fever-test \ run.beir.splade-pp-ed.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fever-test \ run.beir.splade-pp-ed.fever.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-fever.contriever-msmarco \ --topics beir-v1.0.0-fever-test \ --output run.beir.contriever-msmarco.fever.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fever-test \ run.beir.contriever-msmarco.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fever-test \ run.beir.contriever-msmarco.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fever-test \ run.beir.contriever-msmarco.fever.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-fever.bge-base-en-v1.5 \ --topics beir-v1.0.0-fever-test \ --output run.beir.bge-base-en-v1.5.faiss.fever.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fever-test \ run.beir.bge-base-en-v1.5.faiss.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fever-test \ run.beir.bge-base-en-v1.5.faiss.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fever-test \ run.beir.bge-base-en-v1.5.faiss.fever.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-fever.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-fever-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-fever-test \ --output run.beir.cohere-embed-english-v3.0.fever.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-fever-test \ run.beir.cohere-embed-english-v3.0.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-fever-test \ run.beir.cohere-embed-english-v3.0.fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-fever-test \ run.beir.cohere-embed-english-v3.0.fever.txt
climate-fever 0.165 0.425 0.213 0.436 0.230 0.521 0.237 0.575 0.312 0.636 0.259 0.581
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-climate-fever.flat \ --topics beir-v1.0.0-climate-fever-test \ --output run.beir.bm25-flat.climate-fever.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-climate-fever-test \ run.beir.bm25-flat.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-climate-fever-test \ run.beir.bm25-flat.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-climate-fever-test \ run.beir.bm25-flat.climate-fever.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-climate-fever.multifield \ --topics beir-v1.0.0-climate-fever-test \ --output run.beir.bm25-multifield.climate-fever.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-climate-fever-test \ run.beir.bm25-multifield.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-climate-fever-test \ run.beir.bm25-multifield.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-climate-fever-test \ run.beir.bm25-multifield.climate-fever.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-climate-fever.splade-pp-ed \ --topics beir-v1.0.0-climate-fever.test.splade-pp-ed \ --output run.beir.splade-pp-ed.climate-fever.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-climate-fever-test \ run.beir.splade-pp-ed.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-climate-fever-test \ run.beir.splade-pp-ed.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-climate-fever-test \ run.beir.splade-pp-ed.climate-fever.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-climate-fever.contriever-msmarco \ --topics beir-v1.0.0-climate-fever-test \ --output run.beir.contriever-msmarco.climate-fever.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-climate-fever-test \ run.beir.contriever-msmarco.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-climate-fever-test \ run.beir.contriever-msmarco.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-climate-fever-test \ run.beir.contriever-msmarco.climate-fever.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-climate-fever.bge-base-en-v1.5 \ --topics beir-v1.0.0-climate-fever-test \ --output run.beir.bge-base-en-v1.5.faiss.climate-fever.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-climate-fever-test \ run.beir.bge-base-en-v1.5.faiss.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-climate-fever-test \ run.beir.bge-base-en-v1.5.faiss.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-climate-fever-test \ run.beir.bge-base-en-v1.5.faiss.climate-fever.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-climate-fever.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-climate-fever-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-climate-fever-test \ --output run.beir.cohere-embed-english-v3.0.climate-fever.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-climate-fever-test \ run.beir.cohere-embed-english-v3.0.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-climate-fever-test \ run.beir.cohere-embed-english-v3.0.climate-fever.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-climate-fever-test \ run.beir.cohere-embed-english-v3.0.climate-fever.txt
scifact 0.679 0.925 0.665 0.908 0.704 0.935 0.677 0.947 0.741 0.967 0.718 0.963
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-scifact.flat \ --topics beir-v1.0.0-scifact-test \ --output run.beir.bm25-flat.scifact.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scifact-test \ run.beir.bm25-flat.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scifact-test \ run.beir.bm25-flat.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scifact-test \ run.beir.bm25-flat.scifact.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-scifact.multifield \ --topics beir-v1.0.0-scifact-test \ --output run.beir.bm25-multifield.scifact.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scifact-test \ run.beir.bm25-multifield.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scifact-test \ run.beir.bm25-multifield.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scifact-test \ run.beir.bm25-multifield.scifact.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-scifact.splade-pp-ed \ --topics beir-v1.0.0-scifact.test.splade-pp-ed \ --output run.beir.splade-pp-ed.scifact.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scifact-test \ run.beir.splade-pp-ed.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scifact-test \ run.beir.splade-pp-ed.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scifact-test \ run.beir.splade-pp-ed.scifact.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-scifact.contriever-msmarco \ --topics beir-v1.0.0-scifact-test \ --output run.beir.contriever-msmarco.scifact.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scifact-test \ run.beir.contriever-msmarco.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scifact-test \ run.beir.contriever-msmarco.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scifact-test \ run.beir.contriever-msmarco.scifact.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-scifact.bge-base-en-v1.5 \ --topics beir-v1.0.0-scifact-test \ --output run.beir.bge-base-en-v1.5.faiss.scifact.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scifact-test \ run.beir.bge-base-en-v1.5.faiss.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scifact-test \ run.beir.bge-base-en-v1.5.faiss.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scifact-test \ run.beir.bge-base-en-v1.5.faiss.scifact.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-scifact.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-scifact-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-scifact-test \ --output run.beir.cohere-embed-english-v3.0.scifact.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-scifact-test \ run.beir.cohere-embed-english-v3.0.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-scifact-test \ run.beir.cohere-embed-english-v3.0.scifact.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-scifact-test \ run.beir.cohere-embed-english-v3.0.scifact.txt

CQADupStack Results

The "cqadupstack (average)" row above represents the average across the individual CQADupStack datasets, shown below.

BM25 Flat BM25 MF SPLADE Contriever BGE Cohere
nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100 nDCG@10 R@100
android 0.380 0.683 0.371 0.689 0.390 0.740 0.425 0.750 0.508 0.845 0.500 0.832
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-android.flat \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.bm25-flat.cqadupstack-android.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-flat.cqadupstack-android.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-android.multifield \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.bm25-multifield.cqadupstack-android.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.bm25-multifield.cqadupstack-android.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-android.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-android.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-android.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.splade-pp-ed.cqadupstack-android.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-android.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.contriever-msmarco.cqadupstack-android.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.contriever-msmarco.cqadupstack-android.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-android.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-android-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-android.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-android.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-android-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-android-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-android-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-android.txt
english 0.345 0.576 0.332 0.584 0.408 0.695 0.433 0.694 0.486 0.759 0.491 0.757
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-english.flat \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.bm25-flat.cqadupstack-english.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-flat.cqadupstack-english.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-english.multifield \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.bm25-multifield.cqadupstack-english.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.bm25-multifield.cqadupstack-english.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-english.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-english.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-english.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.splade-pp-ed.cqadupstack-english.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-english.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.contriever-msmarco.cqadupstack-english.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.contriever-msmarco.cqadupstack-english.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-english.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-english-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-english.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-english.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-english-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-english-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-english-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-english.txt
gaming 0.482 0.765 0.442 0.757 0.496 0.813 0.528 0.848 0.597 0.904 0.605 0.900
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gaming.flat \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.bm25-flat.cqadupstack-gaming.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-flat.cqadupstack-gaming.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gaming.multifield \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.bm25-multifield.cqadupstack-gaming.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bm25-multifield.cqadupstack-gaming.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gaming.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-gaming.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-gaming.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.splade-pp-ed.cqadupstack-gaming.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-gaming.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.contriever-msmarco.cqadupstack-gaming.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.contriever-msmarco.cqadupstack-gaming.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-gaming.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gaming.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-gaming.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-gaming-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-gaming-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gaming-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gaming.txt
gis 0.290 0.612 0.290 0.646 0.315 0.632 0.302 0.627 0.413 0.768 0.392 0.744
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gis.flat \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.bm25-flat.cqadupstack-gis.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-flat.cqadupstack-gis.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gis.multifield \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.bm25-multifield.cqadupstack-gis.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bm25-multifield.cqadupstack-gis.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-gis.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-gis.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-gis.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.splade-pp-ed.cqadupstack-gis.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-gis.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.contriever-msmarco.cqadupstack-gis.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.contriever-msmarco.cqadupstack-gis.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-gis.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-gis.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-gis.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-gis-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-gis-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-gis-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-gis.txt
mathematica 0.202 0.488 0.205 0.521 0.238 0.580 0.235 0.573 0.316 0.692 0.304 0.667
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-mathematica.flat \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.bm25-flat.cqadupstack-mathematica.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-flat.cqadupstack-mathematica.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-mathematica.multifield \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.bm25-multifield.cqadupstack-mathematica.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bm25-multifield.cqadupstack-mathematica.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-mathematica.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-mathematica.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-mathematica.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.splade-pp-ed.cqadupstack-mathematica.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-mathematica.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.contriever-msmarco.cqadupstack-mathematica.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.contriever-msmarco.cqadupstack-mathematica.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-mathematica.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-mathematica.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-mathematica.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-mathematica-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-mathematica-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-mathematica-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-mathematica.txt
physics 0.321 0.633 0.325 0.649 0.360 0.720 0.416 0.762 0.472 0.808 0.438 0.784
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-physics.flat \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.bm25-flat.cqadupstack-physics.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-flat.cqadupstack-physics.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-physics.multifield \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.bm25-multifield.cqadupstack-physics.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bm25-multifield.cqadupstack-physics.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-physics.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-physics.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-physics.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.splade-pp-ed.cqadupstack-physics.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-physics.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.contriever-msmarco.cqadupstack-physics.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.contriever-msmarco.cqadupstack-physics.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-physics.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-physics.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-physics.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-physics-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-physics-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-physics-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-physics.txt
programmers 0.280 0.559 0.296 0.619 0.340 0.658 0.357 0.719 0.424 0.786 0.437 0.789
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-programmers.flat \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.bm25-flat.cqadupstack-programmers.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-flat.cqadupstack-programmers.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-programmers.multifield \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.bm25-multifield.cqadupstack-programmers.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bm25-multifield.cqadupstack-programmers.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-programmers.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-programmers.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-programmers.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.splade-pp-ed.cqadupstack-programmers.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-programmers.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.contriever-msmarco.cqadupstack-programmers.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.contriever-msmarco.cqadupstack-programmers.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-programmers.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-programmers.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-programmers.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-programmers-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-programmers-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-programmers-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-programmers.txt
stats 0.271 0.534 0.279 0.572 0.299 0.589 0.309 0.586 0.373 0.673 0.352 0.643
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-stats.flat \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.bm25-flat.cqadupstack-stats.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-flat.cqadupstack-stats.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-stats.multifield \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.bm25-multifield.cqadupstack-stats.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bm25-multifield.cqadupstack-stats.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-stats.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-stats.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-stats.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.splade-pp-ed.cqadupstack-stats.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-stats.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.contriever-msmarco.cqadupstack-stats.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.contriever-msmarco.cqadupstack-stats.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-stats.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-stats.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-stats.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-stats-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-stats-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-stats-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-stats.txt
tex 0.224 0.469 0.209 0.495 0.253 0.516 0.221 0.498 0.311 0.649 0.308 0.624
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-tex.flat \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.bm25-flat.cqadupstack-tex.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-flat.cqadupstack-tex.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-tex.multifield \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.bm25-multifield.cqadupstack-tex.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bm25-multifield.cqadupstack-tex.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-tex.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-tex.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-tex.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.splade-pp-ed.cqadupstack-tex.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-tex.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.contriever-msmarco.cqadupstack-tex.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.contriever-msmarco.cqadupstack-tex.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-tex.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-tex.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-tex.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-tex-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-tex-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-tex-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-tex.txt
unix 0.275 0.542 0.279 0.572 0.317 0.621 0.326 0.616 0.422 0.780 0.406 0.754
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-unix.flat \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.bm25-flat.cqadupstack-unix.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-flat.cqadupstack-unix.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-unix.multifield \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.bm25-multifield.cqadupstack-unix.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bm25-multifield.cqadupstack-unix.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-unix.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-unix.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-unix.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.splade-pp-ed.cqadupstack-unix.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-unix.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.contriever-msmarco.cqadupstack-unix.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.contriever-msmarco.cqadupstack-unix.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-unix.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-unix.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-unix.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-unix-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-unix-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-unix-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-unix.txt
webmasters 0.306 0.582 0.301 0.610 0.317 0.636 0.339 0.703 0.407 0.777 0.407 0.749
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-webmasters.flat \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.bm25-flat.cqadupstack-webmasters.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-flat.cqadupstack-webmasters.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-webmasters.multifield \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.bm25-multifield.cqadupstack-webmasters.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bm25-multifield.cqadupstack-webmasters.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-webmasters.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-webmasters.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-webmasters.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.splade-pp-ed.cqadupstack-webmasters.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-webmasters.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.contriever-msmarco.cqadupstack-webmasters.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.contriever-msmarco.cqadupstack-webmasters.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-webmasters.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-webmasters.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-webmasters.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-webmasters-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-webmasters-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-webmasters-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-webmasters.txt
wordpress 0.248 0.515 0.256 0.553 0.273 0.595 0.253 0.577 0.355 0.705 0.343 0.694
BM25 Flat BM25 MF SPLADE Contriever BGE Cohere Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-wordpress.flat \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.bm25-flat.cqadupstack-wordpress.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-flat.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-wordpress.multifield \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.bm25-multifield.cqadupstack-wordpress.txt \ --output-format trec \ --hits 1000 --bm25 --remove-query --fields contents=1.0 title=1.0 Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bm25-multifield.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.lucene \ --threads 16 --batch-size 128 \ --index beir-v1.0.0-cqadupstack-wordpress.splade-pp-ed \ --topics beir-v1.0.0-cqadupstack-wordpress.test.splade-pp-ed \ --output run.beir.splade-pp-ed.cqadupstack-wordpress.txt \ --output-format trec \ --hits 1000 --impact --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.splade-pp-ed.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class contriever --encoder facebook/contriever-msmarco \ --index beir-v1.0.0-cqadupstack-wordpress.contriever-msmarco \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.contriever-msmarco.cqadupstack-wordpress.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.contriever-msmarco.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --encoder-class auto --encoder BAAI/bge-base-en-v1.5 --l2-norm \ --query-prefix "Represent this sentence for searching relevant passages:" \ --index beir-v1.0.0-cqadupstack-wordpress.bge-base-en-v1.5 \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.bge-base-en-v1.5.faiss.cqadupstack-wordpress.txt Command to generate run: python -m pyserini.search.faiss \ --threads 16 --batch-size 512 \ --index beir-v1.0.0-cqadupstack-wordpress.cohere-embed-english-v3.0 \ --topics beir-v1.0.0-cqadupstack-wordpress-test \ --encoded-queries cohere-embed-english-v3.0-beir-v1.0.0-cqadupstack-wordpress-test \ --output run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt \ --hits 1000 --remove-query Evaluation commands: python -m pyserini.eval.trec_eval \ -c -m ndcg_cut.10 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.100 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt python -m pyserini.eval.trec_eval \ -c -m recall.1000 beir-v1.0.0-cqadupstack-wordpress-test \ run.beir.cohere-embed-english-v3.0.cqadupstack-wordpress.txt

References

Programmatic Execution

All experimental runs shown in the above table can be programmatically executed based on the instructions below. To list all the experimental conditions:

python -m pyserini.2cr.beir --list-conditions

These conditions correspond to the table rows above.

For all conditions, just show the commands in a "dry run":

python -m pyserini.2cr.beir --all --display-commands --dry-run

To actually run all the experimental conditions:

python -m pyserini.2cr.beir --all --display-commands

With the above command, run files will be placed in the current directory. Use the option --directory runs/ to place the runs in a sub-directory.

To show the commands for a specific condition:

python -m pyserini.2cr.beir --condition bm25-flat --display-commands --dry-run

This will generate exactly the commands for a specific condition above (corresponding to a row in the table).

To actually run a specific condition:

python -m pyserini.2cr.beir --condition bm25-flat --display-commands

Again, with the above command, run files will be placed in the current directory. Use the option --directory runs/ to place the runs in a sub-directory.

Finally, to generate this page:

python -m pyserini.2cr.beir --generate-report --output beir.html

The output file beir.html should be identical to this page.