Skip to content

Latest commit

 

History

History
307 lines (226 loc) · 31.8 KB

File metadata and controls

307 lines (226 loc) · 31.8 KB


logo of awesome repository

awesome-pretrained-models-for-information-retrieval

A curated list of awesome papers related to pre-trained models for information retrieval (a.k.a., pre-training for IR). If there are any papers I missed, please let me know! And any feedback and contributions are welcome!

Pre-training for IR

For people who want to acquire some basic & advanced knowledge about neural models for information retrieval and try some neural models by hand, we refer readers to the below awesome NeuIR survey and the text-matching toolkit MatchZoo-py:

Survey Papers

First Stage Retrieval

Sparse Retrieval

Neural term re-weighting

Query or document expansion

Sparse representation learning

Dense Retrieval

Hard negative sampling

Late interaction and multi-vector representation

Knowledge distillation

Jointly learning retrieval and indexing

Domain adaptation

Pre-training tailored for dense retrieval

Dense retrieval in open domain QA

Combining Sparse Retrieval and Dense Retrieval

Re-ranking Stage

Basic Usage

Discriminative ranking models

Representation-focused
Interaction-focused

Generative ranking models

Hybrid ranking models

Long Document Processing Techniques

Passage score aggregation

Passage representation aggregation

Designing new architectures

Improving Efficiency

Decoupling the interaction

Knowledge distillation

Early exit

Re-weighting Training Samples

Query Expansion

Partial Fine-tuning

Pre-training Tailored for Re-ranking

Cross-lingual Retrieval

Jointly Learning to Retrieve and Re-rank

Model-based IR System

Multimodal Retrieval

Unified Single-stream Architecture

Multi-stream Architecture Applied on Input

Other Resources

Some Retrieval Toolkits

Other Resources About Pre-trained Models in NLP

Surveys About Efficient Transformers