Extremely Long User History Modeling at Instagram, When Compact Embeddings Outperform Large Models in Hybrid RAG, and More!
Vol.107 for Jun 02 - Jun 08, 2025
Stay Ahead of the Curve with the Latest Advancements and Discoveries in Information Retrieval.
This week’s newsletter highlights the following research:
When Compact Embeddings Outperform Large Models in Hybrid RAG, from SAP
Evaluating Multi-Stage vs. Simple Retrieval Strategies for Long-Context Language Models, from Stanford University
Practical Token Reduction for Large-Scale Visual Document Retrieval, from Ma et al.
Qwen3 Embedding: Advancing Text Embedding and Reranking Through Foundation Models, from Alibaba
Scaling User History Length in Production Recommendation Systems, from Meta
An Agentic Approach to Multimodal Research Report Generation, from Zhejiang University
Learning Step-by-Step Reasoning-Retrieval Strategies for Enhanced Question Answering, from Fudan University
Maintaining Dense Retriever Performance in Dynamic Corpora via Gradient-Based OOD Detection, from Ko et al.
A Two-Stage Training Framework for Efficient Small Language Model Reranking, from Mixedbread AI
Scalable Intent Classification Through Retrieval-Augmented Generation, from Amazon
Keep reading with a 7-day free trial
Subscribe to Top Information Retrieval Papers of the Week to keep reading this post and get 7 days of free access to the full post archives.