Blog: https://blog.reachsumit.com/
Newsletter: https://recsys.substack.com/
🔗 recsys.substack.com/p/understand...
Presents a cost-effective 32B parameter deep research agent achieving expert-level performance through atomic capability training and progressive optimization from mid-training to RL.
📝 arxiv.org/abs/2512.20491
👨🏽💻 github.com/stepfun-ai/S...
Presents a cost-effective 32B parameter deep research agent achieving expert-level performance through atomic capability training and progressive optimization from mid-training to RL.
📝 arxiv.org/abs/2512.20491
👨🏽💻 github.com/stepfun-ai/S...
Proposes a framework with multi-hop multimodal knowledge graphs and a pruning mechanism to improve audio-visual reasoning in multimodal large language models.
📝 arxiv.org/abs/2512.20136
Proposes a framework with multi-hop multimodal knowledge graphs and a pruning mechanism to improve audio-visual reasoning in multimodal large language models.
📝 arxiv.org/abs/2512.20136
Presents a retrieval-augmented framework that enhances prompt learning by decoupling knowledge from memorization through a knowledge-store of training data.
📝 arxiv.org/abs/2512.20145
👨🏽💻 github.com/zjunlp/Promp...
Presents a retrieval-augmented framework that enhances prompt learning by decoupling knowledge from memorization through a knowledge-store of training data.
📝 arxiv.org/abs/2512.20145
👨🏽💻 github.com/zjunlp/Promp...
Proposes a hashing method that integrates inherent group information to learn hash codes, improving both accuracy in sparse settings and efficiency for large-scale recommendations.
📝 arxiv.org/abs/2512.20172
Proposes a hashing method that integrates inherent group information to learn hash codes, improving both accuracy in sparse settings and efficiency for large-scale recommendations.
📝 arxiv.org/abs/2512.20172
Introduces an autonomous memory-retrieval controller that transforms retrieve-then-answer pipelines into closed-loop processes using evidence-gap tracking.
📝 arxiv.org/abs/2512.20237
👨🏽💻 github.com/Leagein/memr3
Introduces an autonomous memory-retrieval controller that transforms retrieve-then-answer pipelines into closed-loop processes using evidence-gap tracking.
📝 arxiv.org/abs/2512.20237
👨🏽💻 github.com/Leagein/memr3
Presents a parallel learning framework that captures multi-level positive and negative feedback signals from behavioral sequences.
📝 dl.acm.org/doi/10.1145/...
👨🏽💻 github.com/lhybq/PPN-ARE
Presents a parallel learning framework that captures multi-level positive and negative feedback signals from behavioral sequences.
📝 dl.acm.org/doi/10.1145/...
👨🏽💻 github.com/lhybq/PPN-ARE
Introduces a framework that stabilizes agentic search through symbolic action protocols and compact context management.
📝 arxiv.org/abs/2512.20458
👨🏽💻 github.com/ShootingWong...
Introduces a framework that stabilizes agentic search through symbolic action protocols and compact context management.
📝 arxiv.org/abs/2512.20458
👨🏽💻 github.com/ShootingWong...
Finds that MLP layers in LLM-based retrievers are highly redundant while attention layers remain critical, enabling substantial compression through coarse-to-fine MLP pruning.
📝 arxiv.org/abs/2512.20612
👨🏽💻 github.com/Yibin-Lei/Ef...
Finds that MLP layers in LLM-based retrievers are highly redundant while attention layers remain critical, enabling substantial compression through coarse-to-fine MLP pruning.
📝 arxiv.org/abs/2512.20612
👨🏽💻 github.com/Yibin-Lei/Ef...
Presents a multimodal graph-based RAG method that automatically constructs knowledge graphs from visual documents, enabling cross-modal reasoning for better content understanding.
📝 arxiv.org/abs/2512.20626
Presents a multimodal graph-based RAG method that automatically constructs knowledge graphs from visual documents, enabling cross-modal reasoning for better content understanding.
📝 arxiv.org/abs/2512.20626
Introduces a simple measure independent of total relevant documents and evaluates retrieval quality metrics against LLM-based response quality judgments across multiple datasets.
📝 arxiv.org/abs/2512.20854
Introduces a simple measure independent of total relevant documents and evaluates retrieval quality metrics against LLM-based response quality judgments across multiple datasets.
📝 arxiv.org/abs/2512.20854
NVIDIA releases a 30B parameter MoE model with competitive accuracy and 3.3x higher throughput while supporting 1M token contexts.
📝 arxiv.org/abs/2512.20848
🤗 huggingface.co/nvidia/NVIDI...
NVIDIA releases a 30B parameter MoE model with competitive accuracy and 3.3x higher throughput while supporting 1M token contexts.
📝 arxiv.org/abs/2512.20848
🤗 huggingface.co/nvidia/NVIDI...
NVIDIA introduces a family of models (Nano, Super, Ultra) using hybrid Mamba-Transformer MoE architecture with up to 1M token context and state-of-the-art reasoning performance.
📝 arxiv.org/abs/2512.20856
NVIDIA introduces a family of models (Nano, Super, Ultra) using hybrid Mamba-Transformer MoE architecture with up to 1M token context and state-of-the-art reasoning performance.
📝 arxiv.org/abs/2512.20856
Introduces a propensity scoring method using sigmoid functions on logarithmic item frequency to improve recommendation diversity while maintaining accuracy.
📝 arxiv.org/abs/2512.20896
👨🏽💻 github.com/cars1015/IPS...
Introduces a propensity scoring method using sigmoid functions on logarithmic item frequency to improve recommendation diversity while maintaining accuracy.
📝 arxiv.org/abs/2512.20896
👨🏽💻 github.com/cars1015/IPS...
Alibaba introduces a reasoning-enhanced framework that leverages LLMs to address knowledge poverty and systemic blindness in recommender systems.
📝 arxiv.org/abs/2512.21257
Alibaba introduces a reasoning-enhanced framework that leverages LLMs to address knowledge poverty and systemic blindness in recommender systems.
📝 arxiv.org/abs/2512.21257
Ant Group introduces a family of code embedding models using Pooling by Multihead Attention to break information bottlenecks in code retrieval.
📝 arxiv.org/abs/2512.21332
👨🏽💻 github.com/codefuse-ai/...
Ant Group introduces a family of code embedding models using Pooling by Multihead Attention to break information bottlenecks in code retrieval.
📝 arxiv.org/abs/2512.21332
👨🏽💻 github.com/codefuse-ai/...
Huawei proposes a bounded lag synchronous operation for distributed recommender systems that improves both latency and throughput in inference-only DLRM runs.
📝 arxiv.org/abs/2512.19342
Huawei proposes a bounded lag synchronous operation for distributed recommender systems that improves both latency and throughput in inference-only DLRM runs.
📝 arxiv.org/abs/2512.19342
Enables non-reasoning models to transfer reasoning strategies by restructuring retrieved evidence into coherent reasoning chains.
📝 arxiv.org/abs/2512.18329
Enables non-reasoning models to transfer reasoning strategies by restructuring retrieved evidence into coherent reasoning chains.
📝 arxiv.org/abs/2512.18329
Etsy unifies multimodal and multi-view learning for e-commerce search, using factorized transport to efficiently align primary and non-primary images with text views.
📝 arxiv.org/abs/2512.18117
Etsy unifies multimodal and multi-view learning for e-commerce search, using factorized transport to efficiently align primary and non-primary images with text views.
📝 arxiv.org/abs/2512.18117
Proposes greedy and hybrid tree construction methods that reduce identifier building time to 2-8% while maintaining or improving retrieval quality.
📝 arxiv.org/abs/2512.18434
👨🏽💻 github.com/joshrosie/re...
Proposes greedy and hybrid tree construction methods that reduce identifier building time to 2-8% while maintaining or improving retrieval quality.
📝 arxiv.org/abs/2512.18434
👨🏽💻 github.com/joshrosie/re...
Analyzes why MLLMs underperform in zero-shot multimodal retrieval, revealing text-dominated spaces and misaligned feature components.
📝 arxiv.org/abs/2512.19115
Analyzes why MLLMs underperform in zero-shot multimodal retrieval, revealing text-dominated spaces and misaligned feature components.
📝 arxiv.org/abs/2512.19115
Presents a dynamic RAG approach that uses pre-training corpus statistics to determine when to retrieve.
📝 arxiv.org/abs/2512.19134
👨🏽💻 github.com/ZhishanQ/QuC...
Presents a dynamic RAG approach that uses pre-training corpus statistics to determine when to retrieve.
📝 arxiv.org/abs/2512.19134
👨🏽💻 github.com/ZhishanQ/QuC...
Presents a hierarchical vector index system that achieves up to 9.64x higher throughput by using balanced partition granularity and accuracy-preserving recursive construction.
📝 arxiv.org/abs/2512.17264
Presents a hierarchical vector index system that achieves up to 9.64x higher throughput by using balanced partition granularity and accuracy-preserving recursive construction.
📝 arxiv.org/abs/2512.17264
Introduces a benchmark for evaluating AI-generated research reports with 50 tasks across 13 domains, combining expert-grounded rubrics and document-level fact-checking.
📝 arxiv.org/abs/2512.17776
Introduces a benchmark for evaluating AI-generated research reports with 50 tasks across 13 domains, combining expert-grounded rubrics and document-level fact-checking.
📝 arxiv.org/abs/2512.17776
Snap introduces a method that trains ID-based and text-based sequential recommendation models independently, then combines them through ensembling to leverage their complementary strengths.
📝 arxiv.org/abs/2512.17820
Snap introduces a method that trains ID-based and text-based sequential recommendation models independently, then combines them through ensembling to leverage their complementary strengths.
📝 arxiv.org/abs/2512.17820
Presents a benchmark of partition-aware collaborative filtering methods, revealing that FPSR models remain competitive but don't consistently outperform block-aware baselines.
📝 arxiv.org/abs/2512.17015
Presents a benchmark of partition-aware collaborative filtering methods, revealing that FPSR models remain competitive but don't consistently outperform block-aware baselines.
📝 arxiv.org/abs/2512.17015