Stop Using Giant LLMs for Everything: Why NVIDIA Research Says Small Language Models (SLMs) Are the Future of AI Agents 01-19
Google's "Free Lunch" for LLMs: How Prompt Repetition Fixes Attention Bottlenecks with Zero Latency 01-19
Is RLHF Killing Creativity? How Verbalized Sampling Mitigates Mode Collapse and Unlocks LLM Diversity 01-03
Beyond Self-Consistency: How CER Boosts LLM Reasoning by Leveraging "Process Confidence" (ACL 2025) 01-03
Stop AI Hallucinations Early: How Meta's DeepConf Uses Token Confidence to Cut Inference Costs by 80% 12-28
rStar Deep Dive: How MCTS & Mutual Reasoning Boost LLaMA2-7B Accuracy from 12% to 64% Without Fine-tuning 12-10