Thinking Slow, Fast: Scaling Inference Compute with Distilled Reasoners
arXiv:2502.20339v1 Announce Type: cross Abstract: Recent advancements have demonstrated that the performance of large language models (LLMs) can be significantly […]
Thinking Slow, Fast: Scaling Inference Compute with Distilled Reasoners Lire l’article »