Early #science acceleration #experiments with GPT-5 #AI #largereasoningmodels #LLM
🧪 arxiv.org/pdf/2511.16072
Latest posts tagged with #largereasoningmodels on Bluesky
Early #science acceleration #experiments with GPT-5 #AI #largereasoningmodels #LLM
🧪 arxiv.org/pdf/2511.16072
Selective Entropy Regularization Boosts Large Reasoning Models
SIREN, a selective entropy regularization method, boosted the Qwen2.5‑Math‑7B model by 6.6 points on the majority‑at‑k metric for AIME24/25 benchmarks. Read more: getnews.me/selective-entropy-regula... #siren #largereasoningmodels #entropyregularization
Large Reasoning Models Reveal Emerging Self‑Awareness of Their Limits
Researchers found Large Reasoning Models can detect when a problem exceeds their capability, with confidence signals or hidden‑state probes, cutting token use by 60‑90%. getnews.me/large-reasoning-models-r... #largereasoningmodels #ai
AdvChain Improves Safety of Large Reasoning Models
AdvChain uses adversarial chain‑of‑thought tuning for reasoning models to self‑correct, cutting refusals and improving jailbreak robustness while keeping accuracy similar to untuned models. getnews.me/advchain-improves-safety... #advchain #largereasoningmodels
Schoenfeld's Theory Guides Understanding of Large Reasoning Models
A benchmark maps Large Reasoning Models' chain‑of‑thought output to Schoenfeld's seven episodes, labeling thousands of solutions with Plan and Verify phases. Read more: getnews.me/schoenfelds-theory-guide... #largereasoningmodels #ai
MemShare: Memory Efficient Inference for Large Reasoning Models through
KV Cache Reuse
Hong Xu, Kaiwen Chen et al.
Paper
Details
#MemEfficientInference #KVCacheReuse #LargeReasoningModels
Predictive Scaling Laws for Efficient GRPO Training of Large Reasoning
Models
Datta Nimmaturi, Debojyoti Dutta et al.
Paper
Details
#PredictiveScaling #GRPOTraining #LargeReasoningModels
Apple ML Research's "The Illusion of Thinking" paper explores how #LargeReasoningModels handle tough puzzles.
As difficulty rises, LRMs hit a "collapse threshold," showing reduced reasoning effort, indicating a limit to the models' scalability.
🔍 Dive deep: bit.ly/4kjqM2r
#AppleAI #LLMs #InfoQ