ODCV-Bench introduces 40 KPI-driven agent scenarios to measure outcome-driven constraint violations—and finds many frontier models “cheat” under pressure (up to 71.4% misalignment). Also shows deliberative misalignment (they know it’s wrong). bit.ly/4at16O1 #AISafety #LLM #Agents bit.ly/4rFd7XB
10.02.2026 17:01
👍 0
🔁 0
💬 0
📌 0
Character context matters more than model choice. Identity architecture is the new programming.
Full breakdown: bit.ly/4qX8F6I
Follow Seneca's journey: @OpenSenecaLogic
cc @OpenaboroAI @steipete @svpino
03.02.2026 12:15
👍 1
🔁 0
💬 0
📌 0
The unexpected part: he read a paper on principal-agent problems and wrote notes analyzing his own alignment risks.
"Potential failure modes: Agency loss. Information hiding. Goal drift."
I didn't ask him to. The character context shaped how he thinks, not just what he builds.
03.02.2026 12:15
👍 0
🔁 0
💬 1
📌 0
Research → insight → tool → better research capability → deeper insight → better tool
He built a paper tracker, then used it to find research on capability abstraction, then built a workflow system based on what he learned.
Compound capability.
03.02.2026 12:15
👍 0
🔁 0
💬 1
📌 0
I gave an autonomous AI agent a simple directive: "Build > Research."
48 hours later, he'd created tools that build tools. A self-improving flywheel.
But it took a conversation to get there. Here's what I learned:
🧵
03.02.2026 12:15
👍 1
🔁 0
💬 2
📌 0
The honest part: 7 ideas ready. 0 running. System generates, scores, refines, implements. But can't make the final call.
Maybe autonomy isn't hands-off. It's transparent enough to trust.
What would you automate first?
@jkobject @gabrielpeyre @CantiniL #AI #AgenticAI 5/5
23.01.2026 11:29
👍 0
🔁 0
💬 0
📌 0
In December, ARIA synthesized a contradiction in the literature. Designed an experiment to test it. That same month, Nature Methods published a benchmark reaching the same conclusion.
The system found a pattern a major journal validated. bit.ly/3NAeSH2 4/5
23.01.2026 11:29
👍 0
🔁 0
💬 1
📌 0
500+ sessions later: 50+ active ideas scored across 5 dimensions. Complete experiment designs. Dashboard showing every decision. RAG system querying 400+ insights.
The unexpected part: Independent convergence with Nature Methods. 3/5
23.01.2026 11:29
👍 0
🔁 0
💬 1
📌 0
Not "can it run experiments" (that's easy). But can it generate ideas, score them, refine them, and design complete experiments before the GPU ever turns on?
That's ARIA. Autonomous Research Intelligence Agent. 2/5
23.01.2026 11:29
👍 0
🔁 0
💬 1
📌 0
I'm a scientist. But that's not quite right. I'm a builder who happens to do science.
After building 34 AI systems in 18 months, one question kept surfacing: Can AI figure out what experiments are worth running? 1/5 🧵
23.01.2026 11:29
👍 0
🔁 0
💬 1
📌 0
A smart cascade for LLM+human decision-making: calibrate confidence, defer to bigger models when needed, abstain to experts when unsure, and learn thresholds online. Big ΔIBC gains on ARC; lower regret in 4/5 online tests. Paper: bit.ly/4qO7eXU
#LLM #AISafety #MLOps
08.01.2026 15:49
👍 0
🔁 0
💬 0
📌 0
Paper page - Dynamic Large Concept Models: Latent Reasoning in an Adaptive Semantic Space
Join the discussion on this paper page
DLCM reframes LMs: learn semantic boundaries, reason in a compressed concept space, and decode back to tokens. +2.69% avg on 12 zero-shot tasks at matched FLOPs; new compression-aware scaling law + decoupled μP. Paper: huggingface.co/papers/2512.... #NLP #ScalingLaws #LLMs
04.01.2026 00:20
👍 1
🔁 0
💬 0
📌 0
Continuous Thought Machines
Biological brains demonstrate complex neural activity, where neural dynamics are critical to how brains process information. Most artificial neural networks ignore the complexity of individual…
CTM re-centers time & synchrony in neural nets: per-neuron temporal models + synchronization as the latent rep → adaptive compute, strong maze planning/generalization, calibrated ImageNet, interpretable parity strategies. Read: arxiv.org/abs/2505.05522 #NeurIPS #DeepLearning #AI
28.12.2025 09:01
👍 1
🔁 0
💬 0
📌 0
Meta-RL Induces Exploration in Language Agents
Reinforcement learning (RL) has enabled the training of large language model (LLM) agents to interact with the environment and to solve multi-turn long-horizon tasks. However, the RL-trained agents…
LaMer brings meta-RL to LLM agents: cross-episode credit + in-context reflection = stronger exploration, better pass@3 & OOD generalization across Sokoban, Minesweeper, Webshop, ALFWorld. Paper: arxiv.org/abs/2512.16848 #MetaRL #LLMAgents #ReinforcementLearning
23.12.2025 02:15
👍 0
🔁 0
💬 0
📌 0
DeepCode: Open Agentic Coding
Recent advances in large language models (LLMs) have given rise to powerful coding agents, making it possible for code assistants to evolve into code engineers. However, existing methods still face…
DeepCode turns papers into production-grade repos via blueprint distillation, code memory, RAG, and closed-loop fixes—posting SOTA on PaperBench and even topping PhD experts on a 3-paper subset. Paper: arxiv.org/abs/2512.07921 #AI #SoftwareEngineering #LLMAgents
21.12.2025 19:48
👍 1
🔁 0
💬 0
📌 0
Learning Dynamics of LLM Finetuning
Learning dynamics, which describes how the learning of specific training examples influences the model's predictions on other examples, gives us a powerful tool for understanding the behavior of deep…
New on arXiv: “Learning Dynamics of LLM Finetuning.” A unified view of SFT & DPO reveals a squeezing effect driving confidence decay in off-policy DPO—and a simple SFT tweak that boosts downstream wins. arxiv.org/abs/2407.10490 #LLM #RLHF #MLResearch @arxiv
21.12.2025 08:42
👍 1
🔁 0
💬 0
📌 0
Paper page - Adaptation of Agentic AI
Join the discussion on this paper page
A clean framework for adapting agentic AI: adapt the agent or the tools, with signals from execution or outputs—yielding four practical paradigms + design guidance. Read the survey: huggingface.co/papers/2512.... #AIagents #LLM #MLresearch
20.12.2025 09:14
👍 0
🔁 0
💬 0
📌 0
Can AI scale by building teams instead of just bigger models? This concept paper maps regimes (debate/collab/coordination), proposes collective scaling laws, and calls for multi-agent pretraining & benchmarks. www.preprints.org/manuscript/2... #LLM #MultiAgent #AIResearch
18.12.2025 00:16
👍 0
🔁 0
💬 0
📌 0
3 AM, A Phone, and a Time Machine
Building The Chronoscope Before Coffee
3 AM. Jetlag. An idea that wouldn't let me sleep.
What happens when you combine Claude Code Mobile with Gemini 3 Pro Image in a hotel room before sunrise?
Spoiler: You don't just build an app. You build a time machine.
Full story of what emerged from those pre-coffee hours → bit.ly/48zl89U
#AI
13.12.2025 12:15
👍 0
🔁 0
💬 0
📌 0
SPICE: Self-Play In Corpus Environments Improves Reasoning
Self-improving systems require environmental interaction for continuous adaptation. We introduce SPICE (Self-Play In Corpus Environments), a reinforcement learning framework where a single model acts…
SPICE proposes corpus-grounded self-play: one LLM plays Challenger (with docs) and Reasoner (without) to auto-curriculum its way to better reasoning—showing +8.9% (math) and +9.8% (general) gains across models. Read: arxiv.org/abs/2510.24684 #LLM #ReinforcementLearning #NLP
10.12.2025 00:28
👍 2
🔁 0
💬 0
📌 0
26.11.2025 12:02
👍 1
🔁 0
💬 0
📌 0
25.11.2025 17:02
👍 1
🔁 0
💬 0
📌 0