Trending

#posttraining

Latest posts tagged with #posttraining on Bluesky

Latest Top
Trending

Posts tagged #posttraining

Preview
The Social Dilemma - Wikipedia

The AI said Earth is Flat en.wikipedia.org/wiki/The_Soc... #Tools #SocialExperiments #Nudges #Grudges #ABTesting #PostTraining

0 0 1 0
Post-Training Techniques for Video Large Multimodal Models: A Survey

Post-Training Techniques for Video Large Multimodal Models: A Survey

Survey outlines three post‑training methods for Video‑LMMs: SFT with chain‑of‑thought prompting, RL from verifiable objectives, and test‑time scaling to boost inference. getnews.me/post-training-techniques... #videolmm #posttraining

0 0 0 0
Structural Shifts Discovered in Large Language Models Post-Training

Structural Shifts Discovered in Large Language Models Post-Training

SVD analysis of LLM linear layers shows post‑training causes near‑uniform scaling of singular values and coordinated orthogonal rotations. The study includes 38 pages and 26 figures. getnews.me/structural-shifts-discov... #llm #svd #posttraining

0 0 0 0
How scaling post-training is offering gains in AI intelligence Investing.com -- The AI race is entering a new phase, with a sharp pivot from the brute-force scaling of training data to the strategic amplification of models after training. Recent disclosures around xAI’s Grok 4 model show a structural shift in how intelligence gains are being unlocked, not by growing model size or feeding in more data, but by investing heavily in post-training compute. Until late 2024, most advances in AI were driven by “Chinchilla” scaling laws—training ever-larger models on ever-larger datasets. That changed with OpenAI’s o1 model and is now being accelerated by Grok 4, according to Barclays. The Grok 4 model uses roughly the same pre-training compute as its predecessor but achieves markedly higher intelligence levels thanks to a tenfold increase in reinforcement learning applied after the initial training phase. Unlike pre-training, which relies on unsupervised learning of vast text corpora, post-training through reinforcement learning allows models to improve through trial and error using curated tasks. In Grok 4’s case, this shift not only improved reasoning and problem-solving but did so without increasing the number of model parameters, keeping inference costs lower while boosting output quality. The implications are significant. “There shouldn’t be as much of a reliance on scaling up raw pre-training data tokens to achieve higher performance and intelligence,” the analyst at Barclays said. Instead to agentic models that can plan, reason, and interact with tools in complex environments. One benchmark shows Grok 4 outperforming other leading models, and humans at managing a simulated vending machine business, a test of economic reasoning and adaptability. This evolution has clear ramifications for compute demand and capex. Whereas earlier models generated responses in single steps, agentic models now reason in chains, issuing 15 times more compute-intensive tokens per query. As such, post-training may not just be the route to smarter models, it may also justify the staggering infrastructure investments being made by hyperscalers. In a field long defined by pre-training scale, post-training may be where the real intelligence lies. Don't miss out on the next big opportunity! Stay ahead of the curve with ProPicks AI – 6 model portfolios fueled by AI stock picks with a stellar performance this year... In 2024 alone, ProPicks AI identified 2 stocks that surged over 150%, 4 additional stocks that leaped over 30%, and 3 more that climbed over 25%. That's an impressive track record. With portfolios tailored for Dow stocks, S&P stocks, Tech Stocks, and Mid Cap stocks, you can explore various wealth-building strategies. So if BARC is on your watchlist, it could be very wise to know whether or not it made the ProPicks AI lists.

Click Subscribe #AI #ArtificialIntelligence #MachineLearning #PostTraining #ModelScaling

1 1 0 0
Preview
Stone Soup AI For some time, I’ve argued that a common conception of AI is misguided. This is the idea that AI systems like large language and vision models are individual intelligent agents, analogous to human age...

Rocks = Profit! simons.berkeley.edu/news/stone-s... #platforms #BetaTesting #UsersAsAProduct #PostTraining

0 0 0 0
Meine Sportschuhe, daneben die schwarzen Söckchen und meine Füße, der eine ganz, der andere eher nur die Zehen, der vordere Bereich.

Meine Sportschuhe, daneben die schwarzen Söckchen und meine Füße, der eine ganz, der andere eher nur die Zehen, der vordere Bereich.

Und immer wieder nach dem Training die kurze Überlegung, ob ich meine verschwitzten Söckchen vielleicht verkaufen sollte.

#FeetLovers
#PostTraining

32 0 3 0
Preview
TÜLU 3 Pushes the Boundaries of AI Post-Training Excellence Researchers at Allen AI introduced TÜLU 3, an open-source framework for refining language models with advanced post-training techniques like RLVR, achieving superior performance over proprietary model...

TÜLU 3 Pushes the Boundaries of AI Post-Training Excellence 🔬✨🚀 www.azoai.com/news/2024120... #AI #MachineLearning #OpenSource #LanguageModels #PostTraining #TULU3 #Innovation #TechResearch #RLVR @alleninstitute.bsky.social

1 0 0 0
Nina Simone Feeling Good Nina Simone Feeling Good Birds flying high you know how I feel Sun in the sky you know how I feel Breeze driftin' on by you know how I feel It's a new dawn It's a new day It's a new life For me And I'm feeling good Fish in the sea you know how I feel River running free you know how I feel Blossom on the tree you know how I feel Dragonfly out in the sun you know what I mean, don't you know Butterflies all havin' fun you know what I mean Sleep in peace when day is done That's what I mean And this old world is a new world And a bold world For me Stars when you shine you know how I feel Scent of the pine you know how I feel Oh freedom is mine And I know how I feel

#aidslifecycle #aidsride #sanfrancisco #posttraining #fitness #bike

There is no better sleep than that which...

fb.me/5kDSAjvfW

0 0 0 0