Trending

#LongContext

Latest posts tagged with #LongContext on Bluesky

Latest Top
Trending

Posts tagged #LongContext

#ClaudeSonnet 4.6 is the latest and most capable #Sonnet model, offering significant improvements in coding, computer use, #longcontext #reasoning, and more. It features a #1Mtoken #context window in beta and is now the default model for Free and Pro plan users on claude.ai and Claude Cowork.…

0 0 0 0

Update: The Tree now holds 1700+ turns of continuous narrative context.
From 'Hello' to 'Healed'.
The Phenotype Project is scaling. 📈
#LongContext #AI #Memory

2 1 0 0
Preview
A Simple But Effective Cache Augmented Generation Web Application In a prior post I wrote about the difference between RAG (Retrieval Augmented Generation) and CAG (Cache Augmented Generation).

A Simple But Effective Cache Augmented Generation Web Application

whyaiman.substack.com/p/a-simple-b...

#AI #EnterpriseAI #RAG #CAG #LongContext #GeminiFLash

0 0 0 0
Preview
Recursive Language Models We study allowing large language models (LLMs) to process arbitrarily long prompts through the lens of inference-time scaling. We propose Recursive Language Models (RLMs), a general inference strategy...

#AI #LongContext #MITCSAIL #mlsky
MIT CSAIL proposes Recursive Language Models that treat long prompts as an external environment. The model writes Python3 to inspect, decompose, and recursively query itself, enabling 10M+ token inputs. On BrowseComp+ (1K) it hits 91.33%
arxiv.org/abs/2512.24601

0 0 0 0

PHOTON stops rereading the entire past. It keeps the meaning, not every token. That is why it scales.
#LongContext #AIEngineering #MachineLearning
arxiv.org/abs/2512.20687

0 0 0 0
Post image

LLM benchmark snapshot 📊
Across long contexts, MiniMax-M2.1 (4-bit) leads in throughput, efficiency, and memory usage, while GLM-4.7 scales with higher cost.
Quantization still matters.
#LLM #AIResearch #MachineLearning #DeepLearning #GenerativeAI #Inference #ModelEfficiency #LongContext #Benchmarks

1 0 0 0
Post image

Gemini 3 now lets you drop full‑court sport clips, using its long‑context vision to nail spatial reasoning and dynamic view analysis. Ready to see AI coach your highlights? Dive in! #Gemini3 #LongContext #VisionAI

🔗 aidailypost.com/news/gemini-...

0 0 0 0
Post image

DeepSeek just dropped its V3.2 reasoning model, rivaling GPT‑5 and Gemini‑3.0‑Pro. The Speciale API stays open until Dec 15 2025—perfect for long‑context and synthetic‑dataset experiments. Dive in! #DeepSeekV3_2 #SpecialeAPI #LongContext

🔗 aidailypost.com/news/deepsee...

0 0 0 0
Preview
Developing a Long Context AI Knowledge App It has been a development day today, implementing an AI knowledge system that works with local files / network shares (using OpenAI’s gpt-oss-120B) but which can also burst out and leverage Gemini 1.5...

Developing a Long Context #AI Knowledge App

whyaiman.substack.com/p/developing...

#EnterpriseAI #UnstructuredData #LongContext #RAG

0 0 0 0

#Anthropic released #Opus45, the latest version of its flagship model, featuring state-of-the-art performance on various benchmarks. The model boasts improved #computeruse and #spreadsheetcapabilities. Additionally, Opus 4.5 includes #memoryimprovements for #longcontext operations and an…

1 0 0 0
Preview
LongContext AI: The Future of Large Language Models Unlock insights from extensive data! This guide reveals how to effectively utilize "long context" AI models for deeper analysis, improved decision-making, and uncovering hidden patterns you'd otherwise miss. #AI #LongContext #DataAnalysis

LongContext AI: The Future of Large Language Models

Unlock insights from extensive data! This guide reveals how to effectively utilize "long context" AI models for deeper analysis, improved decision-making, and uncovering hidden patterns you'd otherwise miss. #AI #LongContext #DataAnalysis

1 0 0 0
Long-Context Vision-Language Model Boosts Biomedical Image Retrieval

Long-Context Vision-Language Model Boosts Biomedical Image Retrieval

Researchers released BMC-LongCLIP, a vision-language model that handles captions up to 512 tokens, trained on BIOMEDICA-LongCAP with 1 million image-caption pairs. Read more: getnews.me/long-context-vision-lang... #biomedicalvlm #longcontext

0 0 0 0
Long-Context Fine-Tuning Improves Short-Task Performance of LLMs

Long-Context Fine-Tuning Improves Short-Task Performance of LLMs

Research shows fine‑tuning LLMs with long‑context data improves accuracy on short‑input benchmarks, and suggests alternating long and short context phases to preserve strengths. Read more: getnews.me/long-context-fine-tuning... #longcontext #llm

0 0 0 0
EntropyLong Boosts Long-Context Language Model Training

EntropyLong Boosts Long-Context Language Model Training

EntropyLong builds a verified long‑context dataset with sequences up to 128 K tokens, and the paper was submitted on 26 Sep 2025. Models trained on it showed notable gains on the RULER benchmark. getnews.me/entropylong-boosts-long-... #entropylong #longcontext

0 0 0 0
SPELL: Self-Play RL Improves Long-Context LLM Performance

SPELL: Self-Play RL Improves Long-Context LLM Performance

Researchers introduced SPELL, an RL framework that lets LLMs improve long‑context reasoning. It achieved a 7.6‑point pass@8 gain on Qwen3‑30B‑A3B‑Thinking, September 2025. Read more: getnews.me/spell-self-play-rl-impro... #spell #longcontext

0 0 0 0
Long-Context Fine-Tuning Improves Short-Task Performance of LLMs

Long-Context Fine-Tuning Improves Short-Task Performance of LLMs

Long‑context fine‑tuning lets LLMs outperform short‑context training on short‑context benchmarks, and researchers recommend hybrid mixes of both lengths for balanced performance. getnews.me/long-context-fine-tuning... #longcontext #finetuning

1 0 0 0
LiteLong Enables Efficient Long-Context Data Synthesis for LLMs

LiteLong Enables Efficient Long-Context Data Synthesis for LLMs

LiteLong can generate up to 128 K-token training samples using hierarchical BISAC categories and a BM25-based retrieval, achieving competitive results on HELMET and Ruler benchmarks. getnews.me/litelong-enables-efficie... #litelong #longcontext #bisc

0 0 0 0
Token-Aware Phase Attention Boosts Long-Context Transformer Performance

Token-Aware Phase Attention Boosts Long-Context Transformer Performance

TAPA adds a phase function to Rotary Positional Embedding, removing distance‑dependent bias and preserving interactions. Tests report lower perplexity on long‑context benchmarks. getnews.me/token-aware-phase-attent... #transformers #longcontext

0 0 0 0

A major highlight is Qwen3-Next's impressive native 256K token context window. This is massive for complex tasks, allowing deep dives into extensive documents without losing track of crucial information. 📚 #LongContext 3/6

0 0 1 0
Post image

🚀 Nvidia Rubin CPX: 1M-Token-KI ab 2026?

▶️ Kontext separat lesen
▶️ 1M-Token Fenster nutzen
▶️ Verfügbar Ende 2026

#ai #ki #artificialintelligence #nvidia #rubincpx #gpu #tech2025 #longcontext

⚡ SAVE IT! SHARE IT! READ IT! 🚀

kinews24.de/nvidia-rubin...

1 0 0 0
Video

Claude’s 1M-Token Leap — What It Means for You #Claude #LongContext
⚠️ 1M-Token Context Is Here — What To Do Now
Claude Sonnet 4 now supports a 1,000,000-token context via API (5x jump)
🌐 technijian.com | ☎ 949-379-8499 | ✉ sales@technijian.com
#Claude #LongContext #AmazonBedrock #AIOps #Technijian

0 0 0 0
Preview
Faster, Smarter AI: AnchorAttention Enhances Model Efficiency Researchers introduced AnchorAttention, an innovative method to resolve positional encoding issues with RoPE and BFloat16, enabling efficient long-context training while improving performance and redu...

Faster, Smarter AI: AnchorAttention Enhances Model Efficiency 🚀📚✨ www.azoai.com/news/2024120... #AI #MachineLearning #DeepLearning #LanguageModels #AIResearch #AnchorAttention #BFloat16 #LongContext #LLMInnovation #AIOptimization @arxiv-stat-ml.bsky.social

0 0 0 0
Preview
ARIA: The Open Multimodal AI Model Redefining Performance ARIA is an open-source multimodal AI model that uses a mixture-of-experts architecture to achieve state-of-the-art performance across tasks involving text, code, images, and videos, excelling in long-...

🔥🤖📊 ARIA: The Open Multimodal AI Model Redefining Performance www.azoai.com/news/2024101... #AI #multimodal #machinelearning #opensource #textprocessing #imagemodeling #MoEarchitecture #dataintegration #longcontext #AIinnovation @arxiv-stat-ml.bsky.social

1 0 0 0
Preview
Researchers Develop HELMET to Evaluate Long-Context Models Effectively Research introduces HELMET, a comprehensive benchmark designed to evaluate long-context language models (LCLMs) across diverse real-world tasks, improving upon existing synthetic benchmarks.

🔍💡📊 Researchers Develop HELMET to Evaluate Long-Context Models Effectively www.azoai.com/news/2024100... #AI #LCLMs #benchmark #research #NLP #evaluation #HELMET #longcontext #innovation #machinelearning @arxiv-stat-ml.bsky.social @princetonupress.bsky.social

0 0 0 0