Trending

#ModelServing

Latest posts tagged with #ModelServing on Bluesky

Latest Top
Trending

Posts tagged #ModelServing

Preview
KServe joins CNCF as an incubating project KServe, the leading standardized AI inference platform on Kubernetes, has been accepted as an incubating project by the Cloud Native Computing Foundation (CNCF).

KServe joins CNCF as an incubating project

www.redhat.com/en/blog/kserve-joins-cnc...

#RedHat #Kubernetes #OpenShift #OpenShiftAI #RedHatAI #CNCF #KServe #Inference #ModelServing

1 1 0 0
Preview
KServe joins CNCF as an incubating project KServe, the leading standardized AI inference platform on Kubernetes, has been accepted as an incubating project by the Cloud Native Computing Foundation (CNCF).

KServe joins CNCF as an incubating project

www.redhat.com/en/blog/kser...

#RedHat #Kubernetes #OpenShift #OpenShiftAI #RedHatAI #CNCF #KServe #Inference #ModelServing

1 0 0 0

💙

#KServe #CNCF #OpenSource #ModelServing #AI #MLOps #CloudNative #Kubeflow #Kubernetes #k8s @kubefloworg.bsky.social

2 0 0 0

This is a big step for the KServe community, and we’re excited about the road ahead in making cloud-native model serving more accessible and production-ready for everyone.

#KServe #CNCF #OpenSource #ModelServing #AI #MLOps #CloudNative @cncf.io @kubernetes.io @kubefloworg.bsky.social

2 0 0 0

Big thanks to everyone contributing code, reviews, and ideas — this integration is shaping up to be a game-changer for 𝗞𝘂𝗯𝗲𝗿𝗻𝗲𝘁𝗲𝘀-𝗻𝗮𝘁𝗶𝘃𝗲 𝗟𝗟𝗠 𝘀𝗲𝗿𝘃𝗶𝗻𝗴. Stay tuned for next release!

#KServe #llmd #GenerativeAI #MLOps #Kubernetes #ModelServing #AIInfrastructure

2 0 0 0
Preview
State of the Model Serving Communities - August 2025 Most recent updates from several AI/ML model inference communities that our team at Red Hat AI is contributing to.

State of the Model Serving Communities - August 2025 by @terrytangyuan
inferenceops.substack.com/p/state-of-the-model-ser...

#OpenSource #Kubernetes #AI #Inference #ModelServing #RedHat

0 1 0 0
Preview
State of the Model Serving Communities - August 2025 Most recent updates from several AI/ML model inference communities that our team at Red Hat AI is contributing to.

State of the Model Serving Communities - August 2025 by @terrytangyuan.xyz

inferenceops.substack.com/p/state-of-t...

#OpenSource #Kubernetes #AI #Inference #ModelServing #RedHat

2 1 0 0
Accelerate Machine Learning Model Serving With FastAPI and Redis Caching

Ускорьте обслуживание моделей машинного обучения с помощью FastAPI и кэширования Redis

Вы когда-нибудь ждали слишком долго, чтобы модель вернула прогнозы? Мы все были в этом месте. Машинные модели, особенно большие и сложные, могут быть болезненно медленными при…

#ai #machinelearning #modelserving

1 0 0 0
Accelerate Machine Learning Model Serving With FastAPI and Redis Caching Ever waited too long for a model to return predictions? We have all been there. Machine learning models, especially the large, complex ones, can be painfully slow to serve in real time. Users, on the other hand, expect instant feedback. That’s where latency becomes a real problem. Technically speaking, one of the biggest problems is […]

Accelerate Machine Learning Model Serving With FastAPI and Redis Caching

Ever waited too long for a model to return predictions? We have all been there. Machine learning models, especially the large, complex ones, can be painfully slow to serve in real time. Use…

#ai #machinelearning #modelserving

2 0 0 0
Preview
Presentation: Scaling Large Language Model Serving Infrastructure at Meta

Представление: Масштабирование инфраструктуры для обслуживания больших языковых моделей в Meta.

Е (Шарлотта) Ци рассматривает проблемы инфраструктуры для обслуживания больших языковых моделей (LLM): соответствие требованиям и скорость (Model Runners, KV-кэш и распределенн…

#llm #meta #modelserving

0 0 0 0
Preview
Presentation: Scaling Large Language Model Serving Infrastructure at Meta Ye (Charlotte) Qi overviews LLM serving infrastructure challenges: fitting & speed (Model Runners, KV cache, and distributed inference), production complexities (latency optimization and continuous evaluation), and effective scaling strategies (heterogeneous deployment and autoscaling). Learn key concepts for robust LLM deployment. By Ye Qi

Presentation: Scaling Large Language Model Serving Infrastructure at Meta

Ye (Charlotte) Qi overviews LLM serving infrastructure challenges: fitting & speed (Model Runners, KV cache, and distributed inference), production complexities (latency optimization and continuous …

#llm #meta #modelserving

1 0 0 0
Image showing the AI Agents Stack for November 2024, featuring categories such as Vertical Agents, Agent Hosting & Serving, Observability, Agent Frameworks, Memory, Tool Libraries, Sandboxes, Model Serving, and Storage. Each category lists various AI tools and platforms, like Decagon, LangGraph, Amazon Bedrock Agents, LangSmith, AutoGen, MemGPT, Chroma, Pinecone, and OpenAI. The image highlights key components in the AI ecosystem, reflecting advancements in artificial intelligence infrastructure and agent development.

Image showing the AI Agents Stack for November 2024, featuring categories such as Vertical Agents, Agent Hosting & Serving, Observability, Agent Frameworks, Memory, Tool Libraries, Sandboxes, Model Serving, and Storage. Each category lists various AI tools and platforms, like Decagon, LangGraph, Amazon Bedrock Agents, LangSmith, AutoGen, MemGPT, Chroma, Pinecone, and OpenAI. The image highlights key components in the AI ecosystem, reflecting advancements in artificial intelligence infrastructure and agent development.

🚀 Exploring the latest #AI stack ! From #VerticalAgents to #ModelServing and #Storage, this stack covers the essential tools & frameworks shaping the future of #ArtificialIntelligence. 📊🤖
#AIAgents #MachineLearning #TechStack #AIInfrastructure #DataScience #MLTools

4 0 0 0
Preview
Release v0.12.0-rc0 · kserve/kserve What's Changed Make storage initializer image configurable by @yuzisun in #3145 chore: Add design doc template links to feature request template by @ckadner in #3155 Increase pytest workers for ko...

🎄 Happy Holidays! KServe v0.12 release candidate is available! Try it out!

github.com/kserve/kserve/releases/t...

#KServe #kubernetes #MLOps #DevOps #CloudNative #Kubeflow #ModelServing #AI #MachineLearning @KnativeProject @LFAIDataFdn @CloudNativeFdn

1 0 0 0
Preview
Release v0.12.0-rc0 · kserve/kserve What's Changed Make storage initializer image configurable by @yuzisun in #3145 chore: Add design doc template links to feature request template by @ckadner in #3155 Increase pytest workers for ko...

🎄 Happy Holidays! KServe v0.12 release candidate is available! Try it out!

github.com/kserve/kserve/releases/t...

#KServe #kubernetes #MLOps #DevOps #CloudNative #Kubeflow #ModelServing #AI #MachineLearning @KnativeProject @CloudNativeFdn

0 0 0 0
Preview
Distributed Machine Learning Patterns Practical patterns for scaling machine learning from your laptop to a distributed cluster.</b> Distributing machine learning systems allow developers to handle extremely large datasets across multiple clusters, take advantage of automation tools, and benefit from hardware accelerations. This book reveals best practice techniques and insider tips for tackling the challenges of scaling machine learning systems. In Distributed Machine Learning Patterns</i> you will learn how to: Apply distributed systems patterns to build scalable and reliable machine learning projects</li> Build ML pipelines with data ingestion, distributed training, model serving, and more</li> Automate ML tasks with Kubernetes, TensorFlow, Kubeflow, and Argo Workflows</li> Make trade-offs between different patterns and approaches</li> Manage and monitor machine learning workloads at scale</li> </ul> Inside Distributed Machine Learning Patterns</i> you’ll learn to apply established distributed systems patterns to machine learning projects—plus explore cutting-edge new patterns created specifically for machine learning. Firmly rooted in the real world, this book demonstrates how to apply patterns using examples based in TensorFlow, Kubernetes, Kubeflow, and Argo Workflows. Hands-on projects and clear, practical DevOps techniques let you easily launch, manage, and monitor cloud-native distributed machine learning pipelines.

🔔 New chapters on model serving and workflow patterns of Distributed Machine Learning Patterns are now available!

👉 http://bit.ly/2RKv8Zo

#MachineLearning #Kubernetes #DistributedSystems #CloudComputing #DeepLearning #DataScience #DevOps #MLOps #CloudNative #ModelServing

0 0 0 0