StarTree announced two new powerful AI-native innovations to its real-time data platform for enterprise workloads: Model Context Protocol (MCP) support: MCP is a standardized way for AI applications to connect with and interact with external data sources and tools. It allows Large Language Models (LLMs) to access real-time insights in StarTree in order to take actions beyond their built-in knowledge. Vector Auto Embedding: Simplifies and accelerates the vector embedding generation and ingestion for real-time RAG use cases based on Amazon Bedrock. These capabilities enable StarTree to power agent-facing applications, real-time Retrieval-Augmented Generation (RAG), and conversational querying at the speed, freshness, and scale enterprise AI systems demand. The StarTree platform now supports: 1) Agent-Facing Applications: By supporting the emerging Model Context Protocol (MCP), StarTree allows AI agents to dynamically analyze live, structured enterprise data. With StarTree’s high-concurrency architecture, enterprises can support millions of autonomous agents making micro-decisions in real time—whether optimizing delivery routes, adjusting pricing, or preventing service disruptions. 2) Conversational Querying: MCP simplifies and standardizes the integration between LLMs and databases, making natural language to SQL (NL2SQL) far easier and less brittle to deploy. Enterprises can now empower users to ask questions via voice or text and receive instant answers, with each question building on the last. This kind of seamless, conversational flow requires not just language understanding, but a data platform that can deliver real-time responses with context. 3) Real-Time RAG: StarTree’s new vector auto embedding enables pluggable vector embedding models to streamline the continuous flow of data from source to embedding creation to ingestion. This simplifies the deployment of Retrieval-Augmented Generation pipelines, making it easier to build and scale AI-driven use cases like financial market monitoring and system observability—without complex, stitched-together workflows.