Qodo, maker of an AI coding platform, today announced the release of Qodo Gen CLI, an agent framework that enables developers to create, customize, and deploy their own AI coding agents. With the framework, creating agents can be done by writing configuration files that add autonomous AI agents throughout the software development life cycle, according to the company’s announcement. Qodo was built to help developers add autonomous coding capabilities to their applications without requiring expertise in AI systems, which can lead to solutions that sync up with an organization’s requirements, the company said. With Qodo Gen CLI, developers can define custom agents and what tools they can access, specify actions that trigger the agents, what instructions guide their behavior and ultimately, what their outputs should be. Along with enabling custom agent creation, Qodo Gen CLI includes pre-built agents for code review, test coverage analysis, and release notes generation. These agents integrate seamlessly with existing development tools through GitHub Actions, GitLab CI, Jenkins, and other CI/CD systems. For advanced use cases, agents can be exposed as Model Context Protocol (MCP) servers, enabling integration with other AI tools and platforms.
OpenAI’s API platform allows developers to express intent, not just configure model flows through built-in capabilities for knowledge retrieval, web search, and function calling for supporting real-world agent workflows
Olivier Godement, Head of Product for OpenAI’s API platform, provided a behind-the-scenes look at how enterprise teams are adopting and deploying AI agents at scale. According to Godement, 2025 marks a real shift in how AI is being deployed at scale. With over a million monthly active developers now using OpenAI’s API platform globally, and token usage up 700% year over year, AI is moving beyond experimentation. Godement emphasized that current demand isn’t just about chatbots anymore. “AI use cases are moving from simple Q&A to actually use cases where the application, the agent, can do stuff for you.” This shift prompted OpenAI to launch two major developer-facing tools in March: the Responses API and the Agents SDK. Some enterprise use cases are already delivering measurable gains. Godement positioned the Responses API as a foundational evolution in developer tooling. Previously, developers manually orchestrated sequences of model calls. Now, that orchestration is handled internally. “The Responses API is probably the biggest new layer of abstraction we introduced since pretty much GPT-3.” It allows developers to express intent, not just configure model flows. “You care about returning a really good response to the customer… the Response API essentially handles that loop.” It also includes built-in capabilities for knowledge retrieval, web search, and function calling—tools that enterprises need for real-world agent workflows. Some enterprise use cases are already delivering measurable gains: Stripe, which uses agents to accelerate invoice handling, reporting “35% faster invoice resolution; ” Box, which launched knowledge assistants that enable “zero-touch ticket triage.” Other high-value use cases include customer support (including voice), internal governance, and knowledge assistants for navigating dense documentation. Godement offered a glimpse into the roadmap. OpenAI is actively working on: Multimodal agents that can interact via text, voice, images, and structured data; Long-term memory for retaining knowledge across sessions; Cross-cloud orchestration to support complex, distributed IT environments. What matters now is building a focused use case, empowering cross-functional teams, and being ready to iterate. The next phase of value creation lies not in novel demos—but in durable systems, shaped by real-world needs and the operational discipline to make them reliable.
Study finds running gen AI models on the phones instead of in the cloud consumed anywhere from 75% to 95% less power, with associated sharp decreases in water consumption and overall carbon footprint
One of the easiest ways to minimize AI’s environmental impact may be to move where the processing is done, per new academic research conducted in partnership with Qualcomm. Running AI on devices instead of in the cloud slashes power consumption of queries by about 90%, the study finds. The industry has long touted the benefits of running models locally on devices instead of in the cloud — not just in energy terms, but also potentially making them cheaper and more private. Researchers at the University of California, Riverside ran a series of experiments comparing the performance of various generative AI models, both in the cloud and on phones powered with Qualcomm chips. Running any of six different models on the phones consumed anywhere from 75% to 95% less power, with associated sharp decreases in water consumption and overall carbon footprint. Qualcomm is also developing an AI simulator and calculator that illustrates, for any given query and user location, what the responses would look like on-device versus the cloud, and how much less power and water they would use. One example — running a coding skills question on the Llama-2-7B model in California — was 94% more power efficient and 96% more water efficient on-device. For all six models in the study, the inference time on the phones, measured in seconds, was higher than in the cloud. Narrowing or eliminating that gap, particularly on the most powerful and popular models, will be crucial to accelerating on-device adoption. For many AI users, the data center in your pocket might be all you need.
SAP Fioneer’s AI agent allows finance teams to generate complex reports using natural language by leveraging the suspense account analysis without the need to share data externally
SAP Fioneer has launched its AI Agent: an expert-built solution designed to intelligently enhance core operations of financial services institutions. By leveraging the suspense account analysis, finance teams can generate complex reports using natural language, significantly reducing manual effort, improving operational efficiency, and achieving considerable time savings. The Fioneer AI Agent delivers intelligence that is integrated into SAP Fioneer’s banking, insurance, and finance solutions, offering contextual, transparent, and actionable use cases without the need for custom development and heavy IT dependance. It empowers financial professionals to interact with data using natural language, eliminating reliance on IT teams and accelerating time to value. Designed for flexibility, the Fioneer AI Agent supports bring-your-own-LLM strategies as well as SAP BTP AI Core LLMs and will integrate with SAP Joule and other agents such as Microsoft Copilot. Integrated and aligned with the SAP strategy, it ensures full compliance with data privacy and auditability standards, making it a trusted solution for institutions seeking to scale AI responsibly and effectively. The first release of the Fioneer AI Agent lays the foundation for banks and insurers to automate processes, gain real-time insights, and make smarter decisions using natural language and without the need to share data externally. The Fioneer AI Agent is generally available now as an add-on for SAP Fioneer S/4HANA products in Banking, Insurance, and Finance.
Crusoe’s modular data centers enable rapid deployments with diverse power sources for edge inference by integrating all necessary infrastructure into a single, portable unit
Crusoe has launched Crusoe Spark™, a prefabricated modular AI factory designed to bring powerful, low-latency compute to the network’s edge. The modular data centers integrate all necessary infrastructure, including power, cooling, remote monitoring, fire suppression, and racks supporting the latest GPUs, into a single, portable unit. Crusoe Spark enables rapid deployments with diverse power sources for on-prem AI, edge inference, and AI capacity expansion needs, with units delivered as fast as three months. AI at the edge is transforming industries by enabling real-time decision-making and intelligence directly where data is generated, without the latency and bandwidth limitations of a remote cloud system. AI-optimized modular data centers integrate all necessary infrastructure—including power, cooling, remote monitoring, fire suppression, and racks that support the latest GPUs—into a single, portable unit. Crusoe Spark enables rapid deployments with diverse power sources for on-prem AI, edge inference, AI capacity expansion needs, with units delivered as fast as three months. AI at the edge is transforming industries by enabling real-time decision-making and intelligence directly where data is generated, without the latency and bandwidth limitations of a remote cloud system. This capability is critical for applications including autonomous vehicles needing instant reactions, real-time patient monitoring in healthcare, predictive maintenance in manufacturing, and smart city infrastructure optimizing traffic flow and public safety. This rapidly expanding market is driven by the explosive growth of IoT devices and the demand for immediate, localized AI insights.
Zerve and Arcee AI solution to enable users to automate AI model selection within their existing workflows by intelligently selecting between SLMs and LLMs based on input complexity, cost, domain relevance, and other variables
Zerve, the agent-driven operating system for Data & AI teams, announced a partnership with Arcee AI, a language model builder to bring model optimization and automation capabilities to the Zerve platform, enabling data science and AI professionals to build faster, smarter, and more efficient AI workflows at scale. Through the new partnership and integration, Zerve and Arcee AI enable users to automate AI model selection within their existing workflows using an OpenAI-compatible API, without incurring infrastructure overhead. Arcee Conductor enhances AI pipeline efficiency for users by intelligently selecting between SLMs and LLMs based on input complexity, cost, domain relevance, and other variables. This collaboration allows data science and AI engineering teams to: Optimize model usage by routing tasks to the most appropriate model, improving accuracy and runtime performance; Enhance automation by combining Conductor’s routing with the Zerve Agent’s dynamic workflow control; Maintain seamless integration through plug-and-play compatibility with existing Zerve environments; Cut costs by deploying lightweight, lower-cost models where applicable.
Anysphere’s new agent orchestration tools allow developers to send natural language prompts from a mobile or web-based browser directly to the background agents, instructing them to perform tasks like writing new features or fixing bugs
Well-funded AI startup Anysphere Inc. is expanding beyond its viral generative AI code editor and into “agentic AI” with the launch of new web and mobile browser-based orchestration tools for coding agents. With its new application, developers can send natural language prompts from a mobile or web-based browser directly to the background agents, instructing them to perform tasks like writing new features or fixing bugs. Using the web app, developers can also monitor fleets of agents that are busy working on different tasks, check their progress and register those that have been completed within the underlying codebase. Anysphere explained that developers can instruct its AI agents to complete tasks via the web app, and if they’re unable to do so, they can seamlessly switch to the IDE to take over and see what’s caused it to come unstuck. Each of its agents has its own shareable link, which developers can click on to see its progress.
Zango Global’s AI agents can read and interpret regulations with a high degree of accuracy, integrate it directly into a company’s day-to-day operations and respond to inquiries or draft consulting reviews complete with citations
Zango Global raised $4.8 million in seed funding led by Nexus Venture Partners to provide artificial intelligence agents to financial firms and banks, with an aim to transform how they deal with regulatory compliance. Zango uses AI agents, a type of artificial intelligence software that can make decisions, do research and achieve specific goals with a degree of autonomy. Agents are designed to carry out tasks with minimal or no human oversight, while adapting to changing circumstances. This allows them to continuously integrate knowledge, including regulatory information, so they can respond to inquiries or draft consulting reviews complete with citations. The company said its large language models and AI agents don’t just read and interpret regulations with a high degree of accuracy. They can integrate directly into a company’s day-to-day operations. In one example given by Zango, a bank involved with a regulator had a process that would have taken 48 hours, reduced to under four hours using the agentic AI platform. Using the platform, the company said, aiming to remain compliant and launching a new product or service can be as simple as spinning up an agent and asking: “I want to launch a lending product in X market. What do I need to do?” The agents will go to work, track down all the necessary resources and produce research, compliance requirements, records, citations, an impact assessment and a gap analysis helpful for future-proofing the product.
OPAQUE Systems integrates confidential computing with popular data and AI tools, to process fully encrypted data from ingestion to inference by enforcing cryptographically verifiable privacy, secure code execution, and auditable proof of compliance
OPAQUE Systems, the industry’s first Confidential AI platform, announced the availability of its secure AI solution on the Microsoft Azure Marketplace. By integrating confidential computing with popular data and AI tools, OPAQUE lets enterprises process sensitive data fully encrypted, from ingestion to inference, without costly code rewrites or specialized cryptographic skills. Most confidential computing solutions focus on encrypting data in use and verifying the basic infrastructure, such as applications running in Confidential Virtual Machines. However, OPAQUE goes significantly further by enforcing privacy, security, and compliance policies from data ingestion to inference. OPAQUE capabilities provide comprehensive coverage, which means customers safely deploy classic analytics/ML and advanced AI agents on their most valuable, confidential data, without compromising on sovereignty or compliance. By keeping sensitive information encrypted even during analysis and inference, organizations gain cryptographically verifiable privacy, protection against unapproved agents or code execution, and auditable proof of compliance at every step. This robust coverage frees enterprises to innovate at scale by using its differentiated, proprietary data while minimizing regulatory risk on a single platform. OPAQUE is the only platform that meets these needs in three critical phases.
Adaptive Computer’s no-code web-app platform lets non-programmers build full-featured apps that include payments (via Stripe), scheduled tasks, and AI features such as image generation, speech synthesis simply by entering a text prompt
Startup Adaptive Computer wants non-programmers to be using full-featured apps that they’ve created themselves, simply by entering a text prompt into Adaptive’s no-code web-app platform. To be certain, this isn’t about the computer itself or any hardware — despite the company’s name. The startup currently only builds web apps. For every app it builds, Adaptive Computer’s engine handles creating a database instance, user authentication, file management, and can create apps that include payments (via Stripe), scheduled tasks, and AI features such as image generation, speech synthesis, content analysis, and web search/research. Besides taking care of the back-end database and other technical details, Adaptive apps can work together. For instance, a user can build a file-hosting app and the next app can access those files. Founder Dennis Xu likens this as more like an “operating system” rather than a single Web app. He says the difference between more established products and his startup is that the others were originally geared toward making programming easier for programmers. “We’re building for the everyday person who is interested in creating things to make their own lives better.”