Coder debuts self-hosted, model-agnostic coding agents
For engineers, designers & product people. Stay up to date with free daily digest.
TLDR: Coder ships self-hosted coding agents, AWS adds AgentCore quality tuning, and SoundHound launches a self-learning agent platform.
Coder ships self-hosted, model-agnostic coding agents for enterprises
Coder Technologies has released a beta of Coder Agents, a self-hosted, AI model agnostic agent architecture that runs on the same infrastructure as existing Coder workspaces. The launch targets enterprises that want AI coding workflows with tight governance over infrastructure, data, and model choices, as of 2026-05-07.
The pitch is that teams can plug in their preferred large language models, keep source code and telemetry on their own networks, and orchestrate AI-assisted development as part of existing dev environments. If you are in a regulated org fighting SaaS code assistants, this is squarely for you. The flip side: it is still a beta, so expect integration gaps, rough edges, and evolving security guidance.
If Coder executes, this could become a default substrate for in-house coding agents instead of everyone rolling custom VS Code plus RAG stacks. Also covered by: GlobeNewswire duplicate listing
AWS AgentCore adds built-in agent quality optimization preview
Amazon Web Services is previewing agent quality optimization for AWS AgentCore so teams can generate improvement recommendations from production traces, validate with batch evaluation and A/B tests, and then ship changes more safely. The workflow explicitly targets quality drift in deployed AI agents as prompts, models, and user behavior change, as of 2026-05-07.
For anyone running support bots, internal copilots, or transactional agents on AWS, this offers a more opinionated loop: observe real traffic, propose changes, then validate across both offline metrics and live experiments. It is especially relevant if you already use Amazon Bedrock or Nova models and are tired of ad hoc dashboards and homegrown evaluation scripts. The limitation: this is a preview and mostly tied into AWS tooling, so multi-cloud or on-prem stacks will get less value.
The interesting bit is AWS leaning into production agent observability and auto-improvement as a first-class concern rather than a sidecar project.
SoundHound OASYS promises self-learning, orchestrated agent teams
SoundHound AI has launched OASYS, described as a self-learning orchestrated agentic AI platform where AI builds AI by automatically creating, coordinating, evaluating, and improving agents over time. The platform can dynamically select and orchestrate multiple AI agents in a single interaction across systems and channels, as of 2026-05-07.
The focus is enterprise voice and conversational use cases: think multilingual, multi-modal support flows spread across IVR, mobile apps, and in-car systems. For teams building complex task chains, OASYS offers lifecycle management rather than just a framework SDK. The marketing around "world’s first" and "AI builds AI" is heavy, and there are no public benchmarks yet, so treat the self-learning claims cautiously until you see concrete tuning and safety controls.
If you own a fleet of domain-specific bots that are brittle and siloed, OASYS is worth tracking as a possible orchestration and improvement layer. Also covered by: AI Business
Quick Hits
Singular Bank helps bankers move fast with ChatGPT and Codex Singular Bank reports saving bankers 60 to 90 minutes per day on meeting prep, portfolio analysis, and follow ups using an internal ChatGPT and Codex powered assistant. Useful case study if you are designing high leverage, narrow-scope agents inside regulated financial workflows.
Uber uses OpenAI to help people earn smarter and book faster Uber is rolling out OpenAI powered assistants and voice features for both drivers and riders across its marketplace, highlighting large scale real time decision support patterns and voice-first UX.
Intelligence-driven message defense and insights using Amazon Bedrock AWS shows how to use Amazon Nova models in Amazon Bedrock for message screening and sentiment analysis so you can both block risky content and mine support traffic for product insights.
How Hapag-Lloyd uses Amazon Bedrock to transform customer feedback Hapag-Lloyd details a generative AI pipeline on Bedrock plus Elasticsearch, LangChain, and LangGraph that turns free text feedback into structured insights, a solid reference architecture for enterprise RAG style analytics.
Live blog: Code w/ Claude 2026 Simon Willison live blogs Anthropic’s Code w/ Claude keynote with concrete notes on new coding features and workflows, handy if you rely on Claude for agentic development.
Show HN: agent-skills-eval A GitHub project (stars not listed yet) that provides an evaluation harness to test whether custom agent skills actually improve outputs, useful if you are adding tools to agents and want hard numbers.
Agent-harness-kit scaffolding for multi-agent workflows A provider agnostic scaffold for multi agent workflows using Model Context Protocol (MCP), targeting teams that want structure around orchestration without committing to a single LLM vendor.
vLLM V0 to V1: Correctness Before Corrections in RL ServiceNow AI walks through how vLLM is used as the inference engine inside PipelineRL and why logprob correctness is critical for stable reinforcement learning signals in production training loops.
Vibe coding and agentic engineering are getting closer than I'd like Simon Willison reflects on AI coding from a recent podcast, raising concerns that loosely specified "vibe" coding patterns are bleeding into agentic engineering in ways that make systems harder to reason about and test.
More from the Digest
For engineers, designers & product people. Stay up to date with free daily digest.