AI++ // getting real time, building skills, and what did raccoons do to get lumped in with goblins?


Working with LLMs is weird, but I never thought it would be as weird as OpenAI having to specifically tell their models not to talk about goblins, gremlins, raccoons, trolls, ogres, or pigeons. It raises so many questions. Thankfully after someone spotted the instructions in the Codex base instructions, OpenAI did give an explanation as to where the goblins came from. They never mentioned why raccoons and pigeons got caught up in the fantasy creature fascination though.

In this edition of AI++ there are many tips on writing Agent Skills, a look at the latest on building Agent harnesses, and we check out some real time speech models.

As a reminder, we are working to move this newsletter over to IBM systems. So the next time you receive it, it may look a little different. Watch out for that!

Phil Nash

Developer relations engineer for IBM

🛠️ Building with AI, Agents & MCP

Skills

Writing good Agent Skills is important if you are building things yourself, in a team, or building something that others could consume through a Skill. Phil Schmid’s 8 Tips for Writing Agent Skills is a good place to start, and the team at Perplexity went even deeper on how they Design, Refine, and Maintain Agent Skills. I also liked the approach that Pulse took in describing how to collaborate on Skills as a team.

Harnesses

Agent harnesses are the biggest topic of conversation right now, with much work being done on how to turn an LLM loop and some tools into the very capable agents that we see. Addy Osmani shared a roundup on harness engineering, and Vivek Trivedy at LangChain wrote about the anatomy of an agent harness. Meanwhile, the team behind the Astro web framework released a harness framework called Flue that makes it easy to build your own harness.

Real time models

There is a section for new model news below, but I thought that it was interesting to see OpenAI and Thinking Machines, a lab started by former OpenAI CTO Mira Murati, both release real time models within the last week. Thinking Machines described their new model as an interaction model that takes part in conversations like a human would, listening in real time, interjecting when required, and able to do the same thing with vision too. Sadly the model isn’t available to experiment with yet, but the demos look very interesting.

OpenAI are touting similar skills with the new GPT-Realtime-2 model, though I don’t think it can interject in the same way. They also announced GPT-Realtime-Translate which can live translate from 70 languages into 13 output languages, and GPT-Realtime-Whisper which live transcribes speech to text.

I also learned last week while at Twilio’s SIGNAL conference about Deepgram’s Flux model. Flux was released in October last year, but is helping to drive voice AI use cases as it was trained to understand conversations, particularly turn-taking.

It’s an exciting time to be building voice interfaces to your agents.

🧠 New models

🗞️ Other news

🧑‍💻 Code & Libraries

  • Steve Yegge announced Gas City a framework in which you can build your own Gas Town multi-agent orchestration system
  • Agent Vault is an open-source credential broker that sits between agents and the APIs they call
  • I came across a couple of agentic memory projects: Stash and agentmemory
  • Cursor launched a TypeScript SDK that you can build your own coding agents with
  • OpenUI is a toolkit for building generative UI into your agent

🔦 Langflow Spotlight

Back when Langflow released version 1.9 one of the most exciting things was the Langflow Assistant, an in app assistant that allows you to generate components from natural language. Check out the video of it in action in this post.

Enjoy this newsletter? Forward it to a friend.

2755 Augustine Dr, 8th Floor, Santa Clara, CA 95054
Unsubscribe · Preferences

AI++ newsletter

Subscribe for all the latest news for developers on AI, Agents and MCP curated by the Langflow team.

Read more from AI++ newsletter

Is a token crunch coming? This week GitHub paused sign-us for GitHub Copilot Pro, Pro+ and Student plans, tightened up their usage limits, and removed Opus from their Pro plans. And today, Anthropic seemed to remove Claude Code from new Pro plans, though that has been reversed quickly. In general, while this is only seeming to affect individual plans related to coding agents, it could point to an inflection point where AI companies start considering how their pricing matches up to their...

The big news last week was that Anthropic mistakenly leaked the source code of Claude Code by leaving source maps in the package. Part of the source code referenced Claude Mythos, which has been properly announced this week as a model that’s incredibly good at finding software bugs and creating security exploits. It’s so good that it’s only being shared with 40 partners as part of Project Glasswing. Here’s a quick heads up, this newsletter is going to be changing format soon. As Langflow...

Jensen Huang has declared on a podcast that we have reached AGI. For a very specific definition of AGI that probably doesn't agree with what you might think AGI is. One would have thought that the afterglow of NVIDIA GTC would have provided enough hype for at least the rest of the month. Meanwhile, for those building agents, there has been a lot of talk about CLIs and Agent Skills, and this week we focus on evaluating skills to make sure they do what they are supposed to. WebMCP has been an...