AI++ // MCP's birthday presents, prompt caching, LLM JSON output, and much more


Happy birthday MCP! πŸ₯³ The world's fastest growing protocol was released on 26th November 2024 and has captivated developers and users alike. I am certain that everyone reading this newsletter has used MCP in one way or another, and will be happy to hear that there is plenty of work going on to keep improving and evolving the protocol.

In the newsletter this week we have stories on prompt caching, JSON outputs, product evals, and the evolution of LLM extensions that has brought us to the state of MCP today.

​Phil Nash​
Developer relations engineer for
Langflow​

πŸ› οΈ Building with AI, Agents & MCP

MCP's anniversary brings new features

The MCP blog celebrated one year since the original spec release with an updated spec including support for task based workflows for long-running tasks, simplified authorization flows, and extensions. One of the new extensions is MCP Apps, which builds on MCP-UI and the OpenAI Apps SDK to bring interactive interfaces to your agents.

Simplifying authorization is important for MCP servers, as this extensive article on MCP Auth shows.

Data ingestion for LLMs

This week on the Langflow blog we look at how the open-source document processor Docling can be used in Langflow to easily turn PDFs into Markdown with just one component. Docling is awesome and it comes out of the box with Langflow, making it easy to parse files, chat with them, ingest them into vector databases, or anything you choose to do with the parsed data.

How to work with LLMs

This deep look into prompt caching will help you to understand both how it works and how you can take advantage of it to decrease latency and costs.

Here's a bunch of tips about getting accurate JSON output from models. Note, sequential calling is no more accurate, but way slower, and pay attention to output order and naming.

Agents beat workflows and single loops beat subagents. These are just two of the eight learnings from one year of agents at PostHog.

Label some data, align an LLM evaluator, run the eval harness with each change. Those are the steps for using product evals.

New models!

There are so many releases, this might need a permanent section in the newsletter. In the last two weeks we've seen the release of Claude Opus 4.5, DeepSeek 3.2, and Mistral 3.

Mistral 3 is a multi-modal model that includes a 3B size that can run in the browser, check out the demo on HuggingFace here.

πŸ—žοΈ Other news

πŸ§‘β€πŸ’» Code & Libraries 

  • ​MCP Bundles (MCPB) was previously known as DXT (Desktop Extensions) and has been taken over by the MCP team.
  • ​Tokenflood is a load testing tool for LLMs, allowing you to test latency across a number of different parameters. Careful when running against hosted LLM providers where you pay per token, it could cost you!
  • ​LLM Council is a project built by Andrej Karpathy that uses multiple LLMs as a council to answer questions and critique each others answers.

πŸ”¦ Langflow Spotlight 

Once you've launched an agent, observability is really important to ensure that agent is behaving. Langflow supports sending tracing data to 6 different services, Arize, Langfuse, LangSmith, LangWatch, Opik and Traceloop, all you need to do is set the correct environment variables for your chosen service and you'll be able to observe the behavior of your agents in production.

πŸ—“οΈ Events 

December 4th, Sao Paulo, Brazil

Get down to the Langflow Meetup Sao Paulo to meet the founders of Langflow and learn about how people are using Langflow in production today.

Enjoy this newsletter? Forward it to a friend.

2755 Augustine Dr, 8th Floor, Santa Clara, CA 95054
​Unsubscribe Β· Preferences​

AI++ newsletter

Subscribe for all the latest news for developers on AI, Agents and MCP curated by the Langflow team.

Read more from AI++ newsletter

There has been a flurry of new frontier models dropping over the last week that you can already use in your applications. Gemini 3 was released today, and Grok 4.1 and GPT-5.1 both arrived last week. This week we're also learning a lot of lessons from how coding agents are built, including building a coding agent in Langflow if you want hands-on experience with your own. There's also much debate over the efficiency of MCP and whether other tools fit the job better. Phil NashDeveloper...

The topic of security, specifically around prompt injection, is often raised and then dropped with a bit of a shrug as the path to a solution isn't very clear. Thankfully there are people out there thinking hard about it. In AI++ today, there are articles from Meta and Perplexity on this, with ways to mitigate the issue that we should all read and learn from. We've also got news of some great AI events coming up, including the online OpenRAG Summit, along with news of introspective AI models,...

If you like building agents that get work done, you're in for a treat in this newsletter. CUGA is a new agent framework that is topping benchmarks and using all sorts of cunning under the hood to help you build better agents that can execute complex tasks. There is also news on model releases, code execution sandboxes, and the latest podcast episode from The Flow, all on OAuth and MCP. Phil NashDeveloper relations engineer for Langflow πŸ› οΈ Building with AI, Agents & MCP IBM Research releases...