AI++ // MCP's birthday presents, prompt caching, LLM JSON output, and much more


Happy birthday MCP! πŸ₯³ The world's fastest growing protocol was released on 26th November 2024 and has captivated developers and users alike. I am certain that everyone reading this newsletter has used MCP in one way or another, and will be happy to hear that there is plenty of work going on to keep improving and evolving the protocol.

In the newsletter this week we have stories on prompt caching, JSON outputs, product evals, and the evolution of LLM extensions that has brought us to the state of MCP today.

​Phil Nash​
Developer relations engineer for
Langflow​

πŸ› οΈ Building with AI, Agents & MCP

MCP's anniversary brings new features

The MCP blog celebrated one year since the original spec release with an updated spec including support for task based workflows for long-running tasks, simplified authorization flows, and extensions. One of the new extensions is MCP Apps, which builds on MCP-UI and the OpenAI Apps SDK to bring interactive interfaces to your agents.

Simplifying authorization is important for MCP servers, as this extensive article on MCP Auth shows.

Data ingestion for LLMs

This week on the Langflow blog we look at how the open-source document processor Docling can be used in Langflow to easily turn PDFs into Markdown with just one component. Docling is awesome and it comes out of the box with Langflow, making it easy to parse files, chat with them, ingest them into vector databases, or anything you choose to do with the parsed data.

How to work with LLMs

This deep look into prompt caching will help you to understand both how it works and how you can take advantage of it to decrease latency and costs.

Here's a bunch of tips about getting accurate JSON output from models. Note, sequential calling is no more accurate, but way slower, and pay attention to output order and naming.

Agents beat workflows and single loops beat subagents. These are just two of the eight learnings from one year of agents at PostHog.

Label some data, align an LLM evaluator, run the eval harness with each change. Those are the steps for using product evals.

New models!

There are so many releases, this might need a permanent section in the newsletter. In the last two weeks we've seen the release of Claude Opus 4.5, DeepSeek 3.2, and Mistral 3.

Mistral 3 is a multi-modal model that includes a 3B size that can run in the browser, check out the demo on HuggingFace here.

πŸ—žοΈ Other news

πŸ§‘β€πŸ’» Code & Libraries 

  • ​MCP Bundles (MCPB) was previously known as DXT (Desktop Extensions) and has been taken over by the MCP team.
  • ​Tokenflood is a load testing tool for LLMs, allowing you to test latency across a number of different parameters. Careful when running against hosted LLM providers where you pay per token, it could cost you!
  • ​LLM Council is a project built by Andrej Karpathy that uses multiple LLMs as a council to answer questions and critique each others answers.

πŸ”¦ Langflow Spotlight 

Once you've launched an agent, observability is really important to ensure that agent is behaving. Langflow supports sending tracing data to 6 different services, Arize, Langfuse, LangSmith, LangWatch, Opik and Traceloop, all you need to do is set the correct environment variables for your chosen service and you'll be able to observe the behavior of your agents in production.

πŸ—“οΈ Events 

December 4th, Sao Paulo, Brazil

Get down to the Langflow Meetup Sao Paulo to meet the founders of Langflow and learn about how people are using Langflow in production today.

Enjoy this newsletter? Forward it to a friend.

2755 Augustine Dr, 8th Floor, Santa Clara, CA 95054
​Unsubscribe Β· Preferences​

AI++ newsletter

Subscribe for all the latest news for developers on AI, Agents and MCP curated by the Langflow team.

Read more from AI++ newsletter

Working with LLMs is weird, but I never thought it would be as weird as OpenAI having to specifically tell their models not to talk about goblins, gremlins, raccoons, trolls, ogres, or pigeons. It raises so many questions. Thankfully after someone spotted the instructions in the Codex base instructions, OpenAI did give an explanation as to where the goblins came from. They never mentioned why raccoons and pigeons got caught up in the fantasy creature fascination though. In this edition of...

Is a token crunch coming? This week GitHub paused sign-us for GitHub Copilot Pro, Pro+ and Student plans, tightened up their usage limits, and removed Opus from their Pro plans. And today, Anthropic seemed to remove Claude Code from new Pro plans, though that has been reversed quickly. In general, while this is only seeming to affect individual plans related to coding agents, it could point to an inflection point where AI companies start considering how their pricing matches up to their...

The big news last week was that Anthropic mistakenly leaked the source code of Claude Code by leaving source maps in the package. Part of the source code referenced Claude Mythos, which has been properly announced this week as a model that’s incredibly good at finding software bugs and creating security exploits. It’s so good that it’s only being shared with 40 partners as part of Project Glasswing. Here’s a quick heads up, this newsletter is going to be changing format soon. As Langflow...