LangChain Unveils LangGraph Cloud and Self-Improving Evaluators in Latest Update

Jessie A Ellis  Jun 29, 2024 10:12  UTC 02:12

2 Min Read

LangChain has announced the launch of LangGraph Cloud in closed beta, alongside several other significant updates, according to the LangChain Blog. This new addition aims to enhance agent workflows by offering scalable, fault-tolerant deployment capabilities.

LangGraph Cloud: A New Era for Agent Workflows

LangGraph Cloud is designed to provide a seamless deployment experience for LangGraph agents. Users can deploy with a single click and benefit from integrated tracing and monitoring features in LangSmith. The platform also includes a studio for debugging agent failure modes, enabling quick iteration and improvement.

This builds on the latest stable release of LangGraph v0.1, which supports human-in-the-loop collaboration and first-class streaming. Interested users can join the waitlist for LangGraph Cloud.

Self-Improving Evaluators in LangSmith

LangSmith introduces self-improving evaluators, a significant enhancement for those using the “LLM-as-a-Judge” approach to evaluate outputs from LLM applications. Users can now correct LLM evaluator feedback, which gets stored as few-shot examples to improve the LLM-as-a-Judge without manual prompt tweaking. This ensures more accurate and reliable testing.

Additional Updates in LangSmith

LangSmith also features new capabilities such as PII masking, custom models in the LangSmith Playground, and the ability to store model configurations when saving prompts. These updates aim to streamline the user experience and enhance functionality.

LangChain Enhancements

LangChain itself has seen improvements, including a universal model initializer for Python that allows users to initialize any common chat model with one line of code. Additionally, a new utility for trimming messages has been introduced, which is particularly useful for stateful or complex applications.

Community and Ecosystem Updates

LangChain continues to engage with its community through events and integrations. An upcoming meetup in Austin, TX, is scheduled for July 10, where enthusiasts can learn more about LangChain and its applications. Additionally, LangChain has been integrated with various partner features, including Anthropic Sonnet 3.5, Firefunction-v2 by Fireworks, and llama.cpp.

For developers looking to build voice-based LLM apps, LangChain has also integrated with Vocode. The company has been recognized by Redpoint, being named to their 2024 InfraRed 100 list.

Real-World Use Cases

LangChain has shared several customer stories showcasing the practical applications of its technologies. For instance, Factory improved its iteration speed by 2x using LangSmith, while Cisco's Adam Lucek discussed the importance of standardized evaluations in AI workflows.

For those new to LangChain, resources such as the LangChain primer by Lakshya Agarwal and the LangChain Masterclass for Beginners by Brandon Hancock are available. These resources provide comprehensive guides on how to leverage LangChain for building powerful AI applications.

For more detailed information, visit the LangChain Blog.



Read More