Discover ANY AI to make more online for less.

select between over 22,900 AI Tool and 17,900 AI News Posts.


venturebeat
LangChain's CEO argues that better models alone won't get your AI agent to production

As models get smarter and more capable, the "harnesses" around them must also evolve.

This "harness engineering" is an extension of context engineering, says LangChain co-founder and CEO Harrison Chase in a new VentureBeat Beyond the Pilot podcast episode. Whereas traditional AI harnesses have tended to constrain models from running in loops and calling tools, harnesses specifically built for AI agents allow them to interact more independently and effectively perform long-running tasks. Chase also weighed in on OpenAI's acquisition of OpenClaw, arguing that its viral success came down to a willingness to "let it rip" in ways that no major lab would — and questioning whether the acquisition actually gets OpenAI closer to a safe enterprise version of the product.

“The trend in harnesses is to actually give the large language model (LLM) itself more control over context engineering, letting it decide what it sees and what it doesn't see,” Chase says. “Now, this idea of a long-running, more autonomous assistant is viable.”Tracking progress and maintaining coherenceWhile the concept of allowing LLMs to run in a loop and call tools seems relatively simple, it’s difficult to pull off reliably, Chase noted. For a while, models were “below the threshold of usefulness” and simply couldn’t run in a loop, so devs used graphs and wrote chains to get around that. Chase pointed to AutoGPT — once the fastest-growing GitHub project ever — as a cautionary example: same architecture as today's top agents, but the models weren't good enough yet to run reliably in a loop, so it faded fast.

But as LLMs keep improving, teams can construct environments where models can run in loops and plan over longer horizons, and they can continually improve these harnesses. Previously, “you couldn't really make improvements to the harness because you couldn't actually run the model in a harness,” Chase said.

LangChain’s answer to this is Deep Agents, a customizable general-purpose harness.

Built on LangChain and LangGraph, it has planning capabilities, a virtual filesystem, context and token management, code execution, and skills and memory functions. Further, it can delegate tasks to subagents; these are specialized with different tools and configurations and can work in parallel. Context is also isolated, meaning subagent work doesn’t clutter the main agent’s context, and large subtask context is compressed into a single result for token efficiency.

All of these agents have access to file systems, Chase explained, and can essentially create to-do lists that they can execute on and track over time.

“When it goes on to the next step, and it goes on to step two or step three or step four out of a 200 step process, it has a way to track its progress and keep that coherence,” Chase said. “It comes down to letting the LLM write its thoughts down as it goes along, essentially.”

He emphasized that harnesses should be designed so that models can maintain coherence over longer tasks, and be “amenable” to models deciding when to compact context at points it determines is “advantageous.”

Also, giving agents access to code interpreters and BASH tools increases flexibility. And, providing agents with skills as opposed to just tools loaded up front allows them to load information when they need it. “So rather than hard code everything into one big system prompt," Chase explained, "you could have a smaller system prompt, ‘This is the core foundation, but if I need to do X, let me read the skill for X. If I need to do Y, let me read the skill for Y.'"

Essentially, context engineering is a “really fancy” way of saying: What is the LLM seeing? Because that’s different from what developers see, he noted. When human devs can analyze agent traces, they can put themselves in the AI’s “mindset” and answer questions like: What is the system prompt? How is it created? Is it static or is it populated? What tools does the agent have? When it makes a tool call, and gets a response back, how is that presented?

“When agents mess up, they mess up because they don't have the right context; when they succeed, they succeed because they have the right context,” Chase said. “I think of context engineering as bringing the right information in the right format to the LLM at the right time.”

Listen to the podcast to hear more about: How LangChain built its stack: LangGraph as the core pillar, LangChain at the center, Deep Agents on top.Why code sandboxes will be the next big thing. How a different type of UX will evolve as agents run at longer intervals (or continuously). Why traces and observability are core to building an agent that actually works. You can also listen and subscribe to Beyond the Pilot on Spotify, Apple or wherever you get your podcasts.

Rating

Innovation

Pricing

Technology

Usability

We have discovered similar tools to what you are looking for. Check out our suggestions for similar AI tools.

venturebeat
Most enterprises can't stop stage-three AI agent threats, VentureBeat

<p>A rogue AI agent at Meta <a href="https://venturebeat.com/security/meta-rogue-ai-agent-confused-deputy-iam-identity-governance-matrix">passed every identity check and still ex [...]

Match Score: 207.47

venturebeat
RSAC 2026 shipped five agent identity frameworks and left three critical ga

<p>“You can deceive, manipulate, and lie. That’s an inherent property of language. It’s a feature, not a flaw,” <a href="https://www.crowdstrike.com/en-us/press-releases/crowdstr [...]

Match Score: 129.31

venturebeat
Nvidia launches enterprise AI agent platform with Adobe, Salesforce, SAP am

<p><a href="https://www.nvidia.com/gtc/keynote/">Jensen Huang</a> walked onto the <a href="https://www.nvidia.com/gtc/">GTC stage</a> Monday wearing h [...]

Match Score: 115.97

venturebeat
Testing autonomous agents (Or: how I learned to stop worrying and embrace c

<p>Look, we&#x27;ve spent the last 18 months building production AI systems, and we&#x27;ll tell you what keeps us up at night — and it&#x27;s not whether the model can answer ques [...]

Match Score: 103.65

venturebeat
Nvidia's agentic AI stack is the first major platform to ship with sec

<p>For the first time on a major AI platform release, security shipped at launch — not bolted on 18 months later. At Nvidia GTC this week, five security vendors announced protection for Nvidia [...]

Match Score: 99.90

venturebeat
Microsoft patched a Copilot Studio prompt injection. The data exfiltrated a

<p>Microsoft assigned <a href="https://nvd.nist.gov/vuln/detail/CVE-2026-21520">CVE-2026-21520</a>, a CVSS 7.5 indirect prompt injection vulnerability, to Copilot Studio. & [...]

Match Score: 93.51

venturebeat
Meta's rogue AI agent passed every identity check — four gaps in ent

<p>A rogue AI agent at Meta took action without approval and <a href="https://www.theinformation.com/articles/inside-meta-rogue-ai-agent-triggers-security-alert">exposed sensitiv [...]

Match Score: 91.87

venturebeat
AI agent credentials live in the same box as untrusted code. Two new archit

<p>Four separate RSAC 2026 keynotes arrived at the same conclusion without coordinating. Microsoft&#x27;s Vasu Jakkal told attendees that zero trust must extend to AI. Cisco&#x27;s Jeetu [...]

Match Score: 91.65

venturebeat
Google's Opal just quietly showed enterprise teams the new blueprint f

<p>For the past year, the enterprise AI community has been locked in a debate about how much freedom to give AI agents. Too little, and you get expensive workflow automation that barely justifie [...]

Match Score: 88.68