Close Menu
Techora News HubTechora News Hub
    Facebook X (Twitter) Instagram
    Techora News HubTechora News Hub
    • Home
    • Crypto News
      • Bitcoin
      • Ethereum
      • Altcoins
      • Blockchain
      • DeFi
    • AI News
    • Stock News
    • Learn
      • AI for Beginners
      • AI Tips
      • Make Money with AI
    • Reviews
    • Tools
      • Best AI Tools
      • Crypto Market Cap List
      • Stock Market Overview
      • Market Heatmap
    • Contact
    Techora News HubTechora News Hub
    Home»AI News»LangChain's CEO argues that better models alone won't get your AI agent to production
    AI News

    LangChain's CEO argues that better models alone won't get your AI agent to production

    March 7, 2026
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr WhatsApp Email
    LangChain's CEO argues that better models alone won't get your AI agent to production
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email
    binance



    As models get smarter and more capable, the "harnesses" around them must also evolve.

    This "harness engineering" is an extension of context engineering, says LangChain co-founder and CEO Harrison Chase in a new VentureBeat Beyond the Pilot podcast episode. Whereas traditional AI harnesses have tended to constrain models from running in loops and calling tools, harnesses specifically built for AI agents allow them to interact more independently and effectively perform long-running tasks.

    Chase also weighed in on OpenAI's acquisition of OpenClaw, arguing that its viral success came down to a willingness to "let it rip" in ways that no major lab would — and questioning whether the acquisition actually gets OpenAI closer to a safe enterprise version of the product.

    “The trend in harnesses is to actually give the large language model (LLM) itself more control over context engineering, letting it decide what it sees and what it doesn't see,” Chase says. “Now, this idea of a long-running, more autonomous assistant is viable.”

    synthesia

    Tracking progress and maintaining coherence

    While the concept of allowing LLMs to run in a loop and call tools seems relatively simple, it’s difficult to pull off reliably, Chase noted. For a while, models were “below the threshold of usefulness” and simply couldn’t run in a loop, so devs used graphs and wrote chains to get around that. Chase pointed to AutoGPT — once the fastest-growing GitHub project ever — as a cautionary example: same architecture as today's top agents, but the models weren't good enough yet to run reliably in a loop, so it faded fast.

    But as LLMs keep improving, teams can construct environments where models can run in loops and plan over longer horizons, and they can continually improve these harnesses. Previously, “you couldn't really make improvements to the harness because you couldn't actually run the model in a harness,” Chase said.

    LangChain’s answer to this is Deep Agents, a customizable general-purpose harness.

    Built on LangChain and LangGraph, it has planning capabilities, a virtual filesystem, context and token management, code execution, and skills and memory functions. Further, it can delegate tasks to subagents; these are specialized with different tools and configurations and can work in parallel. Context is also isolated, meaning subagent work doesn’t clutter the main agent’s context, and large subtask context is compressed into a single result for token efficiency.

    All of these agents have access to file systems, Chase explained, and can essentially create to-do lists that they can execute on and track over time.

    “When it goes on to the next step, and it goes on to step two or step three or step four out of a 200 step process, it has a way to track its progress and keep that coherence,” Chase said. “It comes down to letting the LLM write its thoughts down as it goes along, essentially.”

    He emphasized that harnesses should be designed so that models can maintain coherence over longer tasks, and be “amenable” to models deciding when to compact context at points it determines is “advantageous.”

    Also, giving agents access to code interpreters and BASH tools increases flexibility. And, providing agents with skills as opposed to just tools loaded up front allows them to load information when they need it. “So rather than hard code everything into one big system prompt," Chase explained, "you could have a smaller system prompt, ‘This is the core foundation, but if I need to do X, let me read the skill for X. If I need to do Y, let me read the skill for Y.'"

    Essentially, context engineering is a “really fancy” way of saying: What is the LLM seeing? Because that’s different from what developers see, he noted. When human devs can analyze agent traces, they can put themselves in the AI’s “mindset” and answer questions like: What is the system prompt? How is it created? Is it static or is it populated? What tools does the agent have? When it makes a tool call, and gets a response back, how is that presented?

    “When agents mess up, they mess up because they don't have the right context; when they succeed, they succeed because they have the right context,” Chase said. “I think of context engineering as bringing the right information in the right format to the LLM at the right time.”

    Listen to the podcast to hear more about:

    • How LangChain built its stack: LangGraph as the core pillar, LangChain at the center, Deep Agents on top.

    • Why code sandboxes will be the next big thing.

    • How a different type of UX will evolve as agents run at longer intervals (or continuously).

    • Why traces and observability are core to building an agent that actually works.

    You can also listen and subscribe to Beyond the Pilot on Spotify, Apple or wherever you get your podcasts.



    Source link

    10web
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email

    Related Posts

    Google-Agent vs Googlebot: Google Defines the Technical Boundary Between User Triggered AI Access and Search Crawling Systems Today

    March 29, 2026

    Seeing sounds | MIT News

    March 28, 2026

    Intercom's new post-trained Fin Apex 1.0 beats GPT-5.4 and Claude Sonnet 4.6 at customer service resolutions

    March 27, 2026

    Family offices turn to AI for financial data insights

    March 26, 2026

    Google Introduces TurboQuant: A New Compression Algorithm that Reduces LLM Key-Value Cache Memory by 6x and Delivers Up to 8x Speedup, All with Zero Accuracy Loss

    March 25, 2026

    How to create “humble” AI | MIT News

    March 24, 2026
    changelly
    Latest Posts

    Google-Agent vs Googlebot: Google Defines the Technical Boundary Between User Triggered AI Access and Search Crawling Systems Today

    March 29, 2026

    the AI influencers that ACTUALLY get you paid

    March 29, 2026

    Peter Schiff Warns Bitcoin Collateral Plan Could Amplify Housing Market Risks

    March 28, 2026

    Stablecoins Will Be Crypto’s “ChatGPT Moment,” Says Ripple

    March 28, 2026

    Bitcoin, Altcoins Give Back March Gains As Investors Cut Risk

    March 28, 2026
    coinbase
    LEGAL INFORMATION
    • Privacy Policy
    • Terms Of Service
    • Social Media Disclaimer
    • DMCA Compliance
    • Anti-Spam Policy
    Top Insights

    BNP Paribas Adds Bitcoin, Ether ETNs for France Retail Users

    March 29, 2026

    The next Bitcoin shock could be where Wall Street finally loses faith and starts selling

    March 29, 2026
    binance
    Facebook X (Twitter) Instagram Pinterest
    © 2026 TechoraNewsHub.com - All rights reserved.

    Type above and press Enter to search. Press Esc to cancel.

    bitcoin
    Bitcoin (BTC) $ 66,503.00
    ethereum
    Ethereum (ETH) $ 2,000.35
    tether
    Tether (USDT) $ 0.999227
    bnb
    BNB (BNB) $ 608.77
    xrp
    XRP (XRP) $ 1.32
    usd-coin
    USDC (USDC) $ 0.999718
    solana
    Solana (SOL) $ 81.87
    tron
    TRON (TRX) $ 0.322906
    figure-heloc
    Figure Heloc (FIGR_HELOC) $ 1.02
    staked-ether
    Lido Staked Ether (STETH) $ 2,265.05