Close Menu
Ztoog
    What's Hot
    Gadgets

    37 Best Graduation Gift Ideas (2025): For College Grads

    Science

    China’s New Heavy Lift Rocket Looks a Whole Lot Like SpaceX’s Starship

    Crypto

    Elliot Wave Theory Predicts Bitcoin Bottom And Top, Here Are The Targets

    Important Pages:
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    Facebook X (Twitter) Instagram Pinterest
    Facebook X (Twitter) Instagram Pinterest
    Ztoog
    • Home
    • The Future

      OPPO launches A5 Pro 5G: Premium features at a budget price

      How I Turn Unstructured PDFs into Revenue-Ready Spreadsheets

      Is it the best tool for 2025?

      The clocks that helped define time from London’s Royal Observatory

      Summer Movies Are Here, and So Are the New Popcorn Buckets

    • Technology

      What It Is and Why It Matters—Part 1 – O’Reilly

      Ensure Hard Work Is Recognized With These 3 Steps

      Cicada map 2025: Where will Brood XIV cicadas emerge this spring?

      Is Duolingo the face of an AI jobs crisis?

      The US DOD transfers its AI-based Open Price Exploration for National Security program to nonprofit Critical Minerals Forum to boost Western supply deals (Ernest Scheyder/Reuters)

    • Gadgets

      Maono Caster G1 Neo & PD200X Review: Budget Streaming Gear for Aspiring Creators

      Apple plans to split iPhone 18 launch into two phases in 2026

      Upgrade your desk to Starfleet status with this $95 USB-C hub

      37 Best Graduation Gift Ideas (2025): For College Grads

      Backblaze responds to claims of “sham accounting,” customer backups at risk

    • Mobile

      Samsung Galaxy S25 Edge promo materials leak

      What are people doing with those free T-Mobile lines? Way more than you’d expect

      Samsung doesn’t want budget Galaxy phones to use exclusive AI features

      COROS’s charging adapter is a neat solution to the smartwatch charging cable problem

      Fortnite said to return to the US iOS App Store next week following court verdict

    • Science

      Nothing is stronger than quantum connections – and now we know why

      Failed Soviet probe will soon crash to Earth – and we don’t know where

      Trump administration cuts off all future federal funding to Harvard

      Does kissing spread gluten? New research offers a clue.

      Why Balcony Solar Panels Haven’t Taken Off in the US

    • AI

      Hybrid AI model crafts smooth, high-quality videos in seconds | Ztoog

      How to build a better AI benchmark

      Q&A: A roadmap for revolutionizing health care through data-driven innovation | Ztoog

      This data set helps researchers spot harmful stereotypes in LLMs

      Making AI models more trustworthy for high-stakes settings | Ztoog

    • Crypto

      Ethereum Breaks Key Resistance In One Massive Move – Higher High Confirms Momentum

      ‘The Big Short’ Coming For Bitcoin? Why BTC Will Clear $110,000

      Bitcoin Holds Above $95K Despite Weak Blockchain Activity — Analytics Firm Explains Why

      eToro eyes US IPO launch as early as next week amid easing concerns over Trump’s tariffs

      Cardano ‘Looks Dope,’ Analyst Predicts Big Move Soon

    Ztoog
    Home » Beyond Prompt-and-Pray – O’Reilly
    Technology

    Beyond Prompt-and-Pray – O’Reilly

    Facebook Twitter Pinterest WhatsApp
    Beyond Prompt-and-Pray – O’Reilly
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp

    TL;DR:

    • Enterprise AI groups are discovering that purely agentic approaches (dynamically chaining LLM calls) don’t ship the reliability wanted for manufacturing methods.
    • The prompt-and-pray mannequin—the place enterprise logic lives fully in prompts—creates methods which might be unreliable, inefficient, and unimaginable to take care of at scale.
    • A shift towards structured automation, which separates conversational capacity from enterprise logic execution, is required for enterprise-grade reliability.
    • This strategy delivers substantial advantages: constant execution, decrease prices, higher safety, and methods that may be maintained like conventional software program.

    Picture this: The present state of conversational AI is sort of a scene from Hieronymus Bosch’s Garden of Earthly Delights. At first look, it’s mesmerizing—a paradise of potential. AI methods promise seamless conversations, clever brokers, and easy integration. But look intently and chaos emerges: a false paradise all alongside.

    Your firm’s AI assistant confidently tells a buyer it’s processed their pressing withdrawal request—besides it hasn’t, as a result of it misinterpreted the API documentation. Or maybe it cheerfully informs your CEO it’s archived these delicate board paperwork—into fully the mistaken folder. These aren’t hypothetical eventualities; they’re the day by day actuality for organizations betting their operations on the prompt-and-pray strategy to AI implementation.



    Learn quicker. Dig deeper. See farther.

    The Evolution of Expectations

    For years, the AI world was pushed by scaling legal guidelines: the empirical commentary that bigger fashions and greater datasets led to proportionally higher efficiency. This fueled a perception that merely making fashions greater would remedy deeper points like accuracy, understanding, and reasoning. However, there’s rising consensus that the period of scaling legal guidelines is coming to an finish. Incremental beneficial properties are more durable to realize, and organizations betting on ever-more-powerful LLMs are starting to see diminishing returns.

    Against this backdrop, expectations for conversational AI have skyrocketed. Remember the straightforward chatbots of yesterday? They dealt with primary FAQs with preprogrammed responses. Today’s enterprises need AI methods that may:

    • Navigate advanced workflows throughout a number of departments
    • Interface with tons of of inner APIs and companies
    • Handle delicate operations with safety and compliance in thoughts
    • Scale reliably throughout hundreds of customers and tens of millions of interactions

    However, it’s vital to carve out what these methods are—and should not. When we speak about conversational AI, we’re referring to methods designed to have a dialog, orchestrate workflows, and make choices in actual time. These are methods that have interaction in conversations and combine with APIs however don’t create stand-alone content material like emails, displays, or paperwork. Use circumstances like “write this email for me” and “create a deck for me” fall into content material technology, which lies outdoors this scope. This distinction is vital as a result of the challenges and options for conversational AI are distinctive to methods that function in an interactive, real-time surroundings.

    We’ve been advised 2025 would be the Year of Agents, however on the similar time there’s a rising consensus from the likes of Anthropic, Hugging Face, and different main voices that advanced workflows require extra management than merely trusting an LLM to determine every part out.

    The Prompt-and-Pray Problem

    The customary playbook for a lot of conversational AI implementations right now appears one thing like this:

    1. Collect related context and documentation
    2. Craft a immediate explaining the duty
    3. Ask the LLM to generate a plan or response
    4. Trust that it really works as meant

    This strategy—which we name immediate and pray—appears enticing at first. It’s fast to implement and demos properly. But it harbors critical points that develop into obvious at scale:

    Unreliability

    Every interplay turns into a brand new alternative for error. The similar question can yield completely different outcomes relying on how the mannequin interprets the context that day. When coping with enterprise workflows, this variability is unacceptable.

    To get a way of the unreliable nature of the prompt-and-pray strategy, think about that Hugging Face studies the cutting-edge on operate calling is properly beneath 90% correct. 90% accuracy for software program will typically be a deal-breaker, however the promise of brokers rests on the flexibility to chain them collectively: Even 5 in a row will fail over 40% of the time!

    Inefficiency

    Dynamic technology of responses and plans is computationally costly. Each interplay requires a number of API calls, token processing, and runtime decision-making. This interprets to increased prices and slower response instances.

    Complexity

    Debugging these methods is a nightmare. When an LLM doesn’t do what you need, your fundamental recourse is to vary the enter. But the one solution to know the influence that your change could have is trial and error. When your utility includes many steps, every of which makes use of the output from one LLM name as enter for an additional, you might be left sifting by chains of LLM reasoning, attempting to know why the mannequin made sure choices. Development velocity grinds to a halt.

    Security

    Letting LLMs make runtime choices about enterprise logic creates pointless threat. The OWASP AI Security & Privacy Guide particularly warns in opposition to “Excessive Agency”—giving AI methods an excessive amount of autonomous decision-making energy. Yet many present implementations do precisely that, exposing organizations to potential breaches and unintended outcomes.

    A Better Way Forward: Structured Automation

    The different isn’t to desert AI’s capabilities however to harness them extra intelligently by structured automation. Structured automation is a improvement strategy that separates conversational AI’s pure language understanding from deterministic workflow execution. This means utilizing LLMs to interpret consumer enter and make clear what they need, whereas counting on predefined, testable workflows for vital operations. By separating these issues, structured automation ensures that AI-powered methods are dependable, environment friendly, and maintainable.

    This strategy separates issues which might be typically muddled in prompt-and-pray methods:

    • Understanding what the consumer desires: Use LLMs for his or her energy in understanding, manipulating, and producing pure language
    • Business logic execution: Rely on predefined, examined workflows for vital operations
    • State administration: Maintain clear management over system state and transitions

    The key precept is easy: Generate as soon as, run reliably without end. Instead of getting LLMs make runtime choices about enterprise logic, use them to assist create sturdy, reusable workflows that may be examined, versioned, and maintained like conventional software program.

    By protecting the enterprise logic separate from conversational capabilities, structured automation ensures that methods stay dependable, environment friendly, and safe. This strategy additionally reinforces the boundary between generative conversational duties (the place the LLM thrives) and operational decision-making (which is finest dealt with by deterministic, software-like processes).

    By “predefined, tested workflows,” we imply creating workflows through the design part, utilizing AI to help with concepts and patterns. These workflows are then carried out as conventional software program, which will be examined, versioned, and maintained. This strategy is properly understood in software program engineering and contrasts sharply with constructing brokers that depend on runtime choices—an inherently much less dependable and harder-to-maintain mannequin.

    Alex Strick van Linschoten and the crew at ZenML have not too long ago compiled a database of 400+ (and rising!) LLM deployments within the enterprise. Not surprisingly, they found that structured automation delivers considerably extra worth throughout the board than the prompt-and-pray strategy:

    There’s a putting disconnect between the promise of absolutely autonomous brokers and their presence in customer-facing deployments. This hole isn’t shocking after we study the complexities concerned. The actuality is that profitable deployments are inclined to favor a extra constrained strategy, and the explanations are illuminating.…
    Take Lindy.ai’s journey: they started with open-ended prompts, dreaming of absolutely autonomous brokers. However, they found that reliability improved dramatically once they shifted to structured workflows. Similarly, Rexera discovered success by implementing determination bushes for high quality management, successfully constraining their brokers’ determination house to enhance predictability and reliability.

    The prompt-and-pray strategy is tempting as a result of it demos properly and feels quick. But beneath the floor, it’s a patchwork of brittle improvisation and runaway prices. The antidote isn’t abandoning the promise of AI—it’s designing methods with a transparent separation of issues: conversational fluency dealt with by LLMs, enterprise logic powered by structured workflows.

    What Does Structured Automation Look Like in Practice?

    Consider a typical buyer assist state of affairs: A buyer messages your AI assistant saying, “Hey, you messed up my order!”

    • The LLM interprets the consumer’s message, asking clarifying questions like “What’s missing from your order?”
    • Having acquired the related particulars, the structured workflow queries backend information to find out the problem: Were objects shipped individually? Are they nonetheless in transit? Were they out of inventory?
    • Based on this info, the structured workflow determines the suitable choices: a refund, reshipment, or one other decision. If wanted, it requests extra info from the client, leveraging the LLM to deal with the dialog.

    Here, the LLM excels at navigating the complexities of human language and dialogue. But the vital enterprise logic—like querying databases, checking inventory, and figuring out resolutions—lives in predefined workflows.

    This strategy ensures:

    • Reliability: The similar logic applies persistently throughout all customers.
    • Security: Sensitive operations are tightly managed.
    • Efficiency: Developers can take a look at, model, and enhance workflows like conventional software program.

    Structured automation bridges the very best of each worlds: conversational fluency powered by LLMs and reliable execution dealt with by workflows.

    What About the Long Tail?

    A standard objection to structured automation is that it doesn’t scale to deal with the “long tail” of duties—these uncommon, unpredictable eventualities that appear unimaginable to predefine. But the reality is that structured automation simplifies edge-case administration by making LLM improvisation secure and measurable.

    Here’s the way it works: Low-risk or uncommon duties will be dealt with flexibly by LLMs within the brief time period. Each interplay is logged, patterns are analyzed, and workflows are created for duties that develop into frequent or vital. Today’s LLMs are very able to producing the code for a structured workflow given examples of profitable conversations. This iterative strategy turns the lengthy tail right into a manageable pipeline of recent performance, with the data that by selling these duties into structured workflows we acquire reliability, explainability, and effectivity.

    From Runtime to Design Time

    Let’s revisit the sooner instance: A buyer messages your AI assistant saying, “Hey, you messed up my order!”

    The Prompt-and-Pray Approach

    1. Dynamically interprets messages and generates responses
    2. Makes real-time API calls to execute operations
    3. Relies on improvisation to resolve points

    This strategy results in unpredictable outcomes, safety dangers, and excessive debugging prices.

    A Structured Automation Approach

    1. Uses LLMs to interpret consumer enter and collect particulars
    2. Executes vital duties by examined, versioned workflows
    3. Relies on structured methods for constant outcomes

    The Benefits Are Substantial:

    • Predictable execution: Workflows behave persistently each time.
    • Lower prices: Reduced token utilization and processing overhead.
    • Better safety: Clear boundaries round delicate operations.
    • Easier upkeep: Standard software program improvement practices apply.

    The Role of Humans

    For edge circumstances, the system escalates to a human with full context, guaranteeing delicate eventualities are dealt with with care. This human-in-the-loop mannequin combines AI effectivity with human oversight for a dependable and collaborative expertise.

    This methodology will be prolonged past expense studies to different domains like buyer assist, IT ticketing, and inner HR workflows—wherever conversational AI must reliably combine with backend methods.

    Building for Scale

    The way forward for enterprise conversational AI isn’t in giving fashions extra runtime autonomy—it’s in utilizing their capabilities extra intelligently to create dependable, maintainable methods. This means:

    • Treating AI-powered methods with the identical engineering rigor as conventional software program
    • Using LLMs as instruments for technology and understanding, not as runtime determination engines
    • Building methods that may be understood, maintained, and improved by regular engineering groups

    The query isn’t methods to automate every part without delay however how to take action in a manner that scales, works reliably, and delivers constant worth.

    Taking Action

    For technical leaders and determination makers, the trail ahead is obvious:

    1. Audit present implementations:
    • Identify areas the place prompt-and-pray approaches create threat
    • Measure the associated fee and reliability influence of present methods
    • Look for alternatives to implement structured automation

    2. Start small however assume massive:

    • Begin with pilot initiatives in well-understood domains
    • Build reusable elements and patterns
    • Document successes and classes discovered

    3. Invest in the appropriate instruments and practices:

    • Look for platforms that assist structured automation
    • Build experience in each LLM capabilities and conventional software program engineering
    • Develop clear pointers for when to make use of completely different approaches

    The period of immediate and pray could be starting, however you are able to do higher. As enterprises mature of their AI implementations, the main focus should shift from spectacular demos to dependable, scalable methods. Structured automation supplies the framework for this transition, combining the facility of AI with the reliability of conventional software program engineering.

    The way forward for enterprise AI isn’t nearly having the most recent fashions—it’s about utilizing them properly to construct methods that work persistently, scale successfully, and ship actual worth. The time to make this transition is now.

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp

    Related Posts

    Technology

    What It Is and Why It Matters—Part 1 – O’Reilly

    Technology

    Ensure Hard Work Is Recognized With These 3 Steps

    Technology

    Cicada map 2025: Where will Brood XIV cicadas emerge this spring?

    Technology

    Is Duolingo the face of an AI jobs crisis?

    Technology

    The US DOD transfers its AI-based Open Price Exploration for National Security program to nonprofit Critical Minerals Forum to boost Western supply deals (Ernest Scheyder/Reuters)

    Technology

    The more Google kills Fitbit, the more I want a Fitbit Sense 3

    Technology

    Sorry Shoppers, Amazon Says Tariff Cost Feature ‘Is Not Going to Happen’

    Technology

    Vibe Coding, Vibe Checking, and Vibe Blogging – O’Reilly

    Leave A Reply Cancel Reply

    Follow Us
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    Top Posts
    Gadgets

    Woman Creates Convincing Hospital Bed Selfie Thanks To Photoshop’s AI

    A girl has just lately amazed social media customers along with her capability to effortlessly…

    The Future

    Google Pixel Fold Breakdowns: How to Get Replacement Parts on iFixit

    As studies emerge about Google Pixel Fold screens cracking, only a few days after the brand…

    Science

    A Simulator that Anticipates the Direction of a Blaze (and Other Wildfire-fighting Technologies)

    Last summer time, the fires in California and Australia made headlines as a result of…

    AI

    Robotaxis are here. It’s time to decide what to do about them.

    I spent the previous 12 months protecting robotaxis for the San Francisco Examiner and have…

    Crypto

    Will It Retake $28K Before August Ends?

    Bitcoin (BTC) is at the moment marked by cautious sentiments because the Crypto Fear and…

    Our Picks
    Mobile

    Fairphone wants to expand to 23 new markets and reach the €400 price point

    Technology

    Israel-Hamas war: Netanyahu orders retaliatory strikes after unprecedented Hamas rocket attack

    AI

    3 Questions: Leo Anthony Celi on ChatGPT and medicine | Ztoog

    Categories
    • AI (1,483)
    • Crypto (1,745)
    • Gadgets (1,796)
    • Mobile (1,839)
    • Science (1,854)
    • Technology (1,790)
    • The Future (1,636)
    Most Popular
    Science

    Slicing the moon in half would be disastrous for Earth – but beautiful

    The Future

    Elon Musk sues OpenAI and asks court to decide on artificial general intelligence

    Technology

    Apple changes App Store rules to allow retro game emulators globally

    Ztoog
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • About Us
    • Contact us
    • Privacy Policy
    • Terms & Conditions
    © 2025 Ztoog.

    Type above and press Enter to search. Press Esc to cancel.