[{"data":1,"prerenderedAt":5990},["ShallowReactive",2],{"blog-articles-for-slug-agentic-ai-dialects-and-the-voice-quality-gap":3},[4,154,342,509,1103,1543,1747,1983,2045,2173,2594,2721,2818,2995,3280,3434,3495,3572,3710,3981,4222,4573,4683,4818,5148,5223,5468,5787],{"id":5,"title":6,"author":7,"body":8,"category":142,"date":143,"description":14,"extension":144,"featured":145,"image":146,"meta":147,"navigation":148,"path":149,"seo":150,"stem":151,"tags":152,"__hash__":153},"blog/blog/15-advanced-features-that-turn-your-chatbots-into-gold.md","15 advanced features that turn your chatbots into gold","Len Debets",{"type":9,"value":10,"toc":134},"minimark",[11,15,20,55,59,91,95,127,131],[12,13,14],"p",{},"Building effective chatbots requires more than just basic conversation capabilities. Here are 15 advanced features that can transform your chatbot from a simple Q&A tool into a powerful business asset.",[16,17,19],"h2",{"id":18},"advanced-ai-features","Advanced AI Features",[21,22,23,31,37,43,49],"ul",{},[24,25,26,30],"li",{},[27,28,29],"strong",{},"Multi-turn conversation management:"," Maintain context across extended conversations",[24,32,33,36],{},[27,34,35],{},"Intent recognition with confidence scoring:"," Understand user intent with high accuracy",[24,38,39,42],{},[27,40,41],{},"Entity extraction and validation:"," Extract and validate important information from user inputs",[24,44,45,48],{},[27,46,47],{},"Sentiment analysis:"," Detect user emotions and adjust responses accordingly",[24,50,51,54],{},[27,52,53],{},"Personalization engine:"," Adapt responses based on user history and preferences",[16,56,58],{"id":57},"integration-capabilities","Integration Capabilities",[21,60,61,67,73,79,85],{},[24,62,63,66],{},[27,64,65],{},"API-first architecture:"," Seamless integration with existing business systems",[24,68,69,72],{},[27,70,71],{},"Real-time data synchronization:"," Access live data from CRM, ERP, and other systems",[24,74,75,78],{},[27,76,77],{},"Multi-channel deployment:"," Deploy across web, mobile, voice, and messaging platforms",[24,80,81,84],{},[27,82,83],{},"Workflow automation:"," Trigger business processes based on conversation outcomes",[24,86,87,90],{},[27,88,89],{},"Third-party service integration:"," Connect with payment systems, booking platforms, and more",[16,92,94],{"id":93},"analytics-and-optimization","Analytics and Optimization",[21,96,97,103,109,115,121],{},[24,98,99,102],{},[27,100,101],{},"Conversation analytics:"," Detailed insights into user interactions and satisfaction",[24,104,105,108],{},[27,106,107],{},"A/B testing framework:"," Test different conversation flows and optimize performance",[24,110,111,114],{},[27,112,113],{},"Performance monitoring:"," Real-time monitoring of bot performance and user satisfaction",[24,116,117,120],{},[27,118,119],{},"Continuous learning:"," Improve responses based on user feedback and interaction patterns",[24,122,123,126],{},[27,124,125],{},"Business intelligence integration:"," Connect chatbot data with business analytics platforms",[16,128,130],{"id":129},"implementation-strategy","Implementation Strategy",[12,132,133],{},"Successfully implementing these advanced features requires a strategic approach that considers your business objectives, user needs, and technical capabilities. Start with the features that will have the most immediate impact on your specific use case.",{"title":135,"searchDepth":136,"depth":136,"links":137},"",2,[138,139,140,141],{"id":18,"depth":136,"text":19},{"id":57,"depth":136,"text":58},{"id":93,"depth":136,"text":94},{"id":129,"depth":136,"text":130},"Product Features","2020-08-20T00:00:00.000Z","md",false,"/blog/3327590-300x200.jpg",{},true,"/blog/15-advanced-features-that-turn-your-chatbots-into-gold",{"title":6,"description":14},"blog/15-advanced-features-that-turn-your-chatbots-into-gold",[],"ZPSQp5K-OuOUoJAkKz8IimWleEYtyZZzLloH8CCjNxw",{"id":155,"title":156,"author":7,"body":157,"category":333,"date":334,"description":161,"extension":144,"featured":145,"image":335,"meta":336,"navigation":148,"path":337,"seo":338,"stem":339,"tags":340,"__hash__":341},"blog/blog/9-things-i-hate-about-ai.md","9 Things I Really Hate About AI",{"type":9,"value":158,"toc":321},[159,162,166,169,172,175,178,184,188,191,194,197,201,204,207,210,214,217,220,252,255,258,262,265,268,271,275,278,281,285,288,291,295,298,301,305,308,311,315,318],[12,160,161],{},"Let's be honest: I think it's great that technology is so embedded in our daily lives. It helps us get knowledge faster, complete tasks more efficiently, gives us inspiration, and occasionally scares the hell out of us with those crazy (fake) videos. I help a lot of companies implement AI, so in the end—it pays my bills. But after spending a ridiculous amount of time with all these new technologies, I feel it's time to reflect on the things I really hate about AI.",[16,163,165],{"id":164},"_1-everyone-is-an-expert-in-ai-its-impossible-to-cut-through-the-noise","1. Everyone is an expert in AI – it's impossible to cut through the noise",[12,167,168],{},"Every time I open LinkedIn, it's a total chaos. AI is everywhere, and suddenly everyone is an expert. It ranges from people sharing their \"best prompt templates,\" to AI \"gurus\" analyzing every tiny update from OpenAI, Google, Anthropic, and others.",[12,170,171],{},"Don't get me wrong—I love that people are enthusiastic. But even as someone who works full-time in this field, it's overwhelming. The average business user doesn't need to know the difference between GPT-4o and Claude Opus, or whether a model uses a mixture-of-experts architecture. What they need is clarity, not chaos.",[12,173,174],{},"The annoying part? The reality is often 180 degrees different from what's being posted. Some claims are just nonsense. Like building a fully automated startup in 10 clicks with n8n and a ChatGPT plugin. Sounds great, but anyone who's built a real business knows that coming up with an idea, designing a logo, and setting up a support inbox does not make you money. Real businesses require differentiation, execution, and a lot of care.",[12,176,177],{},"This reminds me of those \"how to make 100K/month with 2 hours a week\" courses. If it were really that easy, why are they selling courses instead of just doing it?",[179,180,181],"blockquote",{},[12,182,183],{},"We've built a lot of AI solutions running in production for large enterprises over the last 6 years. Trust me: there is no holy grail. Making AI work for your business takes time, customization, and a lot of iteration.",[16,185,187],{"id":186},"_2-ai-is-often-just-wrong-or-doesnt-do-what-its-supposed-to","2. AI is often just wrong or doesn't do what it's supposed to",[12,189,190],{},"AI tools are incredibly powerful—but also incredibly unreliable. Hallucinations remain a major issue. I've seen models confidently generate completely incorrect answers, fabricate citations, invent legal policies, or produce entirely useless summaries.",[12,192,193],{},"You can't just \"plug in\" a large language model and expect it to work flawlessly. Even for relatively simple use cases, you'll need robust validation, fallback logic, and clearly defined boundaries. And no, prompt engineering is not a magic wand that fixes everything.",[12,195,196],{},"The problem is that even the best models can make mistakes on very simple tasks—which can be extremely frustrating. Something as basic as updating a .xlsx file might break the time format or introduce other subtle errors.",[16,198,200],{"id":199},"_3-business-leaders-now-put-ai-in-every-sentence","3. Business leaders now put \"AI\" in every sentence",[12,202,203],{},"I get it, AI is exciting. Transformative, even. But let's be honest: some executives are starting to treat it like a catch-all shortcut to success. Want to sound innovative? Say you're \"exploring AI.\" Need more budget? Pitch an \"AI-powered roadmap.\" Trying to impress investors? Just sprinkle in phrases like \"foundation models\" and \"data strategy\" and watch the heads nod.",[12,205,206],{},"Even Apple (one of the most disciplined tech companies in the world) is stumbling here. Its newly announced \"Apple Intelligence\" has already sparked critical debate for overpromising and underexplaining.",[12,208,209],{},"The problem? AI doesn't magically fix bad business models, lack of execution, or broken processes. If your organization struggles with decision-making, culture, or strategy—AI is just going to make those problems worse, faster.",[16,211,213],{"id":212},"_4-ai-is-too-big-of-a-termit-doesnt-fit-all-the-things-we-call-ai","4. AI is too big of a term—it doesn't fit all the things we call AI",[12,215,216],{},"\"AI\" used to mean something specific. Now it's a placeholder for anything remotely technical. Is it a chatbot? AI. A recommendation algorithm? AI. A basic automation script? AI. A fancy Excel formula? AI.",[12,218,219],{},"Today, AI is a catch-all for several subfields:",[21,221,222,228,234,240,246],{},[24,223,224,227],{},[27,225,226],{},"Machine Learning (ML):"," Algorithms that find patterns in data and make predictions (e.g., spam filters, fraud detection).",[24,229,230,233],{},[27,231,232],{},"Natural Language Processing (NLP):"," Understanding and generating human language (e.g., chatbots, language translation).",[24,235,236,239],{},[27,237,238],{},"Computer Vision:"," Analyzing images or video (e.g., facial recognition, autonomous vehicles).",[24,241,242,245],{},[27,243,244],{},"Robotics:"," Physical machines that perceive and act in the real world.",[24,247,248,251],{},[27,249,250],{},"Expert Systems:"," Rule-based decision systems from earlier AI eras, still used in fields like medicine and finance.",[12,253,254],{},"We've lumped everything under one umbrella, and that creates confusion. Business leaders don't know what's what. Vendors rebrand old features as \"AI-powered\" just to sound modern. And it becomes nearly impossible to have a real conversation about needs and capabilities.",[12,256,257],{},"We need new words—or at least clearer distinctions.",[16,259,261],{"id":260},"_5-it-requires-so-much-gpu-processing-powerlike-insanely-much","5. It requires so much \"GPU\" processing power—like, insanely much",[12,263,264],{},"Let's talk about the elephant in the room: GPUs. Training and running state-of-the-art AI models requires massive computing power, and we're not talking about your standard cloud VM here. We're talking thousands of high-end NVIDIA H100s running in parallel—each costing upwards of €25,000 if you can even get your hands on them.",[12,266,267],{},"OpenAI's GPT-4, Google's Gemini, and Anthropic's Claude are all built on compute infrastructures worth hundreds of millions. Even inference—just running these models at scale—requires huge GPU clusters. For most startups or researchers, that level of power is completely out of reach.",[12,269,270],{},"If you want to fine-tune a smaller open-source model like LLaMA 3 or Mistral, you're looking at significant GPU time. Renting a single A100 on AWS or Azure can cost €2 to €5 per hour—and that's if there's availability. For actual experiments or training jobs, you often need 4, 8, or even 16 GPUs just to get started.",[16,272,274],{"id":273},"_6-the-environmental-impact-is-staggering","6. The environmental impact is staggering",[12,276,277],{},"All those GPUs crunching numbers? They consume enormous amounts of electricity. Training a single large language model can emit as much CO2 as multiple transatlantic flights. And we're not just training one model—we're training thousands, running millions of inference queries daily, and constantly iterating.",[12,279,280],{},"Data centers are expanding rapidly to keep up with AI demand, and most still rely heavily on non-renewable energy sources. While companies like Google and Microsoft are investing in carbon offsets and renewable energy, the net impact is still concerning.",[16,282,284],{"id":283},"_7-copyright-ownership-and-ethical-gray-zones","7. Copyright, ownership, and ethical gray zones",[12,286,287],{},"Who owns AI-generated content? What about the data used to train these models? These questions remain largely unanswered, and it's causing real problems.",[12,289,290],{},"Artists, writers, and creators are rightfully upset that their work was scraped without permission to train models that now compete with them. Legal battles are mounting. Regulations are slow to catch up. And companies are stuck in the middle, unsure whether AI-generated content can even be copyrighted.",[16,292,294],{"id":293},"_8-job-displacement-is-realand-were-not-ready","8. Job displacement is real—and we're not ready",[12,296,297],{},"Yes, AI creates new jobs. But it also eliminates many existing ones, and not everyone can transition easily. Customer service representatives, data entry specialists, junior analysts, and even some creative professionals are already feeling the pressure.",[12,299,300],{},"The typical response—\"people should just upskill\"—is oversimplified. Upskilling takes time, resources, and access to education, all of which aren't equally available. We need better safety nets, retraining programs, and honest conversations about the economic impact.",[16,302,304],{"id":303},"_9-the-lack-of-transparency-and-accountability","9. The lack of transparency and accountability",[12,306,307],{},"When an AI makes a mistake—whether it's denying a loan application, misdiagnosing a patient, or spreading misinformation—who's responsible? The model? The company? The developer? The person who deployed it?",[12,309,310],{},"Most AI systems are black boxes. Even the teams building them often can't explain why a model made a specific decision. This lack of transparency is dangerous, especially in high-stakes scenarios like healthcare, finance, and law enforcement.",[16,312,314],{"id":313},"final-thoughts","Final Thoughts",[12,316,317],{},"Don't get me wrong—AI is transformative. It has the potential to solve real problems and improve countless lives. But pretending it's perfect, risk-free, or universally beneficial does no one any favors.",[12,319,320],{},"What we need is nuance. Honest conversations about trade-offs, capabilities, expectations, and very real limitations. If we want to move forward, we need less hype, more clarity, and above all—more honesty.s",{"title":135,"searchDepth":136,"depth":136,"links":322},[323,324,325,326,327,328,329,330,331,332],{"id":164,"depth":136,"text":165},{"id":186,"depth":136,"text":187},{"id":199,"depth":136,"text":200},{"id":212,"depth":136,"text":213},{"id":260,"depth":136,"text":261},{"id":273,"depth":136,"text":274},{"id":283,"depth":136,"text":284},{"id":293,"depth":136,"text":294},{"id":303,"depth":136,"text":304},{"id":313,"depth":136,"text":314},"AI Technology","2025-05-12T00:00:00.000Z","/blog/1747039600744.png",{},"/blog/9-things-i-hate-about-ai",{"title":156,"description":161},"blog/9-things-i-hate-about-ai",null,"y0zXWwZtLha7XwipwI4DYZSQeolLcVUqEYEYXQ1uG-Q",{"id":343,"title":344,"author":7,"body":345,"category":501,"date":502,"description":349,"extension":144,"featured":145,"image":503,"meta":504,"navigation":148,"path":505,"seo":506,"stem":507,"tags":340,"__hash__":508},"blog/blog/a-day-in-the-life-of-ai-scale-up.md","A day in the life of an AI scale-up",{"type":9,"value":346,"toc":494},[347,350,353,357,360,363,389,392,395,399,402,405,425,428,433,436,440,443,446,466,469,474,478,481,484,488,491],[12,348,349],{},"In the world of AI, where innovation moves at breakneck speed, being the CTO of a fast-growing scale-up is an exhilarating job—but not for the faint of heart. Our company specializes in large language models (LLMs), chatbots, digital humans, and voicebots, and we serve mainly large banks, international companies, and professional enterprises.",[12,351,352],{},"With Blits, we were ahead of the OpenAI curve, starting in 2020 and already implementing LLMs for customers before OpenAI became mainstream. Nearly three years of hands-on experience in this domain is rare and gives us a strategic advantage in delivering superior AI implementations today. Every day brings technical challenges, strategic decision-making, and a relentless focus on security and compliance. Let me take you through a typical day.",[16,354,356],{"id":355},"morning-the-customer-alignment-marathon","Morning: The Customer Alignment Marathon",[12,358,359],{},"The day kicks off early, often with a flood of messages and back-to-back alignment meetings. In an enterprise AI company, 80% of the work revolves around ensuring alignment—on delivery, processes, compliance, and customer expectations. Our clients operate in highly regulated industries with demanding standards, meaning every product update, deployment, or model refinement undergoes rigorous scrutiny.",[12,361,362],{},"While this level of detail can feel painstaking, it ensures we deliver at the highest standard in everything we do. A typical morning includes:",[21,364,365,371,377,383],{},[24,366,367,370],{},[27,368,369],{},"Emails:"," A relentless stream of customer updates, compliance reviews, and internal discussions.",[24,372,373,376],{},[27,374,375],{},"Customer check-ins:"," Ensuring our AI solutions meet evolving enterprise needs.",[24,378,379,382],{},[27,380,381],{},"Project management:"," Aligning with our team to track deliverables and execution.",[24,384,385,388],{},[27,386,387],{},"Compliance reviews:"," Meet with legal, security, and pen-testing teams to ensure AI implementations meet regulatory requirements.",[12,390,391],{},"AI and development is a high-stakes game, especially for large companies. Our reputation—and our customers' trust—depends on our ability to deliver secure, reliable, and compliant AI solutions.",[12,393,394],{},"That's why a deep understanding of security frameworks like ISO, SOC, HIPAA, and GDPR isn't just a nice-to-have—it's a must-have for any tech company. Having helped multiple companies get ISO-certified and get familiar with all kinds of VAPT and Pen-tests, really makes the difference if you talk to these companies as we know what we talk about.",[16,396,398],{"id":397},"midday-deep-work-and-decision-making","Midday: Deep Work and Decision-Making",[12,400,401],{},"Once meetings subside, I shift into focused execution mode. This is when I collaborate with our engineering and data science teams to evaluate model improvements and infrastructure scaling strategies.",[12,403,404],{},"Some key focus areas include:",[21,406,407,413,419],{},[24,408,409,412],{},[27,410,411],{},"Optimizing LLM performance"," – Fine-tuning models for multilingual support and real-time customer interactions.",[24,414,415,418],{},[27,416,417],{},"Scaling infrastructure"," – Managing GPU allocations, optimizing cloud costs, and implementing hybrid on-prem solutions for clients with strict data residency policies.",[24,420,421,424],{},[27,422,423],{},"Innovating responsibly"," – Ensuring AI explainability, reducing bias, and maintaining ethical AI practices.",[12,426,427],{},"Even in a rapidly evolving space, we can't just build fast—we must build responsibly. I learned this the hard way. Early on, we developed custom features for a major electronics conglomerate. Just before the CTO was about to sign the deal, they acquired another startup with a similar product—leaving our months of development effort wasted.",[179,429,430],{},[12,431,432],{},"A business owner saying they want something done is completely different from a VP saying they want something done. In large organizations, there's always someone in the chain of command—or a team of subject matter experts—who needs to approve it first.",[12,434,435],{},"That experience reinforced a critical lesson: building AI isn't just about technical capability; it's about understanding business strategy and customer dynamics.",[16,437,439],{"id":438},"afternoon-the-ai-roadmap-and-strategic-vision","Afternoon: The AI Roadmap and Strategic Vision",[12,441,442],{},"With operational priorities under control, the later part of the day is about shaping the future. In a highly competitive industry, differentiation is key.",[12,444,445],{},"This means focusing on:",[21,447,448,454,460],{},[24,449,450,453],{},[27,451,452],{},"Product roadmap planning"," – Identifying new AI capabilities that drive automation, personalization, and cost efficiency for enterprises.",[24,455,456,459],{},[27,457,458],{},"Talent and hiring"," – Scaling AI teams with not just technical expertise but also a security-first mindset.",[24,461,462,465],{},[27,463,464],{},"Investor, sales, and board discussions"," – Demonstrating the tangible business impact of our AI solutions to stakeholders.",[12,467,468],{},"At this stage, it's all about staying ahead of the curve—anticipating industry shifts and ensuring our company remains a leader in enterprise AI.",[179,470,471],{},[12,472,473],{},"The honest part is that this part of the job always gets too little attention, as other priorities are always present. However when you have the time, this is the 'cool' part of the job everyone dreams about.",[16,475,477],{"id":476},"evening-unplugging","Evening: Unplugging",[12,479,480],{},"After a full day of deep technical discussions and high-stakes decisions, winding down can be the hardest part. Closing your laptop and having enough personal time is super important to keep working at high performance. Having fun at home is super important with the people you love, there is more to life than work.",[12,482,483],{},"And for tomorrow on the agenda: kinda the same, but always different!",[16,485,487],{"id":486},"key-takeaway","Key Takeaway",[12,489,490],{},"For anyone reading this and who is leading AI initiatives for large companies, the key lesson from me to you is: It's mostly not the best technology that sets you apart—that's hygiene in today's market. What matters is knowing how to apply AI effectively in an enterprise environment to get the right people on your train.",[12,492,493],{},"Interestingly, this same principle applies to smaller companies, as they, too, are striving for enterprise-grade quality. Happy to help anyone along.",{"title":135,"searchDepth":136,"depth":136,"links":495},[496,497,498,499,500],{"id":355,"depth":136,"text":356},{"id":397,"depth":136,"text":398},{"id":438,"depth":136,"text":439},{"id":476,"depth":136,"text":477},{"id":486,"depth":136,"text":487},"Company News","2025-02-24T00:00:00.000Z","/blog/1740065406903.png",{},"/blog/a-day-in-the-life-of-ai-scale-up",{"title":344,"description":349},"blog/a-day-in-the-life-of-ai-scale-up","NY5sGFttw7LawJ5zYqeo9F5bzRiGk7tsia9TuEHje10",{"id":510,"title":511,"author":7,"body":512,"category":333,"date":1089,"description":1090,"extension":144,"featured":145,"image":1091,"meta":1092,"navigation":148,"path":1093,"seo":1094,"stem":1095,"tags":1096,"__hash__":1102},"blog/blog/agentic-ai-dialects-and-the-voice-quality-gap.md","Agentic AI Languages and Dialects: Why Voice Quality Is Still the Hard Part",{"type":9,"value":513,"toc":1076},[514,517,520,523,526,559,563,578,585,592,595,615,624,628,639,649,656,674,678,689,700,703,724,728,738,744,754,757,774,777,797,804,808,823,830,837,852,855,859,862,869,884,888,899,906,910,925,932,935,955,959,966,991,995,1010,1017,1028,1038,1042,1045,1055,1061],[12,515,516],{},"Agentic AI is moving fast.",[12,518,519],{},"Dialect-accurate speech is not.",[12,521,522],{},"When an agent is allowed to plan, call tools, and act on behalf of a user, the moment of truth is still often a voice turn: did it understand the caller, and did the caller trust what they heard back? In a banking flow, a misheard account fragment or a transfer amount read with the wrong stress pattern can erase confidence faster than any clever reasoning trace. In telco or government hotlines, the same failure shows up as repeat calls, escalations, and complaints that never mention \"the model,\" only \"the robot.\"",[12,524,525],{},"If that layer fails, the rest of the stack does not matter.",[12,527,528,529,532,533,536,537,540,541,544,545,544,548,544,551,554,555,558],{},"Nothing here is unique to one region on the map. We focus on ",[27,530,531],{},"Arabic and Turkish"," in this article because that is where much of our production depth sits today, but the same pattern shows up across ",[27,534,535],{},"non-Western and structurally different languages"," wherever providers optimize for a \"standard\" label instead of how people actually speak. ",[27,538,539],{},"Chinese"," is an obvious example: Mandarin versus regional speech, tonal accuracy, reading of mixed numerals and Latin fragments, and code-switching in business contexts each stress STT and TTS differently from European languages. Similar dynamics appear for ",[27,542,543],{},"Japanese",", ",[27,546,547],{},"Hindi and other Indic languages",[27,549,550],{},"Southeast Asian languages",[27,552,553],{},"African languages"," with limited vendor focus, and anywhere ",[27,556,557],{},"script, tone, or diglossia"," makes the \"one locale code\" story misleading. If your roadmap is global, assume the long tail until you have measured your own variety.",[16,560,562],{"id":561},"it-supports-arabic-is-not-the-same-as-it-works-in-production","\"It supports Arabic\" is not the same as \"it works in production\"",[12,564,565,566,569,570,573,574,577],{},"Across years of delivery we have worked deeply with ",[27,567,568],{},"Arabic in multiple forms",": Gulf variants such as Saudi and Qatari, ",[27,571,572],{},"Modern Standard Arabic (MSA)",", Libyan, and other regional patterns, alongside languages like ",[27,575,576],{},"Turkish"," and many more. None of these are interchangeable. A team that validates MSA for formal prompts can still fail badly when callers use colloquial Gulf phrasing, or when product names and numbers arrive in a mix of Latin digits and Arabic script.",[12,579,580,581,584],{},"The gap between marketing language lists and ",[27,582,583],{},"usable quality in a specific dialect"," is larger than most buyers expect. Procurement decks tend to show a single \"AR\" row. Production reality is closer to a matrix: which variety, which channel, which accent mix in your actual user base, and which entities (people, places, policies) appear every day.",[12,586,587,588,591],{},"Dialect is not a checkbox. It changes phonology, rhythm, vocabulary, and code-switching behavior. Models trained primarily on one variant often ",[27,589,590],{},"collapse toward a \"generic\" Arabic or English-influenced pronunciation"," under load. That collapse is invisible in a thirty-second demo with clean audio. It shows up in week two of real traffic, when users shorten sentences, overlap with the agent, or code-switch mid-thought.",[12,593,594],{},"What we typically watch for in Arabic-heavy programs includes:",[21,596,597,603,609],{},[24,598,599,602],{},[27,600,601],{},"Variant bleed",", where synthesis or recognition drifts toward MSA or another prestige norm when the user expects Gulf or North African sounds.",[24,604,605,608],{},[27,606,607],{},"Entity fragility",", where personal names, district names, or product strings that are common in one country are rare in public training data.",[24,610,611,614],{},[27,612,613],{},"Script and number mixing",", where users say amounts or IDs in one pattern and the UI or CRM stores them in another, so the model must normalize before it can speak or act correctly.",[179,616,617],{},[12,618,619,620,623],{},"\"Supports Arabic\" on a datasheet answers a sales question. Your pilot answers whether your ",[27,621,622],{},"users"," accept the voice as legitimate.",[16,625,627],{"id":626},"why-general-providers-struggle-and-specialist-voice-vendors-do-too","Why general providers struggle, and specialist voice vendors do too",[12,629,630,631,634,635,638],{},"The large cloud stacks and fast-moving model APIs usually optimize for ",[27,632,633],{},"coverage and average case"," quality. That often means strong performance in high-resource languages and major standardized forms. It is a rational commercial strategy: train where data is abundant, ship where demand is widest. The side effect is that ",[27,636,637],{},"narrow dialects and domain-heavy speech"," sit in the long tail, where error rates are higher and regressions land quietly until a customer notices.",[12,640,641,644,645,648],{},[27,642,643],{},"ElevenLabs and other voice-first providers"," can sound exceptional in the scenarios they emphasize. In our experience, ",[27,646,647],{},"regional Arabic and tight dialect targets still break in predictable ways"," when you leave the happy path: unstable prosody on mixed scripts, weak handling of entities that were never in the training distribution, or gradual drift when conversations get long and messy. The failure mode is rarely \"it does not speak Arabic.\" It is \"it speaks a version of Arabic that your listener tags as wrong, distant, or careless.\"",[12,650,651,652,655],{},"None of this is a knock on innovation. It is a reminder that ",[27,653,654],{},"dialect is a product requirement",", not a locale string. The same provider can shine in one market and frustrate in another, sometimes on the same account, because the test set for the second market was never as deep.",[12,657,658,659,662,663,665,666,669,670,673],{},"The mechanism repeats outside the Arabic and Turkish examples above. ",[27,660,661],{},"Logographic and tonal languages"," punish weak grapheme-to-sound or tone handling in TTS, and punish weak acoustic modeling in STT when users speak quickly or with regional accent. ",[27,664,539],{}," sits in many vendor roadmaps as \"supported,\" yet production teams still fight entity disambiguation (same syllable, different characters), polite versus casual register, and whether synthesis sounds like broadcast Mandarin when the user expects something closer to daily speech in a given city. You do not need Arabic in your product for this article to apply. You need a ",[27,667,668],{},"specific human audience",", a ",[27,671,672],{},"specific channel",", and honesty about whether your stack was validated for them.",[16,675,677],{"id":676},"where-we-see-the-pressure-africa-the-middle-east-and-beyond","Where we see the pressure: Africa, the Middle East, and beyond",[12,679,680,681,684,685,688],{},"We serve a large share of customers across ",[27,682,683],{},"Africa and the Middle East",". Many run ",[27,686,687],{},"production workloads"," where a specific dialect is non-negotiable: banking, telco, public sector, and regulated assistants. The business case is not experimental. It is containment, compliance, accessibility, and brand trust in channels where voice is still the primary interface for large segments of the population.",[12,690,691,692,695,696,699],{},"In several of those programs, ",[27,693,694],{},"local vendors alone could not reach the bar"," for accuracy, consistency, and operational controls once real volume arrived. Local presence helps with regulation and relationships; it does not automatically mean the best acoustic models or the right post-processing for your entities. The fix was rarely \"one more model.\" It was ",[27,697,698],{},"measurement, routing, post-processing, and continuous regression"," against prompts that look like your tickets, not like a textbook.",[12,701,702],{},"Concrete patterns we see in the field:",[21,704,705,712,718],{},[24,706,707,708,711],{},"A ",[27,709,710],{},"retail or fintech assistant"," in the Gulf must handle spontaneous phrasing, not only scripted IVR trees, once marketing promises a \"conversational\" experience.",[24,713,707,714,717],{},[27,715,716],{},"public-sector line"," must read policy numbers and dates aloud without confusing elderly callers who judge trust by sound first.",[24,719,720,723],{},[27,721,722],{},"Expansion from one Arabic market to another"," often reopens quality work: the \"same language\" is not the same acoustic or lexical reality on the ground.",[16,725,727],{"id":726},"why-stt-and-tts-fail-for-different-reasons","Why STT and TTS fail for different reasons",[12,729,730,733,734,737],{},[27,731,732],{},"Speech-to-text (STT)"," and ",[27,735,736],{},"text-to-speech (TTS)"," are often bought as a pair, but they fail for different reasons, and agentic systems stress both ends of the pipe.",[12,739,740,743],{},[27,741,742],{},"STT"," breaks when background noise, overlap, domain vocabulary, and dialectal pronunciation do not match what the acoustic and language models were trained to expect. Short utterances hide errors. Names, numbers, mixed-language phrases, and low-context fragments (\"the one from last week,\" \"same as before\") expose them. In an agentic loop, a wrong transcript becomes wrong tool arguments, wrong retrieval, and wrong follow-up questions, so the cost of a single recognition error is multiplied across turns.",[12,745,746,749,750,753],{},[27,747,748],{},"TTS"," breaks when grapheme-to-sound mapping is wrong for the target variety, when numbers and dates need ",[27,751,752],{},"language-specific reading rules",", and when the model \"smooths\" toward a prestige norm that your users experience as wrong or inauthentic. Users forgive an occasional odd word in chat. They are far less forgiving when a voice that represents your institution mispronounces a place name or reads a currency amount in a way that sounds foreign.",[12,755,756],{},"Typical STT pain points:",[21,758,759,762,768],{},[24,760,761],{},"Noisy environments and mobile microphones that differ from lab recordings.",[24,763,764,767],{},[27,765,766],{},"Rare words"," (medications, legal terms, local brands) that the language model biases toward more common homonyms.",[24,769,770,773],{},[27,771,772],{},"Short confirmations"," (\"yes,\" \"no,\" \"the first one\") where a single phoneme error changes intent.",[12,775,776],{},"Typical TTS pain points:",[21,778,779,785,791],{},[24,780,781,784],{},[27,782,783],{},"Long numbers"," (IBAN-style strings, phone numbers, national IDs) without proper chunking and reading rules.",[24,786,787,790],{},[27,788,789],{},"Foreign names and loanwords"," in the middle of a local sentence.",[24,792,793,796],{},[27,794,795],{},"Prosody drift"," over multi-sentence replies, where the opening sounds fine and the tail sounds flat or \"off.\"",[12,798,799,800,803],{},"Agentic loops make both harder: ",[27,801,802],{},"more turns, more tools, more chances for error to compound",". That is why we treat voice as part of the agent architecture, not as a skin on top.",[16,805,807],{"id":806},"why-we-benchmark-for-every-customer","Why we benchmark for every customer",[12,809,810,811,814,815,818,819,822],{},"We maintain structured evaluation runs across ",[27,812,813],{},"major vendors",", including stacks from the likes of ",[27,816,817],{},"Google, OpenAI, and ElevenLabs",", among others. The goal is not a one-time shootout. It is repeatable comparison on ",[27,820,821],{},"the same prompts, same regions, and same latency constraints",", so when a provider ships a new model or a new endpoint, we can see whether your dialect and your entities still pass.",[12,824,825],{},[826,827],"img",{"alt":828,"src":829},"Average-WER-and-latency-per-modell","/blog/Average-WER-and-latency-per-modell.jpg",[12,831,832,833,836],{},"If you are planning a serious rollout, ",[27,834,835],{},"we walk through the methodology and results in customer engagements",": what we test, how we score, how we weight subjective listening against automated signals, and how we tie all of that to your channels and compliance model.",[179,838,839],{},[12,840,841,842,845,846,848,849,851],{},"The benchmark that matters is the one built from ",[27,843,844],{},"your"," prompts, ",[27,847,844],{}," noise profile, and ",[27,850,844],{}," definition of acceptable.",[12,853,854],{},"That is a different article from \"who won last quarter in abstract.\" Both have their place. This one is about why the problem is hard; the spreadsheet belongs in a room where we can argue thresholds honestly.",[16,856,858],{"id":857},"human-ears-plus-automation","Human ears plus automation",[12,860,861],{},"Quality is not only a number. MOS-style scores and similar metrics still appear in RFPs, but they rarely capture whether a Gulf Arabic speaker will accept a voice as appropriate for a bank, or whether a Turkish user will trust a long readout of terms and conditions.",[12,863,864,865,868],{},"We use ",[27,866,867],{},"native-speaking testers"," to validate subjective fit: does this sound right to someone from that market, not only to a spectrogram? Listening panels are slower than scripts. They catch what automation misses: subtle \"almost right\" failures that tank trust.",[12,870,871,872,875,876,879,880,883],{},"Alongside that, we use ",[27,873,874],{},"benchmark suites and automated reporting"," so regressions show up when a provider ships a new model or when traffic patterns shift. Agentic systems change often; ",[27,877,878],{},"your speech layer needs a feedback loop",", not a one-time pick. The combination is deliberate: humans anchor what \"good\" means in culture; machines anchor whether you ",[27,881,882],{},"kept"," that good after the last deploy.",[16,885,887],{"id":886},"vendor-neutral-comparison-by-design","Vendor-neutral comparison by design",[12,889,890,891,894,895,898],{},"Because we integrate with ",[27,892,893],{},"all major speech and model providers",", we can ",[27,896,897],{},"route the same test content"," through different engines and compare outcomes on quality, latency, and failure modes. We are not tied to a single logo in the slide deck, which means we can recommend a stack that fits your region and your hosting constraints instead of retrofitting your requirements to a preferred vendor.",[12,900,901,902,905],{},"That matters when your procurement team wants option A but your Cairo or Istanbul pilot says otherwise. ",[27,903,904],{},"Neutrality is a feature."," It is also operational hygiene: when one provider degrades on a dialect after an update, you need a path to re-benchmark without replatforming the entire product.",[16,907,909],{"id":908},"numbers-dates-and-small-text-that-breaks-trust","Numbers, dates, and \"small\" text that breaks trust",[12,911,912,913,916,917,920,921,924],{},"Pronunciation rules for ",[27,914,915],{},"digits, currencies, ordinals, and ranges"," differ sharply across languages. In English you might read \"2026\" one way in a date and another in a product name. In Arabic and Turkish contexts, similar strings carry different expectations for pausing, grouping, and formal versus colloquial reading. ",[27,918,919],{},"Chinese, Japanese, and Korean"," introduce their own grouping and reading conventions for numbers and dates, often alongside Latin digits in enterprise data. Most providers optimize post-processing for ",[27,922,923],{},"English and sometimes a single \"standard\" form"," of a major language. Everyone else gets approximate behavior that is \"good enough\" until it is not, usually on the first high-stakes transaction.",[12,926,927,928,931],{},"We have built ",[27,929,930],{},"algorithms and normalization layers"," so models receive text that they can actually say correctly in the target language and variety. It is unglamorous engineering: rules, lexicons, disambiguation, and tests around edge cases. It is often what separates a demo from something people will use daily. A voice assistant that reads \"50,000\" with the wrong grouping or stress can sound like it is unsure of the amount, even when the underlying logic is correct.",[12,933,934],{},"Examples that routinely surface in reviews:",[21,936,937,943,949],{},[24,938,939,942],{},[27,940,941],{},"Account and reference numbers"," read as if they were ordinary integers.",[24,944,945,948],{},[27,946,947],{},"Currency amounts"," where the spoken order of units does not match local habit.",[24,950,951,954],{},[27,952,953],{},"Dates and deadlines"," where month-first versus day-first habits collide with TTS defaults trained on US or UK English.",[16,956,958],{"id":957},"cloud-private-cloud-and-on-prem-each-change-the-menu","Cloud, private cloud, and on-prem each change the menu",[12,960,961,962,965],{},"We deploy in ",[27,963,964],{},"public cloud, private cloud, and on-premises"," environments. Each topology shifts which models are available, what latency looks like, what licensing allows, and how quickly you can fall back when an upstream API changes behavior. An agentic workflow that runs beautifully in a US region with low RTT can feel different when inference and speech endpoints must stay in-country or on your own metal.",[12,967,968,971,972,975,976,979,980,982,983,986,987,990],{},[27,969,970],{},"Open-source speech models"," can be the right answer for sovereignty and cost. They are not automatically the best answer for ",[27,973,974],{},"every"," language. In some cases the strongest commercial API still wins on dialect stability for your target. In others a ",[27,977,978],{},"local or regional open model"," is the pragmatic choice once you factor in data residency and per-minute economics. For ",[27,981,576],{},", we have a clear view of ",[27,984,985],{},"which on-prem style options perform best today"," for specific use cases, without pretending one label fits every workload. The same exercise applies when customers need ",[27,988,989],{},"Chinese or other Asian languages"," on private or air-gapped infrastructure: the best public-cloud demo does not always survive your deployment boundary. The point is not open versus closed in the abstract. It is which combination survives your constraints and your listeners.",[16,992,994],{"id":993},"experience-matters-in-nuances","Experience matters in nuances",[12,996,997,998,1001,1002,1005,1006,1009],{},"We regularly see ",[27,999,1000],{},"version and product-line effects"," that contradict the assumption that \"newer is always better\" for dialects. As one illustration we are comfortable sharing at a high level: in our Arabic dialect evaluations, ",[27,1003,1004],{},"a newer Realtime family line has not consistently beaten an earlier STT-oriented stack",". The details belong in a customer readout, not in a headline. The lesson for buyers is simpler: ",[27,1007,1008],{},"ship dates are not quality guarantees",", especially for dialect.",[12,1011,1012,1013,1016],{},"We also see ",[27,1014,1015],{},"large gaps between providers that market heavily to a region and what native listeners accept as natural",". We treat that as a routing and testing input, not as theater. The goal is a better outcome for the end user, not a public ranking.",[12,1018,1019,1020,1023,1024,1027],{},"Finally, ",[27,1021,1022],{},"dialect drift"," is common: output slowly sounds less like the target variety over a session, or the model reverts toward a safer \"standard\" sound as sentences pile up. That is hard to catch with a single clip. We run ",[27,1025,1026],{},"additional checks"," in long conversations and in regression suites so drift shows up before your customers feel it as inconsistency.",[1029,1030,1035],"pre",{"className":1031,"code":1033,"language":1034,"meta":135},[1032],"language-text","Dialect rule for production agentic voice:\nIf you have not measured your exact variety, with your entities,\non your channel, you do not yet know if it works.\n","text",[1036,1037,1033],"code",{"__ignoreMap":135},[16,1039,1041],{"id":1040},"final-thought","Final thought",[12,1043,1044],{},"Agentic AI raises the ceiling on what software can do autonomously.",[12,1046,1047,1050,1051,1054],{},[27,1048,1049],{},"Dialect-accurate speech is still a bottleneck"," for a huge share of the world's population, because providers optimize for major languages and \"standard\" forms, while users speak in ",[27,1052,1053],{},"specific, lived varieties",".",[12,1056,1057,1058,1054],{},"The path forward is not optimism. It is ",[27,1059,1060],{},"integration breadth, disciplined benchmarking, native validation, and deployment-aware model choice",[12,1062,1063,1064,1067,1068,1071,1072,1075],{},"If your roadmap includes voice in ",[27,1065,1066],{},"regional Arabic, Turkish, Chinese, or any similarly sensitive market",", treat ",[27,1069,1070],{},"language variety"," as ",[27,1073,1074],{},"architecture",", not localization, and measure it like you measure uptime.",{"title":135,"searchDepth":136,"depth":136,"links":1077},[1078,1079,1080,1081,1082,1083,1084,1085,1086,1087,1088],{"id":561,"depth":136,"text":562},{"id":626,"depth":136,"text":627},{"id":676,"depth":136,"text":677},{"id":726,"depth":136,"text":727},{"id":806,"depth":136,"text":807},{"id":857,"depth":136,"text":858},{"id":886,"depth":136,"text":887},{"id":908,"depth":136,"text":909},{"id":957,"depth":136,"text":958},{"id":993,"depth":136,"text":994},{"id":1040,"depth":136,"text":1041},"2026-04-10T00:00:00.000Z","Agentic systems promise autonomous workflows, but speech and dialect quality often fail first. Arabic, Turkish, Chinese, and many other non-Western languages face the same gap between datasheet claims and production trust. This article explains why, and what teams should verify.","/blog/dialect.jpg",{},"/blog/agentic-ai-dialects-and-the-voice-quality-gap",{"title":511,"description":1090},"blog/agentic-ai-dialects-and-the-voice-quality-gap",[1097,1098,1099,1100,1101],"agentic ai","speech ai","dialects","multilingual ai","enterprise ai","0nz21x3gNwTNj6aalPQ4AWjz2wcNrs8agZuIi9Cirow",{"id":1104,"title":1105,"author":7,"body":1106,"category":333,"date":1533,"description":1534,"extension":144,"featured":148,"image":1535,"meta":1536,"navigation":148,"path":1537,"seo":1538,"stem":1540,"tags":1541,"__hash__":1542},"blog/blog/agentic-ai-studio-for-enterprises.md","Introducing the Agentic AI Studio for Enterprises",{"type":9,"value":1107,"toc":1521},[1108,1112,1117,1125,1142,1146,1155,1158,1161,1166,1169,1173,1186,1189,1198,1213,1216,1220,1225,1228,1231,1236,1239,1242,1249,1252,1256,1262,1268,1282,1310,1318,1322,1325,1344,1347,1352,1354,1358,1368,1374,1393,1396,1399,1419,1421,1425,1428,1431,1434,1437,1441,1444,1447,1450,1457,1464,1466,1470,1509],[16,1109,1111],{"id":1110},"from-ai-that-answers-to-ai-that-runs","From AI That Answers to AI That Runs",[12,1113,1114],{},[27,1115,1116],{},"Why agentic AI is the next operating layer of the enterprise.",[12,1118,1119,1120,1124],{},"For years, we've automated ",[1121,1122,1123],"em",{},"steps",". We've automated form submissions, approvals, ticket routing, invoice matching, report generation. If the sequence was predictable, we could script it. If the inputs were structured, we could build logic around them.",[12,1126,1127,1128,1131,1132,1135,1136,1141],{},"And yet, despite all that tooling, ",[27,1129,1130],{},"most enterprise work still relies heavily on human coordination",". Not because the tools are missing. But because traditional automation was built for predictability, and real business rarely is. The friction in organizations today isn't in executing steps. ",[27,1133,1134],{},"It's in coordinating them."," That's where ",[1137,1138,1140],"a",{"href":1139},"/products/agentic-ai","agentic AI"," enters the picture.",[16,1143,1145],{"id":1144},"the-automation-ceiling-weve-been-living-with","The automation ceiling we've been living with",[12,1147,1148,1151,1152],{},[27,1149,1150],{},"Classic automation systems"," work extremely well when the world behaves as expected. They assume a stable process, a fixed order of operations, clearly defined inputs, and a limited set of variations. But ",[27,1153,1154],{},"real workflows don't behave that way.",[12,1156,1157],{},"A request comes in slightly differently. A data source has changed. A policy was updated last week. An exception appears that wasn't considered when the automation was designed. Suddenly, the system stalls and a person steps back in.",[12,1159,1160],{},"Someone checks the ERP. Someone pulls a spreadsheet. Someone compares numbers manually. Someone rewrites the email because the template doesn't quite fit. Someone decides who should handle the issue.",[12,1162,1163],{},[27,1164,1165],{},"Most enterprise work is not step-based. It's judgment-based.",[12,1167,1168],{},"And that is precisely the boundary traditional automation struggles to cross.",[16,1170,1172],{"id":1171},"what-agentic-ai-actually-changes","What agentic AI actually changes",[12,1174,1175,1178,1179,1182,1183],{},[27,1176,1177],{},"Agentic AI"," shifts automation from ",[27,1180,1181],{},"step execution"," to ",[27,1184,1185],{},"outcome orchestration.",[12,1187,1188],{},"Instead of predefining every branch of a process, you define the goal and the constraints. The agent receives that goal, plans how to achieve it, interacts with systems, retrieves data, adapts when conditions change, and continues until the outcome is reached, or until human input is required.",[12,1190,1191,1192,1182,1195,1054],{},"That difference may sound subtle, but architecturally it is profound. We are moving from ",[27,1193,1194],{},"script-based automation",[27,1196,1197],{},"intent-driven systems",[179,1199,1200],{},[12,1201,1202,1203,1206,1209,1210],{},"Traditional automation asks: ",[1121,1204,1205],{},"\"What is the next step?\"",[1207,1208],"br",{},"\nAgentic AI asks: ",[1121,1211,1212],{},"\"What needs to be achieved?\"",[12,1214,1215],{},"That shift enables the automation of workflows that previously felt too messy, too cross-functional, or too dynamic to justify building rigid process logic around them.",[16,1217,1219],{"id":1218},"where-this-shows-up-in-the-real-world","Where this shows up in the real world",[12,1221,1222],{},[27,1223,1224],{},"Consider month-end close in finance.",[12,1226,1227],{},"This is not a single transaction. It is a coordination exercise across systems, reconciliations, exceptions, and reporting requirements. Teams retrieve data from ERP systems, compare figures against bank statements and subledgers, identify discrepancies, draft summaries, and prepare documentation for leadership review. The repetitive elements are predictable. The exceptions are not.",[12,1229,1230],{},"An agentic workflow can operate on schedule, retrieve required datasets, apply predefined financial rules, highlight anomalies, assemble a first draft of the close package, and escalate only when irregularities exceed thresholds. The human role shifts from assembling information to validating and interpreting it. The speed improves. The error rate drops. The audit trail remains intact.",[12,1232,1233],{},[27,1234,1235],{},"Or take sales operations.",[12,1237,1238],{},"When a new lead enters a CRM system, someone typically researches the account, checks internal history, selects the right messaging approach, drafts outreach, and logs the activity. It's not conceptually complex work, but it is coordination-heavy and time-sensitive.",[12,1240,1241],{},"An agentic system can enrich the account with connected data sources, apply the correct industry playbook, draft personalized outreach aligned with brand tone, suggest scheduling options, and log all activity back into the CRM. Sales teams focus on conversation and conversion, not preparation.",[12,1243,1244,1245,1248],{},"In both examples, ",[27,1246,1247],{},"the agent is not replacing expertise. It is absorbing coordination overhead."," That distinction matters.",[1250,1251],"hr",{},[16,1253,1255],{"id":1254},"why-this-moment-is-different","Why this moment is different",[12,1257,1258,1259],{},"AI has been embedded into enterprise software for years. Recommendation engines, predictive analytics, workflow suggestions, none of this is new. ",[27,1260,1261],{},"What changed is reasoning.",[12,1263,1264,1265],{},"Modern foundation models can now plan multi-step sequences, use tools, access APIs, and adjust behavior mid-process. They are not simply predicting the next word in a sentence; they are ",[27,1266,1267],{},"orchestrating interactions across systems.",[12,1269,1270,1271,1275,1276,1182,1279],{},"At the same time, enterprise demand has matured. Organizations are no longer experimenting with isolated ",[1137,1272,1274],{"href":1273},"/products/chat-bots","chatbot"," pilots. They are asking how AI can reshape operating models. The shift underway is from ",[27,1277,1278],{},"AI as interface",[27,1280,1281],{},"AI as infrastructure.",[12,1283,1284,1290,1291,1296,1297,1300,1301,1306,1307],{},[1137,1285,1289],{"href":1286,"rel":1287},"https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai-in-2023-generative-ais-breakout-year",[1288],"nofollow","Research and adoption patterns"," point in one direction: enterprises are increasingly turning to AI to automate complex, multi-step processes that involve reasoning and coordination, not just simple rules. ",[1137,1292,1295],{"href":1293,"rel":1294},"https://www.gartner.com/en/newsroom/press-releases/2024-10-21-gartner-identifies-the-top-10-strategic-technology-trends-for-2025",[1288],"Gartner"," has named agentic AI a top strategic technology trend for 2025, describing it as a ",[1121,1298,1299],{},"\"goal-driven digital workforce that autonomously makes plans and takes actions.\""," ",[1137,1302,1305],{"href":1303,"rel":1304},"https://www.bcg.com/publications/2025/how-agentic-ai-is-transforming-enterprise-platforms",[1288],"BCG"," notes that effective AI agents can accelerate business processes by 30% to 50%, and that ",[1121,1308,1309],{},"\"companies that embrace agentic AI now will gain a competitive edge in productivity, responsiveness, and innovation.\"",[179,1311,1312],{},[12,1313,1314,1315],{},"From answering questions to running processes. That is not a feature enhancement. ",[27,1316,1317],{},"It is an architectural layer.",[16,1319,1321],{"id":1320},"the-risk-hidden-in-the-excitement","The risk hidden in the excitement",[12,1323,1324],{},"The rapid evolution of agentic AI has created a crowded landscape. Every major AI provider is embedding workflow capabilities into their ecosystem. New orchestration platforms appear monthly. This creates opportunity, but also risk.",[12,1326,1327,1328,1331,1332,1335,1336,1339,1340,1343],{},"Enterprises operate under constraints that consumer tools rarely face. ",[27,1329,1330],{},"Data sovereignty"," requirements, ",[27,1333,1334],{},"regulatory frameworks",", internal security policies, ",[27,1337,1338],{},"regional hosting mandates",", and ",[27,1341,1342],{},"language support"," complexities all shape architectural decisions.",[12,1345,1346],{},"Committing deeply to a single AI provider or tightly coupling automation logic to one model ecosystem introduces long-term rigidity. AI models improve rapidly. Pricing changes. Policies evolve. Regulatory conditions shift. If workflows are built around one proprietary stack, adapting later can become expensive and disruptive.",[12,1348,1349],{},[27,1350,1351],{},"Flexibility in AI infrastructure is not simply a technical preference. It is strategic insurance.",[1250,1353],{},[16,1355,1357],{"id":1356},"designing-for-control-without-stagnation","Designing for control without stagnation",[12,1359,1360,1361,1364,1365],{},"When building the ",[1137,1362,1363],{"href":1139},"Agentic AI Studio"," at Blits.ai, our design principle was simple: ",[27,1366,1367],{},"control without limitation.",[12,1369,1370],{},[826,1371],{"alt":1372,"src":1373},"agentic-workflow","/images/agentic-workflow.png",[12,1375,1376,1377,1380,1381,1385,1386,1389,1390],{},"Enterprises need the ability to ",[27,1378,1379],{},"define outcomes in natural language",", connect workflows to existing systems, select appropriate AI models per use case (Blits.ai supports ",[1137,1382,1384],{"href":1383},"/integrations","OpenAI GPT, Google Gemini, Anthropic Claude, Meta Llama, DeepSeek, Mistral AI, and 80+ more","), and maintain full visibility into execution. They need ",[27,1387,1388],{},"auditability",". They need deployment flexibility, cloud, hybrid, or on-premises. And they need ",[27,1391,1392],{},"portability.",[12,1394,1395],{},"Model agnosticism is critical in this context. AI capability is evolving too quickly to anchor enterprise workflows permanently to a single provider. Organizations should be able to evaluate models based on performance, cost, language support, regulatory alignment, and policy fit, and change that selection without redesigning every process.",[12,1397,1398],{},"At the same time, governance cannot be an afterthought. Agentic systems must be inspectable. Enterprises must understand what triggered a workflow, which systems were accessed, how data was used, and where human escalation occurred. Responsible automation requires transparency.",[12,1400,1401,1402,1405,1406,1408,1409,1412,1413,1418],{},"The goal is not to chase novelty. It is to build ",[27,1403,1404],{},"an adaptive automation layer that remains stable"," even as AI capabilities evolve. The ",[1137,1407,1363],{"href":1139},", together with the Knowledge Library, ",[1137,1410,1411],{"href":1383},"Apps & Integrations",", content management, quality assurance, and the broader ",[1137,1414,1417],{"href":1415,"rel":1416},"https://www.blits.ai",[1288],"Blits.ai"," platform, is our answer for enterprises that want to adopt agentic AI without sacrificing control or flexibility.",[1250,1420],{},[16,1422,1424],{"id":1423},"the-operating-model-shift-ahead","The operating model shift ahead",[12,1426,1427],{},"Over the next several years, agentic AI will increasingly become the default mechanism for handling coordination-heavy work inside enterprises. Not because organizations want fewer people, but because they want fewer coordination bottlenecks.",[12,1429,1430],{},"Today, much of knowledge work involves routing, reconciling, formatting, escalating, and synthesizing across disconnected systems. These are precisely the kinds of tasks that agentic systems handle well. Humans remain responsible for interpretation, strategy, and accountability. But they are relieved from stitching systems together manually.",[12,1432,1433],{},"The transition mirrors earlier technological shifts. Databases did not eliminate decision-making; they eliminated manual record-keeping. ERP systems did not remove finance teams; they centralized fragmented processes. Agentic AI extends that progression into dynamic, reasoning-driven coordination.",[12,1435,1436],{},"The enterprises that embrace this shift thoughtfully will redesign workflows around outcomes rather than around tool limitations. Those that treat it as a temporary efficiency add-on may miss the broader architectural implications.",[16,1438,1440],{"id":1439},"a-different-question","A different question",[12,1442,1443],{},"The discussion around AI often centers on capability: Can it do this task? Can it replace this function? The more important question is structural. Should humans still be coordinating this process?",[12,1445,1446],{},"If the answer is no, then the solution is not another dashboard or another integration script. It is an outcome-driven orchestration layer that adapts as conditions change. Agentic AI represents that layer.",[12,1448,1449],{},"The shift from AI that answers to AI that runs is already underway. The strategic choice facing enterprises is not whether it will happen, but how deliberately they design for it.",[12,1451,1452,1453,1456],{},"And design, in this context, means preserving ",[27,1454,1455],{},"flexibility, governance, and control"," while unlocking a new level of operational intelligence.",[179,1458,1459],{},[12,1460,1461],{},[27,1462,1463],{},"That is the frontier.",[1250,1465],{},[16,1467,1469],{"id":1468},"summary-and-next-steps","Summary and next steps",[21,1471,1472,1477,1484,1490,1497],{},[24,1473,1474,1476],{},[27,1475,1177],{}," shifts automation from step execution to outcome orchestration: you define the goal and constraints, the agent plans, interacts with systems, and adapts until the outcome is reached or human input is needed.",[24,1478,1479,1480,1483],{},"Most enterprise work is ",[27,1481,1482],{},"judgment-based",", not step-based; traditional automation hits a ceiling when the process or data changes, agentic AI is built for that boundary.",[24,1485,1486,1489],{},[27,1487,1488],{},"Real-world use"," shows up in finance (e.g. month-end close) and sales (e.g. lead enrichment and outreach); the agent absorbs coordination overhead, humans focus on validation and conversion.",[24,1491,1492,1493,1496],{},"The shift is from ",[27,1494,1495],{},"AI as interface to AI as infrastructure","; flexibility and model agnosticism are strategic insurance against lock-in, sovereignty, and regulatory risk.",[24,1498,1499,1500,1504,1505,1508],{},"The ",[27,1501,1502],{},[1137,1503,1363],{"href":1139}," at Blits.ai is designed for control without limitation: natural-language workflows, ",[1137,1506,1507],{"href":1383},"80+ AI providers",", auditability, and deployment flexibility (cloud, hybrid, on-premises).",[12,1510,1511,1517,1518,1520],{},[27,1512,1513],{},[1137,1514,1516],{"href":1515},"/contact-us","Contact sales"," to get access to the ",[1137,1519,1363],{"href":1139}," and run autonomous workflows on your infrastructure with your models and your data.",{"title":135,"searchDepth":136,"depth":136,"links":1522},[1523,1524,1525,1526,1527,1528,1529,1530,1531,1532],{"id":1110,"depth":136,"text":1111},{"id":1144,"depth":136,"text":1145},{"id":1171,"depth":136,"text":1172},{"id":1218,"depth":136,"text":1219},{"id":1254,"depth":136,"text":1255},{"id":1320,"depth":136,"text":1321},{"id":1356,"depth":136,"text":1357},{"id":1423,"depth":136,"text":1424},{"id":1439,"depth":136,"text":1440},{"id":1468,"depth":136,"text":1469},"2026-02-17T00:00:00.000Z","Self-running AI agents are reshaping enterprise processes. We're releasing the Agentic AI Studio for Blits, a platform-agnostic, enterprise-ready way to design and run autonomous workflows with your own models and data.","/images/blits-agentic-studio.png",{},"/blog/agentic-ai-studio-for-enterprises",{"title":1539,"description":1534},"Introducing the Agentic AI Studio for Enterprise","blog/agentic-ai-studio-for-enterprises",[],"571wYqDZTF_d7egcewfXI6Q5Cz0NnRbubZNBkIeSgZU",{"id":1544,"title":1545,"author":7,"body":1546,"category":333,"date":1738,"description":1739,"extension":144,"featured":145,"image":1740,"meta":1741,"navigation":148,"path":1742,"seo":1743,"stem":1744,"tags":1745,"__hash__":1746},"blog/blog/agentic-pay-and-the-moment-ai-was-allowed-to-spend-money-by-itself.md","Agentic Pay and the Moment AI Was Allowed to Spend Money",{"type":9,"value":1547,"toc":1728},[1548,1551,1554,1557,1560,1563,1567,1570,1573,1576,1595,1602,1606,1609,1612,1615,1618,1622,1625,1633,1636,1640,1643,1646,1649,1652,1656,1659,1662,1665,1673,1677,1703,1707,1710,1713,1716,1719,1722,1725],[12,1549,1550],{},"There is a clear line in AI adoption where curiosity turns into discomfort.",[12,1552,1553],{},"That line is money.",[12,1555,1556],{},"Most people are fine letting an AI explain things, summarize documents, or suggest decisions. The moment you suggest letting it actually spend money, the reaction changes. And for good reason.",[12,1558,1559],{},"Traditional payment systems were never designed for non human actors.",[12,1561,1562],{},"At the same time, the ecosystem is moving in exactly that direction. Agents embedded in experiences like AI Mode in Search, Gemini, or ChatGPT are starting to place real orders: buying products, rebooking travel, renewing software, and managing subscriptions. The question is no longer whether AI will be allowed to spend money—it is how we design the rails so that when it does, it stays inside clear, auditable boundaries.",[16,1564,1566],{"id":1565},"why-payments-break-when-ai-gets-involved","Why payments break when AI gets involved",[12,1568,1569],{},"Payment infrastructure assumes a person is on the other end. Someone who owns a card, confirms intent, and carries legal responsibility. Large language models break every one of those assumptions.",[12,1571,1572],{},"They do not own money. They do not have intent in the human sense. They cannot be held accountable.",[12,1574,1575],{},"Yet they increasingly operate in domains where economic action is unavoidable.",[12,1577,1578,1579,1582,1583,1588,1589,1594],{},"On the commerce side, efforts like Google’s Universal Commerce Protocol (UCP) focus on making the ",[1121,1580,1581],{},"order"," itself machine-readable and agent-friendly, so an LLM can reason about line items, totals, and terms before committing (",[1137,1584,1587],{"href":1585,"rel":1586},"https://developers.googleblog.com/under-the-hood-universal-commerce-protocol-ucp/",[1288],"Under the Hood: Universal Commerce Protocol","). On the app side, OpenAI’s ",[1137,1590,1593],{"href":1591,"rel":1592},"https://developers.openai.com/commerce/",[1288],"Agentic Commerce"," and the Agentic Commerce Protocol (ACP) define how ChatGPT agents can discover purchasable actions, manage checkout flows, and keep users informed while they transact.",[12,1596,1597,1598,1601],{},"Agentic Pay exists to resolve the remaining contradiction on the ",[1121,1599,1600],{},"payments"," layer: how to let agents move money within those commerce protocols, without handing them a blank cheque.",[16,1603,1605],{"id":1604},"what-agentic-pay-really-means","What Agentic Pay really means",[12,1607,1608],{},"Agentic Pay does not give AI access to money.",[12,1610,1611],{},"It gives AI delegated authority.",[12,1613,1614],{},"A human or organization defines the rules. What the agent is allowed to buy. How much it may spend. Under which conditions. For how long. The agent operates strictly within that mandate.",[12,1616,1617],{},"This turns payments from an implicit risk into an explicit capability with boundaries.",[16,1619,1621],{"id":1620},"what-this-looks-like-in-practice","What this looks like in practice",[12,1623,1624],{},"Here are two simple examples that business leaders can map to real workflows:",[21,1626,1627,1630],{},[24,1628,1629],{},"A procurement agent is allowed to reorder specific SKUs from approved vendors when stock drops below a threshold. It has a monthly limit, logs every purchase, and escalates anything outside the contract price.",[24,1631,1632],{},"A travel agent can rebook a flight if a delay exceeds two hours, but only within policy, only for preapproved employees, and only if the cost delta is under 200 USD.",[12,1634,1635],{},"In both cases, the agent is not free to spend. It is executing a narrow mandate with transparent guardrails.",[16,1637,1639],{"id":1638},"governance-first-versus-execution-first","Governance first versus execution first",[12,1641,1642],{},"Different players approach Agentic Pay from different angles.",[12,1644,1645],{},"Google approaches this problem from a governance first perspective with its Agent Payments Protocol, often referred to as AP2. The core idea is traceability. Every action an agent takes must be attributable to a delegation granted by a real entity. Limits are enforced by design, not by convention. Observability is not optional. Combined with UCP, you get a stack where the commerce journey and the payment authorization are both explicit, signed, and provable end to end.",[12,1647,1648],{},"Stripe and OpenAI focus more heavily on execution with the Agentic Commerce Protocol, often referred to as ACP. Their approach fits directly into how LLMs already reason and plan. The model can discover a purchasable action, evaluate constraints, request approval when needed, and execute the transaction without falling back to human oriented checkout flows—exactly the kind of patterns described in the Agentic Commerce guides.",[12,1650,1651],{},"Both approaches solve the same problem from opposite sides. Control versus flow. Agentic Pay sits where they meet: it is the discipline of designing delegated payment rights that plug cleanly into those emerging commerce protocols.",[16,1653,1655],{"id":1654},"what-this-unlocks-for-llms","What this unlocks for LLMs",[12,1657,1658],{},"Once payments become agent native, LLMs cross a critical threshold.",[12,1660,1661],{},"They stop being systems that talk about work and become systems that perform work. Procurement agents that optimize spend continuously. Travel agents that rebook instantly when conditions change. Finance agents that manage recurring obligations without reminders or follow ups. Ecommerce agents that move from “here are some sneakers you might like” to “I have selected the best option under your budget and policy and placed the order using your delegated payment instrument.”",[12,1663,1664],{},"Money turns reasoning into responsibility.",[179,1666,1667],{},[12,1668,1669,1670,1672],{},"“The future of online ordering is that it should feel as streamlined as a McDonald’s drive‑through: clear choices, fast confirmation, and no surprises—only this time, your agents are the ones in the driver’s seat.” ",[1207,1671],{},"\n— Len Debets",[16,1674,1676],{"id":1675},"what-business-leaders-should-care-about","What business leaders should care about",[21,1678,1679,1685,1691,1697],{},[24,1680,1681,1684],{},[27,1682,1683],{},"Speed with control:"," Routine purchases happen faster without losing approval boundaries.",[24,1686,1687,1690],{},[27,1688,1689],{},"Cost discipline:"," Policies become executable code, not PDF guidelines.",[24,1692,1693,1696],{},[27,1694,1695],{},"Audit readiness:"," Every action is attributable, logged, and reviewable.",[24,1698,1699,1702],{},[27,1700,1701],{},"Scale:"," The same rules can govern thousands of micro decisions without extra headcount.",[16,1704,1706],{"id":1705},"the-risks-are-real","The risks are real",[12,1708,1709],{},"Mistakes now have financial consequences. Incentives matter. Security failures are no longer theoretical.",[12,1711,1712],{},"That is why every serious Agentic Pay design is built around limits, reversibility, and auditability. Autonomy without control is not innovation. It is negligence.",[16,1714,1715],{"id":313},"Final thoughts",[12,1717,1718],{},"Agentic Pay is uncomfortable because it forces trust to become explicit, but it is also inevitable. Once intelligence can reason, plan, and act within defined boundaries, organizations will stop routing everything through humans by default.",[12,1720,1721],{},"The question is no longer whether AI will be allowed to spend money.",[12,1723,1724],{},"The question is who designs the rules under which it does.",[12,1726,1727],{},"At Blits, we work with large banks and financial institutions—and partner with one of the major global credit card networks—to design and implement Agentic Pay architectures that meet real-world regulatory, risk, and governance requirements.",{"title":135,"searchDepth":136,"depth":136,"links":1729},[1730,1731,1732,1733,1734,1735,1736,1737],{"id":1565,"depth":136,"text":1566},{"id":1604,"depth":136,"text":1605},{"id":1620,"depth":136,"text":1621},{"id":1638,"depth":136,"text":1639},{"id":1654,"depth":136,"text":1655},{"id":1675,"depth":136,"text":1676},{"id":1705,"depth":136,"text":1706},{"id":313,"depth":136,"text":1715},"2026-01-11T00:00:00.000Z","Why giving AI agents controlled access to payments changes everything. The question is who designs the rules under which it does.","/images/blits-agentic-pay.jpg",{},"/blog/agentic-pay-and-the-moment-ai-was-allowed-to-spend-money-by-itself",{"title":1545,"description":1739},"blog/agentic-pay-and-the-moment-ai-was-allowed-to-spend-money-by-itself",[],"Q5ufvUFNDiBeqAZ_67rScC7XAEHfXCd0t_XrViru2iQ",{"id":1748,"title":1749,"author":7,"body":1750,"category":333,"date":1972,"description":1973,"extension":144,"featured":145,"image":1974,"meta":1975,"navigation":148,"path":1976,"seo":1977,"stem":1978,"tags":1979,"__hash__":1982},"blog/blog/ai-spending-delegation-policies-without-losing-control.md","AI Spending Delegation Policies: How to Let Agents Transact Without Losing Control",{"type":9,"value":1751,"toc":1959},[1752,1755,1758,1761,1764,1767,1771,1774,1779,1783,1788,1791,1795,1798,1802,1805,1809,1812,1816,1819,1823,1826,1829,1937,1941,1944,1946,1949,1952,1955],[12,1753,1754],{},"The moment AI can spend money, governance becomes real.",[12,1756,1757],{},"This is where many organizations hesitate, and rightly so.",[12,1759,1760],{},"But \"no\" is not a strategy. Agents will increasingly operate in procurement, travel, subscriptions, and operational purchasing.",[12,1762,1763],{},"The question is how to define spending authority without introducing unmanaged risk.",[12,1765,1766],{},"{image}",[16,1768,1770],{"id":1769},"delegation-is-not-autonomy","Delegation is not autonomy",[12,1772,1773],{},"A strong policy starts with one principle: an agent never receives open-ended payment rights. It receives a mandate. That mandate should define what can be purchased, in which context, at which budget level, and through which approval path. When those boundaries are explicit, spending becomes governable instead of unpredictable.",[179,1775,1776],{},[12,1777,1778],{},"\"Delegation is not trust by default. Delegation is trust with constraints.\"",[16,1780,1782],{"id":1781},"the-five-controls-every-policy-needs","The five controls every policy needs",[1784,1785,1787],"h3",{"id":1786},"_1-scope-control","1) Scope control",[12,1789,1790],{},"Define approved categories, vendors, and SKUs in a way machines can evaluate deterministically. If an agent cannot map a request to an approved scope, it should fail safely and escalate rather than improvise.",[1784,1792,1794],{"id":1793},"_2-value-control","2) Value control",[12,1796,1797],{},"Set layered spending limits per transaction and per period, and include cumulative exposure across related agents. Most incidents are not single large transactions; they are many small decisions that add up.",[1784,1799,1801],{"id":1800},"_3-context-control","3) Context control",[12,1803,1804],{},"Payments should only execute when required business context is present: a valid trigger, budget ownership, and contract-policy alignment. If context is incomplete, the default behavior should be no transaction.",[1784,1806,1808],{"id":1807},"_4-approval-control","4) Approval control",[12,1810,1811],{},"Approval should be tiered by risk and anomaly profile. Low-risk recurring spend can run automatically within policy, while exceptions and unusual behavior routes to managers or finance.",[1784,1813,1815],{"id":1814},"_5-audit-control","5) Audit control",[12,1817,1818],{},"Every transaction must be reconstructable from delegation source to final settlement result. If you cannot replay who approved what and under which policy version, you do not have operational control.",[16,1820,1822],{"id":1821},"a-policy-model-that-scales","A policy model that scales",[12,1824,1825],{},"Most teams start with static limits and quickly hit edge cases.",[12,1827,1828],{},"A stronger approach is policy-as-code with versioning, test gates, change approvals, and rollback capability. This allows governance to evolve at operational speed without becoming a manual bottleneck.",[1029,1830,1834],{"className":1831,"code":1832,"language":1833,"meta":135,"style":135},"language-yaml shiki shiki-themes material-theme-lighter material-theme material-theme-palenight","delegation_policy:\n  category: \"software_subscriptions\"\n  max_transaction_eur: 500\n  monthly_cap_eur: 5000\n  required_context: [\"cost_center\", \"vendor_whitelist_match\"]\n  approval: \"manager_if_exception\"\n","yaml",[1036,1835,1836,1849,1867,1879,1890,1922],{"__ignoreMap":135},[1837,1838,1841,1845],"span",{"class":1839,"line":1840},"line",1,[1837,1842,1844],{"class":1843},"swJcz","delegation_policy",[1837,1846,1848],{"class":1847},"sMK4o",":\n",[1837,1850,1851,1854,1857,1860,1864],{"class":1839,"line":136},[1837,1852,1853],{"class":1843},"  category",[1837,1855,1856],{"class":1847},":",[1837,1858,1859],{"class":1847}," \"",[1837,1861,1863],{"class":1862},"sfazB","software_subscriptions",[1837,1865,1866],{"class":1847},"\"\n",[1837,1868,1870,1873,1875],{"class":1839,"line":1869},3,[1837,1871,1872],{"class":1843},"  max_transaction_eur",[1837,1874,1856],{"class":1847},[1837,1876,1878],{"class":1877},"sbssI"," 500\n",[1837,1880,1882,1885,1887],{"class":1839,"line":1881},4,[1837,1883,1884],{"class":1843},"  monthly_cap_eur",[1837,1886,1856],{"class":1847},[1837,1888,1889],{"class":1877}," 5000\n",[1837,1891,1893,1896,1898,1901,1904,1907,1909,1912,1914,1917,1919],{"class":1839,"line":1892},5,[1837,1894,1895],{"class":1843},"  required_context",[1837,1897,1856],{"class":1847},[1837,1899,1900],{"class":1847}," [",[1837,1902,1903],{"class":1847},"\"",[1837,1905,1906],{"class":1862},"cost_center",[1837,1908,1903],{"class":1847},[1837,1910,1911],{"class":1847},",",[1837,1913,1859],{"class":1847},[1837,1915,1916],{"class":1862},"vendor_whitelist_match",[1837,1918,1903],{"class":1847},[1837,1920,1921],{"class":1847},"]\n",[1837,1923,1925,1928,1930,1932,1935],{"class":1839,"line":1924},6,[1837,1926,1927],{"class":1843},"  approval",[1837,1929,1856],{"class":1847},[1837,1931,1859],{"class":1847},[1837,1933,1934],{"class":1862},"manager_if_exception",[1837,1936,1866],{"class":1847},[16,1938,1940],{"id":1939},"red-flags-to-avoid","Red flags to avoid",[12,1942,1943],{},"One global payment permission for all agents is usually the first design smell. Other recurring problems are manual approvals without structured logs, drift detection that is never implemented, and emergency exceptions that quietly become permanent.",[16,1945,1041],{"id":1040},[12,1947,1948],{},"Agentic spending is not dangerous because agents act.",[12,1950,1951],{},"It is dangerous when organizations fail to define the mandate clearly.",[12,1953,1954],{},"If your delegation model is explicit, testable, and auditable, AI payments can become one of the highest-value automation layers in the enterprise.",[1956,1957,1958],"style",{},"html pre.shiki code .swJcz, html code.shiki .swJcz{--shiki-light:#E53935;--shiki-default:#F07178;--shiki-dark:#F07178}html pre.shiki code .sMK4o, html code.shiki .sMK4o{--shiki-light:#39ADB5;--shiki-default:#89DDFF;--shiki-dark:#89DDFF}html pre.shiki code .sfazB, html code.shiki .sfazB{--shiki-light:#91B859;--shiki-default:#C3E88D;--shiki-dark:#C3E88D}html pre.shiki code .sbssI, html code.shiki .sbssI{--shiki-light:#F76D47;--shiki-default:#F78C6C;--shiki-dark:#F78C6C}html .light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html.light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}html.dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}",{"title":135,"searchDepth":136,"depth":136,"links":1960},[1961,1962,1969,1970,1971],{"id":1769,"depth":136,"text":1770},{"id":1781,"depth":136,"text":1782,"children":1963},[1964,1965,1966,1967,1968],{"id":1786,"depth":1869,"text":1787},{"id":1793,"depth":1869,"text":1794},{"id":1800,"depth":1869,"text":1801},{"id":1807,"depth":1869,"text":1808},{"id":1814,"depth":1869,"text":1815},{"id":1821,"depth":136,"text":1822},{"id":1939,"depth":136,"text":1940},{"id":1040,"depth":136,"text":1041},"2026-03-31T00:00:00.000Z","Letting AI agents spend money is no longer experimental. This article explains how to design delegation policies that make agentic payments auditable, bounded, and safe for enterprise use.","/images/blog-ai-policies.png",{},"/blog/ai-spending-delegation-policies-without-losing-control",{"title":1749,"description":1973},"blog/ai-spending-delegation-policies-without-losing-control",[1980,1981,1101,1600],"agentic pay","ai governance","IZaMadbOHT6hR6u8Z40rsmZ4QeXQuKrtxoUWaZ1JSWs",{"id":1984,"title":1985,"author":1986,"body":1987,"category":2037,"date":2038,"description":1991,"extension":144,"featured":145,"image":2039,"meta":2040,"navigation":148,"path":2041,"seo":2042,"stem":2043,"tags":340,"__hash__":2044},"blog/blog/benchmarking-conversational-ai-whitepaper.md","Benchmarking Conversational AI – Whitepaper","Paul Coerkamp",{"type":9,"value":1988,"toc":2031},[1989,1992,1996,1999,2003,2017,2021,2024,2028],[12,1990,1991],{},"Our comprehensive research into conversational AI performance reveals significant improvements in chatbot effectiveness across multiple industries and use cases.",[16,1993,1995],{"id":1994},"research-methodology","Research Methodology",[12,1997,1998],{},"Our benchmarking approach evaluates conversational AI systems across multiple dimensions including accuracy, response time, user satisfaction, and business impact metrics.",[16,2000,2002],{"id":2001},"key-findings","Key Findings",[21,2004,2005,2008,2011,2014],{},[24,2006,2007],{},"Average response accuracy improved by 34% with modern AI models",[24,2009,2010],{},"User satisfaction scores increased by 28% in enterprise deployments",[24,2012,2013],{},"Resolution rates for customer service queries improved by 45%",[24,2015,2016],{},"Cost reduction of 60% compared to traditional support channels",[16,2018,2020],{"id":2019},"industry-applications","Industry Applications",[12,2022,2023],{},"The whitepaper covers specific applications across healthcare, finance, retail, and technology sectors, providing actionable insights for organizations looking to implement conversational AI solutions.",[16,2025,2027],{"id":2026},"future-outlook","Future Outlook",[12,2029,2030],{},"Our research indicates continued rapid improvement in conversational AI capabilities, with significant advances expected in the next 12-18 months.",{"title":135,"searchDepth":136,"depth":136,"links":2032},[2033,2034,2035,2036],{"id":1994,"depth":136,"text":1995},{"id":2001,"depth":136,"text":2002},{"id":2019,"depth":136,"text":2020},{"id":2026,"depth":136,"text":2027},"Research","2022-08-25T00:00:00.000Z","/blog/Cover-Benchmarking-Conversational-AI-Whitepaper.jpg",{},"/blog/benchmarking-conversational-ai-whitepaper",{"title":1985,"description":1991},"blog/benchmarking-conversational-ai-whitepaper","HLlh69330I7R6HrHArgom3BKfw0rVe1441Rc8h0iEW8",{"id":2046,"title":2047,"author":1986,"body":2048,"category":333,"date":2160,"description":2161,"extension":144,"featured":145,"image":2162,"meta":2163,"navigation":148,"path":2164,"seo":2165,"stem":2166,"tags":2167,"__hash__":2172},"blog/blog/benchmarking-tts-for-customers-beyond-mos-and-demos.md","Benchmarking TTS for Real Customers: Beyond MOS and Demo Scores",{"type":9,"value":2049,"toc":2148},[2050,2053,2056,2059,2062,2064,2068,2071,2075,2078,2082,2085,2089,2092,2096,2099,2103,2106,2111,2115,2118,2121,2124,2130,2134,2137,2139,2142,2145],[12,2051,2052],{},"Most TTS comparisons look great in demos.",[12,2054,2055],{},"That is exactly the problem.",[12,2057,2058],{},"A controlled demo clip does not tell you how a voice model performs under live traffic, noisy prompts, mixed languages, or strict latency targets.",[12,2060,2061],{},"If you benchmark only MOS-style quality scores, you will optimize for the wrong outcome.",[12,2063,1766],{},[16,2065,2067],{"id":2066},"what-to-measure-instead","What to measure instead",[12,2069,2070],{},"A production-ready benchmark should combine perception, responsiveness, reliability, and business impact in one measurement model. Isolated quality scores are still useful, but they should never be the only success signal.",[1784,2072,2074],{"id":2073},"_1-perceived-experience-metrics","1) Perceived experience metrics",[12,2076,2077],{},"Naturalness and intelligibility matter, but consistency across longer interactions matters even more. In production, users notice pronunciation drift, unstable pacing, and awkward handling of names and domain terminology.",[1784,2079,2081],{"id":2080},"_2-real-time-performance-metrics","2) Real-time performance metrics",[12,2083,2084],{},"Measure time to first audio, end-to-end latency, jitter under load, and barge-in behavior. Users remember responsiveness more than lab-grade acoustic perfection.",[1784,2086,2088],{"id":2087},"_3-reliability-metrics","3) Reliability metrics",[12,2090,2091],{},"Track synthesis failures, retry patterns, regional degradation behavior, and fallback success. Reliability is usually where providers separate once traffic becomes real.",[1784,2093,2095],{"id":2094},"_4-business-impact-metrics","4) Business impact metrics",[12,2097,2098],{},"If a benchmark does not connect to containment, handling time, conversion, or cost-to-serve, it is incomplete. Voice quality without business relevance is still a vanity metric.",[16,2100,2102],{"id":2101},"build-your-test-set-like-a-product-team-not-a-research-team","Build your test set like a product team, not a research team",[12,2104,2105],{},"Use realistic prompts from your own support and sales channels: short and long turns, sensitive conversations, multilingual code-switching, and difficult entities like addresses, IDs, and policy clauses. Generic benchmark text tends to hide the exact failures that hurt customer trust.",[179,2107,2108],{},[12,2109,2110],{},"\"The best benchmark dataset sounds messy because real customers sound messy.\"",[16,2112,2114],{"id":2113},"compare-complete-voice-stacks-not-isolated-tts-models","Compare complete voice stacks, not isolated TTS models",[12,2116,2117],{},"In real assistants, TTS sits in a pipeline with STT, LLM reasoning, tool calls, and channel transport.",[12,2119,2120],{},"A fast TTS model can still feel slow if orchestration is weak. A high-quality model can still fail trust if pronunciation post-processing is poor.",[12,2122,2123],{},"Benchmark end-to-end experience, not just synthesis in isolation.",[1029,2125,2128],{"className":2126,"code":2127,"language":1034,"meta":135},[1032],"Voice KPI target profile (example):\n- Time to first audio: \u003C 500 ms\n- P95 full response: \u003C 2.5 s\n- Synthesis failure rate: \u003C 0.3%\n- Pronunciation critical-term accuracy: > 98%\n",[1036,2129,2127],{"__ignoreMap":135},[16,2131,2133],{"id":2132},"common-benchmarking-mistakes","Common benchmarking mistakes",[12,2135,2136],{},"Teams still make avoidable mistakes: testing only English in multilingual environments, tracking averages without P95/P99, and skipping fallback validation during incidents. Another costly pattern is selecting providers before governance and data constraints are mapped.",[16,2138,1041],{"id":1040},[12,2140,2141],{},"The best TTS engine is not the one that wins one clean demo.",[12,2143,2144],{},"It is the one that keeps sounding natural, fast, and reliable across your real workflows, languages, and peak conditions.",[12,2146,2147],{},"Benchmark for reality, and your production results will follow.",{"title":135,"searchDepth":136,"depth":136,"links":2149},[2150,2156,2157,2158,2159],{"id":2066,"depth":136,"text":2067,"children":2151},[2152,2153,2154,2155],{"id":2073,"depth":1869,"text":2074},{"id":2080,"depth":1869,"text":2081},{"id":2087,"depth":1869,"text":2088},{"id":2094,"depth":1869,"text":2095},{"id":2101,"depth":136,"text":2102},{"id":2113,"depth":136,"text":2114},{"id":2132,"depth":136,"text":2133},{"id":1040,"depth":136,"text":1041},"2026-04-08T00:00:00.000Z","Most text-to-speech evaluations still optimize for demo quality, not business outcomes. This guide explains how to benchmark TTS for real customer experience, latency targets, and operational reliability.","/images/blog-benchmark.png",{},"/blog/benchmarking-tts-for-customers-beyond-mos-and-demos",{"title":2047,"description":2161},"blog/benchmarking-tts-for-customers-beyond-mos-and-demos",[2168,2169,2170,2171],"tts","voice ai","benchmarking","ai performance","CJUuhun8lYwhiZYf2acA4lI96QZcwe8hqFHcP12xCvQ",{"id":2174,"title":2175,"author":7,"body":2176,"category":2580,"date":2581,"description":2582,"extension":144,"featured":145,"image":2583,"meta":2584,"navigation":148,"path":2585,"seo":2586,"stem":2587,"tags":2588,"__hash__":2593},"blog/blog/blits-ai-bug-hunt-game-2026-results.md","What We Learned from Turning Bug Hunting into a Game at Blits.ai",{"type":9,"value":2177,"toc":2567},[2178,2181,2184,2199,2202,2206,2209,2230,2237,2266,2279,2283,2286,2321,2324,2335,2338,2342,2345,2368,2371,2390,2393,2397,2400,2404,2407,2421,2424,2428,2435,2443,2450,2454,2457,2468,2475,2479,2482,2485,2508,2511,2515,2518,2538,2541,2543,2558,2561,2564],[12,2179,2180],{},"In the first week of January we did something a bit unusual at Blits.ai:",[12,2182,2183],{},"We turned bug hunting into a game.",[12,2185,2186,2187,2190,2191,2194,2195,2198],{},"Instead of treating quality as a background activity, we made it explicit, visible, and competitive. Non‑technical colleagues became ",[27,2188,2189],{},"Hunters",". Engineers became ",[27,2192,2193],{},"Fixers",". I played the ",[27,2196,2197],{},"Judge",". And behind the scenes, a Nuxt.js app pulled live data from Azure DevOps to keep score on a real‑time leaderboard.",[12,2200,2201],{},"This post is a recap of what we built, what actually happened in that week, and what we learned.",[16,2203,2205],{"id":2204},"how-the-bug-hunt-game-worked","How the Bug Hunt Game worked",[12,2207,2208],{},"We defined three roles:",[21,2210,2211,2216,2221],{},[24,2212,2213,2215],{},[27,2214,2189],{}," (non‑technical): responsible for finding and documenting bugs in enough detail that someone else could reproduce and fix them.",[24,2217,2218,2220],{},[27,2219,2193],{}," (technical): responsible for taking bugs from “this is broken” to “this is fixed and tested”.",[24,2222,2223,2225,2226,2229],{},[27,2224,2197],{}," (me): responsible for enforcing the rules, validating quality, and occasionally ",[1121,2227,2228],{},"stealing"," points when the process wasn’t followed.",[12,2231,2232,2233,2236],{},"Everything was tracked in Azure DevOps. If a bug had the ",[1036,2234,2235],{},"Bug Hunt Game 2026"," flag, it became part of the game. The Nuxt leaderboard app pulled that data through the Azure DevOps API and calculated points in real time:",[21,2238,2239,2245,2250,2255,2260],{},[24,2240,2241,2244],{},[27,2242,2243],{},"+1 point"," for each valid bug report (Hunter).",[24,2246,2247,2249],{},[27,2248,2243],{}," if the original Hunter verified the fix themselves.",[24,2251,2252,2254],{},[27,2253,2243],{}," for adding a help center article or instructions.",[24,2256,2257,2259],{},[27,2258,2243],{}," for fixing a bug (Fixer).",[24,2261,2262,2265],{},[27,2263,2264],{},"+2 points"," for adding unit tests, e2e tests or recurring validation to prevent regressions.",[12,2267,2268,2269,544,2272,2275,2276,2229],{},"And because every game needs some chaos, the Judge had the power to ",[27,2270,2271],{},"invalidate",[27,2273,2274],{},"bonus",", or ",[27,2277,2278],{},"steal",[16,2280,2282],{"id":2281},"the-actual-numbers-from-week-one","The actual numbers from week one",[12,2284,2285],{},"By the end of the first week of January, the scoreboard looked like this:",[21,2287,2288,2294,2300,2309,2315],{},[24,2289,2290,2293],{},[27,2291,2292],{},"49 bugs"," identified and tracked with the game flag.",[24,2295,2296,2299],{},[27,2297,2298],{},"32 bugs"," resolved and verified as fixed.",[24,2301,2302,2305,2306,1054],{},[27,2303,2304],{},"85 total points"," earned across ",[27,2307,2308],{},"7 active players",[24,2310,2311,2314],{},[27,2312,2313],{},"15 new unit tests"," added to our codebase.",[24,2316,2317,2320],{},[27,2318,2319],{},"8 bugs"," still in validation (fixed but waiting for final confirmation).",[12,2322,2323],{},"On paper, those are just numbers. In practice they meant:",[21,2325,2326,2329,2332],{},[24,2327,2328],{},"We surfaced bugs that would normally sit in someone’s notes or chat history.",[24,2330,2331],{},"We pulled non‑technical colleagues directly into the quality process.",[24,2333,2334],{},"We turned “please add tests” into a concrete, rewarded action.",[12,2336,2337],{},"Instead of a vague “we should improve quality this quarter”, we had a visible, measurable week where quality work was the only thing that mattered.",[16,2339,2341],{"id":2340},"who-ended-up-on-top-and-why","Who ended up on top (and why)",[12,2343,2344],{},"The top three players on the leaderboard were:",[21,2346,2347,2354,2361],{},[24,2348,2349,2350,2353],{},"🥇 ",[27,2351,2352],{},"Player 1",", 23 points (9 bugs reported, 11 fixed, 2 unit tests).",[24,2355,2356,2357,2360],{},"🥈 ",[27,2358,2359],{},"Player 2",", 22 points (15 bugs reported, clear top Hunter).",[24,2362,2363,2364,2367],{},"🥉 ",[27,2365,2366],{},"Player 3",", 14 points (3 bugs reported, 8 fixed, 2 unit tests).",[12,2369,2370],{},"A few interesting patterns:",[21,2372,2373,2384,2387],{},[24,2374,2375,2376,2379,2380,2383],{},"The people who ",[27,2377,2378],{},"reported"," bugs and also ",[27,2381,2382],{},"fixed"," them tended to climb fastest. The “Both” role (Hunter + Fixer) was naturally rewarded.",[24,2385,2386],{},"Pure Hunters still had a strong impact. Yuri didn’t fix bugs, but his reports unlocked a lot of downstream work.",[24,2388,2389],{},"Unit tests quietly made a difference. They are “only” +2 points, but they often decided the difference in tight rankings.",[12,2391,2392],{},"I, as the Judge, was of course on the leaderboard—but always anchored to the bottom and excluded from winning. My role was to keep the game fair, not to win it.",[16,2394,2396],{"id":2395},"what-changed-in-our-behaviour","What changed in our behaviour",[12,2398,2399],{},"The most important outcome was not the points, it was the behaviour they created.",[1784,2401,2403],{"id":2402},"_1-better-bug-reports-from-nontechnical-teammates","1. Better bug reports from non‑technical teammates",[12,2405,2406],{},"Hunters quickly learned that vague “it doesn’t work” tickets didn’t score. To earn points, they needed:",[21,2408,2409,2412,2415,2418],{},[24,2410,2411],{},"A clear description.",[24,2413,2414],{},"Steps to reproduce.",[24,2416,2417],{},"Environment or device information.",[24,2419,2420],{},"Evidence like screenshots or video.",[12,2422,2423],{},"After a few rounds of feedback, the average bug report quality went up dramatically. Fixers spent less time guessing and more time actually fixing.",[1784,2425,2427],{"id":2426},"_2-a-healthier-pressure-to-add-tests","2. A healthier pressure to add tests",[12,2429,2430,2431,2434],{},"Developers already know they ",[1121,2432,2433],{},"should"," add tests. The game turned that into a visible incentive:",[21,2436,2437,2440],{},[24,2438,2439],{},"Add a test → get rewarded.",[24,2441,2442],{},"Skip tests → you can still fix the bug, but you leave points on the table.",[12,2444,2445,2446,2449],{},"By the end of the week we had ",[27,2447,2448],{},"15 new tests"," in the codebase, tied to real bugs that had caused real problems. That’s much more valuable than abstract “test coverage” goals.",[1784,2451,2453],{"id":2452},"_3-shared-language-between-business-and-engineering","3. Shared language between business and engineering",[12,2455,2456],{},"Because everything lived on a single leaderboard, non‑technical and technical people talked about the same objects:",[21,2458,2459,2462,2465],{},[24,2460,2461],{},"“This bug is in validation.”",[24,2463,2464],{},"“You’ll get the verification point when you confirm it’s fixed.”",[24,2466,2467],{},"“I’ve added a test, can you double‑check the behaviour and close it?”",[12,2469,2470,2471,2474],{},"Instead of “devs versus business”, the game pushed everyone into a shared mental model of ",[27,2472,2473],{},"Hunters, Fixers, and a Judge"," working through the same backlog.",[16,2476,2478],{"id":2477},"why-we-built-a-custom-app-instead-of-a-spreadsheet","Why we built a custom app instead of a spreadsheet",[12,2480,2481],{},"We could have run this in a spreadsheet. But we wanted the game to feel real, not administrative.",[12,2483,2484],{},"So we built:",[21,2486,2487,2493,2499,2502],{},[24,2488,707,2489,2492],{},[27,2490,2491],{},"Nuxt.js leaderboard app"," styled like a real game dashboard.",[24,2494,2495,2498],{},[27,2496,2497],{},"Server‑side endpoints"," to pull bugs and player stats directly from Azure DevOps.",[24,2500,2501],{},"Logic to handle edge cases: judges stealing points, verification rules, states like “Tested, Ready for Prod”, and date windows for valid bugs.",[24,2503,707,2504,2507],{},[27,2505,2506],{},"bug list view"," showing who reported what, who is working on it, and which tags (tests, articles) are attached.",[12,2509,2510],{},"The end result was a live scoreboard that people could refresh during the week and watch themselves move up or down. That matters. It turns abstract work into visible progress.",[16,2512,2514],{"id":2513},"what-well-change-for-the-next-round","What we’ll change for the next round",[12,2516,2517],{},"The first edition worked, but there are things we want to adjust:",[21,2519,2520,2526,2532],{},[24,2521,2522,2525],{},[27,2523,2524],{},"Clearer seasons:"," Fixers could still earn points for older bugs fixed during the week. Next time we may limit both reporting and fixing to a stricter time window.",[24,2527,2528,2531],{},[27,2529,2530],{},"More structured rewards:"," In this round the prize structure was simple. Future games could have categories like “Best Bug Report”, “Most Valuable Test”, or “Most Impactful Fix”.",[24,2533,2534,2537],{},[27,2535,2536],{},"Automatic recognition:"," Right now the game is internal. We’d like to surface anonymized stats in more places, like internal dashboards or company‑wide updates.",[12,2539,2540],{},"The most important learning: once you make quality visible and fun, people naturally lean into it.",[16,2542,1715],{"id":313},[12,2544,2545,2546,2549,2550,2553,2554,2557],{},"The Bug Hunt Game started as an experiment, but it did three things very well: it ",[27,2547,2548],{},"pulled non‑technical people"," into the quality loop in a structured way, it ",[27,2551,2552],{},"rewarded testing and documentation"," instead of just closing tickets, and it ",[27,2555,2556],{},"gave us concrete numbers"," to talk about, bugs found, bugs fixed, tests added, players involved.",[12,2559,2560],{},"At Blits we spend most of our time helping enterprises build production‑grade AI and conversational systems. This little internal game reminded us that sometimes the most powerful improvements in quality don’t come from new tools at all, but from changing how people see the work.",[12,2562,2563],{},"And if you can turn that work into a game with a leaderboard, medals, and a slightly evil Judge?",[12,2565,2566],{},"Even better.",{"title":135,"searchDepth":136,"depth":136,"links":2568},[2569,2570,2571,2572,2577,2578,2579],{"id":2204,"depth":136,"text":2205},{"id":2281,"depth":136,"text":2282},{"id":2340,"depth":136,"text":2341},{"id":2395,"depth":136,"text":2396,"children":2573},[2574,2575,2576],{"id":2402,"depth":1869,"text":2403},{"id":2426,"depth":1869,"text":2427},{"id":2452,"depth":1869,"text":2453},{"id":2477,"depth":136,"text":2478},{"id":2513,"depth":136,"text":2514},{"id":313,"depth":136,"text":1715},"Product & Culture","2026-01-08T00:00:00.000Z","In the first week of January we turned quality assurance into a Bug Hunt Game. Here’s what happened when we put Hunters, Fixers, and a Judge on a shared leaderboard powered by Azure DevOps.","/images/blits-bounty-bug-game.png",{},"/blog/blits-ai-bug-hunt-game-2026-results",{"title":2175,"description":2582},"blog/blits-ai-bug-hunt-game-2026-results",[2589,2590,2591,2592],"quality","devops","culture","experimentation","TfwzhC_nUJkmyPgzd_PLFD-PuL7fSumz60P85RCXA6o",{"id":2595,"title":2596,"author":1986,"body":2597,"category":501,"date":2713,"description":2601,"extension":144,"featured":145,"image":2714,"meta":2715,"navigation":145,"path":2716,"seo":2717,"stem":2718,"tags":2719,"__hash__":2720},"blog/blog/blits-ai-joins-llms4eu-project.md","Blits.ai joins the EU-funded LLMs4EU project",{"type":9,"value":2598,"toc":2705},[2599,2602,2605,2609,2612,2615,2618,2622,2625,2628,2642,2645,2649,2652,2655,2658,2662,2665,2668,2671,2675,2678,2681,2692,2695,2699,2702],[12,2600,2601],{},"Large Language Models are reshaping how organizations interact with information, customers, and services. At the same time, an increasingly relevant question is emerging in Europe: who builds these models, on what data, and under which values and regulations?",[12,2603,2604],{},"That question is central to LLMs4EU. Blits.ai participates in this European Union–funded initiative focused on strengthening Europe's position in large-scale language technology.",[16,2606,2608],{"id":2607},"building-european-foundations-for-large-language-models","Building European foundations for Large Language Models",[12,2610,2611],{},"LLMs4EU (Large Language Models for the European Union) is a multi-year project funded under the EU's Digital Europe Programme. Its objective is clear: to develop a shared European ecosystem for training, fine-tuning, evaluating, and deploying large language models that reflect Europe's linguistic diversity, legal frameworks, and societal context.",[12,2613,2614],{},"A core focus of the project is addressing the structural underrepresentation of many European languages in existing foundation models. LLMs4EU aims to make high-quality language models available across a wide range of EU languages, including those with limited commercial incentives in today's global AI landscape.",[12,2616,2617],{},"The project emphasizes open access to models, datasets, and tooling, enabling European organizations to innovate without structural dependence on non-European AI providers. Compliance with GDPR and alignment with the EU AI Act are built into the project's foundations rather than treated as afterthoughts.",[16,2619,2621],{"id":2620},"why-llms4eu-matters","Why LLMs4EU matters",[12,2623,2624],{},"Europe has strong research institutions, advanced enterprises, and a highly capable SME ecosystem. What has been missing is a coordinated approach to foundational AI models that are both technically competitive and structurally aligned with European regulation and values.",[12,2626,2627],{},"LLMs4EU addresses this gap by:",[21,2629,2630,2633,2636,2639],{},[24,2631,2632],{},"Enabling European companies and public institutions to access advanced language models",[24,2634,2635],{},"Supporting multilingual and domain-specific AI applications across sectors such as public services, telecom, energy, tourism, and research",[24,2637,2638],{},"Promoting open, reusable AI components instead of closed, opaque systems",[24,2640,2641],{},"Strengthening Europe's technological autonomy in a rapidly consolidating global AI market",[12,2643,2644],{},"For enterprises operating in regulated environments, this is not theoretical. It directly affects how safely, transparently, and sustainably AI can be deployed in production.",[16,2646,2648],{"id":2647},"a-strong-european-consortium","A strong European consortium",[12,2650,2651],{},"LLMs4EU brings together a large and diverse consortium of more than sixty partners from across Europe, combining academic research, applied innovation, and industrial deployment expertise.",[12,2653,2654],{},"Notable participants include organizations such as TNO, KPN, leading European research institutes, universities, telecom providers, and technology companies. This breadth ensures that the project is grounded not only in research excellence, but also in real-world applicability and scalability.",[12,2656,2657],{},"The consortium structure reflects a deliberate choice: building European language models is not just a technical challenge, but an ecosystem challenge that requires collaboration across disciplines and sectors.",[16,2659,2661],{"id":2660},"blitsais-role-in-the-project","Blits.ai's role in the project",[12,2663,2664],{},"Blits.ai participates in LLMs4EU as a European SME with a strong focus on applied, enterprise-grade generative AI. Our work centers on deploying LLM-based solutions in complex, regulated environments where multilingual support, reliability, explainability, and compliance are essential.",[12,2666,2667],{},"That applied perspective is critical within a project like LLMs4EU. Training models is only one part of the equation; making them usable, governable, and deployable at scale is where lasting impact is created.",[12,2669,2670],{},"Through our participation, we contribute practical experience from real-world deployments while staying closely connected to the evolving European AI research and infrastructure landscape.",[16,2672,2674],{"id":2673},"alignment-with-blitsais-mission","Alignment with Blits.ai's mission",[12,2676,2677],{},"At Blits.ai, we design and deploy generative AI solutions that are built for production use, multilingual by default, and suitable for organizations operating under strict regulatory constraints.",[12,2679,2680],{},"Participation in LLMs4EU aligns directly with that mission. It reinforces our focus on:",[21,2682,2683,2686,2689],{},[24,2684,2685],{},"language models that work across European cultures and languages,",[24,2687,2688],{},"transparent and auditable AI systems,",[24,2690,2691],{},"and architectures that respect European regulation while remaining globally competitive.",[12,2693,2694],{},"Being part of LLMs4EU strengthens our ability to deliver these principles in practice—both for our clients and within the broader European AI ecosystem.",[16,2696,2698],{"id":2697},"looking-ahead","Looking ahead",[12,2700,2701],{},"LLMs4EU represents a long-term investment in Europe's AI capabilities. Over the coming years, it will influence how language models are developed, shared, and adopted across industries and public institutions.",[12,2703,2704],{},"Blits.ai will continue to contribute from an applied, enterprise perspective, ensuring that the outcomes of this initiative translate into deployable, real-world AI systems that serve European organizations effectively and responsibly.",{"title":135,"searchDepth":136,"depth":136,"links":2706},[2707,2708,2709,2710,2711,2712],{"id":2607,"depth":136,"text":2608},{"id":2620,"depth":136,"text":2621},{"id":2647,"depth":136,"text":2648},{"id":2660,"depth":136,"text":2661},{"id":2673,"depth":136,"text":2674},{"id":2697,"depth":136,"text":2698},"2026-01-05T00:00:00.000Z","/blog/llms4eu-blog-image.png",{},"/blog/blits-ai-joins-llms4eu-project",{"title":2596,"description":2601},"blog/blits-ai-joins-llms4eu-project",[],"EHm_V6a3Tu61DQM4xz8YACTUYtqbOnVrK84yU27PGrY",{"id":2722,"title":2723,"author":7,"body":2724,"category":333,"date":2811,"description":2728,"extension":144,"featured":145,"image":2812,"meta":2813,"navigation":148,"path":2814,"seo":2815,"stem":2816,"tags":340,"__hash__":2817},"blog/blog/business-predictions-chatgpt.md","Business predictions based on my years of experience with ChatGPT and its predecessors",{"type":9,"value":2725,"toc":2806},[2726,2729,2734,2737,2741,2744,2747,2750,2753,2756,2760,2763,2768,2771,2776,2779,2783,2786,2803],[12,2727,2728],{},"Over 5 years ago I started a chatbot (ad)venture called Blits.ai because I believed the way we as humans interact with data will change to a more conversation-based approach. With the current ChatGPT hype (generative AI), more and more people are coming to the same conclusion which is great, but we are not clearly there yet.",[179,2730,2731],{},[12,2732,2733],{},"The way we as humans interact with data will change to a more conversation-based approach - 2017",[12,2735,2736],{},"As someone that has been working with GitHub Co-pilot, and GPT-3 for almost 2 years, including client cases and real-world problems, I would like to share my take on the current state, challenges, and ways it could be beneficial to any business.",[16,2738,2740],{"id":2739},"my-predictions-for-the-year-2030-with-generative-ai","My predictions for the year 2030 with generative AI",[12,2742,2743],{},"You wake up in the morning and try to call your accountant for discussing your next tax income statement. You're not calling for a typical tax question, because most of them are now already correctly answered by Google or Bing. In addition, your start to rely on the financial advice that is sent to you via your accounting software. You now need specialized advice that requires complete oversight and insights into your personal or business life, which still requires some human expertise. Most accountants in the world would like to spend 80% of their time on financial advice, now this might be within reach for a lucky few, but the rest are in decline or taken over by online competitors.",[12,2745,2746],{},"Next, you call your notary, which has just spent 80% of his time not writing but validating the automatically created investment terms for your new investors. His prices are still way too high, but because you still need his sign-off and 'expertise' you'll allow it. The notary market has shifted from a manual business (.docx) to a digital business (.ai). The transition had a high similarity with the shift from physical stocks to digital stocks on Wall Street (quick and dirty). Almost all of the consultancy-driven business models have a more value-based approach and can't sell their expertise of public knowledge that well anymore.",[12,2748,2749],{},"While waiting for a train, you read a news article that is clearly written with the help of generative AI. You feel like you've already read this article 10 times but it's highly efficient and gets the story well told. Most new content is in this format, but it's still in some way audited by humans. AI Auditing tools and human validation becomes popular and you start to value creative content more and more, you might even get a subscription to a newspaper that has a slogan like '100% human.",[12,2751,2752],{},"The major tech companies have their daily troubles with anti-trust laws and privacy. Not due to the rise of services like ChatGPT, but due to the fact the incentive to create new original blogs or websites is in decline. 80% of the questions are now directly answered by models like ChatGPT. This results in the fact that 80% less traffic is going to the actual website that created the original content anymore, lowering the need for new content.",[12,2754,2755],{},"We predict therefore that the frequently occurring questions and assignments can be performed much faster and better by generative AI models like ChatGPT, but the customization of your business still requires the opposite.",[16,2757,2759],{"id":2758},"what-is-the-current-challenge","What is the current challenge?",[12,2761,2762],{},"When companies started the implementation of chatbot technology with the use of intent recognition (the smart part of a chatbot), companies quickly found out it's a lot of work to shape every dialog and build natural language models based on data that company context. The upside is that a company can shape the customer journey from a to z with much detail and precision, something most companies need for their clients or customers.",[179,2764,2765],{},[12,2766,2767],{},"In Chatbot technology it takes a lot of effort and money to make make a chatbot conversation feel like a human one.",[12,2769,2770],{},"ChatGPT works the other way around, it's extremely easy to feel like a human conversation. The downside, however, is that it's not aware of you specific context of your use cases and company specifics. Next to that it only has access to publicly available internet data until 2021 and the model must be heavily regulated by humans at OpenAI on abuse to make sure it doesn't end up racist or promoting your competitor.",[179,2772,2773],{},[12,2774,2775],{},"With chatGPT it takes a lot of effort to shape the conversation towards your company context, data and policies.",[12,2777,2778],{},"There are some methods to feed GPT-3 with your company context like GPT-Index but it is very limited and the result is basically a black box for what it will answer, with a probability of it being wrong. This doesn't mean that it's useless, but I don't expect it will be a Google killer soon, as most people still value reliability over convenience.",[16,2780,2782],{"id":2781},"what-can-it-do-for-my-business","What can it do for my business?",[12,2784,2785],{},"As these large language models are improving over time the quality is increasing every day. The more these AI models are used the better they get. A lot of industries are now at the opportunity to start experimenting with technology to improve their business. Let me give some simple examples you can start today.",[21,2787,2788,2791,2794,2797,2800],{},[24,2789,2790],{},"Content writers, marketers, social media influencers, and creators can create content quickly and reduce the risk of 'writer's block'",[24,2792,2793],{},"Accountants, notaries, lawyers, financial advisors, regulators, and financial institutions can use these services to give a quick answer to available public data about regulation (or even help create contracts, deals, or statements faster)",[24,2795,2796],{},"Local businesses can help customers with quick access to recipes, instructions, manuals, and information about their business",[24,2798,2799],{},"Customer service departments can start implementing a more human-like automated response system due to the high rise of LLM models integrated within conversational AI platforms",[24,2801,2802],{},"Developers can be helped in faster writing code with suggestions and auto-complete",[12,2804,2805],{},"We are currently actively helping companies make the most use of this technology and are in search of major consultancy-based companies that want to be a disruptor in this market.",{"title":135,"searchDepth":136,"depth":136,"links":2807},[2808,2809,2810],{"id":2739,"depth":136,"text":2740},{"id":2758,"depth":136,"text":2759},{"id":2781,"depth":136,"text":2782},"2023-02-20T00:00:00.000Z","/blog/1675857460197.jpeg",{},"/blog/business-predictions-chatgpt",{"title":2723,"description":2728},"blog/business-predictions-chatgpt","0BrzqIISdeLc6xTjEoFlD8DHoG8BPyCopK1YkmgKLdc",{"id":2819,"title":2820,"author":1986,"body":2821,"category":333,"date":2984,"description":2985,"extension":144,"featured":145,"image":2986,"meta":2987,"navigation":148,"path":2988,"seo":2989,"stem":2990,"tags":2991,"__hash__":2994},"blog/blog/eu-ai-act-2026-enterprise-readiness-checklist.md","EU AI Act 2026: The Enterprise AI Readiness Checklist Before August",{"type":9,"value":2822,"toc":2969},[2823,2826,2829,2832,2840,2844,2847,2850,2853,2857,2861,2864,2867,2870,2874,2877,2880,2884,2887,2890,2894,2897,2900,2904,2907,2910,2914,2917,2920,2924,2927,2931,2934,2940,2944,2947,2950,2955,2958,2960,2963,2966],[12,2824,2825],{},"Most teams still discuss the EU AI Act as if there is plenty of time left. There is not.",[12,2827,2828],{},"For enterprise leaders, the question is no longer \"Should we prepare?\" The real question is: \"Do we know exactly what needs to be in place before enforcement starts?\"",[12,2830,2831],{},"In this article I will give you a practical readiness checklist you can use across legal, product, engineering, and operations.",[179,2833,2834],{},[12,2835,2836,2839],{},[27,2837,2838],{},"Key message:"," AI compliance is not a document exercise. It is a production architecture exercise.",[16,2841,2843],{"id":2842},"why-most-organizations-are-behind","Why most organizations are behind",[12,2845,2846],{},"Many companies made one strategic mistake: they isolated compliance into legal review instead of operational design.",[12,2848,2849],{},"That creates three predictable outcomes: controls exist only on paper, decision trails are incomplete, and AI deployment scales faster than risk controls.",[12,2851,2852],{},"If that sounds familiar, you are not alone. But this can still be fixed quickly if you focus on execution.",[16,2854,2856],{"id":2855},"the-enterprise-readiness-checklist","The enterprise readiness checklist",[1784,2858,2860],{"id":2859},"_1-classify-your-ai-use-cases-by-risk-level","1) Classify your AI use cases by risk level",[12,2862,2863],{},"Start with an inventory. Not a spreadsheet for optics, but a living map of every AI use case in production, pilot, and procurement.",[12,2865,2866],{},"For each use case, capture purpose, owner, data categories, potential harm if output is wrong, and likely high-risk classification under the Act.",[12,2868,2869],{},"No inventory means no control.",[1784,2871,2873],{"id":2872},"_2-assign-clear-accountability","2) Assign clear accountability",[12,2875,2876],{},"Every AI system needs an accountable owner with real authority. If ownership is shared vaguely across teams, incidents become unmanageable.",[12,2878,2879],{},"In practice, this means naming one owner for outcome quality, one for technical controls and reliability, and one for regulatory alignment.",[1784,2881,2883],{"id":2882},"_3-implement-logging-and-traceability-by-default","3) Implement logging and traceability by default",[12,2885,2886],{},"You should be able to reconstruct any critical AI decision from input context to model version, tool calls, approvals, and final action.",[12,2888,2889],{},"If you cannot replay the decision path, you cannot defend it.",[1784,2891,2893],{"id":2892},"_4-add-human-oversight-where-impact-is-high","4) Add human oversight where impact is high",[12,2895,2896],{},"Human-in-the-loop is not old-fashioned. It is a risk control.",[12,2898,2899],{},"Use approval gates for high-impact actions such as pricing, legal interpretation, compliance recommendations, and financial instructions.",[1784,2901,2903],{"id":2902},"_5-validate-data-governance-and-quality","5) Validate data governance and quality",[12,2905,2906],{},"Most AI errors are data errors in disguise.",[12,2908,2909],{},"Make sure training, retrieval, and tool data are relevant, current, permissioned correctly, and versioned for audits.",[1784,2911,2913],{"id":2912},"_6-build-technical-documentation-teams-will-actually-maintain","6) Build technical documentation teams will actually maintain",[12,2915,2916],{},"Documentation should not be a one-time PDF. It should be generated from delivery workflows and updated each release.",[12,2918,2919],{},"Capture model behavior assumptions, guardrails, known limits, fallback paths, and escalation flows.",[1784,2921,2923],{"id":2922},"_7-test-for-robustness-accuracy-and-abuse","7) Test for robustness, accuracy, and abuse",[12,2925,2926],{},"Do not rely on benchmark screenshots. Run structured evaluations against adversarial prompts, boundary requests, tool failures, and region-language edge cases taken from real workflows.",[1784,2928,2930],{"id":2929},"_8-register-and-monitor-high-risk-systems","8) Register and monitor high-risk systems",[12,2932,2933],{},"If a use case falls under high-risk obligations, prepare registration, conformity checks, and continuous monitoring early. Waiting until procurement or launch week is expensive.",[1029,2935,2938],{"className":2936,"code":2937,"language":1034,"meta":135},[1032],"Minimal readiness artifact set:\n- AI use-case register with risk tier\n- Control ownership matrix\n- Decision trace logging specification\n- Evaluation suite with pass/fail gates\n- Remediation backlog with deadlines\n",[1036,2939,2937],{"__ignoreMap":135},[16,2941,2943],{"id":2942},"a-30-day-execution-plan","A 30-day execution plan",[12,2945,2946],{},"If you want momentum, run this sequence:",[12,2948,2949],{},"Week 1 should focus on inventory and risk classification. Week 2 should lock ownership and control design. Week 3 is where logging, oversight, and evaluations become operational. Week 4 closes the loop with documentation, governance review, and a remediation backlog.",[179,2951,2952],{},[12,2953,2954],{},"\"Compliance velocity comes from operational clarity, not from larger policy documents.\"",[12,2956,2957],{},"The goal is not perfect governance in 30 days. The goal is control that is real, visible, and scalable.",[16,2959,1041],{"id":1040},[12,2961,2962],{},"The winners in enterprise AI will not be the teams with the most demos. They will be the teams that can prove control while shipping fast.",[12,2964,2965],{},"Compliance is not the opposite of innovation.",[12,2967,2968],{},"In 2026, it is the condition for durable innovation.",{"title":135,"searchDepth":136,"depth":136,"links":2970},[2971,2972,2982,2983],{"id":2842,"depth":136,"text":2843},{"id":2855,"depth":136,"text":2856,"children":2973},[2974,2975,2976,2977,2978,2979,2980,2981],{"id":2859,"depth":1869,"text":2860},{"id":2872,"depth":1869,"text":2873},{"id":2882,"depth":1869,"text":2883},{"id":2892,"depth":1869,"text":2893},{"id":2902,"depth":1869,"text":2903},{"id":2912,"depth":1869,"text":2913},{"id":2922,"depth":1869,"text":2923},{"id":2929,"depth":1869,"text":2930},{"id":2942,"depth":136,"text":2943},{"id":1040,"depth":136,"text":1041},"2026-03-11T00:00:00.000Z","Most AI teams are still treating EU AI Act compliance as a legal side project. It is not. This checklist shows what enterprises should implement now to avoid operational disruption and expensive rework.","/images/blog-eu-ai-act.png",{},"/blog/eu-ai-act-2026-enterprise-readiness-checklist",{"title":2820,"description":2985},"blog/eu-ai-act-2026-enterprise-readiness-checklist",[2992,1101,1981,2993],"eu ai act","compliance","54BHvlYkTPZCxKSdGs1_TE-Mee_n2pMxtJGrISfKe1o",{"id":2996,"title":2997,"author":7,"body":2998,"category":333,"date":3273,"description":3002,"extension":144,"featured":145,"image":3274,"meta":3275,"navigation":148,"path":3276,"seo":3277,"stem":3278,"tags":340,"__hash__":3279},"blog/blog/feeding-llms-without-leaking-secrets.md","Feeding LLMs Without Leaking Secrets: A Guide for Companies On How To Add Your Company Data",{"type":9,"value":2999,"toc":3259},[3000,3003,3006,3009,3013,3016,3019,3022,3025,3049,3052,3056,3059,3062,3065,3068,3071,3075,3078,3081,3084,3087,3091,3094,3097,3100,3103,3106,3110,3113,3116,3142,3145,3148,3152,3155,3158,3161,3164,3168,3171,3174,3180,3186,3192,3198,3201,3205,3208,3211,3214,3218,3221,3225,3251,3253,3256],[12,3001,3002],{},"In my previous article 9 Things I Really Hate About AI, I mentioned how everyone suddenly seems to be an AI expert and how that creates a lot of noise. In the next few posts, I'll break down some key AI concepts specifically for business professionals. Why? Because even if you're actively looking for information, much of what's out there is either inaccurate or way too technical for the average manager to make sense of.",[12,3004,3005],{},"My goal is to make these complex topics understandable for anyone who needs to make smarter business decisions. This week, I'm kicking things off with one of the most important ones: how (and why) you should add your company's data to large language models.",[12,3007,3008],{},"Let's start with the underlying question:",[16,3010,3012],{"id":3011},"why-should-you-add-company-data-to-a-large-language-model","Why should you add (company) data to a large language model?",[12,3014,3015],{},"Your data transforms a general LLM into a powerful tool that understands your specific needs and context, ultimately leading to better insights, automation, and competitive advantages.",[12,3017,3018],{},"Imagine a general LLM is like a very smart person who knows a lot about the world from reading countless books and articles. However, they don't know anything specific about your business. Adding your data is like giving that smart person your company's internal documents, customer conversations, product information, and industry-specific reports. This focused information allows the LLM to understand your unique context and provide much more valuable results.",[12,3020,3021],{},"Whether you've added context to a question or uploaded a document to tools like ChatGPT, you've already experienced the power of providing data to these models. This personal approach works well for individual use. However, when building solutions for your customers or employees, a more robust strategy for integrating your company's data is essential. Let's explore the various methods currently being used to effectively feed your data to large language models at scale.",[12,3023,3024],{},"The currently most used methods are:",[3026,3027,3028,3031,3034,3037,3040,3043,3046],"ol",{},[24,3029,3030],{},"Adding information directly to your question",[24,3032,3033],{},"Adding a file to the model (uploading a document)",[24,3035,3036],{},"Function calling (connecting an API)",[24,3038,3039],{},"Retrieval Augmented Generation (RAG)",[24,3041,3042],{},"Cache Augmented Generation (CAG)",[24,3044,3045],{},"Fine-tuning an existing model (create your own LLM)",[24,3047,3048],{},"Training a new foundational model (compete with OpenAI, Anthropic, etc.)",[12,3050,3051],{},"These are ranked from easy to hard. Let's break them down.",[16,3053,3055],{"id":3054},"_1-adding-information-directly-to-your-question","1. Adding information directly to your question",[12,3057,3058],{},"Most people do this: You ask the model your question, followed by extra context (or background) to help the model answer better. This works. It's also quick and dirty. If you need something done fast and you're not worried about security or long-term reuse, it's fine.",[12,3060,3061],{},"But this method has limits. You can only add so much information before the model starts to ignore parts of it. The longer your prompt, the more the model will focus on what came last. So, if you throw in 20 pages of data and ask a question at the end, chances are the first few pages will be ignored.",[12,3063,3064],{},"Also, you're sending potentially sensitive data to a third party (yes, even if they say they don't store it). So this method is fine for brainstorming or playing around, but you don't want your board reports or customer data in here.",[12,3066,3067],{},"These methods (and the next three) are all limited by the model's context window. That's the maximum number of tokens (think of tokens as chunks of words or characters) the model can process in a single request. Depending on which model you're using, that window can be quite small, which means you can't simply dump all your company data in at once.",[12,3069,3070],{},"On top of that, you're charged per token. So every time you send a large prompt, you're paying more. If you try to scale this up across hundreds or thousands of requests, it quickly becomes expensive and inefficient. That's why methods like RAG and CAG exist.",[16,3072,3074],{"id":3073},"_2-adding-a-file-to-the-model-uploading-a-document","2. Adding a file to the model (uploading a document)",[12,3076,3077],{},"This is the method most people try after prompt injection. You upload a document, like a PDF, a policy document, or a user manual and then ask the model questions about it. Tools like ChatGPT (with Pro or Enterprise plans) and Claude make this easy. You upload the document in the chat interface, and the model appears to \"read\" it and answer your questions.",[12,3079,3080],{},"But here's the catch: LLMs don't actually \"read\" documents like humans do. Instead, they break the text into chunks (usually 200–500 words at a time), embed those chunks into a vector format, and then retrieve the most relevant ones when you ask a question. This is often invisible to you, but it's happening behind the scenes.",[12,3082,3083],{},"This feels safer, but it's not. The same risks as above apply. If you're using ChatGPT or any third-party tool, your data goes through their servers. Unless you pay for enterprise-level privacy controls (and read the fine print), this is not where confidential company documents belong.",[12,3085,3086],{},"This method is perfect for quick document review or summarizing files. But if you want to build a company-wide solution (like a smart assistant or internal knowledge bot), you'll need something more robust, Function calling or RAG.",[16,3088,3090],{"id":3089},"_3-function-calling-connecting-an-api","3. Function Calling (Connecting an API)",[12,3092,3093],{},"Function calling is one of the most promising recent developments in LLMs. Instead of trying to make the model guess everything from natural language, you give it structured access to your systems and tools. That means the model doesn't just answer questions, but it can trigger real actions.",[12,3095,3096],{},"Think of it like this: the LLM becomes the brain, and your APIs become the hands. You describe what functions are available (like \"get customer order history\" or \"calculate monthly revenue\"), and the model learns when and how to call them.",[12,3098,3099],{},"You don't need to train the model to know your backend logic. Instead, you define the interface. Then, when someone asks: \"What's the weather in New York City?\" the LLM knows it should trigger the a weather function, to get the temperature for that location. This same logic applies if you connect your company's systems.",[12,3101,3102],{},"OpenAI's GPT, Google's Gemini, and Anthropic's Claude models support function calling out of the box. You can use it to connect the LLM with your CRM, ERP, or support systems. Microsoft Copilot uses similar techniques to integrate with Excel, Outlook, and Teams.",[12,3104,3105],{},"Doing this with open-source models requires a lot more engineering, but can be done with popular meta-frameworks like LangChain or LlamaIndex.",[16,3107,3109],{"id":3108},"_4-retrieval-augmented-generation-rag","4. Retrieval Augmented Generation (RAG)",[12,3111,3112],{},"RAG is the industry standard for adding company data to LLMs at scale. It's how most enterprise AI systems are built today.",[12,3114,3115],{},"Here's how it works:",[3026,3117,3118,3121,3124,3127,3130,3133,3136,3139],{},[24,3119,3120],{},"You take your company's documents, policies, manuals, and data",[24,3122,3123],{},"You break them into chunks (typically 200–500 words)",[24,3125,3126],{},"Each chunk gets converted into a vector embedding (a numerical representation)",[24,3128,3129],{},"These vectors are stored in a vector database (like Pinecone, Weaviate, or Qdrant)",[24,3131,3132],{},"When a user asks a question, their question also gets converted to a vector",[24,3134,3135],{},"The system searches the database for the most similar vectors",[24,3137,3138],{},"Those relevant chunks get fed to the LLM as context",[24,3140,3141],{},"The LLM generates an answer based on that context",[12,3143,3144],{},"The beauty of RAG is that it keeps your data separate from the model. You're not training anything. You're just giving the model real-time access to the information it needs. This means you can update your data without retraining, and you maintain full control.",[12,3146,3147],{},"RAG is what powers most enterprise chatbots, document Q&A systems, and internal knowledge assistants. It's secure (you control the data), scalable (you can add millions of documents), and relatively affordable.",[16,3149,3151],{"id":3150},"_5-cache-augmented-generation-cag","5. Cache Augmented Generation (CAG)",[12,3153,3154],{},"CAG is newer and less common, but it's gaining traction. The idea is to cache frequently used prompts and responses so the model doesn't have to regenerate the same answer repeatedly.",[12,3156,3157],{},"Imagine your support team gets asked the same 50 questions every day. Instead of running those questions through the full LLM pipeline each time, you cache the answers. When a similar question comes in, you serve the cached response instantly.",[12,3159,3160],{},"This reduces costs (fewer API calls), improves speed (no generation time), and ensures consistency (same answer every time). But it requires careful cache management and invalidation strategies.",[12,3162,3163],{},"Some providers like Anthropic have started offering prompt caching as a built-in feature, making this easier to implement.",[16,3165,3167],{"id":3166},"_6-fine-tuning-an-existing-model","6. Fine-tuning an existing model",[12,3169,3170],{},"Fine-tuning means taking a pre-trained model (like GPT-4, Llama, or Mistral) and training it further on your specific data. This is more involved than RAG, but it can produce better results for specialized tasks.",[12,3172,3173],{},"There are different levels of fine-tuning:",[12,3175,3176,3179],{},[27,3177,3178],{},"a) Light fine-tuning (LoRA, QLoRA):"," You don't retrain the entire model. Instead, you add small adapter layers that learn your specific patterns. This is much cheaper and faster.",[12,3181,3182,3185],{},[27,3183,3184],{},"b) Full fine-tuning:"," You retrain the entire model on your data. This requires significant compute power (think dozens of GPUs) and expertise. Most companies don't need this.",[12,3187,3188,3191],{},[27,3189,3190],{},"c) Instruction tuning:"," You fine-tune the model to follow specific instruction formats or domain-specific patterns. This is common for customer service bots or internal tools.",[12,3193,3194,3197],{},[27,3195,3196],{},"d) Deep customization:"," This is close to building your own model. You start with a base model checkpoint (like LLaMA, Mistral, or DeepSeek) and train it further on massive datasets, potentially hundreds of millions of tokens or more.",[12,3199,3200],{},"At this point, you're creating your own model variant. You need serious MLOps. Evaluation pipelines. Guardrails. This is what AI-native companies do. It's powerful, but probably not what your company needs, unless AI is your product.",[16,3202,3204],{"id":3203},"_7-training-a-new-foundational-model","7. Training a new foundational model",[12,3206,3207],{},"Unless you are OpenAI, Google, Mistral, Meta or Anthropic, just don't. This costs tens (or hundreds) of millions. It requires large GPU infrastructure, research, and talent that most companies don't have. OpenAI pays its AI engineers more than some companies pay their CEOs.",[12,3209,3210],{},"It also demands vast amounts of data, which isn't easily accessible without significant resources or a large budget. Bloomberg tackled this challenge by developing its foundational model: BloombergGPT. To train it, they compiled a dataset of 363 billion finance-specific tokens from their proprietary database, along with an additional 345 billion general-purpose tokens from public online sources such as Wikipedia.",[12,3212,3213],{},"Some companies say they've built their own models. Most haven't. They've fine-tuned open ones. Which is fine. But let's not confuse that with building from scratch.",[16,3215,3217],{"id":3216},"breakdown-of-these-methods-and-when-to-use-them","Breakdown of these methods and when to use them",[12,3219,3220],{},"It's not easy to decide what you need for your use case, but here's a quick comparison:",[1784,3222,3224],{"id":3223},"quick-takeaways","Quick Takeaways:",[21,3226,3227,3233,3239,3245],{},[24,3228,3229,3232],{},[27,3230,3231],{},"If you want speed and low cost,"," go with prompt injections or document uploads, but accept low safety and limited quality.",[24,3234,3235,3238],{},[27,3236,3237],{},"If you want enterprise-grade quality and safety,"," start with RAG and function calling.",[24,3240,3241,3244],{},[27,3242,3243],{},"If you're AI-native or working in a specialized domain,"," consider fine-tuning.",[24,3246,3247,3250],{},[27,3248,3249],{},"If you're not OpenAI, DeepMind, or Meta,"," avoid creating your own model.",[16,3252,314],{"id":313},[12,3254,3255],{},"If you want any more info on this, let me know in the comments, or just keep following my regular tech updates. I try to break down complex topics like this in a way that's actually useful, especially for business folks trying to make sense of all the AI noise.",[12,3257,3258],{},"My next article will dive into image and video generation: how it works, what's possible today, and what's just hype.",{"title":135,"searchDepth":136,"depth":136,"links":3260},[3261,3262,3263,3264,3265,3266,3267,3268,3269,3272],{"id":3011,"depth":136,"text":3012},{"id":3054,"depth":136,"text":3055},{"id":3073,"depth":136,"text":3074},{"id":3089,"depth":136,"text":3090},{"id":3108,"depth":136,"text":3109},{"id":3150,"depth":136,"text":3151},{"id":3166,"depth":136,"text":3167},{"id":3203,"depth":136,"text":3204},{"id":3216,"depth":136,"text":3217,"children":3270},[3271],{"id":3223,"depth":1869,"text":3224},{"id":313,"depth":136,"text":314},"2025-05-19T00:00:00.000Z","/blog/1747588091563.png",{},"/blog/feeding-llms-without-leaking-secrets",{"title":2997,"description":3002},"blog/feeding-llms-without-leaking-secrets","MYvZaiMAA4ObsxQv16m7pfWq2Ou1LS1n8_8IR8Q6GsQ",{"id":3281,"title":3282,"author":1986,"body":3283,"category":333,"date":3423,"description":3424,"extension":144,"featured":145,"image":3425,"meta":3426,"navigation":148,"path":3427,"seo":3428,"stem":3429,"tags":3430,"__hash__":3433},"blog/blog/from-single-agent-to-multi-agent-systems-when-to-split-roles.md","From Single Agent to Multi-Agent Systems: When to Split Roles",{"type":9,"value":3284,"toc":3414},[3285,3288,3291,3294,3297,3301,3304,3307,3311,3314,3317,3321,3324,3327,3331,3334,3337,3340,3344,3347,3351,3354,3357,3360,3397,3399,3402,3405,3408,3411],[12,3286,3287],{},"The fastest way to overcomplicate agentic AI is to start with five agents when one would do.",[12,3289,3290],{},"The fastest way to break a scaled workflow is to keep one agent doing everything.",[12,3292,3293],{},"So when should you split roles?",[12,3295,3296],{},"This is one of the most important architecture decisions teams face in 2026.",[16,3298,3300],{"id":3299},"start-with-one-agent-unless-complexity-forces-a-split","Start with one agent unless complexity forces a split",[12,3302,3303],{},"A single-agent setup is usually enough while task scope is narrow, tool access is limited, and decision logic remains simple. In that phase, one agent gives you lower cost, faster debugging, and less coordination overhead.",[12,3305,3306],{},"But once workflows become cross-functional, one agent starts to fail in predictable ways.",[16,3308,3310],{"id":3309},"signals-that-it-is-time-to-move-to-multi-agent","Signals that it is time to move to multi-agent",[12,3312,3313],{},"You are usually ready to split when one agent sees too many tools, prompts become bloated and contradictory, accountability gets blurry, and latency starts drifting under real load. Another clear signal is risk mixing, where low-risk and high-risk decisions happen in the same loop without separable controls.",[12,3315,3316],{},"If three or more are true, role separation usually improves outcomes.",[16,3318,3320],{"id":3319},"the-role-pattern-that-works-in-enterprise-environments","The role pattern that works in enterprise environments",[12,3322,3323],{},"A practical baseline is planner, executor, reviewer. The planner interprets goals and decomposes work, executors perform operational steps through tools, and the reviewer validates quality and policy alignment before high-impact outputs are released.",[12,3325,3326],{},"This pattern is useful because planning quality becomes measurable, execution becomes controllable, and validation becomes auditable.",[16,3328,3330],{"id":3329},"governance-improves-when-roles-are-explicit","Governance improves when roles are explicit",[12,3332,3333],{},"Role separation is not just a technical trick. It is a governance advantage.",[12,3335,3336],{},"Once roles are explicit, controls can also be explicit: planners can have broad read access but no critical write rights, executors can be scoped to limited tool actions, and reviewers can own escalation paths and policy checks.",[12,3338,3339],{},"This design makes oversight practical, especially in regulated workflows.",[16,3341,3343],{"id":3342},"common-multi-agent-mistakes","Common multi-agent mistakes",[12,3345,3346],{},"Common mistakes repeat across teams: role separation without clear contracts, broad tool permissions for every agent, no shared memory strategy, and no arbitration mechanism when outputs conflict. Multi-agent is not \"more intelligence by default.\" It is more coordination, and coordination needs architecture.",[16,3348,3350],{"id":3349},"a-simple-migration-approach","A simple migration approach",[12,3352,3353],{},"Do not rewrite everything at once.",[12,3355,3356],{},"Start with your current single-agent flow. Extract validation into a reviewer role first, then introduce a planner only for complex categories, and add specialist executors where tool misuse appears repeatedly.",[12,3358,3359],{},"This gives you control without creating a distributed mess.",[1029,3361,3365],{"className":3362,"code":3363,"language":3364,"meta":135,"style":135},"language-mermaid shiki shiki-themes material-theme-lighter material-theme material-theme-palenight","flowchart LR\n  G[Goal] --> P[Planner]\n  P --> E[Executor]\n  E --> R[Reviewer]\n  R -->|approved| O[Output]\n  R -->|escalate| H[Human]\n","mermaid",[1036,3366,3367,3372,3377,3382,3387,3392],{"__ignoreMap":135},[1837,3368,3369],{"class":1839,"line":1840},[1837,3370,3371],{},"flowchart LR\n",[1837,3373,3374],{"class":1839,"line":136},[1837,3375,3376],{},"  G[Goal] --> P[Planner]\n",[1837,3378,3379],{"class":1839,"line":1869},[1837,3380,3381],{},"  P --> E[Executor]\n",[1837,3383,3384],{"class":1839,"line":1881},[1837,3385,3386],{},"  E --> R[Reviewer]\n",[1837,3388,3389],{"class":1839,"line":1892},[1837,3390,3391],{},"  R -->|approved| O[Output]\n",[1837,3393,3394],{"class":1839,"line":1924},[1837,3395,3396],{},"  R -->|escalate| H[Human]\n",[16,3398,1041],{"id":1040},[12,3400,3401],{},"The right question is not \"single agent or multi-agent?\"",[12,3403,3404],{},"The right question is: \"Where does specialization create measurable value and lower risk?\"",[12,3406,3407],{},"If you cannot answer that with metrics, stay simple.",[12,3409,3410],{},"If you can, split deliberately and design for traceable coordination.",[1956,3412,3413],{},"html .light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html.light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}html.dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}",{"title":135,"searchDepth":136,"depth":136,"links":3415},[3416,3417,3418,3419,3420,3421,3422],{"id":3299,"depth":136,"text":3300},{"id":3309,"depth":136,"text":3310},{"id":3319,"depth":136,"text":3320},{"id":3329,"depth":136,"text":3330},{"id":3342,"depth":136,"text":3343},{"id":3349,"depth":136,"text":3350},{"id":1040,"depth":136,"text":1041},"2026-03-19T00:00:00.000Z","Many teams move to multi-agent setups too early or too late. This article explains when a single agent is enough, when to split roles, and how to avoid coordination chaos.","/images/blog-multi-agent.png",{},"/blog/from-single-agent-to-multi-agent-systems-when-to-split-roles",{"title":3282,"description":3424},"blog/from-single-agent-to-multi-agent-systems-when-to-split-roles",[1097,3431,1101,3432],"multi agent systems","ai architecture","v65C9toLFt0bBB1-B1j3ElUFOK3r7eUfQZQZvmXX6Hw",{"id":3435,"title":3436,"author":1986,"body":3437,"category":333,"date":3488,"description":3441,"extension":144,"featured":145,"image":3489,"meta":3490,"navigation":148,"path":3491,"seo":3492,"stem":3493,"tags":340,"__hash__":3494},"blog/blog/gpt-4-in-chatbots.md","GPT-4 in Chatbots: Blits.ai Breaks New Ground in Conversational AI",{"type":9,"value":3438,"toc":3483},[3439,3442,3446,3449,3466,3470,3473,3477,3480],[12,3440,3441],{},"We're excited to announce a major milestone in conversational AI technology. Blits.ai now supports GPT-4, bringing unprecedented capabilities to chatbot development and deployment.",[16,3443,3445],{"id":3444},"what-this-means-for-your-business","What This Means for Your Business",[12,3447,3448],{},"GPT-4 integration opens up new possibilities for enterprise-grade conversational AI solutions. Our platform now offers:",[21,3450,3451,3454,3457,3460,3463],{},[24,3452,3453],{},"Enhanced natural language understanding with improved context awareness",[24,3455,3456],{},"More accurate response generation across complex scenarios",[24,3458,3459],{},"Better handling of nuanced conversations and multi-turn dialogues",[24,3461,3462],{},"Improved performance on specialized domain knowledge",[24,3464,3465],{},"Advanced reasoning capabilities for complex problem-solving",[16,3467,3469],{"id":3468},"technical-implementation","Technical Implementation",[12,3471,3472],{},"The integration has been carefully designed to maintain our enterprise-grade security and compliance standards while delivering the enhanced capabilities that GPT-4 provides. Our team has worked extensively to ensure seamless integration with existing workflows.",[16,3474,3476],{"id":3475},"performance-improvements","Performance Improvements",[12,3478,3479],{},"Early testing shows significant improvements in conversation quality, with users reporting more natural and helpful interactions. The enhanced model capabilities allow for more sophisticated use cases across various industries.",[12,3481,3482],{},"This advancement represents our commitment to staying at the forefront of AI technology while ensuring our clients have access to the most powerful conversational AI tools available.",{"title":135,"searchDepth":136,"depth":136,"links":3484},[3485,3486,3487],{"id":3444,"depth":136,"text":3445},{"id":3468,"depth":136,"text":3469},{"id":3475,"depth":136,"text":3476},"2023-07-14T00:00:00.000Z","/blog/gpt-4.webp",{},"/blog/gpt-4-in-chatbots",{"title":3436,"description":3441},"blog/gpt-4-in-chatbots","XJMF26QFeB3RkNa1bIZEDNHq5NagD5VBOdo4z_kC4FU",{"id":3496,"title":3497,"author":1986,"body":3498,"category":333,"date":3565,"description":3502,"extension":144,"featured":145,"image":3566,"meta":3567,"navigation":148,"path":3568,"seo":3569,"stem":3570,"tags":340,"__hash__":3571},"blog/blog/how-practical-is-gpt-3-for-conversational-ai-chatbots.md","How practical is GPT-3 for Conversational AI chatbots?",{"type":9,"value":3499,"toc":3559},[3500,3503,3507,3510,3524,3528,3531,3545,3549,3552,3556],[12,3501,3502],{},"GPT-3 has revolutionized the field of conversational AI, but understanding its practical applications and limitations is crucial for successful implementation.",[16,3504,3506],{"id":3505},"gpt-3-capabilities","GPT-3 Capabilities",[12,3508,3509],{},"GPT-3 demonstrates remarkable abilities in natural language understanding and generation, making it highly suitable for conversational AI applications.",[21,3511,3512,3515,3518,3521],{},[24,3513,3514],{},"Natural conversation flow and context awareness",[24,3516,3517],{},"Ability to handle diverse topics and domains",[24,3519,3520],{},"Improved response quality compared to traditional rule-based systems",[24,3522,3523],{},"Scalability for high-volume interactions",[16,3525,3527],{"id":3526},"practical-considerations","Practical Considerations",[12,3529,3530],{},"While GPT-3 offers significant advantages, there are important practical considerations for enterprise deployment:",[21,3532,3533,3536,3539,3542],{},[24,3534,3535],{},"Cost implications for high-volume usage",[24,3537,3538],{},"Latency considerations for real-time applications",[24,3540,3541],{},"Content moderation and safety requirements",[24,3543,3544],{},"Integration complexity with existing systems",[16,3546,3548],{"id":3547},"best-practices","Best Practices",[12,3550,3551],{},"Successful GPT-3 implementation requires careful planning around prompt engineering, context management, and user experience design. Our experience shows that the most effective deployments combine GPT-3's capabilities with domain-specific knowledge and business logic.",[16,3553,3555],{"id":3554},"conclusion","Conclusion",[12,3557,3558],{},"GPT-3 represents a significant advancement in conversational AI, but successful implementation requires understanding both its capabilities and practical limitations.",{"title":135,"searchDepth":136,"depth":136,"links":3560},[3561,3562,3563,3564],{"id":3505,"depth":136,"text":3506},{"id":3526,"depth":136,"text":3527},{"id":3547,"depth":136,"text":3548},{"id":3554,"depth":136,"text":3555},"2022-01-25T00:00:00.000Z","/blog/GPT-3.jpg",{},"/blog/how-practical-is-gpt-3-for-conversational-ai-chatbots",{"title":3497,"description":3502},"blog/how-practical-is-gpt-3-for-conversational-ai-chatbots","5IGl_eAi99JgCyEsWQlfWRsE4qqC7R2W3rkxVlGLQb0",{"id":3573,"title":3574,"author":7,"body":3575,"category":333,"date":3699,"description":3700,"extension":144,"featured":145,"image":3701,"meta":3702,"navigation":148,"path":3703,"seo":3704,"stem":3705,"tags":3706,"__hash__":3709},"blog/blog/how-to-build-an-ai-control-tower-for-agentic-operations.md","How to Build an AI Control Tower for Agentic Operations",{"type":9,"value":3576,"toc":3687},[3577,3580,3583,3586,3589,3591,3595,3598,3601,3604,3608,3612,3615,3618,3622,3625,3628,3633,3637,3640,3643,3647,3650,3653,3657,3660,3663,3667,3670,3676,3678,3681,3684],[12,3578,3579],{},"Most teams have AI dashboards.",[12,3581,3582],{},"Very few have an AI control tower.",[12,3584,3585],{},"A dashboard shows activity. A control tower shows control.",[12,3587,3588],{},"When agents are running real workflows, you need to see not only what happened, but why it happened, where it failed, and who approved what.",[12,3590,1766],{},[16,3592,3594],{"id":3593},"why-classic-monitoring-is-no-longer-enough","Why classic monitoring is no longer enough",[12,3596,3597],{},"Traditional monitoring tracks uptime, latency, and errors. That still matters, but agentic systems add another layer:",[12,3599,3600],{},"reasoning quality, tool selection behavior, policy compliance, and human intervention events.",[12,3602,3603],{},"Without this visibility, organizations scale blind.",[16,3605,3607],{"id":3606},"what-an-ai-control-tower-should-include","What an AI control tower should include",[1784,3609,3611],{"id":3610},"_1-end-to-end-traces-for-every-workflow","1) End-to-end traces for every workflow",[12,3613,3614],{},"Trace each run from request to final action, including context state, model version, retrieval and tool calls, validation outcomes, and approval events.",[12,3616,3617],{},"This is your operational truth.",[1784,3619,3621],{"id":3620},"_2-decision-level-observability","2) Decision-level observability",[12,3623,3624],{},"You need signals on why an agent selected a path, not just whether the API returned 200.",[12,3626,3627],{},"Track branch changes, retries, confidence drops, and refusal rates as leading indicators of instability.",[179,3629,3630],{},[12,3631,3632],{},"\"If you only measure system health, you miss decision health.\"",[1784,3634,3636],{"id":3635},"_3-risk-aware-alerting","3) Risk-aware alerting",[12,3638,3639],{},"Not every failure deserves the same urgency.",[12,3641,3642],{},"Create alerting tiers by business impact so operations can prioritize real risk: financial and compliance incidents first, then repeated quality degradation, then latency/tool drift, and finally low-impact fallback noise.",[1784,3644,3646],{"id":3645},"_4-human-override-console","4) Human-override console",[12,3648,3649],{},"When risk rises, teams need to pause workflows, block high-risk tools, switch to mandatory approvals, and reroute to safe fallbacks without waiting for engineering releases.",[12,3651,3652],{},"If intervention requires an engineering deploy, your control model is too slow.",[1784,3654,3656],{"id":3655},"_5-kpi-layer-that-links-technical-and-business-outcomes","5) KPI layer that links technical and business outcomes",[12,3658,3659],{},"A control tower should combine technical KPIs (P95/P99, tool errors, rollback rate), quality KPIs (groundedness and policy pass rate), and business KPIs (containment, resolution time, conversion, and cost-to-serve) in one operational surface.",[12,3661,3662],{},"This is where AI operations becomes business operations.",[16,3664,3666],{"id":3665},"a-phased-rollout-model","A phased rollout model",[12,3668,3669],{},"Phase 1 should establish tracing and centralized logs. Phase 2 adds policy checks, risk scoring, and alerting. Phase 3 introduces override workflows and executive reporting. Start with visibility, move to control, then optimize.",[1029,3671,3674],{"className":3672,"code":3673,"language":1034,"meta":135},[1032],"Control Tower Maturity Rule:\nNo autonomous scale-up until traceability, intervention, and risk alerting are all live.\n",[1036,3675,3673],{"__ignoreMap":135},[16,3677,1041],{"id":1040},[12,3679,3680],{},"Agentic AI without a control tower is automated risk.",[12,3682,3683],{},"If leaders cannot answer \"What happened, why, and with what impact?\" in minutes, the system is not production mature.",[12,3685,3686],{},"Control is not friction. It is what allows autonomous systems to run safely at scale.",{"title":135,"searchDepth":136,"depth":136,"links":3688},[3689,3690,3697,3698],{"id":3593,"depth":136,"text":3594},{"id":3606,"depth":136,"text":3607,"children":3691},[3692,3693,3694,3695,3696],{"id":3610,"depth":1869,"text":3611},{"id":3620,"depth":1869,"text":3621},{"id":3635,"depth":1869,"text":3636},{"id":3645,"depth":1869,"text":3646},{"id":3655,"depth":1869,"text":3656},{"id":3665,"depth":136,"text":3666},{"id":1040,"depth":136,"text":1041},"2026-03-23T00:00:00.000Z","Once AI agents start running business workflows, dashboards are not enough. This article explains how to build an AI control tower with observability, alerts, and decision-level traceability.","/images/blog-control-tower.png",{},"/blog/how-to-build-an-ai-control-tower-for-agentic-operations",{"title":3574,"description":3700},"blog/how-to-build-an-ai-control-tower-for-agentic-operations",[3707,1097,1101,3708],"ai observability","ai operations","K3DUqmzayhwsfJuItIuMCO-y1Lyi_90XqwQgjXYefLA",{"id":3711,"title":3712,"author":1986,"body":3713,"category":333,"date":3969,"description":3970,"extension":144,"featured":145,"image":3971,"meta":3972,"navigation":148,"path":3973,"seo":3974,"stem":3975,"tags":3976,"__hash__":3980},"blog/blog/mcp-security-in-production-auth-and-least-privilege.md","MCP Security in Production: Auth, Least Privilege, and Audit Logs",{"type":9,"value":3714,"toc":3956},[3715,3718,3721,3724,3728,3731,3734,3737,3741,3745,3748,3751,3755,3758,3761,3764,3768,3771,3774,3777,3781,3784,3787,3791,3794,3900,3904,3907,3927,3930,3934,3937,3942,3944,3947,3950,3953],[12,3716,3717],{},"MCP solved one big problem: standardizing how models use tools. Now enterprises face the next one: securing that tool layer in production.",[12,3719,3720],{},"Most teams can add an MCP server in a day. Fewer can prove that the right tool was called by the right agent with the right permissions at the right time.",[12,3722,3723],{},"That gap is where risk lives.",[16,3725,3727],{"id":3726},"why-mcp-security-is-now-a-board-level-topic","Why MCP security is now a board-level topic",[12,3729,3730],{},"When an agent can call business tools, the threat model changes immediately.",[12,3732,3733],{},"You are no longer securing only model output. You are securing actions: reading data, triggering transactions, changing records, and executing business workflows.",[12,3735,3736],{},"So the core question becomes simple: can your system enforce and prove delegated authority?",[16,3738,3740],{"id":3739},"the-production-security-baseline","The production security baseline",[1784,3742,3744],{"id":3743},"_1-enforce-strong-authentication-for-every-mcp-endpoint","1) Enforce strong authentication for every MCP endpoint",[12,3746,3747],{},"No anonymous tool access. Use enterprise-grade authentication between clients and MCP servers, and rotate credentials on a predictable schedule.",[12,3749,3750],{},"At minimum, this includes mutual service authentication, expiring credentials, and a reliable revocation path for compromised tokens.",[1784,3752,3754],{"id":3753},"_2-apply-least-privilege-scopes-by-agent-role","2) Apply least-privilege scopes by agent role",[12,3756,3757],{},"Most teams overexpose tools during pilot phase and forget to tighten access later.",[12,3759,3760],{},"Do not grant \"all tools\" to generic agents. A research role should remain read-only, operational roles should get scoped update permissions, and financial roles should sit behind additional approvals.",[12,3762,3763],{},"Permissions should be explicit, minimal, and time-bounded where possible.",[1784,3765,3767],{"id":3766},"_3-add-policy-checks-before-tool-execution","3) Add policy checks before tool execution",[12,3769,3770],{},"Validation should happen before the MCP call is executed, not after.",[12,3772,3773],{},"Check caller identity, tool-level permissions, allowed parameter ranges, and contextual constraints such as region, tenant, and policy state.",[12,3775,3776],{},"This prevents valid tools from being used in invalid contexts.",[1784,3778,3780],{"id":3779},"_4-build-audit-logs-for-investigation-not-vanity-dashboards","4) Build audit logs for investigation, not vanity dashboards",[12,3782,3783],{},"You need traceable records that answer incident questions quickly: who triggered the request, which model selected the tool, what arguments were sent, what result came back, and whether approval was required and captured.",[12,3785,3786],{},"If the log cannot support a forensic timeline, it is not useful.",[1784,3788,3790],{"id":3789},"_5-design-for-safe-failure","5) Design for safe failure",[12,3792,3793],{},"Tool outages and permission errors should fail closed for sensitive actions. Add clear fallbacks and escalation paths so users get safe outcomes when execution cannot proceed.",[1029,3795,3799],{"className":3796,"code":3797,"language":3798,"meta":135,"style":135},"language-json shiki shiki-themes material-theme-lighter material-theme material-theme-palenight","{\n  \"agent_role\": \"finance_executor\",\n  \"tool\": \"issue_payment\",\n  \"scope\": [\"vendor_whitelist\", \"max_amount_1000_eur\"],\n  \"requires_approval\": true\n}\n","json",[1036,3800,3801,3806,3829,3849,3881,3895],{"__ignoreMap":135},[1837,3802,3803],{"class":1839,"line":1840},[1837,3804,3805],{"class":1847},"{\n",[1837,3807,3808,3811,3815,3817,3819,3821,3824,3826],{"class":1839,"line":136},[1837,3809,3810],{"class":1847},"  \"",[1837,3812,3814],{"class":3813},"spNyl","agent_role",[1837,3816,1903],{"class":1847},[1837,3818,1856],{"class":1847},[1837,3820,1859],{"class":1847},[1837,3822,3823],{"class":1862},"finance_executor",[1837,3825,1903],{"class":1847},[1837,3827,3828],{"class":1847},",\n",[1837,3830,3831,3833,3836,3838,3840,3842,3845,3847],{"class":1839,"line":1869},[1837,3832,3810],{"class":1847},[1837,3834,3835],{"class":3813},"tool",[1837,3837,1903],{"class":1847},[1837,3839,1856],{"class":1847},[1837,3841,1859],{"class":1847},[1837,3843,3844],{"class":1862},"issue_payment",[1837,3846,1903],{"class":1847},[1837,3848,3828],{"class":1847},[1837,3850,3851,3853,3856,3858,3860,3862,3864,3867,3869,3871,3873,3876,3878],{"class":1839,"line":1881},[1837,3852,3810],{"class":1847},[1837,3854,3855],{"class":3813},"scope",[1837,3857,1903],{"class":1847},[1837,3859,1856],{"class":1847},[1837,3861,1900],{"class":1847},[1837,3863,1903],{"class":1847},[1837,3865,3866],{"class":1862},"vendor_whitelist",[1837,3868,1903],{"class":1847},[1837,3870,1911],{"class":1847},[1837,3872,1859],{"class":1847},[1837,3874,3875],{"class":1862},"max_amount_1000_eur",[1837,3877,1903],{"class":1847},[1837,3879,3880],{"class":1847},"],\n",[1837,3882,3883,3885,3888,3890,3892],{"class":1839,"line":1892},[1837,3884,3810],{"class":1847},[1837,3886,3887],{"class":3813},"requires_approval",[1837,3889,1903],{"class":1847},[1837,3891,1856],{"class":1847},[1837,3893,3894],{"class":1847}," true\n",[1837,3896,3897],{"class":1839,"line":1924},[1837,3898,3899],{"class":1847},"}\n",[16,3901,3903],{"id":3902},"a-practical-rollout-model","A practical rollout model",[12,3905,3906],{},"Use a three-stage rollout:",[3026,3908,3909,3915,3921],{},[24,3910,3911,3914],{},[27,3912,3913],{},"Pilot control:"," basic auth + minimal scopes + full logging",[24,3916,3917,3920],{},[27,3918,3919],{},"Operational hardening:"," policy engine + approval routing + alerting",[24,3922,3923,3926],{},[27,3924,3925],{},"Governed scale:"," periodic access review + automated compliance reporting",[12,3928,3929],{},"This avoids security theater and keeps teams shipping.",[16,3931,3933],{"id":3932},"common-mistakes-to-avoid","Common mistakes to avoid",[12,3935,3936],{},"Treating MCP security as API gateway setup alone is a common error. Other recurring mistakes include reusing credentials across environments, logging only successful calls, skipping denial and approval records, and leaving emergency debug permissions permanently enabled.",[179,3938,3939],{},[12,3940,3941],{},"\"Speed without control is not innovation. It is delayed incident response.\"",[16,3943,1041],{"id":1040},[12,3945,3946],{},"MCP gives you speed. Security gives you survivability.",[12,3948,3949],{},"If your tool layer is not governed, your agent layer is not production-ready.",[12,3951,3952],{},"The teams that win will be the ones that can move fast and still prove control under pressure.",[1956,3954,3955],{},"html pre.shiki code .sMK4o, html code.shiki .sMK4o{--shiki-light:#39ADB5;--shiki-default:#89DDFF;--shiki-dark:#89DDFF}html pre.shiki code .spNyl, html code.shiki .spNyl{--shiki-light:#9C3EDA;--shiki-default:#C792EA;--shiki-dark:#C792EA}html pre.shiki code .sfazB, html code.shiki .sfazB{--shiki-light:#91B859;--shiki-default:#C3E88D;--shiki-dark:#C3E88D}html .light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html.light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}html.dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}",{"title":135,"searchDepth":136,"depth":136,"links":3957},[3958,3959,3966,3967,3968],{"id":3726,"depth":136,"text":3727},{"id":3739,"depth":136,"text":3740,"children":3960},[3961,3962,3963,3964,3965],{"id":3743,"depth":1869,"text":3744},{"id":3753,"depth":1869,"text":3754},{"id":3766,"depth":1869,"text":3767},{"id":3779,"depth":1869,"text":3780},{"id":3789,"depth":1869,"text":3790},{"id":3902,"depth":136,"text":3903},{"id":3932,"depth":136,"text":3933},{"id":1040,"depth":136,"text":1041},"2026-03-15T00:00:00.000Z","MCP makes tool integration fast, but security decides whether it scales. This guide explains how to secure MCP servers in production with authentication, scoped permissions, and audit-ready logging.","/images/blog-mcp-security.png",{},"/blog/mcp-security-in-production-auth-and-least-privilege",{"title":3712,"description":3970},"blog/mcp-security-in-production-auth-and-least-privilege",[3977,3978,1101,3979],"mcp","ai security","governance","6Bw28dMcuxddFYXXM0KREQCeDawIwX0agqVkYpBpWeM",{"id":3982,"title":3983,"author":7,"body":3984,"category":333,"date":4210,"description":4211,"extension":144,"featured":145,"image":4212,"meta":4213,"navigation":148,"path":4214,"seo":4215,"stem":4216,"tags":4217,"__hash__":4221},"blog/blog/mcp-servers-and-tools.md","MCP Servers and Tools: How LLMs Connect to the Real World",{"type":9,"value":3985,"toc":4200},[3986,3989,3993,3996,3999,4003,4006,4020,4023,4027,4030,4062,4065,4069,4072,4077,4081,4084,4090,4125,4128,4134,4138,4176,4180,4183,4189,4192,4195,4197],[12,3987,3988],{},"The fastest way to make an LLM useful is to give it tools. MCP (Model Context Protocol) servers make that reliable, secure, and scalable. In this post I’ll explain what MCP servers are, how LLMs use them, why they add real value, and how you can enable them in the Blits.ai platform.",[16,3990,3992],{"id":3991},"_1-what-are-mcp-servers","1. What are MCP servers?",[12,3994,3995],{},"An MCP server is a standardized tool endpoint for LLMs. Instead of hardcoding tool integrations for every model and app, an MCP server exposes a set of tools (with schemas, parameters, and descriptions) that any compatible LLM agent can discover and call.",[12,3997,3998],{},"Think of MCP as a “USB‑C port for tools.” The model doesn’t need to know how every tool works internally. It just needs a consistent protocol to discover tools and call them safely.",[16,4000,4002],{"id":4001},"_2-how-llms-use-tools-and-mcp-servers","2. How LLMs use tools and MCP servers",[12,4004,4005],{},"Modern LLMs can decide when to call a tool. When the model sees a request that requires external data or action, it:",[3026,4007,4008,4011,4014,4017],{},[24,4009,4010],{},"Selects the right tool based on its description and schema.",[24,4012,4013],{},"Produces structured arguments.",[24,4015,4016],{},"Calls the MCP server.",[24,4018,4019],{},"Receives a response and continues the conversation.",[12,4021,4022],{},"MCP servers make this flow consistent across providers and tools. The same agent logic can call internal APIs, search indexes, or custom business systems without rewriting the integration layer every time.",[16,4024,4026],{"id":4025},"_3-the-added-value-of-mcp-servers","3. The added value of MCP servers",[12,4028,4029],{},"MCP servers are not just a technical detail. They solve real business problems:",[21,4031,4032,4038,4044,4050,4056],{},[24,4033,4034,4037],{},[27,4035,4036],{},"Standardization:"," One protocol across vendors, models, and tools.",[24,4039,4040,4043],{},[27,4041,4042],{},"Speed:"," Add or swap tools without refactoring your agent logic.",[24,4045,4046,4049],{},[27,4047,4048],{},"Security:"," Keep secrets on the server side and avoid exposing internal APIs directly to the LLM.",[24,4051,4052,4055],{},[27,4053,4054],{},"Governance:"," Centralize tool access, logging, and access controls.",[24,4057,4058,4061],{},[27,4059,4060],{},"Scalability:"," Reuse the same tool set across multiple assistants and channels.",[12,4063,4064],{},"If you are building agents at scale, MCP is the difference between a prototype and a maintainable platform.",[16,4066,4068],{"id":4067},"_4-the-future-of-mcp-servers","4. The future of MCP servers",[12,4070,4071],{},"MCP servers are likely to become a quiet but essential layer in how AI systems connect to real world tools and data. As models grow more capable the value shifts from raw intelligence to reliable orchestration governance and context control. MCP servers will evolve into standardized trust hubs that manage permissions data flow and execution boundaries across many tools and models. This makes AI systems safer more composable and easier to integrate into serious production environments where control and transparency matter as much as capability.",[179,4073,4074],{},[12,4075,4076],{},"As MCP adoption grows, tool integration will become as standardized as web APIs are today.",[16,4078,4080],{"id":4079},"_5-how-we-use-mcp-servers-at-blitsai-and-how-you-can-add-them","5. How we use MCP servers at Blits.ai (and how you can add them)",[12,4082,4083],{},"In the Blits.ai platform, MCP servers are part of the LLM tools configuration. You can add your own MCP server in a few steps:",[12,4085,4086],{},[826,4087],{"alt":4088,"src":4089},"mcp-server1","/images/mcp-server1.png",[3026,4091,4092,4098,4108,4111,4122],{},[24,4093,4094,4095,1054],{},"Open your LLM in the admin panel and go to ",[27,4096,4097],{},"Your LLM tools",[24,4099,4100,4101,4104,4105,1054],{},"Scroll to ",[27,4102,4103],{},"MCP Servers"," and click ",[27,4106,4107],{},"Add new MCP Tool",[24,4109,4110],{},"Provide a display name, URL, and description.",[24,4112,4113,4114,4117,4118,4121],{},"Choose the channel (",[1036,4115,4116],{},"sse"," or ",[1036,4119,4120],{},"http",").",[24,4123,4124],{},"Optionally add headers and params if your MCP server needs authentication or default arguments.",[12,4126,4127],{},"Once saved, the MCP server is available to your LLM and can be toggled on or off. This is how we connect our agents to external systems without hardcoded integrations, and it lets customers bring their own tool stacks into the Blits.ai platform.",[12,4129,4130],{},[826,4131],{"alt":4132,"src":4133},"mcp-server2","/images/mcp-server2.png",[16,4135,4137],{"id":4136},"_6-best-practices-for-mcp-servers","6. Best practices for MCP servers",[21,4139,4140,4146,4152,4158,4164,4170],{},[24,4141,4142,4145],{},[27,4143,4144],{},"Start small:"," Add a few high‑impact tools first (search, data lookup, CRM).",[24,4147,4148,4151],{},[27,4149,4150],{},"Describe tools clearly:"," The model relies on tool descriptions to choose correctly.",[24,4153,4154,4157],{},[27,4155,4156],{},"Use least‑privilege access:"," Only expose what the tool needs to do.",[24,4159,4160,4163],{},[27,4161,4162],{},"Log every call:"," Tool usage is part of your audit trail.",[24,4165,4166,4169],{},[27,4167,4168],{},"Add timeouts and retries:"," Tool failures should not break the conversation.",[24,4171,4172,4175],{},[27,4173,4174],{},"Test with realistic prompts:"," Evaluate how the model selects and uses tools.",[16,4177,4179],{"id":4178},"_7-try-our-demo-mcp-server-football-shirt-customizer","7. Try our demo MCP server (football shirt customizer)",[12,4181,4182],{},"We have a demo MCP server running at:",[12,4184,4185],{},[1137,4186,4187],{"href":4187,"rel":4188},"https://www.blits.ai/mcpdemo",[1288],[12,4190,4191],{},"It’s built together with partner 270degrees to create a product experience, where you can customize a football shirt in real time. It’s a simple example, but it shows the core idea: the LLM calls tools through MCP to change colors, names, and options in a real UI.",[12,4193,4194],{},"If you want to explore how this works in code, check out the demo page.",[16,4196,1715],{"id":313},[12,4198,4199],{},"MCP servers are a missing layer between “LLMs that can talk” and “LLMs that can do.” They make tool integration portable, secure, and scalable. If you’re building agents for real business use‑cases, MCP is not optional anymore, it’s the foundation.",{"title":135,"searchDepth":136,"depth":136,"links":4201},[4202,4203,4204,4205,4206,4207,4208,4209],{"id":3991,"depth":136,"text":3992},{"id":4001,"depth":136,"text":4002},{"id":4025,"depth":136,"text":4026},{"id":4067,"depth":136,"text":4068},{"id":4079,"depth":136,"text":4080},{"id":4136,"depth":136,"text":4137},{"id":4178,"depth":136,"text":4179},{"id":313,"depth":136,"text":1715},"2026-01-27T00:00:00.000Z","MCP servers standardize how LLMs discover and use tools. This guide explains what they are, why they matter, how we use them at Blits.ai, and where the ecosystem is heading.","/images/mcp-server-header.webp",{},"/blog/mcp-servers-and-tools",{"title":3983,"description":4211},"blog/mcp-servers-and-tools",[4218,4219,3977,4220],"llm","tools","agents","c7pr8Cq9b1NmVeAtEzNjfor8AsH21F8wdk2rdpztlpI",{"id":4223,"title":4224,"author":7,"body":4225,"category":333,"date":4562,"description":4563,"extension":144,"featured":145,"image":4564,"meta":4565,"navigation":148,"path":4566,"seo":4567,"stem":4568,"tags":4569,"__hash__":4572},"blog/blog/measure-ai-performance-and-set-the-right-kpis.md","Measure AI Performance and Set the Right KPIs",{"type":9,"value":4226,"toc":4546},[4227,4230,4233,4236,4239,4243,4246,4249,4252,4255,4258,4278,4281,4285,4288,4292,4295,4298,4302,4305,4308,4311,4330,4334,4337,4340,4343,4347,4350,4353,4370,4373,4377,4380,4383,4421,4424,4438,4442,4445,4448,4451,4454,4458,4465,4468,4472,4475,4513,4516,4520,4523,4526,4529,4532,4535,4537,4540,4543],[12,4228,4229],{},"Most AI teams can show a good demo. Fewer can show stable production performance.",[12,4231,4232],{},"That is the gap between \"AI works\" and \"AI delivers business value.\"",[12,4234,4235],{},"If you want reliable outcomes, you need to measure the right things continuously. Not just once during model selection.",[12,4237,4238],{},"In this article I will break down how to define AI KPIs, what key technical metrics mean, and how to combine quality and performance into one operational view.",[16,4240,4242],{"id":4241},"why-ai-kpi-design-matters","Why AI KPI design matters",[12,4244,4245],{},"Without clear KPIs, teams optimize for whatever is easiest to measure: token cost, average response time, or a benchmark screenshot.",[12,4247,4248],{},"But users do not experience averages. They experience waiting times, failures, inconsistent answers, and wrong actions.",[12,4250,4251],{},"For enterprise teams, this is crucial because AI is no longer an experiment running in isolation. It is being connected to customer channels, internal operations, and increasingly to automated workflows. When KPI design is weak, risk is invisible until it becomes expensive: higher support load, lower customer trust, slower compliance cycles, and poor scaling decisions.",[12,4253,4254],{},"Strong KPI design creates a shared language between product, engineering, operations, risk, and leadership. It helps teams decide what \"good\" means before incidents happen, and it makes trade-offs explicit when speed, quality, and cost are in tension.",[12,4256,4257],{},"A good AI KPI framework should connect three layers:",[3026,4259,4260,4266,4272],{},[24,4261,4262,4265],{},[27,4263,4264],{},"User experience"," (speed, consistency, trust)",[24,4267,4268,4271],{},[27,4269,4270],{},"Technical performance"," (latency, reliability, errors)",[24,4273,4274,4277],{},[27,4275,4276],{},"Business outcomes"," (conversion, containment, productivity, cost)",[12,4279,4280],{},"If one layer is missing, you can get false confidence quickly.",[16,4282,4284],{"id":4283},"core-performance-concepts-you-should-track","Core performance concepts you should track",[12,4286,4287],{},"Before diving into specific metrics, one mindset matters: you should measure performance the way users experience it, not the way components are organized internally. A model can benchmark well in isolation and still feel slow or unreliable in production because retrieval, tools, guardrails, and integrations all add friction. That is why these core metrics should always be interpreted end-to-end.",[1784,4289,4291],{"id":4290},"latency","Latency",[12,4293,4294],{},"Latency is the total time between a user request and a usable response.",[12,4296,4297],{},"For AI systems, this often includes multiple steps: retrieval, model inference, tool calls, post-processing, and response delivery. If one component is slow, the full experience feels slow.",[1784,4299,4301],{"id":4300},"p99","P99",[12,4303,4304],{},"P99 is the response time under which 99% of requests complete.",[12,4306,4307],{},"Why it matters: averages can look healthy while real users still suffer on slow tail requests. P99 helps you see that tail risk. In customer-facing AI, tail latency usually drives frustration more than average latency.",[12,4309,4310],{},"In practice, teams should track at least:",[21,4312,4313,4319,4325],{},[24,4314,4315,4318],{},[27,4316,4317],{},"P50"," (typical user experience)",[24,4320,4321,4324],{},[27,4322,4323],{},"P95"," (high-load realism)",[24,4326,4327,4329],{},[27,4328,4301],{}," (worst-case user impact at scale)",[1784,4331,4333],{"id":4332},"ttft-time-to-first-token","TTFT (Time to First Token)",[12,4335,4336],{},"TTFT is how fast the first token appears after a request is sent.",[12,4338,4339],{},"In streaming interfaces, TTFT is a critical perception metric. Even if total completion takes longer, fast first feedback makes the assistant feel responsive and alive.",[12,4341,4342],{},"If your AI assistant supports streaming, TTFT is often as important as full completion latency.",[1784,4344,4346],{"id":4345},"error-rates","Error rates",[12,4348,4349],{},"Error rates represent failed requests as a percentage of total requests.",[12,4351,4352],{},"But you should split this metric, because \"error\" can mean many different things:",[21,4354,4355,4358,4361,4364,4367],{},[24,4356,4357],{},"Provider/API failures",[24,4359,4360],{},"Timeouts",[24,4362,4363],{},"Tool call failures",[24,4365,4366],{},"Policy or guardrail blocks",[24,4368,4369],{},"Parsing/validation failures",[12,4371,4372],{},"The total error rate is useful, but the breakdown tells you where to fix the system.",[16,4374,4376],{"id":4375},"ways-to-measure-quality-not-just-speed","Ways to measure quality (not just speed)",[12,4378,4379],{},"Fast answers are useless when they are wrong. Quality must be measured as rigorously as latency.",[12,4381,4382],{},"Useful quality indicators include:",[3026,4384,4385,4391,4397,4403,4409,4415],{},[24,4386,4387,4390],{},[27,4388,4389],{},"Task success rate:"," Did the user complete the intended goal?",[24,4392,4393,4396],{},[27,4394,4395],{},"Groundedness score:"," Is the answer supported by trusted sources?",[24,4398,4399,4402],{},[27,4400,4401],{},"Hallucination rate:"," How often does the model produce unsupported claims?",[24,4404,4405,4408],{},[27,4406,4407],{},"Human review score:"," Expert rating on correctness, clarity, and safety.",[24,4410,4411,4414],{},[27,4412,4413],{},"Containment rate:"," How often the assistant resolves without human handoff (when that is the goal).",[24,4416,4417,4420],{},[27,4418,4419],{},"CSAT / user feedback:"," Direct signal from real users.",[12,4422,4423],{},"For agentic workflows, include action quality metrics as well:",[21,4425,4426,4429,4432,4435],{},[24,4427,4428],{},"Correct tool selected",[24,4430,4431],{},"Correct parameters passed",[24,4433,4434],{},"Correct outcome achieved",[24,4436,4437],{},"Human override frequency",[16,4439,4441],{"id":4440},"how-to-set-kpis-that-actually-work","How to set KPIs that actually work",[12,4443,4444],{},"Start simple: do not force one KPI template across every AI use case. A customer support assistant and an internal drafting assistant are different products with different risk profiles, so they need different thresholds.",[12,4446,4447],{},"Once you split by use-case tier, set a small KPI set per tier, for example your maximum P99 latency, TTFT target, maximum error rate, and minimum groundedness or task success.",[12,4449,4450],{},"Then make ownership explicit. Decide up front who gets paged when error rates spike, who approves model version changes, and which tests must pass before rollout.",[12,4452,4453],{},"If those owners and decisions are not clear, KPIs quickly become dashboard decoration instead of an operational control system.",[1784,4455,4457],{"id":4456},"example-kpi-quote-for-a-financial-institution","Example KPI quote for a financial institution",[179,4459,4460],{},[12,4461,4462],{},[27,4463,4464],{},"\"For customer-facing banking assistants, our production target is: P99 latency below 2.5 seconds, TTFT below 1200 ms, total error rate below 0.5%, and groundedness above 98% on policy and regulatory answers. Any high-risk financial action requires validation and human approval before execution.\"",[12,4466,4467],{},"This kind of KPI statement is strong because it combines speed, reliability, factual quality, and risk controls in one operational target.",[16,4469,4471],{"id":4470},"operational-model-measure-test-improve","Operational model: measure, test, improve",[12,4473,4474],{},"A practical loop for enterprise AI teams:",[3026,4476,4477,4483,4489,4495,4501,4507],{},[24,4478,4479,4482],{},[27,4480,4481],{},"Instrument every step"," (retrieval, model, tools, validation, output)",[24,4484,4485,4488],{},[27,4486,4487],{},"Benchmark regularly"," across providers and model versions",[24,4490,4491,4494],{},[27,4492,4493],{},"Run regression tests"," on fixed evaluation sets",[24,4496,4497,4500],{},[27,4498,4499],{},"Monitor production metrics"," in real time",[24,4502,4503,4506],{},[27,4504,4505],{},"Route high-risk failures"," to human review",[24,4508,4509,4512],{},[27,4510,4511],{},"Iterate prompts, tools, and policies"," based on evidence",[12,4514,4515],{},"This is how AI systems move from experimentation to dependable operations.",[16,4517,4519],{"id":4518},"how-we-do-this-at-blits","How we do this at Blits",[12,4521,4522],{},"At Blits, we treat performance and quality measurement as a built-in platform capability, not as a side dashboard.",[12,4524,4525],{},"For each AI use case, we measure end-to-end flow performance across the full stack: retrieval, model response, tool calls, and output validation. That gives teams visibility into where latency or failure is actually introduced, instead of blaming one model for a system-level issue.",[12,4527,4528],{},"We continuously benchmark provider-model combinations on the same scenarios and compare results on latency, P99, TTFT, error patterns, and quality outcomes. This makes it possible to switch models or providers based on evidence, while keeping consistent user experience and governance requirements.",[12,4530,4531],{},"For agentic workflows, we add additional control points with guardrails and validations before high-impact actions are executed. This reduces the chance that uncertain model behavior becomes an operational incident.",[12,4533,4534],{},"Most importantly, we link technical KPIs to business KPIs. Faster TTFT is only valuable if it improves containment, conversion, or productivity. Lower error rates are only meaningful if they reduce escalations and rework. That KPI linkage is what turns AI performance data into business decisions.",[16,4536,1041],{"id":1040},[12,4538,4539],{},"AI performance is not one metric. It is a balance between speed, reliability, quality, and business impact.",[12,4541,4542],{},"If you want to scale AI safely, measure what users feel, what the system does, and what the business gets.",[12,4544,4545],{},"Teams that treat KPIs as a core AI capability, not a reporting task, will outperform teams that only optimize for model hype.",{"title":135,"searchDepth":136,"depth":136,"links":4547},[4548,4549,4555,4556,4559,4560,4561],{"id":4241,"depth":136,"text":4242},{"id":4283,"depth":136,"text":4284,"children":4550},[4551,4552,4553,4554],{"id":4290,"depth":1869,"text":4291},{"id":4300,"depth":1869,"text":4301},{"id":4332,"depth":1869,"text":4333},{"id":4345,"depth":1869,"text":4346},{"id":4375,"depth":136,"text":4376},{"id":4440,"depth":136,"text":4441,"children":4557},[4558],{"id":4456,"depth":1869,"text":4457},{"id":4470,"depth":136,"text":4471},{"id":4518,"depth":136,"text":4519},{"id":1040,"depth":136,"text":1041},"2026-03-07T00:00:00.000Z","AI performance is not just model accuracy. In this article I explain the KPIs that matter in production, including latency, P99, TTFT, error rates, and practical ways to measure both quality and business impact.","/images/blits-performance.jpg",{},"/blog/measure-ai-performance-and-set-the-right-kpis",{"title":4224,"description":4563},"blog/measure-ai-performance-and-set-the-right-kpis",[4570,2171,4571,1101],"ai kpi","observability","Q8WhiNJK2EU8pLZsk8uDiKwSdKJCIZALRkZ-frgcnRc",{"id":4574,"title":4575,"author":1986,"body":4576,"category":333,"date":4676,"description":4580,"extension":144,"featured":145,"image":4677,"meta":4678,"navigation":148,"path":4679,"seo":4680,"stem":4681,"tags":340,"__hash__":4682},"blog/blog/opening-the-natural-language-understanding-nlu-blackbox.md","Opening the Natural Language Understanding (NLU) Blackbox",{"type":9,"value":4577,"toc":4669},[4578,4581,4585,4588,4592,4618,4622,4625,4642,4644,4647,4664,4666],[12,4579,4580],{},"Natural Language Understanding (NLU) is often treated as a black box, but understanding how it works is crucial for building effective conversational AI solutions.",[16,4582,4584],{"id":4583},"understanding-nlu","Understanding NLU",[12,4586,4587],{},"NLU is the component of conversational AI that interprets human language and converts it into structured data that machines can process. It's the foundation that enables chatbots to understand user intent and extract relevant information.",[16,4589,4591],{"id":4590},"key-nlu-components","Key NLU Components",[21,4593,4594,4600,4606,4612],{},[24,4595,4596,4599],{},[27,4597,4598],{},"Intent Classification:"," Determines what the user wants to accomplish",[24,4601,4602,4605],{},[27,4603,4604],{},"Entity Extraction:"," Identifies specific pieces of information in user input",[24,4607,4608,4611],{},[27,4609,4610],{},"Context Management:"," Maintains conversation context across multiple turns",[24,4613,4614,4617],{},[27,4615,4616],{},"Confidence Scoring:"," Provides reliability metrics for NLU decisions",[16,4619,4621],{"id":4620},"choosing-the-right-nlu-engine","Choosing the Right NLU Engine",[12,4623,4624],{},"Different NLU engines have different strengths and weaknesses. Key factors to consider include:",[21,4626,4627,4630,4633,4636,4639],{},[24,4628,4629],{},"Accuracy for your specific domain and use case",[24,4631,4632],{},"Training data requirements and customization options",[24,4634,4635],{},"Integration complexity and API availability",[24,4637,4638],{},"Cost structure and scalability considerations",[24,4640,4641],{},"Language support and multilingual capabilities",[16,4643,3548],{"id":3547},[12,4645,4646],{},"Successful NLU implementation requires:",[21,4648,4649,4652,4655,4658,4661],{},[24,4650,4651],{},"Thorough testing with real user data",[24,4653,4654],{},"Continuous monitoring and optimization",[24,4656,4657],{},"Proper training data management",[24,4659,4660],{},"Fallback strategies for low-confidence predictions",[24,4662,4663],{},"Regular model updates and retraining",[16,4665,3555],{"id":3554},[12,4667,4668],{},"Understanding NLU is essential for building effective conversational AI. By demystifying this black box, businesses can make informed decisions about their chatbot architecture and achieve better results.",{"title":135,"searchDepth":136,"depth":136,"links":4670},[4671,4672,4673,4674,4675],{"id":4583,"depth":136,"text":4584},{"id":4590,"depth":136,"text":4591},{"id":4620,"depth":136,"text":4621},{"id":3547,"depth":136,"text":3548},{"id":3554,"depth":136,"text":3555},"2020-06-23T00:00:00.000Z","/blog/Monitoring-300x234.png",{},"/blog/opening-the-natural-language-understanding-nlu-blackbox",{"title":4575,"description":4580},"blog/opening-the-natural-language-understanding-nlu-blackbox","-L4ocI0BjZm6rdZB-TUzuonqn-bMtvvMW70twbpQGAI",{"id":4684,"title":4685,"author":7,"body":4686,"category":333,"date":4807,"description":4808,"extension":144,"featured":145,"image":4809,"meta":4810,"navigation":148,"path":4811,"seo":4812,"stem":4813,"tags":4814,"__hash__":4817},"blog/blog/rag-in-2026-why-enterprise-pipelines-still-fail.md","RAG in 2026: Why Enterprise Pipelines Still Fail and How to Fix Them",{"type":9,"value":4687,"toc":4795},[4688,4691,4694,4697,4700,4702,4706,4710,4713,4716,4720,4723,4726,4730,4733,4736,4740,4743,4746,4750,4753,4756,4761,4765,4768,4771,4777,4781,4784,4786,4789,4792],[12,4689,4690],{},"RAG is no longer new.",[12,4692,4693],{},"Yet many enterprise RAG systems still produce inconsistent answers, weak citations, and unpredictable quality under real load.",[12,4695,4696],{},"So what is going wrong?",[12,4698,4699],{},"In most cases, not the model. The pipeline.",[12,4701,1766],{},[16,4703,4705],{"id":4704},"the-five-failure-patterns-we-keep-seeing","The five failure patterns we keep seeing",[1784,4707,4709],{"id":4708},"_1-retrieval-quality-is-treated-as-a-one-time-setup","1) Retrieval quality is treated as a one-time setup",[12,4711,4712],{},"Teams create an index once, run a few tests, and move on.",[12,4714,4715],{},"But source content changes, metadata drifts, and relevance degrades. Retrieval needs continuous tuning and monitoring.",[1784,4717,4719],{"id":4718},"_2-chunking-strategy-ignores-use-case-semantics","2) Chunking strategy ignores use-case semantics",[12,4721,4722],{},"Chunking by arbitrary token size often breaks meaning.",[12,4724,4725],{},"For policy-heavy and technical content, structure-aware chunking is critical. If context is fragmented, the model fills gaps with plausible noise.",[1784,4727,4729],{"id":4728},"_3-ranking-is-optimized-for-keyword-similarity-not-answer-utility","3) Ranking is optimized for keyword similarity, not answer utility",[12,4731,4732],{},"A retrieved chunk can be relevant and still unhelpful.",[12,4734,4735],{},"Ranking should optimize for answerability, freshness, and authority, not only embedding distance.",[1784,4737,4739],{"id":4738},"_4-citation-behavior-is-optional-instead-of-enforced","4) Citation behavior is optional instead of enforced",[12,4741,4742],{},"If source attribution is not required, models will overconfidently answer from memory patterns.",[12,4744,4745],{},"High-impact use cases should require grounded evidence and explicit uncertainty when evidence is insufficient.",[1784,4747,4749],{"id":4748},"_5-no-regression-evaluation-loop","5) No regression evaluation loop",[12,4751,4752],{},"Many teams evaluate once per quarter while content and prompts change weekly.",[12,4754,4755],{},"Without regression gates, quality slips silently into production.",[179,4757,4758],{},[12,4759,4760],{},"\"RAG failures are usually systems failures wearing a model mask.\"",[16,4762,4764],{"id":4763},"a-practical-reliability-model-for-rag","A practical reliability model for RAG",[12,4766,4767],{},"Use a layered operating model that treats RAG as infrastructure, not prompt craft. The data layer enforces source quality and permissions. The retrieval layer handles chunking, ranking, and hybrid tuning. The generation layer adds bounded prompts and citation enforcement. Validation routes low-confidence outputs and high-risk cases. Operations ties all of this to KPIs, incident handling, and release gates.",[12,4769,4770],{},"This is what turns RAG from a feature into a governed system.",[1029,4772,4775],{"className":4773,"code":4774,"language":1034,"meta":135},[1032],"RAG release gate (example):\ngrounded_answer_rate >= 97%\ncitation_correctness >= 98%\nhallucination_rate_high_risk \u003C= 1%\nretrieval_precision_at_5 >= 90%\n",[1036,4776,4774],{"__ignoreMap":135},[16,4778,4780],{"id":4779},"kpis-that-actually-matter","KPIs that actually matter",[12,4782,4783],{},"Track more than answer acceptance. Grounded answer rate, citation correctness, hallucination rate in high-risk queries, retrieval precision at k, and resolution quality by channel-language pair should be monitored together and tied to business outcomes.",[16,4785,1041],{"id":1040},[12,4787,4788],{},"RAG does not fail because the concept is weak.",[12,4790,4791],{},"It fails when teams treat it as prompt engineering instead of information engineering plus operations.",[12,4793,4794],{},"If you want reliable AI in enterprise settings, your RAG pipeline needs ownership, instrumentation, and continuous evaluation, not one-time setup.",{"title":135,"searchDepth":136,"depth":136,"links":4796},[4797,4804,4805,4806],{"id":4704,"depth":136,"text":4705,"children":4798},[4799,4800,4801,4802,4803],{"id":4708,"depth":1869,"text":4709},{"id":4718,"depth":1869,"text":4719},{"id":4728,"depth":1869,"text":4729},{"id":4738,"depth":1869,"text":4739},{"id":4748,"depth":1869,"text":4749},{"id":4763,"depth":136,"text":4764},{"id":4779,"depth":136,"text":4780},{"id":1040,"depth":136,"text":1041},"2026-04-16T00:00:00.000Z","Retrieval-augmented generation is now standard, yet many enterprise deployments still underperform. This article explains the main failure points and a practical operating model for reliable RAG.","/images/blits-hallucination.jpg",{},"/blog/rag-in-2026-why-enterprise-pipelines-still-fail",{"title":4685,"description":4808},"blog/rag-in-2026-why-enterprise-pipelines-still-fail",[4815,1101,4218,4816],"rag","ai reliability","2-LbDgBhvMC2OmHimPxYgv0BVOOBBJEZ-IWBxVWW-T8",{"id":4819,"title":4820,"author":7,"body":4821,"category":333,"date":5138,"description":5139,"extension":144,"featured":145,"image":5140,"meta":5141,"navigation":148,"path":5142,"seo":5143,"stem":5144,"tags":5145,"__hash__":5147},"blog/blog/text-to-speech-engines-and-why-they-matter.md","Text-to-Speech Engines: The Voice Layer Every AI Product Needs",{"type":9,"value":4822,"toc":5119},[4823,4826,4829,4832,4836,4839,4842,4862,4865,4869,4872,4876,4879,4883,4886,4890,4893,4897,4900,4903,4907,4910,4914,4920,4923,4926,4948,4971,4985,4988,4995,4998,5002,5005,5009,5012,5016,5019,5023,5026,5030,5033,5036,5041,5045,5048,5051,5083,5086,5090,5093,5096,5113,5116],[12,4824,4825],{},"Most teams still treat text-to-speech (TTS) as a final output step. It is not. It is a core product layer.",[12,4827,4828],{},"If your assistant can reason well but sounds robotic, slow, or culturally off, users will not trust it. Voice is where AI gets judged in real life.",[12,4830,4831],{},"In this article I will break down what TTS engines are, what model options exist today, which providers matter, and why speed, quality, region, and dialect should be treated as first-class architecture decisions.",[16,4833,4835],{"id":4834},"_1-what-are-text-to-speech-engines-and-where-are-they-used","1) What are text-to-speech engines, and where are they used?",[12,4837,4838],{},"A text-to-speech engine converts written text into synthetic speech. Modern engines no longer just \"read words\"; they model prosody, pacing, emphasis, and pronunciation so output sounds more human.",[12,4840,4841],{},"You see TTS everywhere:",[21,4843,4844,4847,4850,4853,4856,4859],{},[24,4845,4846],{},"Voice assistants and conversational banking",[24,4848,4849],{},"Contact center automation and IVR modernization",[24,4851,4852],{},"E-learning and accessibility solutions",[24,4854,4855],{},"In-car assistants and infotainment systems",[24,4857,4858],{},"Real-time translation and multilingual customer support",[24,4860,4861],{},"Media, gaming, and dynamic content generation",[12,4863,4864],{},"In practical terms, TTS is often the final mile between model intelligence and human experience. That final mile decides whether the interaction feels natural or not.",[16,4866,4868],{"id":4867},"_2-what-model-solutions-are-there","2) What model solutions are there?",[12,4870,4871],{},"There is no single \"best TTS model.\" There are model families, each with a different trade-off profile.",[1784,4873,4875],{"id":4874},"foundation-and-api-first-tts-models","Foundation and API-first TTS models",[12,4877,4878],{},"These are managed models from major providers. They are fast to integrate, continuously improved, and usually offer broad language coverage. For many teams, this is the best first production path.",[1784,4880,4882],{"id":4881},"custom-domain-voices","Custom domain voices",[12,4884,4885],{},"Some organizations need strict brand voice control, regulated wording style, or persona-specific output. In those cases, teams tune prompts, lexicons, and post-processing pipelines, or train custom voices with specialized vendors.",[1784,4887,4889],{"id":4888},"voice-cloning-and-speaker-adaptation","Voice cloning and speaker adaptation",[12,4891,4892],{},"Voice cloning can deliver strong personalization, but it introduces governance questions immediately: permissions, consent, identity misuse risk, and legal boundaries. Technically powerful, operationally sensitive.",[1784,4894,4896],{"id":4895},"llm-native-speech-generation","LLM-native speech generation",[12,4898,4899],{},"Newer systems combine language reasoning and speech generation more tightly, reducing handoffs between separate modules. This can improve naturalness and reduce latency in certain real-time scenarios.",[12,4901,4902],{},"At the same time, not every LLM includes native TTS, and even when it does, language quality can vary a lot by market and dialect. A model that performs well in English does not automatically perform well in Arabic, Turkish, or mixed-language conversations.",[1784,4904,4906],{"id":4905},"hybrid-stacks","Hybrid stacks",[12,4908,4909],{},"Many enterprise setups are hybrid by design: one engine for low-latency live calls, another for premium voice quality, and a fallback provider for reliability or regional compliance.",[16,4911,4913],{"id":4912},"_3-which-providers-are-there","3) Which providers are there?",[12,4915,4916],{},[826,4917],{"alt":4918,"src":4919},"blits-tts-test","/images/blits-tts-test.png",[12,4921,4922],{},"In our current TTS integration landscape we work across Google, Microsoft, Amazon, IBM, OpenAI, Gemini, ElevenLabs, Deepgram, Murf, Cartesia, and Resemble. The market is mature enough that every provider can produce \"good\" output in a demo. The difference shows up when you move from a demo to production.",[12,4924,4925],{},"The hyperscalers, such as Google, Microsoft, Amazon, and IBM, are usually the safest choice for governance-heavy organizations. They are strong on enterprise controls, regional deployment options, and operational reliability. In our testing context this often translates into predictable performance and easier compliance discussions, but sometimes a less distinctive voice identity for brand-led use cases.",[12,4927,4928,4929,544,4932,544,4935,544,4938,544,4941,1339,4944,4947],{},"Then there are the fast-moving model providers such as OpenAI and Gemini. We currently validate models like ",[1036,4930,4931],{},"tts-1",[1036,4933,4934],{},"tts-1-hd",[1036,4936,4937],{},"gpt-4o-mini-tts",[1036,4939,4940],{},"gemini-2.5-flash-tts",[1036,4942,4943],{},"gemini-2.5-flash-lite-preview-tts",[1036,4945,4946],{},"gemini-2.5-pro-tts",". Their main advantage is speed of innovation and a strong quality/latency balance. The trade-off is operational: model families evolve quickly, so teams need disciplined versioning, regular regression checks, and clear fallback paths.",[12,4949,4950,4951,544,4954,544,4957,1339,4960,4963,4964,733,4967,4970],{},"Voice-specialist providers, especially ElevenLabs, Resemble, and in certain scenarios Murf, often stand out when naturalness and brand voice are the top priority. In our validated set this includes options such as ",[1036,4952,4953],{},"eleven_flash_v2_5",[1036,4955,4956],{},"eleven_multilingual_v2",[1036,4958,4959],{},"eleven_turbo_v2_5",[1036,4961,4962],{},"eleven_v3",", as well as Murf's ",[1036,4965,4966],{},"Gen2",[1036,4968,4969],{},"Falcon",". These providers can deliver impressive voice character and multilingual experiences, but procurement, licensing, and deployment constraints can become the deciding factor in enterprise environments.",[12,4972,4973,4974,4977,4978,544,4981,4984],{},"For real-time conversational systems, latency-focused providers like Deepgram (",[1036,4975,4976],{},"aura",") and Cartesia (",[1036,4979,4980],{},"Sonic2",[1036,4982,4983],{},"Sonic3",") are increasingly relevant. They are designed for responsive interaction loops, where milliseconds matter. The practical question is not only speed, but whether language coverage, long-form stability, and regional requirements match your target markets.",[12,4986,4987],{},"There is also a serious open-source track that many enterprise teams should consider. Running TTS locally can be a major advantage when data cannot leave your network, when you need predictable per-minute costs, or when you want full control over deployment and model behavior. For English, there are now strong open-source options with surprisingly high quality, such as Coqui XTTS v2, Piper, and StyleTTS2. The challenge starts when you move beyond English: multilingual quality and dialect consistency can still be uneven, and production hardening often requires extra engineering around voice selection, pronunciation control, and model tuning.",[179,4989,4990],{},[12,4991,4992],{},[27,4993,4994],{},"That is why the strategic decision is not \"who is best overall.\" The right question is: which provider-model combination is best for this specific language, channel, region, and latency target today, and how quickly can we switch when that answer changes tomorrow.",[12,4996,4997],{},"In practice, this means choosing a complete voice stack, not a single model: speech-to-text, language model, and text-to-speech must be selected and tested together for the target language experience.",[16,4999,5001],{"id":5000},"_4-why-speed-quality-region-and-dialect-are-critical","4) Why speed, quality, region, and dialect are critical",[12,5003,5004],{},"From our work on Saudi Arabic voice experiences, and similar projects across Gulf, Egyptian, and Levantine Arabic dialects, one lesson keeps repeating: voice quality is a system property, not a single model property.",[1784,5006,5008],{"id":5007},"speed-latency","Speed (latency)",[12,5010,5011],{},"In voice conversations, delay kills trust. If responses come back late, users interrupt, repeat, or abandon the flow. Good TTS is not only about waveform quality; it is about response time under real traffic conditions.",[1784,5013,5015],{"id":5014},"quality-naturalness-and-intelligibility","Quality (naturalness and intelligibility)",[12,5017,5018],{},"A voice can be technically clear but still feel synthetic. Users notice rhythm, emphasis, and pronunciation errors immediately, especially in repeated operational flows like banking or support journeys.",[1784,5020,5022],{"id":5021},"region-deployment-and-compliance","Region (deployment and compliance)",[12,5024,5025],{},"For enterprise deployments, region matters as much as model quality. Data residency, cloud constraints, and procurement realities often narrow the viable choices. A \"best model\" that cannot run in your allowed environment is not best for your business.",[1784,5027,5029],{"id":5028},"dialect-local-credibility","Dialect (local credibility)",[12,5031,5032],{},"Dialect consistency is decisive in Arabic deployments. This applies not only to Saudi Arabic, but also to other dialect families where users immediately hear when a system mixes styles. Mixing Modern Standard Arabic and local dialects reduces recognition quality upstream and makes generated speech sound less natural downstream.",[12,5034,5035],{},"When all components in the voice pipeline align on the same dialect, user experience improves quickly: better understanding, better response quality, and fewer conversational breakdowns.",[179,5037,5038],{},[12,5039,5040],{},"In short: the strongest voice systems optimize for the full pipeline, not only for one model benchmark.",[16,5042,5044],{"id":5043},"_5-why-blits-multi-engine-approach-adds-value","5) Why Blits' multi-engine approach adds value",[12,5046,5047],{},"At Blits, voice is built as an orchestration layer, not a lock-in layer. You can connect multiple TTS engines, switch between models, and measure performance per use case.",[12,5049,5050],{},"That creates concrete business value:",[21,5052,5053,5059,5065,5071,5077],{},[24,5054,5055,5058],{},[27,5056,5057],{},"Faster experimentation:"," compare engines per language, channel, and use case.",[24,5060,5061,5064],{},[27,5062,5063],{},"Better outcomes:"," optimize for latency, quality, and dialect fit instead of brand popularity.",[24,5066,5067,5070],{},[27,5068,5069],{},"Vendor resilience:"," avoid being blocked by one provider's pricing or policy changes.",[24,5072,5073,5076],{},[27,5074,5075],{},"Compliance flexibility:"," route workloads to providers that fit regional requirements.",[24,5078,5079,5082],{},[27,5080,5081],{},"Continuous optimization:"," benchmark and improve over time as models evolve.",[12,5084,5085],{},"This is especially relevant for large organizations where voice quality must be consistent across markets while still adapting locally.",[16,5087,5089],{"id":5088},"_6-what-comes-next-in-voice-and-why-this-is-crucial-for-ai","6) What comes next in voice, and why this is crucial for AI",[12,5091,5092],{},"The next wave in AI is not only better text reasoning. It is real-time, multimodal interaction where voice becomes a primary interface.",[12,5094,5095],{},"What to expect next:",[21,5097,5098,5101,5104,5107,5110],{},[24,5099,5100],{},"More real-time speech generation with lower end-to-end latency",[24,5102,5103],{},"Better emotional control and speaking style transfer",[24,5105,5106],{},"Stronger dialect and code-switching support",[24,5108,5109],{},"Tighter integration between LLM reasoning and speech output",[24,5111,5112],{},"More enterprise controls for safety, governance, and auditing",[12,5114,5115],{},"Why this matters: voice is the most human interface we have. If AI is going to operate in customer service, healthcare, finance, public services, and education at scale, the voice layer must be fast, trustworthy, culturally correct, and operationally controllable.",[12,5117,5118],{},"Teams that treat TTS as a strategic infrastructure component today will ship more natural AI products tomorrow.",{"title":135,"searchDepth":136,"depth":136,"links":5120},[5121,5122,5129,5130,5136,5137],{"id":4834,"depth":136,"text":4835},{"id":4867,"depth":136,"text":4868,"children":5123},[5124,5125,5126,5127,5128],{"id":4874,"depth":1869,"text":4875},{"id":4881,"depth":1869,"text":4882},{"id":4888,"depth":1869,"text":4889},{"id":4895,"depth":1869,"text":4896},{"id":4905,"depth":1869,"text":4906},{"id":4912,"depth":136,"text":4913},{"id":5000,"depth":136,"text":5001,"children":5131},[5132,5133,5134,5135],{"id":5007,"depth":1869,"text":5008},{"id":5014,"depth":1869,"text":5015},{"id":5021,"depth":1869,"text":5022},{"id":5028,"depth":1869,"text":5029},{"id":5043,"depth":136,"text":5044},{"id":5088,"depth":136,"text":5089},"2026-03-03T00:00:00.000Z","Text-to-speech is becoming core AI infrastructure. In this article I explain what TTS engines are, which model approaches exist, which providers matter, and why speed, quality, region, and dialect decide real-world success.","/images/blits-voice.jpg",{},"/blog/text-to-speech-engines-and-why-they-matter",{"title":4820,"description":5139},"blog/text-to-speech-engines-and-why-they-matter",[2168,2169,5146],"ai infrastructure","gqctNsqJIf_kGSZgwebvLq-zFuMWQebfRhbz4cUQr-s",{"id":5149,"title":5150,"author":7,"body":5151,"category":501,"date":5216,"description":5155,"extension":144,"featured":145,"image":5217,"meta":5218,"navigation":148,"path":5219,"seo":5220,"stem":5221,"tags":340,"__hash__":5222},"blog/blog/the-best-way-for-chatbots-to-actually-understand-your-customers.md","Blits: The Best Way for Chatbots to Actually Understand Your Customers",{"type":9,"value":5152,"toc":5209},[5153,5156,5160,5163,5166,5170,5173,5176,5179,5183,5186,5189,5193,5196,5199,5202,5206],[12,5154,5155],{},"At Blits.ai we're introducing a new method that reshapes how chatbots work. From the beginning, our company has been focusing on building a chatbot ecosystem that gives companies access to the best performing engines of the market. Our current offering is both a middleware layer and a low-code software platform, that gives our customers access to 40+ cognitive AI services.",[16,5157,5159],{"id":5158},"the-next-level-of-chatbot-automation","The Next Level of Chatbot Automation",[12,5161,5162],{},"Companies can easily create high performing and scalable chat and voice bot experiences with the Blits platform, without leaving our online drag and drop environment. This enables them to achieve maximum performance on any use-case in any language for their chat and voicebots.",[12,5164,5165],{},"In addition to our unique ecosystem, we are introducing the newest addition to our offering: Blits Automate.",[16,5167,5169],{"id":5168},"introducing-blits-automate","Introducing Blits Automate",[12,5171,5172],{},"Blits Automate helps companies automatically select the best underlying conversational AI engines for your chat and voicebots. Meaning you no longer have to search and optimize the performance of your bots across various market services in order to select an engine that understands your customers' questions.",[12,5174,5175],{},"Blits Automate saves time, improves customer satisfaction, and truly enables your company to focus on building conversations that add value to your business.",[12,5177,5178],{},"By automatically scanning your user conversational data and NLP model, our battle-tested algorithm Blits Automate selects the best engines fitted for your use-case and language. By continuous tracking of performance, customer input, and data validation, Blits Automate is able to switch to the best fit at any point in time, making sure your chat and voicebots are understanding your customers' needs, 24/7 and 365 days a year.",[16,5180,5182],{"id":5181},"a-use-case-of-how-blits-automate-adds-value","A Use-Case of How Blits Automate Adds Value",[12,5184,5185],{},"One of our customers has built a wholesale bot that is able to order and re-order food and beverages via WhatsApp messages. In order for the chatbot to correctly identify which amount, type, and brand the customers want to order, the bot has to fully understand all elements in the given sentences. This concept is known as entity detection and is part of every advanced chat or voicebot.",[12,5187,5188],{},"The challenge is that every entity engine works differently, and the results differ quite a lot. By using Blits Automate the platform was able to identify that Facebook's Wit.ai was more than 10x more effective in detecting specific entities compared to Google Dialogflow, IBM Watson, and Microsoft LUIS for this specific use-case. With these improved performance results, handling costs decreased substantially and customer satisfaction improved.",[16,5190,5192],{"id":5191},"no-one-size-fits-all","No One Size Fits All",[12,5194,5195],{},"Simply copying these findings from one chatbot to the next won't work. Every use-case is unique and requires a different set of AI engines working in the background. With Blits you no longer need to spend a lot of time selecting technology and testing the performance to find the optimum for your specific bot.",[12,5197,5198],{},"And if the field of AI changes and a better engine comes along, Blits will automatically apply the latest techniques to make your bot perform better than your competition, leading to less frustration with your customers, more satisfaction, and less cost of human intervention in customer processes.",[12,5200,5201],{},"Blits Automate works autonomously and is compatible with all engines in our platform, ranging from Microsoft, Google, IBM to Nuance and Stanford NLP. This is a must-have for companies that want to make sure their bots are ready for a high amount of users around the world.",[16,5203,5205],{"id":5204},"get-started-today","Get Started Today",[12,5207,5208],{},"Want to know more? Contact us to get a demo of the new Blits Automate, or sign up for a free account to get instant access to our ecosystem with 40+ cognitive and conversational AI services.",{"title":135,"searchDepth":136,"depth":136,"links":5210},[5211,5212,5213,5214,5215],{"id":5158,"depth":136,"text":5159},{"id":5168,"depth":136,"text":5169},{"id":5181,"depth":136,"text":5182},{"id":5191,"depth":136,"text":5192},{"id":5204,"depth":136,"text":5205},"2021-02-11T00:00:00.000Z","/blog/1611759270728.png",{},"/blog/the-best-way-for-chatbots-to-actually-understand-your-customers",{"title":5150,"description":5155},"blog/the-best-way-for-chatbots-to-actually-understand-your-customers","KiGLLsSkt3mwF28k-dZWbff5jIiVxbN7A0Mlpk42qmE",{"id":5224,"title":5225,"author":7,"body":5226,"category":333,"date":5458,"description":5459,"extension":144,"featured":145,"image":4809,"meta":5460,"navigation":148,"path":5461,"seo":5462,"stem":5463,"tags":5464,"__hash__":5467},"blog/blog/the-danger-of-ai-hallucinations-and-how-businesses-should-handle-it.md","The Danger of AI Hallucinations and How Businesses Should Handle It",{"type":9,"value":5227,"toc":5447},[5228,5231,5234,5237,5240,5247,5251,5254,5257,5271,5274,5278,5281,5295,5299,5302,5308,5311,5315,5318,5321,5324,5327,5331,5334,5357,5360,5364,5367,5370,5373,5376,5379,5386,5392,5396,5399,5431,5434,5438,5441,5444],[12,5229,5230],{},"Everyone is excited about AI productivity gains. Fair enough. The gains are real.",[12,5232,5233],{},"But if you are using AI in serious business workflows, there is one problem you cannot treat as a side note: hallucination.",[12,5235,5236],{},"A model can respond with full confidence, clean formatting, and a professional tone while still being wrong. That is exactly what makes this risk dangerous. It does not look like an error when you first read it.",[12,5238,5239],{},"In this article I will break down what hallucination is, where it creates business risk, how to reduce it, and why teams that manage this well actually move faster than teams that ignore it.",[179,5241,5242],{},[12,5243,5244,5246],{},[27,5245,2838],{}," Hallucination is not just a model quality issue. It is an operational risk issue.",[16,5248,5250],{"id":5249},"what-is-an-ai-hallucination","What is an AI hallucination?",[12,5252,5253],{},"A hallucination is when an AI model generates content that sounds plausible but is factually incorrect, fabricated, or unsupported by real data.",[12,5255,5256],{},"This can take different forms:",[21,5258,5259,5262,5265,5268],{},[24,5260,5261],{},"Inventing sources, links, quotes, or legal references",[24,5263,5264],{},"Returning outdated facts as if they are current",[24,5266,5267],{},"Misstating numbers, entities, or timelines",[24,5269,5270],{},"Filling gaps with assumptions when context is missing",[12,5272,5273],{},"The key point is simple: hallucination is not a \"bug you can patch once.\" It is a behavior pattern of probabilistic models. You need architecture and process around it.",[16,5275,5277],{"id":5276},"real-examples-that-hurt-businesses","Real examples that hurt businesses",[12,5279,5280],{},"The easiest way to underestimate hallucinations is to think only about consumer chat use. In enterprise environments, the impact is much bigger because output often drives decisions and operations.",[3026,5282,5283,5286,5289,5292],{},[24,5284,5285],{},"Imagine a customer support assistant that confidently explains the wrong refund policy. One wrong response can be fixed. Ten thousand wrong responses become operational debt, escalations, and churn.",[24,5287,5288],{},"Or take a sales enablement assistant that invents product capabilities in proposal text. It may help a team move faster this week, but it creates legal and trust problems once contracts are signed on incorrect assumptions.",[24,5290,5291],{},"In finance and operations, the risk is even more direct. If an AI summary tool misstates a KPI trend or attributes the wrong root cause, leadership decisions can be made on false signals.",[24,5293,5294],{},"In regulated sectors, hallucinated compliance advice is not just inaccurate, it can become a legal event.",[1784,5296,5298],{"id":5297},"a-simple-hallucination-example","A simple hallucination example",[12,5300,5301],{},"Below is a classic hallucination pattern: the model answers confidently, but the answer is impossible in the real world.",[12,5303,5304],{},[826,5305],{"alt":5306,"src":5307},"AI hallucination example showing confident but incorrect answer","/images/blits-hallucination-example.png",[12,5309,5310],{},"The question asks about crossing the English Channel entirely on foot. The model still returns a precise name, date, and duration, even though the premise is physically wrong. This is exactly why confidence is not the same as correctness.",[16,5312,5314],{"id":5313},"why-hallucinations-happen","Why hallucinations happen",[12,5316,5317],{},"Most business stakeholders ask: \"Why would a smart model make things up?\"",[12,5319,5320],{},"Because the model is optimized to generate the most probable next token sequence, not to guarantee factual correctness in your specific business context.",[12,5322,5323],{},"When context is weak, ambiguous, or missing, the model still tries to complete the task. If retrieval quality is poor, if prompts are vague, or if tools are unavailable, the model often fills the gap with plausible language.",[12,5325,5326],{},"That is why this is not just a model problem. It is a system problem: data quality, retrieval, prompt design, guardrails, and human review all matter.",[16,5328,5330],{"id":5329},"practical-tips-to-reduce-hallucinations","Practical tips to reduce hallucinations",[12,5332,5333],{},"You do not eliminate hallucinations fully. You reduce frequency and impact to acceptable business levels.",[3026,5335,5336,5339,5342,5345,5348,5351,5354],{},[24,5337,5338],{},"Start with grounded generation. Use retrieval-augmented flows so answers are tied to approved internal sources, not only to model memory.",[24,5340,5341],{},"Force citation behavior where relevant. If the model cannot point to source documents, it should explicitly say it is uncertain instead of inventing certainty.",[24,5343,5344],{},"Design prompts for bounded behavior. Ask the model to answer only from provided context and to refuse when evidence is insufficient.",[24,5346,5347],{},"Use tool calling for factual operations. For prices, account status, inventory, policy lookups, and transaction data, call systems of record instead of asking the model to \"remember.\"",[24,5349,5350],{},"Implement confidence and risk routing. Low-confidence or high-impact outputs should go through human review before being sent externally.",[24,5352,5353],{},"Measure hallucination rates as an operational metric. Treat it like any other quality KPI with test sets, regression checks, and release gates.",[24,5355,5356],{},"Separate use cases by risk class. A creative brainstorming assistant can tolerate far more uncertainty than a compliance or financial assistant.",[12,5358,5359],{},"At Blits, we mitigate this risk with guardrails for agents and dedicated validation layers for Agentic AI workflows before high-impact actions are executed.",[16,5361,5363],{"id":5362},"why-this-gets-riskier-in-agentic-and-automated-ai","Why this gets riskier in agentic and automated AI",[12,5365,5366],{},"In a normal chat interface, a hallucination usually becomes a bad answer. In agentic or automated AI, the same hallucination can become a bad action.",[12,5368,5369],{},"If an agent misreads policy context and still executes a workflow, it can send the wrong customer communication, trigger incorrect refunds, update records with false data, or escalate the wrong cases. The issue is no longer just content quality, but operational impact.",[12,5371,5372],{},"This risk grows when systems are fully autonomous, connected to multiple tools, and allowed to run at high speed. Small errors can cascade across APIs and business processes before a human notices.",[12,5374,5375],{},"That is why agentic systems need stronger controls than standard assistants: least-privilege tool access, approval gates for high-impact actions, sandbox testing before production rollout, and full audit trails for every decision and tool call.",[12,5377,5378],{},"The more automated the system, the more important it is to design for safe failure. Agents should pause, ask for clarification, or escalate to a human when confidence drops, instead of \"powering through\" uncertainty.",[179,5380,5381],{},[12,5382,5383],{},[27,5384,5385],{},"In agentic systems, hallucination risk shifts from wrong words to wrong actions.",[12,5387,5388,5391],{},[27,5389,5390],{},"At Blits, this is exactly why we implement guardrails for agents and validations for Agentic AI flows:"," to enforce boundaries, check outputs, and reduce the chance that uncertain model behavior turns into a real operational mistake.",[16,5393,5395],{"id":5394},"the-business-value-of-handling-hallucinations-correctly","The business value of handling hallucinations correctly",[12,5397,5398],{},"Some teams see hallucination control as a brake on innovation. In reality, it is the opposite.",[3026,5400,5401,5407,5413,5419,5425],{},[24,5402,5403,5406],{},[27,5404,5405],{},"Make outputs reliable."," Business teams start trusting automation.",[24,5408,5409,5412],{},[27,5410,5411],{},"Build trust."," Adoption increases across teams and channels.",[24,5414,5415,5418],{},[27,5416,5417],{},"Measure quality."," You scale use cases without multiplying risk and spend less time firefighting.",[24,5420,5421,5424],{},[27,5422,5423],{},"Clarify governance."," Procurement and legal move faster because risk posture is explicit.",[24,5426,5427,5430],{},[27,5428,5429],{},"Design for model switching."," You can adopt better models quickly without rebuilding from scratch.",[12,5432,5433],{},"In short, hallucination management is not only about avoiding mistakes. It is a competitive advantage in enterprise AI execution.",[16,5435,5437],{"id":5436},"a-practical-operating-principle","A practical operating principle",[12,5439,5440],{},"The question is not \"Can hallucinations happen?\" They can.",[12,5442,5443],{},"The better question is: what is your acceptable error threshold per use case, and what controls ensure you stay below it in production?",[12,5445,5446],{},"Teams that answer that question clearly build AI systems that are not only impressive in demos, but dependable in real business operations.",{"title":135,"searchDepth":136,"depth":136,"links":5448},[5449,5450,5453,5454,5455,5456,5457],{"id":5249,"depth":136,"text":5250},{"id":5276,"depth":136,"text":5277,"children":5451},[5452],{"id":5297,"depth":1869,"text":5298},{"id":5313,"depth":136,"text":5314},{"id":5329,"depth":136,"text":5330},{"id":5362,"depth":136,"text":5363},{"id":5394,"depth":136,"text":5395},{"id":5436,"depth":136,"text":5437},"2026-02-25T00:00:00.000Z","AI hallucinations are not just annoying errors, they are business risks. In this article I explain what hallucinations are, show real examples, share practical mitigation tips, and outline the business value of doing this right.",{},"/blog/the-danger-of-ai-hallucinations-and-how-businesses-should-handle-it",{"title":5225,"description":5459},"blog/the-danger-of-ai-hallucinations-and-how-businesses-should-handle-it",[5465,4218,5466,1101],"ai","risk management","Ev-FpJgtJRHU5lAfoNKMbDMJJxBmqtyx8K4pxLGIMSk",{"id":5469,"title":5470,"author":7,"body":5471,"category":333,"date":5778,"description":5779,"extension":144,"featured":148,"image":5780,"meta":5781,"navigation":148,"path":5782,"seo":5783,"stem":5784,"tags":5785,"__hash__":5786},"blog/blog/the-universal-commerce-protocol-and-why-llms-need-a-new-economic-language.md","The Universal Commerce Protocol and Why LLMs Need a New Economic Language",{"type":9,"value":5472,"toc":5764},[5473,5476,5479,5486,5489,5493,5496,5499,5502,5505,5509,5512,5515,5518,5523,5526,5529,5531,5534,5545,5548,5551,5562,5566,5569,5572,5575,5578,5589,5593,5598,5601,5616,5621,5624,5631,5635,5638,5641,5644,5647,5652,5656,5659,5666,5669,5672,5698,5702,5734,5738,5741,5744,5747,5750,5752,5755,5758,5761],[12,5474,5475],{},"Large language models have become surprisingly good at thinking. They can reason through complex problems, weigh options, and explain decisions better than most internal dashboards ever could. Yet the moment money enters the picture, everything slows down.",[12,5477,5478],{},"Ask an LLM to analyze vendors and it shines. Ask it to actually buy something and it immediately runs into walls. Redirects, checkout pages, confirmation steps, unclear pricing, hidden conditions. All of it is built for humans, not machines.",[12,5480,5481,5482,5485],{},"Yet everything in the ecosystem is moving toward a world where LLMs do place orders: automatically reordering supplies, booking travel, or buying software subscriptions from inside tools like Search, Gemini, or ChatGPT. The question is not ",[1121,5483,5484],{},"if"," agents will transact, but whether they will do so safely, transparently, and under proper governance.",[12,5487,5488],{},"This is where the Universal Commerce Protocol comes in.",[16,5490,5492],{"id":5491},"commerce-was-never-designed-for-machines","Commerce was never designed for machines",[12,5494,5495],{},"Modern commerce evolved around human behavior. We browse pages, read product descriptions, compare options, and click buttons. That entire flow assumes a person is sitting behind a screen, making subjective decisions and manually confirming intent.",[12,5497,5498],{},"LLMs do not browse. They plan.",[12,5500,5501],{},"When an LLM tries to interact with commerce today, it has to work around systems that were never meant for it. Developers scrape websites, glue together brittle APIs, or insert humans back into the loop to confirm every step. The result is fragile, slow, and hard to scale.",[12,5503,5504],{},"The real issue is not payments or security. The issue is representation. Commerce is not expressed in a way machines can truly understand.",[16,5506,5508],{"id":5507},"what-the-universal-commerce-protocol-changes","What the Universal Commerce Protocol changes",[12,5510,5511],{},"The Universal Commerce Protocol, usually shortened to UCP, does not try to replace ecommerce platforms. It changes how commerce is exposed.",[12,5513,5514],{},"Instead of pages, it exposes actions. Instead of vague descriptions, it exposes explicit terms. Instead of forcing interpretation, it provides structure.",[12,5516,5517],{},"A purchase is no longer something an LLM guesses its way through. It becomes a clearly defined action with known inputs, constraints, and outcomes.",[179,5519,5520],{},[12,5521,5522],{},"Commerce becomes something models can reason about first and execute second, instead of something they have to blindly click through.",[12,5524,5525],{},"This allows a model to reason before it acts. It can evaluate conditions, understand commitments, and decide whether executing an action makes sense in the context of a broader goal.",[12,5527,5528],{},"That shift may sound subtle, but it is foundational.",[16,5530,1621],{"id":1620},[12,5532,5533],{},"Consider three simple ecommerce cases:",[21,5535,5536,5539,5542],{},[24,5537,5538],{},"A shopping assistant is asked to “find a lightweight carry-on suitcase under $200 that fits airline X requirements.” With UCP, the retailer exposes “add suitcase Y to cart” as a structured action with explicit dimensions, price, delivery options, and return terms. The model does not scrape product pages; it evaluates a set of structured offers and chooses the best fit.",[24,5540,5541],{},"A retail agent helps a customer “replace my running shoes with the latest model in my size.” Instead of guessing through UI flows, it sees bookable actions like “purchase product Z in size 42” with stock status, loyalty benefits, and shipping times attached. It can optimize for price, delivery speed, or member perks in a transparent way.",[24,5543,5544],{},"A promotions agent manages “apply the best valid discount for this cart.” UCP exposes discount codes and promotions as structured capabilities, so the model can reason about which discounts stack, which require specific items, and what the final landed price will be before it commits to checkout.",[12,5546,5547],{},"In all of these examples, the important change is not the model. It is the language commerce uses to talk to the model.",[12,5549,5550],{},"Here is the simplest way to explain it to a business leader:",[21,5552,5553,5556,5559],{},[24,5554,5555],{},"A web page describes a product for a human.",[24,5557,5558],{},"A protocol describes an action for a machine.",[24,5560,5561],{},"The protocol is explicit about price, limits, approvals, and outcomes.",[16,5563,5565],{"id":5564},"from-browsing-to-planning","From browsing to planning",[12,5567,5568],{},"Humans browse first and decide later. LLMs do the opposite. They decide first and then look for the best way to execute that decision.",[12,5570,5571],{},"UCP aligns commerce with that mental model.",[12,5573,5574],{},"Instead of sending an LLM into a maze of pages and forms, it gives the model a structured view of what is possible. The model can compare options, reason about trade offs, and understand consequences without ever pretending to be a human user.",[12,5576,5577],{},"This dramatically reduces hallucination and error in one of the most sensitive domains imaginable.",[12,5579,5580,5581,5584,5585,5588],{},"We are already seeing this shift play out in how platforms standardize commerce for agents. Google’s Universal Commerce Protocol defines capabilities for shopping, checkout, and payments so agents in AI Mode on Search or Gemini can safely complete purchases using instruments like Google Pay, with every authorization backed by cryptographic proof of consent (",[1137,5582,1587],{"href":1585,"rel":5583},[1288],"). In parallel, OpenAI’s ",[1137,5586,1593],{"href":1591,"rel":5587},[1288]," initiative and the Agentic Commerce Protocol (ACP) describe how ChatGPT and other OpenAI-based agents can reason over structured commerce state, invoke tools for product discovery, checkout, and fulfillment, and keep customers informed in real time.",[1784,5590,5592],{"id":5591},"two-concrete-retail-examples","Two concrete retail examples",[12,5594,5595],{},[27,5596,5597],{},"Example 1: Conversational shopping and checkout",[12,5599,5600],{},"Today: A user asks “find a light-weight suitcase for an upcoming trip” in a search box or chatbot. Behind the scenes, systems scrape multiple retailers, guess at availability, and then hand the user a list of links and pages to click through.",[12,5602,5603,5604,733,5607,5610,5611,5615],{},"With UCP: Retailers expose capabilities like ",[1036,5605,5606],{},"dev.ucp.shopping.discovery",[1036,5608,5609],{},"dev.ucp.shopping.checkout",". An agent can discover business profiles, query structured product offers that match constraints (dimensions, price, loyalty benefits), and then invoke a checkout capability with explicit line items, totals, and payment options. As the ",[1137,5612,5614],{"href":1585,"rel":5613},[1288],"Google reference implementation of UCP"," shows, this powers experiences like AI Mode in Search and Gemini where users go from discovery to purchase in a single conversational flow, using instruments such as Google Pay with cryptographically provable consent.",[12,5617,5618],{},[27,5619,5620],{},"Example 2: Cart optimization and discounts",[12,5622,5623],{},"Today: Customers manually copy-paste discount codes, experiment with bundles, and hope the final total matches what they expected. Loyalty perks, member pricing, and coupon rules are spread across fine print and UI edge cases.",[12,5625,5626,5627,5630],{},"With UCP: Discounts become structured extensions on top of checkout. Agents can call a checkout session, then apply discount codes or loyalty benefits via explicit fields, receiving updated totals that show subtotals, discounts, taxes, and final amounts. This mirrors the discount application flow in the UCP samples, where an agent updates a checkout session with a code like ",[1036,5628,5629],{},"10OFF"," and receives a revised order with transparent allocations across line items and subtotals.",[16,5632,5634],{"id":5633},"why-this-matters-for-llms","Why this matters for LLMs",[12,5636,5637],{},"Without a protocol like UCP, LLMs are stuck as advisors. They can recommend but not operate.",[12,5639,5640],{},"Once commerce becomes machine readable, that changes. LLMs can move from suggesting actions to executing them in a controlled and auditable way. That requires stronger reasoning loops, better memory, and clearer boundaries, but those are exactly the directions model architectures are already moving in.",[12,5642,5643],{},"Commerce forces LLMs to grow up.",[12,5645,5646],{},"The practical implication is simple: intelligence and responsibility finally meet.",[179,5648,5649],{},[12,5650,5651],{},"When actions are explicit, governance stops being a spreadsheet problem and becomes a design choice.",[16,5653,5655],{"id":5654},"why-businesses-should-pay-attention","Why businesses should pay attention",[12,5657,5658],{},"This is not just about consumer chatbots ordering sneakers.",[12,5660,5661,5662,5665],{},"It is about operational automation ",[1121,5663,5664],{},"and"," the fact that LLMs are starting to execute real orders inside products like Gemini and ChatGPT, powered by protocols such as UCP and ACP. Procurement, travel management, vendor selection, subscription optimization, and yes, everyday ecommerce will increasingly be driven by agents acting on behalf of teams and customers.",[12,5667,5668],{},"When commerce is exposed as structured actions, those workflows — and the orders themselves — can be delegated safely to systems that reason consistently, enforce policy automatically, and never get tired.",[12,5670,5671],{},"For business leaders, the implications are concrete:",[21,5673,5674,5680,5686,5692],{},[24,5675,5676,5679],{},[27,5677,5678],{},"More control, less friction:"," Policies and constraints become executable rules rather than static PDFs people forget to read.",[24,5681,5682,5685],{},[27,5683,5684],{},"Higher-quality decisions at the edges:"," Small, frequent decisions (renewals, rebookings, minor purchases) are made consistently instead of ad hoc.",[24,5687,5688,5691],{},[27,5689,5690],{},"Better audit trails:"," Every action can be traced back to a structured intent with known inputs and outcomes.",[24,5693,5694,5697],{},[27,5695,5696],{},"A new way to compete:"," Companies that expose their services via UCP-style protocols become easier to integrate, easier to compare, and easier to automate around.",[1784,5699,5701],{"id":5700},"what-this-means-for-business-leaders","What this means for business leaders",[21,5703,5704,5710,5716,5722,5728],{},[24,5705,5706,5709],{},[27,5707,5708],{},"Lower friction",": Fewer manual steps, fewer handoffs, and less time stuck in approvals.",[24,5711,5712,5715],{},[27,5713,5714],{},"Better control",": Every action has explicit terms and guardrails, so the model can only do what is allowed.",[24,5717,5718,5721],{},[27,5719,5720],{},"Auditability",": Each decision and purchase is logged with inputs, constraints, and outcomes.",[24,5723,5724,5727],{},[27,5725,5726],{},"Scalability",": One set of rules can be applied across many vendors and regions.",[24,5729,5730,5733],{},[27,5731,5732],{},"Faster decisions",": Models can evaluate options instantly before spending.",[16,5735,5737],{"id":5736},"the-risks-and-misconceptions","The risks and misconceptions",[12,5739,5740],{},"UCP is not a shortcut around governance. It makes governance enforceable.",[12,5742,5743],{},"If you expose the wrong actions or set weak constraints, the model will execute them faithfully. That is not a model problem. It is a design problem.",[12,5745,5746],{},"The other misconception is that a protocol like this will \"flatten\" differentiation. In reality, it makes the differences between vendors more visible. Pricing, terms, service levels, and commitments are no longer buried in PDFs and fine print. They become compareable and testable.",[12,5748,5749],{},"For organizations that already operate with clear policies and well defined commercial logic, that is an advantage, not a threat.",[16,5751,1715],{"id":313},[12,5753,5754],{},"The Universal Commerce Protocol is not exciting in the way new models are exciting. It does not generate text or images. It does not feel intelligent.",[12,5756,5757],{},"But it is the missing layer between intelligence and execution.",[12,5759,5760],{},"Once that layer exists, LLMs stop talking about work and start doing it.",[12,5762,5763],{},"At Blits, we work with enterprises to design, implement, and govern these agentic commerce flows end‑to‑end—connecting protocols like UCP to your real systems so your agents can safely move from “recommend” to “execute.”",{"title":135,"searchDepth":136,"depth":136,"links":5765},[5766,5767,5768,5769,5772,5773,5776,5777],{"id":5491,"depth":136,"text":5492},{"id":5507,"depth":136,"text":5508},{"id":1620,"depth":136,"text":1621},{"id":5564,"depth":136,"text":5565,"children":5770},[5771],{"id":5591,"depth":1869,"text":5592},{"id":5633,"depth":136,"text":5634},{"id":5654,"depth":136,"text":5655,"children":5774},[5775],{"id":5700,"depth":1869,"text":5701},{"id":5736,"depth":136,"text":5737},{"id":313,"depth":136,"text":1715},"2026-01-19T00:00:00.000Z","Why commerce as we know it does not work for AI and what the Universal Commerce Protocol is trying to fix","/images/bannerUCP-blits.png",{},"/blog/the-universal-commerce-protocol-and-why-llms-need-a-new-economic-language",{"title":5470,"description":5779},"blog/the-universal-commerce-protocol-and-why-llms-need-a-new-economic-language",[],"xRTODaim9jKOOE-WBd44xI4OqCr9-OwaYAaZBQP8ZTk",{"id":5788,"title":5789,"author":7,"body":5790,"category":333,"date":5983,"description":5794,"extension":144,"featured":145,"image":5984,"meta":5985,"navigation":148,"path":5986,"seo":5987,"stem":5988,"tags":340,"__hash__":5989},"blog/blog/what-is-artificial-intelligence.md","At this point, most people are afraid to ask a simple question; what is artificial intelligence (AI)?",{"type":9,"value":5791,"toc":5976},[5792,5795,5798,5802,5805,5834,5837,5841,5844,5849,5869,5874,5879,5899,5904,5907,5911,5914,5939,5944,5947,5950,5953,5957,5960,5963,5966,5970,5973],[12,5793,5794],{},"Someone once told me, stupid questions don't exist. This might be true, but people are still afraid to look stupid by asking questions that are assumed to be general knowledge. We are currently at a point in time, where most people are assumed to know what is meant by the term Artificial Intelligence (AI) and how it works, but less is further from the truth. With the recent boost of generative AI services like ChatGPT and Midjourney, the spark is ever more lit, everyone loves to talk about it but I only know a hand full of people who actually know how it works.",[12,5796,5797],{},"In this article, my goal is to get everyone up-to-speed on the basics of AI and how it works in order to get you familiar enough with this concept so you can use and talk about this technology to your advantage.",[16,5799,5801],{"id":5800},"artificial-intelligence-is-everywhere-around-us-and-quite-likely-here-to-stay","Artificial Intelligence is everywhere around us and quite likely here to stay",[12,5803,5804],{},"With AI services rapidly spreading throughout all industries in the market, it's impossible to ignore this concept any longer. Furthermore, it's not going away anytime soon since our current living standards rely on it.",[21,5806,5807,5810,5813,5816,5819,5822,5825,5828,5831],{},[24,5808,5809],{},"It's everywhere and embedded in our daily life. It's used for your feed on Instagram, showing ads on Google searches, video selection on YouTube and Netflix, the best laundry program on your washing machine, fraud detection at your bank, or the daily groceries prices at your local supermarket, and many more.",[24,5811,5812],{},"Every CEO, CTO, or CIO keeps on talking about the importance of AI",[24,5814,5815],{},"Most companies have done some form of digital transformation to make sure they don't miss the wave of the next dot-com bubble and started training their workforce on these concepts.",[24,5817,5818],{},"A lot of new jobs now exist that didn't exist 10 years ago like Cloud Architect, UX/UI designer, Social Media Manager, or Virtual Reality Developer.",[24,5820,5821],{},"It has been on every manager's agenda one way or another to improve their business for a couple of years now.",[24,5823,5824],{},"Colleges and universities have created large curriculums around this topic, and are creating a huge new workforce of Data Scientists, AI Engineers, Prompt Engineers, Conversational Designers, and Machine Learning experts.",[24,5826,5827],{},"Investors are actively looking for startups that are able to build better AI models, products, and/or services.",[24,5829,5830],{},"An extreme amount of money yearly flows into technology where Gartner forecasts a spend of $4.5 trillion in 2023 in IT spending (growth of 2.4%).",[24,5832,5833],{},"Every day new apps are launched for consumers that use AI technology that is better than what's already available in the market.",[12,5835,5836],{},"All of these signs indicate that AI is not merely a passing fad or trend, but rather a technology that is here to stay and will undoubtedly continue to transform our lives.",[16,5838,5840],{"id":5839},"end-of-the-world-or-the-beginning-of-a-new-revolution","End of the world or the beginning of a new revolution?",[12,5842,5843],{},"Advocates and detractors of AI technology have differing perspectives on the benefits and risks associated with its development and integration into society. Here are some key arguments.",[12,5845,5846],{},[27,5847,5848],{},"Advocates:",[21,5850,5851,5857,5863],{},[24,5852,5853,5856],{},[27,5854,5855],{},"Economic and societal benefits:"," Proponents argue that AI can drive economic growth, create new job opportunities, and lead to societal advancements, such as in healthcare and environmental sustainability.",[24,5858,5859,5862],{},[27,5860,5861],{},"Enhanced decision-making:"," Advocates believe AI can improve human decision-making by analyzing vast amounts of data, identifying patterns, and providing actionable insights.",[24,5864,5865,5868],{},[27,5866,5867],{},"Creative potential:"," Proponents highlight AI's ability to enhance human creativity and generate novel solutions to complex problems.",[179,5870,5871],{},[12,5872,5873],{},"\"AI is the new electricity.\" - Andrew Ng",[12,5875,5876],{},[27,5877,5878],{},"Detractors:",[21,5880,5881,5887,5893],{},[24,5882,5883,5886],{},[27,5884,5885],{},"Job displacement:"," Critics argue that the widespread adoption of AI could lead to job loss, especially in sectors that rely heavily on manual labor and repetitive tasks.",[24,5888,5889,5892],{},[27,5890,5891],{},"Ethical concerns:"," Detractors raise concerns about AI's potential to be biased, discriminatory, or malicious, and the potential for misuse or abuse by humans.",[24,5894,5895,5898],{},[27,5896,5897],{},"Loss of human agency and control:"," Skeptics worry about the possibility of AI becoming too autonomous, leading to a loss of human control and potentially existential threats.",[179,5900,5901],{},[12,5902,5903],{},"\"AI will make jobs kind of pointless.\" - Elon Musk",[12,5905,5906],{},"These perspectives reflect the complex and nuanced nature of the AI debate, which encompasses a wide range of topics, from economic and societal simplifications to ethical and existential concerns. Personally, I think this a trend that can't be stopped but it should be regulated to mitigate the risks of ethics and control.",[16,5908,5910],{"id":5909},"ok-len-enough-context-tell-me-about-artificial-intelligence","Ok Len, enough context, tell me about Artificial Intelligence",[12,5912,5913],{},"Let's explain some basic concepts in Computer Science that might help to put all AI terms in perspective, as they all are used together.",[3026,5915,5916,5922,5927,5933],{},[24,5917,5918,5921],{},[27,5919,5920],{},"Algorithm:"," An algorithm is a step-by-step procedure or set of instructions for solving a specific problem or accomplishing a particular task. In computer science, algorithms are used to process, analyze, and manipulate data, and they form the foundation of all computer programs.",[24,5923,5924,5926],{},[27,5925,226],{}," Machine Learning is a subset of artificial intelligence (AI) that focuses on the development of algorithms and statistical models that enable computers to learn and improve from experience without being explicitly programmed. In other words, ML algorithms analyze data, identify patterns, and make decisions or predictions based on those patterns. Common ML techniques include supervised learning, unsupervised learning, and reinforcement learning.",[24,5928,5929,5932],{},[27,5930,5931],{},"Artificial Intelligence (AI):"," AI is a broad field of computer science that aims to create systems capable of performing tasks that would typically require human intelligence. These tasks include problem-solving, learning, planning, natural language understanding, perception, and even creativity.",[24,5934,5935,5938],{},[27,5936,5937],{},"General AI (Artificial General Intelligence or AGI):"," General AI, also known as Artificial General Intelligence, refers to AI systems that possess the ability to understand, learn, and apply knowledge across a wide range of tasks at a level comparable to human intelligence. Unlike narrow AI, which is designed to perform specific tasks, AGI can adapt and generalize its learning to new situations and challenges. AGI is still a hypothetical concept, and its realization remains an open challenge in AI research.",[179,5940,5941],{},[12,5942,5943],{},"\"The development of full artificial intelligence could spell the end of the human race.\" - Stephen Hawking",[12,5945,5946],{},"At its core, Artificial Intelligence (AI) is the creation of computer programs or machines that can perform tasks that typically require human intelligence. These tasks include learning from experience, reasoning, problem-solving, and understanding natural language.",[12,5948,5949],{},"But in itself, AI is a general term consisting of all techniques that help to achieve this goal. These AI models rely on advanced math and statistics. In addition, custom models can also be used for more specialized tasks, such as medical diagnosis, financial analysis, and even playing games like chess or Go.",[12,5951,5952],{},"The downside of most models is that they act as a black box, you can define parameters but it's impossible to retrace your steps. In other words, you can't easily retrace your steps to understand how the model arrived at a particular prediction. This can be problematic for several reasons. For instance, if the model is making incorrect predictions, it may be difficult to figure out why and make improvements.",[16,5954,5956],{"id":5955},"ok-give-me-a-real-world-use-case","Ok, give me a real-world use case",[12,5958,5959],{},"Netflix uses a sophisticated model recommendation system to suggest new movies and TV shows to users based on their viewing history and preferences. The model works by analyzing large amounts of user data, including which titles they've watched, how long they've watched them for, and how they've rated them. It then uses machine learning algorithms to identify patterns and correlations in this data and make predictions about what users might enjoy in the future.",[12,5961,5962],{},"This model is then used in your Netflix app with your data, creating an overview of what movies or series to watch next. Many applications today require these kinds of AI models to provide users with relevant updates, news, data, or features. Building these models can take some time, but are essential for every new service to be competitive in the market.",[12,5964,5965],{},"The better the model, the better the results. This is the reason why TikTok is so successful is all based on its successful AI model, which is feeding you new 'relevant' videos to keep watching and so out-competing other social media platforms.",[16,5967,5969],{"id":5968},"tldr","TL;DR",[12,5971,5972],{},"In summary, AI technology is based on the idea of creating computer programs or machines that can perform tasks typically requiring human intelligence. Machine learning, and more specifically deep learning, are popular techniques for achieving this goal, using artificial neural networks to learn from data and improve performance over time. With the continued advancement of AI technology, we can expect to see it integrated into an increasing number of applications and industries, transforming the way we live and work.",[12,5974,5975],{},"We are actively helping companies to implement these services in their business offerings in order to be competitive. Let me know if you need help.",{"title":135,"searchDepth":136,"depth":136,"links":5977},[5978,5979,5980,5981,5982],{"id":5800,"depth":136,"text":5801},{"id":5839,"depth":136,"text":5840},{"id":5909,"depth":136,"text":5910},{"id":5955,"depth":136,"text":5956},{"id":5968,"depth":136,"text":5969},"2023-03-28T00:00:00.000Z","/blog/1678649727454.jpeg",{},"/blog/what-is-artificial-intelligence",{"title":5789,"description":5794},"blog/what-is-artificial-intelligence","9Zxb4f_6I4QVI02pLAnNog4hjACMssJpy_IbT4YisZA",1775813889643]