
The GenAI hype machine promises everything: chatbots that understand context, customer experiences so personalized they feel telepathic, and operations so efficient they practically run themselves. However, here’s the uncomfortable truth: most enterprise AI projects fail in the pilot phase, stifled by integration nightmares, data chaos, and the harsh reality of aligning sophisticated models with complex business needs.
Durjoy Patranabish, vice president and head of global business at Tiger Analytics, and one whose team has shepherded GenAI from proof of concept to production for its customers, has seen this movie before. His company’s work with an Asian payment services firm and a U.S.-based innovator exploring Agentic AI reveals the real battleground where GenAI either thrives or dies: the unglamorous world of enterprise integration.
The information tsunami
The problem that GenAI solves isn’t sexy, but it is critical for data science: information fragmentation. Take the Asian payment services firm: their rule-based chatbot was essentially digital paperwork, forcing agents into manual archaeology expeditions across knowledge bases. The U.S. innovator? Their frontline workers were drowning in customer queries, wrestling with complex SOPs, and navigating merchandising guidelines that read like ancient scrolls.
“Newly onboarded workers had to spend a lot of time searching for information on these queries, which would be spread out across multiple systems, and the results that were thrown up could often end up being irrelevant and vague,” Patranabish explains. The result? A perfect storm of unpleasant customer experiences, skyrocketing training costs, and workers who would rather quit than search for answers.
Tiger Analytics knew that throwing GPT-4 at the problem was not the solution. The Asian payment services firm required a financial services chatbot for education and investment decisions and an Agent Copilot to reduce query resolution times. Meanwhile, the U.S. innovator sought an autonomous LLM agent that could process text and images, answering complex operational queries and acting as a digital companion for frontline workers. This agent was to integrate with the company’s mobile application, enabling workers to respond to customer queries quickly.
Is your GenAI POC stuck in the lab? Talk to Tiger Analytics at the Chief Digital & Data Officer Philippine Summit on July 23, 2025. To book your appointment, email Usha at [email protected].
Building trust in the age of hallucinations
Here’s why most GenAI deployments don’t make it out of the POC stage: lack of guardrails. In regulated sectors like financial services, one hallucinated response about competitor products could trigger compliance nightmares. So, the financial services chatbot needed bulletproof boundaries — no competitor queries, strict adherence to approved knowledge bases, and responses that were both concise and comprehensive.
Durjoy Patranabish @ Tiger Analytics: The company’s MLOps strategy focuses on modular architectures, robust MLOps pipelines, and continuous model evaluation paired with AI literacy and governance frameworks to future-proof deployments.
The Tiger Analytics team solved this through surgical prompt engineering and semantic search that retrieves only the “most relevant chunk” of data. For sensitive topics like competitor inquiries, they implemented what is essentially a digital bouncer: “Added guardrails in place to strictly adhere to the knowledge base and don’t answer queries specific to competitors. Implemented few-shot learning in prompt tuning to demonstrate proper handling of such topics and guiding users to ask relevant questions specific to client offerings.”
Think of it as building stored procedures with military-grade access controls. This means only permissible data ever makes it through.
The integration gauntlet
This is where data science theory meets enterprise reality, and the collision isn’t pretty. The financial services project encountered immediate roadblocks: integrating custom text analytics APIs for sentiment analysis with prompt flow, connecting to third-party contact center applications, and ensuring seamless integration without disrupting existing systems.
Tiger Analytics’s answer? Treat integration like building a nervous system for distributed microservices. They established a “robust data and API integration layer through API gateway and management with clear and consistent API contracts.” The architecture philosophy was radical in its simplicity: “We designed the integration components and the GenAI solution as loosely coupled, independent modules or microservices.”
This is more than just sound engineering. Tightly coupled systems in enterprise environments can be future hurdles. One upstream change can cascade through the entire system, taking down production services and generating the kind of 3 am phone calls that end careers.
Data security added another layer of complexity. With sensitive customer information flowing through GenAI models, they “implemented techniques to mask personal data before it’s used by GenAI models,” then subjected everything to rigorous testing—unit tests, integration tests, end-to-end testing — the whole nine yards.
The model selection crucible
Choosing the right LLM for enterprise deployment is like selecting a Formula 1 engine: raw power means nothing if it can’t handle the specific track conditions. The financial services chatbot leveraged Azure OpenAI models, such as GPT-3.5 or 4, while the Agent Companion ran on Gemini 1.5. The decision criteria were brutally practical: “Reasoning capability, Latency, and request tokens are some of the key requirements for an LLM to decide. Hence, we selected GPT4, which fits the requirement.”
But the model selection is just the beginning. Real performance comes from prompt tuning, which is the art of coaxing optimal responses from these digital oracles. The process involves “prompt tuning… to ensure optimal generation of results — intent, response, recommendation, and summarization.” It’s like training a world-class athlete: natural talent gets you in the door, but precision training determines who wins.
Multi-modal complexity at scale
The Agent Companion project pushed into truly complex territory: a multi-modal autonomous agent solution that manages five specialized agents. Each agent owns its domain: the Knowledge Agent for SOPs, the Product Agent for descriptions, the Device Agent for troubleshooting, the Merchandising Agent for shelf analysis, and the Sales Agent for the buying journey. “Each agent is an expert in its domain, having been trained and fine-tuned independently with data and documents specific to that area, which includes multi-modal documents,” Patranabish clarifies.
The data ingestion pipeline resembled building a massive data lake in real-time, comprising 142 SOP documents, 67 product data files, 65 knowledge articles and URLs, 36 product manuals, and 31 technical documents and URLs. All of this had to be chunked, processed, classified, and vectorized for semantic search while maintaining high accuracy, reliability, quick response times, and cost efficiency.
The human problem
The biggest deployment challenges are often not technical, but rather human. Patranabish identifies the “common organizational ‘gaps’” that kill GenAI projects: “lack of change management, unclear AI governance, and limited employee readiness.” Many organizations also “struggle with aligning GenAI use cases to real business needs.”
Tiger Analytics tackles this through intensive stakeholder education, cross-functional collaboration, and co-creating adoption roadmaps. They run workshops to build trust, demystify GenAI, and establish responsible use practices. The most elegant technical solution becomes worthless if the people who need to use it don’t understand it, trust it, or see its value.
The MLOps imperative
The future of GenAI deployment hinges on solving what Patranabish calls the next wave of challenges: “model scalability, data privacy, hallucinations, and continuous governance.” As GenAI adoption expands, maintaining performance and ethical alignment across diverse teams becomes exponentially complex.
Tiger Analytics’s strategy focuses on “modular architectures, robust MLOps pipelines, and continuous model evaluation,” paired with “AI literacy and governance frameworks to future-proof deployments.” Just as DevOps revolutionized software development by integrating operations and development, MLOps is becoming the make-or-break factor for machine learning models in production.
The companies that master this integration — technical excellence, human readiness, and operational discipline — won’t just deploy GenAI successfully. They’ll redefine what’s possible in enterprise AI, transforming experimental technologies into indispensable business infrastructure. Ultimately, the GenAI deployment challenge isn’t about building smarter models. As Tiger Analytics showed, it’s about building smarter organizations that can effectively harness these models. The technology is ready. The question is: Are our organizations ready to embrace it?
Image credit: iStockphoto/Prostock-Studio