• Home
    • Case Studies
    • COMPASS LENS
  • Why Compass
    • Competitive Advantages
    • Enterprise-Grade Security
    • ROI Calculator
    • FAQ
  • About Us
    • CC-Expertise Ltd
    • Privacy Policy
    • Terms of Service
  • Our Blog
  • Contact Us
  • Home
    • Case Studies
    • COMPASS LENS
  • Why Compass
    • Competitive Advantages
    • Enterprise-Grade Security
    • ROI Calculator
    • FAQ
  • About Us
    • CC-Expertise Ltd
    • Privacy Policy
    • Terms of Service
  • Our Blog
  • Contact Us
Request Free Trial
  • Home
    • Case Studies
    • COMPASS LENS
  • Why Compass
    • Competitive Advantages
    • Enterprise-Grade Security
    • ROI Calculator
    • FAQ
  • About Us
    • CC-Expertise Ltd
    • Privacy Policy
    • Terms of Service
  • Our Blog
  • Contact Us
Request a Demo

There’s More AI on the Tin Than in the Can

Contact center AI is everywhere, in pitch decks, demos, and roadmaps. But too often, what’s sold as “intelligence” turns out to be automation wrapped in branding.

This 3-part series started on LinkedIn as a call for honesty and substance. It outlines 24 red flags I’ve seen firsthand:

  • Tools that overpromise and underdeliver
  • AI that helps marketing, but not operations
  • Gaps between evaluation, automation, and outcomes

If you’re a CX leader, a platform buyer, or just tired of AI hype with no ROI, this is for you.

Part 1: Behind the Gloss :
Marketing Smoke vs. Machine Intelligence

AI is everywhere, at least in pitch decks. But behind the buzzwords, too many “AI-powered” platforms are patchworks of rule-based logic dressed up as intelligence. In this first part of our reality check on contact center AI, we examine 8 signs your platform may be more branding than brains:

  1. “Our own AI engine”: sounds impressive, means very little: Scratch the surface of most “proprietary AI” claims, and you’ll find a patchwork of open-source NLP, pre-trained models, and rule-based wrappers. Ask for specifics, and the answers get vague “cloud-native,” “built on Bedrock,” “leveraging LLMs”, with no clarity on what’s actually under the hood. If the engine is truly proprietary, it should show in performance, not just pitch decks. Most of the time? It’s marketing smoke, not machine intelligence.

 

  1. Basic summaries that ignore context : Conversations with multiple agents are still summarized as if they were handled by one voice, one person, one brain. No handoff insight. No agent-level accountability. No understanding of who did what. Just a generic paragraph that tells you a call happened, but not how it unfolded. That’s not a summary, it’s a gloss-over.

 

  1. Empathy and satisfaction scores without transparency : Vendors show you a number, but offer no trail, no justification, no context. You don’t know why the score was high or low, which part of the conversation triggered it, or how to act on it. It’s AI without accountability and you can’t coach what you can’t trust.

 

  1. Outdated sentiment analysis: Tagging each sentence as “positive” or “negative” is a 2010 solution in a 2025 world. Real conversations carry emotions, not isolated tones. Where’s the full-journey sentiment view? The frustration buildup, the confusion spike, the relief at resolution? Most systems still miss the forest and the emotion, for the trees.

 

  1. Automation stuck in limbo: Wrap-up code suggestions. Score recommendations. Pre-filled forms. These aren’t breakthroughs, they’re band-aids. Automation that still relies on human final action isn’t “intelligent”, it’s just AI-assisted manual labor. Without autonomy, you’re just dressing up micromanagement as innovation.

 

  1. Topic analysis with word matching, not understanding : Most systems still match keywords to sentences like it’s a glorified “Find & Replace.” But customers don’t talk in isolated phrases, they share narratives. True topic modeling should follow the arc of the conversation, not just tag sentences in isolation. Understanding beats detection. Yet for many tools, that shift hasn’t arrived.

 

  1. No real conversation intelligence: AI today can predict churn, detect escalation risk, and surface upsell opportunities, not in theory, but in practice. Yet most platforms barely scratch the surface. Instead of analyzing full conversations across context, emotion, and intent, they offer shallow summaries and sentiment tags. Where’s the proactive insight? Where’s the action trigger? If your AI can’t highlight a frustrated VIP customer about to leave, or a product issue mentioned five times in a week, it’s not conversation intelligence. It’s just transcription with branding.

 

  1. Multilingual AI that doesn’t understand culture: Many platforms boast language coverage, but real understanding requires more than translation. Tone, politeness markers, emotion, and escalation triggers vary across cultures. Yet most systems treat every language like English with a thesaurus. Global support deserves global intelligence.

Part 2: From Assistive to Autonomous :
Where AI Fails to Deliver on Action

Not all automation is created equal. Many platforms today confuse suggestion with solution, and assistance with autonomy. In this second part of our AI reality check, let’s look at how AI tools are failing to deliver actual business outcomes, despite claiming to.

  1. Survey fatigue, not smart targeting : Surveys are blasted out by rule, not reason. Every customer, every time, regardless of how the call went. There’s no AI asking: Was this conversation worth surveying? Was the customer emotionally ready? Did we resolve anything at all? The result? Lower response rates, biased feedback, and frustrated customers. Intelligent systems should tailor survey timing and delivery based on conversation outcomes, not checkboxes.

 

  1. “New” metrics that aren’t new : CSAT and NPS have been around for decades. Vendors may repackage them or tweak the scale, but where are the metrics that reflect today’s multichannel, emotionally nuanced, AI-augmented journeys? Emerging metrics like a CX Index (a composite score blending sentiment, escalation risk, responsiveness, and resolution effectiveness) or Customer Voice Index (derived from qualitative feedback, intent volatility, emotion shifts, and agent empathy levels) point to a more accurate picture of experience. You could even imagine Agent Impact Scores, which track how each agent influences customer outcomes over time, not just by resolution, but by tone, empathy, and intent coverage. These aren’t hypothetical, they’re technically feasible. But most platforms stop at vanity scores instead of helping contact centers measure what truly matters.

 

  1. Security theatre without intelligence: Compliance checklists are one thing, proactive, intelligent security is another. While vendors claim enterprise-grade security, few offer AI-powered breach detection, real-time anomaly alerts, or adaptive threat modeling based on behavioral patterns. In an age where conversations include sensitive data, AI should be guarding the gates, not just filing the paperwork.

 

  1. “AI-based routing” that’s not AI at all: Slapping “AI” on static routing logic doesn’t make it intelligent. Most so-called “AI routing” is just rebranded rule trees or weight-based prioritization. True adaptive routing should consider intent, sentiment, urgency, historical context, and even real-time agent fit. Instead, we get if-else logic with a marketing badge. If the routing doesn’t learn or adapt, it’s not AI, it’s automation in disguise.

 

  1. Chatbots in disguise: “Agentic AI” promises autonomy and reasoning, but most chatbots are still glorified forms with pre-scripted branches. Yes, the interface may use a sleek LLM prompt here or there, but the system lacks memory, adaptability, or contextual awareness. If it can’t handle ambiguity, switch domains, or escalate meaningfully, then it’s not a co-pilot, it’s a tour guide with a set route.

 

  1. Voice bots that barely clear the bar: “AI-powered voicebots” often translate to NLU-driven menus with LLMs generating utterances, but little more. A typical setup: You manage intents, add/edit utterances, and manually assign keywords to slots. Yes, the interface might be slick. But if your voicebot needs constant human babysitting to handle new topics or changes, it’s not agentic, it’s just dressed-up scripting. True AI should listen, learn, and adapt, not make you micromanage phrases one-by-one.

 

  1. Agent copilots that overpromise and underdeliver : Everyone’s touting an “agent copilot” these days, but most are little more than dressed-up assistants. They pull snippets from static knowledge bases, surface pre-scripted suggestions, and function best when agents already know what to ask. Step outside their narrow logic, or worse, throw in a bit of nuance, ambiguity, or real-world chaos, and they stall.

These so-called copilots often can’t:

  • Adjust guidance based on customer sentiment or conversation tone.
  • Recall context from earlier in the interaction or previous conversations.
  • Suggest smarter paths when the knowledge base is outdated or incomplete.
  • Adapt on the fly to changing policies, promotions, or scenarios.

The result? Agents spend more time re-asking, correcting, or ignoring AI “help” than benefiting from it. Far from boosting performance, many of these copilots create friction, offering partial answers that need human babysitting, or repeating the same broken advice.

A true copilot should do more than respond, it should anticipate, contextualize, and evolve. That means learning from interactions, adapting to business changes, and supporting agents through complex scenarios, not just handing over FAQs in a shiny wrapper.

If your “copilot” needs a pilot, you’re not flying, you’re fixing.

  1. Insights and analytics… sort of: Most platforms promise “AI-powered analytics” but deliver basic reporting: bar charts of sentiment, lists of top topics, and dashboards that describe what happened, never why. There’s no correlation across variables, no guided investigation, no real decision support. AI-powered analysis should go much further: correlating customer satisfaction with underlying variables like silence or overtalk time, agent empathy, specific conversation topics, or escalation trends. Sadly, most tools stop at surface-level reporting.

True AI-powered analysis should connect the dots:

  • How does customer satisfaction shift with silence or overtalk time?
  • Which topics trigger escalations more often?
  • Where does low agent empathy correlate with churn risk?
  • What are the root causes of negative CX trends, not just their symptoms?

These aren’t moonshot use cases, they’re feasible today. Yet most tools avoid them entirely. Few solution offer guided investigation, root cause identification, or even basic multi-variable correlation across channels, sentiment, or agent behavior. True conversational intelligence, root cause analysis, or guided decision-making is still the exception, not the rule.

If your analytics tool can’t explain the why or suggest what next, it’s not AI. It’s just a rear-view mirror with pretty formatting.

Part 3: The Human Factor :
When “AI” Still Leaves You Doing the Work

AI isn’t just about automating tasks, it’s about amplifying value. But too many platforms use “human-in-the-loop” as an excuse for incomplete automation. In this final part, we look at how evaluation, compliance, and transparency are still stuck in the past, despite all the AI talk.

 

  1. Human-in-the-loop? Misused as a crutch, not a safeguard: Many platforms parade “human-in-the-loop” as proof of responsible AI, but in reality, it’s often a cover for incomplete automation. Instead of empowering oversight, it becomes a manual checkpoint for every output, turning QA teams into babysitters rather than strategic reviewers. But true AI calibration doesn’t need humans to validate every single evaluation. Techniques like sampling, trend-based reviews, and dispute analytics can monitor AI performance at scale. Human-in-the-loop should exist for control, escalation, and exception handling, not to manually score what AI should already be handling end-to-end.

 

  1. Compliance monitoring that’s manual, reactive, or nonexistent: Most platforms still treat compliance as a checkbox exercise, hidden in QA forms or sampled through manual reviews. But policy breaches, disclosure misses, and regulatory violations don’t wait for audits. AI should be systematically reviewing all conversations, flagging potential compliance issues based on configurable rules, not just random samples or agent hunches. Whether it’s missing a required statement or using prohibited language, breaches should be detected consistently, logged transparently, and surfaced through usable channels and integrated alerts, not buried in reports. If your platform can’t automatically catch what your agents are legally obligated to say (or not say), you’re not managing risk, you’re just crossing fingers.

 

  1. Where are the enhancement recommendations? True AI analysis should go beyond describing the problem, it should suggest how to fix it. Whether it’s automating repetitive tasks, recommending process tweaks, or shifting volume to other channels, prescriptive AI is the missing layer in most products today. Real AI use cases should include providing efficiency tips to agents and supervisors, recommending automation opportunities where human effort is wasted, validating whether the contact channel was appropriate or if the interaction could have been deflected to a more suitable one (e.g., chat instead of voice), and analyzing if and how the conversation could have been resolved in self-service mode. It should also highlight what enhancements, flows, intents, or knowledge, are needed to make that self-service viable. Without these forward-looking recommendations, AI isn’t helping you improve; it’s just narrating the status quo.

 

  1. Failed Evaluations Scored but Not Solved, and Missed Opportunities to Leverage All Evaluations for Agent PEPs: Some tools can detect that a call failed quality standards. A few might even justify the failure. But almost none tell you how to prevent it next time.

There’s no proactive coaching, no guided remediation, no structured path to improve. Why stop at identifying the problem when AI can help close the loop? Take it further, use both failed and successful evaluations to automatically generate Personal Enhancement Plans (PEPs) tailored for each agent. What’s typically done once a quarter can be automated weekly using AI, drawing from real evaluated conversations.

These PEPs don’t have to be dull or generic, AI can generate them in different formats based on the quality manager’s preference: a high-level summary, a detailed report, a motivational narrative, a Q&A-style breakdown, a case study format, or even a role-play walkthrough. With just a few clicks, quality managers can deliver consistent, insightful coaching content. Their true value isn’t in tedious data crunching, it’s in inspiring, developing, and empowering the organization’s most valuable asset: its agents.

 

  1. Manual scoring disguised as AI “support”: Why settle for “30% time saved” claims when you can automate 100% of evaluations, without giving up fairness, control, or human insight? Too many platforms use human-in-the-loop as a crutch, offering score suggestions while still expecting QA managers to review, finalize, and document feedback. That’s not automation, it’s assistive labor.

Worse still, some vendors continue to tout features like manual recording annotation as if they’re cutting-edge, when in reality, this was impressive in 2013—not in an era of large-scale, real-time conversation analysis. Labeling timestamps and marking issues by hand is not intelligence, it’s admin work.

True evaluation automation means every conversation is scored, issues are flagged with justifications, and next steps are proposed, all at scale. Human input should be there for coaching, calibration, and challenge resolution, not to compensate for missing capability. If your platform can’t evaluate on its own and still leans on humans to do the heavy lifting, it’s not AI. It’s a to-do list.

 

  1. AI should enhance your service, not your brand image: Too many vendors wear AI like a marketing pin, flashy in the deck, absent in the delivery. They talk about “transformative intelligence,” yet limit its presence to a single dashboard tile or demo-only assistant. Ask where AI is actually embedded, and the list gets thin: no coaching triggers, no root cause detection, no smart alerts, no performance nudges.

Want examples?

  • Agent empathy is scored, but not used to personalize routing, trigger targeted coaching, or anticipate churn.
  • Escalations are everywhere, yet few AI systems detect them proactively, let alone escalate through messaging platforms instead of flooding inboxes. Where’s the workflow?
  • Conversation summaries are auto-generated, but lack deeper analysis, improvement recommendations, or actionable insight.
  • Compliance breaches may be detected, but they’re rarely flagged systematically or routed to the right teams beyond the CCaaS ecosystem.
  • “Human-in-the-loop” is used as a shield, as if full automation and human oversight are mutually exclusive. They’re not. You can automate 100% and still retain control, calibration, and fairness.

These aren’t just missed opportunities, they’re missed outcomes. AI should sit at the heart of your service model, not buried in a demo folder. It should nudge, warn, summarize, suggest, and act, not just decorate your pitch. If the AI doesn’t make your operations smarter, faster, or more adaptive, then it’s not a capability. It’s just cosmetics.

 

  1. “Fair use” that’s barely usable , and token pricing that’s barely fair: Many vendors include so-called “fair use” tiers as part of their AI packaging, but don’t be fooled. These often cover less than 5–10% of actual customer usage needs, serving more as a teaser than a solution. Real-world workloads quickly exceed these thresholds, triggering expensive overage fees or premium add-ons.

To make matters worse, every vendor seems to have their own definition of what a “token” is. When it aligns with standard LLM input/output metrics, that’s fair. But when “tokens” become a fuzzy abstraction, used to mask pricing complexity or inflate costs, it’s no longer about transparency. It’s just buzzword billing. Customers suddenly find themselves facing bundles of features that don’t even leverage AI but are still included under the “AI bundle,” now subject to metered usage that quietly drives up the bottom-line cost. And then vendors wonder why AI adoption isn’t picking up. If your pricing model punishes usage, don’t be surprised when users hold back.

Compare that to platforms or solutions where 70%+ of real-world usage is covered transparently under standard policies, where tokens actually reflect real AI usage and don’t cost you an arm and a leg. That’s the difference between AI access and AI theatre

  1. Locked-in AI that’s not optional : Perhaps the biggest red flag: vendors increasingly bundle AI as a mandatory part of their platforms. If the AI is truly world-class and competitive, why force customers to pay before they can evaluate its worth? Lock-in should never be a substitute for value. Just because a tool comes with a Microsoft, AWS, or Google badge doesn’t mean it’s the best fit for every use case. Brand strength should never override solution fit, capability depth, or actual results. In the end.

Great AI earns its place, it shouldn’t have to be bundled, let alone enforced and disguised as mandatory in your platform licenses, as if it were the only fuel to power your engines.

Wrap up : Bringing the Hype Back to Earth

Let’s Call It What It Is.

Many of today’s “AI-powered” platforms are built to win deals, not to deliver outcomes. They prioritize checkboxes over clarity and optics over substance. The result? Higher costs, longer adoption curves, and disillusioned teams who expected intelligence but got automation in a new outfit.

It’s not just vendors.

Even internal enterprise IT departments are sometimes swept up in the AI hype, confidently committing to solve “all use cases” while struggling to define where to start, how to prioritize, or what meaningful value looks like. This challenge deserves its own spotlight, and we’ll be diving deeper into that in a follow-up post soon.

If you’re a leader evaluating AI in your organization, especially in the contact center space, take your time. Ask hard questions. Look under the hood. Don’t be distracted by the noise. Real AI is measurable, explainable, and adaptive. Anything less is just marketing.

Don’t settle for AI that looks good in a deck but falls short in production. If it doesn’t reduce effort, improve outcomes, or deliver measurable ROI, it’s not AI for the business.

You deserve to get more than what’s written on the tin.

We’ve covered 24 signs your AI platform might be more sizzle than substance. Some are subtle. Others are systemic. But none are unsolvable.

True AI should reduce effort, drive action, and deliver measurable improvement, not just decorate your roadmap.

Your Turn or Join the Conversation:

What AI features have actually delivered value in your experience?

Or what challenges have you faced in adopting AI tools that lived up to the hype?

What’s missing?

Let’s compare notes and bring the hype down to earth.


Efficiently Managing AI Costs: The Power of Real Token-Based Billing

Related Articles


Efficiently Managing AI Costs: The Power of Real Token-Based Billing
Dashboards Are Dead. Long Live Conversational Analytics
Dashboards Are Dead. Long Live Conversational Analytics
Previous Article

CC Expertise Ltd

Chancery Station House, 31-33 High Holborn,

London WC1V 6AX

UK

© Copyright CC Expertise Ltd

Company

Our Blog
About Us
Contact Us