OTP Bank Case Study | Ender Turing
OTP BANK
Banking 11 countries 4,200 agents
Customer Success Story

From sampling 3% of calls to analyzing every conversation

How Central and Eastern Europe's largest banking group rebuilt its quality programme on Ender Turing — covering eight languages, eliminating QA backlog, and surfacing compliance risks in hours instead of weeks.

IndustryBanking & Financial Services
Headcount served4,200 agents · 320 team leads
LanguagesHungarian, Romanian, Bulgarian, Serbian, Croatian, Slovenian, Russian, English
Products deployedSpeech Analytics + Quality Management
IntegrationsGenesys, in-house CRM, Microsoft Power BI
Time to first value28 days from kickoff
📈
Outcomes after 9 months

Measured against the pre-deployment baseline

100%
Calls automatically scored, up from 3%
+18
Customer satisfaction points (CSAT)
−22%
Average handle time across retail desk
+14%
First contact resolution
68h
QA hours saved per specialist per month
The Challenge

Quality at a 4,200-agent scale was being run on samples and spreadsheets

OTP Bank's contact-center quality team was reviewing roughly 3% of calls each month by hand — barely enough to spot recurring issues, and far too late to coach agents while a problem was still fresh.

The team had three structural problems to solve in parallel: scoring at scale, multi-language consistency, and getting actionable signal back to team leads in days, not weeks.

!

Sampling missed almost everything

QA reviewed ~3% of monthly call volume. Recurring compliance gaps and misselling patterns surfaced only during quarterly audits.

!

Eight languages, eight different rubrics

Each subsidiary kept its own scorecard in Excel. Cross-country benchmarking was effectively impossible for senior leadership.

!

Coaching loop was too slow

Average lag between a problem call and the coaching session was 11–14 days. By the time an agent heard the feedback, the moment was gone.

!

Compliance signal was reactive, not proactive

Mandatory disclosures (cooling-off rights, APR statements, data-processing consent) were only verified post-hoc — exposing the bank to regulator findings.

The Solution
🎙️

Real-time transcription in 8 languages

Every inbound and outbound call transcribed end-to-end with speaker separation, diarization, and entity extraction (account numbers, product names, regulatory phrases).

Unified quality scorecard

One template scoring 18 criteria across greeting, identity verification, problem framing, solution offered, empathy, compliance disclosures, and closing — applied consistently across all subsidiaries.

Compliance keyword alerts

Custom rules trigger flags within minutes of call end when mandatory disclosures are missed, sensitive topics arise, or risk language appears.

📊

Manager dashboards by subsidiary

Localized dashboards for each market plus a group-level rollup. Team leads see queue health, top issues, and coaching candidates without leaving their workflow.

🤝

Agent self-review portal

Every agent gets weekly auto-scored calls in their own language with playback timestamps for each evaluation criterion — making feedback specific, not abstract.

One platform, eight languages, every conversation scored

Working with OTP Bank's central operations team, Ender Turing replaced the spreadsheet-based QA process with a single platform connected to Genesys across all subsidiaries and the bank's in-house CRM.

The rollout focused on parity over reinvention: instead of replacing the existing 18-criterion scorecard, Ender Turing's QM module was configured to score that exact rubric automatically — letting team leads keep their existing language while getting coverage across 100% of calls.

For the markets where the bank operates regulated lending (Hungary, Romania, Bulgaria), real-time disclosure detection was built on top, surfacing compliance gaps within minutes of a call ending rather than during a quarterly audit cycle.

Implementation

From kickoff to first measured win in 12 weeks

OTP Bank ran the rollout in four phases. Each phase ended with a measurable gate — no phase started until the previous one was signed off by the operations and compliance teams.

1
Weeks 1–3

Integration & Data Pipeline

Connected Genesys SIPREC streams across three subsidiaries. Mapped CRM metadata. Verified language coverage on a 50,000-call replay set.

GenesysCRM8 langs
2
Week 4

Scorecard Configuration

Translated the bank's existing 18-criterion rubric into ML-trained classifiers. Calibrated against 1,200 manually-scored calls until inter-rater agreement hit 92%.

QA rubricCalibration
3
Weeks 5–8

Pilot in Hungary & Romania

320 agents on the retail line moved to fully-automated scoring. Daily standups with QA leads. Compliance-rule library expanded to 47 patterns.

320 agentsDaily review
4
Weeks 9–12

Group-Wide Rollout

Remaining six subsidiaries onboarded in 4 staggered waves. Group-level dashboard delivered to executive sponsors. First quarterly review held in week 14.

4,200 agents11 countries
The Results

Nine months in, the numbers speak for themselves

All figures measured against the pre-deployment baseline. Methodology and per-market breakdowns available on request.

🎯
+33×
100%
Calls automatically scored
Up from a manual sampling rate of 3%. Every customer interaction now produces a structured QA scorecard, in every language, within minutes of the call ending.
Baseline: 3% of monthly call volume
😊
+18 pts
+18%
Lift in customer satisfaction
Tracked via post-call surveys across all subsidiaries. Largest gains came from retail lending and card-services queues — the lines with the highest coaching activity.
Baseline CSAT: 6.8 → 8.0 (10-point scale)
−22%
−22%
Reduction in average handle time
Driven by faster problem framing and fewer escalations. Agents armed with weekly self-review insights resolve known patterns more confidently.
Baseline AHT: 7m 41s → 6m 00s
🛡️
−74%
−74%
Compliance gap detection lag
From an average of 14 days during quarterly audits to under 4 hours via automated keyword alerts. Risk & Compliance can now act on the same business day.
Baseline lag: 11–14 days
⏱️
+68h/mo
68h
QA hours saved per specialist per month
QA specialists redirected from manual scoring into deeper investigative work and structured agent coaching. Coaching sessions per month doubled with the same headcount.
28-person QA team across the group
🎓
+14%
+14%
First-contact resolution
As coaching feedback reached agents within days of the originating call (instead of weeks), known issues stopped recurring on follow-up contacts.
Baseline FCR: 71% → 85%

Before Ender Turing

  • Manual scoring of ~3% of monthly calls
  • 8 different rubrics across 8 languages
  • 11–14 day delay between issue and coaching
  • Quarterly compliance audits, post-hoc
  • QA specialists buried in scoring backlog
  • No group-level visibility for executives

After Ender Turing

  • 100% automatic scoring across all calls
  • One unified scorecard, language-aware
  • Coaching loop measured in days, not weeks
  • Compliance gaps surfaced within hours
  • QA team focused on coaching and trends
  • Group dashboard live for the executive board
★★★★★

"With Ender Turing, we moved from sampling a few hundred calls per month to analyzing every single conversation. Our coaches now spend their time on what actually matters — helping agents grow, not searching for problems. The shift in how we run quality has been one of the biggest operational wins of the past year."

KM
Krisztina Molnár
Head of Contact Center Operations · OTP Bank Group
Inside the deployment

What changed inside the QA team's daily workflow

Beyond the headline numbers, the day-to-day shape of the QA function changed materially. Manual call scoring used to consume roughly 70% of QA-specialist time. Today, that share is closer to 8% — limited to dispute reviews and edge-case investigations.

The freed capacity went into three places: structured coaching sessions doubled month-over-month; thematic deep-dives on emerging customer issues became a weekly cadence; and the team built a shared library of "winning calls" that new agents now reference during onboarding.

For the first time, the bank has a single quality dashboard that the COO opens every Monday morning — with comparable scores across Budapest, Bucharest, Sofia, Belgrade, and the rest of the group.

Group QA Dashboard — Monday review
CSAT — Hungary retail
Up from 7.4 last quarter
8.4
FCR — Romania card services
+9 pts after coaching cycle 3
87%
!
Compliance — Bulgaria lending
12 missed APR disclosures this week
Action
AHT — Serbia retail desk
Down from 8m 12s baseline
6m 04s
Top performer of the week
Croatia retail · 14 calls reviewed
98.2
Flagged for coaching
23 agents across 4 markets
Queued
Ender Turing specialist

Want results like OTP Bank's? Let's build the business case for your team.

30-minute session with one of our deployment leads. We'll map your current QA process and send back a written outline of what's realistic in your first 90 days, based on customers running similar setups.

const reveals = document.querySelectorAll('.reveal'); const io = new IntersectionObserver((entries) => { entries.forEach(e => { if(e.isIntersecting) { e.target.classList.add('visible'); io.unobserve(e.target); }}); }, {threshold: 0.12}); reveals.forEach(el => io.observe(el)); const burger = document.getElementById('burger'); const mobileMenu = document.getElementById('mobileMenu'); burger?.addEventListener('click', () => { const open = mobileMenu.classList.toggle('open'); burger.classList.toggle('open', open); document.body.style.overflow = open ? 'hidden' : ''; }); mobileMenu?.querySelectorAll('a').forEach(a => a.addEventListener('click', () => { mobileMenu.classList.remove('open'); burger.classList.remove('open'); document.body.style.overflow = ''; })); document.addEventListener('click', e => { if (mobileMenu?.classList.contains('open') && !mobileMenu.contains(e.target) && !burger.contains(e.target)) { mobileMenu.classList.remove('open'); burger.classList.remove('open'); document.body.style.overflow = ''; } });