It is both more critical and more challenging than ever to answer the question, “What is happening in my contact center?” That’s because today, no matter your business, your customer experience (CX) is your business: it’s your signature offering, your differentiator and your key revenue generator. And the contact center is the home of CX. This puts contact center performance under the metaphorical spotlight and microscope, as business leaders expect to have a clear, deep and near-real-time understanding of how contact center performance is helping (or hurting) their business.
But as the simple call center has evolved into the modern, multi-channel contact center, it’s become increasingly difficult to gain a firm grasp on performance. The challenge isn’t just just channel expansion. It’s multiple versions of multiple systems spread across multiple locations. It’s unstructured data that’s unusable in its raw form. It’s correlating organization-wide data. In short, it’s the need to make sense of a trove of data that grows in size and complexity every day.
Facing this challenge, one thing is becoming clear: Most contact centers need a better, smarter reporting solution. They need reporting that breaks down silos to give full visibility across the contact center—and across the entire organization. They need reporting that is faster, more efficient and more accurate than manually managing endless “spreadmarts.” Finally, they need reporting that is more intuitive—with KPIs and data visualizations that anyone in the organization can easily understand.
This guide will give you an end-to-end look at the current challenges and future potential of contact center reporting and analytics solutions, including:
In the simplest sense, reporting shows you what is happening in your contact center. Reporting takes the many streams of raw data flowing into your contact center—from your ACD, IVR and WFM systems, for example—and transforms that data into key performance metrics (KPIs). Common KPIs include:
• First Contact Resolution (FCR)
• Adherence to Schedule
• Customer Effort Score
• Net Promoter Score
In today’s technologically advanced world, just about any reporting tool has the capability to turning the raw data into KPIs. However, the art of reporting lies in two factors:
1) How the data is orchestrated into KPIs: This is the formula, so to speak, that a reporting tool uses to calculate a KPI. A better formula leads to more accurate KPIs. More accurate KPIs offer more relevant and usable information.
2) How KPIs are presented to the end user: If the point of reporting is to make raw data easy to understand, then success depends on how KPIs are visualized. End users should be able to understand KPIs at a glance—and immediately recognize when KPIs signal the need for further analysis.
The second point is even more critical today, as business leaders outside the contact center take an increasing interest in contact center performance. Traditionally, the users (or consumers) of contact center reporting have been supervisors, WFM analysts and contact center leaders. But as organizations increasingly elevate the customer experience (CX) as the core of their business, they’re recognizing that the contact center is the home of CX. Executives and other non-contact-center leaders expect to see clear, accurate metrics of contact center performance—and tie contact center data with other data streams to produce customer-centric, business-wide KPIs.
Reporting and analytics are increasingly discussed in tandem. In fact, they’re often used almost interchangeably. The first thing to understand is that reporting and analytics are not the same thing.
|What it is:||Turns raw data into key performance metrics that show you what is happening in your contact center.||Identifies patterns and trends in the data that represent relevant actionable information.|
|Example:||Reporting can tell you that call volume is up and FCR is down.||Analytics can identify a rapid increase in customers calling about an ecommerce checkout issue.|
|The bottom line:||Reporting helps you see what questions to ask.||Analytics helps you answer those questions.|
As you can see, reporting and analytics are distinct, but interrelated. In a forward-thinking contact center, reporting and analytics feed a continuous improvement loop:
Legacy reporting products are product-specific and fail to deliver the omnichannel and organization-wide visibility you need. But relying on ad hoc reporting—Excel data dumps that turn into out-of-control “spreadmarts—is inefficient and error-prone. Modern, multi-channel contact centers need modern, integrated reporting and analytics solutions that are built to answer complex questions and deliver detailed insights.
While every contact center has unique reporting needs that depend on their line of business, the structure of their organization and the markets they serve, there are several essential KPIs that just about every contact center should be measuring:
First Contact Resolution (FCR): FCR is the percent of contacts that are resolved on the first interaction with the customer. For live calls or web chats, this means that the customer’s issue is resolved before they hang up the phone or end the chat session. For email, the most common standard for FCR is resolution within one business hour of receiving the original customer email. FCR is one of the most important factors in determining overall customer satisfaction.
Adherence to Schedule: Schedule adherence is measured by taking the total time a call center agent is available and dividing it by the time they are scheduled to work, expressed as a percentage. This metric may take into account time spent on breaks or doing non-call related work. Most call centers define a target schedule adherence percentage allowing for cushion time beyond the scheduled break times. Adherence to schedule is, by far, the biggest factor in achieving an ROI from a WFM platform. Ensuring agents are doing what they are supposed to do—when they are supposed to do it—keeps contact center costs down while driving up other performance metrics up.
Customer Effort Score (CES): CES measures the customer experience with a product or service. Customers rank their experience on a seven-point scale ranging from “Very Difficult” to “Very Easy.” This determines how much effort was required to use the product or service and how likely they’ll continue paying for it. In other words, CES demonstrates how easy your organization is to do business with—one of the best predictors of customer satisfaction.
Agent Occupancy: Agent occupancy refers to the percentage of time that call agents spend handling incoming calls against the available or idle time, which is determined by dividing workload hours by staff hours. It is a statistic used in calculating the productivity of a call center. This metric has a direct correlation to both customer and agent satisfaction. Occupancy that is too low leads to unengaged agents and lower customer satisfaction, whereas occupancy that is too high leads to agent burnout as well as low customer satisfaction.
Service Level: Service level measures how accessible an organization is to their customers—defined as the percentage of calls answered within a predefined amount of time, or target time threshold. Consistent service levels are one of the first things customers notice when doing business with an organization.
Forecast Accuracy: Forecast accuracy shows the percent variance between the number of customer contacts (calls, texts, emails, chats, etc.) predicted to arrive during a given period and the number of contacts that the contact center actually receives during that time period. An accurate forecast leads to accurate schedules with the proper levels of staffing. An inaccurate forecast leads to inaccurate schedules, which lead to missed service levels and/or unnecessary staffing expenses. In essence, forecast accuracy is the first step in ensuring you have the right people in the right place at the right time in your contact center.
Abandoned Call Percentage: Abandoned call percentage, also called abandon rate, is the percentage of tasks that are abandoned by the customer—either before speaking with an agent or before completing the intended task. This metric shows the level of frustration your customers have with your organization. High abandoned call rates mean high levels of customer frustration.
Average Call Transfer Rate: This a metric that monitors the percentage of calls transferred to another department, a supervisor, or a different queue. Average call transfer rate also predicts customer frustration, as customers who are transferred frequently will be more frustrated and have lower customer satisfaction.
Net Promoter Score (NPS) and Predictive NPS: NPS is a measure of how loyal customers are to an organization. Customers are asked how likely they are to recommend (or promote) an organization to others, giving a rating between 0 (not at all likely) and 10 (extremely likely). NPS provides a direct way to measure contact center performance—and demonstrates the direct connection between the contact center and the overall brand perception. While NPS is typically measured with a select number of random customer surveys, Predictive NPS uses artificial intelligence and machine learning to provide an NPS value for every customer and every interaction.
Quality Score/Predictive Quality: Quality scores show you how are you agents performing against your own internal metrics. Predictive Quality uses AI-powered analytics to automatically evaluate 100% of interactions. This metric, along with the underlying quality form data, should drive your coaching and training programs throughout the contact center.
To learn more about these critical KPIs—including best practices for calculating them, as well as common mistakes and misconceptions that lead contact centers off track—download Calabrio’s quick-read eBook: A Guide to Measuring—and Using—the Top 10 Contact Center KPIs
The first call centers emerged in the 1960s. These early call centers used simple reporting to understand performance—a balance of quality and efficiency—by answering basic questions, like:
These questions are the roots of key performance metrics (KPIs) that are still essential to contact centers today.
But a lot has changed in the last 60 years. The simple call center has evolved into the multi-channel contact center. Modern contact centers record much more than calls, capturing data on every aspect of contact center operations. That data is increasingly connected with other critical data streams across the organization. This reality requires the use of multiple systems and applications, including:
The basic goal remains the same—contact center performance still boils down to balancing quality and efficiency. But the factors that drive performance have grown exponentially more complex. Moreover, the modern contact center already has all the data it needs to understand performance. The challenge today is how to transform that raw data into meaningful metrics—and how to do it accurately, efficiently and as close to real-time as possible.
As organizations refocus on the customer experience, the contact center becomes a hub of business intelligence. Reporting questions come in constantly—from within the contact center and across the business—and those questions continue to grow more complicated. In the face of these increasing demands, the shortcomings of current reporting systems become glaring flaws:
Typically, most contact centers begin with the standard reports included in their ACD, IVR and WFM systems. However, these out-of-the-box reports usually only cover the basics and likely don’t reflect a given organization’s unique measurement needs based on industries and markets served, customer needs or geographical/cultural differences.
When it comes to reporting and analytics, each of the systems and applications a contact center uses (ACD, IVR, WFM, etc.) creates islands of data. Each has a dedicated database optimized for doing one siloed job and generating one set of insights. To further complicate things, many contact centers are managing multiple locations, and each many have a distinct brand or version of ACD, for example, creating yet more silos.
An ACD’s standard reports are designed to show how the contact center is performing only from the ACD’s point of view. They don’t take into account, for example, how agent scheduling (managed in a WFM system) will affect ACD metrics. In other words, siloed information makes it difficult to answer simple questions, such as, “Are break schedules negatively impacting service levels?” To resolve the situation, most contact centers turn to the trusty old spreadsheet. It’s easy to see why; spreadsheet software (usually Microsoft Excel) is the main tool most managers use to crunch numbers or present data. It’s the tool they know. If contact center managers are lucky, the application or system they’re using kicks out data in a .CSV format, so it can be uploaded into a spreadsheet. If not, it requires manual entry or a laborious cut-and-paste process. This is all to create one single, siloed report. To determine how agent sheduling is impacting service levels, for example, at least two reports must be generated, and data from each laboriously merged manually in a spreadsheet—a time-intensive process that also fraught with opportunities for error.
Business intelligence isn’t a static concept. Leaders don’t want a one-time answer to their questions; they want an ongoing, dynamic look at key CX metrics. So reporting needs that were once handled by a manual, ad hoc report turn into daily reports. Then, weekly, monthly and quarterly reports are tacked on. To save time, the manager just rolls daily reports into a weekly report, weekly reports into monthly reports, and so on. Pretty soon, quarterly contact center metrics are being referenced in annual organization-wide performance reports. Spreadsheets built on spreadsheets that are ultimately built on spreadsheets—each of which has to be manually updated.
Requests pile up, one-off reports turn into daily tasks, the cross-referencing gets more convoluted, and before they know it, contact center managers find themselves overwhelmed by the “spreadmart”—a mind-numbing and costly proliferation of spreadsheets attempting to do the job of a dedicated data mart or data warehouse. In addition to being tedious to produce, the spreadmart does not provide the predictive and prescriptive analytics that every organization craves today. Contact center leaders end up spending the majority of their time updating spreadsheets and manually creating metrics—instead of focusing on how to use information and insights contained in those reports and metrics.
As errors proliferate in these patchwork-quilt spreadmarts, contact centers begin to lose trust in their stats. Data integrity can also erode as underlying business rules change, or as call flow/call handling protocols are reconfigured, resulting in changes to how calls are pegged in the ACD. A common result is “dueling spreadsheets”—reports ostensibly covering the same metrics, but with different data values suggesting different conclusions.
A replacement, upgrade or reconfiguration of any component of the contact center environment might change the way underlying operational data is logged, with ripple effects on all dependent reports. Adding new technologies (for social media tracking, for example) introduce new silos generating new data that must be incorporated into performance reports.
Changes in management or organizational direction will likely necessitate significant and sometimes immediate changes in the contact center, and thus its reporting requirements. Evolving goals within the contact center itself also shift reporting requirements. For example, as workforce engagement emerges as a top strategic goal in many contact centers, new metrics and reports are needed to quantify agent engagement If these changes aren’t made quickly, fundamental disconnects can create constant, unproductive internal struggles between the contact center and upper management. These clashes can occur over anything from a perceived lack of progress to claims’ conflicting goals and inconsistent direction.
When faced with one or more of these challenges, contact center leaders often struggle with what to do first in improving contact center reporting and analytics. Fortunately, this is not uncharted territory—there are established best practices that define two clear paths to transforming your reporting system. The best option depends on the specifics of your contact center, but both options are a dramatic improvement on the status quo.
Reporting frustrations can make a fresh start—building a new reporting system from scratch—an appealing option. But that’s not necessarily the right strategy for every organization. Sometimes key stakeholders are reluctant to begin a process that could eventually lead to sweeping changes down the line. There are several common sources of resistance:
Understandable as all these viewpoints may be, the fact remains that as time goes on and the contact center expands, piece-meal manual reporting becomes more costly and error-prone. The implicit cost of getting good data will ultimately break budgets, and bad data will confound efforts to improve contact center performance.
The good news for all these stakeholders is that you can leverage your existing reporting system to build a new, better, smarter system. In this approach, your current spreadsheet-based reports are re-purposed—not tossed away. Despite their flaws, spreadsheet reports have the virtue of being a very accurate reflection of your contact center’s current reporting needs. As such, they make excellent blueprints to guide the next evolutionary step in the process—eliminating manual effort through automation of data collection and consolidation.
Under this gradual, “start small” approach, the only major decision is choosing the proper enabling technologies to completely eliminate the need for a human to cut and paste data into these reports. In fact, limiting the scope of an initial project to this simple goal alone has two advantages:
1. It’s clearly understood by all stakeholders
2. The resulting benefit in saved time and money is easily quantifiable
With the right technology, these quick wins can be achieved easily while providing the foundation for even bigger gains from a more comprehensive review.
Another benefit of building on your existing foundation is giving analysts and managers breathing space to re-examine and refine metrics and calculations used in existing reports, evaluate whether these need to be adjusted, and to fine-tune report formatting. With new technology eliminating human error from the data consolidation process, managers can even take a retrospective look at previous reporting periods with fresh, accurate data to revise any assumptions or conclusions that may have been made based on erroneous information.
A comprehensive review also provides the opportunity to test existing reports’ relevance to the contact center and its stakeholders. Some reports may have outlived their usefulness. For each existing report, ask some value-testing questions:
These questions can be asked not only of existing reports, but also for the measurements they contain: Metrics, formulas, custom calculations and even system calculations should be validated in this way to make sure all stakeholders agree on their ongoing value and accuracy. Though the verbal definition of service level—x percent of calls answered in y seconds—will be familiar to all concerned, its underlying formula has a half-dozen or more known variations. While it’s not necessary that every stakeholder understand how it’s calculated, it is important that a single formula be applied universally.
Even if you’re an outside-the-box thinker stuck in a slow-and-steady-wins-the-race organization, rest assured that this conservative approach can lead to strong long-term results. With smart technologies powering their current reports, analysts and managers will inevitably discover the richness and variety of new information available in the contact center’s underlying operational systems. Reluctant stakeholders will realize that all the tools they need for broader and deeper reporting and analysis are already at hand; all that’s needed is the determination to take the next step: a formal review.
Some contact centers don’t have the luxury of time required to take the first, more conservative approach. Often, the impetus for the contact center change is a much more time-sensitive issue, such as:
Whatever the reason, managing a more time-sensitive reporting transformation requires an approach focused on a defined end goal. This approach requires more upfront planning and a structured, top-down process centered on key goals that drive every subsequent leg in the journey.
Whether you’ve arrived at a comprehensive review as part of your time-sensitive, strategic initial transformation, or as the next phase following a more conservative, build-on-what-we-have approach, the process is the same. In general, there are five phases in a complete review of contact center reporting and analytics. First, the scope of the project itself must be clearly established. Second, requirements must be gathered and defined. Third comes the design process, which will likely have several iterations. Fourth comes the crucial technology selection, followed by the fifth and final phase: implementation and review.
To stay on track, any formal project needs scope—big or small. A project can be defined as narrowly as “merge average-speed-of answer stats from these three different ACD systems into a line-graph report,” or as broadly as “build agent, team and company-wide scorecards on seven key performance indicators, to be refreshed hourly in an at-a-glance dashboard.” Other aspects to include in project scope include:
There are three critical roles that should be clearly assigned at the outset of the project:
By definition, a review of contact center reporting and analytics is a reassessment of key stakeholders’ requirements. It’s also a great opportunity to make sure the contact center is conforming to the goals of the organization as a whole. The most powerful way to accomplish this is to derive the contact center’s top-level requirements from strategic targets and goals set by the organization’s leaders.
An hour spent with the vice president of sales and marketing, the chief financial officer or the CEO can go a long way towards defining what needs to be accomplished. While this may sound daunting, most high-level strategic goals can easily be boiled down to specific, measurable actions in the contact center.
For instance, if a financial institution’s strategic goal is to move into a new market category (say, home insurance), then the contact center’s goal may be to improve metrics and reporting for cross-selling by customer service agents handling mortgage inquiries. Similarly, for an online retailer of computers that wants to win market share from competitors through better post-sale customer service and technical support, the contact center goal may be to improve customer satisfaction metrics and reporting.
In many cases, strategic goals lie partly or completely outside the contact center’s control. Stakeholders in other functional areas of the organization—marketing, product development, finance, human resources and other back office departments—should be consulted on what feedback or metrics they need from the contact center to refine their own plans and programs.
Finally, there are the contact center’s internal operational goals. Some contact centers want their agents to be more productive, so reporting will have to gather and consolidate accurate data on handle time, talk time or after-call work time. Some want to be more effective in the interactions agents have with customers, so first-contact resolution or customer satisfaction ratings will be key metrics. Most will want to gauge how all the moving parts of the contact center—agents, management and technology—are working together, so service level, average in queue and/or on hold times, average speed-of-answer and abandons will be critical.
When gathering requirements from stakeholders, it’s important not to shackle their choices to lists of known or existing metrics or reports. Rather, ask each stakeholder group to compile a prioritized list of information, expressed in business terms, which they would like to see from the contact center. Not all of these “wish-list” items will be achievable, but the lists will help clarify each stakeholder’s needs and suggest new metrics or measurements that may be possible with new reporting technologies.
Such user-driven consultation also helps with stakeholder buy-in. Asking for requirements to be expressed out loud in business terms—or just plain old English—is useful in another way. Consider the difference in the two questions below:
This question is all too easily answered with a canned report showing each of these metrics in hours, minutes and seconds. These metrics then have to be compared (often manually) to shift times (from another report) before they are of any use.
This question, expressed in plain English, could easily be answered by a single, more meaningful report with trended data shown as proportions (say, 60% inbound, 20% breaks, 10% training, 10% unaccounted for) which a supervisor or coach could understand at a glance—no assembly or decoding required.
Look for ways to consolidate many reports into a few—or just one. If users currently receive multiple reports summarizing the same data (by skill set, queue or application, or by time or geography), make eliminating this redundancy a key requirement of the review. The right reporting solution will provide ways to interact with a report to show (or hide) data as needed, and to navigate through standard hierarchies of data (year/month/week/day) within the same report. It will also allow your contact center’s unique hierarchies (site/team/agent, or custom geographies) to be represented as well.
The requirements list should also include any custom or unique calculations that users or analysts may be performing by hand or in spreadsheets. Every contact center serves a unique customer segment, market, industry and/or geography, and “standard” lists of metrics can never capture everything a given organization needs to measure. A new reporting solution should automate these custom formulas and apply any constants (cost-per-interaction data, loaded labor rates, occupancy or service level targets, or other performance objectives) needed in their calculation or presentation.
Don’t be restricted by the old limits on reporting imposed by operational data silos. If supervisors or coaches need an agent scorecard combining data from ACD, screen capture, quality management and workforce optimization or workforce engagement systems, then add a mock-up report to the requirements. It’s the reporting solution’s job to tap into the relevant data silos as needed—and to present that scorecard as though it came from a single, all-knowing system. Similarly, if analysts have struggled in the past to build end-to-end call audits from raw ACD and/or IVR data, then this should be a determining requirement for a new reporting technology.
With complete requirements and stakeholder buy-in, it’s time to begin designing actual reports. Again, spreadsheets are a useful tool to create mock-up reports. Most stakeholders have the skills to view and at least make cosmetic changes to spreadsheet mock-ups. These should circulate to a design team with at least one representative from each stakeholder group. A report should only move to the implementation phase when all affected stakeholder reps have signed off on it. Each contact center’s report mock-ups will be different, but here are some key guidelines that should pertain to all templates passed to the design phase:
1) Document sources: Beyond layout and formatting, these mock-up reports should include notes about from which operational data systems each metric should be drawn—ACD, IVR, WFM, WEM, CRM, etc.—and which custom formulas or calculations need to be applied in the presentation in the report itself. Cumulative metrics should also be highlighted along with any special rules governing them (fiscal YTD, etc.).
2) Note calculations from source systems: Remember, many of the metrics pulled from operational systems like the ACD are actually calculations, with formulas underlying them. If a new metric is created from the combination of two or more calculated statistics, the operational source of each constituent metric should be noted along with their respective formulas or calculations. This is particularly important in mixed-vendor environments, where a standard metric like service level may be calculated differently by different systems.
3) Check your frequency: Note stakeholders’ required frequency for each report refresh (intraday, daily, weekly, monthly, quarterly,e tc.) and whether previous editions of each report should be maintained separately (i.e., not overwritten when updated) and kept accessible for future reference. Your reporting solution should give you the option of filtering report data based on date, making a historical archive of past reports redundant.
4) Determine distribution and scheduling: Each report mock-up should note who will be receiving or accessing the report and through which distribution channel. Some users may simply require a PDF or spreadsheet attachment via email; others may require access to a web portal for more interaction with report data. Avoid sending reports to stakeholders who don’t need them.
5) Keep it simple: Don’t try to include too many metrics in a single report. Each report should support a single management action or decision, or a group of related decisions. For instance, scorecards used to coach individual agents should include metrics that are under the agent’s control (not-ready time, average talk time, attendance, etc.) and relevant independent variables like calls per hour, but not top-line measures like calls offered, agent utilization or abandon rates.
6) Avoid dueling metrics: Some contact center measures clash with others, pulling managers in multiple directions. If managers decide that first call resolution is a customer service priority, then setting low targets for average talk times will actually hinder agent and team performance. Arbitrary call-time targets will also hamper sales-oriented centers that reward higher revenue-per-call or conversion rate (the percentage of inbound or outbound calls converted to a sale). Avoid this conflict by keeping reports relevant to one actionable decision or group of related decisions.
The first three phases of a comprehensive review of contact center reporting and analytics can—and often should—be completed before meeting with vendors to select enabling technologies. The goal is to define your needs—your ideal list of reports and metrics—and then find a solution that enables you to meet those needs. If a vendor is introduced into the review process too early, their technical limitations may inherently limit the report requirements and design, and the whole project may ultimately fall short of its potential.
There are three solution categories that have evolved to meet the challenge of better reporting and analytics in the contact center:
1. Report extension tools
Sometimes called point-to-point tools, report extension tools are offered by many of the same vendors who manufacture the contact center’s operational hardware and software. These systems almost always feature built-in, standard, “out of the box” reports for a given system like the ACD, as well as report-authoring tools that analysts can use to create their own custom reports. In skilled hands, these tools allow data to be imported from other systems and data stores. Similarly, vendors of real-time dashboard, workforce optimization, workforce engagement and quality management software offer reporting tools that may be used to access other data sources.
2. Performance management suites
These solutions usually offer reporting services as part of a much wider array of applications, including workflow management, coaching, contact type profiling, agent recognition and compensation.
3. Business Intelligence (BI)
BI is a software architecture designed specifically for reporting and analysis of information from operational data systems of any kind—point-of-sale, manufacturing, banking and finance, sales and marketing, or any other computerized system with an underlying database logging transactional data.
Defining your reporting architecture and selecting a vendor is just the front half of the process of transforming your contact center reporting system. An intensive testing and review process is still critical to a successful implementation—even if your vendor takes full responsibility for installing and configuring all the required software, as well as developing all content based on your mocked-up reports,
Testing and review is typically an iterative process:
Early “alpha” versions of each report will be generated with smaple data and distributed to internal analysts for data validation.
Alpha reports get recirculated back to report designers for corrections in calculations or formulas.
Once the numbers are right, designers move on to report formatting tasks like adding charts and graphs, improving the look of tablets and generally improving the user experience.
Once reports have achieved a certain polish, it’s time to expand the resting and review process to include end users. These reviews should consider how the reports look, how their interactive features work, how they print out and most importantly, how useful they are.*
Usually it takes a few testing and review iterations with all stakeholder audiences before reports are ready for implementation and day-to-day use. As with report mock-ups in the design phase, represtatives of each stakeholder group should (literally) sign off on each report before it is put in production.
Once reports go live in the wider contact center, stakeholder should still expect one of two hiccups as final configuration issues are worked out. Representatives from each stakeholder group should document any errors or unpredictable behavior they experience in using the new reports.
* NOTE: Involving these stakeholders too soon in testing can be a mistake. Non-technical users typically can’t add much value to early data validation, and it can be difficult for them to shake a bad first impression made by a report containing ratty or nonsensical data. It’s best to wait to solicit end-user feedback until all they’re reviewing is the day-to-day utility of reports.
If your software vendor is tasked with report design, don’t let their enthusiasm for fancy new technology take your project off course. Just because they can show 16 different metrics represented as speedometers in a dashboard with a 15-minute refresh doesn’t mean they should—unless your requirements process identified that need. The same rule applies to internal “power” users with advanced skill sets who may be understandably excited about what their new tools can do. There will be time enough for groundbreaking innovation once the first wave of reports meet documented requirements.
Whether your contact center takes the build-on-what-you-have approach or immediately undertakes a comprehensive review and rebuild, the ultimate sign of success is the same: when you’re getting the reports you need, filled with accurate data that’s useful to you, in a timely fashion—without being burdened with manual work to patch gaps. At the very least, following the guidance provided here should lead you to a point where you are free to spend more of your time and effort on improving your contact center’s performance, rather that chasing bits of data down dead ends.
But though re-engineering your contact center reporting program will deliver significant and ongoing benefits, there is no true endpoint to this process. The contact center ecosystem continues to expand in size and complexity—with emerging channels like social media, new switching technologies like SIP and new innovations in performance management techniques. This evolution isn’t just constant—it’s accelerating. And these changes will bring new reporting needs and create new challenges in meeting those needs.
Fortunately, these future challenges won’t be all that different from those currently facing many contact centers: bringing together disparate data streams, harmonizing those data streams, and orchestrating that integrated data into relevant, usable reports and KPIs that show you what is happening in your contact center.
To that end, the best practices covered on this page will remain the best practices for tackling the contact center reporting challenges of the future: defining project scope, identifying stakeholders, gathering requirements and creating mock-up reports that express them, choosing a reporting architecture that can meet reporting needs now and in the future, reviewing, testing, and finally implementing a solution. Follow this guidance, and you’ll be on your way to more accurate, more usable reports—and more time to focus on using those reports to drive the performance of your contact center and the success of your business.