This article can also be found in the Premium Editorial Download "BI Trends + Strategies: What predictive analytics pros don't do."
Download it now to read this article plus other related content.
The widespread adoption of predictive analytics has been at the mercy of two opposing forces over the past two decades. Frequent, compelling use cases from the few organizations that have properly implemented predictive analytics projects propelled the discipline into the mainstream. Yet its perceived complexity has slowed adoption.
Let’s take a look at five critical things business intelligence (BI) and analytics professionals often overlook, hence depriving their organizations of the substantial benefits of predictive analytics.
1. Getting started the right way.
Predictive analytics is not another flash-in-the-pan technology. Most Fortune 500 companies have established departments and practices that refine crude data into highoctane intelligence. But few beyond the largest corporations have formalized or even organized their approaches.
Why haven’t more jumped into the game? There many barriers and excuses, most notably:
- Corporate executives don’t believe predictive analytics can deliver net gains;
- Department heads who would have a stake in a project assume it will require a substantial investment in expensive consultants or a bench of professionals with doctorates;
- BI managers don’t want to take on another initiative; and
- The business can’t get ahead of the deluge of data.
Organizations are simply not getting started with predictive analytics the right way. Those that approach it like any other IT or BI project will not unearth meaningful or measurable results, and predictive analytics will be prematurely dismissed.
Yet as the flow of incoming data gains speed, organizations can’t just keep pumping dollars into storing, structuring, cleansing, transforming and transporting data. They need to ultimately uncover the core value: actionable information hidden in growing data stores.
2. Training before doing.
Why not approach predictive analytics like most IT and BI projects—with an emphasis on technical training? Most IT projects focus on fulfilling objectives at the operational level; their functions are more direct and tactical. Predictive and advanced analytics, however, rely heavily on strategic assessment, design and implementation. For practitioners and functional managers entrenched in typical IT projects, this requires a significant shift in mind-set.
Most are surprised to learn that the bulk of the training necessary for predictive analytics success is not technical. The mathematics involved can be very sophisticated, but modern software tools automate the complexity, allowing most business practitioners to build adequate predictive models with little training. Although it’s helpful to have some statistical grounding and a basic appreciation of the capabilities and limitations of various modeling methods, expertise in actual model development has little impact on the success or failure of a project.
But it is imperative that at least one project manager or functional leader be well-versed in a formal, methodical, process-driven approach to predictive analytics— for example, the Cross-Industry Standard Process for Data Mining, or CRISP-DM—at the project level. Unfortunately, there are very few courses out there that provide this emphasis—particularly from a vendor- neutral perspective. But a search on data-mining and predictive-analytics strategic training will produce a few good options with a projectlevel orientation.
3. Designing before building.
If you were building a new house, wouldn’t you first meet with designers and engineers to draw up blueprints? What would your home look like if the builder started construction before fully understanding your needs, preferences and site dynamics? The answer is obvious.
Most organizations start predictive analytics projects by jumping directly in with software and data and then hammer away on models without understanding what they’re building. When predictive analytics is not approached as a well-planned process, practitioners may still end up building very good models, but the models answer the wrong questions and can’t be properly interpreted or implemented.
A number of standardized processes spell out precisely how to plan and implement a successful predictive analytics project. Two of the major ones are vendor-neutral CRISP-DM and SAS-specific SEMMA. But vendors are primarily interested in trumpeting and selling technology. Their marketing efforts have influenced misguided organizations to listen to the loudest voices and start with the easiest-to-obtain resource: software.
Even the few companies that are aware of standardized processes are resistant to start with training and follow a formal process like CRISPDM. When it comes to predictive analytics, nobody wants to have to sell the notion of a comprehensive assessment and resulting project definition. The exercise itself is not sexy. It requires an up-front investment of time, money and effort. And it will not produce a return on investment. But then again, neither will blueprints. They will help ensure that your house doesn’t collapse, though.
4. Putting the problem ahead of the analytics.
Many argue that organizations should start with predictive analytics to uncover unknown insights, relationships or anomalies that may direct subsequent analysis. This approach may uncover a few artifacts of interest, but it rarely moves beyond an unsupervised exploratory exercise.
A 2009 survey of self-proclaimed data-mining practitioners by Rexer Analytics revealed a focus on the wrong part of the problem. The majority of participants cited the performance or accuracy of predictive models as the most important factor in determining success. In the real world, however, practitioners are not rewarded for how well a model conforms to artificial metrics— but rather for how effectively it helps optimize the allocation and use of organizational resources.
While discovery is a function of predictive analytics, the derived information should support the organization’s priorities—and not the other way around. The first step should be establishing and validating strategic priorities across functional teams. Justifying the time to properly design an analytics project in these days of Agile BI and immediate results is not easy, but it’s mandatory. A comprehensive assessment, formal project definition and measurement framework must be established for results that are substantive and sustainable.
5. Avoiding distractions and buzz.
Just as the masses moved far enough down the BI chain to embrace predictive analytics as a formal practice and start deriving value, they were diverted to the next shiny, new thing upon hearing exciting terms like big data and data science.
Seasoned BI professionals will recall how chasing hype can lead to costly exploration of uncharted, underdeveloped and oversold technology.
Vendors riding the coattails of the big data analytics buzz may argue that traditional analytics just won’t get the job done in the new world of ever-swelling data sets. When real-time analysis of high-velocity, multidimensional data is required, there may be a need for computational scalability. But applications that call for in-database analytics or distributed processing are certainly not the norm.
When sampled properly, only a small fraction of the data available is required to adequately represent the solution space—a mathematical term describing the entire area represented by multiple dimensions—and effectively train a predictive model to produce the desired information. Once such models are deployed, running them and scoring large volumes of new data can be done highly efficiently, suiting the vast majority of business applications.
Be wary of the hype. Those who maintain the course toward predictive analytics while gradually building other aspects of their BI chains toward big data scalability are due for early payback.
So don’t cheat the process. Don’t rush to grab software and dive headlong into your data. It’s simply not worth it, and that’s been proven time and time again. If you don’t succeed with predictive analytics on a small-scale pilot program driven by a sound assessment and project definition, then forging ahead into big data analytics will produce nothing more than a big noise.
About the author:
Eric King is president and founder of The Modeling Agency, a consulting and training firm with a focus on data mining and predictive analytics. Email him at firstname.lastname@example.org.
This was first published in September 2012