Share with your friends


Analytics Magazine

Forum: Oil & gas producers need to tame the gusher … of data

July/August 2013

Warren WilsonBy Warren Wilson

Throughout the century-and-a-half since the dawn of the commercial petroleum industry, oilmen have always hoped for the gusher – the big find that would spew enough oil to make them rich. There have always been far more “dry holes” than gushers, however, and the proportion only gets worse as oil and gas become harder to find and more difficult to produce. One thing the industry has in abundance today, however, is data. And just like some of the biggest oil discoveries, the data gusher offers huge promise if it can be tamed. However, the exploration and production (E&P) industry, and the IT companies that support it, have a lot of work ahead to derive maximum value from their growing troves of data.

The first oil wells didn’t rely on data at all. They were sited where oil seeped from the ground naturally, and the challenge – not simple, but not data-dependent – was figuring out how to dig or drill down to the source.

Data grew more important as the industry grew, reserves became harder to exploit and drilling technology evolved. Drillers began keeping paper records of what occurred during each work shift or “tour.” They tracked basic metrics such as the number of feet drilled per hour or day, obstacles encountered, injuries sustained.

Over the years recordkeeping has become steadily more thorough and sophisticated. Today the E&P industry has vastly better tools for every aspect of oilfield operations including: three-dimensional maps of subsurface geological structures and hydrocarbon reservoirs; graphs or “logs” of the wells’ downhole conditions (temperature, pressure, porosity, permeability, etc.); and records of injuries and environmental incidents. New equipment is often fitted with sensors that produce steady streams of data about temperature, vibration and other parameters that indicate whether the asset is operating as it should, or whether it needs service, repair or replacement.

Yet E&P companies still find themselves in the same position as their predecessors and their counterparts in other industries: They are awash in data, but they don’t have all the insight they need. For example, they have copious real-time data from individual wells, but they do not have a good handle on the dynamics of complex unconventional reservoirs. Similarly, they have access to vastly better analytical tools than ever before, but few companies can claim to have fully optimized field operations, production or asset management.

Analytics Tools Must Evolve

E&P companies’ drilling programs have always relied primarily on historical data that describe a given region and its history. Extrapolation and interpolation suggest where to drill next – if you drill between two producing wells, your odds of success are relatively good. Outside known reservoir boundaries, the odds fall off dramatically; so-called “wildcat” wells face a much greater risk of producing no return on many millions of dollars invested. That has always been the biggest gamble in oil exploration – even when armed with the best historical, descriptive data available, you still have to spend large sums up front to drill the well before finding out if your data is any good.

Today the industry’s data and analytics needs are changing, and so are the tools at its disposal. Having found and developed most of the world’s “easy” oil and gas reserves, E&P companies are venturing into even more remote locations and extracting hydrocarbons from unconventional sources. They are tapping into shale rock so impermeable that it won’t give up its gas and oil without first being “fractured” with water, sand and chemicals injected under high pressure. They are tapping sand deposits that contain oil so viscous that it won’t flow without first being diluted with solvents or softened by steam.

These unconventional sources pose unique challenges that require greater precision and real-time analytics – for example, to keep drill bits positioned precisely where they need to be within the shale “pay zone” and to control the placement, composition and pressure of fracking fluids to yield optimal results.

At the same time, E&P companies increasingly need predictive analytics tools, for a variety of purposes. They need to better understand how current production methods will affect long-term yield. They need more accurate predictions of asset behavior to improve continuity in drilling and production while minimizing the costs of spare equipment or service crews. Software vendors are increasingly offering analytics tools that address such needs.

But prediction is just one step forward. The next is so-called “prescriptive” analytics that go beyond merely predicting future behavior, to recommend the best course of action to achieve a desired result. Such capabilities are quite rare today and at a very early stage of development. But they promise to bring new levels of performance, not just in E&P but in many other industries as well, because they track the results of their recommendations and feed those results back into the prescriptive algorithms to produce (in theory, at least) better and better recommendations over time.

Complex Needs Drive Demand for New Solutions

Despite all the potential benefits of advanced analytics in oilfield operations, their adoption is still at a very early stage. That is true of most industries, simply because the technologies themselves are new. Software vendors today are evangelizing their predictive capabilities; only a tiny number yet offer “prescriptive” analytics. Indeed, the term itself is not yet widely known.

An additional factor in IT adoption in E&P is the industry’s insularity. It is a world unto itself, an industry that operates largely outside of everyday view and speaks a specialized language that few outsiders understand. It is an industry driven by geologists and engineers who tend to regard IT as a mere support tool for existing operations.

The idea that IT, and particularly analytical software, can provide strategic and competitive advantage is not widely held. But until it is, E&P companies will not find or produce as much oil and gas as they could, nor will they manage their operations as efficiently and safely as they could. E&P companies owe it to themselves, and to their shareholders, to broaden their traditional view of IT and consider the strategic advantage it can provide. IT vendors, for their part, must state their value propositions in terms that E&P engineers can understand. They must explain, in plain language, how IT can help E&P companies find and extract more oil.

Warren Wilson ( leads Ovum’s Energy team, focusing primarily on IT for upstream Oil & Gas. He has been an analyst for 14 years. He joined Ovum in 2006 when Ovum acquired his former employer, Summit Strategies. At Summit his primary area of responsibility was mobile business applications. On joining Ovum, his research focus shifted to business management applications such as enterprise resource planning, supply chain management, and analytics.

Before becoming an IT analyst, Wilson had been a reporter and editor for U.S. newspapers including the Casper (Wyoming) Star-Tribune, where he covered oil & gas and other energy industries. He majored in geology at Carleton College in Northfield, Minn., and later worked in the oilfield as a roughneck and well logger.

business analytics news and articles



Related Posts

  • 81
    November/December 2014 The marriage of two “natural resources” – hydrocarbons and data – will transform unconventional oil development. By Atanu Basu (right), Daniel Mohan and Marc Marshall Known-knowns, known-unknowns and unknown-unknowns. Donald Rumsfeld’s notable turn of phrase is an apt characterization of where we are with unconventional oil development today.…
    Tags: data, oil, analytics, prescriptive, wells, gas
  • 62
    March/April 2014 By Atanu Basu The United States is re-emerging as an energy superpower. According to the International Energy Agency, by 2016 the U.S. will surpass Saudi Arabia and become the world’s largest oil producer. The domestic energy industry’s recent rise is the result of lower demand through energy efficiency…
    Tags: oil, analytics, data, production, gas, prescriptive
  • 58
    As the Big Data Analytics space continues to evolve, one of the breakthrough technologies that businesses will be talking about in the coming years is prescriptive analytics. The promise of prescriptive analytics is certainly alluring: it enables decision-makers to not only look into the future of their mission critical processes…
    Tags: analytics, prescriptive, data
  • 53
    The information explosion has led organizations to leverage data to improve the overall decision-making process. Organizations are looking to deploy data-driven strategies across their business processes and functions, such as marketing, risk, supply chain and finance. It’s critical now more than ever to have the right people, processes, methodologies, platforms…
    Tags: analytics, prescriptive
  • 49
    November/December 2012 The advent of the “digital oil field” helps produce cost-effective energy while addressing safety and environmental concerns. Finding and producing hydrocarbons is technically challenging and economically risky. By Adam Farris Everyone needs it, few know how we get it, and many feel compelled to slow down efforts to…
    Tags: data, oil, gas, industry


Does negative political advertising actually work?

While many potential voters dread campaign season because of pervasive negative political advertising, a new study has found that negative political advertising actually works, but perhaps not in the way that many may assume. The study “A Border Strategy Analysis of Ad Source and Message Tone in Senatorial Campaigns,” which will be published in the June edition of INFORMS’ journal Marketing Science, is co-authored by Yanwen Wang of the University of British Columbia in Vancouver, Michael Lewis of Emory University in Atlanta and David A. Schweidel of Georgetown University in Washington, D.C. Read more →

Meet Summit, world’s most powerful, smartest scientific supercomputer

The U.S. Department of Energy’s Oak Ridge National Laboratory on June 8 unveiled Summit as the world’s most powerful and smartest scientific supercomputer. With a peak performance of 200,000 trillion calculations per second – or 200 petaflops – Summit will be eight times more powerful than ORNL’s previous top-ranked system, Titan. For certain scientific applications, Summit will also be capable of more than three billion billion mixed precision calculations per second, or 3.3 exaops. Read more →

Employee engagement a top concern affecting customer experience

Employee engagement has surfaced as a major concern in delivering improvements in customer experience (CX), with 86 percent of CX executives in a Gartner, Inc. survey ranking it as having an equal or greater impact than other factors such as project management and data skills. “CX is a people issue,” says Olive Huang, research vice president at Gartner. “In some instances, the best technology investments have been derailed by employee factors, such as a lack of training or incentives, low morale or commitment, and poor communication of goals." Read more →



INFORMS Annual Meeting
Nov. 4-7, 2018, Phoenix


Making Data Science Pay
July 30-31, 12:30 p.m.-5 p.m.

Predictive Analytics: Failure to Launch Webinar
Aug. 18, 11 a.m.

Applied AI & Machine Learning | Comprehensive
Sept. 10-13, 17-20 and 24-25

Advancing the Analytics-Driven Organization
Sept. 17-20, 12-5 p.m. LIVE Online

The Analytics Clinic: Ensemble Models: Worth the Gains?
Sept. 20, 11 a.m. -12:30 p.m.


CAP® Exam computer-based testing sites are available in 700 locations worldwide. Take the exam close to home and on your schedule:

For more information, go to