Tuesday 6 May 2014

THE EVOLUTION OF ANALYTICS - PART 1


THE EVOLUTION OF ANALYTICS - Part 1

An analysis of the Business Intelligence Industry; past, present and predictions for the future.        

History of analytics

There are many examples throughout history that showcase how others have presented their data and findings to the world. The old adage that a picture is worth a thousand words rings true. For as long as people have been analysing data, they have been using visualisations to share their findings. Before the dawn of the computer age, these hand-drawn graphs were highly influential in the political and national arenas of their time. Florence Nightingale’s graphical illustration of the key causes of mortality during war, for example, showed at a glance the deaths of soldiers from preventable causes and led directly to improvements in military hospitals:



Charles Joseph Mindard’s graphic depicting the Russian campaign of 1812 showed the relationship between number of soldiers, falling temperatures and distance by soldiers, allowing for better planning on the part of military advisers:                      

                                                                 

William Playfair’s chart of 1821 compared weekly wages of a good mechanic with the price of a quarter of wheat, showing the decline in buying power of the labour force as part of his book showcasing the causes of the fall of powerful and wealthy nations:




With the onset of the industrial age, data analytics became a vital tool for business. From the first time management exercises conducted by Frederick Winslow Taylor in the late 19th century to the analytics utilised by Henry Ford’s assembly line to measure pacing of production, this field began to command more and more attention.

As computers became more prevalent in business, further developments here led to the creation of systems that would capture and make use of business data such as Enterprise Resource Planning Systems, Customer Relationship Management systems, data warehouses and a variety of hardware and software tools to further aid the cause.[i]

The Evolution of Analytics

Analytics has since grown more and more prominent. Today nearly all organisations have some sort of methodology to track and utilise their data as well as dedicated roles responsible for sustaining and managing this growth.

Analytics 1.0: Highly Scientific and only for the larger players

As more and more businesses embraced the power and competitive advantages that analytics could bring, it became obvious that a deep understanding of important business phenomena gave management a better ability to make decisions about the various processes in the organisation. It was during this era that the Enterprise Data Warehouse began to be used to capture information, with BI software developed to query and report on it.[ii]

BI has been the mainstream word used to describe the organisational intelligence software packages that are used by many companies to connect to their data. These packages come in all sorts of varieties from the simple to configure and implement to the more complicated and powerful. Out of this space a few market leaders emerged including IBM, SAS, SAP, Cognos and Qlikview.

Traditionally, the characteristics of BI environments in organisations were such that:
  • The software was configured, maintained, and administered by IT
  • Few users had broad flexibility to customize or create their own reports (most users were generally limited to pre-defined reports and prompts)
  • The vast majority of the reports generated contained some combination of grid-style data points and basic visuals such as line graphs, bar charts, and pie charts
  • The logic behind the reports was limited to what could be generated through standard SQL programming language constructs

By today’s standards, these limitations would not be acceptable to most users, but the restrictions were mainly a consequence of the power and availability of technology of the time. As the available technology has become more powerful, the demand for the outputs of BI systems has also grown, increasing the need to change how the environment works.

Analytics 2.0: The Social Media revolution and more break-throughs

The next evolution came with the advent of what is commonly referred to as “Big Data”. Firms began to amass large amounts of internet-based social media information in addition to their own internal data,[iii] giving them further insight into their customers. For the organisations that were equipped to properly analyse this data, it proved to be a valuable source of competitive advantage. During this time innovative technologies were created, acquired and mastered and revolutionary ways to handle the data volumes came about in both hardware and software technology.[iv] This includes the creation of Hadoop open software framework, cloud-based software environments, in-memory engines and NoSQL databases to name a few.

Today, the BI system has changed and is no longer governed by IT. Now there is more flexibility, and more users than ever are able to explore the data, discover new insights and share the results.

·         New, powerful BI tools breaking the restriction that users could only access the reports they were given
o   In terms of being able to experiment with new metrics and views of data, users are no longer constrained as they once were. Users now have the ability to explore their data (the term “data discovery” was coined to describe this process). All the while there are controllable limits on what data is accessible due to hardware and security limitations.

·         Advanced visualisations and interactive dashboards
o   No longer just bar and pie charts with static drill-downs – charts, graphs and maps are now explorable and can be linked to real-time data updates enabling faster insights

·         BI tools no longer constrained to just standard SQL programming language logic
o   More and more tools now have advanced analytic techniques including predictive analysis and are not restricted to standard SQL logic.  The use platforms like Hadoop and Teradata expanded the types of processing that can be applied and utilised to expand the avenues that can be explored with data.

This means the modern BI environment is no longer limited to just standard reports provided by a small team (usually connected to an IT department) but rather BI is becoming more of a self-service space, where visuals and interactivity are the norm. It leads to a blurring of the lines between IT and business users.[v]

Analytics 3.0: Empowering more users

Analytics 3.0 is seen as the next stage of the evolutionary chart for this industry and comes about when analytics becomes ingrained in almost all of an organisation’s actions. Regardless of whether that organisation makes or moves or consumes things, or produces or provides services, it will have access to information and data to report and analyse. With Analytics 3.0, the organisation can use the power of data analytics to create more valuable products and services.[vi]

In this era, the concepts of data discovery and exploration become even more important factors for organisational success, leading to greater empowerment for both internal and external users who can now “quickly plug-in, model, and analyse new data sources while still leveraging enterprise metadata and data”.[vii]

Much of the progression has been a consequence of better and faster hardware support systems, the advent of cloud computing and the move towards the availability of highly capable infrastructure via offerings such as infrastructure as a service (IAAS) and platform as a service (PAAS). With the new cloud hosted, browser-based software model, users are no longer reliant on the responsiveness of their organisation’s IT department, something that has traditionally been a source of great consternation for many users.[viii]

BI tools in general have improved and now feature self-service capabilities. The potential benefits of a self-service model include:
  •           Analysts have more time to concentrate on analysing reports as opposed to preparing them
  •           Users are empowered to discover data themselves, rather than relying on a reports team who might not have full understanding of the data
  •           Usability of reports has improved, especially for non-traditional BI users
  •           IT workload is reduced so they can concentrate on addressing any data requests more quickly and efficiently
Examples of self-service BI tools from tools like SuperDataHub, Tableau and Qlikview are shown below:









Data volumes increase dramatically

Recently there has been not only a staggering uptick in the volume of data produced and collected by businesses, but also a steady increase in the awareness of the power of data analytics.

The combined effect is that users are increasingly dissatisfied with prescriptive reports and dashboards that are handed down to them and that do not evolve. This is especially evident where the reports raise new questions, and users are then unable to obtain the answers quickly enough to take advantage of a market condition or situation.[ix]

Self-service analytics can thrive if it can keep up with user demand and is the key towards changing the perception of software from being a cost centre to being a fundamental underpinning to organisational success.

Lessons from the National Statistics Offices and the Open Data Movement

Before moving on it is important to take a look at how the providers of the biggest self-service platforms in the world have handled the growing demand for data and the lessons that can be learned.  Governments around the world have been leading the way in self-service analytics for some time, with the data they release via their National Statistical Organisations (NSOs).

Many of these NSOs have been using some sort of self-service portal to serve up the statistics they collect. This includes the Australian Bureau of Statistics, US Census, Office of National Statistics (UK) and many others. These portals mostly provide large amounts of data for dissemination purposes and use by researchers, statisticians and the public, who can access datasets if they have the appropriate accreditation.

Lessons can be learned from the way these organisations organise the protection of private information and balance that responsibility with their mandate to ensure that enough data is released so that valuable insights can be gained by researchers and the like.

A newcomer to the discussion is the Open Data Movement which, at its very essence, is about the release and dissemination of government datasets. It stems from the ideas of “Government 2.0”, which is defined as “the use of technology to encourage a more open, transparent and engaging form of government, where the public has a greater role in forming policy and has improved access to government information.”[x]

The Open Data Movement encourages the notion that government data should be freely available to everyone to use and republish as they wish, without restrictions from copyright, patents or other mechanisms of control. The goals of the Open Data Movement are similar to those of other "open" movements such as open source, open hardware, open content, and open access.

As John Wilbanks, VP Science at Creative Commons, says “numerous scientists have pointed out the irony that right at the historical moment when we have the technologies to permit worldwide availability and distributed process of scientific data, broadening collaboration and accelerating the pace and depth of discovery, we are busy locking up that data and preventing the use of correspondingly advanced technologies on knowledge.”[xi]

Open data sites are beginning to become more and more prominent around the world. They include the Data.Gov websites seen in many countries including the USA, Australia and UK.

There are great benefits from this model of openness for society and government. These include economic benefits from innovations made using the data as well as social benefits from more transparent governments.

The ability for data to help empower users is important but the potential for inappropriate use must be also considered.  This is especially true where the release of confidential data can lead to severe repercussions. 

To combat this, companies will usually prepare aggregated data to hide any confidential data but often this aggregated data cannot deliver the required abundance of information needed to gain insights.

In most environments, the time taken to protect and confidentialise data is time consuming, so a balance must be struck between what can safely be released and the time and costs required to prepare it.

Back to Analytics: What does all this mean?

Bearing in mind the trend towards self-service analytics, and the lessons learned from the Open Data Movement the next question we might ask is “where do we go from here?”

To attempt to answer this, we might ask what are the concerns of those involved in deploying analytical or BI solutions, and what limitations still exist with current generation software? How would these need to be addressed before it really does become the norm in all businesses and all industries?


MORE TO COME IN PART 2..... STAY TUNED

References mentioned above will all be added in PART 2

No comments:

Post a Comment