big data

2 Takeaways from the 2018 Spring Meetings by the International Monetary Fund (IMF) and the World Bank

Every year the IMF and the World Bank hold a conference-style event that is referred to as the Spring Meetings. These Spring Meetings bring together central bankers, ministers of finance and development, private sector executives and academics to discuss global issues such as global economy, international development and the world’s financial markets.

This year I had the opportunity to attend the 2018 Spring Meetings where discussions were held about threats and opportunities of technological changes as it affects global economies and policies. Here are 2 takeaways from the 2018 Spring Meetings focused on technology and innovation including some of my related articles:

  1. TECHNOLOGICAL ADVANCEMENTS AND JOBS
    •  Industrialization Paradigms
      • Typical Industrialization: Agriculture → Manufacturing → Services
      • Current Industrialization: Agriculture → Services
    • Impacts of Technology
      • Technological Changes → Job loss → Re-skill → New Jobs
      • Some jobs will never be recovered
      • Flow of technology and expertise doesn’t flow easily across countries
      • Even within countries technological impacts are uneven causing inequality
      • A good balance between data privacy and business models is needed that benefits societies at a larger scale
      • Depending upon where innovation (internal or external) to the organizations is can impact society at different levels
      • A good balance of foundations and advance education is needed
      • Specialized knowledge can negatively impact holistic societal impacts
    • Artificial Intelligence (AI)
      • Dystopian Views: AI will take over most human activities and would rule over humans
      • Middle Ground Views: AI will augment and enhance human activities but never replace humans
      • Utopian Views: AI will take over most human activities that would free up time for humans to do other things
    • The Brave New World of Data
      • Data quality issues are borderless
      • Standard data definitions of economic data has to be agreed upon and used
      • Data is being used to build economic policies
      • Data is being used to create multinational economic blocs
      • Data is being used to assess the humming of the global economy
      • Data Standardization and Harmonization àData Transparency àData Accountability
  1. PARTNERSHIPS
    • For economic prosperity, no organization, country, region is an island in of itself
    • Bridges need to be created across, public, private, academic, non-profit and shareholders
    • Regulations are slow to adopt to technological advancements and can be too heavy-handed or light-touch if not properly understood by policy makers
    • Grassroots changes are affecting how governments function and adapt
    • Technology and innovation should have executive level consideration across all branches of government and not just a ministry or a few people

Bonus: IMF’s Innovation Lab (iLab)

IMF has created the iLab whose goal seems to be to look at how technology and innovation is affecting the global economy and economic policies in various countries.

Related Articles:

  1. 5 QUESTIONS TO ASK ABOUT ARTIFICIAL INTELLIGENCE
  2. WHERE IS MY BIG DATA COMING FROM AND WHO CAN HANDLE IT
  3. 5 QUESTIONS TO ASK ABOUT YOUR INFORMATION
  4. 5 QUESTIONS TO ASK ABOUT PREDICTIVE ANALYTICS
  5. 5 QUESTIONS TO ASK ABOUT YOUR BIG DATA
  6. 5 QUESTIONS TO ASK ABOUT PRESCRIPTIVE ANALYTICS
  7. IDENTIFYING ORGANIZATIONAL MATURITY FOR DATA MANAGEMENT
  8. UNDERSTANDING AND APPLYING PREDICTIVE ANALYTICS
  9. 5 OBSERVATIONS ON BEING INNOVATIVE (AT AN ORGANIZATIONAL LEVEL)
  10. 5 OBSERVATIONS ON BEING INNOVATIVE (AT AN INDIVIDUAL LEVEL)
  11. HOW DO YOU COMMUNICATE?
  12. 5 QUESTIONS TO ASK ABOUT YOUR BUSINESS PROCESSES
  13. 75 QUESTIONABLE THOUGHTS ABOUT ORGANIZATIONAL TRANSFORMATION
  14. 35 CONCEPTS THAT AFFECT ORGANIZATIONAL TRANSFORMATION EFFORTS
  15. SPICE FOR BUSINESS TRANSFORMATION
  16. 5 QUESTIONS TO ASK ABOUT YOUR CULTURE

 

World Map Data

Advertisements

Understanding and Applying Predictive Analytics

Executive Summary

This article proposes looking at Predictive Analytics from a conceptual standpoint before jumping into the technological execution considerations. For the implementation aspect, organizations need to assess the following keeping in mind the contextual variances:

Strategies

Tactics

Operations

  • Top Down
  • Bottom Up
  • Hybrid
  • Organizational Maturity
  • Change Management
  • Training
  • Practical Implications
  • Pros and Cons of Technology Infrastructure
  • Providing Enabling Tools to Users
  • Best Practices

Describing Predictive Analytics

Predictive Analytics is a branch of data mining that helps predict probabilities and trends. It is a broad term describing a variety of statistical and analytical techniques used to develop models that predict future events or behaviors. The form of these predictive models varies, depending on the behavior or event that they are predicting. Due to the massive amount of data organizations are collecting, they are turning towards Predictive Analytics to find patterns in this data that could be used to predict future trends. While no data is perfect in predicting what the future may hold there are certain areas where organizations are utilizing statistical techniques supported by information systems at strategic, tactical and operational levels to change their organizations. Some examples of where Predictive Analytics is leveraged include customer attrition, recruitment and supply chain management.

Gartner describes Predictive Analytics as any approach to data mining with four attributes:

  1. An emphasis on prediction (rather than description, classification or clustering)
  2. Rapid analysis measured in hours or days (rather than stereotypical months of traditional data mining)
  3. An emphasis on the business relevance of the resulting insights (no ivory tower analyses)
  4. An (increasing) emphasis on ease of use, thus making tools accessible to business users

The above description highlights some important aspects for organizations to consider namely:

  1. More focus on prediction rather than just information collection and organization. Sometimes in organizations it is observed that information collection becomes the end goal rather than using that information to make decisions.
  2. Timeliness is important otherwise organizations might be making decisions on information that is already obsolete.
  3. Understanding of the end goal is crucial by asking why Predictive Analytics is being pursued and what value it brings to the organization.
  4. Keeping in mind that if the tools are more accessible to business users then they would have a higher degree of appreciation of what Predictive Analytics could help them achieve.

Relationship of Predictive Analytics with Decision Support Systems or Business Intelligence

University of Pittsburg describes Decision Support Systems as interactive, computer-based systems that aid users in judgment and choice activities. They provide data storage and retrieval but enhance the traditional information access and retrieval functions with support for model building and model-based reasoning. They support framing, modeling and problem solving. While Business Intelligence according to Gartner is an umbrella term that includes the applications, infrastructure and tools, and best practices that enable access to and analysis of information to improve and optimize decisions and performance. These descriptions point to the fact that Decision Support Systems or Business Intelligence are used for decision making within organization.

Interestingly, it seems like Predictive Analytics is the underlying engine for Decision Support Systems or Business Intelligence. What this means is the predictive models that result in Predictive Analytics could be under the hood of Decision Support Systems or Business Intelligence. It should be noted that organizations should proceed with caution with regards to the Decision Support Systems or Business Intelligence since if the underlying assumption are incorrect in making the predictive models then the decision making tools would be more harmful then helpful. A balanced approach would be to create expert systems where Decision Support Systems or Business Intelligence is augmented by human judgment and the underlying models are checked and verified periodically.

Implementation Considerations for Predictive Analytics

As the descriptions above have indicated that the aim of Predictive Analytics is to recognize patterns and trends that can be utilized to transform the organization. This requires organizations to firstly educate themselves on what value they want and what can be derived from Predictive Analytics. Predictive Analytics is about business transformation and it needs to show what value it brings to the organization. In this regard, we have to assess people, processes and technologies of the organization in terms of current state (where the organization is right now) and future state (where the organization wants to be). Typically, this revolves around Strategies, Politics, Innovation, Culture and Execution (SPICE) as shown below.

SPICE Factors

SPICE Factors

The assessment of people for Predictive Analytics means to understand what users will be leveraging Predictive Analytics and if they understand that simply relying on Predictive Analytics is not enough but in order to have an effective system they need to be part of the system. This means that the analytics insights need to be augmented by human expertise to make intelligent decisions. The assessment of processes for Predictive Analytics entails looking at how organizations make decisions right now and how future decisions would be made if Predictive Analytics is put into place. This includes having appropriate governance structures in place. The assessment of technology entails looking at what technologies exist within the organization and if they could be leveraged for Predictive Analytics. If not then looking at what Predictive Analytics products are in the market that would work for the organization and are they flexible enough in case the underlying assumptions for the predictive models change and when predictive models become obsolete.

The advanced techniques mentioned in the book, Seven Methods for Transforming Corporate Data into Business Intelligence would be applicable to Predictive Analytics. These methods are:

  1. Data-driven decision support
  2. Genetic Algorithms
  3. Neural Networks
  4. Rule-Based Systems
  5. Fuzzy Logic
  6. Case-Based Reasoning
  7. Machine Learning

Technologies Used for Predictive Analytics

Gartner has been publishing their Magic Quadrant on Business Intelligence and Analytics Platforms since 2006. Due to the increased importance of Predictive Analytics in the marketplace, Gartner decided to create a separate Magic Quadrant for Advanced Analytics Platforms which focuses on Predictive Analytics and published its first version on February 2014. Since it is the first version of the Magic Quadrant, all vendors listed are new and no vendors were dropped.

 

Gartner's Magic Quadrant for Advanced Analytics Platforms

Gartner’s Magic Quadrant for Advanced Analytics Platforms

As we can see from this Magic Quadrant that it includes well-known vendors but also vendors that are not as big or as well-known. It is interesting to note that open-source vendors such as RapidMiner (a Chicago company) and Knime (a European company) are in the same Leaders Quadrant as well-established vendors such as SAS and IBM. While there are some issues with these open-source vendors as stated in the report but perhaps this Magic Quadrant is also an indication of where the next generation of analytics would come from. Due to the very nature of open-source, there are more opportunities for cheaper customization which would give the organizations the flexiblity to be as granular as they want to be. Ofcourse code stablity and lack of proper documentation are issues that organizations need to be cognizant about. Organizations may also want to “try out” these open source tools before they make a big commitment to propertary software to see if Predictive Analytics is something they want to invest heavily in.

Using Predictive Analytics in Specific Industries

There are many industries that utilize Predictive Analytics. The organizations in these industries either use Predictive Analytics to transform their business and/or to address certain areas that they would like to improve upon. Following is a list of some of the industries that utilize Predictive Analytics:

Industry How is Predictive Analytics used?
Retail
  • Customer Retention
  • Inventory Optimization
  • Low-Cost Promotions
Oil and Gas
  • Well and Field Asset Surveillance
  • Production Optimization
  • Equipment Reliability and Maintenance
Automotive
  • Adjust production schedules
  • Tweak marketing campaigns
  • Minimize Inventory
Food
  • Human Resources Allocation
  • Supply Chain Optimization
Healthcare
  • Electronic Health Records
Government
  • Nation-wide Blood Levels
Social Media
  • New Business Models

While there are many examples of industries that have embraced Predictive Analytics but there are other industries that have not fully accepted it as a new reality. These industries have many excuses for not considering Predictive Analytics but typically revolve around scope, quality, cost and fear of the known. However, the tide might be changing for these industries as well since industry bloggers are beginning to insist how Predictive Analytics could be leveraged for competitive advantages.

My Opinion

Predictive Analytics can come in handy in making organizations analytical and becoming a better version of themselves. However, Predictive Analytics can be a deal-breaker if organizations have attempted and failed in the past and for this very reason Predictive Analytics should start as a discussion first. This discussion should revolve around asking which areas need improvements and among other things determine if Predictive Analytics could be something that could help. After a successful Predictive Analytics initiative other areas could be potential candidates as well.

An important thing to note is that Predictive Analytics is an organization-wide initiative that has touch points across the organization and thus the maturity of the organization has to be seriously considered prior to going on a Predictive Analytics journey. No matter how good Predictive Analytics can be for the organization but if the organization is not mature enough and it does not have the right governance, processes and feedback mechanisms in place then it might turn out to be another attempt at glory but nothing to show for it.

References:

  1. Predictive Analytics for Dummies
  2. Seven Methods for Transforming Corporate Data Into Business Intelligence
  3. IBM Journal Paper on A Business Intelligence System by H.P. Luhn
  4. Gartner report (G00258011) Magic Quadrant for Advanced Analytics
  5. Gartner IT Glossary on Predictive Analytics
  6. Gartner IT Glossary on Business Intelligence
  7. SAP Predictive Analytics
  8. Decision Support Systems by Marek J. Druzdzel and Roger R. Flynn
  9. 5 Questions to Ask About Predictive Analytics
  10. 5 Factors for Business Transformation

5 Questions to Ask About Your Big Data

In statistics, a hypothesis is proposed and then data samples are collected to prove or disprove the hypothesis with acceptable confidence levels. For example, lets say that all our customers are aware of all our product lines. Basically, there are two ways of assessing our hypothesis that includes: (1) Proving our hypothesis and (2) Disproving our hypothesis.

The first way to proving our hypothesis is that we communicate with all of our customers and inquire if they know all our product lines. The second way is to communicate with as many customers as possible until we come across any customer that does not know all our product lines. From this example, we can see that if we find even one customer then that disproves our hypothesis. Thus, this is the reason why in statistics, sometimes it is easier to find an exception to disproving a hypothesis rather than to proving it.

Big Data on the other hand, inverts the generally acceptable process from hypothesis then data sample collection to data collection then hypothesis. What this means is that Big Data emphasizes on collecting data first and then coming up with a hypothesis based on patterns found in the data. Generally speaking, when we talk about Big Data, we are concerned with the 3 Vs that include:

  • Volume – Amount of data
  • Velocity – Rate of data analysis
  • Variety – Different data sources

Some have indicated that we need to go beyond just the above three Vs and should also include:

  • Viscosity – Resistance to the flow of data
  • Variability – Changes in the flow changes of data
  • Veracity – Outlier data
  • Volatility – Validity of the data
  • Virality – Speed at which data is shared

I would take the Big Data concept a bit further and introduce:

  • Vitality – General and specific importance of the data itself
  • Versatility – Applicability of data to various situations
  • Vocality – Supporters of data-driven approaches
  • Veto – The ultimate authority to accept or reject Big Data conclusions

For a metrics-driven organization, a possible way to determine the effectiveness of your Big Data initiatives is to do a weighted rating of the Vs based on your organizational priorities. These organizational priorities can range from but not limited to increasing employee retention rates, improving customer experiences, improving mergers and acquisitions activities, making better investment decisions, effectively managing the organization, increasing market share, improving citizens services, faster software development, improving design, becoming more innovative and improving lives. What all of this means is that data is not just data but it is in fact an organization’s most important asset after its people. Since data is now a competitive asset, lets explore some of the ways we can use it:

  • Monte Carlo Simulations – Determine a range of scenarios of outcomes and their probabilities.
  • Analysis of Variance (ANOVA) – Determine if our results change when we change the data
  • Regression – Determine if data is related and can be used for forecasting
  • Seasonality – Determine if data shows the same thing occurring at the same intervals
  • Optimization – Getting the best possible answer from the data
  • Satisficing – Getting a good enough answer from the data

Now that we understand what is Big Data and how it can be used, lets ask the following questions:

Currently

In the Future

Who is capturing data? Who should be capturing data?
What is the lifecycle of your data? What should be the lifecycle of your data?
Where is data being captured? Where should data be captured?
When is data available for analysis? When should data be available for analysis?
Why data is being analyzed? Why data should be analyzed?

Having discussed the positives of Big Data, we have to realize that it is not a panacea and has its negatives as well. Some of the negative ways data can lead to bad decisions includes: (1) Data is correlated but that does not imply cause and effect, (2) Data shows you pretty pictures but that does not imply it is telling you the truth and (3) Biases can affect data anywhere from capturing to analysis to decision-making.

In conclusion, what this means is that the non-distorted quality, understanding and usage of data is the difference between just getting on the Big Data bandwagon or truly understanding how data can fundamentally change your organization.

Big Data Vs

References:

  1. Realizing the Promise of Big Data
  2. Beyond the three Vs of Big Data
  3. 5 Factors for Business Transformation
  4. 5 Questions to Ask About Your Business Processes
  5. 5 Questions to Ask About Your Information
  6. 5 Questions to Ask About Customer Experiences
  7. 5 Observations on Being Innovative (at an organizational level)
  8. Where is my Big Data coming from and who can handle it

Where is My Big Data Coming From and Who Can Handle It

Recently, a reader asked my insights on the article (Data Scientists are the New Rock Stars as Big Data Demands Big Talent).  Here is my response.

It seems like in today’s world people and organizations are somewhat struggling with this big data concept and do not know where to begin. Due to this reason they are collecting everything they can think of in the hopes that one day they will be able to use this data in a meaningful way such as better customer experience, new products/services, better collaboration, increasing revenue etc. This hope approach of “lets collect data and later decide what we can use it for” on the surface might seem sound but last I checked hope is not a strategy. Perhaps this is one of the reasons that even now only <1% of data collected is actually being analyzed. What good is more data when one cannot even make sense of the other 99%+ of data it already has. Are we chasing a ghost?

While it is true that vast amounts of data are and will be generated from financial transactions, medical records, mobile phones and social media to the Internet of Things but there are questions that need to be asked to understand data’s meaningful use:

  1. How will data be managed?
  2. How will data be shared?

I believe that in order to come to a point where data becomes meaningful and useful it would require (broadly speaking) three phases:

  1. Establishment of standards, governance, guidelines. (E.g., open architectures)
  2. Creation of industry specific data exchanges. (E.g., healthcare data exchanges, environment data exchanges etc.)
  3. Creation of cross-industry data exchanges. (E.g., healthcare data exchanges seamlessly interacting with environmental data exchanges etc.)

Additionally, lets keep this in mind that the data we are talking about is data that can be captured by current tools and systems but the data which is perhaps the most difficult to capture is unstructured human data which within organizations is called Institutional Knowledge. This does not reside in a document or a system but in the minds of the people of an organization who understand what needs to be done in order to move things forward.

So, the question becomes, do we really need Data Scientists who have a mix of coding skills with PhDs in scientific disciplines and business sense or do we need someone who is able to connect the dots and have the ability to create the future. The answer is not a simple one. Perhaps you need both. The ability to code should not be the deciding factor but rather the ability to leverage technology and data should be. I agree that there is shortage of people with diverse talent but there is also shortage of people who actually know how to leverage this kind of talent.

Before organizations go on a hiring spree they should consider:

  1. Why do they need a Data Scientist? (E.g., have strategic intent, jumping on the bandwagon etc.)
  2. Who will the Data Scientist report to? (E.g., Board, CEO, CFO, COO, CIO etc.)
  3. Does the organization have the ability to enhance/change its business model? (E.g., making customers happy, leading employees etc.)
  4. Is the Data Scientist really an IT person with advanced skills or does s/he have advanced skills and happens to know how to leverage technology and data?
  5. How often will you measure the relevancy of the data? (E.g., key data indicators)
3 Phases of Big Data Harmonization

3 Phases of Big Data Harmonization