Wayne Eckerson Master Class: Data as a Product

| January 30, 2017

article image
Big data applications are complex. Even simple ones require 10 different technologies--and some require as many as 30! That’s why we’ve created our Business Analytics Master Class. We take the complexity of big data analytics and break it down into “consumable chunks.” Our first set of Master Class videos with industry expert Wayne Eckerson will be released on our blog over the coming weeks. Here's the second video in the series-Data as a Product. Wayne Eckerson: Hi, everyone.  We’re in a data driven economy now.  And data driven organizations are starting to realize that sometimes the data about a product is as valuable, if not more valuable, than the product itself.

Spotlight

REAN Cloud

REAN Cloud LLC, now part of Hitachi Vantara, is a leading Cloud Systems Integrator with deep experience supporting enterprise IT infrastructures and implementing continuous integration, continuous delivery pipelines. With its headquarters at Herndon, VA, the company has implemented complex and highly scalable architectures which support secure, compliant operations in highly regulated industries such as the Financial Services, Healthcare/Life Sciences, Education, and Public Sector verticals. The company’s team has extensive AWS expertise and DevOps experience that ensures quick, secure and reliable launch of clients’ solutions with no capital investments needed to procure hardware or services. REAN Cloud team has worked with global organizations including American Heart Association, Alexion Pharmaceuticals, Ditech Mortgage, Ellucian, Globus Genomics, PierianDx, SAP, SAP NS2, Symantec, Teradata and Veritas. REAN Cloud solutions are bundled with advanced security features to help address

OTHER ARTICLES

DATA CENTRE MARKET EXPECTED TO ACCELERATE OWING TO INCREASING CLOUD DEMAND

Article | February 28, 2020

An enormous amount of data is generated daily through various medium and amid this their storage becomes a great concern for organizations. Currently, two significant styles of data storage capacities are available Cloud and Data Centre.The main difference between the cloud vs. data centre is that a data centre refers to on-premise hardware while the cloud refers to off-premise computing. The cloud stores the data in the public cloud, while a data centre stores the data on company’s own hardware. Many businesses are turning to the cloud. In fact, Gartner, Inc. predicted that the worldwide public cloud services market has grown to 17.5 percent in 2019 to total US$214.3 billion. For many businesses, utilizing the cloud makes sense. While, in many other cases, having an in-house data centre is a better option. Often, maintaining an in-house data centre is expensive, but it can be beneficial to be in total control of computing environment.

Read More

A learning guide to accelerate data analysis with SPSS Statistics

Article | May 20, 2021

IBM SPSS Statistics provides a powerful suite of data analytics tools which allows you to quickly analyze your data with a simple point-and-click interface and enables you to extract critical insights with ease. During these times of rapid change that demand agility, it is imperative to embrace data driven decision-making to improve business outcomes. Organizations of all kinds have relied on IBM SPSS Statistics for decades to help solve a wide range of business and research problems.

Read More

MAKING IOT DATA MEANINGFUL WITH AI-POWERED COGNITIVE COMPUTING

Article | April 1, 2020

Today, the world is all about industry 4.0 and the technologies brought in by it. From Artificial Intelligence (AI) to Big Data Analytics, all technologies are transforming one or the other industries in some ways. AI-powered Cognitive Computing is one such technology that provides high scale automation with ubiquitous connectivity. More so, it is redefining how IoT technology operates.The need for Cognitive computing in the IoT emerges from the significance of information in present-day business. In the brilliant IoT settings of things to come. Everybody from new AI services companies to undertakings to use the information to settle on choices utilizing realities instead of impulses.Cognitive computing uses information and reacts to changes inside it to decide on better options. It is based on explicit gaining from past encounters, contrasted and a standard-based choice framework

Read More

Data Analytics Convergence: Business Intelligence(BI) Meets Machine Learning (ML)

Article | July 29, 2020

Headquartered in London, England, BP (NYSE: BP) is a multinational oil and gas company. Operating since 1909, the organization offers its customers with fuel for transportation, energy for heat and light, lubricants to keep engines moving, and the petrochemicals products. Business intelligence has always been a key enabler for improving decision making processes in large enterprises from early days of spreadsheet software to building enterprise data warehouses for housing large sets of enterprise data and to more recent developments of mining those datasets to unearth hidden relationships. One underlying theme throughout this evolution has been the delegation of crucial task of finding out the remarkable relationships between various objects of interest to human beings. What BI technology has been doing, in other words, is to make it possible (and often easy too) to find the needle in the proverbial haystack if you somehow know in which sectors of the barn it is likely to be. It is a validatory as opposed to a predictory technology. When the amount of data is huge in terms of variety, amount, and dimensionality (a.k.a. Big Data) and/or the relationship between datasets are beyond first-order linear relationships amicable to human intuition, the above strategy of relying solely on humans to make essential thinking about the datasets and utilizing machines only for crucial but dumb data infrastructure tasks becomes totally inadequate. The remedy to the problem follows directly from our characterization of it: finding ways to utilize the machines beyond menial tasks and offloading some or most of cognitive work from humans to the machines. Does this mean all the technology and associated practices developed over the decades in BI space are not useful anymore in Big Data age? Not at all. On the contrary, they are more useful than ever: whereas in the past humans were in the driving seat and controlling the demand for the use of the datasets acquired and curated diligently, we have now machines taking up that important role and hence unleashing manifold different ways of using the data and finding out obscure, non-intuitive relationships that allude humans. Moreover, machines can bring unprecedented speed and processing scalability to the game that would be either prohibitively expensive or outright impossible to do with human workforce. Companies have to realize both the enormous potential of using new automated, predictive analytics technologies such as machine learning and how to successfully incorporate and utilize those advanced technologies into the data analysis and processing fabric of their existing infrastructure. It is this marrying of relatively old, stable technologies of data mining, data warehousing, enterprise data models, etc. with the new automated predictive technologies that has the huge potential to unleash the benefits so often being hyped by the vested interests of new tools and applications as the answer to all data analytical problems. To see this in the context of predictive analytics, let's consider the machine learning(ML) technology. The easiest way to understand machine learning would be to look at the simplest ML algorithm: linear regression. ML technology will build on basic interpolation idea of the regression and extend it using sophisticated mathematical techniques that would not necessarily be obvious to the causal users. For example, some ML algorithms would extend linear regression approach to model non-linear (i.e. higher order) relationships between dependent and independent variables in the dataset via clever mathematical transformations (a.k.a kernel methods) that will express those non-linear relationship in a linear form and hence suitable to be run through a linear algorithm. Be it a simple linear algorithm or its more sophisticated kernel methods variation, ML algorithms will not have any context on the data they process. This is both a strength and weakness at the same time. Strength because the same algorithms could process a variety of different kinds of data, allowing us to leverage all the work gone through the development of those algorithms in different business contexts, weakness because since the algorithms lack any contextual understanding of the data, perennial computer science truth of garbage in, garbage out manifests itself unceremoniously here : ML models have to be fed "right" kind of data to draw out correct insights that explain the inner relationships in the data being processed. ML technology provides an impressive set of sophisticated data analysis and modelling algorithms that could find out very intricate relationships among the datasets they process. It provides not only very sophisticated, advanced data analysis and modeling methods but also the ability to use these methods in an automated, hence massively distributed and scalable ways. Its Achilles' heel however is its heavy dependence on the data it is being fed with. Best analytic methods would be useless, as far as drawing out useful insights from them are concerned, if they are applied on the wrong kind of data. More seriously, the use of advanced analytical technology could give a false sense of confidence to their users over the analysis results those methods produce, making the whole undertaking not just useless but actually dangerous. We can address the fundamental weakness of ML technology by deploying its advanced, raw algorithmic processing capabilities in conjunction with the existing data analytics technology whereby contextual data relationships and key domain knowledge coming from existing BI estate (data mining efforts, data warehouses, enterprise data models, business rules, etc.) are used to feed ML analytics pipeline. This approach will combine superior algorithmic processing capabilities of the new ML technology with the enterprise knowledge accumulated through BI efforts and will allow companies build on their existing data analytics investments while transitioning to use incoming advanced technologies. This, I believe, is effectively a win-win situation and will be key to the success of any company involved in data analytics efforts.

Read More

Spotlight

REAN Cloud

REAN Cloud LLC, now part of Hitachi Vantara, is a leading Cloud Systems Integrator with deep experience supporting enterprise IT infrastructures and implementing continuous integration, continuous delivery pipelines. With its headquarters at Herndon, VA, the company has implemented complex and highly scalable architectures which support secure, compliant operations in highly regulated industries such as the Financial Services, Healthcare/Life Sciences, Education, and Public Sector verticals. The company’s team has extensive AWS expertise and DevOps experience that ensures quick, secure and reliable launch of clients’ solutions with no capital investments needed to procure hardware or services. REAN Cloud team has worked with global organizations including American Heart Association, Alexion Pharmaceuticals, Ditech Mortgage, Ellucian, Globus Genomics, PierianDx, SAP, SAP NS2, Symantec, Teradata and Veritas. REAN Cloud solutions are bundled with advanced security features to help address

Events