Welcome!

Containers Expo Blog Authors: Elizabeth White, Carmen Gonzalez, Pat Romanski, Liz McMillan, Sematext Blog

Related Topics: Containers Expo Blog, Microservices Expo

Containers Expo Blog: Article

Five Ways Data Virtualization Improves Data Warehousing

Data virtualization fills the EDW agility gap

An array of business intelligence (BI), predictive analytics, data and content mining, portals and more tap a growing volume of information sourced from enterprise data warehouses (EDW).  However, significant volumes of business-critical enterprise data resides outside the enterprise data warehouse.  To deliver the most comprehensive information to business decision-makers, IT teams are implementing data virtualization to preserve and extend their existing enterprise data warehouse investments.

This article discusses five integration patterns that combine both enterprise data warehouses and data virtualization to solve real business and IT problems along with examples from Composite Software's data virtualization customers.  The five patterns include:

  1. Data Warehouse Augmentation
  2. Data Warehouse Federation
  3. Data Warehouse Hub and Virtual Data Mart Spoke
  4. Complementing the ETL Process
  5. Data Warehouse Prototyping

Maximizing Value from Enterprise Data Warehouse Investments
Supporting critical, yet ever-changing information requirements in an environment of ever-increasing data volumes and complexity is a challenge well understood by large enterprises and government agencies today.

This inexorable pressure has and will continue to drive the demand for enterprise data warehouses as an array of BI, predictive analytics, data and content mining, portals and other key applications rely on data sourced from enterprise data warehouses.

However, business change often outpaces enterprise data warehouse evolution.  And while useful for physically consolidating and transforming a large portion of enterprise data, significant volumes of enterprise data resides outside the confines of the enterprise data warehouse.  Further, enterprise data warehouses themselves require support throughout their lifecycles, driving demand for solutions that prototype, migrate, extend, federate and leverage enterprise data warehouse assets.

Data virtualization middleware, an advanced version of earlier data federation or enterprise information integration (EII) middleware, complements enterprise data warehouses by providing a range of flexible data integration techniques that preserve, extend and thereby drive greater business value from existing enterprise data warehouse investments.

1. Data Warehouse Augmentation
Organizations overwhelmed by scattered data silos and exponentially growing data volumes have deployed data warehouses to meet many of their reporting requirements.  However, a number of data sources remain outside the warehouse.  Providing users with complete business insight in support of revenue, cost and risk management goals often requires the following:

  • Historical data from the warehouse and up-to-the-minute data from transaction systems or operational data stores;
  • Summarized data from the warehouse and drill-down detail from transaction systems or operational data stores;
  • Master customer, product or employee data from an MDM hub or warehouse and detail from transaction systems or operational data stores; and
  • Internal data from the warehouse and external data from outside sources including cloud computing.

Data virtualization effectively federates data-warehouse information with additional sources, therefore extending existing data warehouse schemas and data.  These complementary views are conducive to adding current data to historical warehouse data, detailed data to summarized warehouse data, and external data to internal warehouse data.

Energy Company Combines Up-to-the-minute and Historical Data - To optimize deployment of repair crews and equipment across more than 10,000 production oil wells, an energy company uses data virtualization to federate real-time crew, equipment and well status data from their wells and SAP's maintenance management system with historical surface, subsurface and business data from their enterprise data warehouse.  The net result is faster repairs for more uptime and thus more revenue.

2. Data Warehouse Federation
A primary reason enterprises implement data warehouses is to overcome the various transaction and analytic system silos typical in most large enterprise and government agencies today.  However, for a number of often pragmatic reasons, the single "enterprise" data warehouse remains elusive.  Instead, for these same reasons, multiple data warehouses and data marts have been developed and deployed, in effect perpetuating, rather than overcoming, the data silo issue.

Optimizing business performance requires data from across these various warehouses and marts.   But physically combining multiple marts and warehouses into a singular and complete enterprise-wide data warehouse is often too costly and time consuming.

Data virtualization federates multiple physical warehouses.  Two examples include combining data from the sales and financial warehouses, or combining two sales data warehouses after a corporate merger. This approach achieves logical consolidation of warehouses by creating an integrated view across them, using abstraction to rationalize the different schema designs.

Investment Bank Federate Financial Trading Data Warehouses - To enable more flexible customer self-service reporting and meet SEC compliance reporting mandates, a prime brokerage uses data virtualization to federate equity, fixed income and other investment positions and trades information from siloed trading data warehouses.  The net result is higher customer satisfaction and lower reporting costs.

3. Data Warehouse Hub and Virtual Spoke
A typical data warehouse pattern is a central data warehouse hub with satellite data marts as spokes around the hub.  These marts use a subset of the warehouse data and are used by a subset of the data warehouse users.   Sometimes these marts are created because the analytic tools require data in a different form than the warehouse.  On the other hand, they may be created to work around the controls provided by the warehouse, and thus act as "rogue" data marts.  Regardless of the reason, every additional mart adds cost and compromises data quality.

Data virtualization provides virtual data marts that eliminate, or at least significantly reduce, the need for physical data marts around the data warehouse hubs.  This approach abstracts the warehouse data to meet specific consuming tool and user query requirements, while still preserving the quality and controls inherent in the data warehouse.

Mutual Fund Manager Eliminates "Rogue" Financial Data Marts - A mutual fund company uses data virtualization to enable more than 150 financial analysts to build portfolio analysis models with MATLAB® and other analysis tools leveraging a wide range of equity financial data from a 10 terabyte financial research data warehouse.  Prior to introducing data virtualization, analysts frequently spawned new satellite data marts with useful data subsets for every new project.  To accelerate and simplify data access and to stop the proliferation of costly, unnecessary physical marts, the firm instead used data virtualization to create virtual data marts formed from a set of robust, reusable views that directly accessed the financial warehouse on demand.  This enables analysts to spend more time on analysis and less on access, thereby improving portfolio returns.  The IT team has also eliminated extra, unneeded marts and all the costs that go with maintaining them.

4. Complementing the ETL Process
Extract, Transform, and Load (ETL) middleware is the tool of choice for loading data warehouses.  However, there are some cases where ETL tools are not the most effective approach.  Some examples include:

  • ETL tools lack interfaces to easily access source data, for example data from packaged applications such as SAP or new technologies such as web services;
  • Readily available, existing virtual views or data services can be reused rather than building new ETL scripts from scratch; and
  • Tight batch windows require access, abstraction and federation activities to be pre-processed and virtually staged in advance of ETL processes.

ETL tools can leverage data virtualization views and data services as inputs to their batch processes, appearing as another data source. This integration pattern also integrates data source types that ETL tools cannot easily access as well as reuse existing views and services, saving time and costs.  Further these abstractions do not require ETL developers to understand the structure of, or interact directly with, actual data sources, significantly simplifying their work and reducing time to solution.

Energy Company Preprocesses SAP Data - To provide the SAP financial data required for their financial data warehouse, an energy company uses data virtualization to access and abstract SAP R/3 FICO data.  This replaces an error-prone, SAP data-expert-intensive, flat-file-extraction process that would not scale across a complex SAP landscape.  The results include more complete and timely data in the financial data warehouse enabling better performance management.

5. Data Warehouse Prototyping
Building a new data warehouse from scratch is a large undertaking that requires significant design, development and deployment efforts.  One of the biggest issues is schema change, a frequent activity early in a warehouse's lifecycle.   This change process requires modification of both the ETL scripts and physical data in the warehouse and thus becomes a bottleneck that slows new warehouse deployments.  This problem does not go away later in the lifecycle; it just lessens as the pace of change slows.

Data virtualization middleware can be the platform for prototype development environment for a new data warehouse.  In this prototype stage, a virtual data warehouse is built, rather than a physical one, saving the time to build the physical warehouse.  This virtual warehouse includes a full schema that is easy to iterate as well as a complete functional testing environment.  Performance testing is somewhat constrained at this stage, however.

Once the actual warehouse is deployed, the views and data services built during the prototype stage still have value.  These are useful for prototyping and testing subsequent warehouse schema changes that arise as business needs or underlying data sources change.

Government Agency Prototypes New Data Warehouses - To reduce data warehousing time-to-solution for new data warehouse projects and changes to existing ones, a government agency uses data virtualization.  The time spent in getting the data right has proven to be four times faster than directly building the ETL and warehouse, even when the subsequent translation of these working views into ETL scripts and physical warehouse schemas is factored in.

Key Takeaways
As data sources proliferate, including many web-based and cloud computing sources outside the traditional enterprise data warehouse, enterprises and government agencies are deploying solutions that combine enterprise data warehouses and data virtualization to deliver the most comprehensive information to decision-makers.  The results are extended life to existing information system investments, greater agility for adding new BI and other analytic technologies, and less disruption from corporate activities such as mergers and acquisitions.

More Stories By Robert Eve

Robert Eve is the EVP of Marketing at Composite Software, the data virtualization gold standard and co-author of Data Virtualization: Going Beyond Traditional Data Integration to Achieve Business Agility. Bob's experience includes executive level roles at leading enterprise software companies such as Mercury Interactive, PeopleSoft, and Oracle. Bob holds a Masters of Science from the Massachusetts Institute of Technology and a Bachelor of Science from the University of California at Berkeley.

@ThingsExpo Stories
"We build IoT infrastructure products - when you have to integrate different devices, different systems and cloud you have to build an application to do that but we eliminate the need to build an application. Our products can integrate any device, any system, any cloud regardless of protocol," explained Peter Jung, Chief Product Officer at Pulzze Systems, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
The cloud promises new levels of agility and cost-savings for Big Data, data warehousing and analytics. But it’s challenging to understand all the options – from IaaS and PaaS to newer services like HaaS (Hadoop as a Service) and BDaaS (Big Data as a Service). In her session at @BigDataExpo at @ThingsExpo, Hannah Smalltree, a director at Cazena, provided an educational overview of emerging “as-a-service” options for Big Data in the cloud. This is critical background for IT and data professionals...
"Once customers get a year into their IoT deployments, they start to realize that they may have been shortsighted in the ways they built out their deployment and the key thing I see a lot of people looking at is - how can I take equipment data, pull it back in an IoT solution and show it in a dashboard," stated Dave McCarthy, Director of Products at Bsquare Corporation, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Fact is, enterprises have significant legacy voice infrastructure that’s costly to replace with pure IP solutions. How can we bring this analog infrastructure into our shiny new cloud applications? There are proven methods to bind both legacy voice applications and traditional PSTN audio into cloud-based applications and services at a carrier scale. Some of the most successful implementations leverage WebRTC, WebSockets, SIP and other open source technologies. In his session at @ThingsExpo, Da...
As data explodes in quantity, importance and from new sources, the need for managing and protecting data residing across physical, virtual, and cloud environments grow with it. Managing data includes protecting it, indexing and classifying it for true, long-term management, compliance and E-Discovery. Commvault can ensure this with a single pane of glass solution – whether in a private cloud, a Service Provider delivered public cloud or a hybrid cloud environment – across the heterogeneous enter...
"IoT is going to be a huge industry with a lot of value for end users, for industries, for consumers, for manufacturers. How can we use cloud to effectively manage IoT applications," stated Ian Khan, Innovation & Marketing Manager at Solgeniakhela, in this SYS-CON.tv interview at @ThingsExpo, held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA.
Today we can collect lots and lots of performance data. We build beautiful dashboards and even have fancy query languages to access and transform the data. Still performance data is a secret language only a couple of people understand. The more business becomes digital the more stakeholders are interested in this data including how it relates to business. Some of these people have never used a monitoring tool before. They have a question on their mind like “How is my application doing” but no id...
@GonzalezCarmen has been ranked the Number One Influencer and @ThingsExpo has been named the Number One Brand in the “M2M 2016: Top 100 Influencers and Brands” by Onalytica. Onalytica analyzed tweets over the last 6 months mentioning the keywords M2M OR “Machine to Machine.” They then identified the top 100 most influential brands and individuals leading the discussion on Twitter.
Information technology is an industry that has always experienced change, and the dramatic change sweeping across the industry today could not be truthfully described as the first time we've seen such widespread change impacting customer investments. However, the rate of the change, and the potential outcomes from today's digital transformation has the distinct potential to separate the industry into two camps: Organizations that see the change coming, embrace it, and successful leverage it; and...
Extracting business value from Internet of Things (IoT) data doesn’t happen overnight. There are several requirements that must be satisfied, including IoT device enablement, data analysis, real-time detection of complex events and automated orchestration of actions. Unfortunately, too many companies fall short in achieving their business goals by implementing incomplete solutions or not focusing on tangible use cases. In his general session at @ThingsExpo, Dave McCarthy, Director of Products...
The Internet of Things (IoT) promises to simplify and streamline our lives by automating routine tasks that distract us from our goals. This promise is based on the ubiquitous deployment of smart, connected devices that link everything from industrial control systems to automobiles to refrigerators. Unfortunately, comparatively few of the devices currently deployed have been developed with an eye toward security, and as the DDoS attacks of late October 2016 have demonstrated, this oversight can ...
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, discussed how research has demonstrated the value of Machine Learning in delivering next generation analytics to impr...
More and more brands have jumped on the IoT bandwagon. We have an excess of wearables – activity trackers, smartwatches, smart glasses and sneakers, and more that track seemingly endless datapoints. However, most consumers have no idea what “IoT” means. Creating more wearables that track data shouldn't be the aim of brands; delivering meaningful, tangible relevance to their users should be. We're in a period in which the IoT pendulum is still swinging. Initially, it swung toward "smart for smar...
20th Cloud Expo, taking place June 6-8, 2017, at the Javits Center in New York City, NY, will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy.
Businesses and business units of all sizes can benefit from cloud computing, but many don't want the cost, performance and security concerns of public cloud nor the complexity of building their own private clouds. Today, some cloud vendors are using artificial intelligence (AI) to simplify cloud deployment and management. In his session at 20th Cloud Expo, Ajay Gulati, Co-founder and CEO of ZeroStack, will discuss how AI can simplify cloud operations. He will cover the following topics: why clou...
Internet of @ThingsExpo, taking place June 6-8, 2017 at the Javits Center in New York City, New York, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @ThingsExpo New York Call for Papers is now open.
"ReadyTalk is an audio and web video conferencing provider. We've really come to embrace WebRTC as the platform for our future of technology," explained Dan Cunningham, CTO of ReadyTalk, in this SYS-CON.tv interview at WebRTC Summit at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Successful digital transformation requires new organizational competencies and capabilities. Research tells us that the biggest impediment to successful transformation is human; consequently, the biggest enabler is a properly skilled and empowered workforce. In the digital age, new individual and collective competencies are required. In his session at 19th Cloud Expo, Bob Newhouse, CEO and founder of Agilitiv, drew together recent research and lessons learned from emerging and established compa...
Data is the fuel that drives the machine learning algorithmic engines and ultimately provides the business value. In his session at Cloud Expo, Ed Featherston, a director and senior enterprise architect at Collaborative Consulting, discussed the key considerations around quality, volume, timeliness, and pedigree that must be dealt with in order to properly fuel that engine.
Everyone knows that truly innovative companies learn as they go along, pushing boundaries in response to market changes and demands. What's more of a mystery is how to balance innovation on a fresh platform built from scratch with the legacy tech stack, product suite and customers that continue to serve as the business' foundation. In his General Session at 19th Cloud Expo, Michael Chambliss, Head of Engineering at ReadyTalk, discussed why and how ReadyTalk diverted from healthy revenue and mor...