Introduction to Data Virtualization: Technology and Use Cases

Introduction to Data Virtualization: Technology and Use Cases




12 June 2019 (9.30-17.30)
Location: Parker Hotel (Diegem)
Presented in English by Rick van der Lans
Price: 720 EUR (excl. 21% VAT)
Register Now »

This event is history, please check out the List of Upcoming Seminars, or send us an email

Check out our related in-house workshops:

 Learning Objectives

What you will learn ?

This seminar provides you with answers to the following questions:

  • How to use Data Virtualization to integrate data in a more agile way
  • How to embed Data Virtualization in Business Intelligence systems
  • How to use Data Virtualization for integrating on-premised and Cloud applications
  • How to migrate to a more agile integration system
  • How Data Virtualization products work
  • How to avoid well-known pitfalls
  • How to learn from real-life experiences with Data Virtualization

Why do we organise this seminar ?

This seminar focuses on Data Virtualization. The technology is explained, advantages and disadvantages are discussed, products are compared, and use cases are discussed.

Data is increasingly becoming a crucial asset for organisations to survive in today’s fast moving business world. In addition, data becomes more valuable if enriched and/or fused with other data. Unfortunately, enterprise data is dispersed by most organisations over numerous systems all using different technologies. To bring all that data together is and has always been a major technological challenge.

In addition, more and more data is available outside the traditional enterprise systems. It's stored in big data platforms, in cloud applications, spreadsheets, simple file systems, in weblogs, in social media systems, and so on. and stored in traditional databases. For each system that requires data from several systems, different integration solutions are deployed. In other words, integration silos have been developed that over time has led to a complex integration labyrinth. The disadvantages are clear:

  • Inconsistent integration specifications
  • Inconsistent results
  • Decreased time to market
  • Increased development costs
  • Increased maintenance costs

The bar for integration tools and technology has been raised: the integration labyrinth has to disappear. It must become easier to integrate data from multiple systems, and integration solutions should be easier to design and maintain to keep up with the fast changing business world.

All these new demands are changing the rules of the integration game, they demand that integration solutions are developed in a more agile way. One of the technologies making this possible today is Data Virtualization.


Who should attend this seminar ?

This seminar is aimed at everyone who needs to stay up-to-date about the latest trends in business intelligence, such as:

  • BI and datawarehousing consultants,
  • datawarehouse and database developers,
  • database specialists and managers,
  • database-experts,
  • consultants,
  • technology planners,
  • BI project managers,
  • information analysts and system analysts

This course is a pre-conference workshop of the BI & Data Analytics Summit, a large conference in Belgium with 2 keynote speakers, 12 case-based presentations, and over 100 participants.

 Full Programme

8.30h - 9.30h
Registration and welcome of the participants with coffee/tea and croissants, and opportunity to network
9.30h
Introduction to Data Virtualization
  • What is data virtualization ?
  • Use case of data virtualization: business intelligence, data science, democratizing of data, master data management, distributed data
  • Differences between data abstraction, data federation, and data integration
  • Open versus closed data virtualization servers
  • Market overview: AtScale, Cirro Data Hub, Data Virtuality, Denodo Platform, FraXses, IBM Data Virtualization Manager for z/OS, Stone Bond Enterprise Enabler, and TIBCO Data Virtualization
 
How Do Data Virtualization Servers Work ?
  • The key building block: the virtual table
  • Integrating data sources via virtual tables
  • Implementing transformation rules in virtual tables
  • Stacking virtual tables
  • Impact analysis and lineage
  • Running transactions – updating data
  • Securing access to data in virtual tables
  • Importing non-relational data, such as XML and JSON documents, web services, NoSQL, and Hadoop data
  • The importance of an integrated business glossary and centralization of metadata specifications
11.00h
Coffee/tea, refreshments and opportunity to network
11.15h
Performance Improving Features
  • Caching of a virtual table for improving query performance, creating consistent report results, or minimizing interference on source systems
  • Differences between full refreshing, incremental refreshing, live refreshing, online refreshing and offline refreshing
  • Different query optimization techniques, including query substitution, pushdown, query expansion, ship joins, sort-merge Joins, statistical data and SQL override
 
Use Case 1: The Logical Data Warehouse Architecture
  • The limitations of the classic data warehouse architecture
  • On-demand versus scheduled integration and transformation
  • Making a BI system more agile with data virtualization
  • The advantages of virtual data marts
  • Strategies for adopting data virtualization
  • Application areas of data virtualization
  • The need for powerful analytical database servers
  • Migrating to a data virtualization-based BI system
 
Use Case 2: Data Virtualization and Master Data Management
  • How can data virtualization help with creating a 360° view of business objects
  • Developing MDM with a data virtualization server – from a stored to a virtual solution
  • On-demand data profiling and data cleansing
13.00u
Lunch Buffet together with the other group @itworks
14.00u
Use Case 3: From the Physical Data Lake to the Logical Data Lake
  • Practical limitations of developing one physical data lake
  • Shortening the data preparation phase of data science with data virtualization
  • Sharing metadata specifications between data scientists
  • Implementing analytical models inside a data virtualization server
 
Use Case 4: Democratizing Enterprise Data
  • Increasing the business value of the data asset by making all the data available to a larger group of users within the organisation
  • The business value of consistent data integration
  • Using lean data integration to make data available for analytics and reporting faster
  • One consistent data view for the entire organisation
  • How the business glossary and search features help business users
  • The coming of the data marketplace
 
Use Case 5: Dealing with Big Data
  • Big data can be too big to move - data can't be transported to the place of integration
  • Data virtualization pushes data processing to where the data is produced
  • Hiding the physical location of the data
  • With data virtualization, the network becomes the database
 
Closing Remarks
  • The Future of Data Virtualization
  • Data virtualization as driving force for data integration
  • Potential new product features
17.15h
Questions, summary and conclusions
17.30h
End of this one-day seminar

This course is a pre-conference workshop of the BI & Data Analytics Summit, a large conference in Belgium with 2 keynote speakers, 12 case-based presentations, and over 100 participants.

 Speakers


Rick van der Lans (R20/Consultancy BV)
R20/Consultancy BV

Rick van der Lans is a highly-respected independent analyst, consultant, author, and internationally acclaimed lecturer specialising in data warehousing, business intelligence, big data, and database technology.

He has presented countless seminars, webinars, and keynotes at industry-leading conferences. For many years, he has served as the chairman of the annual European Enterprise Data and Business Intelligence Conference in London and the annual Data Warehousing and Business Intelligence Summit in The Netherlands.

Rick helps clients worldwide to design their data warehouse, big data, and business intelligence architectures and solutions and assists them with selecting the right products. He has been influential in introducing the new logical data warehouse architecture worldwide which helps organisations to develop more agile business intelligence systems.

Over the years, Rick has written hundreds of articles and blogs for newspapers and websites and has authored many educational and popular white papers for a long list of vendors. He was the author of the first available book on SQL, entitled including Introduction to SQL, which has been translated into several languages with more than 100,000 copies sold. More recently, he published his book Data Virtualization for Business Intelligence Systems.

Questions about this ? Interested but you can't attend ? Send us an email !

-->