• No results found

The Logical Data Warehouse


Academic year: 2021

Share "The Logical Data Warehouse"

Show more ( Page)

Full text









APRIL 18 2016

APRIL 19-20 2016

Incorporating Big Data,

Hadoop and NoSQL

in Data Warehouse and

Business Intelligence


The Logical

Data Warehouse

Design, Architecture,

and Technology


The Logical Data Warehouse



Classic Data Warehouse architectures are made up of a chain of databases. This chain consists of numerous databases, such as the staging area, the central Data Warehouse and several datamarts, and countless ETL programs needed to pump data through the chain. This architecture has served many organizations well. But is it still adequate for all the new user requirements and can new technology be use optimally for data analysis and storage? Integrating self-service BI products with this architecture is not easy and certainly not if users want to access the source systems. Delivering 100% up-to-date data to support operational BI is difficult to im-plement. And how do we embed new storage technologies, such as Hadoop and NoSQL, into the architecture? It is time to migrate gradually to a more flexible architecture in which new data sources can hooked up to the Data Warehouse more quickly, in which self-service BI can be supported correctly, in which OBI is easy to im-plement, in which the adoption of new technology, such as Hadoop and NoSQL, is easy, and in which the pro-cessing of Big Data is not a technological revolution, but an evolution. The architecture that fulfills all these needs is called the Logical Data Warehouse architecture. This architecture, introduced by Gartner, is based on a decoupling of reporting and analyses on the one hand, and data sources on the other hand. The technology to create a Logical Data Warehouse is available, and many organizations have already successfully completed the migration; a migration that is based on a step-by-step process and not on full rip-and-replace approach. In this practical seminar, the architecture is explained and products will be discussed. It discusses how organiza-tions can migrate their existing architecture to this new one. Tips and design guidelines are given to help make this migration as efficient as possible.


In this seminar Rick van der Lans answers the following questions:

• What are the practical benefits of the Logical Data Warehouse architecture and what are the differences with the classical architecture

• How can organizations step-by-step and successfully migrate to this flexible Logical Data Warehouse archi-tecture?

• You will learn about the possibilities and limitations of the various available products • How do data virtualization products work?

• How can big data be added transparently to the existing BI environment? • How can self-service BI be integrated with the classical forms of BI?

• How can users be granted access to 100% up-to-date data without disrupting the operational systems? • What are the real-life experiences of organizations that have already implemented a Logical Data Warehouse?


This seminar is intended for everyone who needs to be aware of developments in the field of Business Intelli-gence and Data Warehousing, such as BI architects, business analysts, Data Warehouse and database de-signers, database experts, consultants, technology planners, project managers, and system analysts. Some knowledge of the classical Data Warehouse architecture is required.




1. Challenges for the Classic Data Warehouse • Integrating Big Data with existing data and making it

available for reporting and analytics

• Supporting self-service BI and self-service data preparation

• Polyglot persistency – processing data stored in Hadoop and NoSQL systems

• Operational Business Intelligence, or analyzing of 100% up-to-date data

2. The Logical Data Warehouse

• The essence: decoupling of reporting and data sources • From batch-integration to on-demand integration of


• The impact on flexibility and productivity – an im-proved time-to-market for reports

• Examples of organizations

3. Implementing a Logical Data Warehouse with data virtualization servers

• Why data virtualization?

• Market overview: Cirro Data Hub, Cisco Information Server, Denodo Platform, Informatica Data Ser-vices, RedHat JBoss Data Virtualization, Rocket, and Stone Bond Enterprise Enabler

• Importing non-relational data, such as XML and JSON documents, Web Services, NoSQL, and Hadoop data • The importance of an integrated business glossary

and centralization of metadata specifications

• Performance improving facilities: Caching of virtual tables, refreshing, and optimization techniques 4. Migrating to a Logical Data Warehouse • An A to Z roadmap

• Guidelines for the development of a Logical Data Warehouse

• Three different methods for modeling: outside-in, in-side-out, and middle-out

• The value of a canonical data model • Considerations for security aspects

• Step by step dismantling of the existing architecture • The focus on sharing of metadata specifications for

integration, transformation, and cleansing

5. Self-Service BI and the Logical Data Warehouse • Why self-service BI can lead to “report chaos” • Centralizing and reusing metadata specifications

with a Logical Data Warehouse

• Upgrading self-service BI into managed self-service BI

• Implementing Gartner’s BI-modal environment 6. Big Data and the Logical Data Warehouse • New data storage technologies for Big Data,

includ-ing Hadoop, MongoDB, Cassandra

• The appearance of the polyglot persistent environ-ment; or each application its own optimal database technology

• Design rules to integrate Big Data and the Data Warehouse seamlessly

• Big Data is too “big” to copy

• Offloading cold data with a logical Data Warehouse 7. Implementing Operational BI with a Logical Data


• Examples of operational reporting and operational analytics

• Extending a Logical Data Warehouse with opera-tional data for real-time analytics

• “Streaming” data in a Logical Data Warehouse • The coupling of data replication and data virtualization 8. The Logical Data Warehouse and the


• Design principles to define data quality rules in a Logical Data Warehouse

• How data preparation can be integrated with a logi-cal Data Warehouse

• Shifting of tasks in the BICC

• Which new development and design skills are im-portant?

• The impact on the entire design and development process


Incorporating Big Data, Hadoop and NoSQL in DW and BI Systems



Big Data, Hadoop, in-memory analytics, self-service BI, Data Warehouse automation, analytical database servers, data virtualization, data vault, operational in-telligence, predictive analytics, and NoSQL are just a few of the new technologies and techniques that have become available for developing BI systems. Most of them are very powerful and allow for development of more flexible and scalable BI systems. But which ones do you pick?

Due to this waterfall of new developments, it’s be-coming harder and harder for organizations to select the right tools. Which technologies are relevant? Are they mature? What are their use cases? These are all valid but difficult to answer questions.

This seminar gives a clear and extensive overview of all the new developments and their inter-relation-ships. Technologies and techniques are explained, market overviews are presented, strengths and weaknesses are discussed, and guidelines and Best Practices are given.

The biggest revolution in BI is evidently Big Data. Therefore, considerable time in the seminar is re-served for this intriguing topic. Hadoop, MapReduce, Hive, NoSQL, SQL-on-Hadoop are all explained. In addition, the relation with analytics is discussed ex-tensively.

This seminar gives you a unique opportunity to see and learn about all the new BI developments. It’s the perfect update for those interested in knowing how to make BI systems ready for the coming ten years.


• In-depth overview of all the modules making up Hadoop, including HDFS, MapReduce, HBase, Storm, and Yarn

• Critical assessment of the SQL-on-Hadoop engines, including

• Application areas of Hadoop in BI systems: sand-box, offloading cold data

• Letting classic reporting and analytical tools access Big Data stored in Hadoop

• Transparently offloading Data Warehouse data to Hadoop using data virtualization servers

• Integrating an Hadoop-based sandbox for Data Scientists in BI systems

• Developing operational Business Intelligence sys-tems using Storm and other event processing en-gines

• Using NoSQL (MongoDB, Cassandra etc) or NewSQL (Clustrix, NuoDB, etc) as transaction sys-tems

• Moving data quality aspects to the business users by using self-service data preparation

• Data modeling for Big Data and Hadoop

• Implementing data lakes and data reservoirs with the right technology

• Comparison and overview of new data storage tech-nology


• Learn about the trends and the technological devel-opments related to Business Intelligence, analytics, Data Warehousing, and Big Data

• Discover the value of Big Data and analytics for or-ganizations

• Learn which products and technologies are winners and which ones are losers

• Learn how new and existing technologies, such as Hadoop, NoSQL and NewSQL, will help you create new opportunities in your organization

• Learn how more agile data Business Intelligence systems can be designed

• Learn how to embed Big Data and analytics in exist-ing Business Intelligence architectures


• Business Intelligence Specialists • Data Warehouse Designers • Business Analysts • Technology Planners • Technical Architects • Enterprise Architects • IT Consultants • IT Strategists • Systems Analysts • Database Developers • Database Administrators • Solutions Architects • Data Architects • IT Managers




1. The Changing World of Business Intelligence • Big Data: Hype or reality?

• Operational Intelligence: does it require online Data Warehouses?

• Data Warehouses in the Cloud • Self-service BI

• The business value of analytics

2. Overview of Analytical SQL Database Servers • Are classic SQL database servers more suitable for

Data Warehousing?

• Important performance improving features: column-oriented storage, in-database analytics

• Market overview of analytical SQL database servers 3. Hadoop explained

• The relationship between Big Data and analytics • The Hadoop software stack explained, including

HDFS, MapReduce, YARN, Hive, Storm, Sqoop, Flume, and HBase

• The balancing act: productivity versus scalability • Making Big Data available to a larger audience with

SQL-on-Hadoop engines, such as Apache Drill and Hive, CitusDB, Cloudera Impala, Hadapt, IBM BigSQL, JethroData, MemSQL, Pivotal HawQ, ScleraDB and Splice Machine

4. NoSQL explained

• Classification of NoSQL database servers: key-val-ue stores, document stores, column-family stores and graph data stores

• Market overview: CouchDB, Cassandra, Cloudera, MongoDB, and Neo4j

• Strong consistency or eventual consistency? • Why an aggregate data model?

5. Data Virtualization for Agile BI systems and Lean Integration

• Data Virtualization offers on-demand data integration • Seamlessly integrating Big Data and the Data Ware


• Market overview: Cirro Data Hub, Cisco/Composite Information Server, Denodo Platform, Informatica Data Services, RedHat Jboss Data Virtualization, Rocket, and Stone Bond Enterprise Enabler

• Importing non-relational data, such as XML docu-ments, Web services, NoSQL and Hadoop data, and unstructured data

• Differences between Data Virtualization and with data blending

6. New Business Intelligence Architectures

• Discussion of different BI architectures, including Kimball’s Data Warehouse Bus, Architecture, Inmon’s Corporate Information Factory, DW 2.0, the Federa ted Architecture, the Centralized Warehouse Archi -tecture, the Data Virtualization Archi-tecture, and the BI in the Cloud Architecture

• Do we still need Data Marts?

• What is the role of Master Data Management in BI architectures?

• Using data vault to create more flexible Data Ware -houses

• Data Warehouse automation to create Data Ware -houses and Data Marts faster

7. Operational Business Intelligence • Analytics at the speed of business

• Different forms of operational BI: operational report-ing, operational analytics, and embedded analytics • What is time-series analysis?

• Integrating operational and historical data

• The role of data replication, rule engines, complex event processing and ESBs

8. New Forms of Reporting and Analytics • Mobile BI, Exploratory analysis, self-service BI • Collaborative Analytics: the marriage of social

net-works and BI

• Tools for embedded analytics

• Investigative Analytics and the Data Scientist • R as the new open source platform for Analytics 9. Summary and Conclusions




Rome April 18, 2016

Residenza di Ripetta - Via di Ripetta, 231 Registration fee: € 700


Rome April 19-20 2016

Residenza di Ripetta - Via di Ripetta, 231 Registration fee: € 1300


Special price for the delegates who attend both seminars: € 1900

If anyone registered is unable to attend, or in case of cancellation of the seminar, the general conditions mentioned before are applicable.

first name ... surname ... job title ... organisation ... address ... postcode ... city ... country ... telephone ... fax ... e-mail ...

Send your registration form with the receipt of the payment to:

Technology Transfer S.r.l.

Piazza Cavour, 3 - 00193 Rome (Italy) Tel. +39-06-6832227 - Fax +39-06-6871102 info@technologytransfer.it

www.technologytransfer.it Stamp and signature



PARTICIPATION FEE The Logical Data Warehouse


Incorporating Big Data, Hadoop and NoSQL in DW and BI Systems


Special price for the delegates who attend both seminars: €1900

The fee includes all seminar documentation, luncheon and coffee breaks. VENUE Residenza di Ripetta Via di Ripetta, 231 Rome (Italy) SEMINAR TIMETABLE 9.30 am - 1.00 pm 2.00 pm - 5.00 pm HOW TO REGISTER

You must send the registration form with the receipt of the payment to:

TECHNOLOGY TRANSFER S.r.l. Piazza Cavour, 3 - 00193 Rome (Italy) Fax +39-06-6871102

within April 4, 2016 PAYMENT

Wire transfer to:

Technology Transfer S.r.l. Banca: Cariparma Agenzia 1 di Roma IBAN Code: IT 03 W 06230 03202 000057031348 BIC/SWIFT: CRPPIT2P546 GENERAL CONDITIONS DISCOUNT

The participants who will register 30 days before the seminar are entitled to a 5% discount.

If a company registers 5 participants to the same seminar, it will pay only for 4. Those who benefit of this discount are not entitled to other discounts for the same seminar.


A full refund is given for any cancellation received more than 15 days before the seminar starts. Cancellations less than 15 days prior the event are liable for 50% of the fee. Cancellations less than one week prior to the event date will be liable for the full fee.


In the case of cancellation of an event for any reason, Technology Transfer’s liability is limited to the return of the registration fee only.


Rick van der Lans is an independent consultant, author and lecturer. He specializes in Business Intelligence, Data Warehousing, and Service Oriented Architectures. He is managing director of R20/Consultancy. Mr. van der Lans is an internationally acclaimed lecturer. For the last 12 years, he has been presenting professionally, and has lectured in many of the European countries, South America, the USA, and in Australia. He has pre-sented many keynote speeches at international events. Mr. van der Lans is chairman of the Database Systems Show (organised annually in The Netherlands since 1984), he is columnist for two major newspapers in the Benelux, called Computable and DataNews. Additionally, he is advisor for magazines such as Software Release Magazine and Database Magazine. His popular books, including “Introduction to SQL” and “The SQL Guide to Oracle”, have been translated into many languages and have sold over 100,000 copies. Recently, he has published a very successful book on presentation skills.


Related documents

classificaon scheme/fileplan E-Discovery Seng agreed corporate retenon policies Long term archive Enterprise search Back-conversion of exisng paper records Implemenng records

As shown in Figure 4, we find very interesting patterns (which are also prevalent in other cases): 1) first two sentences (blue frame), which contain the main topic and idea of

The majority of them (49.2% of all first booking prompted participants) finally enter the offsetting program again as prompted participants in the second booking decision... If

Big Data Analytics – Every company needs both a Data Warehouse and a Discovery Platform to maximize big data business value. Big Data Management – Hadoop for landing storing

Figure 13: Reference Architecture Hive Hive bI tools Hadoop Data warehouse analytics Archive Management Data exchange (SAP Data Services) Data warehouse/database (SAP HANA®,

External Trane Devices Required Tracer Summit™, Tracer 100 System or Tracer Chiller Plant Control Ice-Making Systems Controls An ice-making option may be ordered with the

» Objective 1: Partner with IT and Media Services to reconfigure the first floor to allow more space for collaborative learning and to increase student engagement with

Three tribunal Chairs and three medical members were interviewed for the research about their views and experiences of using medical evidence in hearing incapacity benefit

A mediation model was tested with family stress and paternal sensitivity as mediating variables in the association between father’s lifetime depression or anxiety disorder and (1)

According to a recent Ventana Research report on big data technology, only 22% of 163 organizations that Ventana polled last year were using Hadoop, and 45% said they had no plans

Oracle ERP & CRM Solutions on Exadata Advanced Analytics, In- Memory, Big Data SQL Oracle Database Data Warehouse on Exadata ODI Big Data Connectors ODI..

In conclusion, for the studied Taiwanese population of diabetic patients undergoing hemodialysis, increased mortality rates are associated with higher average FPG levels at 1 and

Franckx, Laurent and Van Hyfte, Annick and Bogaert, Sarah and Vermoote, Stijn and Hunt, Alistair. ARCADIS

Players can create characters and participate in any adventure allowed as a part of the D&D Adventurers League.. As they adventure, players track their characters’

Strategy #4: Review and update comprehensive park plan Action Step #1: Identify financial costs of plan components Action Step #2: Research funding options. Action Step #3:

¶2 On appeal the City contends that the District Court erred in upholding the PCD’s decision that a certain channel adjacent to the Yellowstone River is part of the

Advanced Papers Foundation Certificate Trust Creation: Law and Practice Company Law and Practice Trust Administration and Accounts Trustee Investment and Financial

Result of MicroStrategy’s Big Data Analytics accessing data stored in Hadoop and a relational warehouse. Sales Associate and

Full-scale dynamic analysis of an innovative rockfall fence under impact using the discrete element method: from the local scale to the structure scale.. Full-scale dynamic analysis

Phonetic analysis of unsupervised state-based interpolation between Regional Standard Austrian German (RSAG) and Bad Goisern dialect (GOI)..

The main wall of the living room has been designated as a "Model Wall" of Delta Gamma girls -- ELLE smiles at us from a Hawaiian Tropic ad and a Miss June USC

patient should be carried out on a spontaneous speech sample. As a rule, speech intelligibility is markedly more impaired in the spontaneous speech flow than it is in a

Advances in medical treatment, management, diagnosis and surgical palliation have improved the quality and longevity of children born with Congenital Heart Disease. As