october 17-19 in portland, or - data modeling · october 17-19 in portland, or. ... er/studio sig...

35
October 17-19 in Portland, OR

Upload: vuongduong

Post on 16-May-2018

219 views

Category:

Documents


1 download

TRANSCRIPT

October 17-19 in Portland, OR

Data Modeling Zone 2016

Monday, October 17

Foundational

Data Modeling

Agile and

Communication

Big Data and

NoSQL

Hands-On and

Case Studies

Advanced

Data Modeling

DA

MA

CD

MP

Te

stin

g –

Ta

ke

an

y C

DM

P e

xa

m w

he

n y

ou

wa

nt, in

clu

din

g th

e d

ata

mo

de

ling

ex

am

! (Gre

en

Rm

)

7:00-9:00 Breakfast in the Governor Ballroom

8:30-11:30 Data Modeling

Fundamentals

Steve

Hoberman, Steve Hoberman

& Associates,

LLC Billiard, Pg 1

Data Quality

for Data

Modelers

Sue Geuens, President of

DAMA

International Fireside, Pg 1

An Overview of

NoSQL

Database

Joe Celko

Library, Pg 1

Hands-On CA

ERwin Data

Modeler Primer

Jeff Harris, Sandhill

Card, Pg 3

Analytic

Architecture

Modernization

– A Practical

Approach

Eddie Sayer, Teradata

Renaissance,

Pg 3

11:45-12:00 Welcome and Announcements Governor Ballroom

12:00-1:00 KEYNOTES:

Nike United

Ryan Smith and Suzanne Kimble, Nike

The Blockchain Billionaire

Steve Hoberman

1:00-2:15 Lunch in the Governor Ballroom

2:15-3:15

UML Made

Easy!

Norman

Daoust, Daoust

Associates Renaissance,

Pg 5

How to talk

Data to Non-

Data People

Jill Camper, DST Systems,

Inc. Card, Pg 6

Crossing the

Unstructured

Barrier

Bill Inmon, Forest Rim

Technologies Library, Pg 6

When little

meets BIG:

Practical Data

Modeling for

Analytics

Stacie Benton, Microsoft Billiard, Pg 6

Data Modeling

for Mobile

Data

Collection

Stephen

Gonzales, Reformatix,

LLC Fireside, Pg 7

3:15-3:45 Afternoon Snacks

3:45-4:45 Case Talk -

Data Modeling

by Example

Marco Wobben, BCP Software Fireside, Pg 12

Empowering

Business Users

to Lead with

Data

Denise

McInerney, Intuit Library, Pg 8

Enrichment

Modeling:

Dimensional

Modeling for

NoSQL

Datastores

Jim Cannaliato, Leidos Billiard, Pg 13

Creating

Ontologies from

your Existing

E/R Data

Models

Nicolas

Leveque, Deutsche Bank Card, Pg 9

Advanced

Aspects of

UML Data

Modeling

Michael

Blaha, Modelsoft

Consulting Renaissance,

Pg 10

5:00-6:00 Welcome Reception in the Governor Ballroom

Data Modeling Zone 2016

Tuesday, October 18

Foundational

Data Modeling

Agile and

Communication

Big Data and

NoSQL

Hands-On and

Case Studies

Advanced

Data Modeling DA

MA

CD

MP

Te

stin

g –

Ta

ke

an

y C

DM

P e

xa

m w

he

n y

ou

wa

nt, in

clu

din

g th

e d

ata

mo

de

ling

ex

am

! (Gre

en

Rm

)

7:00-9:00 Breakfast in the Governor Ballroom

7:45-8:15

SIGs

Bring Your

Model to Life

Gertjan Vlug, Attunity Library, Pg 10

“Zen with Len”

Meditation with

Len Silverston, Universal Data

Models Renaissance,

Pg 11

WhereScape –

Data

Warehouse

Automation

Douglas Barrett,

WhereScape

Billiard, Pg 30

ER/Studio SIG

Ron Huizenga, Idera Card, Pg 12

erwin

Modeling SIG

Danny

Sandwell, ERwin Fireside, Pg 31

8:30-11:30 FoCuSeD™

Business Data

Modeling Made

Easy

Gary Rush, MGR Consulting,

Inc. Billiard, Pg 13

Practical Tools

for the Human

Side of Data

Modeling

Len Silverston, Universal Data

Models Renaissance,

Pg 14

Data Science

Primer

Daniel D.

Gutierrez, AMULET

Analytics Library, Pg 15

Business-

Driven Data

Architecture - A

Practical,

Model-based

Approach

(Hands On)

Ron Huizenga, Idera Card, Pg 15

Only 362

days to

DMZ 2017!!

11:45-12:00 Welcome and Announcements Governor Ballroom

12:00-1:00 KEYNOTE: How To Punch A Shark In The Face, and other stuff

Gary Hall, Jr., Olympic Gold Medalist and Data Professional

1:00-2:15 Lunch in the Governor Ballroom

2:15-3:15

Modeling Time

Petr Olmer, GoodData Fireside, Pg 16

Beyond Data

Models: A

Metadata

Model for the

Data Curator

Dave Wells, Infocentric Renaissance,

Pg 17

Taxonomies

and Ontologies

Bill Inmon, Forest Rim

Technologies Library, Pg 17

How to compose

a fact-based

model into any

kind of schema

Clifford Heath, Infinuendo Card, Pg 27

Bridging

disparate IT

systems with

an Enterprise

Data Model

Brian Shive, Microsoft Billiard, Pg 18

3:15-3:45 Afternoon Snacks

3:45-4:45 Introduction to

Dimensional

Data Models

Petr Olmer, GoodData Fireside, Pg 19

The Importance

of Fun in Data

Modeling and

Success!

Len Silverston, Universal Data

Models Renaissance,

Pg 19

Better Data

Science

conversations

through

Conceptual

Data Modeling

Asoka Diggs, Intel Corp. Card, Pg 20

Caesars

Enterprise Data

Model

Patti Lee, Caesars

Entertainment

and Michael

Blaha, Modelsoft

Consulting Billiard, Pg 21

Scales,

Measurement

and Encoding

Scheme

Joe Celko

Library, Pg 21

5:00-6:30 Betting on Data Modeling with Wild Bill’s Casino! (Governor Ballroom)

Data Modeling Zone 2016

Wednesday, October 19

Foundational

Data Modeling

Agile and

Communication

Big Data and

NoSQL

Hands-On and

Case Studies

Advanced

Data Modeling

DA

MA

CD

MP

Te

stin

g –

Ta

ke

an

y C

DM

P e

xa

m w

he

n y

ou

wa

nt, in

clu

din

g th

e d

ata

mo

de

ling

ex

am

! (Gre

en

Rm

)

7:00-9:00 Breakfast in the Governor Ballroom

8:00-11:00 Introduction to

Linked Data

and the

Semantic Web

Cody Burleson, Base22. Library, Pg 29

Data

Management

Maturity – Why

We Need It and

How It Can

Propel You to

DM Leadership

Melanie Mecca, CMMI Institute Fireside, Pg 22

Modern Data

Governance:

Implications of

Agile, Big Data,

and Cloud

Dave Wells, Infocentric Renaissance,

Pg 23

Hands-On

PowerDesigner

Jeff Giles, Sandhill

Card, Pg 23

Advanced

Data Modeling

Challenges

Workshop

Steve

Hoberman, Steve Hoberman

& Associates,

LLC Billiard, Pg 24

11:15-12:15 Conceptual vs.

Logical vs.

Physical Data

Modeling - a

Contrarian

View

Gordon

Everest,

University of

Minnesota

Card, Pg 25

Even non-

relational

databases have

relationships

Pascal

Desmarets, Hackolade

Renaissance,

Pg 24

Machine

Learning

Primer

Daniel D.

Gutierrez, AMULET

Analytics Library, Pg 26

Only 362

days to

DMZ 2017!!

Essential Data

Modeling

David C. Hay,

Essential

Strategies, Inc.

Billiard, Pg 26

12:15-1:15 Lunch in the Governor Ballroom

1:15-4:15

Business

Information

Modeling using

the fact-based

approach

Clifford Heath, Infinuendo Renaissance,

Pg 27

Agile for Data

Professionals

Larry Burns, PACCAR Card, Pg 28

Data Modeling

in the NoSQL

World

Ted Hills, LexisNexis Library, Pg 8

Only 362

days to

DMZ 2017!!

UML in Depth

Norman

Daoust, Daoust

Associates Billiard, Pg 29

Data Modeling Zone 2016

Page 1

Data Modeling Fundamentals

Steve Hoberman, Steve Hoberman &

Associates, LLC

Assuming no prior knowledge of data modeling, we

start off with an exercise that will illustrate why

data models are essential to understanding

business processes and business requirements.

Next, we will explain data modeling concepts and

terminology, and provide you with a set of

questions you can ask to quickly and precisely

identify entities (including both weak and strong

entities), data elements (including keys), and

relationships (including subtyping). We will also

explore each component on a data model and

practice reading business rules. We will discuss

the three different levels of modeling (conceptual,

logical, and physical), and for each explain both

relational and dimensional mindsets.

Steve Hoberman has trained more than 10,000

people in data modeling since 1992. Steve is known

for his entertaining and interactive teaching style

(watch out for flying candy!), and organizations

around the globe have brought Steve in to teach his

Data Modeling Master Class, which is

recognized as the most comprehensive data

modeling course in the industry. Steve is the author

of nine books on data modeling, including the

bestseller Data Modeling Made Simple. One of

Steve’s frequent data modeling consulting

assignments is to review data models using his

Data Model Scorecard® technique. He is the founder

of the Design Challenges group and recipient of the

2012 Data Administration Management

Association (DAMA) International Professional

Achievement Award.

Data Quality for Data Modelers

Sue Geuens, President of DAMA

International

Data Quality is not generally a priority when you

start data modeling. The focus is on defining your

conceptual model or understanding of the business

requirements; parlaying that into a decent logical

model and then handing over to the DBAs or

physical DB modelers.

Unfortunately, that focus is ignoring the fact that

Data Quality is a primary driver in being able to

use the data the business has captured, created

and stored to provide meaningful business

intelligence that drives accurate and timely

business decisions.

Sue’s almost 20 years in data stands her in good

stead. She has been involved in many projects of

data modeling, designing and understand very

large databases and systems; has run a couple of

data quality projects and most recently finished a 2

year contract to implement a Data Governance

program at SA’s largest mobile operator. She is

currently on a 6 months project at the same

company designing and implementing a KPI

Metric model for the Commercial Operations

division (including Online and Self Service) and

this project has managed to unearth many data

quality anomalies.

This workshop will help data modelers understand

how to consider data quality actually MUST fit

into any data model – be it at the conceptual level

or right down in the nuts and bolts of the physical

model.

Typical discussions will be around the dimensions

of data quality, how to keep strict controls on the

data as you start to develop your models,

understanding how to get the business to specify

their data at a “Fit for Purpose” level – enabling

data modelers to manage their models and to build

them keeping the quality of the data as a key

priority. Further discussions will include why Sue

believes that primary keys, foreign keys, clearly

defined relationships and data attributes all

contribute to appropriate data quality. Finally, a

discussion on measuring how your data models

Data Modeling Zone 2016

stand up against good data quality and

governance. You should leave this workshop with a

clear understanding of what changes you may need

to bring to your future data modeling efforts to

improve the quality of the data your business

requires to make solid and innovative business

decisions for the future.

Sue is a Senior Data Management Specialist who

has been customer facing for the past 18 years.

During this time she has focused specifically in the

financial (banking, insurance, pensions) and

telecommunications sectors, gaining immense

knowledge and expertise in both. Each year she

attends a number of Data Management conferences

giving presentations both locally and overseas. Her

initial step into the world of data came about in the

form of designing and implementing the first

registration system for the NHBRC. Since then she

has moved on to various businesses and

enterprising, always working toward Data Quality

and Integrity, which is her passion. Sue was elected

President of DAMA SA during January 2009 and

was the driving force behind the Inaugural Meeting

which was held on 18th February 2009 at

Vodaworld in Midrand. Just completed

implementing Data Governance at a large SA

Telco, Sue has moved her focus to responding to the

many challenges facing SA companies with their

data. Sue has just been voted in as the DAMA I

President for the 2014/ 2015 term.

An Overview of NoSQL Database

Joe Celko

The traditional SQL database evolved from file

systems. It assumes that all data can be expressed

in scalar values in the columns of the rows of a

table, just as file systems used fields in records of

files. But the truth is that not all data fits into a

traditional SQL model.

The storage models have changed. Columnar

databases have no rows, but keep tables as

columns which can be assembled into rows if

needed. This means that query optimizations are

very different. The MapReduce Model and Key-

Value Stores do not have tables at all. Cloud

storage is not like traditional, local file systems.

Parallel processing and brute force replace clever

indexing optimization.

Traditional SQL assumes that the data is static;

think about common business applications like

accounting and transaction processing. Changes to

the data are explicit actions done by the database.

Actions against the data are taken in other tiers of

the application. But Streaming Databases (also

known as Complex Event models) assume that

data is constantly changing on its own. This would

include sensor monitoring, stocks, commodity

trades, and so forth. These events occur so fast

that responses have to be built into the database.

Graph Databases do not have a table concept at

all! They model pure relationships, expressed as a

graph structure. SQL assumes that we know the

relationships in the data. Graph databases

discover the relationships. They are used for social

network analysis, patterns in various kinds of

flows. Think about a SQL query to find the

minimal set of “the cool kids” in a social network

whose adoption of a product makes it into the next

trend.

Textbases or document databases have no concept

of structured data. Instead of syntax, they get

meaning from semantics. For example, Nexis,

Lexis and Westlaw are used by researchers and

lawyers to locate documents. Think about trying to

write an SQL query to determine if a TV dinner

meets the contractual obligation of a company to

provide off-site employees with a meal equivalent

to a fast-food order.

One of Dr. Codd’s 12 rules for RDBMS systems

was that there must be a linear language to control

Data Modeling Zone 2016

Page 3

and access the database. But Geographical Data

(GIS) deal with maps, so they are 2 or 3

dimensional by nature. Their query languages are

often graphic tools. For example, you can draw a

polygon on a map and then ask how many people

live inside it.

SQL is not always the right answer; it depends on

the question!

Joe Celko served 10 years on ANSI/ISO SQL

Standards Committee and contributed to the SQL-

89 and SQL-92 Standards. He is author of eight

books on SQL for Morgan-Kaufmann: SQL for

Smarties (1995, 1999, 2005, 2010), SQL Puzzles &

Answers (1997, 2006), Data & Databases (1999)

and Trees & Hierarchies in SQL (2004), SQL

Programming Style (2005) and Analytics & OLAP

in SQL (2005) and Thinking in Sets (2008). He has

written over 1200 columns in the computer trade

and academic press, mostly dealing with data and

databases.

Hands-On CA ERwin Data Modeler

Primer

Jeff Harris, Sandhill

This is an intensive short course on the CA ERwin

Data Modeler toolset. It is designed to help users

understand some of the more advanced features of

the tool and how to get better usage of your

existing CA ERwin Data Modeler investment. It

will assist users in being more productive and

leverage the re-usability features of ERwin Data

Modeler. This will be an instructor led course with

examples, illustrations and student exercises.

An existing knowledge of the CA ERwin Data

Modeler is advised, as this course will not be

covering the basics, though for those that are

interested in investing in the tool, we welcome you

to attend to understand the power of the tool.

What you will learn:

The benefits and workings of Domains,

The benefits and workings of Model

Naming Standards,

Also covering some dynamics of Name

Hardening,

The benefits and workings of Template

Models,

And we will explore some aspects of how to

work towards Data Modeling Standards to

achieve Data Governance.

Jeff Harris is Sandhill Consultants’ technical

services manager in North America and comes with

over 20 years of IT experience and has specialized

in Data Modeling & Data Architecture for more

than 11 years. He headed up the Sandhill

Consultants’ business in South Africa and due to

the great success he achieved there, was invited to

come to the USA to bring the same passion and

drive to replicate the same growth here. His

experience has led him to work on key projects for

some large corporations in the implementation of

ERwin, associated products, data modeling

methodologies and frameworks. These projects

involving larger geographical footprints, ranging

from South Africa, New Zealand and the USA has

brought satisfaction to several clients in all

industries, ranging from the banking industry,

government sectors and petroleum industry.

Data Modeling Zone 2016

Analytic Architecture Modernization

– A Practical Approach

Eddie Sayer, Teradata

The data and analytics landscape is evolving at an

astonishing pace. The number of ‘big data’

technology alternatives is staggering. The

momentum of open-source Hadoop is undeniable.

The marketplace offers the promise of modernized

analytic architectures to increase analytic agility,

optimize TCO and drive greater business value.

Success requires your role evolve from

conventional ‘data architect’ to ‘data ecosystem

architect’. This workshop provides a structure to

help you make the transition. We will explore

practical steps for modernizing your analytic

architecture by leveraging emerging technologies

and employing data architecture patterns for

acquisition, provisioning, integration and access.

Best practices will be discussed for architecting the

modern-day analytic ecosystem by implementing

data lakes and data products, as well as

integrating heterogeneous data components, such

as relational databases, NoSQL databases and

open-source Hadoop. You will leave the session not

only better educated on analytic architecture key

themes and concepts, but armed with a practical

approach for modernizing the analytic architecture

in your own organization.

For over two decades, Eddie has been helping large

organizations gain sustainable competitive

advantage with data. He has worked at length in

various roles including enterprise data

management, enterprise architecture, data

modeling and data warehousing.

Eddie joined Teradata in 2008 and has since

conducted numerous engagements with clients,

helping to set direction for data management. Prior

to joining Teradata, Eddie was a Data Architect at

CheckFree Corporation, the largest provider of e-

billing and electronic bill payment in the US.

Previously, Eddie held similar positions at Macys

Systems & Technology and Inacom Corporation.

Eddie is a founding member of the Georgia chapter

of Data Management International (DAMA) and is

a frequent speaker at industry events.

Nike United

Ryan Smith and Suzanne Kimble, Nike

No results were guaranteed, in a recent effort to

align all business functions to a single future state

vision for product data. Yet in partnership with

new business leadership and with an appreciation

for the strategic value of information, the business

and tech teams successfully prioritized and

accomplished this. Asking “what if our product

data were as good as our products,” they persisted

to arrive at broad acceptance of a substantially

improved common logical data model to describe

Nike’s innovative and iconic product lines. The

team will speak to how this future vision was

developed, elevated as a priority business

conversation, and brought to agreement and

relevance for the enterprise. Ryan Smith is an

Information Architecture Director and Suzanne

Kimble is a Data Architect, both dedicated to

Product Line Planning and Merchandising.

Ryan Smith is an Information Architecture

Director focused on Nike’s Line Planning and

Data Modeling Zone 2016

Page 5

Merchandising data. With a prior background in

Healthcare and Automotive, he now seeks to

maximize the value of enterprise modeling and

well-managed product master data, to enable the

product vision to be clearly understood and

executed throughout the company.

Suzanne Kimble is a Data Architect, also at Nike,

with prior experience at Con-way and Mentor

Graphics. She has over 20 years of IT experience,

including data architecture, user experience design

and technical training.

The Blockchain Billionaire

Steve Hoberman

Expect a game changer at least once a decade. A

technology that disrupts the status quo and ideally

improves our lives in terms of efficiencies or

general well-being. Blockchain is this decade’s

game changer. We’ve past the hype cycle and now

both startups and well-established organizations

are focusing on practical uses of a replicated

general ledger beyond Bitcoin. Learn about

blockchain and its three levels of maturity. Play a

game to understand blockchain’s benefits. Identify

the opportunities available to us within data

management. Overcome the challenges that lie

ahead to become the next blockchain billionaire!

Steve Hoberman has trained more than 10,000

people in data modeling since 1992. Steve is known

for his entertaining and interactive teaching style

(watch out for flying candy!), and organizations

around the globe have brought Steve in to teach his

Data Modeling Master Class, which is

recognized as the most comprehensive data

modeling course in the industry. Steve is the author

of nine books on data modeling, including the

bestseller Data Modeling Made Simple. One of

Steve’s frequent data modeling consulting

assignments is to review data models using his

Data Model Scorecard® technique. He is the founder

of the Design Challenges group and recipient of the

2012 Data Administration Management

Association (DAMA) International Professional

Achievement Award.

UML Made Easy!

Norman Daoust, Daoust Associates

An introduction to the thirteen UML diagram

types and their relationship to data modeling.

We’ll focus on those most relevant to data

professionals. The presentation includes examples

of each of the thirteen diagram types from a case

study.

Attendees will learn:

which UML diagram type is closest to a

data model

which UML diagram type includes entity

names from your data model

which UML diagram type visually

illustrates the allowable state changes of

an entity from your data model

when to use each of the diagram types

Norman Daoust founded his consulting company

Daoust Associates, www.DaoustAssociates.com in

2001. His clients have included the Centers for

Disease Control and Prevention (CDC), the

Veteran’s Health Administration, the Canadian

Institute for Health Information, a Fortune 500

software company, and several start-ups. He has

been an active contributor to the healthcare

industry standard data model, the Health Level

Data Modeling Zone 2016

Seven (HL7) Reference Information Model (RIM)

since its inception. He enjoys introducing data and

process modeling concepts to the business analysis

community and conducting business analysis

training courses. Norman’s book, “UML

Requirements Modeling for Business Analysts”

explains how to adapt UML for analysis purposes.

How to talk Data to Non-Data People

Jill Camper, DST Systems, Inc.

Do you ever find it frustrating to talk about data to

those who just don’t seem to know how to “talk

data”? This session will give specific tips on how to

talk to data to non-data people so that you can

both be on the same page and get your projects and

ideas approved.

Jill has over 16 years of data management

experience in the financial services industry,

including high profile conversions. For the past 7

years she has been focused on data design in both

the mainframe and open systems world including

traditional RDMBS design as well as Data

Warehouse, Star Schemas, and BI oriented designs.

She loves educating people on things data and the

importance of data in our everyday lives. Jill has

her bachelor’s degree in Psychology and her MBA

and is also a Certified Data Management

Professional.

Crossing the Unstructured Barrier

Bill Inmon, Forest Rim Technologies

The most exciting advances in technology have

been made in the arena of incorporating textual

data into the corporate decision making process.

This presentation addresses the reality of textual

exploitation of medical records, call center

information, restaurant and hotel feedback

analysis, and other arenas where text is found.

Bill Inmon – the “father of data warehouse” – has

written 53 books published in nine languages. Bill’s

latest adventure is the building of technology

known as textual disambiguation – technology that

reads raw text in a narrative format and allows the

text to be placed in a conventional database so that

it can be analyzed by standard analytical

technology, thereby creating unique business value

for Big Data/unstructured data. Bill was named

by ComputerWorld as one of the ten most

influential people in the history of the computer

profession. For more information about textual

disambiguation refer to www.forestrimtech.com.

When little meets BIG: Practical Data

Modeling for Analytics

Stacie Benton, Microsoft

More organizations are trying to get value using

Big Data but questioning the return on investment

in new technologies, processes, and skills they may

not have. Understanding and defining an

organization’s “little” data can make the difference

between finding insight or drowning in data

swamps. This session will cover what makes big

Data Modeling Zone 2016

Page 7

data different, define the complementary

differences between data modeler/analyst and data

scientist, and show practical examples on how data

modeling can accelerate the value from Big Data

trends, including a case study on how data

modeling on projects can help focus efforts and

help drive important business decisions and

impact.

Stacie Benton, CDMP, is currently an Information

Architect for Microsoft’s Enterprise Services

Business. She also recently spent time as the

Enterprise Information Architect for the Bill &

Melinda Gates Foundation working on building a

data culture as well as launching a Data

Governance program. She spent a tour of duty in

the consulting world as well as many years at

Microsoft, slaying dragons, both transactional and

BI, across the enterprise as well as in the Finance,

Sales, Services, HR, and Marketing spaces. She

spent years as a program manager in Microsoft IT

and has the awards and scars to prove it. She loves

talking data and process in the same sentence as

well as realizing the full potential of data through

good data management.

Data Modeling for Mobile Data

Collection

Stephen Gonzales, Reformatix, LLC

We can easily see how today’s mobile devices and

smartphones improve our productivity in the

workplace by allowing immediate interaction with

our organizations’ information assets whether in

the office or not. Besides the basic checking of

email and calendar via mobile device, common

business desktop applications such as timekeeping,

expense reporting have gone mobile as well! You

can be assured that more business applications

will become mobile-enabled through apps or web-

based applications.

So how can the discipline of data modeling adapt

in this new world of mobile computing? This

session will provide alternative insights into the

ubiquitous entity-relationship diagram by focusing

on often-overlooked model characteristics. You will

be introduced to table categorization, comparative

strengths of relationships, and table affinity

concepts. Additionally, a set of model analysis

techniques will be introduced that leverage the

aforementioned concepts to help identify atomic

business transactions to help get into the mobile

application perspective. Finally, the topic of data

storage for mobile data collection will be covered

whether using traditional RDBMS’s or persisting

structured data in a big data environment.

You will learn:

How data modelers can participate in the

development process of mobile applications

How highly-normalized data models

provide an advantage for efficient data

collection

How to leverage often overlooked entity

and relationship characteristics to gain a

deeper understanding of the data model in

terms of natural data flow

Stephen L. Gonzales has worked in the IT field for

nearly 20 years with both large and small

businesses, government and non-profits, including

Dell, ADP/Digital Motorworks, Habitat for

Humanity RE-store, Waste Management, The State

of Texas, Oracle Retail, Advanced Micro Devices,

and General Motors.

Data Modeling Zone 2016

Stephen has extensive professional experience with

software development methodologies, database

design, data analysis, data cleansing and

migration efforts. Having worked in both software

development and data warehouse environments,

Stephen has focused on designing efficient data

capture and storage solutions. He began his data

modeling career in 1996 as a staff consultant with

Information Engineering Systems Corp. (now

Visible Systems Corp.). With an active interest in

model-driven rapid application development

strategies, he founded Reformatix, LLC in 2011 to

pursue that aim.

Empowering Business Users to Lead

with Data

Denise McInerney, Intuit

The most valuable people in your organization

combine business acumen with data savviness. But

these data heroes are rare. Denise McInerney

describes how she has empowered business users

at her company to make better decisions with data

and explains how you can do the same thing in

your organization.

Businesses are only as good as the best decisions

they make. The business users who drive these

decisions are all working to make the best

decisions they can. Having access to the right data,

in the right way, can propel these business experts

in the right direction, and equip them to make

decisions that will make the business successful.

Data professionals with the technical know-how to

empower these users are true partners in any

organization’s success. Denise McInerney describes

how she has empowered business users at her

company to make better decisions with data and

explains how you can do the same thing in your

organization.

Attendees will learn how to:

Maximize their existing investments in

people and tooling

Meet their data consumers at their

starting point

Visualize their business from the

perspective of their business users

Inspire their business users to help drive

their data journey

Bridge the gap between the technologist

and business users views of the data

Create an environment of quality data that

inspires confidence in business users

Get past “tool snobbery”

Denise McInerney is a data professional with over

17 years of experience. Denise began her career as a

database administrator managing and developing

databases for online transactional systems. She

now works as a data architect at Intuit, where she

designs and implements BI and analytics solutions

with a focus on enabling the work of analysts and

business users. Denise is active in technical-

community organizations and has spoken at user

groups and conferences over two dozen times. She is

currently serving as the vice president of marketing

for PASS, an international organization of data

professionals. She is also a Microsoft Most

Valuable Professional. Denise has been a

passionate proponent of advancing the cause of

women in technology for over a decade. She has

organized events, spoken, and written on the topic.

She founded the Women in Technology chapter

of PASS and is active in the Tech Women @ Intuit

organization.

Data Modeling in the NoSQL World

Ted Hills, LexisNexis

The venerable entity-relationship (E-R) modeling

technique is very mature for SQL databases, but

Data Modeling Zone 2016

Page 9

doesn't translate well to many of the NOSQL

databases. This interactive workshop will

introduce the Concept and Object Modeling

Notation (COMN, pronounced "common"), which

can represent the new structures that NOSQL

DBMSs support, while still supporting traditional

SQL DBMSs. It also has more expressive power for

modeling the real-world entities that databases are

about.

Topics to be covered include:

modeling notation for containment and

multi-valued attributes

how to model schema-less databases

how to model the problem space and the

solution space, and the mapping between

them.

Ted has been active in the Information Technology

industry since 1975, moving gradually up through

device drivers and operating systems to

communications software, applications, and finally

information architecture. Past employers include

Bank of America, AT&T Bell Laboratories, Dow

Jones, and Bloomberg. At LexisNexis, Ted co-leads

the work of establishing enterprise data

architecture standards and governance processes,

working with data models and business and data

definitions for both structured and unstructured

data. Prior to joining LexisNexis, Ted was the

Enterprise Information Architecture Executive for

Bank of America, where he led several enterprise

reference data and data warehouse projects. Ted’s

work spans the spectrum from conceptual design

through database and software implementation.

Ted has always been an active researcher, with

interests in software and data integration, data

modeling notations, and improving the

expressiveness of languages while keeping them

efficient and type-safe.

Creating Ontologies from your

Existing E/R Data Models

Nicolas Leveque, Deutsche Bank

During the last two generations, visionary people

like Codd, Chen, Kent, Blaha, Wells, Inmon,

Hoberman contributed to the development and

success of the Entity Relationship model and its

implementation by modern RDBMS vendors.

During this period, thousands of students, scholars

and professionals were taught the three data

modeling perspectives (Conceptual, Logical and

Physical), and data management has become one

of the most lucrative, job creating, and ever

changing facet of the IT industry.

The last couple of years, our industry has seen the

advent of a new modeling technique that facilitates

the integration of heterogeneous data sources by

resolving semantic heterogeneity between them:

Ontology.

This presentation will explain in detail how to

leverage all of the numerous E/R Data models

produced and turn them into an OWL ontology

model. This session will explain the links between

ontology and E/R model constructs, and will help

you with learning and creating your first ontology

from their existing E/R data models.

So, if, as with Deutsche Bank, your company has

started embracing Linked Data and ontology

modeling, come and learn how to re-use what you

did and turn your E/R Data Models into a basic

OWL ontology.

Data Modeling Zone 2016

Nicolas Leveque is Head of Physical Data

Architecture, Deutsche Bank, and a Data Architect

with 20 years experience mainly within financial

institutions including UBS, Barclays, Societe

Generale, Capital Group and Deutsche Bank. He

specializes in Enterprise Data Architecture, Data

Modeling and in bridging the gap between High

Level architect and implementation teams. Still a

hands-on technician, he doesn’t believe in one-

solution-fits-all, but strives to find the right

solution depending on business and non-functional

requirements. He also looks forward and sees how

new practices and technologies could help to solve

data problems with a different and better mindset.

Advanced Aspects of UML Data

Modeling

Michael Blaha, Modelsoft Consulting

The UML class model is the UML model that is

pertinent to data modeling. It has its advantages

and disadvantages. In general, a data model

expressed in the UML notation excels at reaching

business audiences and also helps to foster deep,

abstract thinking. The primary downsides are that

the UML notation doesn’t address database design

and is less familiar to database developers.

This session will presume that the audience is

familiar with basic UML concepts. We will take a

deep dive and explore some advanced aspects. In

particular we will discuss qualified associations,

association classes, aggregation, composition, and

operations. We will explain their precise meaning,

discuss corresponding database designs, and give

compelling business examples.

Michael Blaha is a consultant, author, and trainer

who specializes in conceiving, architecting,

modeling, designing, and tuning databases. He has

worked with dozens of organizations around the

world. Blaha has authored seven U.S. patents,

seven books, many articles, and two video courses.

He received his doctorate from Washington

University in St. Louis and is an alumnus of GE

Global Research in Schenectady, New York. You

can find out more about him at superdataguy.com.

Bring Your Model to Life

Gertjan Vlug, Attunity

Organizations can implement their Business

Intelligence solution automatically, based on

corporate data models and business rules. No

programming or coding is needed anymore, nor is

technical modeling of optimized structures for

Data Warehouses and Data Marts Databases. This

saves time and money, and increases quality and

consistency. In general, modelers are working on

Staff-level. They define the model, business rules,

meta-data, etc. They hand it over to different IT

teams, including BI / DW teams. The BI / DW

teams start to work and, at best, use the document

from Staff as a guideline. They often lose track

with the model and start making changes without

any alignment to the model anymore.

Model-Driven Data Warehousing solves these

issues. Based on the model, the complete

implementation (Design of Data Warehouse (3NF

and/or Data Vault) and Data Marts (Star-

schemas), plus the generation and execution of

ETL-code) is automated. And when changes need

to apply, a simple “update” of the model adjusts

the Data Warehouse automatically.

Creating a logical data model gives you a

Data Warehouse in just a few clicks. With

consistent management information available,

technical complexity and challenges are no longer

delaying your business advantage. Better

alignment between the business and IT can also be

achieved through this model-driven approach.

From a Corporate perspective this will bring more

Data Modeling Zone 2016

Page 11

control and consistency at the implementation of

business rules, meta-data standards and correct

definitions of KPIs. From a developer’s perspective

the backlog of work is decreased, unsatisfied and

impatient customers are no longer frustrated, and

you don’t have to redo your own work over and

over for each change. The technical, time-

consuming and error-prone activities are

automated for you.

You finally have achieved an agile Data

Warehouse: Flexible, Incremental, Fast, Cost

effective. With shorter time to market due to the

efficiency and productivity improvements, you gain

a significant competitive advantage. Data

modeling not only ensures that your business

intelligence solution will be built more quickly and

effectively, but that the underlying information

that drives them is correct and fully documented.

Usher in a new era in Data Warehouse automation

by automatically designing and generating the

Enterprise Data Warehouse code for Third Normal

Form and/or Data Vault as well as Star Schema

data marts and staged data for Business

Intelligence projects.

Gertjan Vlug is the director of Data Warehouse

Automation at Attunity where he heads the

corporate development of next generation of ETL

tools that are fundamentally changing how

analytics platforms are built. Gertjan has a long

history of driving innovations in Business

Intelligence (BI) dating back to 1980’s. His

involvement in BI industry includes executive

management roles at Cognos (now part of IBM),

Business Objects (now part of SAP) and several

European consulting agencies. As an entrepreneur

he built 3 different companies whose focus were to

solve BI challenges. His latest, BIReady, was

recently acquired by Attunity. Prior to starting

BIReady, Gertjan was the founder of

QuickIntelligence and VisionProof.

“Zen with Len”

Meditation, Qigong (moving meditation),

and Talking Meditation (The what,

why and how of meditation)

Len Silverston, Universal Data Models

Prepare yourself and your mind for the day so you

can make the most of it! Come invigorate yourself,

reduce stress, develop your mind, and learn about

and practice meditation.

Len Silverston, who is not only a data modeling

thought leader, but is also a fully ordained Zen

priest and spiritual teacher, will provide this brief

overview of what meditation is, why it is

important, how to meditate, and lead a sitting

meditation and moving meditation (Qigong)

session. This will be an enlightening, wonderful

session to start your day in a relaxed and receptive

state of mind!

Len Silverston is a best-selling author, consultant,

and a fun and top rated speaker in the field of data

modeling, data governance, as well as human

behavior in the data management industry, where

he has pioneered new approaches to effectively

tackle enterprise data management. He has helped

many of the largest organizations world-wide as

well as small organizations, to integrate their data,

Data Modeling Zone 2016

systems and even their people.He is well known for

his work on “Universal Data Models”, that are

described in his The Data Model Resource Book

series (Volumes 1, 2, and 3), (Volume 1 was rated

#12 on the Computer Literacy Best Seller List) and

these books have been translated into multiple

languages. He is the winner of the DAMA (Data

Administration Management Association)

International Professional Achievement Award for

2004 and the DAMA International Community

Award for 2006. He has received the highest

speaker ratings at many international conferences

and is dedicated to being of the greatest service to

his audiences.

Enrichment Modeling:

Dimensional Modeling for

NoSQL Datastores

Jim Cannaliato, Leidos

Many have rejoiced at the arrival of "schema free"

NoSQL data stores, appearing to bring joy to

developers by relieving them of the burden of doing

data modeling. However, reports on the death of

data modeling have been greatly exaggerated.

Data modeling approaches such as dimensional

data modeling and star schemas were incredibly

useful in relational stores, and are now with

Enrichment Modeling, are even more powerful

than ever in NoSQL data stores. Learn how to go

beyond traditional dimensional modeling for

BigData stores, such as Hadoop, MongoDb or

ElasticSearch, and see firsthand examples of this

in action with DigitalEdge.

Jim Cannaliato is a Technical Fellow at Leidos,

and previously a Vice President at SAIC. He has

over 30 years of software development and

architecture experience, building complex

intelligence systems for our nation’s defense. For

the past ten years, he has been building large scale

database systems using a dimensional data

modeling approach. He received his Bachelor of

Science in Applied Science and Technology in

Computer Science from Thomas Edison State

college, and is a certified SCRUM Master.

ER/Studio SIG

Ron Huizenga, Idera

Hear the latest features available in ER/Studio

and ask the Idera team any questions you have on

ER/Studio!

Ron Huizenga is the Senior Product Manager for

the Idera ER/Studio product family. Ron has over

30 years of experience as an IT executive and

consultant in Enterprise Data Architecture,

Governance, Business Process Reengineering and

Improvement, Program/Project Management,

Software Development and Business Management.

CaseTalk - Data Modeling by Example

Marco Wobben, BCP Software

Q: Data modeling is described as a craft and once

completed the results may even seem artful. Yet

outsiders may see data modeling as abstract, time

consuming or even unnecessary. In many cases the

data modeler interviews business experts, studies

piles of requirements, talks some more, and then,

hocus pocus, presents a diagram with boxes, crows

feet, arrows, etc… Then the slow process begins to

Data Modeling Zone 2016

Page 13

keep the diagrams up to date, explain what the

diagrams behold, and sometimes even data

modelers themselves may get lost while

maintaining a growing set of data models and

requirements.

A: Fact based information modeling is the very

opposite of abstract. Fact based information

modeling uses natural language which expresses

facts that are intelligible for both business and

technical people. It does not require people to

understand the modeler’s magical language of

boxes and arrows. Although models can be

presented in several diagramming notations, they

can be validated in natural language at all times.

This gives both data modelers, technically skilled

people, and business people the benefit of having a

well documented and grounded data model.

Therefore the method of Fact Oriented Modeling, is

also known as "Data Modeling by Example".

Presentation Highlights:

key elements of fact oriented modeling;

data modeling with facts;

visualizing the model;

validating and verbalizing;

transforming and generating output (E.g.:

SQL, Relational, UML, XSD,

PowerDesigner, etc.).

Marco Wobben is director of BCP Software and has

been developing software well over 30 years. He has

developed a wide range of applications from

financial expert software, software to remotely

operate bridges, automating DWH generating and

loading, and many back- and front office and web

applications. For the past 10 years, he is product

manager and lead developer of CaseTalk, the CASE

tool for fact based information modeling, which is

widely used in universities in the Netherlands and

across the globe.

FoCuSeD™ Business Data Modeling

Made Easy

Gary Rush, MGR Consulting, Inc.

This interactive session is geared to enable data

analysts to facilitate Data Modeling sessions with

business clients. Gary will show you how to

facilitate, step-by-step, a data modeling workshop

and what skills or tools are needed at each step. It

is a brief summary of Gary’s book, FoCuSeD Data

Modeling Made Easy. Attendees will learn:

How to build a Data Model with business

clients who have never seen a data model.

How to use the modeling session to clarify

and re-engineer the business.

How Active Listening affects the model and

how to effectively listen to your clients.

How to harness the collective knowledge of

your clients to build a data model that they

embrace. How to make the model truly

their model.

Gary Rush, IAF CPF, Founder and President of

MGR Consulting, Inc., attended the U.S. Naval

Academy and is a former Chair of the International

Association of Facilitators (IAF). He is a recognized

leader in the field of Facilitation and Facilitator

training, managing projects since 1980, facilitating

since 1983, and providing Facilitator training since

1985; and continues to be the leading edge in the

industry by continuing as a practicing Facilitator.

Data Modeling Zone 2016

As a Facilitator Trainer, he teaches FoCuSeD. He

teaches specific “how to” with an understanding of

the “why” to perform as an effective Facilitator; he

provides detailed Facilitator and process tools,

enhances his training through effective learning

activities, and, as an IAF CPF Assessor, he covers

the IAF Core Facilitator Competencies and what

students need to do to achieve them. As a

Facilitator, he improves client business

performance through effective application of

exceptional facilitation processes and he is highly

skilled at engaging participants and guiding them

to consensus. Gary has written numerous “how to”

books, including the FoCuSeD Facilitator Guide – a

comprehensive reference manual sharing his step-

by-step process so that students can replicate his

practices. His alumni often tell us how much Gary

has changed their lives.

Practical Tools for the Human Side of

Data Modeling

Len Silverston, Universal Data Models

Success in data modeling requires skills in

communications, gaining sponsorship and buy-in,

negotiation, understanding needs, and leadership.

Yet many data modeling classes do not focus on

these essential aspects of effective data modeling.

This interactive, informative and fun class focuses

on the human elements of data modeling. Len

Silverston, best-selling data modeling author,

leads this seminar and provides tools, techniques,

principles, interactive exercises and case studies

that illustrate the importance of the human side to

data modeling.

This seminar will address the behavioral side of

data modeling in many different contexts such as

data modeling for big data, dimensional modeling,

traditional data modeling and more.

Participants of this seminar will learn about:

Inevitable human scenarios that data

modelers are bound to face and how to

address them

Personal, political and cultural factors that

are critical to successful data modeling

Practical tools, techniques, and principles

to enable successful data modeling such as

keys in gaining and sustaining funding,

realizing a common vision, how to get to

common semantics, developing trust,

communications, involvement, leadership,

and managing conflict

Stories of how using these principles were

the key to success for some organizations

and how other organizations had huge

challenges stemming from human

behavioral issues

This class will share fundamental and practical

tools while being an informative and fun

experience – come join it!

Len Silverston is a best-selling author, consultant,

and a fun and top rated speaker in the field of data

modeling, data governance, as well as human

behavior in the data management industry, where

he has pioneered new approaches to effectively

tackle enterprise data management. He has helped

many of the largest organizations world-wide as

well as small organizations, to integrate their data,

systems and even their people. He is well known for

his work on “Universal Data Models”, that are

described in his The Data Model Resource Book

series (Volumes 1, 2, and 3), (Volume 1 was rated

#12 on the Computer Literacy Best Seller List) and

these books have been translated into multiple

languages. He is the winner of the DAMA (Data

Administration Management Association)

International Professional Achievement Award for

2004 and the DAMA International Community

Award for 2006. He has received the highest

speaker ratings at many international conferences

and is dedicated to being of the greatest service to

his audiences.

Data Modeling Zone 2016

Page 15

Data Science Primer

Daniel D. Gutierrez, AMULET Analytics

Data science involves understanding and

preparing the data, defining the statistical

learning model, and following the Data Science

Process. Statistical learning models can assume

many shapes and sizes, depending on their

complexity and the application for which they are

designed. The first step is to understand what

questions you are trying to answer for your

organization. The level of detail and complexity of

your questions will increase as you become more

comfortable with the data science process.

In this session, I will cover the most important

steps in the data science process – a general

formula followed by data scientists in striving to

achieve best practices with a data science project:

understanding the goal of the project, data access,

data munging, exploratory data analysis, feature

engineering, model selection, model validation,

data visualization, communicate the results and

deploy the solution to production.

Daniel D. Gutierrez is a practicing data scientist

through his Santa Monica, Calif. consulting firm

AMULET Analytics. Daniel also serves as

Managing Editor for insideBIGDATA.com where

he keeps a pulse on this dynamic industry. He is

also an educator and teaches classes in data

science, machine learning and R for universities

and large enterprises. Daniel holds a BS degree in

mathematics and computer science from UCLA.

Business-Driven Data Architecture -

A Practical, Model-based Approach

(Hands-On)

Ron Huizenga, Idera

We are in the midst of a re-awakening, in which

the business value of data is once again being

recognized. This is occurring against a backdrop of

continuously evolving technology, an explosion of

diverse data sources, and exponential growth in

volume. To truly enable our organizations, we

require a practical, business-driven approach to

effectively understand, define and utilize data

effectively. We will discuss a practical approach to

business-driven data architecture and modeling

with several concepts including Business Data

Objects, data and business process model

structure, change management, and business

collaboration. This will be a hands-on session

utilizing examples and demos with ER/Studio

Enterprise Team Edition.

Ron Huizenga is the Senior Product Manager for

the Idera ER/Studio product family. Ron has over

30 years of experience as an IT executive and

consultant in Enterprise Data Architecture,

Governance, Business Process Reengineering and

Improvement, Program/Project Management,

Software Development and Business Management.

KEYNOTE: How To Punch A Shark In

The Face, and other stuff

Gary Hall, Jr., Olympic Gold Medalist and

Data Professional

Gary Hall, Jr. is a washed up swimmer, inducted

into the Olympic Hall of Fame. He has ten

Data Modeling Zone 2016

tarnished Olympic medals. As an Advisory Group

member to the Aspen Institute’s Sports & Society

Project Play program, Gary leads an initiative that

aims to extend health insurance incentives for

youth sport enrollment. He is the Member and

Alliance Manager for T1D Exchange, an innovative

model accelerating clinical research through the

use of data. He oversees a partnership with the

American College of Sports Medicine addressing

the management of chronic conditions in sport and

physical activity. He was recently published in the

British Journal of Sports Medicine as a co-author

of the International Olympic Committee consensus

statement on Youth Athlete Development. He

serves on the Leadership Board for the National

Youth Sports Health & Safety Institute, the

Advisory Board to the University of Arizona

Department of Surgery and the International

Children’s Board of Sanford Health.

Gary is a world recognized patient advocate,

having twice testified before Congress on health

care related issues. In April, he addressed a select

group of 300 scientists, ethicists and researchers at

the Vatican conference on cellular therapy,

immediately following the Pope’s address. Also,

while spearfishing in the Florida Keys, Gary

punched the shark that attacked his sister.

Gary will talk about the Olympic experience, as

Team USA returns from Rio. He’ll share how a

right brained athlete with extensive chlorine

exposure came to appreciate the value of research

and data. And he’s agreed to tell the shark story.

Modeling Time

Petr Olmer, GoodData

I will provide an introduction to the challenges of

modeling, managing, and querying temporal data

in relational and document-oriented databases.

Current, uni-temporal and bi-temporal data will be

covered, as well as working with valid time and

transaction time, temporal relations, logic, TSQL2,

and SQL:2011. See examples of temporal features

of SQL Server, PostgreSQL, and MongoDB.

Petr Olmer studied multi agent systems, artificial

intelligence, and declarative programming. He saw

big data for the first time while working at

Computer Centre at CERN, The European

Laboratory for Particle Physics. Today he works at

GoodData, building tools and defining architecture

and methodology for BI implementations.

Data Modeling Zone 2016

Page 17

Beyond Data Models: A Metadata

Model for the Data Curator

Dave Wells, Infocentric

As the world of data management grows and

changes, the role of data curator becomes

increasingly visible and significant. A curator is a

person responsible for management and

custodianship of a collection of things. Just as a

museum curator oversees a museum’s asset

collection, or a gallery curator is responsible for a

collection of art objects, a data curator has similar

responsibility for a collection of data assets.

Curation – whether art, artifacts, or data – implies

active and ongoing management of assets through

their entire lifecycle of interest and usefulness.

Acquisition, description, preservation, location,

appropriate use, presentation, and sharing are all

curation responsibilities.

Curating is an asset management activity, and

asset management requires data about the assets

under management. For data asset management,

that data is metadata. Current metadata

approaches, however, offer more support for data

creators and consumers than for data curators. To

support the increasingly important role of data

curator we must extend current metadata models

to include information such as data origins,

interest communities, sharing constraints,

presentation guidelines and practices, preservation

considerations, and more.

In this presentation you will learn:

The what and why of curating data

The role of the data curator

The relationships of curation and governance

The kinds of metadata that are needed for data

curation

The positioning of curator metadata into a

comprehensive metadata model.

Dave Wells is actively involved in information

management, business management, and the

intersection of the two. As a consultant he provides

strategic guidance and mentoring for Business

Intelligence, Performance Management, and

Business Analytics programs - the areas where

business effectiveness, efficiency, and agility are

driven. As an educator he plans curriculum,

develops courses, and teaches for organizations

such as TDWI and eLearningCurve. On a personal

level, Dave is a continuous learner, currently

fascinated with understanding how we think, both

individually and organizationally. He studies and

practices systems thinking, critical thinking, lateral

thinking, and divergent thinking, and he now

aspires to develop deep understanding and

appreciation for the art and science of innovation.

Taxonomies and Ontologies

Bill Inmon, Forest Rim Technologies

Professionals have long recognized the need for

abstraction of data. And the world of unstructured

data is no different. But abstraction of data in the

unstructured world takes a different form -

taxonomies and ontologies.

This session is an introduction to taxonomies and

ontologies and how they apply to the world of

unstructured data.

Data Modeling Zone 2016

Bill Inmon – the “father of data warehouse” – has

written 53 books published in nine languages. Bill’s

latest adventure is the building of technology

known as textual disambiguation – technology that

reads raw text in a narrative format and allows the

text to be placed in a conventional database so that

it can be analyzed by standard analytical

technology, thereby creating unique business value

for Big Data/unstructured data. Bill was named

by ComputerWorld as one of the ten most

influential people in the history of the computer

profession. For more information about textual

disambiguation refer to www.forestrimtech.com.

Bridging disparate IT systems with

an Enterprise Data Model

Brian Shive, Microsoft

Most businesses store data in multiple places.

They may have a SQL Server store for their

General Ledger, a NoSQL store for customer

relationship management, a DB2 store for

enterprise resource management, a SharePoint file

store for contracts and an Oracle store for orders,

deliveries and billing. To present all this data in a

single integrated view typically involves large

development time and costs for extracting,

transforming and loading a business intelligence

warehouse. Then the data must be transformed

into star and snowflake schemas for reporting.

This adds significant latency to the data in the

reports. This commonly used approach is

reasonable for business intelligence questions that

look across broad sections of the business and over

broad periods of time. Stale data is not a problem

when queries look across years of history.

Unfortunately this approach is not good for

business queries that must present near real-time

data across multiple data stores.

This presentation will examine a hybrid solution to

near real-time cross data store reporting that

combines data virtualization tools and graphs.

Data virtualization tools such as Composite

Software and Informatica Data Services are

industry standard solutions to near real-time

reporting. The new design element presented is the

use of graphs to stitch together multiple data

stores into a single web of linked data keys. This

new element allows for the data virtualization

tools to access data in parallel, allowing for near

real-time high performance across an almost

unlimited set of data stores.

This new design element is not easy to design and

implement. Graphs can represent unlimited

complexity and can be difficult to design,

implement and test. In this presentation we will

examine how to design a key-only graph that

unites separate data stores into a single web of

related data keys. We will look at the various

flavors of graphs such as: directed, non-directed,

cyclic and acyclic graphs and the challenges

involved in each flavor combination. We will

examine the use of graph relationships where each

relationship is “typed” allowing for dynamic

extension of the graph to new business uses with

minimum changes to the graph code. Further the

use of “typed” relationships allows for more control

over graph traversal when exploding a graph.

The design approach presented will include

enterprise data models and value chain models as

the basis for graph design. This enterprise view of

data and process is necessary to stitch together

multiple systems and data stores in a way that

meets user needs for near real-time reporting. This

design approach to near real-time reporting has

been implemented at Boeing, the Federal Aviation

Agency and Microsoft.

Data Modeling Zone 2016

Page 19

Brian started his data modeling career in the late

1970s while working as a consultant to the

relational database gurus at IBM. Brian learned

from John Zachman at IBM how to use the

discipline of engineering when designing data.

Brian works at Microsoft where during his 18 years

he has served as Microsoft Corporate Data

Administrator, Enterprise Architecture Lead

Information Architect, Principal Architect,

Development Manager and most-fun-one Developer.

He spent 16 years with Boeing IT. Brian also

worked as Solar Energy Designer, Executive of Boy

Scouts of America, musician, comedian and poet

and janitor. Brian and his wife and two children

live in the Seattle area. He teaches Aramaic in his

Methodist church and can be seen on YouTube

sounding at times like Jimi Hendrix. He is working

on a book of poetry and loves teaching data

modeling, database design and data integration.

The human brain and the behavior it elicits have

provided Brian with years of study in neurology,

psychology, sociology and history. He is the author

of the novel, “Data Engineering”.

Introduction to Dimensional Data

Models

Petr Olmer, GoodData

Description of basic techniques used in

dimensional data modeling including star schemas

and grain. Learn how to model business processes

in transaction, snapshot, and accumulating fact

tables. We will discuss these aspects of dimension

design: conformed dimensions and performance

optimization.

Introductory session–no previous dimensional

modeling knowledge required.

Petr Olmer studied multi agent systems, artificial

intelligence, and declarative programming. He saw

big data for the first time while working at

Computer Centre at CERN, The European

Laboratory for Particle Physics. Today he works at

GoodData, building tools and defining architecture

and methodology for BI implementations.

The Importance of Fun in Data

Modeling and Success!

Len Silverston, Universal Data Models

A key principle in success is that when we enjoy

what we do, there is much more likelihood of

success.

This session will illustrate this principle in the

context of fun during data modeling and how this

can lead to success. The presentation will show

this in many different ways. For example, you will

learn about the ‘science of fun’ and what actually

happens where there is a lot of fun, real life data

modeling stories comparing very serious data

modeling efforts with much more fun efforts and

what the outcomes were, principles of fun, and

most importantly, we will demonstrate this

principle in this class, and have a load of fun in

the presentation!

Participants of this session will gain:

Information and tools regarding the

‘science of fun’, what happens when fun is

involved, and what we can do to increase

fun and success

Principles of fun

Stories regarding data modeling efforts

that involved fun and what happened

Data Modeling Zone 2016

A REALLY FUN experience in this interactive and

fun presentation!

Len Silverston is a best-selling author, consultant,

and a fun and top rated speaker in the field of data

modeling, data governance, as well as human

behavior in the data management industry, where

he has pioneered new approaches to effectively

tackle enterprise data management. He has helped

many of the largest organizations world-wide as

well as small organizations, to integrate their data,

systems and even their people. He is well known for

his work on “Universal Data Models”, that are

described in his The Data Model Resource Book

series (Volumes 1, 2, and 3), (Volume 1 was rated

#12 on the Computer Literacy Best Seller List) and

these books have been translated into multiple

languages. He is the winner of the DAMA (Data

Administration Management Association)

International Professional Achievement Award for

2004 and the DAMA International Community

Award for 2006. He has received the highest

speaker ratings at many international conferences

and is dedicated to being of the greatest service to

his audiences.

Better Data Science conversations

through Conceptual Data

Modeling

Asoka Diggs, Intel Corp.

As a data management professional, I’ve long been

impressed at how quickly and efficiently we can

articulate the scope of the data for a project using a

conceptual data model. The modeling technique

and the graphical representations used to

articulate agreed understanding of what the data

of interest is for the problem at hand, and the

relationships among that data, enable teams of

people with a range of skills and points of view to

contribute effectively and efficiently to what the

data of interest are, what things mean, and

thereby articulate the scope for a business

problem.

As a developing data scientist, I’ve been frustrated

with how chaotic discussions about business

problems can be. What specifically is the problem

we think needs solving? What are the data of

interest? What do we know, what do we think is

true, and what are we investigating? What are the

business problems and how are we framing them

as data mining problems? I know from experience

that without a well framed problem, analysis is

unable to proceed. I also believe that framing a

business problem as a data mining problem is as

much as 50% of the work involved in advanced

analytics.

Can we apply the skill of ER modeling to the data

science process, and not only enhance our

understanding of the data of interest, but even

help with framing business problems as data

mining problems? I believe that we can.

In this session we will discuss an extension to

conceptual data modeling that can be used

immediately by ER modelers to contribute early in

the advanced analytics lifecycle, to more quickly

articulate the business problem being tackled, the

data of interest that apply to the business problem,

and graphically represent the specific focus of the

analytics effort. This may also be a good reason for

the data scientists to learn a little bit about ER

modeling.

In particular, we will cover:

The advanced analytics lifecycle and where

this technique fits

Framing a business problem as a data

mining problem

Changes in the meaning of graphical

symbols in this view of conceptual data

modeling

Data Modeling Zone 2016

Page 21

And we’ll work at least one generic

example to see the idea in practice

After 15+ years of experience in a variety of data

management disciplines, including database

administration, ER modeling, ETL development,

and data architecture, I am transitioning into the

world of analytics / data science. It turns out that

there IS something even more fun than ER

modeling, and I’ve found it. Today, my primary

interest is in the organizational transformation

involved in adopting analytics as a source of

competitive advantage. How does an organization

get there? What needs to be done? What

organization design and leadership changes are

needed to get the most benefit from analytics? I now

spend my work life practicing these new analytic

modeling skills, and teaching others how to

participate and contribute to predictive analytic

projects.

Caesars Enterprise Data Model

Patti Lee, Caesars Entertainment, and

Michael Blaha, Modelsoft Consulting

In 2014 Caesars constructed an enterprise data

model for the entire company. This was a major

undertaking with one person year of effort. The

model was built with the support of top IT and

business management. The authors will present an

overview of the EDM and dive deep into several

areas. We will discuss objectives, EDM inputs,

EDM deliverables, the construction process, and

governance. This session will appeal to attendees

who are interested in advanced data models and

their commercial use. It will provide guidance for

those constructing their own EDMs. Attendees will

obtain insight into the industry of hospitality and

gaming.

Patricia Lee is the Director of Data Architecture for

Caesars Entertainment and an evangelist for data

governance. Patti has a unique breadth of business

knowledge obtained from years of designing and

integrating a broad spectrum of systems, including

the award winning Total Rewards loyalty program.

Michael Blaha is a consultant, author, and trainer

who specializes in conceiving, architecting,

modeling, designing, and tuning databases. He has

worked with dozens of organizations around the

world. Blaha has authored seven U.S. patents,

seven books, many articles, and two video courses.

He received his doctorate from Washington

University in St. Louis and is an alumnus of GE

Global Research in Schenectady, New York. You

can find out more about him at superdataguy.com.

Scales, Measurement and Encoding

Scheme

Joe Celko

Most DBAs are fair on normalization and good on

the mechanics of their SQL product. But they do

not know anything about how to actually design

Data Modeling Zone 2016

the data going into that database and its encoding!

Would you really like to use Roman Numerals

today? On a scale from 1 to 10, what color is your

favorite letter of the alphabet? Did you know that

measurement theory was invented in 1947?

First we will talk about measurement theory, then

classify types of scales. But how do we encode such

things for a database? How do we add verification

(check digits and regular expressions) and

validation?

Finally we will discuss examples of actual

encoding for the same problem. For example,

compare the Canadian, US, and UK postal codes.

There is a set of ISO Standards that every DBA

should know.

Joe Celko served 10 years on ANSI/ISO SQL

Standards Committee and contributed to the SQL-

89 and SQL-92 Standards. He is author of eight

books on SQL for Morgan-Kaufmann: SQL for

Smarties (1995, 1999, 2005, 2010), SQL Puzzles &

Answers (1997, 2006), Data & Databases (1999)

and Trees & Hierarchies in SQL (2004), SQL

Programming Style (2005) and Analytics & OLAP

in SQL (2005) and Thinking in Sets (2008). He has

written over 1200 columns in the computer trade

and academic press, mostly dealing with data and

databases.

Data Management Maturity – Why We

Need It and How It Can Propel You to

DM Leadership

Melanie Mecca, CMMI Institute

Our industry is continually building capabilities

based on its considerable accomplishments over

the past decades. Some of the (roughly) sequential

milestone markers that most organization share

include: data design, data administration, data

architecture / warehousing, data quality and

governance, MDM, and predictive analytics using

both structured and unstructured data.

So why haven’t organizations attained DM

perfection? As we know, the data layer in the vast

majority of organizations grew project by project,

typically to meet specific needs of a line of

business. Best practices were not usually shared,

useful work products languished in project

repositories, etc. – and above all, there was no

universal mandate to manage data as a critical

corporate asset.

The Data Management Maturity (DMM)SM Model’s

primary goals are to accelerate organization-wide

DM programs by: providing a sound reference

model to quickly evaluate capabilities, strengths

and gaps; accelerating business engagement;

launching a collaborative vision / strategy; and

identifying key initiatives to extend existing

capabilities while building new ones – leading to

efficiency, cost savings, creativity, and improved

data quality.

In this seminar, we’ll address:

Data Management Capabilities and

Maturity Evaluation

The DMM in action – interactive exercises

Case study examples – how organizations

accelerate their progress

How to leverage Data Management

Maturity to empower your career.

Ms. Mecca, CMMI Institute’s Director of Data

Management Products and Services, was the

managing author of the Data Management

Maturity (DMM) SM Model. Her team created a

business-centric method for assessing an

organization’s capabilities via the DMM, and she

leads Assessments for organizations in multiple

industries.

Data Modeling Zone 2016

Page 23

She directed development of three successive

courses leading to the Enterprise Data

Management Expert certification, web-based

learning, and DM topic seminars. In 30+ years

solving enterprise data challenges, Ms. Mecca

advocates measuring data management

capabilities as the quickest path to empowering

data management and to achieve business value,

and is devoted to advancing our industry through

data management education.

Modern Data Governance:

Implications of Agile, Big Data, and

Cloud

Dave Wells, Infocentric

Traditional data governance practices need to

adapt to the realities of today’s data management

practices. It is as simple as ABC — Agile, Big data,

and Cloud. Each brings new challenges for

governance. Do agile and governance conflict or

can they coexist? The answer, of course, is that

they must coexist. The challenge is in adapting

both agile and governance processes to be

compatible. Big Data certainly brings new

governance challenges: governing data from

external and open sources, questions of data

quality and traceability, data ethics, and more.

Cloud implementations obviously amplify

governance concerns for data privacy and security.

Beyond the obvious, cloud also brings new

governance questions related to services

provisioning, service level agreements, quality

management, integration, availability, and much

more.

Dave Wells is actively involved in information

management, business management, and the

intersection of the two. As a consultant he provides

strategic guidance and mentoring for Business

Intelligence, Performance Management, and

Business Analytics programs - the areas where

business effectiveness, efficiency, and agility are

driven. As an educator he plans curriculum,

develops courses, and teaches for organizations

such as TDWI and eLearningCurve. On a personal

level, Dave is a continuous learner, currently

fascinated with understanding how we think, both

individually and organizationally. He studies and

practices systems thinking, critical thinking, lateral

thinking, and divergent thinking, and he now

aspires to develop deep understanding and

appreciation for the art and science of innovation.

Hands-On PowerDesigner

Jeff Giles, Sandhill

After a quick overview to PowerDesigner, we will

exploit the more powerful features in the tool,

including model generation. In this half-day

workshop you will learn how to maximize your

investment in PowerDesigner by leveraging this

often-misunderstood capability. Build information

architectures, enterprise data models, and many

more data connections.

What you will learn:

What is Link n’ Synch

Top Down Generations

Bottom Up Generations

Establishing the correct level of model

abstraction

Understand the ERD to UML connection

Linking Process to Data

Data Modeling Zone 2016

Lineage and Impact Analysis

Jeff Giles is Principal Architect at Sandhill

Consultants. Prior to that he has worked for Sybase

Inc and SAP. He has over 15 years experience in

Information Technology and is recognized as a

Certified PowerDesigner Professional. Jeffrey has

been a guest lecturer on modeling Enterprise

Architecture at the Boston University School of

Management. He has been involved in modeling

various perspectives of process, data, systems, and

technology.

Advanced Data Modeling Challenges

Workshop

Steve Hoberman, Steve Hoberman &

Associates, LLC

After you are comfortable with data modeling

terminology and have built a number of data

models, often the way to continuously sharpen

your skills is to take on more challenging

assignments. Join us for a half day of tackling real

world data modeling scenarios. We will complete at

least ten challenges covering these four areas:

NoSQL data modeling

Agile and data modeling

Abstraction

Advanced relational and dimensional

modeling

Join us as in groups as we solve and discuss a set

of model scenarios.

Steve Hoberman has trained more than 10,000

people in data modeling since 1992. Steve is known

for his entertaining and interactive teaching style

(watch out for flying candy!), and organizations

around the globe have brought Steve in to teach his

Data Modeling Master Class, which is

recognized as the most comprehensive data

modeling course in the industry. Steve is the author

of nine books on data modeling, including the

bestseller Data Modeling Made Simple. One of

Steve’s frequent data modeling consulting

assignments is to review data models using his

Data Model Scorecard® technique. He is the founder

of the Design Challenges group and recipient of the

2012 Data Administration Management

Association (DAMA) International Professional

Achievement Award

Even non-relational databases have

relationships

Pascal Desmarets, Hackolade

The JSON-based dynamic-schema nature (aka

schemaless nature) of NoSQL document data

stores is a fantastic opportunity for application

developers: flexibility, fast and easy evolution,

ability to start storing and accessing data with

minimal effort and setup.

With increased scale and complexity of the data, is

it still good enough to find data structures

described tacitly in the application code? With a

bigger team -- including analysts, architects,

designers, developers, and DBAs – don’t the

different stakeholders want to engage in a fruitful

dialog about the evolution of the application and

the data?

A database model (or map) helps evaluate design

options beforehand, think through the implications

of different alternatives, and recognize potential

hurdles before committing sizable amounts of

development effort. A database model helps plan

ahead, in order to minimize later rework. In the

end, the modeling process accelerates development,

increases quality of the application, and reduces

execution risks.

Data Modeling Zone 2016

Page 25

But the traditional methods and tools of the old

‘relational’ world are no longer valid for NoSQL. As

a matter of fact, this argument has sometimes

been used as a justification against the adoption

and implementation of NoSQL solutions.

In this session, after a review of the different

modeling alternatives, you will learn how to

reverse-engineer an existing instance so you can

enrich the documentation and constraints. You

will also learn how to develop a model from scratch

to produce a Mongoose schema or MongoDB 3.2

validator script.

Pascal Desmarets has been designing applications

and databases for a couple of decades, but has

embraced MongoDB and its dynamic schema

approach. In the course of designing a social

network built on MongoDB, he realized that he was

missing some of the tools previously at his disposal

in the old relational world. After a thorough

inventory of the different solutions available on the

market, he embarked on the design and

development of a new application targeted at

analysts, solution designers, architects, developers,

and database administrators. Users can visually

design, model, define, and create documentation for

Mongo databases. More information is available at

http://hackolade.com.

Conceptual vs. Logical vs. Physical

Data Modeling - a Contrarian View

Gordon Everest, University of Minnesota

Here we challenge the traditional view of these

being levels of data models. First, by looking at the

dictionary meaning of these three terms. Then

presenting an alternative view. We conclude that

a conceptual model, as commonly viewed, is not a

different type of model, but rather an abstracted

view of a more detailed underlying model. Thus, it

is a matter of presentation not of modeling. We

also conclude that all data models are logical

models since they are developed according to some

logical data modeling scheme. There is also

confusion about what elements of a data model are

moving into physical storage and

implementation. If these three terms do not help

us distinguish data models, what do we call a data

model which includes all of the semantic detail of

the user domain being modeled, but without

including any constructs or elements which are

really in the physical implementation realm. Is a

relational data model logical or physical?

Dr. Everest is Professor Emeritus of MIS and

Database in the Carlson School of Management at

the University of Minnesota. With early

“retirement”, he continues to teach as an adjunct.

His Ph.D. dissertation at the Univ of Pennsylvania

Wharton School entitled “Managing Corporate

Data Resources” became the text from McGraw-

Hill, “Database Management: Objectives, System

Functions, and Administration” in 1986 and

remained in print until 2002!

Gordon has been teaching all about databases,

data modeling, database management systems,

database administration, and data warehousing

since he joined the University in 1970. Students

learn the theory of databases, gain practical

experience with real data modeling projects, and

with hands-on use of data modeling tools and

DBMSs. Besides teaching about databases, he has

Data Modeling Zone 2016

helped many organizations and government

agencies design their databases. His approach

transfers expertise to professional data architects

within those organizations by having them

participate in and observe the conduct of database

design project meetings with the subject matter

experts. He is a frequent speaker at professional

organizations such as DAMA.

Machine Learning Primer

Daniel D. Gutierrez, AMULET Analytics

Machine learning can be thought of as a set of tools

and methods that attempt to infer patterns and

extract insight from enterprise data assets. The

subject of machine learning is one that has

matured considerably over the past several years.

Machine learning has grown to be the facilitator of

the field of data science, which is, in turn, the

facilitator of big data. In this session, I will provide

a high-level overview of the field by examining the

two primary types of statistical learning:

supervised learning and unsupervised learning.

Supervised learning is the most common type,

often associated with predictive analytics. We’ll

discuss two classes of supervised algorithms to

make predictions: regression and classification.

Next, we’ll discuss the most common type of

unsupervised algorithm: clustering to discover

previously unknown patterns within the data.

Daniel D. Gutierrez is a practicing data scientist

through his Santa Monica, Calif. consulting firm

AMULET Analytics. Daniel also serves as

Managing Editor for insideBIGDATA.com where

he keeps a pulse on this dynamic industry. He is

also an educator and teaches classes in data

science, machine learning and R for universities

and large enterprises. Daniel holds a BS degree in

mathematics and computer science from UCLA.

Essential Data Modeling

David C. Hay, Essential Strategies, Inc.

The history of the Information Technology

Industry has been marked by the effort to

apply new technologies to the world they are

to serve. The technologies have introduced

new structures to the way problems are

solved, but it is not always easy to connect

those structures to that world.

Among the problems is that of making sure

the “real-world” problem is properly

understood in the first place—in terms that

could be addressed.

The attempt to understand the “essence” of a

domain being addressed in fact pre-dates the

IT industry—by about 2500 years. (Ontology

was the branch of Greek philosophy concerned

with identifying what exists.) This

presentation will show how the effort has been

undertaken through the last five decades.

This includes a timeline of the essential

aspects of each technology (relational data

bases, essential data flow diagrams, the

semantic web, etc.), plus an example of an

“essential” enterprise data model.

A veteran of the Information Industry since the

days of punched cards, paper tape, and teletype

Data Modeling Zone 2016

Page 27

machines, Dave Hay has been producing data

models to support strategic information planning

and requirements planning for over thirty years.

He has worked in a variety of industries, including,

among others, power generation, clinical

pharmaceutical research, oil refining, banking, and

broadcast. He is President of Essential Strategies

International, a consulting firm dedicated to

helping clients define corporate information

architecture, identify requirements, and plan

strategies for the implementation of new systems.

Dave is the author of the book, Data Model

Patterns: Conventions of Thought, published by

Dorset House, Requirements Analysis: From

Business Views to Architecture, from Prentice Hall,

and Data Model Patterns: A Metadata Map, from

Morgan Kaufmann.

Most recently, he wrote Enterprise Model Patterns:

Describing the World, and UML and Data

Modeling: A Reconciliation, both from Technics

Publications.

He has spoken numerous times to annual DAMA

International Conferences (both in the United

States and Europe), annual conferences for various

Oracle user groups conferences, and numerous local

chapters of both data administration and database

management system groups.

He may be reached at [email protected]

or (713) 464-8316. Many of his works can be found

at http://articles.essentialstrategies.com.

How to compose a fact-based model

into any kind of schema

Clifford Heath, Infinuendo

Fact-based modeling is not just a powerful

business communication tool, it is the one true ring

of power for the data modeler. By working with a

fully decomposed schema in “elementary form”, we

avoid making a commitment to any particular kind

of composite schema. This presentation will show

how different composition rules applied to a fact-

based schema will generate an entity-relational,

star, snowflake, hierarchical, object-oriented, XML

or graph schema.

Clifford Heath is a computer research scientist who

has long experience in the design and

implementation of enterprise-scale software

products, and in the use of fact-based modeling. He

is a Certified Data Management Professional

(CDMP) at Masters level, has published a number

of papers in peer-reviewed scientific journals, holds

several patents, is the creator of the Constellation

Query Language and is a participant in the Fact

Based Modeling Working Group. Clifford has

frequently presented at chapter meetings of the

Data Management Association as well as at the

NATO CAX Forum and the European Space

Agency, and is CTO and founder at Infinuendo.

Business Information Modeling using

the fact-based approach

Clifford Heath, Infinuendo

Most modeling languages are effective for

communicating only with people who have been

trained to read and interpret the details. However,

fact-based languages are deeply rooted in natural

speech. They build a complete vocabulary for the

business – not just a glossary of terms – and can be

used directly with untrained business personnel.

This radically improves the depth and breadth of

the business dialogue.

As each new fact is encountered, the modeler adds

appropriate elements to an Object Role Model

diagram, which makes it easy to verbalize to

natural language. Alternatively, the new fact can

be expressed directly in the structured natural

language form of the Constellation Query

Data Modeling Zone 2016

Language, which is immediately readable to

business users but can also be understood by the

computer. This makes it possible to describe any

factual situation precisely, to ask or answer any

factual question, and to elaborate detailed business

rules about possible or allowable situations.

This presentation will introduce you to the

approach and to these two languages, peeking at

the respective software tools which generate

physical schemas. A group exercise will challenge

you to apply it to model a sample problem. The

attendee will learn to use verbalization and fact-

based analysis to understand any subject material,

and to communicate it at a deep conceptual level

with those less expert in the field.

Clifford Heath is a computer research scientist who

has long experience in the design and

implementation of enterprise-scale software

products, and in the use of fact-based modeling. He

is a Certified Data Management Professional

(CDMP) at Masters level, has published a number

of papers in peer-reviewed scientific journals, holds

several patents, is the creator of the Constellation

Query Language and is a participant in the Fact

Based Modeling Working Group. Clifford has

frequently presented at chapter meetings of the

Data Management Association as well as at the

NATO CAX Forum and the European Space

Agency, and is CTO and founder at Infinuendo.

Agile for Data Professionals

Larry Burns, PACCAR

In recent years, there’s been intense debate about

how (or whether) the principles of Agile

development can/should be applied to data

management work (including data modeling and

database development). Now the Agile debate has

shifted to BI development, raising questions of

whether an incremental approach can be applied to

enterprise-wide data work.

Larry Burns, author of Building the Agile

Database, has been in the vanguard of Agile Data

for over a decade. In his current role as Data and

BI Architect for a global Fortune 500 company, he

is also applying Agile principles to the

development of his company’s BI architecture. In

this workshop, Larry will be providing answers to

the questions that all Data and BI professionals

have about Agile, including:

What is Agile (hint: you might think you know,

but you’re probably wrong), and why is it

important?

What are the implications of Agile to IT

projects, including Data and BI development?

What is the relationship of Agile to Enterprise

Architecture (e.g., the Zachman Framework)?

What is the impact of Agile on architecture and

design? On data quality?

How can Agile principles be applied to Data

and BI initiatives?

How can data modeling and database design be

done in an Agile manner?

How can patterns be used in an Agile approach

to Data and BI development?

How can data virtualization help achieve data

agility?

What is the importance of an “Agile Attitude”?

Larry Burns is the domain architect for Data and

BI at a global Fortune 500 company, where he is

involved in enterprise-level data modeling and BI

development. He also teaches data management

and database development to the company’s

application developers. He was a contributor to

DAMA International’s Data Management Body of

Knowledge (DAMA-DMBOK), and a featured

columnist for TDAN.com. He was formerly an

Data Modeling Zone 2016

Page 29

instructor and advisor in the certificate program

for Data Resource Management at the University of

Washington in Seattle. He is the author of Building

the Agile Database, published by Technics

Publications.

Introduction to Linked Data and the

Semantic Web

Cody Burleson, Base22

In his 2009 TED talk, Tim Berners-Lee (a.k.a. The

Father of the Web) urged us onward toward a

compelling vision for the next Web - a "Semantic

Web".

"It's called Linked Data," he said. "I want you to

make it. I want you to demand it."

But what is Linked Data? Why is it valuable? How

does it work? And how can you use it to advance

your organization with the work you do?

Briefly described, Linked Data, is a method of

modeling and publishing structured data so that it

can be interlinked and become more useful

through semantic queries. Linked Data proposes to

turn the existing Web of hyperlinked documents,

which are mostly made for people to read, into a

Web of linked data that better enables computers

and people to work in cooperation. It simplifies the

publishing and integration of data and it can solve

a host of classic I.T. problems within your

organization and on a global scale. Linked Data

can change the way you think about how to model,

create, share, and integrate data.

In this gentle, step-wise, and comprehensive

presentation, you'll learn essential concepts - the

most important things you need to know. Those

new to this topic will walk away with a fresh new

perspective and a whole new toolbox for solving

problems. Those familiar to the topic are sure to

glean new insights, best-practices, and lessons-

learned. In addition to learning about the "what,

why and how", you will learn about:

Essential W3C standards such as the

Resource Description Framework (RDF)

The SPARQL Query Language (SQL for

the Semantic Web)

OWL - The Web Ontology Language and

Ontology Modeling

“Triple-store” Graph Databases

Pre-existing (reusable) Vocabularies

Linked Open Data sources that you can

find and leverage on the Web

Free, Open Source, and Commercial Tools

Available for Use.

Cody Burleson is a founder, Enterprise Web

Architect, and Director of Product Innovations at

Base22, a consulting and systems integration firm

that builds enterprise-class web solutions for some

of the world’s most recognized brands. Their motto

is “enterprise web evolution”. Cody has been an

avid practitioner and advocate for Linked Data

and semantic computing for over sixteen years -

contributing several important works to the field.

He is a member of the World Wide Web Consortium

and an active participant in the W3C's Linked

Data Platform Working Group. When not chasing

his xenophobic dog through miles of cactus covered

Texas country, he designs web solutions that

leverage Linked Data and Semantic Web

technologies for better outcomes.

UML in Depth

Norman Daoust, Daoust Associates

An in-depth look at those UML diagram types of

most importance to data professionals: use case,

activity, class, object, state machine, timing,

sequence, communication and package. The

presentation includes best practice guidelines and

tips for each of these diagram types. They will be

illustrated with examples from a case study. We

Data Modeling Zone 2016

will briefly illustrate how to model services and

their operations for Service Oriented Architecture

(SOA).

Note: This is not an introductory session.

Attendees should be familiar at least with use case

and class diagrams.

Attendees will learn:

for each of the listed diagram types:

modeling tips, diagram layout tips, naming

guidelines

the relationships between the different

diagram types

how these diagram types can assist data

professionals in their work

Norman Daoust founded his consulting company

Daoust Associates, www.DaoustAssociates.com in

2001. His clients have included the Centers for

Disease Control and Prevention (CDC), the

Veteran’s Health Administration, the Canadian

Institute for Health Information, a Fortune 500

software company, and several start-ups. He has

been an active contributor to the healthcare

industry standard data model, the Health Level

Seven (HL7) Reference Information Model (RIM)

since its inception. He enjoys introducing data and

process modeling concepts to the business analysis

community and conducting business analysis

training courses. Norman’s book, “UML

Requirements Modeling for Business Analysts”

explains how to adapt UML for analysis purposes.

WhereScape – Data Warehouse

Automation

Douglas Barrett, WhereScape

This session will discuss what data warehouse

automation is and show how it can be applied

using WhereScape tools to automate data

warehouse design, development and operations.

Data Warehouse Automation is both an alternative

to and in support of self service BI tools.

WhereScape 3D is a data warehouse design

automation tool. 3D stands for data driven design.

WhereScape 3D be used to discover source systems

and then derive target models such as star

schemas and data vaults. Once a target design

looks good it can then be exported to WhereScape

RED.

WhereScape RED is a data warehouse

development automation tool. RED is used to

accelerate project delivery and drive agile

development of your data warehouse. Building /

maintaining the data warehouse is often the

lengthy, fraught part of BI projects. WhereScape

RED is an IDE for building and managing data

warehouses, marts, stores and vaults. Together

with 3D, RED provides data warehouse

automation from design to delivery.

This session will discuss how WhereScape tools

can be used to drive data warehouse projects using

an agile approach to get the best outcome - using

collaboration with the domain experts (eg BAs) and

iterations of a working / populated subject area.

Douglas has worked for WhereScape in the US for

the last 4 years, and globally for 12. He has worked

in data warehousing on SQLServer since SQL 7.

Douglas has given talks t SQL Saturdays, TDWI

events and Tech Eds.

WhereScape is the leading provider of automation

software for planning, building and extending

Microsoft SQL Server data marts, data warehouses

and analytic systems. More than 400 SQL Server

customers use WhereScape to accelerate and

automate development while delivering a fully

documented SQL Server solution. To learn more

about how hundreds of SQL Server customers have

developed solutions with WhereScape in weeks

instead of months, click here.

Data Modeling Zone 2016

Page 31

erwin Modeling SIG

Danny Sandwell, erwin

Come and meet the erwin team and join a

discussion on all things erwin. In April 2016 erwin

Modeling became the standalone entity, erwin

Inc.. It has been an been an exciting 6 months

establishing this new entity, defining our vision

and strategy, expanding our solution set in the

first major step of realizing that vision, all the

while continuing to deliver the market leading

data modeling solution for our customers. Hear

about:

erwin Inc Vision and Strategy

erwin Modeling latest releases

erwin Inc expansion into the Enterprise

Architecture space with acquisition of

Corso Agile EA

NEW erwin Cloud Core Bundle

Plans for the Future

Additionally we are excited to facilitate an open

conversation, answers all your questions and

provide meeting place for you to meet and connect

with your fellow data management professionals.

With more than 25 years’ experience in the IT

industry, Danny has been with erwin for over 16

years. His industry experience includes various

roles in data administration, database design,

business intelligence, metadata management and

application development. Danny’s roles with erwin

include technical presales consulting, business

development, product management and business

strategy. Each role has given him a deep insight

into a broad range of issues facing organizations as

they plan, develop and manage their data

architecture and strategic information delivery

infrastructure. Danny is responsible for the

strategy, messaging, and strategic alliances for

erwin Modeling.