Data Management Complete Study Guide
Data Management Complete Study Guide
Data Management
Examination Guide
Data Exam Series Vol. 1
Acknowledgement is made for permission to use the full exam outline from the jointly
developed DAMA International Institute for Certification of Computing Professionals
(ICCP) Data Management exam outline, copyright 2006 ICCP.
Written by Diane C. Johnson, PMP, for DAMA International & DAMA International
Foundation.
Published by DAMA International & DAMA International Foundation, Bellevue, WA,
U.S.A.
Data Management Examination Guide
Data Exam Series, Vol. 1
To order copies, please contact
DAMAi@dama.org
PO Box 5786
Bellevue, WA 98006-5786
1-425-562-2636
www.dama.org
office@iccp.org
or call 1.800.843.8227
2350 E. Devon Avenue,
Suite 115, Des Plaines,
IL 60018 USA
www.iccp.org
Page 2 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Table of Contents
Introduction......................................................................................................................... 4
How Do You Obtain a CDMP? ...................................................................................... 5
CDMP Examination Criteria........................................................................................... 5
Additional CDMP Certification Criteria......................................................................... 6
Recommended Exams Based on Candidates Work Experience.................................... 8
Preparation for Taking Exams ........................................................................................ 9
Taking CDMP Exams ..................................................................................................... 9
Professional Development / Recertification ................................................................. 10
CDMP Contact Information.......................................................................................... 10
ICCP Data Management Specialty Examination Outline................................................. 12
How To Read The Data Management Subject Outline ................................................ 12
Data Management Exam Subject Outline..................................................................... 13
1.0 Data Management Function........................................................................................ 21
Overview....................................................................................................................... 21
Topics............................................................................................................................ 21
Questions....................................................................................................................... 22
Quick Answers.............................................................................................................. 33
Detailed Answers .......................................................................................................... 34
2.0 Data & Metadata Infrastructures Creation / Maintenance .......................................... 42
Overview....................................................................................................................... 42
Topics............................................................................................................................ 42
Questions....................................................................................................................... 43
Quick Answers.............................................................................................................. 51
Detailed Answers .......................................................................................................... 52
3.0 Data Analysis and Modeling....................................................................................... 57
Overview....................................................................................................................... 57
Topics............................................................................................................................ 57
Questions....................................................................................................................... 58
Quick Answers.............................................................................................................. 76
Detailed Answers .......................................................................................................... 77
4.0 Data / Metadata Infrastructure Management .............................................................. 88
Overview....................................................................................................................... 88
Topics............................................................................................................................ 88
Questions....................................................................................................................... 89
Quick Answers.............................................................................................................. 96
Detailed Answers .......................................................................................................... 97
5.0 Information Quality Management............................................................................. 102
Overview..................................................................................................................... 102
Topics.......................................................................................................................... 102
Questions..................................................................................................................... 103
Quick Answers............................................................................................................ 114
Detailed Answers ........................................................................................................ 115
Selected Bibliography..................................................................................................... 122
Page 3 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Introduction
The Certified Data Management Professional (CDMP) credential validates knowledge
and experience of Data Management Professionals. CDMP Credentials can be a doorway
to opportunities either measuring your standing by demonstrating Mastery level or
providing a starting point for new professions through a Practitioner designation.
The Certified Data Management Professional (CDMP) credential is awarded to those
who qualify based on a combination of criteria including education, experience and testbased examination of professional level knowledge. To maintain certified status and
continued use of the credential, an annual recertification fee along with a 3-year cycle of
continuing education and professional activity is required. The Data Management
Association International (DAMA) authorizes the Certified Data Management
Professional certification program and granting of the CDMP designation in partnership
with the Institute for Certification of Computing Professionals (ICCP), which administers
testing and recertification.
The ICCP Data Management exam is meant to be an experience exam, meaning that it
tests what you know at the time. This study guide is meant to be a refresher to test taking
and the concepts behind data management. You can focus on the sections that you need
to learn, or take the practise exam to see where your strengths lie. The study guide is
broken down into the five major sections of the exam:
1.0 Data Management Function
2.0 Data & Metadata Infrastructures Creation / Maintenance
3.0 Data Analysis and Modeling
4.0.Data / Metadata Infrastructure Management
5.0 Information Quality Management
The DAMA International Foundation welcomes feedback on this Study Guide, as
revisions will occur in the future. We encourage you to let us know how you are using
these materials and how they might be improved. Your comments can be sent to:
Vice President of Education, VP_Education_Services@DAMA.org
Page 4 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Credential Earned
CDMP Practitioner Certificate
CDMP Mastery Certificate
The CDMP Practitioner certification is awarded to professionals who scored above 50%
on all three exams. These individuals can contribute as a team member on assigned tasks
for they have a working knowledge of concepts, skills and techniques in a particular data
specialization.
The CDMP Mastery certification is awarded to professionals who scored 70% or higher
on all three exams. These individuals have the ability to lead and mentor a team of
professionals as they have mastered the concepts, skills and practices of their data
specialization.
Exams may be retaken to improve your score and go from the Practitioner to the Mastery
certificate level. You may be able to substitute select vendor certifications for up to one
specialty exam.
Page 5 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
CDMP Practitioner
Certificate
2
CDMP Mastery
Certificate
4+
Yes
Yes
Yes
Yes
Page 6 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Work Experience
Sample qualifying position/role titles:
VP, Director, or Manager of Data Management
Data Architect, Data Administrator, Data Analyst, Data Modeler
Data Specialist, Database Administrator, Data Warehousing Analyst
Systems Architect, Systems Analyst, Project Manager, Project Leader
Business Analyst, Repository Analyst, Repository Architect
Professional Examinations
The CDMP requires three ICCP exams: IT Core, one specified data oriented exam, and
one other exam. If you already passed one or more ICCP exams, these exams can be
used toward a CDMP if considered current by ICCP standards, and the exams are listed
within your CDMP area of specialization. For information on your status, contact the
ICCP.
If you wish to demonstrate expertise in exam specialty areas specifically, the ICCP will
issue Expert (Proficiency) Certificates for each specialty exam passed at 70% or higher.
If you wish to know how these exams were developed, go to
(http://www.iccp.org/iccpnew/iwg2.html). These exams are product and vendor neutral,
and international in scope.
Page 7 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Mgmt
Architecture
Data
Analysis
& Design
DBA
Data
Warehousing
Metadata /
Repository
Mgmt
Data /
Information
Quality
(Future)
Reqd
Reqd
Reqd
Reqd
Reqd
Reqd
Reqd
Database
Administration
Data Warehousing
C
C
IT Core
Specialty Exams
Data Management
Integrated Project
Mgmt
IT Management
Systems
Development
Object Oriented
Analysis & Design
Systems Security
Future ICCP
Exams
Business
Intelligence &
Analytics
C (future: e.g.
MIT or Berkeley
DQ programs)
Page 8 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 9 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
A DAMA chapter can set up exam sessions during their chapter meetings. What is
needed is a volunteer proctor from the chapter. A proctor is an individual authorized by
ICCP to oversee the writing of an exam by an ICCP exam taker. This person must meet
specific guidelines (http://www.iccp.org/iccpnew/testing.html) and be willing to
supervise the exam taker. The ICCP reserves the right to reject proposed proctors.
Contact office@iccp.org or phone 847.299.4227 or 800.843.8227 if you require
assistance in determining an appropriate proctor.
The exams run off the USB drive of an individuals laptop. There are 110 questions with
110 being scored and 10 are beta questions to complete in 90 minutes. You will not know
which type of question you are answering. Questions and possible distracters (answers)
are randomly listed in a different order for each exam taker. Therefore, although this
guide contains sample questions that allow for all or none of the above type answers
meant for study purposes, you will not find this type of answer to choose from on the
actual exam.
Computer based testing allows for the immediate scoring after the exam is taken. An
ICCP Performance Profile is then available for downloading, and one will be sent later to
the individual by the ICCP. This Profile shows your exam strengths and weaknesses.
Page 11 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Outline elements with numeric level leaders imply inclusivity. Concepts not
within the numbered structure will not be tested.
Outline elements with a bullet leader are examples to clarify the content of a
numbered element, and are not necessarily inclusive.
Numbers in parentheses after an element name indicate the number of questions
in the exam, which will be presented on the subject indicated by the element name
and all subordinate elements. These allocations are guidelines established by the
Test Management Council, and are followed as closely as possible in selecting
questions for the exam. There are 100 multiple-choice questions on each exam
version and this outline reflects this total.
The characters D# after an element name indicate the target depth of
questions to be posed on the subject indicated by the element name and all
subordinate elements. The depths of knowledge are defined as follows:
D1
D2
D3
D4
D5
D6
Recognition
Knowing what a concept is called.
Differentiation Knowing the external differences between a concept and a
neighboring concept.
Description
Knowing the external characteristics of a concept.
Usage
Knowing how to use instances of the concept and why.
Structure
Knowing the internal structure of the concept its
components and the relationships among these components.
Construction
Knowing how to put together instances of the concept
tailored to specific purposes.
Page 12 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
1.1 Planning
(8)
1.1.1.
1.1.2.
D4
1.2.2.
(2)
D3
(8)
D3
Page 13 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
1.3.2.
1.3.3.
1.3.4.
1.3.5.
1.3.6
Page 14 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
2.0 Data & Metadata Infrastructures Creation / Maintenance (15) Section 2 Total
2.1 Planning for Data & Metadata
2.1.1
2.1.2.
D4
Architectures
Enterprise Data
Data Sourcing
Data Distribution
Data Integration
Change Authorization
Zachman Framework
Data Processing Architectures (i.e. client-server, distributed data, etc.)
Metadata Architectures
Data Architecture Methods
Information Engineering
Enterprise Architecture Planning
Data Life Cycle
Data Reengineering
Prototyping
2.2.2.
2.2.3.
(6)
(9)
Data
Database Management Systems (DBMS & ODBMS)
Data modeling tools
Extract, transform, and load (ETL) tools
Metadata & Descriptive Information
Data dictionaries
Data directories
Data encyclopaedias
Metadata registries (e.g. ISO/IEC 11179)
Metadata repositories
Data Issues
Business intelligence technologies (OLAP, Data Mining, etc.)
Data management and the Internet / Intranet
Data management and unstructured data
Page 15 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
D3
3.1.2.
(6)
D5
3.2.2.
3.2.3.
3.2.4.
(21)
Page 16 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
D5
3.3.2.
3.3.3.
(10)
D5
Page 17 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
D5
(5)
Page 18 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
D3
D3
5.1.1. Definition
Data and Information
Information quality
Data definition as Information Product Specification
Data definition quality
Information architecture (data model) quality
Business drivers
Costs of nonquality information
5.1.2. Information Quality Characteristics
Conformance to definition
Completeness
Validity
Accuracy
Precision
Non duplication
Consistency of redundant data
Timeliness
Usefulness
Objectivity (of presentation)
Presentation clarity
5.1.3 Data Definition (or Information Product Specification) Quality
Characteristics
Properly formed name, in accordance with approved naming convention
standard
Standard, single enterprise abbreviations for new development
Name appropriate to knowledge workers
Correct, clear, and complete definition
Business term (used in data definition) defined in glossary
Correctly specified value domain and definition (of code values)
Properly defined data value type (not just alphanumeric, etc., but domain
type (corresponding to class words, e.g., data, code, amount, ID, etc.)
Correct, complete, and useful business rule specification
5.2. Information Quality Assessment / Audit
(4)
Page 19 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
D4
(8)
Page 20 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
D3
Topics
Data Management Planning
Data Management Scope & Charter
Data Management Plans
Policies / Standards / Processes / Procedures / Guidelines
Data Management Organization
Types of Staff Training
Communication
Roles & Responsibilities
Data Administration
Metadata Administration
Database Administration
Data Warehouse Administration
Information Stewardship
Configuration Management
Page 21 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Questions
1.1. Planning
1.1.1
12. Which one of the following is not a benefit of a strong data portfolio management
program?
A. Maximize value of IT data investments while minimizing the risk
B. Improve communication and alignment between technology and business.
C. Encourage reuse of data throughout the organization.
D. Allow planners to schedule resources more efficiently.
13. Which one of the following is least desirable benefit of Enterprise data / information
framework?
A. Provides enterprise-wide definitions of concepts and data.
B. Provides a scoping tool for new initiatives.
C. Reduces data redundancy by providing transparency as to the meaning of data
items
D. Encourages re-use and consistent data structures across the enterprise
14. What party would not be considered when creating a relationship management plan?
A. Vendor / Supplier
B. Customer
C. Employee
D. President
15. A Relationship Management Plan when dealing with vendors / suppliers should be
part of which overall strategy?
A. Procurement Strategy
B. Quality Management Strategy
C. Enterprise Architecture Strategy
D. IT Strategy
16. Which area does a Quality Management Plan does not address?
A. Quality policies and procedures.
B. Roles, responsibilities and authorities.
C. Description of quality system.
D. Meta-metamodel
17. Which one of the following is not true when describing Capability Maturity Model
Integration (CMMI)?
A. Model framework to assess data and process maturity.
B. Model framework to determine priorities.
C. Model framework to institute process and data improvement.
D. Defines six levels of process maturity.
Page 24 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
18. What level is organizations CMMI maturity, if the data management requirements
are not being met?
A. Level 0
B. Level 1
C. Level 2
D. Level 3
19. What level is organizations CMMI maturity, if the data management requirements
are being met?
A. Level 0
B. Level 1
C. Level 2
D. Level 3
20. What level is organizations CMMI maturity, if the data management requirements
are being managed and tracked?
A. Level 1
B. Level 2
C. Level 3
D. Level 4
21. What level is organizations CMMI maturity, if the data management requirements
meet EIA Standard 859 Industry Standard for Data Management that includes nine high
level Data Management Principles?
A. Level 1
B. Level 2
C. Level 3
D. Level 4
22. Which one is not the purpose of the data management configuration management
plan?
A. Identify and describe the overall policies and methods for Configuration
Management.
B. Establish and provide the basis for a uniform and concise Configuration
Management practice
C. Manage the data for its entire lifecycle.
D. Retain data commensurate with value.
Page 25 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
23. Who is responsible for developing and implementing Data-Management planning for
projects, for ensuring that the activities are completed according to agreed standards and
timelines and for coordinating ongoing data management to support the business?
A. Data Manager
B. Data Analyst
C. Database Administrator
D. Business Manager
Page 26 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
27. What is the following statement: Aggregation of data values is appropriate for fields
with a large numbers of values, such as dates, age, and geographic areas; it is the primary
method used to collapse a dataset in order to create tables with no small numbers as
denominators or numerators in cells?
A. Policy
B. Standard
C. Procedure
D. Guideline
28. What is the following statement: Data custodians are responsible for creating and
maintaining metadata for their datasets?
A. Policy
B. Standard
C. Procedure
D. Guideline
29. Which of the following is the best answer for the definition of cost when following
the metadata procedure, to state what the concept is, not only what it is not.
A. Total spent for goods or services including money and time and labor.
B. Cost is a price paid.
C. Costs, which are not related to external costs.
D. Direct cost to the business owner of those items, which will be sold to customers.
1.2.
Organization
1.2.1. Types of Staff Training
30. Which one of the following is not appropriate for an orientation of the data
environment for new employees?
A. Acronym list.
B. Customer Service Policy.
C. Data Policy and Procedure.
D. WWW Design and Programming.
31. When embarking on continuing education for required skills or retraining, which
training method is least desirable?
A. Mentoring with another employee.
B. Workshops and seminars.
C. Classroom or computer based courses.
D. Booklets and information sheets.
Page 27 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
1.2.2. Communication
32. Which one of the following is the least beneficial when promoting data services?
A. Communicating data services and benefits.
B. Publishing newsletters and web site news.
C. Customer education and training.
D. Convening a meeting of the Organizations Management Team.
1.3.
1.3.1.
33. Who is responsible for identifying and analyzing information needs for the enterprise
or business area, and develops and maintains data architecture?
A. Data Administrator
B. Manager, Data Administration
C. Data Administration Consultant
D. Database Administrator
34. Which one is not a responsibility of the Data Administrator?
A. Identify and analyze customer information needs.
B. Develop and maintain data architecture.
C. Develop and maintain strategic data plan.
D. Provide approval authority over metadata policies and design.
35. When hiring a Data Administrator which skill is the least preferred?
A. Relational Database experience.
B. Logical and Physical Data Modeling.
C. Project Management experience.
D. Strong written and oral communication skills.
36. Which role would a Data Administrator not typically interact?
A. Business Analyst
B. CEO
C. Repository Administrator
D. Management
Page 28 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
37. Who is responsible for planning, organizing, directing and controlling data
definition, data use, and ensure data availability for the enterprise?
A. Data Administrator
B. Manager, Data Administration
C. Metadata Administrator
D. Database Administrator
Page 29 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 30 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
47. When hiring a Data Warehouse Administrator which skill is the least preferred?
A. Relational Database, related utilities and tools experience.
B. Ability to calculate Data Warehouse return on investment, costs and benefits.
C. Expert in data structure including parallel data structure.
D. Logical and Physical Data Modeling.
Page 31 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
1.3.6
Configuration Management
53. A new data model is created and rolled into Production. Which process is
responsible for registering the modification in the Configuration Management Database
(CMDB)?
A. Change Management
B. Configuration Management
C. Problem Management
D. Release Management
54. Which of the following is a Configuration Item (CI)?
A. Organization Structure
B. Data Model
C. An incident
D. A process
55. Which one is not a discipline of Data Management Configuration?
A. Status Accounting
B. Collection
C. Approval
D. Distribution
56. Which item is not a responsibility of the Configuration and Data Management team?
A. Management of all documentation and specifications.
B. Configuration and data management of programs.
C. Maintaining requirements of deliverables through the data change process.
D. Providing storage, retrieval, distribution, and management of program data.
Page 32 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Quick Answers
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
A
D
D
A
A
A
C
D
A
D
C
D
B
D
A
D
D
A
B
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
36.
37.
38.
B
C
D
A
A
B
C
D
A
A
D
D
D
A
D
C
B
B
C
39.
40.
41.
42.
43.
44.
45.
46.
47.
48.
49.
50.
51.
52.
53.
54.
55.
56.
D
C
D
B
D
D
D
D
B
A
B
D
A
B
B
B
A
A
Page 33 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Detailed Answers
1. Answer: A. To facilitate the development, management, and use of the data resources
as a vital asset. The primary mission of Data Management is to facilitate the
development, management, and use of the data resources as a vital asset. The services of
Data Management comprise of the following: Database maintenance and enhancement
for production application systems and Data analysis and modeling for projects in
planning or analysis.
2. Answer: D. Provide a cost effective and robust document and content management
capabilities, workflow and business process management capabilities. Objectives of Data
Management are: Maintain the physical integrity and efficiency of data resources;
education about the benefits of and methods for enhancing data quality; and provide the
architecture and guidelines for documenting and implementing data resources.
3. Answer: D. Economies of scale in purchasing. The scope of Data Management
function include: Requirements analysis and modeling; Enterprise-wide data
coordination, integration, stewardship and use; and data security and quality.
4. Answer: A. Support for warehouse initiatives. Data Management services include data
maintenance and enhancement for production application systems; Database design for
projects in development; and Data analysis and modeling for projects in planning or
analysis.
5. Answer: A. Information Technology Strategy. A Data Management Plan is usually
part of the overall Information Technology Strategy. The Information Technology
Strategy leads to Infrastructure Strategy, Information Technology Architecture Plan and
Application Infrastructure Plan.
6. Answer: A. Information Technology Strategy. A Data Management Plan is usually
part of the overall Information Technology Strategy. The Information Technology
Strategy leads to Infrastructure Strategy, Information Technology Architecture Plan and
Application Infrastructure Plan.
7. Answer: C. Implement facilities and tools for managing metadata resources. Data
Management Plans are high level and describe the roles and resources of program staff,
define future direction of data management activities in a work plan and the development
of a quality management plan.
8. Answer: D. Project Change Committee. Data Management oversight committees have
been called: Data Standards Approval Committee; Data Management Process Change
Management Committee; and Enterprise Data Management Board of Directors. Project
Change Committee refers to changes made to project scope, time, or cost.
9. Answer: A. To ensure metadata management controls have achieved intended results.
The purpose of conducting a metadata management audit is to ensure metadata
Page 34 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 36 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
a uniform and concise Configuration Management practice; and Manage the data for its
entire lifecycle.
23. Answer: A. Data Manager. A data manager is responsible for developing and
implementing Data-Management planning for projects, for ensuring that the activities are
completed according to agreed standards and timelines and for coordinating ongoing data
management to support the business, known as the data management standards
management plan.
24. Answer: A. Policy. The sentence is a Policy statement: Data archives must include
easily accessible information about the data holdings, including quality assessments,
supporting ancillary information, and guidance and aids for locating and obtaining the
data. A Policy is a prescribed or proscribed course of action or behavior, which is to be
followed with respect to the acquisition, deployment, implementation or use of
information technology resources. It is not a standard, as it does not outline a specific
technical approach. It is not a procedure, as it does not offer a set of administrative
instructions for implementation of a policy or standard. It is not guideline that should
offer a detailed plan or explanation to guide you in setting standards or determining a
course of action.
25. Answer: B. Standard. The sentence is a Standard statement: Contact Information
offers data groupings that are used to describe a point of contact, address, and
communication information. Standard(s) is a prescribed or proscribed specific technical
approach, solution, methodology, product or protocol which must be adhered to in the
design, development, implementation or upgrade of data architecture. Standards are
intended to establish uniformity in data. Standards should be designated as either
"preferred" or "mandatory". It is not a procedure, as it does not offer a set of
administrative instructions for implementation of a policy or standard. It is not a
guideline, which should offer a detailed plan or explanation to guide you in setting
standards or determining a course of action.
26. Answer: C. Procedure. The sentence is a Procedure statement: To keep the hard
drives from getting full, please back-up your data. Procedure is a set of administrative
instructions for implementation of a policy or standard. It is not a guideline, which should
offer a detailed plan or explanation to guide you in setting standards or determining a
course of action.
27. Answer: D. Guideline. The sentence is a Guideline statement: Aggregation of data
values is appropriate for fields with a large numbers of values, such as dates, age, and
geographic areas; it is the primary method used to collapse a dataset in order to create
tables with no small numbers as denominators or numerators in cells. A guideline offers
a detailed plan or explanation to guide you in setting standards or determining a course of
action.
28. Answer: A. Policy. The sentence is a Metadata Policy statement: Data custodians are
responsible for creating and maintaining metadata for their datasets A Policy is a
Page 37 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
include developing and enforcing standards for metadata through the review of
definitions. Also, assist in developing procedures and data management policies that
ensure the integrity, consistency and control of the enterprise's data resource.
35. Answer: C. Project Management experience. When hiring a Data Administrator the
skill that may be desirable but least preferred in the above list is Project Management
experience. Typically, skills for Data Administrators are: Relational Database
experience, logical and physical data modeling, strong written and oral communication
skills, strong analysis skills and prior work experience.
36. Answer: B. CEO. The Data Administrator would not typically interact with the CEO
in an organization. The Data Administrator would interact with Business Analysts,
Repository Administrator, Management and Users/Customers of the data.
37. Answer: B. Manager, Data Administration. The Manager, Data Administration is
responsible for planning, organizing, directing and controlling data definition, data use,
and ensure data availability for the enterprise. The Data Administrator identifies and
analyzes information needs for the enterprise or business area, and develops and
maintains data architecture plus the strategic data plan. The Metadata Administrator is
responsible for creating, administrating and enforcing of standards, guidelines and
procedures for the use of metadata plus metadata query and analysis tools.
38. Answer: C. Metadata Administrator. The Metadata Administrator is responsible for
creating, administrating and enforcing of standards, guidelines and procedures for the use
of metadata plus metadata query and analysis tools. The Manager, Data Administration
is responsible for planning, organizing, directing and controlling data definition, data use,
and ensure data availability for the enterprise. A Database Administrator conducts data
store modeling, design, implementation, and operation.
39. Answer: D. Provide final review and approval authority over data design for an
application system. The Metadata Administrator role would: Establish and maintain the
metadata architecture; Provide approval authority over metadata policies and design; and
Maintain repository security profiles in addition to Metadata tool administration.
40. Answer: C. Metadata Analyst. In a company with a Metadata team, the Metadata
Analyst would collect the requirements and design the metadata solution. The
Metamodeler would convert the requirements into metamodels. The Metadata
Administrator is responsible for creating, administrating and enforcing of standards,
guidelines and procedures for the use of metadata plus metadata query and analysis tools.
The Manager, Data Administration is responsible for planning, organizing, directing and
controlling data definition, data use, and ensuring data availability for the enterprise.
41. Answer: D. Perform application tuning and performance monitoring. A Database
Administrator would be responsible for: Establish and maintain sound backup and
recovery policies and procedures; Implement and maintain database security (create and
maintain users and roles, assign privileges); Perform database tuning and performance
Page 39 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
monitoring; Capacity planning; Perform general technical trouble shooting and give
consultation to development teams.
42. Answer: B. CEO. The Database Administrator would not typically interact with the
CEO in an organization. The Database Administrator would interact with Business
Analysts, Data Administrator, Management and Users/Customers of the data.
43. Answer: D. Network security administration. When hiring a Database Administrator
the skill that may be desirable but least preferred in the above list is Network security
administration experience. Typically, skills for Database Administrators are: Relational
Database; related utilities and tools experience, physical data modeling; ability to perform
both Relational Database and Operating System performance tuning and monitoring; and
prior work experience.
44. Answer: D. Database Administrator. In an organization, the Database Administrator
would develop the referential integrity constraint scripts. The Data Analyst would work
with the Database Administrator to link the logical to physical data model.
45. Answer: D. Database Administrator. In an organization, the Database Administrator
would have the responsibility to recover the physical database in the event of a power
disruption.
46. Answer: D. Installing the Operating System on the Data Warehouse server. A Data
Warehouse Administrator would be responsible for: Data Warehouse data modeling and
design; Data Warehouse implementation and refresh; Data Access administration;
Perform application performance monitoring; Perform general technical trouble shooting
and give consultation to development and metadata teams.
47. Answer: B. Ability to calculate Data Warehouse return on investment, costs and
benefits. When hiring a Data Warehouse Administrator the skill that may be desirable
but least preferred in the above list is the ability to calculate Data Warehouse return on
investment, costs and benefits. Typically, skills for Data Warehouse Administrators are:
Relational Database; related utilities and tools experience, logical and physical data
modeling; Expert in data structure including parallel data structure; Extract Transform
and Load tool experience; and prior work experience.
48. Answer: A. Business Information Steward. Business Information Steward would
review and approve data definitions and domain value specifications for business data.
Other responsibilities would include: validating business rules and keeping the domain
values current across the Enterprise. Managerial Information Steward is responsible for
setting information policy and creating information measures for either the organization
or a specific department or business area. Physical Data Trustee is accountable for the
integrity of the physical data assets. An Information Producer is accountable for the
content of the information.
Page 40 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 41 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Topics
Data Architecture Methods
Architectures
Data Architecture Methods
Tools and Technology Types
Data
.
Metadata & Descriptive Information
.
Data Issues
Page 42 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Questions
2.1
2.1.1
Architectures
Page 43 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
6. Which one of the following requires combining and matching information from
different sources, and resolving a variety of conflicts?
A. Data Sourcing
B. Data Distribution
C. Data Integration
D. Enterprise Data
7. What is a fundamental principle in Change Authorization of Architectures?
A. Single point of authorization.
B. Single point of access.
C. Private key encryption on data.
D. Standard for communication.
8. Which of the following is not a feature of the client in client-server architecture?
A. Passive
B. Active
C. Sending requests
D. Waits until reply arrives
9. Which of the following is not a feature of the server in client-server architecture?
A. Passive
B. Active
C. Waiting for requests
D. On requests serves them and send a reply
10. What is the best name for a network called if the networks consists clients,
application servers which process data for the clients, and database servers, which store
data for the application servers?
A. 2-tier Architecture
B. 3-tier Architecture
C. n-tier Architecture
D. Multi-tier Architecture
11. Which one of the following is not a source of metadata?
A. Case Tools
B. Applications
C. Physical Database
D. Company Directory
Page 44 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
12. Which type of analysis is needed when beginning Metadata solution architecture?
A. Metadata record
B. Metadata flows
C. Metadata Categorization
D. Metadata user
13. What is best definition of the Zachman Framework?
A. A 36-cell Matrix.
B. A Normalized schema.
C. A good analytical tool.
D. Specific to methods/tools.
Page 45 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 46 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
2.2.
2.2.1. Data
23. What is the definition of a Database Management System (DBMS)?
A. Controls the organization, storage and retrieval of data in a database.
B. A modeling language to define the schema.
C. Inverted list management.
D. Supports the database query language to interactively access data.
24. Which one is not a common DBMS model?
A. Hierarchical
B. Network
C. Relational
D. File
25. Which one is not a function of a DBMS?
A. A modeling language to define the schema
B. A database query language
C. Transaction method that ensures Atomicity, Consistency, Isolation, and Durability
(ACID)
D. RAID Disk arrays.
26. What is the definition of an Object Database Management System (ODBMS)?
A. Controls the organization, storage and retrieval of data in a database.
B. A modeling language to define the schema
C. Inverted list management
D. Supports the database query language to interactively access data.
27. Which one is not a function of an ODBMS?
A. Object Definition Language (ODL)
B. Object Query Language (OQL)
C. C++ and Java Binding.
D. Structured Query Language (SQL).
28. Which function does Extract, Transform, and Load (ETL) tool does not involve in
the process in data warehousing?
A. Extracting data from data sources.
B. Transforming data to fit business requirements.
C. Transforming metadata to fit business requirements.
D. Loading data into the data warehouse.
Page 47 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
29. The first part of an ETL process is to extract the data from what type of common data
source formats?
A. Relational database
B. Flat File
C. IMS
D. C++
30. Which one is not a typical function of the transformation process in ETL tools?
A. Translating code values
B. Deriving new calculated values
C. Joining or merging data from multiple sources
D. DDL SQL statements with SQL variations
31. Which of the following is not a type of load function of an ETL tool in the data
warehouse?
A. Overwrite old information
B. Insert new records
C. Update old record and Insert new record
D. Insert audit trail records.
32. Which of the following is not true of a Data Modeling Tool?
A. Specific to a DBMS
B. Produce a diagram summarizing the results of your data modeling efforts
C. Generate a database schema from a model.
D. Diagram of referential integrity constraints.
Page 48 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 49 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
39. What is an integrated, virtual holding area with vendor-independent input, access,
and structure; used to directly store metadata and/or metadata-based gateways to external
metadata?
A. Data dictionaries
B. Data directories
C. Data encyclopaedias
D. Metadata repositories
Page 50 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Quick Answers
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
C
D
B
D
B
C
A
A
B
B
D
D
B
D
C
A
D
B
A
D
A
D
23.
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
36.
37.
38.
39.
40.
41.
42.
43.
A
D
D
A
D
C
D
D
D
A
B
C
D
D
B
C
D
A
A
D
C
Page 51 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Detailed Answers
1. Answer: C. Economies of scale in purchasing Case tools. The benefits of an Enterprise
Data Architecture are that it organizes data around the enterprises data subject to create
shared data resources, an integrated view of enterprise data that enables organizational
change.
2. Answer: D. Design stability and data object abstraction and generalization. An
Enterprise Data Architecture should be created for design stability and data object
abstraction and generalization. Enterprise Data Architecture treats data like an
information asset; it is not application specific but enterprise specific. Enterprise Data
Architecture is typically created in more than one iteration.
3. Answer: B. Determine the index for the data mart. The source data should be
architected to determine the source of data needed, diagram the source data, and
determine the method for extraction and delivery.
4. Answer: D. Determine the monthly flat file transmission protocol. The goal of Source
Data Architecture is to ensure that the source data is extracted only once, define the scope
and implementation of the data warehouse and oversee the construction of the enterprise
data warehouse.
5. Answer: B. Data Distribution. Data Distribution is targeted towards the efficient
delivery of the proper information to the proper recipients. In Data Distribution data can
be streamed or supplied depending on the requirements of the communication.
6. Answer: C. Data Integration. Data Integration requires combining and matching
information in different sources, and resolving a variety of conflicts. XML is becoming a
de facto data integration standard.
7. Answer: A. Single point of authorization. A fundamental principle in Change
Authorization of Architectures is a single point of change authorization. Every change
must run the same process and authorization prior to changes are implemented.
8. Answer: A. Passive. In a client-server architecture, the features of the client are:
Active (Master), sending request, and waiting until reply arrives.
9. Answer: B. Active. In a client-server architecture, the features of the client are:
Passive (Slave), waiting for requests, and on requests serves them and send a reply.
10. Answer: B. 3-tier Architecture. If the networks consists of clients, application servers
which process data for the clients, and database servers which store data for the
application servers it is known as a 3-tier Architecture.
11. Answer: D: Company Directory. Metadata has many sources including but not
limited to: Tools, Applications and Software Packages.
Page 52 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
12. Answer: D. Metadata user. When beginning Metadata solution architecture, the
following type of analysis should be carried out: Process Flow, Metadata Flow, Metadata
Record Identification, and Metadata Categorization. Metadata occurs at the input and
output of processes the tools like case tools and applications gather the metadata. The
Metadata Record Identification identifies the data needed in a metadata solution and their
origins. The last step is to categorize the metadata for display and usage.
13. Answer: B. A Normalized schema. The Zachman Framework is a normalized schema
that is a good analytical tool. It is not a 36-cell matrix. The Zachman Framework logic is
neutral to methods and tools. Each cell of the Framework is unique and primitive.
14. Answer: D. Greater accounting staff effectiveness. Enterprise Architecture Planning
has the following benefits: consistency and compatibility of systems, interoperability
between systems and databases, economies of scale in purchasing and developing
systems, reduced overall system costs, and greater IT staff effectiveness.
15. Answer: C. Maintain a secure infrastructure and IT support for networks and
distributed systems. Enterprise Architecture Planning addresses: data management,
application environment and development toolsets, middle-ware and transaction
management, Web delivery environment, operating systems and other system software,
network environment, and hardware server and client environments.
16. Answer: A. Information Technology Strategy. An Enterprise Architecture Plan is
usually part of the overall Information Technology Strategy. The Information
Technology Strategy leads to Infrastructure Strategy, Information Technology
Architecture Plan and Application Infrastructure Plan.
17. Answer: D. Shred. The data life cycle phases are: Create/Store, Retrieve,
Modify/Update, Read/Use, Transport, Archive and Delete. The data lifecycle is the
process of managing data throughout its lifecycle from conception until disposal, within
the constraints of the data policy.
18. Answer: B. Data Life cycle management is an approach to managing an
organization's data. The definition of data life cycle management is an approach to
managing an organization's data that involves procedures and practices as well as
applications.
19. Answer: A. Information Engineering. Information Engineering is defined as "an
integrated and evolutionary set of tasks and techniques that enhance business
communication throughout an enterprise enabling it to develop people, procedures and
systems to achieve its vision".
20. Answer: D. Data Warehousing. Information Engineering has many purposes,
including organization planning, business re-engineering, application development,
information systems planning and systems re-engineering.
Page 53 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 54 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
multiple sources, generating surrogate key values and summarizing values. The DDL
SQL statement with SQL Variations is a DB mechanism to create tables for example.
31. Answer: D. Insert audit trail records. Typical load functions of an ETL tool in a data
warehouse are to: Overwrite old information, insert new records and update old record
and insert a new record.
32. Answer: A. Specific to a DBMS. Data modeling tools are RDBMS-independent.
Data Modeling Tools: Produce a diagram summarizing the results of your data modeling
efforts; Generate a database schema from a model; and Diagram of referential integrity
constraints.
33. Answer: B. Database that tracks data element definitions. A data dictionary is a
database that tracks data element definitions. An encyclopedia is an internal database
that store information tracked, developed, and maintained by a predefined set of singlevendor tools. A data staging area is a set of cleansed, organized, and transaction level
data. Instances of characters could represent any time of file or data store and is not
specific to a data dictionary.
34. Answer: C. A repository or database of information. A data dictionary is a repository
or database of information. Data dictionaries can be used as a white page application and
network information service or as an LDAP directory service.
35. Answer: D. Internal database that store information tracked, developed, and
maintained by a predefined set of single-vendor tools. A data encyclopedia is an Internal
database that store information tracked, developed, and maintained by a predefined set of
single-vendor tools. A data dictionary is a database that tracks data element definitions.
A data staging area is a set of cleansed, organized, and transaction level data.
36. Answer: D. XML. XML is a format. A Data Registry is defined as an automated
resource used to describe, document, protect, control and access informational
representations of an enterprise. Typically the following is held in a data registry:
Standardized information in a pre-defined model, Metadata, system metadata, system
engineering; and Reference information. Standards for models and templates for data
and metadata registries already exist for example, the ISO 11179 standard for Metadata
Registries, and ebXML for XML registries.
37. Answer: B. UDDI registry with a proprietary repository. A federated SOA
deployment requires a standards-based registry-repository the choices involve two
standards, UDDI and ebXML Registry. A UDDI registry offers a subset of capabilities
offered by an ebXML Registry. Published in a UDDI registry are pointers to service
artifacts such as WSDL. Published in an ebXML Registry is not just pointers to service
artifacts, but also the actual artifact. Thus, an ebXML registry-repository can be used for
governance of any type of service artifacts throughout their life cycles.
Page 55 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 56 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Topics
Data / Metadata Analysis & Design
Fact Finding Techniques
Requirements Definition and Management
Data Model Components
Logical Data Model
Dimensional Warehouse
Object Oriented / UML
Data Representations in Process Models
Data / Metadata Model Management
Types of Data Models
Scope of Model and Metadata
Data Model Support
Page 57 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Questions
3.1
Data / Metadata Analysis & Design
3.1.1 Fact Finding Techniques
1. What type of fact-finding technique works best when dealing with numerous
divisions?
A. Interviewing
B. Surveys, questionnaires
C. JAD session
D. Legacy systems analysis
2. What are the benefits of a JAD Workshop?
1. Communication and combined knowledge
2. Build consensus and ownership
3. Improve design quality
4. Design cross-functional solutions
A.
B.
C.
D.
1&2
2&3
3&4
All of the above
3. What type of fact-finding technique minimizes time and assists in narrowing scope?
A. Interviewing
B. Surveys, questionnaires
C. JAD session
D. Legacy systems analysis
4. What type of fact-finding technique is a systematic attempt to collect information
from a person?
A. Interviewing
B. Surveys, questionnaires
C. JAD session
D. Legacy systems analysis
5. What type of fact-finding technique is always used in data warehousing projects?
A. Interviewing
B. Surveys, questionnaires
C. JAD session
D. Legacy systems analysis
Page 58 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 60 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
3.2
Data Model Components
3.2.1 Logical Data Modeling
17. What are the major components in an Entity-Relationship diagram?
A. Attributes, relationships, and associations;
B. Object types, relationships, associations, and supertype/subtypes;
C. Object types and transitions, associations, and supertype/subtypes;
D. States and transitions;
18. What does the following Entity-relationship diagram describe?
Negotiates
Price
Buyer
Agent
Seller
1. Real estate agent negotiates price between buyer and seller.
2. Buyer negotiates price with seller, through real estate agent
A.
B.
C.
D.
1.
2.
1&2
Neither 1 or 2
Page 61 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Manager
1.
2.
3.
4.
A.
B.
C.
D.
Contract
Employee
Hourly
Employee
Employee
Manager
Contract Employee
Hourly Employee
1
2
3&4
2, 3 & 4
1&2
1&3
1, 2, & 3
1, 2, 3, & 4
Page 62 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
A.
B.
C.
D.
Organization
Unit
25. What is the relationship between the Primary Key and Foreign Keys?
A. One-to-one;
B. One-to-many;
C. Many-to-Many;
D. Foreign Keys do not relate entities;
26. What is the Attribute that uniquely identifies an entity is called?
A. Entity Type
B. Entity Occurrence
C. Primary Key
D. Foreign Key
Page 63 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
27. What is the difference between an Entity Type and Entity Occurrence?
A. An Entity Type is something that exists and is capable of being described and an
entity occurrence is a relationship;
B. An Entity Type is the definition and the entity occurrence is an instance of the
Entity;
C. An Entity Type is a physical object type and the entity occurrence is the project.
D. There is no difference.
28. What data model is the Normalization Process applied?
A. Conceptual Data Model
B. Logical Data Model
C. Physical Data Model
D. Metadata Data Model
29. What is the objective of the Normalization process?
A. To identify the one best place an attribute belongs.
B. To organize the physical design of the data model into tables and columns.
C. To organize columns based on the mathematical principles of set theory.
D. To assign an attribute to multiple entities.
30. What are the attributes of a data model in First Normal Form?
A. All repeating groups have been eliminated
B. Every attribute describes completely that entity and not an entity identified by
only part of the primary identifier.
C. Data items that do not describe the entire primary key of the entity are eliminated.
D. Identified restrictions that apply to the data and its relationships.
31. What are the attributes of a data model in Second Normal Form?
A. All repeating groups have been eliminated
B. Every attribute describes completely that entity and not an entity identified by
only part of the primary identifier.
C. Data items that do not describe the entire primary key of the entity are eliminated.
D. Identified restrictions that apply to the data and its relationships.
32. What are the attributes of a data model in Third Normal Form?
A. All repeating groups have been eliminated
B. Every attribute describes completely that entity and not an entity identified by
only part of the primary identifier.
C. Data items that do not describe the entire primary key of the entity are eliminated.
D. Identified restrictions that apply to the data and its relationships.
Page 64 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
33. What are the two principle types of static relationships in a class diagram?
A. Primary Key and Foreign Key.
B. Association and Subtype.
C. Cardinality and Optionality.
D. One-to-One and One-to-Many.
Page 65 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 66 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Exactly one.
Many.
Optional.
Number specified.
Page 67 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
49. In the following diagram, what does the association show in terms of multiplicity?
*
A.
B.
C.
D.
Exactly one.
Many.
Optional.
Number specified
50. In the following diagram, what does the association show in terms of multiplicity?
0..1
A.
B.
C.
D.
Exactly one.
Many.
Optional.
Number specified
51. In the following diagram, what does the association show in terms of multiplicity?
1..10
A.
B.
C.
D.
Exactly one.
Many (zero or more)
Optional (zero or one)
Number specified
A.
B.
C.
D.
Aggregation.
Composition.
Ordered Role.
Not an association type.
Page 68 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 69 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
3.3
Data / Metadata Model Management
3.3.2 Types of Data Models
60. What data model is composed of subject areas, relationships, and subject area
definitions?
A. Conceptual Data Model
B. Logical Data Model
C. Physical Data Model
D. Dimensional Data Model
61. What is the difference between a Conceptual Data Model and Enterprise Data
Model?
A. A Conceptual Data Model is a type of Business Model while an Enterprise Data
Model is a physical data model.
B. A Conceptual Data Model describes the whole enterprise business subject areas
while an Enterprise Data Model is a decomposition of subject area entities.
C. A Conceptual Data Model is a logical data model while an Enterprise Data model
is a physical data model.
D. A Conceptual Data Model is a concept applied to the enterprise while an
Enterprise Data Model is applied to databases.
62. In a typical organization, the order of creation of data models is in which of the following
orders?
A. Conceptual, Enterprise, Logical, Physical
B. Enterprise, Conceptual, Logical, Physical
C. Logical, Conceptual, Enterprise, Physical
D. Physical, Logical, Enterprise, Conceptual
63. What data model is composed of tables and columns?
A. Conceptual Data Model
B. Logical Data Model
C. Physical Data Model
D. Dimensional Data Model
64. What data model is geared to a decision support environment?
A. Conceptual Data Model
B. Logical Data Model
C. Physical Data Model
D. Dimensional Data Model
Page 70 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 71 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Create/Store
Modify/Update
Delete
Shred
Page 72 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
77. A data model represents financial data for an organization. What is scope of data
model?
A. Enterprise
B. Business Area
C. Project Oriented
D. Subject Area
78. A project has begun to track costs for starting up operations in Plant A.
scope of data model?
A. Enterprise
B. Business Area
C. Project Oriented
D. Subject Area
What is
Page 73 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
87. Which of the following is not a reason for using Model Merge feature?
A. To create an enterprise data model.
B. To bring individual data models together in a group.
C. Two previously unrelated projects have merged.
D. Comparing two data models to detect changes.
Page 74 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
88. Which one of the following does not apply when importing Customer data model
with New Customer data model?
A. Create a new data model
B. Merge New Customer into Customer data model
C. Merge Customer into New Customer data model
D. The functionality is not allowed.
89. Which one is not typically represented in the breadth of data models in data
modeling tools?
A. Enterprise Data Model
B. Logical Data Model
C. Physical Data Model
D. Business Process Model
90. What is the benefit of linkages and mappings between enterprise, logical, and
physical data models?
A. To define the different purposes of the data models in the application
development process.
B. To maintain links between the different data models.
C. Synchronize changes between data models.
D. Applying transformation functions to the data models.
Page 75 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Quick Answers
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
C
D
A
A
D
B
A
A
B
A
D
D
D
D
B
C
B
C
D
A
D
B
A
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
36.
37.
38.
39.
40.
41.
42.
43.
44.
45.
46.
A
B
C
B
B
A
A
B
C
B
C
D
C
B
A
A
A
A
B
A
A
A
D
47.
48.
49.
50.
51.
52.
53.
54.
55.
56.
57.
58.
59.
60.
61.
62.
63.
64.
65.
66.
67.
68.
69.
C
A
B
C
D
A
A
C
A
A
D
A
A
A
B
A
C
D
A
D
B
A
A
70.
71.
72.
73.
74.
75.
76.
77.
78.
79.
80.
81.
82.
83.
84.
85.
86.
87.
88.
89.
90.
B
A
D
C
B
D
B
D
C
A
B
D
D
A
A
C
B
D
D
D
C
Page 76 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Detailed Answers
1. Answer: C. JAD Session: Joint Application Design session is a method for performing
analysis that brings specific parties together within a workshop environment. Surveys
and questionnaires must be made up of closed questions and typically involve
clarification and discussion after the results are tabulated. Interviews are best done in
small groups. Legacy system analysis involves examining and probing the legacy
systems.
2. Answer: D. All of the above are the benefits of a JAD Workshop.
3. Answer: A. Surveys, questionnaires. Surveys and questionnaire is an effective
technique to get opinions from a wide variety of stakeholders in an organization. Surveys
and questionnaires must be made up of closed questions and typically involve
clarification and discussion after the results are tabulated. Joint Application Design
session is a method for performing analysis that brings specific parties together within a
workshop environment to collect requirements. Interviews are best done in small groups
and can be used when scope is unknown. Legacy system analysis involves examining
and probing the legacy systems.
4. Answer: A. Interviewing. Surveys and questionnaire are typically done electronically.
Joint Application Design is a workshop environment to collect requirements. Interviews
are best done in small groups where heuristic questions can be asked. Legacy system
analysis involves examining and probing the legacy systems and may or may not involve
discussion with staff.
5. Answer: D. Legacy systems analysis. Legacy system analysis is used in data
warehousing projects to perform source(legacy system) to target(data warehouse)
transformations.
6. Answer: B. Open-ended questions. Open-ended questions cannot be answered with a
simple yes or no response and thus encourage the interviewee to provide more
information. Closed-ended questions give a yes or no answer. Leading questions put the
interviewees opinion into the question and do not give an opportunity for the interviewer
to answer without bias.
7. Answer: A. When the interviewer wants to gain a broad based view on an issue that
needs to be explored. The unstructured interview is used when the interviewer wants to
explore an issue and facilitates description of domain in a way that is easy for the
interviewee.
8. Answer: A. Evaluation of current environment and documentation. After gathering
user requirements for a new system, the next best step is an evaluation of the current
environment and documentation to complete the current state assessment. Next, a future
state can be derived with the linkages to the current state if needed. A gap analysis is
Page 77 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
then conducted to compare its current state with its future state to determine the variance
between business requirements and current capabilities.
9. Answer: B. Future state. After creating a current state environment evaluation, the
next best step is to complete the future state assessment. Next, a gap analysis is then
conducted to compare its current state with its future state to determine the variance
between business requirements and current capabilities.
10. Answer: A. Gap analysis. After creating a current state environment evaluation and
future state creation, the next best step is to complete the gap analysis. A gap analysis
compares current state with future state to determine the variance between business
requirements and current capabilities.
11. Answer: D. It specifies a Pre or Post condition of an entity. Business rules are put in
business terms not in terms of conditions on entities. A business rule:
Is a statement that defines some facet of the business.
Asserts business structure or controls or influences performance of the business
and can be applied across the organization.
Is at the lowest level and cannot be decomposed further or it would lose business
meaning.
12. Answer: D. A customer with preferred status should have its orders filled as soon as
possible. Business rules should contain the words: must; must not; should; should not;
only; only if. The following did not qualify as business rules due to the following
reasons:
When a failure is reported, an expeditor is assigned by the maintenance department
who sends the failure form to the service desk for scheduling. Ordering of events in
the business rule is declarative. Business rules should be procedural.
If Acct_num is between 0 and 5000 then the customer is a member of the branch that
may deposit money. Business rules should not contain technology nomenclature but
be solely about the business.
A customer places an order. This rule can be decomposed into further rules.
13. Answer: D. Rule Engine. The typical business rule creation process is discovery,
validation and documentation. Business rules are discovered as part of a formal
requirement gathering process during the initial stages of design. Once they are
documented, they are validated to ensure consistency and non-conflicting business rules.
Finally, they are documented. In some organizations, software packages are used to store
business rules.
14. Answer: D. ABAP. ABAP is a language used in SAP. If the situation merits,
business rules can be expressed in a very formal language, for example: UML, Z
notation, Business Rules Markup Language, Business Process Execution Language
(BPEL) and Business Process Modeling Notation (BPMN).
Page 78 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
15. Answer: B. To ensure the system performs as it should. The benefit of Requirements
tracking and management to implementation is to ensure the system performs as it
should. The requirements should be classified in a matrix that has a listing of the
requirements for the entire project. The type of information about each requirement in
the matrix should consist of the following information: the number assigned to the
requirement, a brief description, the date submitted to project, and the tracking of the
requirement as it relates to development. The matrix can determine multi-modal
requirements.
16. Answer: C. To clarify dependencies of information flow. The matrix has the
following structure: the columns in the matrix represent the major processes, and the
rows the Entities and Attributes. Elements of the matrix thus represent interactions
between a Process and Data. Matrix elements marked with an X represent data/process
interactions.
17. Answer: B. Entity-relationship diagrams consist of two major components: Object
types and relationships. Object types represent a collection, or set, or objects that can be
identified uniquely, and can be described by one or more facts (represented by a
rectangular box). Relationships represent a set of connections, or associates, between the
object types (represented by diamonds). Associations represent a relationship that we
need to maintain information. The subtype/supertype indicator consists of an object type
and one or more subcategories; connected by a relationship.
18. Answer: C. The diagram shows that both descriptions are valid.
19. Answer: D. Manager, Contract Employee and Hourly Employee are examples of
subtypes. Employee is the general category and the subcategories are: Manager,
Contract Employee and hourly employee.
20. Answer: A. Attributes are an atomic fact or characteristic, which describes an entity.
It possible to differentiate between an entity and attribute by examining whether it can
stand-alone and hold meaning. For example Street Name only makes sense when it
resides in the context of Employee. Street Name is an attribute of Employee entity an as
such modifies Employee. Employee is an entity. One occurrence of Employee entity
might be the employee Johnson. Attributes such as Street Name, City, and Phone
Number are attributes, which describe or modify Employee Johnson.
21. Answer: D: 1, 2, 3, & 4 Attributes roles identify, describe and relate attributes.
22. Answer: B. Cardinality can be defined as the relative number of occurrences, which
may exist between a pair of entities. There are three kinds of relationships: one-to-one;
one-to-many; and many-to-many. In a one-to-one relationship between two entities, at
any one time there exists only one occurrence of the entity (Customer and
Employee). In a one-to-many relationship between two entities, at any one time, there
may exist multiple occurrences of the entity Customer for each entity of Customer
Page 79 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
describe the entire primary key of the entity are eliminated. A simple rhyme to remember
the ordering is "the key, the whole key and nothing but the key", so help me Codd.
31. Answer: B. Every attribute describes completely that entity and not an entity
identified by only part of the primary identifier. A data model in First Normal Form has
all repeating groups eliminated. A data model in Second Normal Form has every
attribute describes completely that entity and not an entity identified by only part of the
primary identifier. A data model in Third Normal Form has data items that do not
describe the entire primary key of the entity are eliminated. A simple rhyme to remember
the ordering is "the key, the whole key and nothing but the key", so help me Codd.
32. Answer: C. Data items that do not describe the entire primary key of the entity are
eliminated. A data model in First Normal Form has all repeating groups eliminated. A
data model in Second Normal Form has every attribute describes completely that entity
and not an entity identified by only part of the primary identifier. A data model in Third
Normal Form has data items that do not describe the entire primary key of the entity are
eliminated. A simple rhyme to remember the ordering is "the key, the whole key and
nothing but the key", so help me Codd.
33. Answer: B. Association and Subtype. The two principle types of static relationships
in a class diagram are association and subtype. Primary key and foreign key, cardinality
and optionality, and one-to-one and one-to-many describe data models.
34. Answer: C. Fact tables and dimension tables. Every dimensional model is composed
of one table with a multipart key called a fact table and a set of tables called dimension
tables that describe the dimensions of the fact table. Examples of dimension tables are:
Time, Store, Product, Customer, and Employee while the fact table could be: Sales. The
Data Warehouse Bus Architecture may be defined as: A master suite of conformed
dimensions and to standardize the definitions of facts. Process, data stores, flows and
terminators are part of data flow diagrams. Entities, data and relationships are part of
data modeling. Processes, data stores, relationships and flows are a combination of data
flow diagrams and data modeling.
35. Answer: D. dollars_sold is a fact attribute. The most useful facts in a fact table are
numeric, additive and continuously valued. Continuously valued means that every time
the attribute is sampled, it can take on different values. Dimension tables, most often
contain descriptive textual information. Clerk_grade, address and promotion_name
would all be dimensions.
36. Answer: C. Snowflaking removes low-cardinality textual attributes from dimension
tables and places them in joined secondary dimension table. Snowflaking may be used
both logically and physically. For example Customer may be separated from Customer
Type in two dimension tables.
37. Answer: B. OLAP. Dimensional Modeling is best used in OLAP type applications
for browsing, performance, and user understandability. OLTP systems or online
Page 81 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
transaction processing systems use a normalized data model. HPC or High Performance
Computing uses a normalized model that typically resides in memory for faster
transaction time. Web is similar to OLTP in the data model used.
38. Answer: A. type 1. In a Dimensional Model, when tracking changes in a slowly
changing dimensional table, the old value is discarded and has no significance is regarded
as type 1. Type 1 overwrites the old record and does not track changes. Type 2 tracks full
changes and partitions history of the dimension table. Type 3 tracks old and new
definitions on the same record. Cross-dimensional attribute is an attribute that describes
an attribute but may also be counted and could reside in either a fact or dimension table.
39. Answer: A. type 2. In a Dimensional Model, when tracking changes in a slowly
changing dimensional table, the old value is recorded and has significance is regarded as
type 2. Type 2 tracks full changes and partitions history of the dimension table. Type 1
overwrites the old record and does not track changes. Type 3 tracks old and new
definitions on the same record. Cross-dimensional attribute is an attribute that describes
an attribute but may also be counted and could reside in either a fact or dimension table.
40. Answer: A. type 3. In a Dimensional Model, when tracking changes in a slowly
changing dimensional table, the old value and new value are equally important is
regarded as type 3. Type 1 overwrites the old record and does not track changes. Type 2
tracks full changes and partitions history of the dimension table. Type 3 tracks old and
new definitions on the same record. Cross-dimensional attribute is an attribute that
describes an attribute but may also be counted and could reside in either a fact or
dimension table.
41. Answer: A. In the Customer Dimension, a Slowly Changing Type 2 Dimension, the
key would be Customer_Key, which is a new generated, meaningless, key or surrogate
key. The type 2 response requires the use of a surrogate key to fully track the changes of
the record. Type One Dimension is overwritten. Intelligence in the key like Time or
Version numbers should not be used when creating keys.
42. Answer: B. Class. The basis of modularity and structure in object-oriented modeling
is called class. A class is a grouping of data and behaviour for a concept. Entity and
Relationship are concepts from data modeling not object-oriented modeling. An object is
an instance of a class. Each object has its own data, though the code within a class.
43. Answer: A. An instance of a class. An object is an instance of a class. Each object
has its own data, though the code within a class.
44. Answer: A. Encapsulation. The term for a type of privacy applied to the data and
some methods of a call is known as encapsulation. Inheritance is a method of
generalization that creates subtypes. Abstraction is the ability of a function to have
different specifications. Polymorphism is the ability of objects belonging to different
types to respond to method calls of methods of the same name, each one according to the
right type-specific behavior.
Page 82 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
referential integrity between tables that have a relationship. There are typically three
triggering EVENTS that cause trigger to 'fire':
INSERT event (as a new record is being inserted into the database).
UPDATE event (as a record is being changed).
DELETE event (as a record is being deleted).
57. Answer: D. Supertype. The following are typical types of referential integrity
triggers: Identifying, non-identifying (allowing nulls and non-null), and subtype. The
typical actions that the triggers can conduct are: CASCADE, RESTRICT, SET NULL
(non-identifying-nulls allowed), SET DEFAULT (non-identifying), and NONE.
58. Answer: A. Stored procedures can accept parameters while Triggers cannot. Stored
Procedure is a program which, like a Trigger is physically stored in a database. A stored
procedure and trigger can both be used to enforce Referential Integrity and simply data
management.
59. Answer: A. UML standardizes representation of object oriented analysis and design.
While all of the other statements are true, the overall benefit of UML is that it
standardizes representation of the object oriented analysis and design.
60. Answer: A. A conceptual model is composed of subject areas, relationships, and
subject area definitions. A Conceptual Model is a type of Business Model.
61. Answer: B. A Conceptual Data Model describes the whole enterprise business subject
areas while an Enterprise Data Model is a decomposition of subject area entities. A
Conceptual Model is a high-level starting-point for design and construction activities leading
to implemented information systems that fulfill important business needs. The Enterprise Data
Model is at a lower level of detail than the conceptual model. A typical conceptual model of a
whole enterprise might consist of 7-9 subject areas, representing major business subject areas.
An Enterprise Data Model (EDM) while still not containing all entities or all relationship will
have many entities for each subject area.
62. Answer: A. Conceptual, Enterprise, Logical, Physical. In a typical organization, the order
of creation of data models is Conceptual, Enterprise, Logical, Physical.
63. Answer: C. A Physical data model is composed of tables and columns. Physical data
models are representations of models that specify database or file structure implementations.
64. Answer: D. A Dimension Data Model is geared to decision support environments.
Data is modeled for retrieval of large amounts of data. Design for high volume retrieval is
coupled with specialized administration skills and techniques, and often specialized
dimensional database management systems.
65. Answer: A. Conceptual Data Model. A Conceptual Data Model describes the whole
enterprise business subject areas.
Page 84 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
66. Answer: D. Enterprise Data Model. An Enterprise Data Model (EDM) while still not
containing all entities or all relationship will have many entities for each subject area.
67. Answer: B. Entities and Relationships. An Enterprise Data Model is comprised of
Entities and Relationships that are organized into Subject Areas. The model may include
attributes of the Entities.
68. Answer: A. Data models that specify one or more other data models. Metamodels are
the details behind the metadata that depict metadata relationships.
69. Answer: A. Metamodels are data models that specify other data models, while metametamodels defines ontology.
70. Answer: B. Meta Data Coalition (MDC). Meta Data Coalition is a group that defined
the Metadata Interchange Specification. CDIF, CWM, MDIS are all industry standards.
71. Answer: A. Standard data models for an industry that may be used off the shelf.
Standard data models are widely used in an industry and shared among different
companies.
72. Answer: D. Shred. The data life cycle phases are: Create/Store, Retrieve,
Modify/Update, Read/Use, Transport, Archive and Delete. The data lifecycle is the
process of managing data throughout its lifecycle from conception until disposal, within
the constraints of the data policy.
73. Answer: C. Causes internal gridlock and inconsistencies. The Enterprise wide data
model known as a Conceptual Data Model describes the whole enterprise business
subject areas, supports the entire enterprise, a wide audience, provides a data picture of
the business that is capable of being easily extended to capture new requirements. The
Enterprise wide data model leads to and integrated data picture that breaks down
inconsistencies and promotes data knowledge, use and sharing.
74. Answer: B. Corporation. The Enterprise wide data model, known as the Conceptual
Data Model describes the whole corporation not just a business unit, geographic unit or
functional area.
75. Answer: D. The enterprise data model will frequently change as new requirements are
determined. The enterprise data model should remain stable unless for example, a new
company is acquired. The enterprise wide data model is driven by the business, which
encompasses areas of concern or importance to the corporation. The corporate data
architect owns the enterprise wide data model.
76. Answer: B. Business Area. The scope of a data model that represents Marketing data
for an organization is business area. The data model may be specific to Marketing area
needs or may have access restricted to the Marketing group. The data model would be a
part of the enterprise data model.
Page 85 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
77. Answer: D. Subject Area. The scope of a data model that represents financial data for
an organization is subject area. Every area in the organization needs financial
information of some sort, not just the Finance group. The Finance group will see the
financial data for an entire organization.
78. Answer: C. Project Oriented. The scope of a data model that tracks costs for starting
up operations in Plant A is Project Oriented. The data will only need to be collected
while the project is in progress, and analyzed after the closure of the project.
79. Answer: A. Forward Engineering. Forward Engineering the data model is already
created and the data modeling tool is creating the database scripts and will apply the
scripts to create a database and/or tables in the database. Reverse Engineering captures
information from a database or script file to create a physical data model. Creating
Logical and Physical data model involves using the GUI to create a data model that
conforms to business requirements. Splitting models of older version into separate logical
and physical data models can create more manageable data models.
80. Answer: B. Reverse Engineering. Reverse Engineering creates a data model from an
existing database. Forward Engineering the data model is already created and the data
modeling tool is creating the database scripts and will apply the scripts to create a
database and/or tables in the database. Creating Logical and Physical data model
involves using the GUI to create a data model that conforms to business requirements.
Importing a prior version data model is looking at an older version of the data model.
81. Answer: D. Create the Enterprise Data Model. Reverse Engineering benefits are:
maintaining database, changing database types, analyzing the differences in databases.
Reverse Engineering functionality cannot create the Enterprise Data Model unless the
database has enterprise wide data.
82. Answer: D. Improving. Forward Engineering, Reverse Engineering and Comparisons
are all typical Data Model tool functions. Improving is a function that does not exist in
data modeling tools. Forward Engineering the data model is already created and the data
modeling tool is creating the database scripts and will apply the scripts to create a
database and/or tables in the database. Forward Engineering is also referred to as
exporting. Reverse Engineering creates a data model from an existing database. The
comparison function allows for comparisons between the data model and database; select
the objects to compare and selectively import or export changes.
83. Answer: A. Keep data model and database synchronized. The benefit of Comparison
function in data modeling tools is that it allows the data modeler or database
administrator to keep the data model and database synchronized. The comparison
function allows for comparisons between the data model and database; select the objects
to compare and selectively import or export changes. These are features of the
comparison function. The main benefit is the data model and database is synchronized.
Page 86 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
84. Answer: A. Forward Engineering. Forward Engineering the data model is already
created and the data modeling tool is creating the database scripts and will apply the
scripts to create a database and/or tables in the database. Forward Engineering is also
referred to as exporting.
85. Answer: C. Versioning. Versioning records who made the change and when the
change was made to provide full project audit trail thus enabling rollback options when
comparing versions. Change control is a function that allows knowing the impact of
change before saving. Model Merge is simply merging two data models together.
Submodeling breaks a data model into smaller models (e.g. by subject area) for ease of
use.
86. Answer: B. Model Merge. Model Merge supports merging of data models in an
enterprise. Versioning records who made the change and when the change was made to
provide full project audit trail thus enabling rollback options when comparing versions.
Change control is a function that allows knowing the impact of change before saving.
Submodeling breaks a data model into smaller models (e.g. by subject area) for ease of
use.
87. Answer: D. Comparing two data models to detect changes. Model Merge features
can be used to bring two data models together. Model Merge can: create an enterprise
data model, bring individual data models together in a group, and merge two previously
unrelated projects. If a data modeler wanted to compare two data models to detect
changes, they would use a Comparison function if they were previously linked.
88. Answer: D. The functionality is not allowed. When importing and merging data
models, data modelers can typically: create a new data model or update one of the
existing data models.
89. Answer: D. Business Process Model. Although some tools offer linkages between
data model and business process tools and have synchronization options, the tools are still
typically separate. The benefit of this synchronization is it verifies the data models
support the business processes and vice versa. Data Modeling tools have the breadth of
data models that represent: Enterprise Data Model; Logical Data Model; and Physical
Data Model.
90. Answer: C. Synchronize changes between data models. The benefit of linkages and
mapping between enterprise, logical and physical data models is to synchronize changes
between data models. The data models do not need to be stored in the same file, but need
to maintain links between the different data models. Different data models are used as
they serve different purposes in the application development process. A transformation
function may include items like many-to-many where the relationship is dissolved by an
identifying entity.
Page 87 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Topics
Standards, Policies, Procedures, Guidelines
Standards Management Process
Data Models
Data Elements
Data Security and Privacy
Data Security Principles
Data Security Policy Types
Page 88 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Questions
4.1 Standards, Policies, Procedures, Guidelines
1. What is the name of an entity?
A. Customer
B. Customers
C. Customer_Name
D. Name
2. Which is the best name for an entity describing employee information:
A. EmployeeTable
B. Employee_Table
C. EmployeeTbl
D. Employee
3. Which attributes follows best practises in naming?
A. Social-insurance-number
B. Social-insurance
C. Social-insurance-code
D. Social-insurance-numbers
4. When naming primary keys attributes follows best practises in naming?
A. Customer-Ident
B. Customer-Ids
C. Customer-Id
D. Customer-Identifier
5. Which of the following should be used when naming a relationship between an
employee and a manager?
A. Supervises
B. Supervise
C. Supervisor
D. Is a
6. Which of the relationships capture a hierarchal relationship between employees:
A. Has a
B. Supervises
C. Is a
D. Manages
Page 89 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Cleaning
Redundancy resolution
Business rule enforcement
Random sampling.
Page 90 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
12. Which one of the following is not a data element representation types:
A. Amount
B. Code
C. Date
D. Customer
13. A data element name that conforms to the ISO/IEC 11179 metadata registry naming
convention does not have one of the following:
A. Object
B. Property
C. Representation term
D. Process Definition
14. In the ISO/IEC 11179 metadata is defined as:
A. Data about data
B. Data that defines and describes other data
C. DNA of the data
D. All information that is not the data itself.
15. What is the process for creating metadata when using a metadata registry?
A. Attributing, Classifying, Defining, and Registering
B. Attributing, Classifying, Defining, and Maintaining
C. Creation, Approval, Enforcement, Registering
D. Creation, Approval, Enforcement, Maintenance
16. Which of the following is not considered a sampling technique of data element audits:
A. Random Sampling
B. Systematic Sampling
C. Cluster Sampling
D. Standard Deviation.
17. In a data element audit, the data element must be:
A. Valid and accurate
B. Conformance of data values to its domain and business rules
C. Complete
D. Not null
Page 91 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 92 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
B. Confidential
C. Restricted
D. Classified
29. Which of the following applies the most when classifying data in a data warehouse?
A. Only the summarized financial information needs to be confidential.
B. Only detailed financial records need to be confidential.
C. Data classification depends on the data content.
D. All data in a data warehouse should be available.
30. Audit trail inspection is a classified under what type of security monitoring?
A. Proactive
B. Reactive
C. Offensive
D. Defensive
31. When data is considered mission critical, what is the class and data availability
required, according to the Storage Networking Industry Association (SNIA) based upon
Five 9s?
A. Class 2 - 99% data availability
B. Class 3 - 99.9% data availability
C. Class 4 - 99.99% data availability
D. Class 5 - 99.999% data availability
32. When data is classified as Class 1 90% data availability, what is the business
classification of the data according to the Storage Networking Industry Association?
A. Not important to Operations
B. Important for Productivity
C. Business Important Information
D. Business Vital Information
33. Which of the following statements is the not true when data content is controlled in
an organization?
A. Business policies or regulatory rules require some/all data be retained X period
of time.
B. May need to prove records stored are trustworthy at later date
C. High data value to an organization need to be accessible, available and protected
D. Data availability is Class 1 90% data availability.
Page 94 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
34. Which of the following statements is the not true when data content is non-controlled
in an organization?
A. No business rules or regulations are requiring this data be kept for X period of
time.
B. Business just needs to keep the data archived and accessible
C. High data value to an organization needs to be accessible, available and protected.
D. Med-Low data value to an organization needs to be accessible and available.
Page 95 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Quick Answers
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
A
D
A
D
A
C
C
D
A
B
D
D
D
B
A
D
A
A
A
C
A
A
A
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
C
B
A
A
D
C
D
D
A
D
C
Page 96 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Detailed Answers
1. Answer: A. Customer. Naming an entity should follow a standard and uniform
approach. Entity Names should be simple, clear, and expressed in business terms.
Entities should be in noun or adjective noun format; singular; in business terms; and not
process specific. Customers are plural. Customer_Name is not in business terms. Name
is not clear to which name it is referring.
2. Answer: D. Employee. The best name for an entity describing employee information
is simply Employee. Incorrect names include: EmployeeTable, Employee_Table, and
EmployeeTbl. Employee_Table is not in business terms. EmployeeTbl is not clear.
EmployeeTable is not in business terms.
3. Answer: A. Social-insurance-number. An attribute should be in the singular,
consistent, and clearly defined in business terms. Properly defined attributes should
define the data domain type like date, time, amount, code, name, quantity and
description. Social-insure is unclear domain type. Social-insurance-numbers is plural.
Social-insurance-code does not define the contents.
4. Answer: D. Customer-Identifier. An attribute should be in the singular, consistent, and
clearly defined in business terms. Completely spelling out the attribute is preferred. If
using an abbreviation, they should be as clear as possible and used consistently across the
enterprise. If using abbreviations, whenever possible, use industry standard
abbreviations.
5. Answer: A. Supervises. Relationships capture how two or more entities are related to
one another. Relationships should be named after plural verbs.
6. Answer: C. Is a. Is a captures a hierarchal relationship between employees and the
different classes of employees or subtypes. In this relationship, the new class or object
has data or behavior aspects that are not part of the inherited class.
7. Answer: C. Repeat customers do not need a credit check. Business rules describe the
operations, definitions and constraints for governing policies. The business rule needs to
define in clear and concise business terms and easily applied.
8. Answer: D. Random sampling. Data integrity rules in a data warehouse environment
include: cleaning, redundancy resolution and business rule enforcement. Random
sampling is a technique and tool that may be used to conduct checks on the data.
9. Answer: A. ISO/IEC forms the specialized system for worldwide standardization,
while ANSI that administers and coordinates the U.S. voluntary
standardization and conformity assessment system. ISO (the International
Organization for Standardization) and IEC (the International Electro technical
Commission) form worldwide standards through technical committees. ANSI (The
Page 97 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Page 98 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
19. Answer: A. Use data warehouse record and metadata to navigate back to the record
in the source system. Prior to the source data being transformed to the data warehouse, it
should be standardized, cleansed, completed, enhanced, consolidated and summarized
where needed. The knowledge workers should be able to analyze the degree to which the
data agrees with original source. Complete the data in a data warehouse by adding postal
information and the ability to consolidate the data should be done prior to loading the
data into the data warehouse. The knowledge worker should be told of the data quality in
the data warehouse. In analyzing the data, they should not have to conduct a baseline
assessment of the quality of the data.
20. Answer: C. Employee - an individual who has entered into or works under (or, where
the employment has ceased, worked under) a contract of employment. Metadata
principles are: state the essential meaning of the concept; be precise and unambiguous
(Answer B); contain only commonly understood abbreviations(Answer C); be concise; be
able to stand alone; be expressed without embedding rationale, functional usage, domain
information, or procedural information; avoid circular reasoning (Answer A); and use the
same terminology and consistent logical structure for related definitions.
21. Answer: A. Agreement. The need for data security can be derived from the three
principles of: Accountability, Authorization, and Availability.
22. Answer: A. Accountability. Accountability is the concept that every user must be
responsible for their actions, so that in the event of any questionable activity or breach of
policy, a specific user can be identified. The specific security services that support
accountability are identification, authentication, and auditing. Identification refers to a
security service that recognizes a claim of identity by comparing a userid offered with
stored security information. Authentication refers to a security service that verifies the
claimed identity of the user, for example a password. Auditability refers to a security
service that records information of potential security significance. Authorization is a
concept that access to data and system resources should be limited to a need to know
basis, and that specific users must be specifically allowed such access. Availability is the
concept that system and data resources must be accessible whenever they are needed.
23. Answer: A. Authorization is a concept that access to data and system resources
should be limited to a need to know basis, and that specific users must be specifically
allowed such access. Access control refers to a security service that allows or denies a
user request based on privilege, group information, or context. The specific security
services that support authorization are access control and confidentiality. Confidentiality
refers to a security service that prevents disclosure of information to unauthorized parties
while the information is in use or in transit, or while the information is being stored or
destroyed. Accountability is the concept that every user must be responsible for their
actions, so that in the event of any questionable activity or breach of policy, a specific
user can be identified. Availability is the concept that system and data resources must be
accessible whenever they are needed.
Page 99 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
24. Answer: C. Availability. Availability is the concept that system and data resources
must be accessible whenever they are needed. The necessity for availability is dependent
upon your particular business proposition. The specific security service that supports
availability is integrity. Integrity refers to a security service that guarantees data has not
been altered, deleted, repeated, or rearranged during transmission, storage, processing, or
recovery. Accountability is the concept that every user must be responsible for their
actions, so that in the event of any questionable activity or breach of policy, a specific
user can be identified. Authorization is a concept that access to data and system
resources should be limited to a need to know basis, and that specific users must be
specifically allowed such access.
25. Answer: B. Management of the data asset accessibility. The main objective of the
data steward responsibilities with respect to defining the security policy is the
management of the data asset accessibility. The data steward does not ensure compliance
to the security policy, define the data quality standards, or define the security
requirements, controls and mechanisms applicable to all data assets.
26. Answer: A. Define the security requirements, controls and mechanisms applicable to
all data assets. The trustee is entrusted with the administration of the data assets.
27. Answer: A. Provides privacy protection in relation to personal data. OECD
Guidelines on the Protection of Privacy and Transborder Flow of Data provides privacy
protection in relation to personal data. The Guidelines apply to personal data, no matter
if the company is public or private sectors, because of the potential detriment to civil
liberties. The guidelines specify how to collect, store, process or disseminate personal
information.
28. Answer: D. Classified. The typical security classified levels are top secret, secret,
confidential, restricted and the lowest level of unclassified. The levels determine the
impact on national security if the data was to be made public. Corporations typically
have a similar type of classification structure.
29. Answer: C. Data classification depends on the data content. When classifying data in
a data warehouse, the data content needs to be evaluated. In some cases the detailed data
should be confidential and in others, the summarized data. The challenge in data privacy
is to share data while protecting the identifiable information.
30. Answer: D. Defensive. Audit trail inspection is classified under a Defensive type of
security monitoring. Other types of defensive monitoring are: Role Based Access
Security and Process Rights Management. Offensive monitoring includes: Provisioning
and Federated Identity Management.
31. Answer: D. Class 5 99.999% data availability. Mission critical data availability
according to the Storage Networking Industry Association based upon Five 9s is Class 5 99.999%. The five classes are:
Class 1 - 90% data availability: Not important to Operations
Page 100 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Topics
Information Quality Principles
Definition
Information Quality Characteristics
Data Definition (or Information Product Specification) Quality Characteristics
Information Quality Assessment / Audit
Quality Assessment Characteristics
Quality /Cost Measurement
Information Quality Improvement
Data Corrective Maintenance
Data Movement Control
Information Quality Process Improvement
Information Quality Culture Transformation
Questions
5.1 Information Quality Principles
1. What is the difference between data and information?
A. Data is the representation of facts, information is data in context.
B. Data is the context and Information is the content.
C. Data is the value and Information is a valuable enterprise resource.
D. Data is meaningful and information has context.
2. What is Information Quality?
A. Quality data that enables knowledge workers to answer their questions.
B. Correctness or accuracy of the data and the degree of usefulness and value data
has to the organization.
C. Quality data values in a data attribute field.
D. Valid and meaningful information that the enterprise can make decisions upon.
3. Who benefits the most from Information Quality?
A. Knowledge Workers
B. Management
C. Data Modeler
D. Data Base Administrator
4. What type of data is required when the consequences of nonquality cause major
process failure?
A. Complete Quality data
B. Accurate data
C. Scientific data
D. Zero-defect data
5. According to Larry English, what are the components that make up information?
A. Information = Data + Definition
B. Information = Data + Definition + Quality
C. Information = Data + Definition + Presentation
D. Information = Data + Definition + Content
6. What is Data Definition Quality?
A. The definition, quality, and accuracy that govern data.
B. The definition, domain value set, and business rules that govern data.
C. The information about the data that meets manager expectations.
D. The information that departmental knowledge workers depend on.
13. Which of the following depends on a reliable and precise recording of information?
A. Definition Conformance
B. Completeness
C. Accuracy
D. Precision
14. Which one of the following does nonduplication principles lead to the least?
A. Identical values in multiple records.
B. Duplicate records of a single event.
C. Same data maintained in many independent, distributed databases.
D. Its not a problem if there are standards and controls.
15. Which of the following describe when data is semantically equivalent?
A. Consistency of Redundant data
B. Timeliness
C. Usefulness
D. Objectivity
16. What does Information Float describe?
A. The timetable to gather data.
B. Length of time from when data is known, until it is available for a specific process
or use.
C. The average time required for data to be disseminated in the organization.
D. The degree of variance in the data.
17. In a decision support system, which of the following is considered most useful?
A. Tabular data
B. Data list
C. Graphic presentation (bar chart, etc)
D. It depends on the type of data
18. What is presentation clarity?
A. The degree the knowledge worker can understand the meaning of the data through
the presentation.
B. Statements of fact with a neutral point of view including reporting without bias,
and emphasis on initutive presentation to the knowledge worker.
C. The ability to see the legends, rows and columns in a bar chart.
D. The redesign producing high quality data content or presentation.
19. Which one is a statement of fact with a neutral point of view, reporting without bias,
and emphasis on initutive presentation to the knowledge worker called?
A. Consistency of Redundant data
B. Timeliness
C. Usefulness
D. Objectivity
20. What are the primary inherent information quality characteristics?
A. Information Product Specification, Data Definitions, Data Models.
B. Definition, domain value set, and business rules that govern data.
C. Definition conformance, completeness, validity, accuracy, precision,
nonduplication, and consistency of redundant data.
D. Accessibility, timeliness, contextual clarity, derivation integrity, usability,
completeness.
21. What are the primary pragmatic information quality characteristics?
A. Information Product Specification, Data Definitions, Data Models.
B. Definition, domain value set, and business rules that govern data.
C. Definition conformance, completeness, validity, accuracy, precision,
nonduplication, and consistency of redundant data.
D. Accessibility, timeliness, contextual clarity, derivation integrity, usability,
completeness.
22. Which definition for Employee-start-date conforms best to the Attribute definition
quality principles?
A. The date an employee was hired.
B. The date a new employee was hired.
C. The date that an employee first started with the company regardless of location.
D. Tells when the employee first came to work
23. Using data name and definition consistency principles, which attributes best
describes: The date a service is started with the customer?
A. Product-start-date
B. Service-start-date
C. Service-release-date
D. Data-Service-release-date
24. Which of the following is the most appropriate to demonstrate Entity Type Name
Clarity?
A. Retail Customer
B. Customers
C. Buying Customers
D. RTL_CUST
25. Which of the following is the most appropriate to demonstrate Attribute Name
Clarity?
A. Retail Customer Name
B. Customers Identities
C. Buying Customers Names
D. RTL_CUST_NAME
26. Which one of the following is not known as a domain type?
A. Date
B. Time
C. Amount
D. Customer
27. For a cross-reference attribute, what is the most appropriate abbreviation following
acronym clarity principles?
A. Xref
B. Cross-ref
C. X-reference
D. Crss-rfrnc
28. Why should names be appropriate to the knowledge workers?
A. Attributes names are consistent where facts context are equivalent.
B. Attributes names are equivalent to ease dissemination of information.
C. Attribute names are accurate to express the meaning of the fact being defined.
D. Attribute names are clear and concise business terms.
29. What role in an organization is responsible for the enterprise wide glossary?
A. Business Steward
B. Knowledge Worker
C. Data Modeler
D. Database Administrator
30. What role in an organization is responsible for the business term definition?
A. Knowledge Worker
B. Data Modeler
C. Database Administrator
D. Subject Matter Expert
31. Where should business terms be defined?
A. Glossary
B. Data Model
C. Data Dictionary
D. Database
32. Which of the following is not true for Business rules:
A. Based on technical or existing system limitations.
B. Expressed in business terms.
C. Complete and specific.
D. Defines an aspect of business to take action.
47. What area is considered part of defining summary and derived data, adding data from
external sources, consolidating data?
A. Source Data Cleansing
B. Data Conversion
C. Data Scrubbing
D. Best Practice
48. Which group should conduct the quality audit and control of data movement
procedures?
A. Internal Audit
B. Knowledge Workers
C. Data Conversion Specialists
D. Information Steward
49. Which one of the following is not a cost of information quality?
A. Nonquality information costs.
B. Information quality assessments/audits.
C. Information quality process improvements and prevention.
D. Incorrect decisions made on poor quality information.
50. Which of the following is a proactive technique in Information Process Quality
Improvement?
A. Fix the symptoms.
B. Ignore Problem signs until they become issues.
C. Analyze root cause and eliminate the cause.
D. Following Best Practise in IT System Management
51. Which one of the following techniques is not used in root cause analysis for
determining quality issues?
A. Cause-and-effect diagram
B. Interrelationship diagram
C. Current reality tree
D. Value chain relationship diagram
52. Which step is not involved in the Shewhart cycle?
A. Plan and Do
B. Check
C. Act
D. Refine
53. When implementing an Information Quality Culture Transformation which one of the
following is most important to do well?
A. Training
B. Management Buy-in
C. Define a Methodology
D. Data Definition quality assessment process
54. Who is accountable for the integrity of the processes and quality of information?
A. Knowledge Steward
B. Managerial Information Steward
C. Process Steward
D. Business Information Steward
55. What is the first step an enterprise should take when embarking on an information
quality program?
A. Conduct an Information Quality Management Maturity Assessment and gap
analysis.
B. Create a vision, mission, and objectives for the information quality program.
C. Appoint an Information Quality Leader.
D. Conduct a customer satisfaction survey.
56. Which state is the least mature in the Information Quality Management Maturity
Assessment?
A. Uncertainty
B. Awakening
C. Enlightenment
D. Wisdom and Certainty
57. Which state is characterized by knowing a data quality problem exists but not
knowing what to do about it?
A. Uncertainty
B. Awakening
C. Enlightenment
D. Wisdom and Certainty
58. Which state has adopted a commitment to quality and implements the 14-point
program?
A. Uncertainty
B. Awakening
C. Enlightenment
D. Wisdom and Certainty
Page 112 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
Quick Answers
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
A
B
A
D
C
B
A
A
A
D
D
B
C
C
A
B
D
A
D
C
D
C
B
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
36.
37.
38.
39.
40.
41.
42.
43.
44.
45.
46.
A
A
D
B
A
A
D
A
A
B
D
A
D
D
D
B
B
B
C
A
D
A
B
47.
48.
49.
50.
51.
52.
53.
54.
55.
56.
57.
58.
59.
60.
61.
C
D
D
C
D
D
A
B
A
A
B
C
C
D
A
Detailed Answers
1. Answer: A. Data is the representation of facts, information is data in context. Data is
the raw material and information is the finished product.
2. Answer: B. Correctness or accuracy of the data and the degree of usefulness and value
data has to the organization. Information Quality needs to have inherent quality (data
accuracy) and pragmatic quality (usefulness and value to support the enterprise process
that enable accomplishing enterprise objectives). Data needs to support the knowledge
workers decision-making process or else it holds no value to the organization.
3. Answer: A. Knowledge Workers. Knowledge Workers benefit the most from
Information Quality because they require data to do their jobs to the benefit of the endcustomer.
4. Answer: D. Zero-defect data. Zero-defect data is required when the consequences of
nonquality cause major process failures. An example of data that must be accurate is
domain reference data like: Medical diagnosis codes.
5. Answer: C. Information = Data + Definition + Presentation. As defined by Larry
English, the three components that make up information are meaning (definition) of a fact
(data) in a context (presentation).
6. Answer: B. The definition, domain value set, and business rules that govern data. Data
Definition quality is the degree to which the data definition describes the meaning of the
data and meets the needs of all stakeholders to understand the data and the context.
7. Answer: A. Information Product Specification. To produce consistent high-quality
information an information product specification is needed. The specification states
clearly and definitely the requirements along with acceptable product variations.
8. Answer: A. Data standards quality, data definition quality and information architecture
quality. Data Definition Quality or Information Product Specification Quality is the
specification for building well designed information architecture like manufacturing a
Product.
9. Answer: A. The name of the data plainly conveys the meaning of the objects named.
10. Answer: D. To be consistent with the meaning of actual data values with its data
definition. Definition conformance is comprised of data definition quality, validity and
accuracy.
11. Answer: D. Precision. Precision is the characteristic of having the right level of
granularity in the data values. For example, measurement of uptime of a system(99.999
or 99.5 ) needs to allow for a finer breakdown of values instead of 99% available.
24. Answer: A. Retail Customer. Retail Customer is the most appropriate to demonstrate
Entity Type Name Clarity. Entity type name clarity is easily understood by the
knowledge worker and represents the objects. The characteristics of Entity Name Type
Clarity are: singular nouns, business terms and easily comprehended by the knowledge
worker.
25. Answer: A. Retail Customer Name. Retail Customer Name is the most appropriate to
demonstrate Attribute Name Clarity. Attribute Name Clarity is easily understood by the
knowledge worker and represents the facts. The characteristics of Attribute Name Clarity
are: business terms, easily associated with the entity type and easily comprehended by
the knowledge worker.
26. Answer: D. Customer. Domain type consistency (also known as class word) in an
attribute represents the type of data stored. For example, Start-Date attribute has a
domain type of date and the valid values would be a subset of all possible dates. Typical
domain types include: date, time, amount, identifier (id), amount, code, name, quantity,
percent, rate, and description.
27. Answer: B. Cross-ref. The most appropriate abbreviation for acronym clarity for the
term cross-reference is Cross-ref. When using abbreviations, they should be documented
in a single, enterprise-wide standards abbreviation list that is used consistently throughout
the enterprise. Rules of thumb for creating abbreviations are: use industry-standards or
universally accepted abbreviations where applicable, use short abbreviations without loss
of meaning, and always use the first letter of the term.
28. Answer: A. Attributes names are consistent where facts context are equivalent.
Names should be appropriate or consistent across the enterprise even across different
formats of presentation and storage formats.
29. Answer: A. Business Steward. A Business Steward in the enterprise is responsible
for keeping the enterprise wide glossary current in an organization. The responsibilities
include adding, changing and deleting the definitions as needed.
30. Answer: D. Subject Matter Expert. A Subject Matter Expert or Business Information
Steward in the enterprise is responsible for keeping the business term definitions current
in an organization. The responsibilities include adding, changing and deleting the
definitions as needed.
31. Answer: A. The business terms should be defined in the glossary that is enterprise
wide. The glossary can take on several forms.
32. Answer: A. Based on technical or existing system limitations. Business rules should
not be due to a technical or existing system limitation. Business rules should be
expressed in business terms, defines an aspect of the business to take action, complete
and specific, plus defines who, what, when, why and how and identities any exceptions.
Page 117 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
33. Answer: B. To measure the quality of data, either in physical form (file, database,
spreadsheet) or output from a process. The information quality assessment is conducted
for the benefit of the knowledge workers.
34. Answer: D. Determine which processes should be retired. The purpose of
information quality assessments is evaluating the processes and data, certify the data, and
providing feedback plus measuring against the baseline to calculate the costs of
nonquality.
35. Answer: A. Linear Regression. Linear Regression is not a valid sampling technique
when conducting an Information Quality assessment.
36. Answer: D. Usability. Usability is not a concern of the data assessment test. Data
assessment tests measure: Validity of business rule conformance, Timeliness,
Nondupliction, Accuracy to surrogate source including derivation integrity, and
Consistency of data.
37. Answer: D. Business Glossary List. The Information Quality Report deduces and
reports on the data assessment using Pareto Diagrams, Bar Chart, Statistical Control
Charts and outputs from Information quality analysis software.
38. Answer: D. Data residing in a single sharable database. The cost of redundancy is
part of the cost formula of information quality in the value basis component. Redundant
costs occur when the data is contained in multiple databases. There is a cost to capture
and control all the multiple databases and of the inconsistent or inaccurate data to the
organization. Data residing in a database has only potential value. Information value
occurs through usage only.
39. Answer: B. Cost to define the interface to acquire customer data. The cost of data is
comprised of two areas: cost basis and value basis. Cost basis is the cost of developing
and maintaining infrastructure. It is made up of the cost to define information
requirements, develop information, application, and technology architectures; and to
design and build applications and databases. Value basis uses the information to add
value for the enterprise.
40. Answer: B. Knowledge workers re-verifying the data. Process failure information
costs result in spent costs, liability and exposure costs, and recovery costs.
41. Answer: B. Information enables the enterprise to accomplish its mission and goals.
The enterprise business performance objectives need to be aligned to measuring
information quality.
42. Answer: C. Fix the process that produces the defective data by identifying root cause.
Data cleansing fixes the data. Information process quality improvements fix the process
that produces the defective data. The process is typically iterative that involves the
Page 118 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
cycles of: planning, implementing, assessing and rollout (Shewhart cycle of Plan-DoCheck-Act or PDCA).
43. Answer: A. Data Reengineering and cleansing process. The process that takes
existing data, which is defective and brings the data to suitable levels of quality, is known
as data reengineering and cleansing. Data reengineering is similar to reverse engineering
but only looks at the data not the application or system. It mainly focuses on how the
data is used in an organization and can work backwards to the data models. Data
reengineering often results in a deeper understanding of data assets of the enterprise and
may lead to areas like: data consolidation, data architecture, and data acquisition
strategies. Data Cleansing is the act of identifying and correcting data. Correcting data
involves cleaning up data that is incorrect, out-of-date, redundant, incomplete, or
formatted incorrectly.
44. Answer: D. Presenting data. Data Architects will consider source data cleansing, data
conversions, and data scrubbing when embarking on an Information Product
Improvement project. These are known as the three data cleansing areas: Source Data
Cleansing, Data Conversions and Data Scrubbing.
45. Answer: A. Source Data Cleansing. Source data cleansing improves existing data
quality where the data is initially stored.
46. Answer: B. Data Conversion. Data conversion is the act of mapping the source to
target and improving the quality of data by correcting, standardizing, de-duplicating,
completing and formatting.
47. Answer: C. Data Scrubbing. Data scrubbing is an act of defining summary and
derived data, adding data from external sources, consolidating data.
48. Answer: D. Information Steward. Information Steward should conduct quality audit
and control of data movement procedures. They should get input from internal audit,
knowledge workers and data conversion specialists.
49. Answer: D. Incorrect decisions made on poor quality information. There are three
categories of information quality costs: Nonquality information costs (process, rework,
lost and missed opportunity costs); Information quality assessments/audits; and
Information quality process improvements and prevention.
50. Answer: C. Analyze root cause and eliminate the cause. Information Process Quality
Improvement has two elemental processes: reactive and proactive. Proactive process
involves conducting root cause analysis. Root Cause Analysis identifies not only what
and how an event occurred, but also why it happened. Only when an analysis of why an
event or failure occurred will corrective measures are found.
51. Answer: D: Value chain relationship diagram. Value chain relationship diagram is
not a type of root cause analysis technique. Cause-and-effect diagram (Ishikawa or
Page 119 of 122
Copyright 2006 by DAMA International & DAMA International Foundation. All rights reserved.
fishbone diagram) breaks down causes into detailed categories so they can be organized
into related factors to identify root cause. Interrelationship diagram quantifies the
relationships between factors and classifies causes leading to root cause. Current reality
tree classifies interdependent relationships between effects leading to the determination
of root cause.
52. Answer: D. Refine. The Shewhart cycle of Plan-Do-Check-Act or PDCA because
known as the Deming cycle, is the foundation to improve information process quality.
53. Answer: A. Training. Training is most essential when implementing an Information
Quality Culture Transformation for both management and staff. The training needs to
cover why quality is fundamental to the enterprise and how to achieve quality. When
defining training requirements, identify the role and their responsibilities toward
information quality for example: general information, policies and processes, usage, and
information management principles. Next define their training requirements and learning
objectives for each role.
54. Answer: B. Managerial Information Steward. The Managerial Information Steward is
accountable for the integrity of the processes and quality of information. Knowledge
steward is accountable for the use of information. Process Steward is accountable for the
definition of a business process. Business Information is accountable for validating the
definition of data.
55. Answer: A. Conduct an Information Quality Management Maturity Assessment and
gap analysis. When embarking on an information quality program, the first step an
enterprise should take is to conduct an Information Quality Management Maturity
Assessment and gap analysis to determine the current state of the organization and where
they would like to be in the future.
56. Answer: A. Uncertainty. The Uncertainty state is Stage 1 and the least mature in the
Information Quality Management Maturity Assessment. In the Uncertainty stage,
Information quality is not considered a management tool. When issues occur they are
dealt with in a reactive manner.
57. Answer: B. Awakening. The Awakening state is Stage 2 and is characterized by
knowing a data quality problem exists but not knowing what to do about it in the
Information Quality Management Maturity Assessment. In the Awakening stage,
Information quality issues have been identified but management does not commit to their
resolution. When issues occur they are cleaned up rather than fixed at the source and
dealt with in a tight scope.
58. Answer: C. Enlightenment. The Enlightenment state is Stage 3 and is characterized
by adopting a commitment to quality and implementing the 14-point program in the
Information Quality Management Maturity Assessment. In the Enlightenment stage,
Quality Improvement Program is implemented with communication and resolution.
Selected Bibliography
Brackett, Michael, DATA SHARING: USING A COMMON DATA ARCHITECTURE,
John Wiley, 1994, ISBN 04711309931.
DAMA International & DAMA Chicago Standards Committee, DATA
MANAGEMENT ASSOCIATION: GUIDELINES TO IMPLEMENTING DATA
RESOURCE MANAGEMENT, DAMA International, 4th edition, 2002.
English, Larry P., IMPROVING DATA WAREHOUSE AND BUSINESS
INFORMATION QUALITY, John Wiley & Sons, 1999, ISBN: 0471253839.
Fowler, Martin, UML DISTILLED: APPLYING THE STANDARD OBJECT
MODELING LANGUAGE, Addison-Wesley, 1997, ISBN: 0-201-32563-2.
Inmon, W.H. BUILDING THE DATA WAREHOUSE, John Wiley, 2002, ISBN 0-471081302.
Kimball, Ralph, THE DATA WAREHOUSE TOOLKIT, John Wiley, 1996, ISBN: 0471-15337-0.
Marco, David, BUILDING AND MANAGING THE META DATA REPOSITORY: A
FULL LIFECYCLE GUIDE, John Wiley, 2002, ISBN: 0471355232.
McFadden, Fred R., Hoffer, Jeffrey A., and Prescott, Mary B. MODERN DATABASE
MANAGEMENT, Fifth Edition, Addison-Wesley, 1999, ISBN 0-8053-6054-9.
Simsion, Graeme and Witt, Graham, DATA MODELING ESSENTIALS, Third Edition,
Morgan Kaufman, 2004, ISBN: 0126445516.
Tannenbaum, Adrienne, METADATA SOLUTIONS: USING METAMODELS,
REPOSITORIES, XML, AND ENTERPRISE PORTALS TO ACHIEVE
INFORMATION ON DEMAND, Addison-Wesley, 2001, ISBN: 0201719762.
Watson, Richard T. DATA MANAGEMENT: DATABASES AND ORGANIZATIONS,
John Wiley & Sons, 2002, ISBN 0-471-41845-5.