What statement is most accurate about master data metadata?
Includes a sample of content
Does little to improve fit-for-purpose choices on when and where to apply the ' data
Secures the content
Provides the who. what, and where context about master data content
Can either be related to technical or business perspectives of content, but not
Master data metadata provides crucial information about the master data, offering context and supporting its management and use within the organization.
Who, What, and Where Context:
Metadata provides descriptive information about the master data, including details about who created or modified the data, what the data represents, and where it is used.
This contextual information is essential for understanding the origins, purpose, and usage of the master data.
Includes a Sample of Content:
While metadata might include examples or samples of the data, this is not its primary purpose.
Improving Fit-for-Purpose Choices:
Metadata helps improve the application and governance of master data by providing context and supporting data management decisions.
Securing the Content:
Metadata itself is not primarily focused on security, though it can support data governance and access control processes.
Technical or Business Perspectives:
Metadata can encompass both technical and business perspectives, providing a holistic view of the data's context and usage.
What characteristics does Reference data have that distinguish it from Master Data?
It is more volatile and needs to be highly structured
It is always data from an outside source such as a governing body
It always has foreign database keys to link it to other data
It is less volatile, less complex, and typically smaller than Master Data sets
It provides data for transactions
Reference data and master data are distinct in several key characteristics. Here’s a detailed explanation:
Reference Data Characteristics:
Stability: Reference data is generally less volatile and changes less frequently compared to master data.
Complexity: It is less complex, often consisting of simple lists or codes (e.g., country codes, currency codes).
Size: Reference data sets are typically smaller in size than master data sets.
Master Data Characteristics:
Volatility: Master data can be more volatile, with frequent updates (e.g., customer addresses, product details).
Complexity: More complex structures and relationships, involving multiple attributes and entities.
Size: Larger in size due to the detailed information and numerous entities it encompasses.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Which of the following is true about MDM?
Manages master data formally with a high degree of diligence and collaboration
Master data is not managed without a formal MDM program
Once master data is published by a MDM hub. it no longer is considered master ' data
MDM programs have a definitive life span
A MDMprogram must include a MDM software application
MDM (Master Data Management) is characterized by formal management with a high degree of diligence and collaboration. Here’s why:
Formal Management:
Structured Processes: MDM involves structured processes for managing master data, including data governance, data quality management, and data stewardship.
Policies and Standards: Establishes and enforces policies and standards to ensure data consistency, accuracy, and integrity.
Collaboration:
Cross-Functional Teams: Requires collaboration across different departments, including IT, business units, and data governance teams.
Stakeholder Involvement: Engages various stakeholders in the data management process, ensuring that master data meets the needs of the entire organization.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Which of the following Is a characteristic of a probabilistic matching algorithm?
A score is assigned based on weight and degree of match
Each variable to be matched is assigned a weight based on its discriminating power
Individual attribute matching scores arc used to create a match probability percentage.
All answers are correct
Following the matching process there are typically records requiring manual review and decisioning.
Probabilistic matching algorithms assign a score based on the weight and degree of match, assign weights to variables based on their discriminating power, and use individual attribute matching scores to create a match probability percentage. Additionally, after the matching process, some records typically require manual review and decisioning to ensure accuracy. Therefore, all provided characteristics describe the nature of probabilistic matching algorithms accurately.
References:
DAMA-DMBOK: Data Management Body of Knowledge (2nd Edition), Chapter 11: Reference and Master Data Management.
"Master Data Management and Data Governance" by Alex Berson and Larry Dubov
What is a registry as it applies to Master Data?
An index that points to Master Data in the various systems of record
Any data available during record creation
Reconciled versions of an organization's systems
A starling point for matching and linking new records
A system to identify how data is used for transactions and analytics
A registry in the context of Master Data Management (MDM) is a centralized index that maintains pointers to master data located in various systems of record. This type of architecture is commonly referred to as a "registry" model and allows organizations to create a unified view of their master data without consolidating the actual data into a single repository. The registry acts as a directory, providing metadata and linkage information to the actual data sources.
References:
DAMA-DMBOK2 Guide: Chapter 10 – Master and Reference Data Management
"Master Data Management: Creating a Single Source of Truth" by David Loshin
A catalog where products are organized by category is an example of?
A meronomy
A marketing mix
A taxonomy
A metadata repository
A catalog where products are organized by category is an example of a taxonomy. Here’s why:
Definition of Taxonomy:
Classification System: Taxonomy refers to the practice and science of classification. It involves organizing items into hierarchical categories based on their relationships and similarities.
Example: In the context of a product catalog, taxonomy is used to classify products into categories and subcategories, making it easier to browse and find specific items.
Application in Product Catalogs:
Categorization: Products are grouped into logical categories (e.g., Electronics, Clothing, Home Appliances) and subcategories (e.g., Smartphones, Laptops, Televisions).
Navigation and Search: Helps users navigate the catalog efficiently and find products quickly by narrowing down categories.
References:
Data Management Body of Knowledge (DMBOK), Chapter 9: Data Architecture
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
The Data Architecture design of an MDM solution must resolve where to leverage what type of relationships?
Traceable relationships and/or lineage relationships
Data Acquisition relationships
Affiliation relationships and/or parent-child relationships
Hub and spoke relationships
Ontology relationships and/or epistemologyrelationships
Data Architecture in MDM Solutions:The design of a Master Data Management (MDM) solution involves defining and managing relationships between data entities.
Types of Relationships:
Traceable relationships and/or lineage relationships:These are important for understanding data provenance and transformations but are more relevant to data governance and data lineage tracking.
Data Acquisition relationships:These pertain to how data is sourced and collected, rather than how master data entities are related.
Affiliation relationships and/or parent-child relationships:These are crucial in MDM as they define how entities are related in hierarchical and associative contexts, such as customer relationships, organizational hierarchies, and product categorizations.
Hub and spoke relationships:This refers to the architecture model for MDM systems rather than the type of data relationship.
Ontology relationships and/or epistemology relationships:These are more abstract and pertain to the nature and categorization of knowledge, not specifically to the functional relationships in MDM.
Conclusion:The correct answer is "Affiliation relationships and/or parent-child relationships" as these are essential for defining and managing master data relationships in an MDM solution.
References:
DMBOK Guide, sections on Data Architecture and Master Data Management.
CDMP Examination Study Materials.
The format and allowable ranges of Master Data values are dictated by:
Business rules
Semantic rules
Processing rules
Engagement rules
Database limitations
The format and allowable ranges of Master Data values are primarily dictated by business rules.
Business Rules:
Business rules define the constraints, formats, and permissible values for master data based on the organization’s operational and regulatory requirements.
These rules ensure that data conforms to the standards and requirements necessary for effective business operations.
Semantic Rules:
These rules pertain to the meaning and context of the data but do not directly dictate formats and ranges.
Processing Rules:
These rules focus on how data is processed but not on the allowable values or formats.
Engagement Rules:
These rules govern interactions and workflows rather than data formats and ranges.
Database Limitations:
While database limitations can impose constraints, they are typically secondary to the business rules that drive data requirements.
Which of the following isNOT part of MDM Lifecycle Management?
Establishing recovery and backup rules
Reconciling and consolidating data
Identifying multiple instances of the same entity
Identifying improperly matched or merged instances of data
Maintaining cross-references to enable information integration
Master Data Management (MDM) lifecycle management encompasses the processes and practices involved in managing master data throughout its lifecycle, from creation to retirement. It ensures that master data remains accurate, consistent, and usable.
Reconciling and Consolidating Data:
This process involves merging data from multiple sources to create a single, unified view of each master data entity.
It ensures that duplicate records are identified and consolidated, maintaining data consistency.
Identifying Multiple Instances of the Same Entity:
This involves detecting and resolving duplicate records to ensure that each master data entity is uniquely represented.
Tools and algorithms are used to identify potential duplicates based on matching criteria.
Identifying Improperly Matched or Merged Instances of Data:
This step involves reviewing and correcting any errors that occurred during the matching or merging process.
Ensures that data integrity is maintained and that merged records accurately represent the underlying entities.
Maintaining Cross-References to Enable Information Integration:
Cross-references link related data entities across different systems, enabling seamless information integration.
This ensures that data can be consistently accessed and used across the organization.
Establishing Recovery and Backup Rules (NOT part of MDM Lifecycle Management):
While important for overall data management, recovery and backup rules pertain more to data protection and disaster recovery rather than the specific processes of MDM lifecycle management.
Key processing steps for successful MDM include the following steps with the exception of which processing step?
Data Indexing
Data Acquisition
Data Sharing & Stewardship
Entity Resolution
Data Model Management
Key processing steps for successful MDM typically include:
Data Acquisition: The process of gathering and importing data from various sources.
Data Sharing & Stewardship: Involves ensuring data is shared appropriately across the organization and that data stewards manage data quality and integrity.
Entity Resolution: Identifying and linking data records that refer to the same entity across different data sources.
Data Model Management: Creating and maintaining data models that define how data is structured and related within the MDM system.
Excluded Step - Data Indexing: While indexing is a critical database performance optimization technique, it is not a primary processing step specific to MDM. MDM focuses on consolidating, managing, and ensuring the quality of master data rather than indexing, which is more about search optimization within databases.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Business entities are represented by entity instances:
In the form technical capabilities
In the form of business capabilities
In the form of files
in the form of data/records
In the form of domains
Business entities are represented within an organization through various forms, primarily as data or records within information systems.
Technical Capabilities:
While technical capabilities support the management and usage of business entities, they are not the representation of the entities themselves.
Business Capabilities:
Business capabilities describe the functions and processes that an organization can perform, but they do not represent individual business entities.
Files:
Files can contain data or records, but they are not the direct representation of business entities.
Data/Records:
Business entities are captured and managed as data or records within databases and information systems.
These records contain the attributes and details necessary to uniquely identify and describe each business entity.
Domains:
Domains refer to specific areas of knowledge or activity but are not the direct representation of business entities.
Management of Reference and Master data is aimed to reduce cost and risk of having disparate data mainly caused by:
Organicgrowth of systems and data, isolated systems, mergers and acquisitions
High number of legacy applications and lack of expertise to evolve or replace them
Lack of appropriate processes to assure data availability and accuracy
Migration to new technology platforms and evolution of landscape
Poor or non-existent data documentation available for developers and business analysts
Management of Reference and Master Data aims to mitigate the challenges of disparate data, which typically arise from:
Organic Growth:
Unplanned Expansion: Over time, organizations often develop new systems and applications organically, leading to isolated and redundant data stores.
Inconsistent Data: These disparate systems often result in inconsistent and unreliable data.
Isolated Systems:
Siloed Applications: Independent systems that do not communicate effectively with each other can lead to multiple versions of the same data.
Lack of Integration: Without proper integration, data consistency and quality suffer.
Mergers and Acquisitions:
Combining Systems: Mergers and acquisitions introduce the challenge of integrating different data systems and standards.
Data Redundancy: Newly acquired systems often come with their own data sets, leading to redundancy and conflicts.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
One of the main guiding principles for Reference and Master Data is the one related to ownership, which states that:
Reference Data ownership belongs to IT while Master Data ownership belongs to the Business
Reference and Master Data ownership is usually owned by a specific department
Reference and Master Data typically belong to the organization, not to a particular application or department
Reference and Master Data cannot include purchased data
Reference and Master Data ownership falls into Data Governance Office
Ownership is a crucial principle in managing Reference and Master Data. Here’s an in-depth look at why:
Organizational Ownership:
Unified Responsibility: Reference and Master Data are assets that span across various functions and departments within an organization.
Consistency and Accuracy: Ensuring that data ownership is attributed to the organization prevents silos and ensures data is consistently accurate and available across all departments.
Data Governance: Proper governance frameworks ensure that data is managed in a way that meets the organization’s needs and complies with relevant regulations and standards.
Avoiding Departmental Silos:
Cross-functional Use: Different departments use and rely on Reference and Master Data, so ownership by a single department can lead to conflicts and inconsistencies.
Holistic Management: Centralized ownership enables holistic data management practices, enhancing data quality and usability across the organization.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Should both in-house and commercial tools meet ISO standards for metadata?
Yes. at the very least they should provide guidance
No. each organization needs to develop their own standards based on needs
Adhering to ISO standards for metadata is important for both in-house and commercial tools for the following reasons:
Standardization:
Uniformity: ISO standards ensure that metadata is uniformly described and managed across different tools and systems.
Interoperability: Facilitates interoperability between different tools and systems, enabling seamless data exchange and integration.
Guidance and Best Practices:
Structured Approach: Provides a structured approach for defining and managing metadata, ensuring consistency and reliability.
Compliance and Quality: Ensures compliance with internationally recognized best practices, enhancing data quality and governance.
References:
ISO/IEC 11179: Information technology - Metadata registries (MDR)
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
The MDM process step responsible for determining whether two references to real world objects refer to the same object or different objects is known as:
Data Model Management
Data Acquisition
Entity Resolution
Data Sharing & Stewardship
Data Validation. Standardization, and Enrichment
Entity resolution is a critical step in the MDM process that identifies whether different data records refer to the same real-world entity. This ensures that each entity is uniquely represented within the master data repository.
Data Model Management:
Focuses on defining and maintaining data models that describe the structure, relationships, and constraints of the data.
Data Acquisition:
Involves gathering and bringing data into the MDM system but does not deal with resolving entities.
Entity Resolution:
This process involves matching and linking records from different sources that refer to the same entity. Techniques such as deterministic matching (based on exact matches) and probabilistic matching (based on similarity scores) are used.
Entity resolution helps in deduplication and ensuring a single, unified view of each entity within the MDM system.
Data Sharing & Stewardship:
Focuses on managing data access and ensuring that data is shared responsibly and accurately.
Data Validation, Standardization, and Enrichment:
Ensures data quality by validating, standardizing, and enriching data but does not directly address entity resolution.
Master Data Management resolves uncertainty by clearly stating that;
To have master data you must focus resources properly
Some entities [master entities) are more important than others
Only those entities in the Enterprise Data Model are considered Master Data.
All entities arc equal across an enterprise and need to be managed
Data elements must be stored in a repository before they are considered master data
Master Data Management (MDM) aims to establish a single, reliable source of key business data (master data). The correct answer here is B, which states that "Some entities [master entities) are more important than others."
Definition of Master Data:Master data refers to the critical data that is essential for operations in a business, such as customer, product, and supplier information.
Significance in MDM:MDM focuses on identifying and managing these key entities because they are vital for business processes and decision-making. This is why these entities are considered more important than others.
Resolution of Uncertainty:By emphasizing the importance of master entities, MDM reduces ambiguity around which data should be prioritized and managed meticulously, ensuring consistency and accuracy across the enterprise.
References:
DAMA-DMBOK: Data Management Body of Knowledge, 2nd Edition.
CDMP Study Guide
For MDMs. what is meant by a classification scheme?
Codes that represent a controlled set of values
A vocabulary view covering a limited range of topics
Descriptive language used to control objects
A way of classifying unstructured data
In Master Data Management (MDM), a classification scheme refers to a structured way of organizing data by using codes that represent a controlled set of values. These codes help in categorizing and standardizing data, making it easier to manage, search, and analyze.
References:
DAMA-DMBOK: Data Management Body of Knowledge (2nd Edition), Chapter 11: Reference and Master Data Management.
"Master Data Management and Data Governance" by Alex Berson and Larry Dubov.
The biggest challenge to implementing Master Data Management will be:
The inability to get the DBAs to provide their table structures
Defining requirements for master data within an application
the disparity between sources
Complex queries
Indexes and foreign keys
Implementing Master Data Management (MDM) involves several challenges, but the disparity between data sources is often the most significant.
Disparity Between Sources:
Different systems and applications often store data in varied formats, structures, and standards, leading to inconsistencies and conflicts.
Data integration from disparate sources requires extensive data cleansing, normalization, and harmonization to create a single, unified view of master data entities.
Data Quality Issues:
Variability in data quality across sources can further complicate the integration process. Inconsistent or inaccurate data must be identified and corrected.
Defining Requirements for Master Data:
While defining requirements is crucial, it is typically a manageable step through collaboration with business and technical stakeholders.
DBA Cooperation:
Getting Database Administrators (DBAs) to share table structures can pose challenges, but it is not as critical as dealing with disparate data sources.
Complex Queries and Indexes:
While important for performance optimization, complex queries and indexing issues are more technical hurdles that can be resolved with appropriate database management practices.
Managing Master Data involves:
Managing transaction data
Managing process models
Managing database keys
Managing structured and unstructured data
Managing security risks
Managing Master Data involves several key activities, primarily focusing on:
Structured and Unstructured Data:
Structured Data: Managing well-defined data types, such as relational databases, where data is organized into tables and fields.
Unstructured Data: Handling data that does not have a predefined format or structure, such as emails, documents, and multimedia files.
Comprehensive Management:
Data Integration: Ensuring that data from various sources, both structured and unstructured, is integrated into the master data repository.
Data Quality: Implementing processes and tools to maintain high data quality for both structured and unstructured data.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
All organizations have master data even if it is not labelled Master Data.
True
False
All organizations possess master data, even if it is not explicitly labeled as such. Here’s why:
Definition of Master Data:
Core Business Entities: Master data refers to the critical entities around which business transactions are conducted, such as customers, products, suppliers, and accounts.
Business Operations: Every organization maintains records of these entities to support business operations, decision-making, and reporting.
Implicit Existence:
Unlabeled Data: Organizations may not explicitly label this data as “Master Data,” but it exists within various systems, databases, and spreadsheets.
Examples: Customer lists, product catalogs, employee records, and financial accounts.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Is there a standard tor defining and exchanging Master Data?
Yes, ISO 22745
No. every corporation uses their own method
Yes. it is called ETL
No. there are no standards because not everyone uses Master Data
ISO 22745 is an international standard for defining and exchanging master data.
ISO 22745:
This standard specifies the requirements for the exchange of master data, particularly in industrial and manufacturing contexts.
It includes guidelines for the structured exchange of information, ensuring that data can be shared and understood across different systems and organizations.
Standards for Master Data:
Standards like ISO 22745 help ensure consistency, interoperability, and data quality across different platforms and entities.
They provide a common framework for defining and exchanging master data, facilitating smoother data integration and management processes.
Other Options:
ETL:Refers to the process of Extract, Transform, Load, used in data integration but not a standard for defining master data.
Corporation-specific Methods:Many organizations may have their own methods, but standardized frameworks like ISO 22745 provide a common foundation.
No Standards:While not all organizations use master data, standards do exist for those that do.
The 3 primary categories of components in a MDM framework are:
People, process, & technology
Structure, ETL, & storage
Integration, quality, & governance
Program, project, task
Matching, linking. & verification
The three primary categories of components in a Master Data Management (MDM) framework are people, process, and technology. Here’s a detailed breakdown:
People:
Roles and Responsibilities: Involves defining roles such as data stewards, data owners, and data governance committees who are responsible for managing and overseeing master data.
Skills and Training: Ensuring that the individuals involved have the necessary skills and training to manage master data effectively.
Process:
Data Governance: Establishing policies, procedures, and standards for managing master data to ensure its accuracy, consistency, and reliability.
Data Lifecycle Management: Processes for creating, maintaining, and retiring master data.
Technology:
MDM Tools and Platforms: Utilizing technology solutions to support the management of master data, including data integration, data quality, and data management platforms.
Infrastructure: Ensuring the necessary technical infrastructure is in place to support MDM activities.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Which of the following is NOT a Reference & Master Data activity?
Evaluate and Assess Data Sources
Manage the Lifecycle
Establish Governance Policies
Model Data
Define Architectural Approach
Activities related to Reference & Master Data typically include managing the lifecycle, establishing governance policies, modeling data, and defining architectural approaches. However, evaluating and assessing data sources is generally not considered a core activity specific to Reference & Master Data management. Here's a detailed explanation:
Core Activities:
Manage the Lifecycle: Involves overseeing the entire lifecycle of master data, from creation to retirement.
Establish Governance Policies: Setting up policies and procedures to govern the management and use of master data.
Model Data: Creating data models that define the structure and relationships of master data entities.
Define Architectural Approach: Developing the architecture that supports master data management, including integration and data quality frameworks.
Excluded Activity:
Evaluate and Assess Data Sources: While this is an important activity in data management, it is more relevant to data acquisition and integration rather than the ongoing management of reference and master data.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Depending on the granularity and complexity of what the Reference Data represents. it may be structured as a simple list, a cross-reference or a taxonomy.
True
False
Reference data can be structured in various ways depending on its granularity and complexity.
Simple List:
Reference data can be a simple list when it involves basic, discrete values such as country codes or product categories.
Cross-Reference:
When reference data needs to map values between different systems or standards, it can be structured as cross-references. For example, mapping old product codes to new ones.
Taxonomy:
For more complex hierarchical relationships, reference data can be structured as a taxonomy. This involves categorizing data into parent-child relationships, like an organizational hierarchy or biological classification.
What MDM style allows data to be authored anywhere?
Consolidation
Centralized style
Persistent
Registry style
Coexistence
Master Data Management (MDM) styles define how and where master data is managed within an organization. One of these styles is the "Coexistence" style, which allows data to be authored and maintained across different systems while ensuring consistency and synchronization.
Coexistence Style:
The coexistence style of MDM allows master data to be created and updated in multiple locations or systems within an organization.
It supports the integration and synchronization of data across these systems to maintain a single, consistent view of the data.
Key Features:
Data Authoring: Data can be authored and updated in various operational systems rather than being confined to a central hub.
Synchronization: Changes made in one system are synchronized across other systems to ensure data consistency and accuracy.
Flexibility: This style provides flexibility to organizations with complex and distributed IT environments, where different departments or units may use different systems.
Benefits:
Enhances data availability and accessibility across the organization.
Supports operational efficiency by allowing data updates to occur where the data is used.
Reduces the risk of data silos and inconsistencies by ensuring data synchronization.
Which of the following is NOT an example of Master Data?
A categorization of products
A list of account codes
Planned control activities
A list of country codes
Currency codes
Planned control activities are not considered master data. Here’s why:
Master Data Examples:
Categories and Lists: Master data typically includes lists and categorizations that are used repeatedly across multiple business processes and systems.
Examples: Product categories, account codes, country codes, and currency codes, which are relatively stable and broadly used.
Planned Control Activities:
Process-Specific: Planned control activities pertain to specific actions and checks within business processes, often linked to operational or transactional data.
Not Repeated Data: They are not reused or referenced as a stable entity across different systems.
References:
Data Management Body of Knowledge (DMBOK), Chapter 7: Master Data Management
DAMA International, "The DAMA Guide to the Data Management Body of Knowledge (DMBOK)"
Where is the most time/energy typically spent tor any MDM effort?
Subscribing content from the MDM environment
Designing the Enterprise Data Model
Vetting of business entities and data attributes by Data Governance process
Publishing content to the MDM environment
Securing funding for the MDM effort
In any Master Data Management (MDM) effort, the most time and energy are typically spent on vetting business entities and data attributes through the Data Governance process. This step ensures that the data is accurate, consistent, and adheres to defined standards and policies. Itinvolves significant collaboration and decision-making among stakeholders to validate and approve the data elements to be managed.
References:
DAMA-DMBOK: Data Management Body of Knowledge (2nd Edition), Chapter 11: Reference and Master Data Management.
"Master Data Management and Data Governance" by Alex Berson and Larry Dubov.
A key capability to quickly onboard new data suppliers and subscribers to a MDM solution is which of the following?
Data format and transfer flexibility
Source system conformance to a single standard data input format
Requiring only delta loads of changed data attributes
Encrypting all personal information
Subscriber conformance to a single standard data output format
Definitions and Context:
MDM Solution: This involves tools and processes to manage master data within an organization to ensure a single source of truth.
Onboarding Data Suppliers and Subscribers: This process involves integrating new data sources (suppliers) and distributing data to various applications or users (subscribers).
Explanation:
A key capability for onboarding is the flexibility in data format and transfer methods because different data suppliers may use various formats and protocols.
Ensuring flexibility allows the MDM system to easily adapt to different data sources and meet the needs of diverse data consumers, thereby facilitating quick and efficient onboarding.
References:
DAMA-DMBOK: Data Management Body of Knowledge, 2nd Edition, Chapter 11: Master and Reference Data Management.
The Open Group, "TOGAF Series Guide: The Data Management Capability Assessment Model (DCAM)".
What activity is helpful in mapping source system data for MDM efforts?
Data profiling
ETL toolset
Process modeling
Data dictionary
Data modeling
Data profiling is a crucial activity in mapping source system data for MDM efforts. Data profiling involves analyzing data from source systems to understand its structure, content, and quality. Key steps include:
Data Assessment: Evaluating the data to identify patterns, inconsistencies, and anomalies.
Data Quality Analysis: Measuring the quality of data in terms of accuracy, completeness, consistency, and uniqueness.
Metadata Extraction: Extracting metadata to understand data definitions, formats, and relationships.
Data Cleansing: Identifying and correcting data quality issues to ensure that the data is suitable for integration into the MDM system.
By performing data profiling, organizations can gain insights into the current state of their data, identify potential issues, and develop strategies for data integration and quality improvement.
References:
DAMA-DMBOK: Data Management Body of Knowledge, 2nd Edition.
"Data Quality: The Accuracy Dimension" by Jack E. Olson.
When 2 records are not matched when they should have been matched, this condition is referred to as:
False Positive
A True Positive
A False Negative
A True Negative
An anomaly
Definitions and Context:
False Positive: This occurs when a match is incorrectly identified, meaning records are deemed to match when they should not.
True Positive: This is a correct identification of a match, meaning records that should match are correctly identified as matching.
False Negative: This occurs when a match is not identified when it should have been, meaning records that should match are not matched.
True Negative: This is a correct identification of no match, meaning records that should not match are correctly identified as not matching.
Anomaly: This is a generic term that could refer to any deviation from the norm and does not specifically address the context of matching records.
Explanation:
The question asks about a scenario where two records should have matched but did not. This is the classic definition of aFalse Negative.
In data matching processes, this is a critical error because it means that the system failed to recognize a true match, which can lead to fragmented and inconsistent data.
References:
DAMA-DMBOK: Data Management Body of Knowledge, 2nd Edition, Chapter 11: Master and Reference Data Management.
ISO 8000-2:2012, Data Quality - Part 2: Vocabulary.
Copyright © 2014-2024 Certensure. All Rights Reserved