Skip to content
Browse
BABOK Guide
BABOK Guide
10. Techniques
Introduction 10.1 Acceptance and Evaluation Criteria 10.2 Backlog Management 10.3 Balanced Scorecard 10.4 Benchmarking and Market Analysis 10.5 Brainstorming 10.6 Business Capability Analysis 10.7 Business Cases 10.8 Business Model Canvas 10.9 Business Rules Analysis 10.10 Collaborative Games 10.11 Concept Modelling 10.12 Data Dictionary 10.13 Data Flow Diagrams 10.14 Data Mining 10.15 Data Modelling 10.16 Decision Analysis 10.17 Decision Modelling 10.18 Document Analysis 10.19 Estimation 10.20 Financial Analysis 10.21 Focus Groups 10.22 Functional Decomposition 10.23 Glossary 10.24 Interface Analysis 10.25 Interviews 10.26 Item Tracking 10.27 Lessons Learned 10.28 Metrics and Key Performance Indicators (KPIs) 10.29 Mind Mapping 10.30 Non-Functional Requirements Analysis 10.31 Observation 10.32 Organizational Modelling 10.33 Prioritization 10.34 Process Analysis 10.35 Process Modelling 10.36 Prototyping 10.37 Reviews 10.38 Risk Analysis and Management 10.39 Roles and Permissions Matrix 10.40 Root Cause Analysis 10.41 Scope Modelling 10.42 Sequence Diagrams 10.43 Stakeholder List, Map, or Personas 10.44 State Modelling 10.45 Survey or Questionnaire 10.46 SWOT Analysis 10.47 Use Cases and Scenarios 10.48 User Stories 10.49 Vendor Assessment 10.50 Workshops

3. Techniques

3.6 Data Mapping

Guide to Business Data Analytics

.1    Purpose

Data mapping is used to consolidate data from one or more sources to a destination to create a meaningful set of data with a standardized format. This ensures that data can be accessed and used for business reporting and analysis.
.2 Description

Organizations collect a lot of data from many different data points, stored in varied formats. During data sourcing, there is often a need to migrate one or more of these data sources into a single data repository where all data has a consistent format. This allows for easy access to the data and reporting across the organization. Data mapping establishes a relationship between the source and the target repository, allowing data migration to happen, despite potentially disparate data formats.

Data mapping is used to support:

  • data migration, where source data is moved to a new data repository, and
  • data integration, where source data is merged with an existing data repository.

Note: For data integration to be possible, at a minimum, the data models of both source and target should be the same, even if the data schemas are different.

Data mapping happens at the attribute-level. During data migration and integration, attention is given to:

  • the selected attributes that will be migrated from source to target.
  • creation of new attributes in the target repository to allow migration from the source.
  • data type and size of the target attribute that is defined to ensure it can accommodate the source attribute data. In some instances, it may be determined the data size of the target attribute should be the same as the source data size, while in other instances the data size for the target attribute may need to be larger so it can accommodate future expansion. However, the target data size should never be smaller than the source data size as that may result in missing data.
  • new custom-defined attributes that involve some form of manipulation or calculation (for example, tax amount, product name based on product code).
Data integration also requires the analysis of the common attributes between source and target repositories and the nature of their relationship (for example, one-to-one and zero-to-many).

The complexity of data mapping may depend on several factors, including the degree of disparity between the source and target repositories, and any existing data hierarchy.
3.6.1 Elements

.1    Source

The repository providing the original data is referred to as the source. When analyzing the source, analysts consider the:
  • format of the source (for example, delimited file, spreadsheet, database entity),
  • attributes of interest or potential interest, and
  • data type and data size of the attributes.
.2    Target

The repository receiving the data is referred to as the target. When analyzing the target, analysts consider the:
  • format of the source (for example, delimited file, spreadsheet, database entity),
  • new attributes that need to be created within the target to align with the source, with similar data type and data sizes,
  • source attributes that need to be transformed in the target and their corresponding transformation rule. These could be based on a business rule or a business requirement (for example, the "TransactionDatetime" data attribute in the source has a datetime format of mm/dd/yyyy hh:mm:tt but in the target, only the date should be brought in for easy querying purposes. So, the TransactionDate attribute in the target will have the date format mm/dd/yyyy), and
  • creation of new custom fields (for example, adding a calculation, combining attributes, or formatting content based on a logic).
.3    Data Mapping Specification

Data mapping is often completed using a spreadsheet where target attributes are mapped to source attributes with applicable transformation logic. The following is a basic data map.

Target
Source Transformation Rules
Entity Name Attribute Name Data Type Attribute(s) Direct Map? Logic
T_Employe e Emp_ID Varchar (10) Employee.ID Y  
T_Employe e Emp_Name Varchar (50) Employee.FirstName, Employee.LastName N Concatenate (Employee.FirstName, “ “,EmployeeLastName)
T_Employe e Emp_DepNo Number (2) Employee.Departnment Number Y  
T_Employe e Emp_StartDate Date Employee.StartDate Y Only bring the date component with format YYYY-MM-DD
3.6.2 Usage Considerations

.1    Strengths

  • Provides a meticulous approach to ensuring smooth migration and integration of data between varied data platforms.
  • Provides data traceability, which is particularly beneficial when resolving data issues.
  • Easy to learn and use.
  • Enables creation of a standardized, business-focused data repository, which in turn allows for consistency and ease in data access and reporting.
  • Helps identify any data quality issues within the source or the target.
.2    Limitations

  • Requires careful attention to detail to avoid encountering data loss, redundancies, or other unexpected errors.
  • Data mapping work, particularly when done manually, can be time- consuming.
  • Needs updating as soon as changes are made in either source or target