Beruflich Dokumente
Kultur Dokumente
27/08/2013
Confidentiality Statement
Include the confidentiality statement within the box provided. This has to be legally
approved
Confidentiality and Non-Disclosure Notice
The information contained in this document is confidential and proprietary to TATA
Consultancy Services. This information may not be disclosed, duplicated or used for any
other purposes. The information contained in this document may not be released in
whole or in part outside TCS for any purpose without the express written permission of
TATA Consultancy Services.
Abstract
SAP Business Objects Data Services (BODS) is a software tool designed by Business Object (a company that got
acquired by SAP in 2007). This tool pull data from any system/database/tables apply changes to modify the data
and load the data in to any other system/database. This process in known as Extraction, transformation and
Loading (ETL)
This training material will provide guidance to beginners to understand BODS architecture, components, objects,
transformations, mappings, Job execution, scheduling and monitoring of logs and error.
Table of Content
1.
2.
3.
4.
5.
3.2
BODS Objects................................................................................................................. 10
3.3
3.4
3.5
4.2
4.3
4.4
Figure 1: ETL
Local Repository - Repository is like a database that stores predefined system objects and user defined objects
including source/target metadata and transformation rules. This is a mandatory repository for BODS functioning.
Central Repository - A central repository is an optional component that can be used to support multi-user
development. The central repository provides a shared object library allowing developers to check objects in and
out of their local repositories.
Job Server and Engine Job Server starts the data movement engine processes to perform data extraction,
transformation and movement.
Access Server It facilitates real time job execution by passing messages between web applications and the
Data Services - Job Server and engines
Project - A Project is the highest-level object in Designer. A Project is a single-use objects that allows us to group
and organize Jobs in Designer. Only one project can be open and visible in the Project Area at a time.
Jobs - composed of work flows and/or data flows. A job is the smallest unit of work that can Schedule
independently for Execution. Jobs must be associated with project to display logs in Admin console. Also, Job
wont be displayed in the Job list of that repository in admin console if it is not associated with Project
Work Flows - is the collection of several Data flows into a sequence. A Work flow orders Data flows and the
operations that support them. It also defines the interdependencies between data flows. Work flows can be used
to define strategies for error handling or to define conditions for running the Data flows. A workflow is optional.
Data Flow - is the process by which source data is transformed into target data. It describes how to process a
task.
Transforms - are the in built transformation objects available in DS for transforming source data as per business
rules. The following is a list of available transforms. The transforms that you can use depend on the software
package that you have purchased. If a transform belongs to a package that you have not purchased, it is greyed
out and cannot be used in a job.
Transform Category
Transform
Description
Data Integrator
Data_Transfer
Allows a data flow to split its processing into two sub data
flows and push down resource consuming operations to
the database server.
Generates a column filled with date values based on the
start and end dates and increment that you provide.
Generates an additional "effective to" column based on the
primary keys "effective date."
Flattens hierarchical data into relational tables so that it can
participate in a star schema. Hierarchy flattening can be
both vertical and horizontal.
Converts rows flagged as UPDATE to UPDATE
plus INSERT, so that the original
Values are preserved in the target. You
specify in which column to look for updated
data
Generates new keys for source data, starting from a value
based on existing keys in the table you specify
Sorts input data, maps output data, and resolves
Before- and after-images for UPDATE rows.
Date_Generation
Effective_Date
Hierarchy_Flattening
History_Preserving
Key_Generation
Map_CDC_Operation
Data Quality
Pivot (Columns to
Rows)
Reverse Pivot (Rows to
Columns)
XML_Pipeline
Associate
Country ID
Data Cleanse
Match
Table_Comparison
USA Regulatory
Address Cleanse
User-Defined
Platform
Case
Map_Operation
Merge
Query
Script - A Script is a single-use object that is used to call functions and assign values in a workflow. To apply
decision-making and branch logic to work flows DI scripting language is used.
Once the Database type is selected, below screen appears. Provide the credentials.
Click on OK to create Datastore. Create the datastore for Source and Target Database.
To import the table, right click on Datastore name -> Import by Name.
BODS, Batch job and Real Time Job both can be created. Batch jobs are those that run in
batches at a predefined time and after a predefined time period (frequency). Any batch job in
BODS basically contains one or more data flow or workflow. A workflow can contain one or
more data flows.
Data flow is a single logical unit where the whole logic to transport data from one schema to
other, is specified. A data-flow, being a logical unit, cannot execute on its own. It must be
encapsulated inside a batch job in order to execute it. Data flows can also be grouped under
one or different workflows and those workflows can, in turn, be executed through the batch job.
Below are the steps to show a simple mapping to extract data from SAP table and load into target oracle database.
Steps 1 - First, import Source and Target Tables in Data Services .For ex: source Table is DD07T and target table is ZDD07T (Target
Table should be created in Target Database before importing)
Go to Data Store > Expand the Source Data store Name -> Right Click on Tables -> Import by Name .
Provide the Table Name and click on Import. Instance of Table is now available in Data Services for pulling the data. Similarly, click on
Target Data Store and import the target table.
Double click on Data Flow. Select ABAP Data flow .Below screen appears.
Click on Data Store tab and drag the source table DD07T
Drag the Query transforms from the tool Palette .Connect the Source table with Query Transform.
Double click on Query Transform. Drag the fields from Schema In to Schema Out.
Change the name of column in Schema out according to the description of the column in Source table, if required. Here, Column names
should be same as the columns in Target Table.
Drag the Data Transport transform and join it with Query transform.
Double Click on Data Transport transform and provide the name of .dat file
Now go to Data Flow, add Query transform to add audit information like LOAD_DATE. Double click on Query transform and map all
columns from Schema in to Schema out.
Go to Datastores tab and click on Target database datastore and import target table.
Now Go to Job Tab and add the job to the Project Area.
Now For Execution, Double click the job and click on Execute and OK.
Trace, Monitor,
Click on Administrator
Click on status and Select the repository where Job was created.
Select the Job which is to be scheduled and Click on Add the Schedule
Provide the
Schedule Name
Select the
Time for
Go to Repository Schedules, Select the Schedule and Press Activate. Schedule will execute only when it is in Active State.
Click on Execute.
5. BODS Benefits
Below are the benefits that Data Services Provide:
Single Infrastructure for the data movement to enable faster and lower cost implementation
Integrate data across many systems and reuse that data for many purposes
Implement Pre-packaged data solutions for fast deployment and quick ROI.
Customizes and manages data access and uniquely combines industry-leading
technologies for delivering data to analytic, supply-chain management, customer
relationship management, and Web applications.
Contact
For more information, contact gsl.cdsfiodg@tcs.com (Email Id of ISU)
IT Services
Business Solutions
Consulting
All content / information present here is the exclusive property of Tata Consultancy Services Limited (TCS). The content /
information contained here is correct at the time of publishing. No material from here may be copied, modified, reproduced,
republished, uploaded, transmitted, posted or distributed in any form without prior written permission from TCS.
Unauthorized use of the content / information appearing here may violate copyright, trademark and other applicable laws,
and could result in criminal or civil penalties. Copyright 2011 Tata Consultancy Services Limited