Beruflich Dokumente
Kultur Dokumente
Ans. Normalization is used to transform non tabular data structure into tabular format mainly required for
COBOL data files.
Ans. Export your mapping to XML file and then Import it to Client machine. I am assuming here the client
and your system as same version and environment for Informatica.
5. When i was running workflow manager i.e, when i right click the task icon and
clicking start workflow from task, i am getting a message like 'cannot connect to the
server'.
Can any body guide me, how to get the output.
Ans. First check out whether the Informatica service is running or not. Sometimes the server address will
change everytime.so it is suggestible to resolve the server (server-server configuration) and then start the
service. Also check out the repository server configuration...
6. In Source 1 to 100 numbers? How can we load the even number in One target
and odd number in another target? explain the mapping? How can we draw and
which transformations we r using? and explain logic?
Ans. Use a Router after the source , if the port have the number 1 to 1000 divided by 2
returns 0 then it belongs to Group 1 else Group 2. Send group 1 to one target, group 2 to 1
target
7. What is the difference between dynamic cache lookup and uncached lookup?
Ans. Uncached lookup: The informatica server doesn’t build a cache for temporary storage of data.
Whenever it needs to refer to the look up table, it scans the source directly.
Dynamic Cache: The informatica server builds a cache of the source when the workflow runs for the first
time and it updates the cache records dynamically after each row it loads to the target. Means, if a target
row gets updated, the dynamic cache also gets updated automatically after that particular row committed in
the target.
Classification : Confidential
8. What r two types of processes that informatica runs the session?
Ans. Load manager Process: Starts the session, creates the DTM process, and sends post-
session email when the session completes.
The DTM process. Creates threads to initialize the session, read, write, and transform data,
and handle pre- and post-session operations.
9. What r the new features of the server manager in the informatica 5.0?
Ans. U can use command line arguments for a session or batch. This allows U to change the
values of session parameters, and mapping parameters and mapping variables.
Parallel data processing: This feature is available for power center only. If we use the
informatica server on a SMP system, U can use multiple CPU’s to process a session
concurrently.
Process session data using threads: Informatica server runs the session in two processes.
Explained in previous question.
10. Which tool U use to create and manage sessions and batches and to monitor and
stop the informatica server?
Ans. Session - A Session Is A set of instructions that tells the Informatica Server How And
When To Move Data From Sources To Targets. After creating the session, we can use either
the server manager or the command line program pmcmd to start or stop the session. Batches
- It Provides A Way to Group Sessions For Either Serial Or Parallel Execution By The Informatica Server.
There Are Two Types Of Batches:
Sequential - Run Session One after the Other. concurrent - Run Session At The Same Time.
12. What r the different options used to configure the sequential batches?
Ans.Two-options
Run the session only if previous session completes successfully. Always runs the session.
Ans. If you configure a session in a sequential batch to stop on failure, you can run recovery
starting with the failed session. The Informatica Server completes the session and then runs
the rest of the batch. Use the Perform Recovery session property
Classification : Confidential
14. How to recover sessions in concurrent batches?
Ans. If multiple sessions in a concurrent batch fail, you might want to truncate all targets and
run the batch again. However, if a session in a concurrent batch fails and the rest of the
sessions complete successfully, you can recover the session as a standalone session.
To recover a session in a concurrent batch:
1.Copy the failed session using Operations-Copy Session.
2.Drag the copied session outside the batch to be a standalone session.
3.Follow the steps to recover a standalone session.
4.Delete the standalone copy.
Ans. A standalone session is a session that is not nested in a batch. If a standalone session
fails, you can run recovery using a menu command or pmcmd. These options are not available
for batched sessions.
Ans. You can use a Command task to call the shell scripts, in the following ways:
1. Standalone Command task. You can use a Command task anywhere in the workflow or
worklet to run shell commands.
2. Pre- and post-session shell command. You can call a Command task as the pre- or post-
session shell command for a Session task. For more information about specifying pre-session
and post-session shell commands
18. When we create a target as flat file and source as oracle.. How can i specify first
rows as column names in flat files...
Ans. use a pre sql statement....but this is a hardcoding method...if you change the column names or put in
extra columns in the flat file, you will have to change the insert statement
Classification : Confidential
20.how do we do unit testing in informatica?
how do we load data in informatica ?
1. Quantitative testing
2.Qualitative testing
Steps.
Once the session is succeeded the right click on session and go for statistics tab.
There you can see how many number of source rows are applied and how many number of rows loaded in
to targets and how many number of rows rejected. This is called Quantitative testing.
If once rows are successfully loaded then we will go for qualitative testing.
Steps
1.Take the DATM(DATM means where all business rules are mentioned to the corresponding source
columns) and check whether the data is loaded according to the DATM in to target table. If any data is not
loaded according to the DATM then go and check in the code and rectify it.
Ans. Factless Fact Tables are the fact tables with no facts or measures(numerical data). It contains only the
foreign keys of corresponding Dimensions.
Ans. You can use the joiner transformation by setting as outer join of either master or detail.
23. How can I get distinct values while mapping in Informatica in insertion?
Ans. You can add an aggregator before insert and group by the fields that need to be distinct.
IN the source qualifier write the query with distinct on key column
Ans. Changed Data Capture (CDC) helps identify the data in the source system that has
changed since the last extraction. With CDC, data extraction takes place at the same time
the insert, update, or delete operations occur in the source tables, and the change data is
stored inside the database in change tables. The change data, thus captured, is then made
available to the target systems in a controlled manner.
Classification : Confidential
25. How can we store previous session logs
Ans. Just run the session in time stamp mode then automatically session log will not overwrite current
session log.
We can do this way also. Using $PMSessionlogcount (specify the number of runs of the session log to
save)
Ans. informatica server is the main server component in informatica product family. Which is
responsible for reads the data from various source system and transforms the data according
to business rule and loads the data into the target table
Ans. u can also do it this way. connect to both the repositories, open the respective folders. keep the
destination repository as active. from the navigator panel just drag and drop the mapping to the work area. it
will ask whether to copy the mapping, say YES. its done.
28. my source is having 1000 rows. i have brought 300 records into my ODS. so next
time I want to load the remaining records. so i need to load from 301 th record.
when ever I start the work flow again it will load from the beginning. how do we
solve this problem.
ie by changing the RESET option in the properties tab of your Sequence Generator Transformation.
2) You can use a filter transformation and set a condition there as rownum>300
3) You can also use variable to store the rownum of the final row you loaded in the target.
From next time you can use this variable to load the rest of the data to the target.
4) You can do a lookup on the target table and check for the rows already present there. Hence the first 300
records will not be reloaded to the target.
Ans. A fact table is always DENORMALISED table. It consists of data from dimension table (Primary Key's)
and Fact table has Foreign keys and measures.
30. How can we join 3 database like Flat File, Oracle, Db2 in Informatrica..
If you want to view the data on source/target we can preview the data but with some limitations.
Classification : Confidential
32. how u will create header and footer in target using informatica?
Ans. If you are focus is about the flat files then one can set it in file properties while creating a mapping or at
the session level in session properties
Ans. In the designer go to the main menu and one can see the export/import options.
Import the exported mapping in to the production repository with replace options.
Ans. You can generate Power Center Metadata Reporter from a browser on any workstation, even a
workstation that does not have Power Center tools installed.
Ans. Before loading to target , use an aggregator transformation and make use of group by
function to eliminate the duplicates on columns.
Or u can use sorter transformation or Rank Transformation to eliminate duplicates from flat
files.
Ans. The Partitioning Option increases Power Center’s performance through parallel data
processing, and this option provides a thread-based architecture and automatic data
partitioning that optimizes parallel processing on multiprocessor and grid-based hardware
environments.
37.what is surrogate key ? In ur project in which situation u has used? explain with
example?
Ans. A surrogate key is system generated/artificial key /sequence number or A surrogate key
is a substitution for the natural primary key. It is just a unique identifier or number for each
row that can be used for the primary key to the table. The only requirement for a surrogate
primary key is that it is unique for each row in the table it is useful because the natural
primary key (i.e. Customer Number in Customer table) can change and this makes updates
more difficult. but In my project, I felt that the primary reason for the surrogate keys was to
record the changing context of the dimension attributes.(particularly for sod )The reason for
them being integer and integer joins are faster.
38. what is the diff b/w STOP & ABORT in INFORMATICA session level ?
Abort: WE cant restart the session. We should truncate all the pipeline after that start the session
Classification : Confidential
39. if the workflow has 5 session and running sequentially and 3rd session has been
failed how can we run again from only 3rd to 5th session?
Ans. If multiple sessions in a concurrent batch fail, you might want to truncate all targets and
run the batch again. However, if a session in a concurrent batch fails and the rest of the
sessions complete successfully, you can recover the session as a standalone session. To
recover a session in a concurrent batch:1.Copy the failed session using Operations-Copy
Session.2.Drag the copied session outside the batch to be a standalone session.3.Follow the
steps to recover a standalone session.4.Delete the standalone copy.
Ans. Its a big data warehouses OR centralized data warehousing OR the old style of warehouse.
Its a single enterprise data warehouse (EDW) with no associated data marts or operational data store (ODS)
systems.
41. how to load the data from people soft hrm to people soft erm using informatica?
2.Import the source and target from people soft using ODBC connections
3.Define connection under "Application Connection Browser" for the people soft source/target in workflow
manager .
select the proper connection (people soft with oracle,sybase,db2 and informix)
42. What is exact use of 'Online' and 'Offline' server connect Options while defining
Work flow in Work flow monitor? . The system hangs when 'Online' server connect
option. The Informatica is installed on a Personal laptop.
Ans. When the repo is up and the PMSERVER is also up, workflow monitor always will be connected on-
line.
When PMserver is down and the repo is still up we will be prompted for an off-line connection with which we
can just monitor the workflows
previously ran.
43. in real-time which one is better star schema or snowflake star schema the
surrogate will be linked to which columns in the dimension table.
Ans. In real time only star schema will implement because it will take less time and surrogate key will there
in each and every dimension table in star schema and this surrogate key will assign as foreign key in fact
table.
Classification : Confidential
44. what is the mapping for unit testing in Informatica, are there any other testing in
Informatica, and how we will do them as a etl developer. How do the testing people
will do testing are there any specific tools for testing.
Ans. In informatica there is no method for unit testing. There are two methods to test the mapping.
1. But we have data sampling. set the data sampling properties for session in workflow manager for
specified number of rows and test the mapping.
2. Use the debugger and test the mapping for sample records.
45. What are the properties should be notified when we connect the flat file source
definition to relational database target definition?
If its can be executed without performance issues then normal load will work
If its huge in GB they NWAY partitions can be specified at the source side and the target side.
Ans. yes, we can stop the batches using server manager or pmcmd command.
47. what are the real time problems generally come up while doing/running
mapping/any transformation?
Ans. may be you will encounter with connection failure, other then that i don't think so,
conserve will handle all the syntax errors and Invalid mappings.
48. What will happen if you are using Update Strategy Transformation and your
session is configured for "insert"?
Ans. if u r using a update strategy in any of ur mapping, then in session properties u have to set treat source
rows as Data Driven. if u select insert or update or delete, then the info server will not consider UPD for
performing any DB operations.
ELSE
u can use the UPD session level options. instead of using a UPD in mapping just select the update in treat
source rows and update else insert option. this will do the same job as UPD. but be sure to have a PK in the
target table.
49. how to get two targets T1 containing distinct values and T2 containing duplicate
values from one source S1.
Ans. Use filter transformation for loading the target with no duplicates. and for the other
transformation load it directly from source.
Classification : Confidential
50. Where is the cache stored in informatica?
51. In a joiner transformation, you should specify the source with fewer rows as the
master source. Why?
Ans. Joiner transformation compares each row of the master source against the detail source. The fewer
unique rows in the master, the fewer iterations of the join comparison occur, which speeds the join process.
52. 1)What are the various test procedures used to check whether the data is loaded
in the backend, performance of the mapping, and quality of the data loaded in
INFORMATICA.
1) Check in the workflow monitor status, whether no. of records in source and no. of actual records loaded
are equal
Ans. It displays the updated information about the session in the monitor window. The monitor
window displays the status of each session when you poll the Informatica server.
54. Partitioning, Bitmap Indexing (when to use), how will the bitmap indexing will
effect the performance.
Ans. Bitmap indexing a indexing technique to tune the performance of SQL queries. The
default type is B-Tree indexers which is of high cardinality (normalized data). You can use
bitmap indexers for de-normalized data or low cardinalities. The condition is the amount of
DISTINCT rows should be less than 4% of the total rows. If it satisfies the given condition then
bitmap indexers will optimize the performance for this kind of tables.
55. where do we use MQ series source qualifier, application multi group source
qualifier. just give an example for a better understanding.
Ans. We can use a MQSeries SQ when we have a MQ messaging system as source (queue).
When there is need to extract data from a Queue, which will basically have messages in XML format, we will
use a JMS or a MQ SQ depending on the messaging system. If you have a TIBCO EMS Queue, use a JMS
source and JMS SQ and an XML Parser, or if you have a MQ series queue, then use a MQ SQ which will be
associated with a Flat file or a Cobol file.
Ans. Junk Dimension a Dimension is called junk dimension if it contains attribute which are
rarely changed or modified. example In Banking Domain , we can fetch four attributes
accounting to a junk dimensions like from the Overall_Transaction_master table tput flag tcmp
flag del flag advance flag all these attributes can be a part of a junk dimensions.
Classification : Confidential
57. Explain about incremental aggregation with an example?
Ans. Incremental aggregation is specially used for tune the performance of the aggregator. It
captures the change each time (incrementally) you run the transformation and then performs
the aggregation function to the changed rows and not to the entire rows. This improves the
performance because you are not reading the entire source, each time you run the session.
Ans. Rank:
1
2<--2nd position
2<--3rd position
4
5
Same Rank is assigned to same totals/numbers. Rank is followed by the Position. Golf game ususally
Ranks this way. This is usually a Gold Ranking.
Dense Rank:
1
2<--2nd position
2<--3rd position
3
4
Same ranks are assigned to same totals/numbers/names. the next rank follows the serial number.
Ans. Data sources ---> Data acquisition ---> Warehouse ---> Front end tools ---> Metadata
management ---> Data warehouse operation management
Ans. Most complex logic we use is de-normalization. We don’t have any De-normalizer transformation in
INformatica. So we will have to use an aggregator followed by an expression. Apart from this, we use most
of the complexicity in expression transformation involving lot of nested IIF's and Decode
statements...another one is the union transformation and joiner.
61. In the source, if we also have duplicate records and we have 2 targets, T1- for
unique values and T2- only for duplicate values. How do we pass the unique values
to T1 and duplicate values to T2 from the source to these 2 different targets in a
single mapping?
function)--->t2
Classification : Confidential
62. what are the enhancements made to Informatica 7.1.1 version when compared to
6.2.2 version?
Ans. 1.union & custom transformation 2.Lookup on flat file 3. using PMCMD command.
4.Exporting independent and dependent repository objects 5. Version controlling 6.Data profiling
7.Supporting of 64MB Architecture. 8.Idap Authentication.
63. What is the difference between Power Centre and Power Mart?
64. What is the logic will you implement to load the data in to one fact from 'n'
number of dimension tables?
Ans. To load data into one fact table from more than one dimension tables . firstly u need to
create a fact table and dimension tables. later load data into individual dimensions by using
sources and transformations(aggregator, sequence generator, lookup) in mapping designer
then to the fact table connect the surrogate to the foreign key and the columns from
dimensions to the fact.
65. Can i use a session Bulk loading option that time can i make a recovery to the
session?
Ans. If the session is configured to use in bulk mode it will not write recovery information to
recovery tables. So Bulk loading will not perform the recovery as required.
Ans. You should configure the mapping with the least number of transformations and
expressions to do the most amount of work possible. You should minimize the amount of data
moved by deleting unnecessary links between transformations.
For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup transformations),
limit connected input/output or output ports. Limiting the number of connected input/output or output ports
reduces the amount of data the transformations store in the data cache.
You can also perform the following tasks to optimize the mapping:
Classification : Confidential
• Optimize expressions. You should configure the mapping with the least number of
transformations and expressions to do the most amount of work possible. You should
minimize the amount of data moved by deleting unnecessary links between
transformations.
67. What are mapping parameters and variables in which situation we can use it?
Ans. If we need to change certain attributes of a mapping after every time the session is run, it will be very
difficult to edit the mapping and then change the attribute. So we use mapping parameters and variables
and define the values in a parameter file. Then we could edit the parameter file to change the attribute
values. This makes the process simple.
69. If a session fails after loading of 10,000 records in to the target. How can u load
the records from 10001 th record when u run the session next time in informatica
6.1?
Ans. Running the session in recovery mode will work, but the target load type should be normal. If its bulk
then recovery won’t work as expected.
71. If u had to split the source level key going into two separate tables. One as
surrogate and other as primary. Since informatica does not guarantee keys are
loaded properly (order!) into those tables. What are the different ways you could
handle this type of situation?
73. How do I import VSAM files from source to target? Do I need a special plug-in?
Ans. using power exchange tool convert VSAM file to oracle tables then do mapping as usual
to the target table.
74. If you have four lookup tables in the workflow. How do you troubleshoot to
improve performance?
Ans. there r many ways to improve the mapping which has multiple lookups.
1) We can create an index for the lookup table if we have permissions (staging area).
Classification : Confidential
2) Divide the lookup mapping into two (a) dedicate one for insert means: source - target,, these r new rows
. only the new rows will come to mapping and the process will be fast. (b) Dedicate the second one to
update: source=target, these r existing rows. Only the row which exists already will come into the mapping.
75. How to retrieve the records from a rejected file. Explain with syntax or example.
Ans. there is one utility called "reject Loader" where we can find out the reject records. And
able to refine and reload the rejected records...
76. Which objects are required by the debugger to create a valid debug session?
Ans. We can create a valid debug session even without a single break-point. But we have to give valid
database connection details for sources, targets, and lookups used in the mapping and it should contain
valid mapplets (if any in the mapping).
77. Suppose session is configured with commit interval of 10,000 rows and source
has 50,000 rows. Explain the commit points for Source based commit and Target
based commit. Assume appropriate value wherever required.
Ans. Source based commit will commit the data into target based on commit interval. so, for every 10,000
rows it will commit into target.
Target based commit will commit the data into target based on buffer size of the target. i.e., it commits the
data into target when ever the buffer fills. Let us assume that the buffer size is 6, 000.So, for every 6,000
rows it commits the data.
Ans. Yes, we can use Informatica for cleansing data. Some time we use stages to cleansing the data. It
depends upon performance again else we can use expression to cleansing data.
For example an field X have some values and other with Null values and assigned to target field where
target field is not null column, inside an expression we can assign space or some constant value to avoid
session failure.
The input data is in one format and target is in another format, we can change the format in expression.
We can assign some default values to the target to represent complete set of data in the target.
Ans. When you create a version of a folder referenced by shortcuts, all shortcuts continue
to reference their original object in the original version. They do not automatically update
to the current folder version.
For example, if you have a shortcut to a source definition in the Marketing folder, version
1.0.0, then you create a new folder version, 1.5.0, the shortcut continues to point to the
source definition in version 1.0.0.
Classification : Confidential
Maintaining versions of shared folders can result in shortcuts pointing to different
versions of the folder. Though shortcuts to different versions do not affect the server, they
might prove more difficult to maintain. To avoid this, you can recreate shortcuts pointing
to earlier versions, but this solution is not practical for much-used objects. Therefore,
when possible, do not version folders referenced by shortcuts.
Ans. A Staging area in a DW is used as a temporary space to hold all the records from the source system.
So more or less it should be exact replica of the source systems except for the load strategy where we use
truncate and reload options.
So create using the same layout as in your source tables or using the Generate SQL option in the
Warehouse Designer tab.
Bottom up
82. What is the difference between summary filter and detail filter?
Ans. summary filter can be applied on a group of rows that contain a common value. where as
detail filters can be applied on each and every rec. of the data base.
83. What are the difference between view and materialized view?
Ans. A view do no derive the change made to it master table after the view is created.
A materialized view immediately carries the change done to its master table even after the materialized view
is created.
A materialized view is a database object that contains the results of a query. They
are local copies of data located remotely, or are used to create summary tables
based on aggregations of a table's data. Materialized views, which store data based
on remote tables are also, know as snapshots.
A materialized view can query tables, views, and other materialized views.
Collectively these are called master tables (a replication term) or detail tables (a
data warehouse term).
Ans. Time dimension is one of important in Data warehouse. Whenever u generated the
report, that time u access all data from thro time dimension.
Classification : Confidential
Fields : Date key, full date, day of week, day , month, quarter, fiscal year
85. What r the tasks that Load manger process will do?
Ans. Manages the session and batch scheduling, locking and reading the session, reading the parameter
file, verifies permissions and privileges, creating log files.
86. To provide support for Mainframes source data, which files r used as a source
definitions?
87. How the informatica server sorts the string values in Rank transformation?
Ans. When Informatica Server runs in UNICODE data movement mode, then it uses the
sort order configured in session properties.
Ans. Based on which port you want generate Rank is known as rank port, the generated
values are known as rank index.
89. What is difference between partitioning of relational target and partitioning of file
targets?
Ans. If u partitions a session with a relational target informatica server creates multiple
connections to the target database to write target data concurrently. If u partition a session
with a file target the informatica server creates one target file for each partition can configure
session properties to merge these target files.
Ans. An active data warehouse provides information that enables decision-makers within an organization
to manage customer relationships nimbly, efficiently and proactively. Active data warehousing is all about
integrating advanced decision support with day-to-day-even minute-to-minute-decision making in a way that
increases quality of those customer touches which encourages customer loyalty and thus secure an
organization's bottom line. The marketplace is coming of age as we progress from first-generation "passive"
decision-support systems to current- and next-generation "active" data warehouse implementations.
91. what are the advantages data mining over traditional approaches?
Ans. Data Mining is used for the estimation of future. For example, if we take a company/business
organization, by using the concept of Data Mining, we can predict the future of business in terms of Revenue
(or) Employees (or) Customers (or) Orders etc.
Traditional approaches use simple algorithms for estimating the future. But, it does not give accurate results
when compared to Data Mining.
Classification : Confidential
92. What is the main difference between Inmon and Kimball philosophies of data
warehousing?
Ans. RalfKimball: he follows bottom-up approach i.e., first create individual Data Marts from the existing
sources and then create Data Warehouse.
BillImmon: he follows top-down approach i.e., first create Data Warehouse from the existing sources and
then create individual Data Marts.
Ans. Junk dimension: Grouping of Random flags and text attributes in a dimension and moving
them to a separate sub dimension.
Degenerate Dimension: Keeping the control information on Fact table ex: Consider a
Dimension table with fields like order number and order line number and have 1:1 relationship
with Fact table, In this case this dimension is removed and the order information will be
directly stored in a Fact table in order eliminate unnecessary joins while retrieving order
information..
Ans. Basic diff is E-R modeling will have logical and physical model. Dimensional
model will have only physical model.
Ans. A BUS Schema or a BUS Matrix? A BUS Matrix (in Kimball approach) is to identify
common Dimensions across Business Processes; ie: a way of identifying Conforming
Dimensions.
Ans. Incremental loading means loading the ongoing changes in the OLTP. Aggregate table contains the
[measure] values, aggregated /grouped/summed up to some level of hierarchy.
Ans. Real-time data warehousing is a combination of two things: 1) real-time activity and 2)
data warehousing. Real-time activity is activity that is happening right now. The activity could
be anything such as the sale of widgets. Once the activity is complete, there is data about it.
Data warehousing captures business activity data. Real-time data warehousing captures
business activity data as it occurs. As soon as the business activity is complete and there is
data about it, the completed activity data flows into the data warehouse and becomes
Classification : Confidential
available instantly. In other words, real-time data warehousing is a framework for deriving
information from data as the data becomes available.
Ans. level of granularity defines the extent of detail. As an example, let us look at
geographical level of granularity. We may analyze data at the levels of COUNTRY, REGION,
TERRITORY, CITY and STREET. In this case, we say the highest level of granularity is STREET.
Ans. Degenerate Dimensions: If a table contains the values, which r neither dimension nor measures is
called degenerate dimensions. Ex : invoice-id, empno .
Ans. Additive: A measure can participate arithmetic calculations using all or any dimensions.
Semi additive: A measure can participate arithmetic calculations using some dimensions.
Ex: temperature
ANS. An updatable set of integrated operational data used for enterprise- wide tactical decision making.
Contains live data, not snapshots, and has minimal history retained.
Ans. Full Load is the entire data dump load taking place the very first time.
Gradually to synchronize the target data with source data, there are further 2 techniques:-
Refresh load - Where the existing data is truncated and reloaded completely.
Incremental - Where delta or difference between target and source data is dumped at regular
intervals. Timestamp for previous delta load has to be maintained.
Ans. Status codes provide error handling for the Integration Service during a workflow.
Classification : Confidential
106. In a sequential batch how can i stop/start the particular session?
1 st up all if u r working in windows environment. its easy u can just click on particular task on that batch and
make it run ... be sure that remaining all are in suspend mode
If u are working in UNIX environment change status of each session using pmcmd commands.
Ans. With a test load, the Integration Service reads and transforms data without writing
to targets. The Integration Service generates all session files, and performs all pre- and
post-session functions, as if running the full session.
The Integration Service writes data to relational targets, but rolls back the data when the
session completes. For all other target types, such as flat file and SAP BW, the
Integration Service does not write data to the targets.
108. How the Informatica Server will identify the string as date?
Ans. External loaders can increase session performance by loading information directly from a file or pipe
rather than running the SQL commands to insert the same data into the database.
Ans. Bulk loading improves the performance of a session that inserts a large amount of data to the target
database.
Ans. When bulk loading to Microsoft SQL Server or Oracle targets, define a large commit interval to reduce
the number of bulk load transactions and increase performance.
Ans. If the source changes incrementally and you can capture changes, you can configure the session to
process those changes. This allows the Integration Service to update the target incrementally, rather than
forcing it to process the entire source and recalculate the same data each time you run the session.
Ans. A grid is an alias assigned to a group of nodes that run sessions and workflows. When you run a
workflow on a grid, you improve scalability and performance by distributing Session and Command tasks to
service processes running on nodes in the grid.
Classification : Confidential
115. What r the tasks done by DTM?
Ans. When you create a session with multiple partitions, the Integration Service creates
cache for the Aggregator, Joiner, Lookup, and Rank transformations. It creates a separate
cache for each partition, and each partition works with the rows needed by that partition.
As a result, the Integration Service requires a portion of total cache memory for each
partition. When you run a session, the Integration Service accesses the cache in parallel
for each partition.
Ans. If the volume of data grows or you add more CPUs, you might need to adjust
partitioning so the session run time does not increase. When you use dynamic
partitioning, you can configure the partition information so the Integration Service
determines the number of partitions to create at run time.
Ans. You can optimize session performance by using the database partitioning partition
type for source and target databases. When you use source database partitioning, the
Integration Service queries the database system for table partition information and fetches
data into the session partitions. When you use target database partitioning, the Integration
Service loads data into corresponding database partition nodes.
Indicator files are two types: a) ROW wise Indicator file b) COLUMN wise Indicator file.
Classification : Confidential
121. What is Push down Optimization?
Ans. When you run a session configured for pushdown optimization, the Integration Service analyzes the
mapping and writes one or more SQL statements based on the mapping transformation logic.
Ans. The XML Parser transformation is similar to an XML source definition. When the Integration Service
processes an XML Parser transformation, it reads a row of XML data, parses the XML, and passes data
through output groups.
Ans. Use an XML Generator transformation to combine input that comes from several sources to create an
XML document.
Ans. Stores the information, or metadata used by informatica Server and Client tools.
Ans. Metadata describes different types of objects such as mappings, or transformations, that u can create
or modify using informatica client tools.
Ans. It uses all the repository objects to extract, transform and load data.
127. I tried to view the session log, but it contains garbage characters.
Ans. Enable the Session Log in UTF-8 option on the Configuration tab of the Informatica
Server setup program.
128. I installed and registered the Informatica Server, but I get an internal error when I try to start the
service.
Ans. Verify that the Repository Server is running. You must start the Repository Server
before you can start the Informatica Server.
129. I started the Informatica Server, but it does not connect to the repository.
Ans. Verify that the Informatica Server and the system on which you installed the
Informatica Server are configured to connect to the Repository Server. Also verify that
the Repository Server is running.
130. The Informatica Server does not start after rebooting the machine.
Ans. In the Services configuration, make sure the Startup Type is set to automatic.
Classification : Confidential
131.What r the different types of threads available in DTM?
Ans. Pre and post sessions, writer, reader, transformation, and mapping threads.
132. Where will the informatica Server creates the reject files.
133. I have a section in a parameter file for a session, but the Informatica Server
does not seem to read it.
Ans. In the parameter file, folder and session names are case-sensitive. Make sure to enter
folder and session names exactly as they appear in the Workflow Manager. Also, use the
appropriate prefix for all user-defined session parameters.
134. I am trying to run a workflow with a parameter file and one of the sessions
keeps failing.
Ans. The session might contain a parameter that is not listed in the parameter file. The
Informatica Server uses the parameter file to start all sessions in the workflow. Check the
session properties, and then verify that all session parameters are defined correctly in the
parameter file.
135. I am trying to set up a session to load data to an external loader, but I cannot
select an external loader connection in the session properties.
Ans. Check your mapping to make sure you did not configure it to load to a flat file target.
In order to use an external loader, you must configure the mapping to load to a DB2,
Oracle, Sybase IQ, or Teradata relational target. When you create the session, select a file
writer in the Writers settings of the Targets tab in the session properties. Then open the
Connections settings and select an external loader connection.
136. What r the out put files that the informatica server creates during the session
running?
Ans. Indicator file, output file, performance detail log, session log, session details,
workflow log, workflow details, control file, reject files, cache files.
Ans. If you use a flat file as a target, you can configure the Informatica Server to create
an indicator file for target row type information. For each target row, the indicator file
contains a number to indicate whether the row was marked for insert, update, delete, or
reject.
Classification : Confidential
138. What is control File?
Ans. When you run a session that uses an external loader, the Informatica Server creates
a control file and a target flat file. The control file contains information about the target
flat file such as data format and loading instructions for the external loader.
Ans. If the session writes to a target file, the Informatica Server creates the target file
based on file target definition. By default, the Informatica Server names the target file
based on the target table: target_table.out.
141. Techniques of Error Handling - Ignore, Rejecting bad records to a flat file,
loading the records and reviewing them (default values)
Ans. Rejection of records either at the database due to constraint key violation or the
informatica server when writing data into target table. These rejected records we can find in
the bad files folder where a reject file will be created for a session. we can check why a record
has been rejected. And this bad file contains first column a row indicator and second column a
column indicator.
These row indicators or of four types
D-valid data,
O-overflowed data,
N-null data,
T- Truncated data,
And depending on these indicators we can changes to load data successfully to target.
142.What are the various methods of getting incremental records or delta records
from the source systems?
Ans. getting incremental records from source systems to target can be done
by using incremental aggregation transformation
Classification : Confidential
144. What are the modules in Power Mart?
Ans. Yes
Ans. Code Page is used to identify characters that might be in different languages. If you are
importing Japanese data into mapping, u must select the Japanese code page of source data.
Aggregator Transformation: If u use sorted ports u cannot partition the associated source.
Joiner Transformation: U cannot partition the master source for a joiner transformation.
Ans. In case of stored procedure transformation procedure will be compiled and executed in a
relational data source need data base connection to import the stored procedure in to u’r
maping.Where as in external procedure transformation procedure or function will be executed
out side of data source.Ie u need to make it as a DLL to access in u r maping.No need to have
data base connection in case of external procedure transformation.
Ans. informatica server connects source data and target data using native
ODBC drivers
Again it connect to the repository for running sessions and retrieving metadata information
Source------>informatica server--------->target
Classification : Confidential
150. What is the difference between connected and unconnected stored procedures?
Ans.
Unconnected
Run a stored procedure once during your mapping, such as pre- or post-session.
Connected or
Run a stored procedure every time a row passes through the Stored Procedure transformation. Unconnected
Run a stored procedure based on data that passes through the mapping, such as when a specific Unconnected
port does not contain a null value.
Connected or
Pass parameters to the stored procedure and receive a single output parameter. Unconnected
Pass parameters to the stored procedure and receive multiple output parameters.
Connected or
Note: To get multiple output parameters from an unconnected Stored Procedure transformation, Unconnected
you must create variables for each output parameter. For details, see Calling a Stored Procedure
From an Expression.
Unconnected
Call multiple times within a mapping.
151. Suppose session is configured with commit interval of 10,000 rows and source has 50,000 rows. Explain the
commit points for Source based commit and Target based commit. Assume appropriate value wherever
required.
Ans. Source based commit will commit the data into target based on commit interval.so, for
every 10,000 rows it will commit into target.
Target based commit will commit the data into target based on buffer size of the target.i.e; it
commits the data into target when ever the buffer fills. Let us assume that the buffer size is
6,000.So,for every 6,000 rows it commits the data.
152. Which objects are required by the debugger to create a valid debug session?
Ans. We can create a valid debug session even without a single break-point. But we have
to give valid database connection details for sources, targets, and lookups used in the
mapping and it should contain valid mapplets (if any in the mapping).
153. How does the server recognize the source and target databases?
Ans. By using ODBC connection. if it is relational. if is flat file FTP connection. See we can
make sure with connection in the properties of session both sources & targets.
Classification : Confidential
154. How to lookup the data on multiple tables.
Ans. When you create lookup transformation that time INFA asks for table name so you can choose source,
target, import and skip. So click skip and the use the sql override property in properties tab to join two tables
for lookup.
155. How do you handle decimal places while importing a flat file into informatica?
Ans. while importing flat file definition just specify the scale for a numeric data type. in the
mapping, the flat file source supports only number data type(no decimal and integer). In the
SQ associated with that source will have a data type as decimal for that number port of the
source?
Source ->number data type port ->SQ -> decimal datatype.Integer is not supported. Hence
decimal is taken care.
Ans. Partition points mark the thread boundaries in a source pipeline and divide the pipeline into stages.
158. How to create a custom transformation, can u give a real-time example where
exactly u has used it. Just give some explanation why u used the custom
transformation.
Ans. By using Custom transformation we create the single input ports and multiple output ports
or multiple input ports and single input otherwise multiple input or output ports. When we
clicking the custom transformation one window will opens we have select active custom or
passive custom transformation, here one thing this is one transformation it acts like active and
passive.
160. How can we join the tables if the tables have no primary and foreign key relation
and no matching port to join?
Ans. without common column or common data type we can join two sources using dummy
ports.
Classification : Confidential
3. Use Joiner transformation to join the sources using dummy port (Use join conditions).
161. 1] In certain mapping there are four targets tg1, tg2, tg3 and tg4.
Tg1 has a primary key, tg2 foreign key referencing the tg1's primary key, tg3 has
primary key that tg2 and tg4 refers as foreign key, tg2 has foreign key referencing
primary key of tg4 ,the order in which the informatica will load the target?
2] How can I detect aggregate transformation causing low performance?
Ans. T1 and T3 are being the master table and don't have any foreign key reference to other
table will be loaded first. Then T3 will be loaded as it's master table T3 is already been loaded.
and at the end T2 will be loaded as it's all master table T1, T3, and T2 to whtch it refers has
been already loaded.
To optimize the aggregator transformation, you can use the following options.
Ans. Union Transformation: in Union Transformation, we may combine the data from two (or)
more sources. Assume, Table-1 contains '10' rows and Table-2 contains '20' rows. If we
combine the rows of Table-1 and Table-2, we will get a total of '30' rows in the Target. So, it is
definitely an Active Transformation.
Ans. Just run the session in time stamp mode then automatically session log will not overwrite
current session log. (OR)
Go to Session-->right click -->Select Edit Task then Go to -->Config. Object then set the
property
164. How do you create single lookup transformation using multiple tables?
Ans. 1.Write a override sql query. Adjust the ports as per the sql query.
2. For connected lkp transformation1>create the lkp transformation.2>go for
skip.3>manually enter the ports name that u want to lookup.4>connect with the i/p port
from src table.5>give the condition6>go for generate sql then modify according to u'r
requirement validate it will work....
3. Just we can create the view by using two table then we can take that view as
lookup table
Classification : Confidential
165. What is the architecture of any Data warehousing project? What is the flow?
Ans. 1) The basic step of data warehousing starts with data modeling. i.e creation dimensions and facts.
2) Data warehouse starts with collection of data from source systems such as OLTP,CRM,ERPs etc
4)by the end of ETL process target databases(dimensions, facts) are ready with data which
accomplishes the business rules.
5) Now finally with the use of Reporting tools(OLAP) we can get the information which is used
for decision support.
Ans. Yes we can do it in Unix Server but not in Windows based server.
167. What about rapidly changing dimensions? Can u analyze with an example?
Ans. rapidly changing dimensions are those in which values are changing continuously and
giving lot of difficulty to maintain them.
i am giving one of the best real world example which i found in some website while browsing.
Go through it. i am sure you like it.
So a better option would be shift those attributes into a fact table as facts, which solves the
problem.
Ans. The Repository Agent is a multi-threaded process that fetches, inserts, and updates
metadata in the repository database tables. The Repository Agent uses object locking to
ensure the consistency of metadata in the repository.
Bottleneck in ETL Processing is the point, by which the performance of the ETL Process is
slower.
Classification : Confidential
When ETL Process is in progress, first thing login to workflow monitor and observe
performance statistic. I.e. observe processing rows per second. In SSIS and Data stage,
when you run the job, you can see at every level how many rows per second is processed
by the server.
Mostly bottleneck occurs at source qualifier – during fetching data from source, joiner,
aggregator, Lookup – Cache Building Session.
Ans. The informatica Server saves the cache files for every session and reuses for the next session, by that
the query on the table will be reduced, so there will be some performance increment.
Ans. In time based scheduling, the jobs run at the specified time. In some situations, we've to
run a job based on some events like if a file arrives then only the job has to run whatever the
time it is. In such cases, event based scheduling is used.
172. What is the new lookup port in look-up transformation and explain?
Ans. Sometimes you need to create a generated key for a column in the target table. For
lookup ports with an Integer or Small Integer data type, you can associate a generated key
instead of an input port. To do this, select Sequence-ID in the Associated Port column.
When you select Sequence-ID in the Associated Port column, the Informatica Server
generates a key when it inserts a row into the lookup cache. Map the lookup/output ports
to the target to ensure that the lookup cache and target are synchronized.
Ans. Toad is an application development tool built around an advanced SQL and PL/SQL editor.
Using Toad, you can build and test PL/SQL packages, procedures, triggers, and functions. You
can create and edit database tables, views, indexes, constraints, and users. The Schema
Browser and Project Manager provide quick access to database objects.
Ans. No, it’s not possible to validate all the mappings in the repository at once.
Ans. "referential Integrity" is all about foreign key relationship between tables. Need to check
for the primary and foreign key relationship and the existing data if any. (See if child table has
any records which are pointing to the master table records that are no more in master table.)
Ans. Used to update the particular records in the target level using where clauses.
Classification : Confidential
177. What is auxiliary mapping?
Ans. Auxiliary mapping is used to reflect change in one table when ever there is a change in
the other table.
Example:
In Siebel we have S_SRV_REQ and S_EVT_ACT table; let’s say that we have a image table
defined for S_SRV_REQ from where our mappings read data. Now if there is any change in
S_EVT_ACT then it won’t be captured in S_SRV_REQ if our mappings are using image table for
S_SRV_REQ. To overcome this we define a mapping between S_SRV_REQ and S_EVT_ACT
such that if there is any change in second it will be reflected as an update in the first table.
Ans. Authentication requests validate user names and passwords to access the Power Center
repository. You can use the following authentication requests to access the Power Center
repositories: Login Logout The Login function authenticates user name and password for a
specified repository. This is the first function a client application should call before calling any
other functions. The Logout function disconnects you from the repository and its Power Center
server connections. You can call this function once you are done calling Metadata and Batch
Web Services functions to release resources at the Web Services Hub.
Ans. Use sequence generator and set the 'Increment by' property in that with 2.
180.In my source table 1000 rec's are there. I want to load 501 rec to 1000 rec into
my Target table?
Ans. Connect a sequence generator next to source qualifier into the expression. Use the filter
transformation to filter rows like where nextval >500.
181. What is the diff b/w source qualifier transformation and application source qualifier
transformation?
Ans. Source qualifier transformation is used to extract the data from one or more tables from
RDBMS of the same instance or from a single flat file system.
Application Source Qualifier transformation is used to extract the data from Application
sources like ERP.
182. What is the optimum method to load five flat files with same structures into
target? What is meant by file repository?
Ans. File repository is a collection of files with source file type as Indirect.
Just open a empty notepad and give the paths of files like
1. C: Flatfilesemp.txt 2.
D: Flatfilesemp2.txt 3. G: Flatfilesemp3.txt and u can give any
numbers. Be careful when u set the session properties “Set source file type as Indirect”.
Classification : Confidential
183. What are all the new features of informatica 8.1?
Ans. 1. Java Custom Transformation support 2. HTTP transformation support 3. New name of
Superglue as Metadata manager 4. New name of Power Analyzer as Data Analyzer 5. Support
Grid Computing 6. Push down Optimization.
Ans. U can transfer file from one server to other..............In UNIX there is an utility XCOMTCP
which transfer file from one server to other. But lot of constraints they are for this..... U need
to mention target server name and directory name where u need to send. The server directory
should have write permeation.....Check in detail in UNIX by typing MAN XCOMTCP command
which guides u i guess.
Ans. Unconnected lookup is used for updating Slowly Changing Dimensions...so it is used to
determine whether the rows are already in the target or not, but it doesn't change the no. of
rows ...so it is passive.
Connected lookup transformations are used to get a related value based on some value or to
perform a calculation.....in either case it will either increase no. of columns or not...but
doesn't change row count...so it is passive.
In lookup SQL override property we can add a WHERE statement to the default SQL
statement, but it doesn't change no. of rows passing through it, it just reduces the no. of rows
included in the cache.
Ans. Generally any mapping run from UNIX i.e. UNIX shell script calls work flow. Work flow
intern calls mapping. Suppose i am generating a flat file in UNIX from informatica mapping, if
I need to keep track of file when it has come today or yesterday...... For this append the
target table name with current date time stamp. In order to achieve u can use Unix shell script
in post session command. A small example i have given ...............U can do a lot ...........If
you are using doing MLOAD or Fload.Here even u can sort the flat file generated before doing
Mload or Fload.
187. Is
it possible to execute work flows in different repositories at the same time
using the same informatica server?
Ans. Yes
188. Explain about Informatica server process that how it works relates to mapping variables?
Ans. First load manager starts the session and it performs verifications and validations
about variables and manages post session tasks such as mail.
then it creates DTM process.
this DTM in turn creates a master thread, which creates remaining threads.
master thread creates
Classification : Confidential
read thread
write thread
transformation thread
pre and post session thread etc...
Finally DTM hand over’s to the load manager after writing into the target.
189. What does Check-In and Check-Out option refer to in the mapping designer?
Ans. Check-In and Check-Out refers to Versioning your Mapping. It is like maintaining the
changes you have made. It is like using VSS or CVS. When you right-click you mapping, you
have a option called Versioning if you have got that facility enabled.
190. Wheredo the records goes which does not satisfy condition in filter
transformation?
Ans. The most common performance bottleneck occurs when the Informatica Server writes to a target
database. You can identify target bottlenecks by configuring the session to write to a flat file target. If the
session performance increases significantly when you write to a flat file, you have a target bottleneck. If
your session already writes to a flat file target, you probably do not have a target
bottleneck. You can optimize session performance by writing to a flat file target local to
the Informatica Server. Causes for a target bottleneck may include small check point intervals, small
database network packet size, or problems during heavy loading operations. For details about eliminating a
target bottleneck.
Ans. Performance bottlenecks can occur when the Informatica Server reads from a source
database. If your session reads from a flat file source, you probably do not have a source
bottleneck. You can improve session performance by setting the number of bytes the
Informatica Server reads per line if you read from a flat file source. If the session reads from
relational source, you can use a filter transformation, a read test mapping, or a database query to identify
source bottlenecks.
Ans. You can identify mapping bottlenecks by using a Filter transformation in the mapping. You can add a
Filter transformation in the mapping before each target definition. Set the filter condition to false so that no
data is loaded into the target tables. If the time it takes to run the new session is the same as the original
session, you have a mapping bottleneck. You can also identify mapping bottlenecks by using performance
details.
Ans. You can identify a session bottleneck by using the performance details. The Informatica Server creates
performance details when you enable Collect Performance Data in the Performance settings on the
Properties tab of the session properties. Performance details display information about each Source
Classification : Confidential
Qualifier, target definition, and individual transformation. All transformations have some basic counters that
indicate the number of input rows, output rows, and error rows. Small cache size, low buffer memory, and
small commit intervals can cause session bottlenecks. For details on eliminating session bottlenecks.
195. If I add a column in my flat file source how can load to the target which already contains data.
Ans. Alter the target table with new field and re-import the target and source. Add one lookup on target
table, and take the newly added field return port and create one filter, check lookup return value is null those
null records only populate into target.
196. What are the different types of repositories available from 6.1 version onwards?
Ans. A) Local Repository b) Global Repository c) Stand Alone Repository d) Versioned Repository
Ans. $: These are system variables like $ Bad File, $ Input File, $ Output File, $ DB Connection.
198. How do you recover a session or folder if you accidentally dropped them?
Ans. If you dropped your session or folder we can get it back if it has proper backup.
we can also get it back when you have same folders in other environment, I mean to say,
generally, we will have three environments and won't have back up for the same, there three
are dev/stage and production. If you dropped in any one or two of them would get it
recovered by copying objects from.
But If you intentionally dropped all three environments' folders or sessions then we cannot get
them back.
199. How do you load flat files of same structure into a target table?
Ans. You can do it through file list. You have create a file containing path of all file. And u
have to give that file name in session properties and select indirect.
200. Which transformation you can use to get n+1 record in Target database from N records source?
Ans. Normalizer
201. How do we extract SAP data Using Informatica? What is ABAP? What are IDOCS?
Ans. We can extract SAP data by using informatica with the powerConnect option.
We need to install and configure the powerConnect tool . Import the Source in to the Source
Analyzer. Here PowerConnect is the gateway between SAP & Informatica.Then develops the
mapping in the mapping designer. Next we need to generate the ABAP code for that mapping,
then only the Informatice can be pull the data from SAP.
Classification : Confidential
202. We have one source and 3 targets with same fields i.e. empno, ename in all 3 targets and
source, then the 1st row in source should go to 1st target, 2nd row in source should go to 2nd target
and the 3rd row in source should go to 3rd target, 4th row again to 1st target, 5th row to target2 6th
row to target 3 like this we need to transfer data how is it possible?
Ans. First Load Data from source To One dummy table in target with SEQ Gen T/R.
in SEQ Gen T/R Set min value 1 and max value 3 and check the reset option.
It will load the data with new column Values 1,2,3,1,2,3...........
Then use that dummy table as a source And use a router T/R in This mapping pipeline.
put Condition In router T/R, with three groups and load data in three different Targets.
Ans. It is a intersection table between a dimension table & fact table with a multipart key
capturing a many-to-many relationship.
205. How to send a mail to users with the output file data?
Ans. It can be possible by using %a in Session Components POST SESSION EMAIL tab.
206. What are the benefits of using of 64-bit platform instead of 32-bit?
Ans. With 32-bit the max memory you can have for a session is 2^32
With 64-bit the max memory you can have 2^64 - more memory means faster execution of
session.
207. What are the main differences between the informatica version 7 and version 8?
Ans. The new Version Informatica 8.1 has got some additional feature's like:1) Mapping Generation
Option2) Push Down Optimisation3) Data Cleanse and Match4) Data Federation (EII)5) Enterprise Grid6)
High Availability7) Unstructured Data And there are some updations in 1) Data Profiling2) Partitioning3)
Real-Time Loading.
209. What is the primary purpose of using surrogate key? Will those be used in fact tables? If not
then what is the purpose in using it?
Ans. Surrogate key is used for indexing purpose. i.e. as a primary key for a table. You can
say that why can't we use primary key. But the problem with pk is it may take more or less
memory i.e. it may contain characters but sk contains only number. So comparison of
numbers is easy when comparing with characters.
Yes sk will be used in fact tables to establish a relationship between dimensions.
Classification : Confidential
210. How do you take only duplicate rows in target table in Informatica?
Ans. you use rank transformation make range according the field which representing duplicasy
rows that have more then 1 rank put get only those rows in target table.
211. What is the difference between direct mode and indirect mode in relates to session properties?
Ans. Indirect and direct mode in session properties are related to flat file. If you are giving the
flat file name directly then it is direct mode. If you have 2 or more files where data is present
and you want the data to be loaded from all of these files then, you select indirect and give
that file name where you save all file name of the actual flat file.
212. i have EMP table with eno, ename as columns with duplicate rows.
Ex : eno ename
1 kiran
2 sena
3 naveen
1 kiran
4 dev
3 naveen
1 kiran
5 Loki
3 Naveen
1 kiran
6 shashi
1 kiran
etc...
So I want to find out the third occurrence of that duplicate row. I don’t know where that row is. There
r some millions of rows and so many duplicates for each row.
Ans. 1. Select distinct (ename) from emp_tab where empno in (select empno from (select
empno, count (*) from emp_tab group by empno having count (*)>2));
2. Select * from emp where rowid > select max(rowid) from emp group by eno,ename
3. Select * from t_name where rowid in (select rowid from t_name group by empno
having count (*) =3);
214. How a custom transformation works and what are the options and properties need to be set?
Ans. To extend the power center functionality by using outside developed procedures. You can create
Custom transformations with multiple input and output groups.
Ans. Associate lookup ports with either an input/output port or a sequence ID. The Power
Center Server uses the data in the associated ports to insert or update rows in the lookup
cache. If you associate a sequence ID, the Power Center Server generates a primary key for
inserted rows in the lookup cache.
Classification : Confidential
216. What is the use of NewLookupRow in Lookup transformation when u enables the dynamic
lookup cache?
Ans. The Designer adds this port to a Lookup transformation configured to use a dynamic
cache. Indicates with a numeric value whether the Power Center Server inserts or updates the
row in the cache, or makes no change to the cache. To keep the lookup cache and the target
table synchronized, you pass rows to the target when the NewLookupRow value is equal to 1
or 2.
Ans. Dimension table having more number of .attributes (dimension) where as dimension is single attribute.
218. What is the difference between Informatica server, Repository Server, Power center Server?
Ans. Both the Informatica Server and Power Centre server are same meaning in
Informatica.Its working to work proceedings for sessions and workflows in workflow Manager
Windows. The Repository server of Informatica is only to connect the Repository through the
Informatica Repository administration console and also connect to multiple repositories at the
same time. After Repository Server runned the Power Centre Server can work.
Ans. power channel helps to transfer large amount of encrypted and compressed data over
LAN, WAN through firewalls, transfers files over FTP.
Ans. Ifyou want to save and reuse the cache files, you can configure the transformation to
use a persistent cache. Use a persistent cache when you know the lookup table does not
change between session runs.
Ans. You can also share the cache between multiple Lookup transformations by using a
persistent lookup cache and naming the cache files. You can share one cache between
Lookup transformations in the same mapping or across mappings.
Ans. You can instruct the Informatica Server to rebuild the lookup cache if you think that
the lookup table changed since the last time the Informatica Server built the cache. When
you want to rebuild a lookup cache, use the Recache from Database option.
Classification : Confidential
224. What are the differences between physical and logical data model?
Ans. The main important aspect is physical data model is created after logical data
model. For designing a logical data model one need not have through knowledge in
any specific RDBMS like oracle, Informix and so on. This is because logical data
model is created after having through study of the business entities which would be
involved in the application. The same can be represented by using E-R diagram also
called as Entity Relationship diagram which gives a diagrammatic representation of
the study made on the business application. The main aspect or quality a person
must have to build a good logical data model is through understanding of the
business, thorough knowledge about the application being modeled and very good
understanding of relational theory for presenting the understanding in a structured
manner. After the logical data model is built as specified above the physical data
model comes into picture. The logical data model built is used by business authors
and analysts and after a through revision and review the physical data model is
created. For creating a physical data model one must have good understanding of a
specific RDBMS in which it would be built. This is because physical data model is
represented using primary keys, foreign keys, triggers, referential integrity,
constraints and so on that is specific for any RDMBS.
225. What is the command to find out which shell you are running?
Ans. PS command is used to find out which shell you are running.
227. if we r using aggregator we enabled sorted input but the records r not sorted what
happen?
Ans. This technique is used when you want a performance boost. There will be an aggregator
cache created and will be indexed. Although you have not given a sorted input it will still work fine
as it is indexed in cache. Please note this is not like a page fault where you have to bring the
page from the secondary memory when not found in primary memory.
228. Draw a workflow diagram for this situation. If s3 is true session s4 and s5 should concurrently
Run.if any one s4 or s5 true it should run s1.If s1 success it should execute s2 otherwise exit.can
any one give me suggestion.
Classification : Confidential
7.Connect it to S2 using the same PreviousTaskStatus =
Succeeded.
229. What is mean by grouping of condition column in lookup transformation?
Ans. Its a validation rule If the data is altered outside the company
firewall, the checksum will automatically detect the violation and deny
validation of the data.
232. Which T/r we can use it mapping parameter and mapping variable?
and which one is reusable for any mapping mapping parameter or mapping
variable?
233. One flat file it contains some data but i want to don’t want to
load first and last record.how it is? Can u tell me complete logic?
Ans. When you are using Flat File source you usually get the header
files separately from the data files. You may get source data in more
than one file. In such cases you give the paths of all the data files
in one file for eg say list.txt. This file is called the list file. In
the session edit task window instead of giving the source filename you
enter the name of the list file - list.txt and in source file type you
make it indirect.
Classification : Confidential
235. I have flat file source. i want to load the maximum salary of each
deptno into target. What is the mapping flow?
238. How DTM buffer size and buffer block size are related?
239. I have flat file source. I have two targets t1,t2. I want to load
the odd no. of records into t1 and even no. of records into t2. What is
the procedure and what t/r's are involved and what is the mapping flow?
Or
Classification : Confidential
240. What is casual dimension?
For example:
When the cash register rings in a retail store, a fact table record is
created for each line item on the sales ticket. The obvious dimensions
of this fact table record are product, store, customer, sales ticket,
and time. At a bank ATM, a fact table record is created for every
customer transaction. The dimensions of this fact table record are
financial service, ATM location, customer, transaction type, and time.
When the telephone rings, the phone company creates a fact table record
for each "hook event." A complete call-tracking data warehouse in a
telephone company records each completed call, busy signal, wrong
number, and partially dialed call. In all three of these cases, a
physical event takes place, and the data warehouse responds by storing
a fact table record. However, the physical events and the corresponding
fact table records are more interesting than simply storing a small
piece of rev enue. Each event represents a conscious decision by the
customer to use the product or the service. A good marketing person is
fascinated by these events. Why did the customer choose to buy the
product or use the service at that exact moment? If we only had a
dimension called "Why Did The Customer Buy My Product Just Now?" our
data warehouses could answer almost any marketing question. We call a
dimension like this a "causal" dimension, because it explains what
caused the event.
242. We have a parameter file in UNIX location where we have .txt files
and those file will be used as source in informatica. I cannot use
source file name directly as file name will keep on changing in Unix
location. I need to define $$InputFile as parameter. Can anybody send
me the parameter file and the steps to handle this?
Ans. eg:
basu.txt(prm)
[foldername.sessname]
$$inputfileabc.domainname&path(root/aaa/bbb/prm.txt
Classification : Confidential
give remove all other options
have only input conn:$$inputfileabc
243. How do u tune queries?
245. What is the filename which you need to configure in UNIX while
installing informatica?
Ans. pmserver.cfg
Normalizer Transformation
XML targets.
247. In unconnected lookup , what are the other transformations , that can be
used in place of that expression transformation ?
Classification : Confidential
When the data in the file, may be tab separated, is incorrect. I.e. if
there is an extra tab for a set of records and this is making the text
data come under a numeric column, this violates the datatype and the
source file is rejected.
251. How the informatica server sorts the string values in Rank
transformation?
Ans. When the informatica server runs in the ASCII data movement
mode it sorts session data using Binary sort order. If you
configure the session to use a binary sort order, the informatica
server calculates the binary value of each string and returns the
specified number of rows with the highest binary values for the
string.
Ans. After you gather all the relevant information about the subject
area, the next step in the design process is to determine the
granularity of the fact table. To do this you must decide what an
individual low-level record in the fact table should contain. The
components that make up the granularity of the fact table correspond
directly with the dimensions of the data model. Thus, when you define
the granularity of the fact table, you identify the dimensions of the
data model.
Or
254. How can we eliminate duplicate rows from flat file, explain?
Classification : Confidential
256. What are the new features of informatica 7.1?
257. In update strategy target table or flat file which gives more
performance? Why?
Ans. Flat file give better performance bcz Retrieval of data from flat
file is fast then relation database.
258. If u can't find what you are looking for in lookup table, how do
you handle?
259. Suppose data are coming from different locations and those data
will not change. Is there any need to use surrogate key?
Ans. Yes, we should use the surrogate key, here we are getting data
from different locations means every one have one primary key, while
transforming the data into target that time more than two key not in
use so if you use surrogate key it will identified the duplicate fields
in dimensional table.
Ans. bit map index is an index in which the data will be in the form of
bites, by this we can retrieve the data very fastly
btree index is an index which is default index as normal indexes.
Classification : Confidential
262. How to we create data mart?
Ans. This technique is used when you want a performance boost. There
will be an aggregator cache created and will be indexed. Although you
have not given a sorted input it will still work fine as it is indexed
in cache. Please note this is not like a page fault where you have to
bring the page from the secondary memory when not found in primary
memory.
Ans. Yes, you can load data from a pdf to a table using
UDO[Unstructured Data Option] Transformation in Informatica from PC811
onwards.
267.
Classification : Confidential
Classification : Confidential
Classification : Confidential
Classification : Confidential