Sie sind auf Seite 1von 46

LATEST INFORMATICA FAQ’S

1. The architectural difference between informatica 6.1 and 7.1.


Ans. Few differences or the enhancements in 7.1, which i have noticed are Versioning
in 7.1, Recovery, Can lookup flat files.

2. Can u tell how normalizer works with an example?

Ans. Normalization is used to transform non tabular data structure into tabular format mainly required for
COBOL data files.

3. How do we deliver our mapping to client after creating in our system?

Ans. Export your mapping to XML file and then Import it to Client machine. I am assuming here the client
and your system as same version and environment for Informatica.

4. How will you use surrogate key?

Ans. Surrogate keys are implemented by using the Sequence generator.

5. When i was running workflow manager i.e, when i right click the task icon and
clicking start workflow from task, i am getting a message like 'cannot connect to the
server'.
Can any body guide me, how to get the output.

Ans. First check out whether the Informatica service is running or not. Sometimes the server address will
change everytime.so it is suggestible to resolve the server (server-server configuration) and then start the
service. Also check out the repository server configuration...

6. In Source 1 to 100 numbers? How can we load the even number in One target
and odd number in another target? explain the mapping? How can we draw and
which transformations we r using? and explain logic?

Ans. Use a Router after the source , if the port have the number 1 to 1000 divided by 2
returns 0 then it belongs to Group 1 else Group 2. Send group 1 to one target, group 2 to 1
target

7. What is the difference between dynamic cache lookup and uncached lookup?

Ans. Uncached lookup: The informatica server doesn’t build a cache for temporary storage of data.
Whenever it needs to refer to the look up table, it scans the source directly.

Dynamic Cache: The informatica server builds a cache of the source when the workflow runs for the first
time and it updates the cache records dynamically after each row it loads to the target. Means, if a target
row gets updated, the dynamic cache also gets updated automatically after that particular row committed in
the target.

Classification : Confidential
8. What r two types of processes that informatica runs the session?

Ans. Load manager Process: Starts the session, creates the DTM process, and sends post-
session email when the session completes.
The DTM process. Creates threads to initialize the session, read, write, and transform data,
and handle pre- and post-session operations.

9. What r the new features of the server manager in the informatica 5.0?

Ans. U can use command line arguments for a session or batch. This allows U to change the
values of session parameters, and mapping parameters and mapping variables.

Parallel data processing: This feature is available for power center only. If we use the
informatica server on a SMP system, U can use multiple CPU’s to process a session
concurrently.

Process session data using threads: Informatica server runs the session in two processes.
Explained in previous question.

10. Which tool U use to create and manage sessions and batches and to monitor and
stop the informatica server?

Ans. Informatica server manager.

11. What is Session and Batches?

Ans. Session - A Session Is A set of instructions that tells the Informatica Server How And
When To Move Data From Sources To Targets. After creating the session, we can use either
the server manager or the command line program pmcmd to start or stop the session. Batches
- It Provides A Way to Group Sessions For Either Serial Or Parallel Execution By The Informatica Server.
There Are Two Types Of Batches:

Sequential - Run Session One after the Other. concurrent - Run Session At The Same Time.

12. What r the different options used to configure the sequential batches?

Ans.Two-options

Run the session only if previous session completes successfully. Always runs the session.

13. How can u recover the session in sequential batches?

Ans. If you configure a session in a sequential batch to stop on failure, you can run recovery
starting with the failed session. The Informatica Server completes the session and then runs
the rest of the batch. Use the Perform Recovery session property

To recover sessions in sequential batches configured to stop on failure:

1.In the Server Manager, open the session property sheet.


2.On the Log Files tab, select Perform Recovery, and click OK.
3.Run the session.
4.After the batch completes, open the session property sheet.
5.Clear Perform Recovery, and click OK.
If you do not clear Perform Recovery, the next time you run the session, the Informatica
Server attempts to recover the previous session.
If you do not configure a session in a sequential batch to stop on failure, and the remaining
sessions in the batch complete, recover the failed session as a standalone session.

Classification : Confidential
14. How to recover sessions in concurrent batches?

Ans. If multiple sessions in a concurrent batch fail, you might want to truncate all targets and
run the batch again. However, if a session in a concurrent batch fails and the rest of the
sessions complete successfully, you can recover the session as a standalone session.
To recover a session in a concurrent batch:
1.Copy the failed session using Operations-Copy Session.
2.Drag the copied session outside the batch to be a standalone session.
3.Follow the steps to recover a standalone session.
4.Delete the standalone copy.

15. How to recover the standalone session?

Ans. A standalone session is a session that is not nested in a batch. If a standalone session
fails, you can run recovery using a menu command or pmcmd. These options are not available
for batched sessions.

To recover sessions using the menu:


1. In the Server Manager, highlight the session you want to recover.
2. Select Server Requests-Stop from the menu.
3. With the failed session highlighted, select Server Requests-Start Session in Recovery Mode
from the menu.

To recover sessions using pmcmd:


1.From the command line, stop the session.
2. From the command line, start recovery.

16. How do we call shell scripts from informatica?

Ans. You can use a Command task to call the shell scripts, in the following ways:
1. Standalone Command task. You can use a Command task anywhere in the workflow or
worklet to run shell commands.
2. Pre- and post-session shell command. You can call a Command task as the pre- or post-
session shell command for a Session task. For more information about specifying pre-session
and post-session shell commands

17. How to load time dimension?

Ans. one way to do this by using Pl/Sql

18. When we create a target as flat file and source as oracle.. How can i specify first
rows as column names in flat files...

Ans. use a pre sql statement....but this is a hardcoding method...if you change the column names or put in
extra columns in the flat file, you will have to change the insert statement

19. How many types of facts and what are they?

Ans. there are Additive Facts, Semi-Additive, Non-Additive, Accumulating Facts,


Factless facts, Periodic fact table, Transaction Fact table.

Classification : Confidential
20.how do we do unit testing in informatica?
how do we load data in informatica ?

Ans. Unit testing are of two types

1. Quantitative testing

2.Qualitative testing

Steps.

1.First validate the mapping

2.Create session on the mapping and then run workflow.

Once the session is succeeded the right click on session and go for statistics tab.

There you can see how many number of source rows are applied and how many number of rows loaded in
to targets and how many number of rows rejected. This is called Quantitative testing.

If once rows are successfully loaded then we will go for qualitative testing.

Steps

1.Take the DATM(DATM means where all business rules are mentioned to the corresponding source
columns) and check whether the data is loaded according to the DATM in to target table. If any data is not
loaded according to the DATM then go and check in the code and rectify it.

This is called Qualitative testing.

This is what a developer will do in Unit Testing.

21. Why and where we are using factless fact table?

Ans. Factless Fact Tables are the fact tables with no facts or measures(numerical data). It contains only the
foreign keys of corresponding Dimensions.

22. What transformation you can use in place of lookup?...

Ans. You can use the joiner transformation by setting as outer join of either master or detail.

23. How can I get distinct values while mapping in Informatica in insertion?

Ans. You can add an aggregator before insert and group by the fields that need to be distinct.

IN the source qualifier write the query with distinct on key column

24. What is change data capture?

Ans. Changed Data Capture (CDC) helps identify the data in the source system that has
changed since the last extraction. With CDC, data extraction takes place at the same time
the insert, update, or delete operations occur in the source tables, and the change data is
stored inside the database in change tables. The change data, thus captured, is then made
available to the target systems in a controlled manner.

Classification : Confidential
25. How can we store previous session logs

Ans. Just run the session in time stamp mode then automatically session log will not overwrite current
session log.

We can do this way also. Using $PMSessionlogcount (specify the number of runs of the session log to
save)

26. How to define Informatica server?

Ans. informatica server is the main server component in informatica product family. Which is
responsible for reads the data from various source system and transforms the data according
to business rule and loads the data into the target table

27. How to move the mapping from one database to another?

Ans. u can also do it this way. connect to both the repositories, open the respective folders. keep the
destination repository as active. from the navigator panel just drag and drop the mapping to the work area. it
will ask whether to copy the mapping, say YES. its done.

28. my source is having 1000 rows. i have brought 300 records into my ODS. so next
time I want to load the remaining records. so i need to load from 301 th record.
when ever I start the work flow again it will load from the beginning. how do we
solve this problem.

Ans. 1)By using Sequence Generator Transformation u can do it

ie by changing the RESET option in the properties tab of your Sequence Generator Transformation.

2) You can use a filter transformation and set a condition there as rownum>300

3) You can also use variable to store the rownum of the final row you loaded in the target.
From next time you can use this variable to load the rest of the data to the target.

4) You can do a lookup on the target table and check for the rows already present there. Hence the first 300
records will not be reloaded to the target.

29. Is a fact table normalized or de-normalized?

Ans. A fact table is always DENORMALISED table. It consists of data from dimension table (Primary Key's)
and Fact table has Foreign keys and measures.

30. How can we join 3 database like Flat File, Oracle, Db2 in Informatrica..

Ans. Using Joiner Transformation.

31. How do we analyze the data at database level?

Ans. Data can be viewed using Informatica's designer tool.

If you want to view the data on source/target we can preview the data but with some limitations.

We can use data profiling too.

Classification : Confidential
32. how u will create header and footer in target using informatica?

Ans. If you are focus is about the flat files then one can set it in file properties while creating a mapping or at
the session level in session properties

33. How to export mappings to the production environment?

Ans. In the designer go to the main menu and one can see the export/import options.

Import the exported mapping in to the production repository with replace options.

34. How to Generate the Metadata Reports in Informatica?

Ans. You can generate Power Center Metadata Reporter from a browser on any workstation, even a
workstation that does not have Power Center tools installed.

35. how can we eliminate duplicate rows from flat file?

Ans. Before loading to target , use an aggregator transformation and make use of group by
function to eliminate the duplicates on columns.

Or u can use sorter transformation or Rank Transformation to eliminate duplicates from flat
files.

36. what is Partitioning ? where we can use Partition? what is advantages? Is it


necessary?

Ans. The Partitioning Option increases Power Center’s performance through parallel data
processing, and this option provides a thread-based architecture and automatic data
partitioning that optimizes parallel processing on multiprocessor and grid-based hardware
environments.

37.what is surrogate key ? In ur project in which situation u has used? explain with
example?

Ans. A surrogate key is system generated/artificial key /sequence number or A surrogate key
is a substitution for the natural primary key. It is just a unique identifier or number for each
row that can be used for the primary key to the table. The only requirement for a surrogate
primary key is that it is unique for each row in the table it is useful because the natural
primary key (i.e. Customer Number in Customer table) can change and this makes updates
more difficult. but In my project, I felt that the primary reason for the surrogate keys was to
record the changing context of the dimension attributes.(particularly for sod )The reason for
them being integer and integer joins are faster.

38. what is the diff b/w STOP & ABORT in INFORMATICA session level ?

Ans. Stop: We can Restart the session

Abort: WE cant restart the session. We should truncate all the pipeline after that start the session

Classification : Confidential
39. if the workflow has 5 session and running sequentially and 3rd session has been
failed how can we run again from only 3rd to 5th session?

Ans. If multiple sessions in a concurrent batch fail, you might want to truncate all targets and
run the batch again. However, if a session in a concurrent batch fails and the rest of the
sessions complete successfully, you can recover the session as a standalone session. To
recover a session in a concurrent batch:1.Copy the failed session using Operations-Copy
Session.2.Drag the copied session outside the batch to be a standalone session.3.Follow the
steps to recover a standalone session.4.Delete the standalone copy.

40. What is meant by EDW?

Ans. Its a big data warehouses OR centralized data warehousing OR the old style of warehouse.
Its a single enterprise data warehouse (EDW) with no associated data marts or operational data store (ODS)
systems.

41. how to load the data from people soft hrm to people soft erm using informatica?

Ans. Following are necessary

1.Power Connect license

2.Import the source and target from people soft using ODBC connections

3.Define connection under "Application Connection Browser" for the people soft source/target in workflow
manager .

select the proper connection (people soft with oracle,sybase,db2 and informix)

and execute like a normal session.

42. What is exact use of 'Online' and 'Offline' server connect Options while defining
Work flow in Work flow monitor? . The system hangs when 'Online' server connect
option. The Informatica is installed on a Personal laptop.

Ans. When the repo is up and the PMSERVER is also up, workflow monitor always will be connected on-
line.

When PMserver is down and the repo is still up we will be prompted for an off-line connection with which we
can just monitor the workflows

previously ran.

43. in real-time which one is better star schema or snowflake star schema the
surrogate will be linked to which columns in the dimension table.

Ans. In real time only star schema will implement because it will take less time and surrogate key will there
in each and every dimension table in star schema and this surrogate key will assign as foreign key in fact
table.

Classification : Confidential
44. what is the mapping for unit testing in Informatica, are there any other testing in
Informatica, and how we will do them as a etl developer. How do the testing people
will do testing are there any specific tools for testing.

Ans. In informatica there is no method for unit testing. There are two methods to test the mapping.

1. But we have data sampling. set the data sampling properties for session in workflow manager for
specified number of rows and test the mapping.

2. Use the debugger and test the mapping for sample records.

45. What are the properties should be notified when we connect the flat file source
definition to relational database target definition?

Ans. 1.File is fixed width or delimited

2.Size of the file.

If its can be executed without performance issues then normal load will work

If its huge in GB they NWAY partitions can be specified at the source side and the target side.

3.File reader, source file name etc .....

46. can batches be copied/stopped from server manager?

Ans. yes, we can stop the batches using server manager or pmcmd command.

47. what are the real time problems generally come up while doing/running
mapping/any transformation?

Ans. may be you will encounter with connection failure, other then that i don't think so,
conserve will handle all the syntax errors and Invalid mappings.

48. What will happen if you are using Update Strategy Transformation and your
session is configured for "insert"?

Ans. if u r using a update strategy in any of ur mapping, then in session properties u have to set treat source
rows as Data Driven. if u select insert or update or delete, then the info server will not consider UPD for
performing any DB operations.

ELSE

u can use the UPD session level options. instead of using a UPD in mapping just select the update in treat
source rows and update else insert option. this will do the same job as UPD. but be sure to have a PK in the
target table.

49. how to get two targets T1 containing distinct values and T2 containing duplicate
values from one source S1.

Ans. Use filter transformation for loading the target with no duplicates. and for the other
transformation load it directly from source.

Classification : Confidential
50. Where is the cache stored in informatica?

Ans. Informatica Server

51. In a joiner transformation, you should specify the source with fewer rows as the
master source. Why?

Ans. Joiner transformation compares each row of the master source against the detail source. The fewer
unique rows in the master, the fewer iterations of the join comparison occur, which speeds the join process.

52. 1)What are the various test procedures used to check whether the data is loaded
in the backend, performance of the mapping, and quality of the data loaded in
INFORMATICA.

Ans. Some of the steps could be :

1) Check in the workflow monitor status, whether no. of records in source and no. of actual records loaded
are equal

2) Check for the duration for a workflow to succeed

3)Check in the session logs for data loaded.

53. what is polling?

Ans. It displays the updated information about the session in the monitor window. The monitor
window displays the status of each session when you poll the Informatica server.

54. Partitioning, Bitmap Indexing (when to use), how will the bitmap indexing will
effect the performance.

Ans. Bitmap indexing a indexing technique to tune the performance of SQL queries. The
default type is B-Tree indexers which is of high cardinality (normalized data). You can use
bitmap indexers for de-normalized data or low cardinalities. The condition is the amount of
DISTINCT rows should be less than 4% of the total rows. If it satisfies the given condition then
bitmap indexers will optimize the performance for this kind of tables.

55. where do we use MQ series source qualifier, application multi group source
qualifier. just give an example for a better understanding.

Ans. We can use a MQSeries SQ when we have a MQ messaging system as source (queue).

When there is need to extract data from a Queue, which will basically have messages in XML format, we will
use a JMS or a MQ SQ depending on the messaging system. If you have a TIBCO EMS Queue, use a JMS
source and JMS SQ and an XML Parser, or if you have a MQ series queue, then use a MQ SQ which will be
associated with a Flat file or a Cobol file.

56. What is meant by Junk Attribute in Informatica?

Ans. Junk Dimension a Dimension is called junk dimension if it contains attribute which are
rarely changed or modified. example In Banking Domain , we can fetch four attributes
accounting to a junk dimensions like from the Overall_Transaction_master table tput flag tcmp
flag del flag advance flag all these attributes can be a part of a junk dimensions.

Classification : Confidential
57. Explain about incremental aggregation with an example?

Ans. Incremental aggregation is specially used for tune the performance of the aggregator. It
captures the change each time (incrementally) you run the transformation and then performs
the aggregation function to the changed rows and not to the entire rows. This improves the
performance because you are not reading the entire source, each time you run the session.

58. Difference between Rank and Dense Rank?

Ans. Rank:
1
2<--2nd position
2<--3rd position
4
5

Same Rank is assigned to same totals/numbers. Rank is followed by the Position. Golf game ususally
Ranks this way. This is usually a Gold Ranking.

Dense Rank:
1
2<--2nd position
2<--3rd position
3
4

Same ranks are assigned to same totals/numbers/names. the next rank follows the serial number.

59. What are the hierarchies in DWH?

Ans. Data sources ---> Data acquisition ---> Warehouse ---> Front end tools ---> Metadata
management ---> Data warehouse operation management

60. Explain real time complain mappings or complex transformations in Informatica.

Ans. Most complex logic we use is de-normalization. We don’t have any De-normalizer transformation in
INformatica. So we will have to use an aggregator followed by an expression. Apart from this, we use most
of the complexicity in expression transformation involving lot of nested IIF's and Decode
statements...another one is the union transformation and joiner.

61. In the source, if we also have duplicate records and we have 2 targets, T1- for
unique values and T2- only for duplicate values. How do we pass the unique values
to T1 and duplicate values to T2 from the source to these 2 different targets in a
single mapping?

Ans. use this sequence to get the result.

source--->sq--->exp-->sorter(with enable select distinct check box)--->t1

--->aggregator(with enabling group by and write count

function)--->t2

Classification : Confidential
62. what are the enhancements made to Informatica 7.1.1 version when compared to
6.2.2 version?

Ans. 1.union & custom transformation 2.Lookup on flat file 3. using PMCMD command.

4.Exporting independent and dependent repository objects 5. Version controlling 6.Data profiling
7.Supporting of 64MB Architecture. 8.Idap Authentication.

63. What is the difference between Power Centre and Power Mart?

Ans. power center power mart

no. of repository n No. n No.

applicability high end WH low & mid range WH

global repository supported not supported

local repository supported supported

ERP support available not available

64. What is the logic will you implement to load the data in to one fact from 'n'
number of dimension tables?

Ans. To load data into one fact table from more than one dimension tables . firstly u need to
create a fact table and dimension tables. later load data into individual dimensions by using
sources and transformations(aggregator, sequence generator, lookup) in mapping designer
then to the fact table connect the surrogate to the foreign key and the columns from
dimensions to the fact.

65. Can i use a session Bulk loading option that time can i make a recovery to the
session?

Ans. If the session is configured to use in bulk mode it will not write recovery information to
recovery tables. So Bulk loading will not perform the recovery as required.

66. How do you configure mapping in informatica.

Ans. You should configure the mapping with the least number of transformations and
expressions to do the most amount of work possible. You should minimize the amount of data
moved by deleting unnecessary links between transformations.

For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup transformations),
limit connected input/output or output ports. Limiting the number of connected input/output or output ports
reduces the amount of data the transformations store in the data cache.

You can also perform the following tasks to optimize the mapping:

• Configure single-pass reading.


• Optimize data type conversions.
• Eliminate transformation errors.
• Optimize transformations.

Classification : Confidential
• Optimize expressions. You should configure the mapping with the least number of
transformations and expressions to do the most amount of work possible. You should
minimize the amount of data moved by deleting unnecessary links between
transformations.

67. What are mapping parameters and variables in which situation we can use it?

Ans. If we need to change certain attributes of a mapping after every time the session is run, it will be very
difficult to edit the mapping and then change the attribute. So we use mapping parameters and variables
and define the values in a parameter file. Then we could edit the parameter file to change the attribute
values. This makes the process simple.

68. What is meant by complex mapping?

Ans. complex mapping means having more business rules.

69. If a session fails after loading of 10,000 records in to the target. How can u load
the records from 10001 th record when u run the session next time in informatica
6.1?

Ans. Running the session in recovery mode will work, but the target load type should be normal. If its bulk
then recovery won’t work as expected.

70. Can we run a group of sessions without using workflow manager?

Ans. Using PMCMD command.

71. If u had to split the source level key going into two separate tables. One as
surrogate and other as primary. Since informatica does not guarantee keys are
loaded properly (order!) into those tables. What are the different ways you could
handle this type of situation?

Ans. Foreign Keys

72. What is data merging, data cleansing, and sampling?

Ans. data merging: it is process of integrated multiple Input sources


into single output with similar structure and data type.
Data cleansing: to eliminate the inconsistent data
Sampling: it is the process, orbitarly reading the data from
group of records.

73. How do I import VSAM files from source to target? Do I need a special plug-in?

Ans. using power exchange tool convert VSAM file to oracle tables then do mapping as usual
to the target table.

74. If you have four lookup tables in the workflow. How do you troubleshoot to
improve performance?

Ans. there r many ways to improve the mapping which has multiple lookups.

1) We can create an index for the lookup table if we have permissions (staging area).

Classification : Confidential
2) Divide the lookup mapping into two (a) dedicate one for insert means: source - target,, these r new rows
. only the new rows will come to mapping and the process will be fast. (b) Dedicate the second one to
update: source=target, these r existing rows. Only the row which exists already will come into the mapping.

3) We can increase the cache size of the lookup.

75. How to retrieve the records from a rejected file. Explain with syntax or example.

Ans. there is one utility called "reject Loader" where we can find out the reject records. And
able to refine and reload the rejected records...

76. Which objects are required by the debugger to create a valid debug session?

Ans. We can create a valid debug session even without a single break-point. But we have to give valid
database connection details for sources, targets, and lookups used in the mapping and it should contain
valid mapplets (if any in the mapping).

77. Suppose session is configured with commit interval of 10,000 rows and source
has 50,000 rows. Explain the commit points for Source based commit and Target
based commit. Assume appropriate value wherever required.

Ans. Source based commit will commit the data into target based on commit interval. so, for every 10,000
rows it will commit into target.

Target based commit will commit the data into target based on buffer size of the target. i.e., it commits the
data into target when ever the buffer fills. Let us assume that the buffer size is 6, 000.So, for every 6,000
rows it commits the data.

78. Can Informatica be used as a Cleansing Tool? If Yes, give example of


transformations that can implement a data cleansing routine

Ans. Yes, we can use Informatica for cleansing data. Some time we use stages to cleansing the data. It
depends upon performance again else we can use expression to cleansing data.

For example an field X have some values and other with Null values and assigned to target field where
target field is not null column, inside an expression we can assign space or some constant value to avoid
session failure.

The input data is in one format and target is in another format, we can change the format in expression.

We can assign some default values to the target to represent complete set of data in the target.

79. Briefly explain the Versioning Concept in Power Center 7.1.

Ans. When you create a version of a folder referenced by shortcuts, all shortcuts continue
to reference their original object in the original version. They do not automatically update
to the current folder version.

For example, if you have a shortcut to a source definition in the Marketing folder, version
1.0.0, then you create a new folder version, 1.5.0, the shortcut continues to point to the
source definition in version 1.0.0.

Classification : Confidential
Maintaining versions of shared folders can result in shortcuts pointing to different
versions of the folder. Though shortcuts to different versions do not affect the server, they
might prove more difficult to maintain. To avoid this, you can recreate shortcuts pointing
to earlier versions, but this solution is not practical for much-used objects. Therefore,
when possible, do not version folders referenced by shortcuts.

80. How to create the staging area in your database.

Ans. A Staging area in a DW is used as a temporary space to hold all the records from the source system.
So more or less it should be exact replica of the source systems except for the load strategy where we use
truncate and reload options.

So create using the same layout as in your source tables or using the Generate SQL option in the
Warehouse Designer tab.

81. Compare Data Warehousing Top-Down approach with Bottom-up approach.

Ans. top down

ODS-->ETL-->Data warehouse-->Data mart-->OLAP

Bottom up

ODS-->ETL-->Data mart-->Data warehouse-->OLAP

82. What is the difference between summary filter and detail filter?

Ans. summary filter can be applied on a group of rows that contain a common value. where as
detail filters can be applied on each and every rec. of the data base.

83. What are the difference between view and materialized view?

Ans. A view do no derive the change made to it master table after the view is created.

A materialized view immediately carries the change done to its master table even after the materialized view
is created.

A materialized view is a database object that contains the results of a query. They
are local copies of data located remotely, or are used to create summary tables
based on aggregations of a table's data. Materialized views, which store data based
on remote tables are also, know as snapshots.

A materialized view can query tables, views, and other materialized views.
Collectively these are called master tables (a replication term) or detail tables (a
data warehouse term).

84. What is a time dimension? Give an example.

Ans. Time dimension is one of important in Data warehouse. Whenever u generated the
report, that time u access all data from thro time dimension.

eg. employee time dimension

Classification : Confidential
Fields : Date key, full date, day of week, day , month, quarter, fiscal year

85. What r the tasks that Load manger process will do?

Ans. Manages the session and batch scheduling, locking and reading the session, reading the parameter
file, verifies permissions and privileges, creating log files.

86. To provide support for Mainframes source data, which files r used as a source
definitions?

Ans. COBOL Copy-book files.

87. How the informatica server sorts the string values in Rank transformation?

Ans. When Informatica Server runs in UNICODE data movement mode, then it uses the
sort order configured in session properties.

88. What is the Rank-index in Rank transformation?

Ans. Based on which port you want generate Rank is known as rank port, the generated
values are known as rank index.

89. What is difference between partitioning of relational target and partitioning of file
targets?

Ans. If u partitions a session with a relational target informatica server creates multiple
connections to the target database to write target data concurrently. If u partition a session
with a file target the informatica server creates one target file for each partition can configure
session properties to merge these target files.

90. What is active data warehousing?

Ans. An active data warehouse provides information that enables decision-makers within an organization
to manage customer relationships nimbly, efficiently and proactively. Active data warehousing is all about
integrating advanced decision support with day-to-day-even minute-to-minute-decision making in a way that
increases quality of those customer touches which encourages customer loyalty and thus secure an
organization's bottom line. The marketplace is coming of age as we progress from first-generation "passive"
decision-support systems to current- and next-generation "active" data warehouse implementations.

91. what are the advantages data mining over traditional approaches?

Ans. Data Mining is used for the estimation of future. For example, if we take a company/business
organization, by using the concept of Data Mining, we can predict the future of business in terms of Revenue
(or) Employees (or) Customers (or) Orders etc.

Traditional approaches use simple algorithms for estimating the future. But, it does not give accurate results
when compared to Data Mining.

Classification : Confidential
92. What is the main difference between Inmon and Kimball philosophies of data
warehousing?

Ans. RalfKimball: he follows bottom-up approach i.e., first create individual Data Marts from the existing
sources and then create Data Warehouse.

BillImmon: he follows top-down approach i.e., first create Data Warehouse from the existing sources and
then create individual Data Marts.

93. What is junk dimension?


What is the difference between junk dimension and degenerated dimension?

Ans. Junk dimension: Grouping of Random flags and text attributes in a dimension and moving
them to a separate sub dimension.

Degenerate Dimension: Keeping the control information on Fact table ex: Consider a
Dimension table with fields like order number and order line number and have 1:1 relationship
with Fact table, In this case this dimension is removed and the order information will be
directly stored in a Fact table in order eliminate unnecessary joins while retrieving order
information..

94. What is Difference between E-R Modeling and Dimensional Modeling?

Ans. Basic diff is E-R modeling will have logical and physical model. Dimensional
model will have only physical model.

E-R modeling is used for normalizing the OLTP database design.

Dimensional modeling is used for de-normalizing the ROLAP/MOLAP design.

95. What is BUS Schema?

Ans. A BUS Schema or a BUS Matrix? A BUS Matrix (in Kimball approach) is to identify
common Dimensions across Business Processes; ie: a way of identifying Conforming
Dimensions.

96. What are the possible data marts in Retail sales?

Ans. Product information, sales information

97. What is incremental loading?

Ans. Incremental loading means loading the ongoing changes in the OLTP. Aggregate table contains the
[measure] values, aggregated /grouped/summed up to some level of hierarchy.

98. What is real time data-warehousing?

Ans. Real-time data warehousing is a combination of two things: 1) real-time activity and 2)
data warehousing. Real-time activity is activity that is happening right now. The activity could
be anything such as the sale of widgets. Once the activity is complete, there is data about it.

Data warehousing captures business activity data. Real-time data warehousing captures
business activity data as it occurs. As soon as the business activity is complete and there is
data about it, the completed activity data flows into the data warehouse and becomes

Classification : Confidential
available instantly. In other words, real-time data warehousing is a framework for deriving
information from data as the data becomes available.

99. What does level of Granularity of a fact table signify?

Ans. level of granularity defines the extent of detail. As an example, let us look at
geographical level of granularity. We may analyze data at the levels of COUNTRY, REGION,
TERRITORY, CITY and STREET. In this case, we say the highest level of granularity is STREET.

100. What is degenerate dimension table?

Ans. Degenerate Dimensions: If a table contains the values, which r neither dimension nor measures is
called degenerate dimensions. Ex : invoice-id, empno .

101. What are the Different methods of loading Dimension tables?

Ans. Conventional Load:


Before loading the data, all the Table constraints will be checked against the data.

Direct load :( Faster Loading)


All the Constraints will be disabled. Data will be loaded directly. Later the data will be checked
against the table constraints and the bad data won't be indexed.

102. Where do we use semi and non additive facts?

Ans. Additive: A measure can participate arithmetic calculations using all or any dimensions.

Ex: Sales profit

Semi additive: A measure can participate arithmetic calculations using some dimensions.

Ex: Sales amount

Non Additive: A measure can’t participate arithmetic calculations using dimensions.

Ex: temperature

103. What is ODS (operation data source).

ANS. An updatable set of integrated operational data used for enterprise- wide tactical decision making.
Contains live data, not snapshots, and has minimal history retained.

104. What is Full load & Incremental or Refresh load?

Ans. Full Load is the entire data dump load taking place the very first time.
Gradually to synchronize the target data with source data, there are further 2 techniques:-
Refresh load - Where the existing data is truncated and reloaded completely.
Incremental - Where delta or difference between target and source data is dumped at regular
intervals. Timestamp for previous delta load has to be maintained.

105. What is Status code?

Ans. Status codes provide error handling for the Integration Service during a workflow.

Classification : Confidential
106. In a sequential batch how can i stop/start the particular session?

Ans. U can do in 2 ways...

1 st up all if u r working in windows environment. its easy u can just click on particular task on that batch and
make it run ... be sure that remaining all are in suspend mode

If u are working in UNIX environment change status of each session using pmcmd commands.

107. What is Test Load?

Ans. With a test load, the Integration Service reads and transforms data without writing
to targets. The Integration Service generates all session files, and performs all pre- and
post-session functions, as if running the full session.
The Integration Service writes data to relational targets, but rolls back the data when the
session completes. For all other target types, such as flat file and SAP BW, the
Integration Service does not write data to the targets.

108. How the Informatica Server will identify the string as date?

Ans. With the Sysdate.

109. What is external loading?

Ans. External loaders can increase session performance by loading information directly from a file or pipe
rather than running the SQL commands to insert the same data into the database.

110. What is bulk loading?

Ans. Bulk loading improves the performance of a session that inserts a large amount of data to the target
database.

111. What is commit Interval?

Ans. When bulk loading to Microsoft SQL Server or Oracle targets, define a large commit interval to reduce
the number of bulk load transactions and increase performance.

112. What is Normal Load?

Ans. The Integration Service loads targets normally.

113. What is Incremental Aggregation?

Ans. If the source changes incrementally and you can capture changes, you can configure the session to
process those changes. This allows the Integration Service to update the target incrementally, rather than
forcing it to process the entire source and recalculate the same data each time you run the session.

114. What is Grid?

Ans. A grid is an alias assigned to a group of nodes that run sessions and workflows. When you run a
workflow on a grid, you improve scalability and performance by distributing Session and Command tasks to
service processes running on nodes in the grid.

Classification : Confidential
115. What r the tasks done by DTM?

Ans. The DTM process the following Tasks given below.

• 1. Retrieves and validates session information from the repository.


• Creates the session log.
• Runs pre-session shell commands, stored procedures and SQL.
• Runs post-session stored procedures, SQL, and shell commands.
• Sends post-session email.
• Adds partitions to the session when the session is configured for dynamic
partitioning.

116. What is Cache partitioning?

Ans. When you create a session with multiple partitions, the Integration Service creates
cache for the Aggregator, Joiner, Lookup, and Rank transformations. It creates a separate
cache for each partition, and each partition works with the rows needed by that partition.
As a result, the Integration Service requires a portion of total cache memory for each
partition. When you run a session, the Integration Service accesses the cache in parallel
for each partition.

117. What is Dynamic Partitioning?

Ans. If the volume of data grows or you add more CPUs, you might need to adjust
partitioning so the session run time does not increase. When you use dynamic
partitioning, you can configure the partition information so the Integration Service
determines the number of partitions to create at run time.

118. What is Database Partitioning?

Ans. You can optimize session performance by using the database partitioning partition
type for source and target databases. When you use source database partitioning, the
Integration Service queries the database system for table partition information and fetches
data into the session partitions. When you use target database partitioning, the Integration
Service loads data into corresponding database partition nodes.

119. What are the different types of Commits?

Ans. Target-based commit, Source-based commit, User-defined commits

120. What is a control files?

Ans. These r nothing but indicator files.

Indicator files are two types: a) ROW wise Indicator file b) COLUMN wise Indicator file.

Row wise indicator files are: Insert, Update etc.

Column wise indicator files are: Null, Delete etc.

Classification : Confidential
121. What is Push down Optimization?

Ans. When you run a session configured for pushdown optimization, the Integration Service analyzes the
mapping and writes one or more SQL statements based on the mapping transformation logic.

122. What is XML Parser Transformation?

Ans. The XML Parser transformation is similar to an XML source definition. When the Integration Service
processes an XML Parser transformation, it reads a row of XML data, parses the XML, and passes data
through output groups.

123. What is XML Generator Transformation?

Ans. Use an XML Generator transformation to combine input that comes from several sources to create an
XML document.

124. What is the repository Server?

Ans. Stores the information, or metadata used by informatica Server and Client tools.

125. What is the Metadata?

Ans. Metadata describes different types of objects such as mappings, or transformations, that u can create
or modify using informatica client tools.

126. What is Informatica Server?

Ans. It uses all the repository objects to extract, transform and load data.

127. I tried to view the session log, but it contains garbage characters.

Ans. Enable the Session Log in UTF-8 option on the Configuration tab of the Informatica
Server setup program.

128. I installed and registered the Informatica Server, but I get an internal error when I try to start the
service.

Ans. Verify that the Repository Server is running. You must start the Repository Server
before you can start the Informatica Server.

129. I started the Informatica Server, but it does not connect to the repository.

Ans. Verify that the Informatica Server and the system on which you installed the
Informatica Server are configured to connect to the Repository Server. Also verify that
the Repository Server is running.

130. The Informatica Server does not start after rebooting the machine.

Ans. In the Services configuration, make sure the Startup Type is set to automatic.

Classification : Confidential
131.What r the different types of threads available in DTM?

Ans. Pre and post sessions, writer, reader, transformation, and mapping threads.

132. Where will the informatica Server creates the reject files.

Ans. $PMBadFileDir server variable directory

133. I have a section in a parameter file for a session, but the Informatica Server
does not seem to read it.

Ans. In the parameter file, folder and session names are case-sensitive. Make sure to enter
folder and session names exactly as they appear in the Workflow Manager. Also, use the
appropriate prefix for all user-defined session parameters.

134. I am trying to run a workflow with a parameter file and one of the sessions
keeps failing.

Ans. The session might contain a parameter that is not listed in the parameter file. The
Informatica Server uses the parameter file to start all sessions in the workflow. Check the
session properties, and then verify that all session parameters are defined correctly in the
parameter file.

135. I am trying to set up a session to load data to an external loader, but I cannot
select an external loader connection in the session properties.

Ans. Check your mapping to make sure you did not configure it to load to a flat file target.
In order to use an external loader, you must configure the mapping to load to a DB2,
Oracle, Sybase IQ, or Teradata relational target. When you create the session, select a file
writer in the Writers settings of the Targets tab in the session properties. Then open the
Connections settings and select an external loader connection.

136. What r the out put files that the informatica server creates during the session
running?

Ans. Indicator file, output file, performance detail log, session log, session details,
workflow log, workflow details, control file, reject files, cache files.

137. What is Indicator File?

Ans. If you use a flat file as a target, you can configure the Informatica Server to create
an indicator file for target row type information. For each target row, the indicator file
contains a number to indicate whether the row was marked for insert, update, delete, or
reject.

Classification : Confidential
138. What is control File?

Ans. When you run a session that uses an external loader, the Informatica Server creates
a control file and a target flat file. The control file contains information about the target
flat file such as data format and loading instructions for the external loader.

139. What is output File?

Ans. If the session writes to a target file, the Informatica Server creates the target file
based on file target definition. By default, the Informatica Server names the target file
based on the target table: target_table.out.

140. What is the configuration management in informatica?

Ans. Configuration management consists of controlling an enterprise's intellectual


property - data, processes, people, and the relationships between them - for manufactured
goods, hardware, software, or any industry requiring traceability.

141. Techniques of Error Handling - Ignore, Rejecting bad records to a flat file,
loading the records and reviewing them (default values)

Ans. Rejection of records either at the database due to constraint key violation or the
informatica server when writing data into target table. These rejected records we can find in
the bad files folder where a reject file will be created for a session. we can check why a record
has been rejected. And this bad file contains first column a row indicator and second column a
column indicator.
These row indicators or of four types
D-valid data,
O-overflowed data,
N-null data,
T- Truncated data,
And depending on these indicators we can changes to load data successfully to target.

142.What are the various methods of getting incremental records or delta records
from the source systems?

Ans. getting incremental records from source systems to target can be done
by using incremental aggregation transformation

143. What is a three tier data warehouse?

Ans. The 3 tiers are:


1. Data tier - bottom tier - consists of the database
2. Application tier - middle tier - consists of the analytical server
3. Presentation tier - tier that interacts with the end-user

Classification : Confidential
144. What are the modules in Power Mart?

Ans. 1. Power Mart Designer


2. Server
3. Server Manager
4. Repository
5. Repository Manager

145. Can Informatica load heterogeneous targets from heterogeneous sources?

Ans. Yes

146. What is Code Page used for?

Ans. Code Page is used to identify characters that might be in different languages. If you are
importing Japanese data into mapping, u must select the Japanese code page of source data.

147. What r the transformations that restricts the partitioning of sessions?

Ans. Advanced External procedure transformation and external procedure


transformation: This transformation contains a check box on the properties tab to all
partitioning.

Aggregator Transformation: If u use sorted ports u cannot partition the associated source.

Joiner Transformation: U cannot partition the master source for a joiner transformation.

Normalizer transformation and XML targets.

148.What is difference between stored procedure transformation and external


procedure transformation?

Ans. In case of stored procedure transformation procedure will be compiled and executed in a
relational data source need data base connection to import the stored procedure in to u’r
maping.Where as in external procedure transformation procedure or function will be executed
out side of data source.Ie u need to make it as a DLL to access in u r maping.No need to have
data base connection in case of external procedure transformation.

149. Explain the informatica Architecture in detail

Ans. informatica server connects source data and target data using native

ODBC drivers

Again it connect to the repository for running sessions and retrieving metadata information

Source------>informatica server--------->target

Repository REPOSITORY Server Repository Admin. Console

Classification : Confidential
150. What is the difference between connected and unconnected stored procedures?

Ans.

Run a stored procedure before or after your session. Unconnected

Unconnected
Run a stored procedure once during your mapping, such as pre- or post-session.

Connected or
Run a stored procedure every time a row passes through the Stored Procedure transformation. Unconnected

Run a stored procedure based on data that passes through the mapping, such as when a specific Unconnected
port does not contain a null value.

Connected or
Pass parameters to the stored procedure and receive a single output parameter. Unconnected

Pass parameters to the stored procedure and receive multiple output parameters.
Connected or
Note: To get multiple output parameters from an unconnected Stored Procedure transformation, Unconnected
you must create variables for each output parameter. For details, see Calling a Stored Procedure
From an Expression.

Run nested stored procedures. Unconnected

Unconnected
Call multiple times within a mapping.

151. Suppose session is configured with commit interval of 10,000 rows and source has 50,000 rows. Explain the
commit points for Source based commit and Target based commit. Assume appropriate value wherever
required.

Ans. Source based commit will commit the data into target based on commit interval.so, for
every 10,000 rows it will commit into target.

Target based commit will commit the data into target based on buffer size of the target.i.e; it
commits the data into target when ever the buffer fills. Let us assume that the buffer size is
6,000.So,for every 6,000 rows it commits the data.

152. Which objects are required by the debugger to create a valid debug session?

Ans. We can create a valid debug session even without a single break-point. But we have
to give valid database connection details for sources, targets, and lookups used in the
mapping and it should contain valid mapplets (if any in the mapping).

153. How does the server recognize the source and target databases?

Ans. By using ODBC connection. if it is relational. if is flat file FTP connection. See we can
make sure with connection in the properties of session both sources & targets.

Classification : Confidential
154. How to lookup the data on multiple tables.

Ans. When you create lookup transformation that time INFA asks for table name so you can choose source,
target, import and skip. So click skip and the use the sql override property in properties tab to join two tables
for lookup.

155. How do you handle decimal places while importing a flat file into informatica?

Ans. while importing flat file definition just specify the scale for a numeric data type. in the
mapping, the flat file source supports only number data type(no decimal and integer). In the
SQ associated with that source will have a data type as decimal for that number port of the
source?

Source ->number data type port ->SQ -> decimal datatype.Integer is not supported. Hence
decimal is taken care.

156. What are partition points?

Ans. Partition points mark the thread boundaries in a source pipeline and divide the pipeline into stages.

157. What is the gap analysis?

Ans. 1. BRD (Business Requirement Document)-BA


2. SSSD (Source System Study Document)-BA

BRD consists of the requirements of the client.


SSSD consists of the source system study.
The source that does not meet the requirements specified in BRD using the source given in
the SSSD is treated as gap analysis. Or in one word the difference between 1 and 2 is called
gap analysis.

158. How to create a custom transformation, can u give a real-time example where
exactly u has used it. Just give some explanation why u used the custom
transformation.

Ans. By using Custom transformation we create the single input ports and multiple output ports
or multiple input ports and single input otherwise multiple input or output ports. When we
clicking the custom transformation one window will opens we have select active custom or
passive custom transformation, here one thing this is one transformation it acts like active and
passive.

159. What happens if you try to create a shortcut to a non-shared folder?

Ans. It only creates a copy of it.

160. How can we join the tables if the tables have no primary and foreign key relation
and no matching port to join?

Ans. without common column or common data type we can join two sources using dummy
ports.

1. Add one dummy port in two sources.

2. In the expression transformation assign ‘1’ to each port.

Classification : Confidential
3. Use Joiner transformation to join the sources using dummy port (Use join conditions).

161. 1] In certain mapping there are four targets tg1, tg2, tg3 and tg4.
Tg1 has a primary key, tg2 foreign key referencing the tg1's primary key, tg3 has
primary key that tg2 and tg4 refers as foreign key, tg2 has foreign key referencing
primary key of tg4 ,the order in which the informatica will load the target?
2] How can I detect aggregate transformation causing low performance?

Ans. T1 and T3 are being the master table and don't have any foreign key reference to other
table will be loaded first. Then T3 will be loaded as it's master table T3 is already been loaded.
and at the end T2 will be loaded as it's all master table T1, T3, and T2 to whtch it refers has
been already loaded.

To optimize the aggregator transformation, you can use the following options.

• Use incremental aggregation


• Sort the ports before you perform aggregation
• Avoid using aggregator transformation after update strategy, since it might be
confusing.

162. How is the union transformation active transformation?

Ans. Union Transformation: in Union Transformation, we may combine the data from two (or)
more sources. Assume, Table-1 contains '10' rows and Table-2 contains '20' rows. If we
combine the rows of Table-1 and Table-2, we will get a total of '30' rows in the Target. So, it is
definitely an Active Transformation.

163. How can we store previous session logs?

Ans. Just run the session in time stamp mode then automatically session log will not overwrite
current session log. (OR)

Go to Session-->right click -->Select Edit Task then Go to -->Config. Object then set the
property

Save Session Log By --Runs


Save Session Log for These Runs --->To Number of Historical Session logs you want.

164. How do you create single lookup transformation using multiple tables?

Ans. 1.Write a override sql query. Adjust the ports as per the sql query.
2. For connected lkp transformation1>create the lkp transformation.2>go for
skip.3>manually enter the ports name that u want to lookup.4>connect with the i/p port
from src table.5>give the condition6>go for generate sql then modify according to u'r
requirement validate it will work....
3. Just we can create the view by using two table then we can take that view as
lookup table

Classification : Confidential
165. What is the architecture of any Data warehousing project? What is the flow?

Ans. 1) The basic step of data warehousing starts with data modeling. i.e creation dimensions and facts.

2) Data warehouse starts with collection of data from source systems such as OLTP,CRM,ERPs etc

3) Cleansing and transformation process is done with ETL(Extraction Transformation


Loading) tool.

4)by the end of ETL process target databases(dimensions, facts) are ready with data which
accomplishes the business rules.

5) Now finally with the use of Reporting tools(OLAP) we can get the information which is used
for decision support.

166. In workflow can we send multiple emails?

Ans. Yes we can do it in Unix Server but not in Windows based server.

167. What about rapidly changing dimensions? Can u analyze with an example?

Ans. rapidly changing dimensions are those in which values are changing continuously and
giving lot of difficulty to maintain them.

i am giving one of the best real world example which i found in some website while browsing.
Go through it. i am sure you like it.

Description of a rapidly changing dimension by that person:

I'm trying to model a retailing case. I'm having a SKU dimension


of around 150,000 unique products which is already a SCD Type 2 for
some attributes. In addition I'm willing to track changes of the sales
and purchase price. However these prices change almost daily for quite
a lot of these products leading to a huge dimensional table and requiring continuous
updations.

So a better option would be shift those attributes into a fact table as facts, which solves the
problem.

168. What is the repository agent?

Ans. The Repository Agent is a multi-threaded process that fetches, inserts, and updates
metadata in the repository database tables. The Repository Agent uses object locking to
ensure the consistency of metadata in the repository.

169. What is bottleneck in informatica?

Ans. Bottleneck in Informatica

Bottleneck in ETL Processing is the point, by which the performance of the ETL Process is
slower.

Classification : Confidential
When ETL Process is in progress, first thing login to workflow monitor and observe
performance statistic. I.e. observe processing rows per second. In SSIS and Data stage,
when you run the job, you can see at every level how many rows per second is processed
by the server.

Mostly bottleneck occurs at source qualifier – during fetching data from source, joiner,
aggregator, Lookup – Cache Building Session.

Removing bottleneck is performance tuning.

170. Where persistent cache will be stored?

Ans. The informatica Server saves the cache files for every session and reuses for the next session, by that
the query on the table will be reduced, so there will be some performance increment.

171. What is the event-based scheduling?

Ans. In time based scheduling, the jobs run at the specified time. In some situations, we've to
run a job based on some events like if a file arrives then only the job has to run whatever the
time it is. In such cases, event based scheduling is used.

172. What is the new lookup port in look-up transformation and explain?

Ans. Sometimes you need to create a generated key for a column in the target table. For
lookup ports with an Integer or Small Integer data type, you can associate a generated key
instead of an input port. To do this, select Sequence-ID in the Associated Port column.

When you select Sequence-ID in the Associated Port column, the Informatica Server
generates a key when it inserts a row into the lookup cache. Map the lookup/output ports
to the target to ensure that the lookup cache and target are synchronized.

173. What TOAD and for what purpose it will be used?

Ans. Toad is an application development tool built around an advanced SQL and PL/SQL editor.
Using Toad, you can build and test PL/SQL packages, procedures, triggers, and functions. You
can create and edit database tables, views, indexes, constraints, and users. The Schema
Browser and Project Manager provide quick access to database objects.

174. How we validate all the mappings in the repository at once?

Ans. No, it’s not possible to validate all the mappings in the repository at once.

175. What is referential Integrity error? How will u rectify it?

Ans. "referential Integrity" is all about foreign key relationship between tables. Need to check
for the primary and foreign key relationship and the existing data if any. (See if child table has
any records which are pointing to the master table records that are no more in master table.)

176. What is the use of update override in target table?

Ans. Used to update the particular records in the target level using where clauses.

Classification : Confidential
177. What is auxiliary mapping?

Ans. Auxiliary mapping is used to reflect change in one table when ever there is a change in
the other table.

Example:

In Siebel we have S_SRV_REQ and S_EVT_ACT table; let’s say that we have a image table
defined for S_SRV_REQ from where our mappings read data. Now if there is any change in
S_EVT_ACT then it won’t be captured in S_SRV_REQ if our mappings are using image table for
S_SRV_REQ. To overcome this we define a mapping between S_SRV_REQ and S_EVT_ACT
such that if there is any change in second it will be reflected as an update in the first table.

178. What is authenticator?

Ans. Authentication requests validate user names and passwords to access the Power Center
repository. You can use the following authentication requests to access the Power Center
repositories: Login Logout The Login function authenticates user name and password for a
specified repository. This is the first function a client application should call before calling any
other functions. The Logout function disconnects you from the repository and its Power Center
server connections. You can call this function once you are done calling Metadata and Batch
Web Services functions to release resources at the Web Services Hub.

179. How to create primary key only on odd numbers?

Ans. Use sequence generator and set the 'Increment by' property in that with 2.

180.In my source table 1000 rec's are there. I want to load 501 rec to 1000 rec into
my Target table?

Ans. Connect a sequence generator next to source qualifier into the expression. Use the filter
transformation to filter rows like where nextval >500.

181. What is the diff b/w source qualifier transformation and application source qualifier
transformation?

Ans. Source qualifier transformation is used to extract the data from one or more tables from
RDBMS of the same instance or from a single flat file system.

Application Source Qualifier transformation is used to extract the data from Application
sources like ERP.

182. What is the optimum method to load five flat files with same structures into
target? What is meant by file repository?

Ans. File repository is a collection of files with source file type as Indirect.

Just open a empty notepad and give the paths of files like

1. C: Flatfilesemp.txt 2.
D: Flatfilesemp2.txt 3. G: Flatfilesemp3.txt and u can give any
numbers. Be careful when u set the session properties “Set source file type as Indirect”.

Classification : Confidential
183. What are all the new features of informatica 8.1?

Ans. 1. Java Custom Transformation support 2. HTTP transformation support 3. New name of
Superglue as Metadata manager 4. New name of Power Analyzer as Data Analyzer 5. Support
Grid Computing 6. Push down Optimization.

184. How to FTP a file to a remote server?

Ans. U can transfer file from one server to other..............In UNIX there is an utility XCOMTCP
which transfer file from one server to other. But lot of constraints they are for this..... U need
to mention target server name and directory name where u need to send. The server directory
should have write permeation.....Check in detail in UNIX by typing MAN XCOMTCP command
which guides u i guess.

185. How a LOOKUP is passive?

Ans. Unconnected lookup is used for updating Slowly Changing Dimensions...so it is used to
determine whether the rows are already in the target or not, but it doesn't change the no. of
rows ...so it is passive.

Connected lookup transformations are used to get a related value based on some value or to
perform a calculation.....in either case it will either increase no. of columns or not...but
doesn't change row count...so it is passive.

In lookup SQL override property we can add a WHERE statement to the default SQL
statement, but it doesn't change no. of rows passing through it, it just reduces the no. of rows
included in the cache.

What is the purpose of using UNIX commands in informatica? Which UNIX


186.
commands are generally used with informatica?

Ans. Generally any mapping run from UNIX i.e. UNIX shell script calls work flow. Work flow
intern calls mapping. Suppose i am generating a flat file in UNIX from informatica mapping, if
I need to keep track of file when it has come today or yesterday...... For this append the
target table name with current date time stamp. In order to achieve u can use Unix shell script
in post session command. A small example i have given ...............U can do a lot ...........If
you are using doing MLOAD or Fload.Here even u can sort the flat file generated before doing
Mload or Fload.

187. Is
it possible to execute work flows in different repositories at the same time
using the same informatica server?

Ans. Yes

188. Explain about Informatica server process that how it works relates to mapping variables?

Ans. First load manager starts the session and it performs verifications and validations
about variables and manages post session tasks such as mail.
then it creates DTM process.
this DTM in turn creates a master thread, which creates remaining threads.
master thread creates

Classification : Confidential
read thread
write thread
transformation thread
pre and post session thread etc...

Finally DTM hand over’s to the load manager after writing into the target.

189. What does Check-In and Check-Out option refer to in the mapping designer?

Ans. Check-In and Check-Out refers to Versioning your Mapping. It is like maintaining the
changes you have made. It is like using VSS or CVS. When you right-click you mapping, you
have a option called Versioning if you have got that facility enabled.

190. Wheredo the records goes which does not satisfy condition in filter
transformation?

Ans. Session Logs.

191. How to identify Target bottlenecks?

Ans. The most common performance bottleneck occurs when the Informatica Server writes to a target
database. You can identify target bottlenecks by configuring the session to write to a flat file target. If the
session performance increases significantly when you write to a flat file, you have a target bottleneck. If
your session already writes to a flat file target, you probably do not have a target
bottleneck. You can optimize session performance by writing to a flat file target local to
the Informatica Server. Causes for a target bottleneck may include small check point intervals, small
database network packet size, or problems during heavy loading operations. For details about eliminating a
target bottleneck.

192. How to identify Source bottlenecks?

Ans. Performance bottlenecks can occur when the Informatica Server reads from a source
database. If your session reads from a flat file source, you probably do not have a source
bottleneck. You can improve session performance by setting the number of bytes the
Informatica Server reads per line if you read from a flat file source. If the session reads from
relational source, you can use a filter transformation, a read test mapping, or a database query to identify
source bottlenecks.

193. How to identify Mapping bottlenecks?

Ans. You can identify mapping bottlenecks by using a Filter transformation in the mapping. You can add a
Filter transformation in the mapping before each target definition. Set the filter condition to false so that no
data is loaded into the target tables. If the time it takes to run the new session is the same as the original
session, you have a mapping bottleneck. You can also identify mapping bottlenecks by using performance
details.

194. How to identify Session bottlenecks?

Ans. You can identify a session bottleneck by using the performance details. The Informatica Server creates
performance details when you enable Collect Performance Data in the Performance settings on the
Properties tab of the session properties. Performance details display information about each Source

Classification : Confidential
Qualifier, target definition, and individual transformation. All transformations have some basic counters that
indicate the number of input rows, output rows, and error rows. Small cache size, low buffer memory, and
small commit intervals can cause session bottlenecks. For details on eliminating session bottlenecks.

195. If I add a column in my flat file source how can load to the target which already contains data.

Ans. Alter the target table with new field and re-import the target and source. Add one lookup on target
table, and take the newly added field return port and create one filter, check lookup return value is null those
null records only populate into target.

196. What are the different types of repositories available from 6.1 version onwards?

Ans. A) Local Repository b) Global Repository c) Stand Alone Repository d) Versioned Repository

197. What is the meaning of $, $$, $$$ in Informatica?

Ans. $: These are system variables like $ Bad File, $ Input File, $ Output File, $ DB Connection.

$$: User Defined Variables

$$$: $$$ SessionstartTime

198. How do you recover a session or folder if you accidentally dropped them?

Ans. If you dropped your session or folder we can get it back if it has proper backup.
we can also get it back when you have same folders in other environment, I mean to say,
generally, we will have three environments and won't have back up for the same, there three
are dev/stage and production. If you dropped in any one or two of them would get it
recovered by copying objects from.
But If you intentionally dropped all three environments' folders or sessions then we cannot get
them back.

199. How do you load flat files of same structure into a target table?

Ans. You can do it through file list. You have create a file containing path of all file. And u
have to give that file name in session properties and select indirect.

200. Which transformation you can use to get n+1 record in Target database from N records source?

Ans. Normalizer

201. How do we extract SAP data Using Informatica? What is ABAP? What are IDOCS?

Ans. We can extract SAP data by using informatica with the powerConnect option.
We need to install and configure the powerConnect tool . Import the Source in to the Source
Analyzer. Here PowerConnect is the gateway between SAP & Informatica.Then develops the
mapping in the mapping designer. Next we need to generate the ABAP code for that mapping,
then only the Informatice can be pull the data from SAP.

Classification : Confidential
202. We have one source and 3 targets with same fields i.e. empno, ename in all 3 targets and
source, then the 1st row in source should go to 1st target, 2nd row in source should go to 2nd target
and the 3rd row in source should go to 3rd target, 4th row again to 1st target, 5th row to target2 6th
row to target 3 like this we need to transfer data how is it possible?

Ans. First Load Data from source To One dummy table in target with SEQ Gen T/R.
in SEQ Gen T/R Set min value 1 and max value 3 and check the reset option.
It will load the data with new column Values 1,2,3,1,2,3...........
Then use that dummy table as a source And use a router T/R in This mapping pipeline.
put Condition In router T/R, with three groups and load data in three different Targets.

203. Which is primary server process in informatica?


1. Repository manager 2.load manager 3.workflow manager 4.workflow designer

Ans. Load Manager

204. What is the Bridge Dimension?

Ans. It is a intersection table between a dimension table & fact table with a multipart key
capturing a many-to-many relationship.

205. How to send a mail to users with the output file data?

Ans. It can be possible by using %a in Session Components POST SESSION EMAIL tab.

206. What are the benefits of using of 64-bit platform instead of 32-bit?

Ans. With 32-bit the max memory you can have for a session is 2^32

With 64-bit the max memory you can have 2^64 - more memory means faster execution of
session.

207. What are the main differences between the informatica version 7 and version 8?

Ans. The new Version Informatica 8.1 has got some additional feature's like:1) Mapping Generation
Option2) Push Down Optimisation3) Data Cleanse and Match4) Data Federation (EII)5) Enterprise Grid6)
High Availability7) Unstructured Data And there are some updations in 1) Data Profiling2) Partitioning3)
Real-Time Loading.

208. How do you store Historical data in a DWH?

Ans. Through SCD’s.

209. What is the primary purpose of using surrogate key? Will those be used in fact tables? If not
then what is the purpose in using it?

Ans. Surrogate key is used for indexing purpose. i.e. as a primary key for a table. You can
say that why can't we use primary key. But the problem with pk is it may take more or less
memory i.e. it may contain characters but sk contains only number. So comparison of
numbers is easy when comparing with characters.
Yes sk will be used in fact tables to establish a relationship between dimensions.

Classification : Confidential
210. How do you take only duplicate rows in target table in Informatica?

Ans. you use rank transformation make range according the field which representing duplicasy
rows that have more then 1 rank put get only those rows in target table.

211. What is the difference between direct mode and indirect mode in relates to session properties?

Ans. Indirect and direct mode in session properties are related to flat file. If you are giving the
flat file name directly then it is direct mode. If you have 2 or more files where data is present
and you want the data to be loaded from all of these files then, you select indirect and give
that file name where you save all file name of the actual flat file.

212. i have EMP table with eno, ename as columns with duplicate rows.

Ex : eno ename
1 kiran
2 sena
3 naveen
1 kiran
4 dev
3 naveen
1 kiran
5 Loki
3 Naveen
1 kiran
6 shashi
1 kiran

etc...
So I want to find out the third occurrence of that duplicate row. I don’t know where that row is. There
r some millions of rows and so many duplicates for each row.

Ans. 1. Select distinct (ename) from emp_tab where empno in (select empno from (select
empno, count (*) from emp_tab group by empno having count (*)>2));

2. Select * from emp where rowid > select max(rowid) from emp group by eno,ename

3. Select * from t_name where rowid in (select rowid from t_name group by empno
having count (*) =3);

214. How a custom transformation works and what are the options and properties need to be set?

Ans. To extend the power center functionality by using outside developed procedures. You can create
Custom transformations with multiple input and output groups.

215. What is an Associated Port in Lookup Transformation?

Ans. Associate lookup ports with either an input/output port or a sequence ID. The Power
Center Server uses the data in the associated ports to insert or update rows in the lookup
cache. If you associate a sequence ID, the Power Center Server generates a primary key for
inserted rows in the lookup cache.

Classification : Confidential
216. What is the use of NewLookupRow in Lookup transformation when u enables the dynamic
lookup cache?

Ans. The Designer adds this port to a Lookup transformation configured to use a dynamic
cache. Indicates with a numeric value whether the Power Center Server inserts or updates the
row in the cache, or makes no change to the cache. To keep the lookup cache and the target
table synchronized, you pass rows to the target when the NewLookupRow value is equal to 1
or 2.

217. What is the difference between DIMENSION and DIMENSION TABLE?

Ans. Dimension table having more number of .attributes (dimension) where as dimension is single attribute.

218. What is the difference between Informatica server, Repository Server, Power center Server?

Ans. Both the Informatica Server and Power Centre server are same meaning in
Informatica.Its working to work proceedings for sessions and workflows in workflow Manager
Windows. The Repository server of Informatica is only to connect the Repository through the
Informatica Repository administration console and also connect to multiple repositories at the
same time. After Repository Server runned the Power Centre Server can work.

219. What is Power Channel in informatica?

Ans. power channel helps to transfer large amount of encrypted and compressed data over
LAN, WAN through firewalls, transfers files over FTP.

220. What is the difference between OLTP and ODS?

Ans. OLTP is online transaction processing systems and ODS os


Operational database system.
In OLTP we can save the current data, it depends on the day
to day transactions and it stores the day to day data.

In ODS we can store data for a month also and it is not


Restricted to a specific day or transaction.

221. What is persistent cache?

Ans. Ifyou want to save and reuse the cache files, you can configure the transformation to
use a persistent cache. Use a persistent cache when you know the lookup table does not
change between session runs.

222. What is shared cache?

Ans. You can also share the cache between multiple Lookup transformations by using a
persistent lookup cache and naming the cache files. You can share one cache between
Lookup transformations in the same mapping or across mappings.

223. What is Recache from Database Option?

Ans. You can instruct the Informatica Server to rebuild the lookup cache if you think that
the lookup table changed since the last time the Informatica Server built the cache. When
you want to rebuild a lookup cache, use the Recache from Database option.

Classification : Confidential
224. What are the differences between physical and logical data model?

Ans. The main important aspect is physical data model is created after logical data
model. For designing a logical data model one need not have through knowledge in
any specific RDBMS like oracle, Informix and so on. This is because logical data
model is created after having through study of the business entities which would be
involved in the application. The same can be represented by using E-R diagram also
called as Entity Relationship diagram which gives a diagrammatic representation of
the study made on the business application. The main aspect or quality a person
must have to build a good logical data model is through understanding of the
business, thorough knowledge about the application being modeled and very good
understanding of relational theory for presenting the understanding in a structured
manner. After the logical data model is built as specified above the physical data
model comes into picture. The logical data model built is used by business authors
and analysts and after a through revision and review the physical data model is
created. For creating a physical data model one must have good understanding of a
specific RDBMS in which it would be built. This is because physical data model is
represented using primary keys, foreign keys, triggers, referential integrity,
constraints and so on that is specific for any RDMBS.

225. What is the command to find out which shell you are running?

Ans. PS command is used to find out which shell you are running.

226. How to we create data mart?

Ans. Data mart is a subset of data warehouse and it supports a Particular


region,business unit or business functions. Data warehouses and data marts are built on
dimensional Data modeling where fact tables are connected with Dimension tables. It is
designed for a particular line of Business, such as sales, marketing, or finance.

227. if we r using aggregator we enabled sorted input but the records r not sorted what
happen?

Ans. This technique is used when you want a performance boost. There will be an aggregator
cache created and will be indexed. Although you have not given a sorted input it will still work fine
as it is indexed in cache. Please note this is not like a page fault where you have to bring the
page from the secondary memory when not found in primary memory.

228. Draw a workflow diagram for this situation. If s3 is true session s4 and s5 should concurrently
Run.if any one s4 or s5 true it should run s1.If s1 success it should execute s2 otherwise exit.can
any one give me suggestion.

Ans. 1. Run S3.


2. Connect it to S4 and S5. Check if PreviousTaskStatusS3 =
succeeded in the connecting link.
3. Once S3 is successfully executed it will start S4 and S5
4. Use a decision task and input that from S4 and S5. In
the decision use
PreviousTaskStatusS4 = succeeded or PreviousTaskStatusS5 =
succeeded
5. This will allow you to move ahead incase either of S4 or
S5 is succeeded.
6.Connect it to S1 using the same PreviousTaskStatus =
Succeeded.

Classification : Confidential
7.Connect it to S2 using the same PreviousTaskStatus =
Succeeded.
229. What is mean by grouping of condition column in lookup transformation?

Ans. grouping of condition column means- the columns which we have


included in lookup conditions, the lookup source should data should be
grouped on that column. This is mandatory for flat files, otherwise the
session will fail.

230. What is the term PIPELINE in informatica?

Ans. pipeline is used in the context of partitioning the source so that


the dtm process is executed i a less time. to make informatica server
read, transform n load the data into the targets in a relatively less
duration.

231. What is checksum terminology in informatica? Where do you use it?

Ans. Its a validation rule If the data is altered outside the company
firewall, the checksum will automatically detect the violation and deny
validation of the data.

232. Which T/r we can use it mapping parameter and mapping variable?
and which one is reusable for any mapping mapping parameter or mapping
variable?

Ans. It is seq gen, filter, expression in which u can use mapping


parameter and variable. Mapping parameter is reusable. U simply change
the value of the parameter in the parameter file.

233. One flat file it contains some data but i want to don’t want to
load first and last record.how it is? Can u tell me complete logic?

Ans. We can implement like this


1) Take sequence generator transformation and get sequence
of no's for each and every record
2) Then in expression take two output port's like out_min
and out_max.Assign Min and Max functions to corresponding
output ports then u will get Min and max value of the
records.
3) Take filter transformation and filter those two records in
filter transformation.

234. what is file list concept in informatica?

Ans. When you are using Flat File source you usually get the header
files separately from the data files. You may get source data in more
than one file. In such cases you give the paths of all the data files
in one file for eg say list.txt. This file is called the list file. In
the session edit task window instead of giving the source filename you
enter the name of the list file - list.txt and in source file type you
make it indirect.

Classification : Confidential
235. I have flat file source. i want to load the maximum salary of each
deptno into target. What is the mapping flow?

Ans. We can use an aggregator to group by on deptno and create a


newport to find the max(salary) and load dept no and salary, we’ll get
unique dept no and the max salary.

236. To eliminate duplicate records the following query is used.

Ans. select * from TABLE where COLUNA in (


Select COLUNA from TABELA group by COLUNA having 1 < count
(COLUNA) );

237.To get nth record who holds the maximum salary.

Ans. SELECT * FROM Employee E1 WHERE


(N-1) = (SELECT COUNT(DISTINCT(E2.Salary))
FROM Employee E2
WHERE E2.Salary > E1.Salary)

238. How DTM buffer size and buffer block size are related?

Ans. The number of buffer blocks in a session = DTM Buffer Size /


Buffer Block Size. Default settings create enough buffer blocks for 83
sources and targets. If the session contains more than 83, you might
need to increase DTM Buffer Size or decrease Default Buffer Block Size.

239. I have flat file source. I have two targets t1,t2. I want to load
the odd no. of records into t1 and even no. of records into t2. What is
the procedure and what t/r's are involved and what is the mapping flow?

Ans. 1)Drag source to mapping designer


2) Take a router transformation .consider EMP table in which
i am using EMPNO. In group1, assign the condition as
mod(empno,2)=0 which gives even numbers and in group2,
assign the condition as mod(empno,2)!=0 which gives odd numbers.
3) Connect group1 to one target and group2 to another target.

Or

1. Load ur source table into source analyzer.

2. Generate the target tables.

3. In the mapping designer take 2 filter transformations and


write these queries in each of the filter transformations

for even no of records:- select * from <tablename> where


(rowid,1) in (select rowid, mod(rownum,2) from <tablename>)

for odd no of records:- select * from <tablename> where


(rowid,0) in (select rowid, mod(rownum,2) from <tablename>)\

4. Connect the output ports of this transformation to their


respective output tables.

Classification : Confidential
240. What is casual dimension?

Ans. One of the most interesting and valuable dimensions in a data


warehouse is one that explains why a fact table record exists. In most
data warehouses, you build a fact table record when something happens.

For example:

When the cash register rings in a retail store, a fact table record is
created for each line item on the sales ticket. The obvious dimensions
of this fact table record are product, store, customer, sales ticket,
and time. At a bank ATM, a fact table record is created for every
customer transaction. The dimensions of this fact table record are
financial service, ATM location, customer, transaction type, and time.
When the telephone rings, the phone company creates a fact table record
for each "hook event." A complete call-tracking data warehouse in a
telephone company records each completed call, busy signal, wrong
number, and partially dialed call. In all three of these cases, a
physical event takes place, and the data warehouse responds by storing
a fact table record. However, the physical events and the corresponding
fact table records are more interesting than simply storing a small
piece of rev enue. Each event represents a conscious decision by the
customer to use the product or the service. A good marketing person is
fascinated by these events. Why did the customer choose to buy the
product or use the service at that exact moment? If we only had a
dimension called "Why Did The Customer Buy My Product Just Now?" our
data warehouses could answer almost any marketing question. We call a
dimension like this a "causal" dimension, because it explains what
caused the event.

241. What is the difference between view and materialized views?

Ans. view is a logical or virtual table it doesn't have data on


its own, but materialized view has a physical structure it stores data
in local machine or on its own. materialized view can be refreshed
automatically or manually. but in view, if any changes happened in the
base tables, we want to reflect the same in the view means view has
to issue the select statement again to the database.

242. We have a parameter file in UNIX location where we have .txt files
and those file will be used as source in informatica. I cannot use
source file name directly as file name will keep on changing in Unix
location. I need to define $$InputFile as parameter. Can anybody send
me the parameter file and the steps to handle this?

Ans. eg:
basu.txt(prm)

[foldername.sessname]
$$inputfileabc.domainname&path(root/aaa/bbb/prm.txt

now in session properties


give ths parameter file name: basu.txt

& under mapping tab in session properties

Classification : Confidential
give remove all other options
have only input conn:$$inputfileabc
243. How do u tune queries?

Ans. U can tune the queries by creating indexes on columns and


eliminating key constraints, also give sorting in sql itself.

244. Why do u use shortcuts in informatica?

Ans. Shortcut is a concept of reusability. If there is a mapping that


can be reused across several folders, create it in one folder and use
shortcuts of it in other folders. Thus, if you have to make change, you
can do it in main mapping which reflects in shortcut mappings
automatically.

245. What is the filename which you need to configure in UNIX while
installing informatica?

Ans. pmserver.cfg

246. What are the transformations that restrict the partitioning of


sessions?

Ans. Advanced External procedure transformation and External


procedure transformation: This transformation contains a
check box on the properties tab to allow partitioning.

Aggregator Transformation: If you use sorted ports you can


not partition the associated source

Joiner Transformation: you can not partition the master


Source for a joiner transformation

Normalizer Transformation

XML targets.

247. In unconnected lookup , what are the other transformations , that can be
used in place of that expression transformation ?

Ans. Filter, expression and update strategy transformations.

248. What is difference between partitioning of relational target and


partitioning of file targets?

Ans. If you partition a session with a relational target informatica


server creates multiple connections o the target database to write
target data concurently.If you partition a session with a file target
the informatica server creates one target file for each partition. You
can configure session properties to merge these target files.

249. In which circumstances that informatica server creates Reject


files?

Ans. When it encounters the DD_Reject in update strategy transformation.


Violates database constraint
Filed in the rows was truncated or overflowed.

Classification : Confidential
When the data in the file, may be tab separated, is incorrect. I.e. if
there is an extra tab for a set of records and this is making the text
data come under a numeric column, this violates the datatype and the
source file is rejected.

250. Why we use partitioning the session in informatica?

Ans. Partitioning achieves the session performance by reducing


the time period of reading the source and loading the data into target.

251. How the informatica server sorts the string values in Rank
transformation?

Ans. When the informatica server runs in the ASCII data movement
mode it sorts session data using Binary sort order. If you
configure the session to use a binary sort order, the informatica
server calculates the binary value of each string and returns the
specified number of rows with the highest binary values for the
string.

252. How can you improve session performance in aggregator


transformation?

Ans. The aggregator stores data in the aggregate cache until it


completes aggregate calculations. When u run a session that uses an
aggregator transformation, the informatica server creates index and
data caches in memory to process the transformation. If the informatica
server requires more space, it stores overflow values in cache files.

253. What is fact table granularity?

Ans. After you gather all the relevant information about the subject
area, the next step in the design process is to determine the
granularity of the fact table. To do this you must decide what an
individual low-level record in the fact table should contain. The
components that make up the granularity of the fact table correspond
directly with the dimensions of the data model. Thus, when you define
the granularity of the fact table, you identify the dimensions of the
data model.

Or

The level of details to be stored in fact table is termed as


granularity.
Eg: for a Retail store, the granularity for sales fact is as that of
Point Of Sales i.e., each transaction occurs, the data is stored in the
fact table.

254. How can we eliminate duplicate rows from flat file, explain?

Ans. using sorter transformation for distinct in a flat files.

255. Can we generate reports in informatica? How?

Ans. Yes we can, by using Informatica Metadata driven reporting Tool.

Classification : Confidential
256. What are the new features of informatica 7.1?

Ans. Lookup on flat file is possible.


Union Transformation is present.
Version Control.
LDAP Authentication.
Supporting of 64mb architecture.

257. In update strategy target table or flat file which gives more
performance? Why?

Ans. Flat file give better performance bcz Retrieval of data from flat
file is fast then relation database.

258. If u can't find what you are looking for in lookup table, how do
you handle?

Ans. This Depend on the situation,


To check the value is there or not there two ways
1. Use not found function (if possible)
2. Check the Key value from lkp as NULL
Than based on the requirement generate error/Put default
Value.

259. Suppose data are coming from different locations and those data
will not change. Is there any need to use surrogate key?

Ans. Yes, we should use the surrogate key, here we are getting data
from different locations means every one have one primary key, while
transforming the data into target that time more than two key not in
use so if you use surrogate key it will identified the duplicate fields
in dimensional table.

260. What is data purging?

Ans. Occasionally, it is necessary to remove large amounts of data from


a data warehouse. A very common scenario is the rolling window
discussed previously, in which older data is rolled out of the data
warehouse to make room for new data.

However, sometimes other data might need to be removed from a data


warehouse. Suppose that a retail company has previously sold products
from Company ABC, and that Company ABC has subsequently gone out of
business. The business users of the warehouse may decide that they are
no longer interested in seeing any data related to Company ABC, so this
data should be deleted. This process is data purging.

261. What is the difference between Bitmap and Btree index?

Ans. bit map index is an index in which the data will be in the form of
bites, by this we can retrieve the data very fastly
btree index is an index which is default index as normal indexes.

Classification : Confidential
262. How to we create data mart?

Ans. Data mart is a subset of data warehouse and it supports particular


region, business unit or business function. Data warehouses and data
marts are built on dimensional data modeling where fact tables are
connected with dimension tables. It is designed for a particular line
of business, such as sales, marketing, or finance.

263. If we r using aggregator we enabled sorted input but the records r


not sorted what happen?

Ans. This technique is used when you want a performance boost. There
will be an aggregator cache created and will be indexed. Although you
have not given a sorted input it will still work fine as it is indexed
in cache. Please note this is not like a page fault where you have to
bring the page from the secondary memory when not found in primary
memory.

264. What is the difference between Informatica7.1 & informatica 8.1?

Ans. 1)power center connect for SAP Net Weaver BW Option


2) SQL Transformation is added
3) Service Oriented Architecture
4) Grid concept is additional feature
5) Random file name can generate in target
6) Command line programs: Infacmd and infasetup new commands were added.
7) Java Transformation is added feature
8) Concurrent cache creation and faster index building are additional
feature in lookup transformation.

265. What is mean by grouping of condition column in lookup


transformation?

Ans. Grouping of condition column means- the columns which we have


included in lookup conditions, the lookup source data should be grouped
on that column. This is mandatory for flat files, otherwise the session
will fail.

266. Can we load data from a pdf to a table using informatica.

Ans. Yes, you can load data from a pdf to a table using
UDO[Unstructured Data Option] Transformation in Informatica from PC811
onwards.

267.

Classification : Confidential
Classification : Confidential
Classification : Confidential
Classification : Confidential

Das könnte Ihnen auch gefallen