You are on page 1of 6

What r the join types in joiner transformation?

Normal (Default) -- only matching rows from both master and detail Master outer -- all detail rows and only matching rows from master Detail outer -- all master rows and only matching rows from detail Full outer -- all rows from both master and detail (matching or non matching)

A mapplet is a reusable object that represents a set of transformations. Mapplet can be designed using mapping designer in informatica power center. A mapplet is a reusable object which contains one or more than one transformation which is used to populate the data from source to target based on the business logic and we can use the same logic in different mappings without creating the mapping again.

Transformations can be active or passive. An active transformation can change the number of rows that pass through it such as a Filter transformation that removes rows that do not meet the filter condition. A passive transformation does not change the number of rows that pass through it such as an Expression transformation that performs a calculation on data and passes all rows through the transformation. An Unconnected transforamtion is not connected to other transformations in the mapping. Connected transforamation is connected to other transforamtions in the mapping. An unconnected transformation cant be connected to another transformation. but it can be called inside another transformation. Connect Transformation : A transformation which participates in the mapping data flow Connected transformation can receive multiple inputs and provides multiple outputs Unconnected : A unconnected transformation does not participate in the mapping data flow It can receive multiple inputs and provides single output.
Reusable transformations can be used in multiple mappings. When u need to incorporate this transformation into maping,U add an instance of it to maping.Later if U change the definition of the transformation ,all instances of it inherit the changes.Since the instance of reusable t ransforamation is a pointer to that transforamtion,U can change the transforamation in the transformation developer,its instances automatically reflect these changes.

What r the unsupported repository objects for a mapplet?

Cobol sources
Joiner transformations Normalizer transformations Non reusable sequence generator transformations. Pre or post session stored procedures Target defintions Power mart 3.5 style Look Up functions

XML sources
IBM MQ source definitions

======================================= l Source definitions. Definitions of database objects (tables views synonyms) or files that provide source data. l Target definitions. Definitions of database objects or files that contain the target data. l Multi-dimensional metadata. Target definitions that are configured as cubes and dimensions. l Mappings. A set of source and target definitions along with transformations containing business logic that you build into the transformation. These are the instructions that the Informatica Server use s to transform and move data. l Reusable transformations. Transformations that you can use in multiple mappings. l Mapplets. A set of transformations that you can use in multiple mappings. l Sessions and workflows. Sessions and workflows store information about how and when the Informatica Server moves data. A workflow is a set of instructions that describes how and when to run tasks related to extracting transforming and loading data. A session is a type of task that you can put in a workflow. Each session corresponds to a single mapping. Aggregate cache contains data values while aggregate calculations are being performed. Aggregate cache is made up of index cache and data cache. Index cache contains group values and data cache consists of row values. ======================================= When server runs the session with aggregate transformation it stores data in memory until it completes the aggregation When u partition a source the server creates one memory cache and one disk cache for each partition .it routes the data from one partition to another based on group key values of the transformation
Use lookup transformation in ur mapping to lookup data in a relational table,view,synonym. Informatica server queries the look up table based on the lookup ports in the transformation.It compares the lookup transformation port values to lookup table column values based on the look up condition.

For Ex : Suppose the source contains only Empno but we want Empname also in the mapping.Then instead of adding another tbl which contains Empname as a source we can Lkp the table and get the Empname in target.

Use a Lookup transformation in your mapping to look up data in a relational table view or synonym. Import a lookup definition from any relatio nal database to which both the Informatica Client and Server can connect. You can use multiple Lookup transformations in a mapping.
What r two types of processes that informatica runs the session? Load manager Process: Starts the session, creates the DTM process, and sends post-session email when the session completes. The DTM process. Creates threads to initialize the session, read, write, and transform data, and handle pre- and post-session operations.

When the workflow start to run Then the informatica server process starts Two process:load manager process and DTM Process; The load manager process has the following tasks 1. lock the workflow and read the properties of workflow 2.create workflow log file 3. start the all tasks in workfolw except session and worklet 4.It starts the DTM Process. 5.It will send the post session Email when the DTM abnormally terminated The DTM process involved in the following tasks 1. read session properties 2.create session log file 3.create Threades such as master thread r ead write transformation threateds 4.send post session Email. 5.run the pre and post shell commands . 6.run the pre and post stored procedures.

Summary filter can be applied on a group of rows that contain a common value, where as Detail filters can be applied on each and every row of the data base.

Top Down Approach ODS-->ETL-->Datawarehouse -->Datamart -->OLAP Bottom up Approach ODS-->ETL-->Datamart-->Datawarehouse -->OLAP Power center server -power center server does the extraction from the source and loaded it to the target. Repository server -it takes care of the connection between the power center client and repository. Repository -it is a place where all the metadata information is stored.repository server and power center server access the repository for managing the data.

======================================= Polling : It displays the updated information about the session in the monitor window. The monitor window displays the status of each session when you poll the Informatica server.

Cache is stored in the Informatica server memory and over flowed data is stored on the disk in file format which will be automatically deleted after the successful completion of the session run. If you want to store that data you have to use a Persistent cache.

A surrogate key is system genrated/artificial key /sequence number or A surrogate key is a substitution for the natural primary key.It is just a unique identifier or number for each row that can be used for the primary key to the table. The only requirement for a surrogate primary key is that it is unique for each row in the tableI it is useful because the natural primary key (i.e. Customer Number in Customer table) can change and this makes updates more difficult, but In my project I felt that the primary reason for the surrogate keys was to record the changing context of the dimension attributes.(particulaly for scd )The reason for them being integer and integer joins are faster. Unlike other , ======================================= Surrogate key is a Unique identifier for eatch row it can be used as a primary key for DWH.The DWH does not depends on primary keys generated by OLTP systems for internally identifying the recods. When the new record is inserting into DWH primary keys are autimatical ly generated such type od keys are called SURROGATE KEY. Advantages : 1. Have a flexible mechanisam for handling S.C.D's 2. we can save substantial storage space with integer valued surrogate keys.

How to move the mapping from one database to another? Do you mean migration between repositories? There are 2 ways of doing this. 1. Open the mapping you want to migrate. Go to File Menu - Select 'Export Objects' and give a name - an XML file will be generated. Connect to the repository where you want to migrate and then select File Menu - 'Import Objects' and select the XML file name. 2. Connect to both the repositories. Go to the source folder and select mapping name from the object navigator and select 'copy' from 'Edit' menu. Now go to the target folder and select 'Paste' from 'Edit' menu. Be sure you open the target folder. ======================================= u can also do it this way. connect to both the repositories open the respective folders. keep the destination repository as active. from the naviga tor panel just drag and drop the mapping to the work area. it will ask whether to copy the mapping say YES. its done. ======================================= if we go by the direct meaning of your question....there is no need for a new mapping for a new database you just need to change the connections in the workflow manager to run the mapping on another database.

======================================= A materialized view provides indirect access to table data by storing the results of a query in a separate schema object unlike an ordinary view which does not take up any storage space or contain data. Materialized views are schema objects that can be used to summarize precompute replicate and distribute data. E.g. to construct a data warehouse. The definition of materialized view is very near to the concept of Cubes where we keep summarized data. But cubes occupy space. Coming to datamart that is completely different concept. Datawarehouse contains overall view of the organization. But datamart is spe cific to a subjectarea like Finance etc... we can combine different data marts of a compnay to form datawarehouse or we can split a datawarehouse into different data marts. The Repository Agent is a multi-threaded process that fetches inserts and updates metadata in the repository database tables. The Repository Agent uses object locking to ensure the consistency of metadata in the repository.

PMCMD means program command line utility. It is a program command line utility to communicate with informatica server. PMCMD performs following tasks: 1)start and stop batches and sessions 2)recovery sessions 3)stops the informatica 4)schedule the sessions by shell scripting 5)schedule the sessions by using operating system schedule tools like CRON

Tracing level in the case of Informatica specifies the level of detail of information that can be recorded in the session log file while executing the workflow. 4 types of tracing levels supported: 1. Normal: It specifies the initialization and status info rmation and summarization of the success rows and target tows and the information about the skipped rows due to transformation errors. 2. Terse specifies Normal + Notification of data 3. Verbose Initialisation: In addition to the Normal tracing specifies t he location of the data cache files and index cache files that are treated and detailed transformation statistics for each and every transformation within the mapping. 4. Verbose data: Along with verbose initialisation records each and every record processed by the Informatica server For better performance of mapping execution the tracing level should be specified as TERSE Verbose initialisation and verbose data are used for debugging purpose. =======================================

You might also like