64
Cognos Interview Questions 1.Explain how to create power play reports To create a power play report 1)You have to create a IQD file from impromptu by saving IMR report as IQD 2)Use Frame Work manager model and externalize the query subject you want to use in powerplay transformer and power play When you have the IQD file published to a location such as UNC Go to power play transformer model and select new, select the data source name, and IQD file location either published from FM or Impromptu (saved as IQD) You will see all the query items in the explorer then you can specify the dimensions like time and measures. Then generate categories for the dimensions and measures .Right click the cube and select create power cube, after that view the cube in cognos power play explorer. We can also publish the cube to PPES(Power play enterprise server) Publish the cube to the upfront Use power play web explorer to view the cube 2. what type problems we can face in general at report running time The most common problems are - 1. No Data Appears in the Report (to remove this check data source or package) 2. Unexpected or incorrect values appears in reports, may be report is running with limited data 3. The report filter does not work; values are not coming in filter option 4. Report is not able to open in excel, cvs or xml 3.How can i test reports in cognos? In cognos report net by the validate report option a report can be tested. If there will be any error, it will specify the error, unless it will give message -'report specification is valid'. 4.How can i schedule reports in cognos By using Cognos Scheduler, one can schedule the running reports in Impromptu to execute and save it in desired format. By using Cognos MACRO script language the reports can be executed and distributed to recipients by using mail applications. Compiled Cognos Macros can be scheduled using Cognos Scheduler. 5.What is the difference between a cascading report and drill thru report? Why do we go for drill thru report? Cascading report works based on the condition,but drill thru work based on the data item what we select as a drill thru options. 6.What is meant by Junk Dimension? A dimension which does not changes the grain level is called junk diamension.grain lowest level of reporting. 7. What is meant by Junk Dimension?How do you perfom while running the report? Where will you see the time Is LOOP CONSTRAINTS occur in COGNOS, if yes how to resolve that and please tell me how the loops occur in cognos?

Interview Questions for DWH Full Allignment

  • Upload
    nandy39

  • View
    39

  • Download
    8

Embed Size (px)

DESCRIPTION

DWH

Citation preview

Page 1: Interview Questions for DWH Full Allignment

Cognos Interview Questions

1.Explain how to create power play reports

To create a power play report1)You have to create a IQD file from impromptu by saving IMR report as IQD2)Use Frame Work manager model and externalize the query subject you want to use in powerplay transformer and power playWhen you have the IQD file published to a location such as UNCGo to power play transformer model and select new, select the data source name, and IQD file location either published from FM or Impromptu (saved as IQD) You will see all the query items in the explorer then you can specify the dimensions like time and measures. Then generate categories for the dimensions and measures .Right click the cube and select create power cube, after that view the cube in cognos power play explorer. We can also publish  the cube to PPES(Power play enterprise server)Publish the cube to the upfront Use power play web explorer to view the cube

2.  what type problems we can face in general at report running time  The most common problems are -1. No Data Appears in the Report (to remove this check data source or package)2. Unexpected or incorrect values appears in reports, may be report is running with limited data3. The report filter does not work; values are not coming in filter option4. Report is not able to open in excel, cvs or xml 

3.How can i test reports in cognos?

 In cognos report net by the validate report option a report can be tested. If there will be any error, it will specify the error, unless it will give message  -'report specification is valid'.

4.How can i schedule reports in cognosBy using Cognos Scheduler, one can schedule the running reports in Impromptu to execute and save it in desired format. By using Cognos MACRO script language the reports can be executed and distributed to recipients by using mail applications. Compiled Cognos Macros can be scheduled using Cognos Scheduler.

5.What is the difference between a cascading report and drill thru report? Why do we go for drill thru report?Cascading report works based on the condition,but drill thru work based on the data item what we select as a drill thru options.

6.What is meant by Junk Dimension?A dimension which does not changes the grain level is called junk diamension.grain lowest level of reporting.

7.  What is meant by Junk Dimension?How do you perfom while running the report?Where will you see the time  Is LOOP CONSTRAINTS occur in COGNOS, if yes how to resolve that and please tell me how the loops occur in cognos?Problem Description

what are looping joins and how do you resolve them?Solution DescriptionLooping joins could potentially return incorrect data.An example of a looping join, B -> CA -> < > -> DE -> FWhen you select an item from table A and D, Impromptu will try to choose the shortest path, e.g. if A -> D existed, then this is the path Impromptu will take. But in the above situation, the two paths are equal and so Impromptu has to make a choice, e.g. "A -> B -> C -> D" OR "A -> E -> F -> D". Impromptu makes it's choice based on how the catalog was constructed, which cannot be altered once it's created; order of the tables in the catalog.The two paths could return different results depending on the relationship between the tables in the path.

Page 2: Interview Questions for DWH Full Allignment

The report would be fine IF Impromptu chose the expected path. The choice is not always right.

Eliminating looping joins prevents Impromptu from making the wrong choice. To eliminate looping joins, you can either break unnecessary joins, e.g. reports do not need a join between table F and D, e.g.

B -> C -> DA -> < E -> F

But if you need all the joins, use ALIAS tables to break the looping join. Add an alias table for table A and break the join from table A and E, e.g.

A -> B -> C> -> DAlias A -> E -> F

Both solutions could affect existing reports. Title: Looped joinsCreated: Nov 04, 1999Applies To: Impromptu - 2.0

Impromptu - 3.01Impromptu - 3.03Impromptu - 3.04Impromptu - 3.5Impromptu - 4.0Impromptu - 5.0Impromptu - 7.1

Problem Description

Under the Joins dialog on the Analyze Tab it states a Loop join is present. What does this mean and how can it be resolved?

Solution Description

A Loop Join occurs when there are multiple paths between database tables. An example of this is A joins to B and B joins to C and C joins to A.

The proper definition of join strategies in an Impromptu catalog is crucial to the success of an ad-hoc reporting environment. Impromptu shelters the user from having to know any of the technical information about the database, including name, location, table and column names, and join strategies. The Impromptu Administrator must be very thorough in their definition and testing of the join strategies. Impromptu provides an ability to analyze the joins and determine any anomalies. The most common is the Loop Join.

The implications of the loop join are that there is no way to predetermine which of the various join paths will be used by Impromptu when creating the SQL. SQL is dynamically generated for each report as it is created and before it executes. For example, to create a report using columns from tables A and C, we could join from A=>B=>C or directly from A=>C. In some cases, both of these joins would result in the same data being retrieved. However, in other cases it may result in different data. Impromptu will always try to use the shortest route in joining multiple tables. It will also try to use the tables that are already included in the query, rather than including an additional table.

There is no hard and fast rule to resolving Loop Joins. There are four basic resolutions:1. Break the join2. Create alias tables with different join strategies3. Use the join expression editor to specify the join4. Modify SQLEach of these resolutions is done for a different reason and may have some issues associated with it. Determine the best resolution for your situation by analyzing the data with regards to the results required from the join structure.

Page 3: Interview Questions for DWH Full Allignment

Example:

The join structure looks like this:

A = BA = CB = C

This is producing incorrect results. To resolve this issue, make table C an alias to omit the loop in the join structure and this will result in data displaying correctly.

Correct Join Structure:

A = BA = C aliasB = C

8.Use this query i am retriving all years OCt data from 01-10-2004 to 30-10-2007i need to restrict this query to current date and current yearYou have a function called 'extract' in cognos.Ex:- extract(month, the date field). By giving like this you will get month. So you can keep a filter to restrict the rows only for October.

9.How to show the data reported horizontally:(For example:)employee skill1 a1 b1 c2 d2 e2 f

Report result:1 abc2 defAssuming 3 records per grouped item:

1. Group on employee2. Create a running count based on the skill field.3. Create 3 calculated columns based on the count field.Call them skill1, skill2, skill3:

if (count = 1) then (skill) else nullif (count = 2) then (skill) else nullif (count = 3) then (skill) else null

4. Create 3 more calculated columns using the maximum function. Call them maxskill1, maxskill2, maxskill3

maximum (skill1)maximum (skill2)maximum (skill3)

5. Group on employee on maxskill1 on maxskill2 on maxskill36. Report employee axskill1 maxskill2 maxskill3

10.How to pass multiple values from picklist prompt to sub report filterThe sub-report only includes the first value.When the sub-report query runs, it checks for the first row in the Customer Name column and shows only information for that customer. If you want a sub-report to show information for another row in the column,

Page 4: Interview Questions for DWH Full Allignment

place the main report in a form frame that shows only one row at a time. When you insert the sub-report into the form frame as well, it changes as you click through the rows in the main report. For example, the main and sub-report above are both in a form frame that shows only one row of the Customer Name column at a time. Each time you scroll to another customer name, the sub-report shows only information for that customer."

11.How can I create a dynamic column name in Cognos1. Create a calculated column which contains the information that the header is to contain, such as "Report for year 1999" (concatenated text and date to string sub string extraction). 2. Highlight the report, and then right-click. 3. Select Properties, and then click the Headers/Footers tab. 4.Clear the Column Title Header check box. This will remove the headers from your columns. 5. Reinsert the rest of the column headers; insert text will work. For the dynamic column, from the Insert menu, click Data and select the calculated column you created and insert it into the report.

Page 5: Interview Questions for DWH Full Allignment

Data warehouse Concepts

1.Why fact table is in normal form?A fact table consists of measurements of business requirements and foreign keys of dimensions tables as per business rules.

2.What is Difference between E-R Modeling and Dimentional Modeling.Basic diff is E-R modeling will have logical and physical model. Dimensional model will have only physical model. E-R modeling is used for normalizing the OLTP database design.Dimensional modeling is used for de-normalizing the ROLAP/MOLAP design.

3.What is conformed fact?Conformed dimensions are the dimensions which can be used across multiple Data Marts in combination with multiple facts tables accordingly

4.What are the methodologies of Data Warehousing.Most of the time, we use Mr. Ralph Kimball methodologies for datawarehousing design.Two kind of schema star and snow flake.

5.What is Data warehosuing Hierarchy?HierarchiesHierarchies are logical structures that use ordered levels as a means of organizing data. A hierarchy can be used to define data aggregation. For example, in a time dimension, a hierarchy might aggregate data from the month level to the quarter level to the year level. A hierarchy can also be used to define a navigational drill path and to establish a family structure.Within a hierarchy, each level is logically connected to the levels above and below it. Data values at lower levels aggregate into the data values at higher levels. A dimension can be composed of more than one hierarchy. For example, in the product dimension, there might be two hierarchies--one for product categories and one for product suppliers.Dimension hierarchies also group levels from general to granular. Query tools use hierarchies to enable you to drill down into your data to view different levels of granularity. This is one of the key benefits of a data warehouse.When designing hierarchies, you must consider the relationships in business structures. For example, a divisional multilevel sales organization.Hierarchies impose a family structure on dimension values. For a particular level value, a value at the next higher level is its parent, and values at the next lower level are its children. These familial relationships enable analysts to access data quickly.LevelsA level represents a position in a hierarchy. For example, a time dimension might have a hierarchy that represents data at the month, quarter, and year levels. Levels range from general to specific, with the root level as the highest or most general level. The levels in a dimension are organized into one or more hierarchies.Level RelationshipsLevel relationships specify top-to-bottom ordering of levels from most general (the root) to most specific information. They define the parent-child relationship between the levels in a hierarchy.Hierarchies are also essential components in enabling more complex rewrites. For example, the database can aggregate an existing sales revenue on a quarterly base to a yearly aggregation when the dimensional dependencies between quarter and year are known.

6.What is surrogate key ? where we use it expalin with examplesSurrogate key is a substitution for the natural primary key.It is just a unique identifier or number for each row that can be used for the primary key to the table. The only requirement for a surrogate primary key is that it is unique for each row in the table.

Data warehouses typically use a surrogate, (also known as artificial or identity key), key for the dimension tables primary keys. They can use Infa sequence generator, or Oracle sequence, or SQL Server Identity values for the surrogate key.

It is useful because the natural primary key (i.e. Customer Number in Customer table) can change and this

Page 6: Interview Questions for DWH Full Allignment

makes updates more difficult.

Some tables have columns such as AIRPORT_NAME or CITY_NAME which are stated as the primary keys (according to the business users) but ,not only can these change, indexing on a numerical value is probably better and you could consider creating a surrogate key called, say, AIRPORT_ID. This would be internal to the system and as far as the client is concerned you may display only the AIRPORT_NAME.

7.Importance of Surrogate Key in Data warehousing?Surrogate Key is a Primary Key for a Dimension table. Most importance of using it is it is independent of underlying database. i.e Surrogate Key is not affected by the changes going on with a database. 

8.What is the flow of loading data into fact & dimensional tables?Fact table - Table with Collection of Foreign Keys corresponding to the Primary Keys in Dimensional table. Consists of fields with numeric values. Dimension table - Table with Unique Primary Key. Load - Data should be first loaded into dimensional table. Based on the primary key values in dimensional table, the data should be loaded into Fact table.

9.What is a linked cube?A cube can be stored on a single analysis server and then defined as a linked cube on other Analysis servers. End users connected to any of these analysis servers can then access the cube. This arrangement avoids the more costly alternative of storing and maintaining copies of a cube on multiple analysis servers. linked cubes can be connected using TCP/IP or HTTP. To end users a linked cube looks like a regular cube.

10.What is meant by metadata in context of a Datawarehouse and how it is important?Metadata or Meta Data Metadata is data about data. Examples of metadata include data element descriptions, data type descriptions, attribute/property descriptions, range/domain descriptions, and process/method descriptions. The repository environment encompasses all corporate metadata resources: database catalogs, data dictionaries, and navigation services. Metadata includes things like the name, length, valid values, and description of a data element. Metadata is stored in a data dictionary and repository. It insulates the data warehouse from changes in the schema of operational systems. Metadata Synchronization The process of consolidating, relating and synchronizing data elements with the same or similar meaning from different systems. Metadata synchronization joins these differing elements together in the data warehouse to allow for easier access.

11.Differentiate Primary Key and Partition Key? Primary Key is a combination of unique and not null. It can be a collection of key values called as composite primary key. Partition Key is a just a part of Primary Key. There are several methods of partition like Hash, DB2, Random etc..While using Hash partition we specify the Partition Key. 

12.What are the possible data marts in Retail sales.?Product information, sales information

13.What is  degenerate dimension table?In simple terms, the column in a fact table that does not map to any dimensions, neither it s a measure column. for e.g Invoice no, Invoice_line_no in fact table will be a degenerate dimension (columns), provided if you don’t have a dimension called invoice.

14.What is the main differnce between schema in RDBMS and schemas in DataWarehouse?RDBMS Schema * Used for OLTP systems * Traditional and old schema * Normalized * Difficult to understand and navigate * Cannot solve extract and complex problems * Poorly modelled  DWH Schema * Used for OLAP systems * New generation schema * De Normalized * Easy to understand and navigate 

Page 7: Interview Questions for DWH Full Allignment

* Extract and complex problems can be easily solved * Very good model

15.Difference between Snow flake and Star Schema. What are situations where Snow flake Schema is better than Star Schema to use and when the opposite is true?Star schema contains the dimesion tables mapped around one or more fact tables.It is a denormalised model.No need to use complicated joins.Queries results fastly.Snowflake schema It is the normalised form  of Star schema.contains indepth joins ,bcas the tbales r splitted in to many pieces.We can easily do modification directly in the tables.We hav to use comlicated joins ,since we hav more tables .There will be some delay in processing the Query .

16. What is a CUBE in datawarehousing concept?Cubes r muti-dimensional view of dw or data marts. it is designed in a logical way to drill, slice-n-dice. Every part of the cube is a logical representation of the combination of facts-dimension attribs.

17. Is it correct/feasible develop a Data Mart using an ODS?The ODS is technically designed to be used as the feeder for the DW and other DM's -- yes.  It is to be the source of truth.

18.Why should you put your data warehouse on a different system than your OLTP systemOLTP system stands for on-line transaction processing.  These are used to store only daily transactions as the changes have to be made in as few places as possible. OLTP do not have historical data of the organization Datawarehouse will contain the historical information about the organization

19.Why are OLTP database designs not generally a good idea for a Data WarehouseOLTP cannot store historical information about the organization. It is used for storing the details of daily transactions while a datawarehouse is a huge storage of historical information obtained from different datamarts for making intelligent decisions about the organization.

20.What are conformed dimensionsConformed Dimensions are the Dimensions which are common to two cubes .say CUBE-1 contains F1,D1,D2,D3 and CUBE-2 contains F2,D1,D2,D4 are the Facts and Dimensions ,here D1,D2 are the Conformed Dimensions  

21.What are Semi-additive and factless facts and in which scenario will you use such kinds of fact tablesA fact less fact table captures the many-to-many relationships between dimensions, but contains no numeric or textual facts. They are often used to record events or coverage information. Common examples of fact less fact tables include: - Identifying product promotion events (to determine promoted products that didn’t sell) - Tracking student attendance or registration events - Tracking insurance-related accident events - Identifying building, facility, and equipment schedules for a hospital or university

22.What is SCD1 , SCD2 , SCD3The value of dimensions is used change very rarely, That is called Slowly Changing dimensionsHere mainly 31) SCD1: Replace the old values overwrite by new values2) SCD2: Just Creating Additional records3) SCD3: It’s maintain just previous and recentIn the SCD2 again 31) Versioning2) Flag value3) Effective Date rangeVersioning: Here the updated dimensions inserted in to the target along with version numberThe new dimensions will be inserted into the target along with Primary keyFlagvalue:The updated dimensions insert into the target along with 0and new dimensions inset into the target along with 1

ORSCD Type 1, the attribute value is overwritten with the new value, obliterating the historical attribute values. For example, when the product roll-up changes for a given product, the roll-up attribute is merely updated with the current value. 

Page 8: Interview Questions for DWH Full Allignment

SCD Type 2,a new record with the new attributes is added to the dimension table. Historical fact table rows continue to reference the old dimension key with the old roll-up attribute; going forward, the fact table rows will reference the new surrogate key with the new roll-up thereby perfectly partitioning history.  SCDType 3, attributes are added to the dimension table to support two simultaneous roll-ups - perhaps the current product roll-up as well as “current version minus one”, or current version and original.

23.What is VLDBThe perception of what constitutes a VLDB continues to grow. A one terabyte database would normally be considered to be a VLDB.

24.What are non-additive factsNon-additive facts are facts that cannot be summed up for any of the dimensions present in the fact table. Example: temparature,bill number...etc

25.What are slowly changing dimensionsIf the data in the dimension table happen to change very rarely,then it is called as slowly changing dimension.Ex: changing the name and address of a person, this happens rarely.

26.What does level of Granularity of a fact table signifyIn simple terms, level of granularity defines the extent of detail. As an example, let us look at geographical level of granularity. We may analyze data at the levels of COUNTRY, REGION, TERRITORY, CITY and STREET. In this case, we say the highest level of granularity is STREET.

27.Which columns go to the fact table and which columns go the dimension tableThe Aggregation or calculated value columns will go to Fact Table and details information will go to dimensional table.

28.What is ODS?ODS means Operational Data Store It is used to store current data through transactional web applications, Sap, MQ series Current data means particular data from one date into one date. ODS contains 30-90 data.

29.What is Normalization, First Normal Form, Second Normal Form , Third Normal FormNormalization: The process of decomposing tables to eliminate data redundancy is called Normalization.  1 N.F :- The table should contain scalar or atomic values. 2 N.F :- Table should be in 1N.F + No partial functional dependencies 3 N.F :-Table should be in 2 N.F + No transitive dependencies

30.What is real time data-warehousing?Real-time data warehousing is a combination of two things: 1) real-time activity and 2) data warehousing. Real-time activity is activity that is happening right now. The activity could be anything such as the sale of widgets. Once the activity is complete, there is data about it. Data warehousing captures business activity data. Real-time data warehousing captures business activity data as it occurs. As soon as the business activity is complete and there is data about it, the completed activity data flows into the data warehouse and becomes available instantly. In other words, real-time data warehousing is a framework for deriving information from data as the data becomes available. 

31.What are modeling tools available in the MarketModeling Tool Vendor 

Erwin Computer Associates ER/Studio Embarcadero Power Designer Sybase Oracle Designer Oracle

32.What is a general purpose scheduling tool?General purpose of scheduling tool may be cleansing and loading data at specific given time.

33.What is a lookup table?reference table can be otherwise called as lookup table

Page 9: Interview Questions for DWH Full Allignment

34.What is a dimension table?A dimensional table is a collection of hierarchies and categories along which the user can drill down and drill up. it contains only the textual attributes.

35.What is Fact table?A table in a data warehouse whose entries describe data in a fact table. Dimension tables contain the data from which dimensions are created.

36.What are the various Reporting tools in the Market?INEA MS-Excel  Business Objects (Crystal Reports)  Cognos (Impromptu, Power Play)  Micro strategy  MS reporting services  Informatic Power Analyzer  Actuate  Hyperion (BRIO)  Oracle Express OLAP  Proclarity  SAS

37.What are the vaious ETL tools in the Market?1. Informatica Power Center 2. Ascential Data Stage 3. ESS Base Hyperion 4. AbIntio 5. BO Data Integrator 6. SAS ETL 7. MS DTS 8. Oracle OWB 9. Pervasive Data Junction 10. Cognos Decision Stream

38.What is ETLETL is an abbreviation for "Extract, Transform and Load”. This is the process of extracting data from their operational data sources or external data sources, transforming the data which includes cleansing, aggregation, summarization, integration, as well as basic transformation and loading the data into some form of the data warehouse.

39.What is the Difference between OLTP and OLAPCurrent data Short database transactions Online update/insert/delete Normalization is promoted High volume transactions Transaction recovery is necessary   OLAP Current and historical data Long database transactions Batch update/insert/delete Denormalization is promoted Low volume transactions Transaction recovery is not necessary 

40.What are Aggregate tables?Aggregate table contains the  summary of existing warehouse data which is grouped to certain levels of dimensions.Retrieving the required data from the actual table, which have millions of records will take more time and also affects the server performance.To avoid this we can aggregate the table to certain required level and can use it.This tables reduces the load in the database server and increases the performance of the query and can retrieve the result very fastly. E.g Yearly, monthly sales information.

Page 10: Interview Questions for DWH Full Allignment

41.What are the Different methods of loading Dimension tables?Conventional Load: Before loading the data, all the Table constraints will be checked against the data.  Direct load:(Faster Loading) All the Constraints will be disabled. Data will be loaded directly. Later the data will be checked against the table constraints and the bad data won't be indexed.

42.What is Dimensional Modeling?Dimensional Modeling is a design concept used by many data warehouse designers to build their data warehouse. In this design model all the data is stored in two types of tables - Facts table and Dimension table. Fact table contains the facts/measurements of the business and the dimension table contains the context of measurements i.e., the dimensions on which the facts are calculated.

43.What is ER DiagramER - Stands for entitity relationship diagrams. It is the first step in the design of data model which will later lead to a physical database design of possible a OLTP or OLAP database.

44.What are  Data MartsData Mart is a segment of a data warehouse that can provide data for reporting and analysis on a section, unit, department or operation in the company, e.g. sales, payroll, production. Data marts are sometimes complete individual data warehouses which are usually smaller than the corporate data warehouse.

45.What is a Data Warehousing?Data Warehouse is a repository of integrated information, available for queries and analysis. Data and information are extracted from heterogeneous sources as they are generated....This makes it much easier and more efficient to run queries over data that originally came from different sources. Typical relational databases are designed for on-line transactional processing (OLTP) and do not meet the requirements for effective on-line analytical processing (OLAP). As a result, data warehouses are designed differently than traditional relational databases.

Page 11: Interview Questions for DWH Full Allignment

Informatica

1.What are Target Options on the Servers?Target Options for File Target type are FTP File, Loader and MQ. There are no target options for ERP target type Target Options for Relational are Insert, Update (as Update), Update (as Insert), Update (else Insert), Delete, and Truncate Table

2.How do you identify existing rows of data in the target table using lookup transformation?Can identify existing rows of data using unconnected lookup transformation.

3.What are Aggregate transformation?The Aggregator transformation allows you to perform aggregate calculations, such as averages and sums. The Aggregator transformation is unlike the Expression transformation, in that you can use the Aggregator transformation to perform calculations on groups

4.What are various types of Aggregation?Various types of aggregation are SUM, AVG, COUNT, MAX, MIN, FIRST, LAST, MEDIAN, PERCENTILE, STDDEV, and VARIANCE.

5.What are 2 modes of data movement in Informatica Server?The data movement mode depends on whether Informatica Server should process single byte or multi-byte character data. This mode selection can affect the enforcement of code page relationships and code page validation in the Informatica Client and Server.

a) Unicode - IS allows 2 bytes for each character and uses additional byte for each non-ascii character (such as Japanese characters)b) ASCII - IS holds all data in a single byte

The IS data movement mode can be changed in the Informatica Server configuration parameters. This comes into effect once you restart the Informatica Server.

6..What is Code Page Compatibility?Compatibility between code pages is used for accurate data movement when the Informatica Sever runs in the Unicode data movement mode. If the code pages are identical, then there will not be any data loss. One code page can be a subset or superset of another. For accurate data movement, the target code page must be a superset of the source code page. Superset - A code page is a superset of another code page when it contains the character encoded in the other code page, it also contains additional characters not contained in the other code page.Subset - A code page is a subset of another code page when all characters in the code page are encoded in the other code page.

7.What is Code Page used for?Code Page is used to identify characters that might be in different languages. If you are importing Japanese data into mapping, u must select the Japanese code page of source data.

8.What is Router transformation?Router transformation allows you to use a condition to test data. It is similar to filter transformation. It allows the testing to be done on one or more conditions. Router transformation is use to load data in multiple targets depending on the test condition.

9.What is Load Manager?While running a Workflow,the PowerCenter Server uses the Load Manager process and the Data Transformation Manager Process (DTM) to run the workflow and carry out workflow tasks.When the PowerCenter Server runs a workflow, the Load Manager performs the following tasks:  1. Locks the workflow and reads workflow properties. 2. Reads the parameter file and expands workflow variables. 3. Creates the workflow log file. 4. Runs workflow tasks. 5. Distributes sessions to worker servers. 6. Starts the DTM to run sessions. 7. Runs sessions from master servers. 

Page 12: Interview Questions for DWH Full Allignment

8. Sends post-session email if the DTM terminates abnormally.  When the PowerCenter Server runs a session, the DTM performs the following tasks: 1. Fetches session and mapping metadata from the repository. 2. Creates and expands session variables. 3. Creates the session log file. 4. Validates session code pages if data code page validation is enabled. Checks query conversions if data code page validation is disabled. 5. Verifies connection object permissions. 6. Runs pre-session shell commands. 7. Runs pre-session stored procedures and SQL. 8. Creates and runs mapping, reader, writer, and transformation threads to extract,transform, and load data. 9. Runs post-session stored procedures and SQL. 10. Runs post-session shell commands. 11. Sends post-session email.

10.What is Data Transformation Manager?After the load manager performs validations for the session, it creates the DTM process. The DTM process is the second process associated with the session run. The primary purpose of the DTM process is to create and manage threads that carry out the session tasks.

· The DTM allocates process memory for the session and divide it into buffers. This is also known as buffer memory. It creates the main thread, which is called the master thread. The master thread creates and manages all other threads.· If we partition a session, the DTM creates a set of threads for each partition to allow concurrent processing.. When Informatica server writes messages to the session log it includes thread type and thread ID. Following are the types of threads that DTM creates:

Master thread - Main thread of the DTM process. Creates and manages all other threads. Mapping thread - One Thread to Each Session. Fetches Session and Mapping Information. Pre and Post Session Thread-One Thread each to Perform Pre and Post Session Operations.reader thread-One Thread for Each Partition for Each Source Pipeline. WRITER THREAD-One Thread for Each Partition if target exist in the source pipeline write to the target. TRANSFORMATION THREAD - One or More Transformation Thread For Each Partition11.What is Session and Batches?

Session - A Session Is A set of instructions that tell the Informatica Server How and When to Move Data from Sources to Targets. After creating the session, we can use either the server manager or the command line program pmcmd to start or stop the session.Batches - It Provides A Way to Group Sessions for Either Serial or Parallel Execution by the Informatica Server. There Are Two Types Of Batches : Sequential - Run Session One after the Other. concurrent - Run Session At The Same Time.

12.What is a source qualifier?When you add a relational or a flat file source definition to a mapping, you need to connect it to a Source Qualifier transformation. The Source Qualifier represents the rows that the Informatica Server reads when it executes a session

13.Why we use lookup transformations?Lookup Transformations can access data from relational tables that are not sources in mapping. With Lookup transformation, we can accomplish the following tasks:

Get a related value-Get the Employee Name from Employee table based on the Employee ID Perform Calculation.Update slowly changing dimension tables - We can use unconnected lookup transformation to determine whether the records already exist in the target or not.

14.While importing the relational source defintion from database,what are the meta data of source U import?Source name Database location Column names Data types Key constraints

Page 13: Interview Questions for DWH Full Allignment

15.How many ways you can update a relational source defintion and what r they?Two ways 1. Edit the definition 2. Reimport the definition

16.Where should U place the flat file to import the flat file defintion to the designer?There is no such restriction to place the source file. In performance point of view its better to place the file in server local source folder. if you need path please check the server properties available at workflow manager. It doesn't mean we should not place in any other folder, if we place in server source folder by default source will be selected at time session creation.

17.To provide support for Mainframes source data,which files r used as a source definitions?COBOL Copy-book files

18.Which transformation should u need while using the cobol sources as source defintions?Normalizer transformation which is used to normalize the data. Since cobol sources r oftenly consists of Demoralized data.

19.How can U create or import flat file definition in to the warehouse designer?U can not create or import flat file defintion in to warehouse designer directly. Instead U must analyze the file in source analyzer, then drag it into the warehouse designer. When U drag the flat file source definition into warehouse designer workspace, the warehouse designer creates a relational target defintion not a file defintion.If u want to load to a file, configure the session to write to a flat file. When the informatica server runs the session, it creates and loads the flat file.

20.What is the maplet?Maplet is a set of transformations that you build in the maplet designer and U can use in multiple mapings.For Ex:Suppose we have several fact tables that require a series of dimension keys.Then we can create a mapplet which contains a series of Lkp transformations to find each dimension key and use it in each fact table mapping instead of creating the same Lkp logic in each mapping. Set of transforamations where the logic can be reusable.

21.What is a transforamation?A transformation is repository object that pass data to the next stage (i.e to the next transformation or target) with/with out modifying the data

22.What r the designer tools for creating tranformations?Mapping designer Transformation developer Mapplet designer

23.What r the active and passive transforamtions?Transformations can be active or passive. An active transformation can change the number of rows that pass through it, such as a Filter transformation that removes rows that do not meet the filter condition. A passive transformation does not change the number of rows that pass through it, such as an Expression transformation that performs a calculation on data and passes all rows through the transformation.

24.What r the connected or unconnected transforamations?An unconnected transforamtion is not connected to other transformations in the mapping.Connected transforamation is connected to other transforamtions in the mapping.

25.How many ways u create ports?Two ways 1.Drag the port from another transformation 2.Click the add button on the ports tab.

Page 14: Interview Questions for DWH Full Allignment

26.What r the reusable transforamtions?Reusable transformations can be used in multiple mappings. When u need to incorporate this transformation into mapping add an instance of it to maping.Later if U change the definition of the transformation, all instances of it inherit the changes. Since the instance of reusable transformation is a pointer to that transformation can change the transformation in the transformation developer, its instances automatically reflect these changes this feature can save U great deal of work.27.What r the methods for creating reusable transforamtions?Two methods 1.Design it in the transformation developer. 2. Promote a standard transformation from the mapping designer. After U adds a transformation to the mapping, U can promote it to the status of reusable transformation. Once U promotes a standard transformation to reusable status can demote it to a standard transformation at any time. If u change the properties of a reusable transformation in mapping can revert it to the original reusable transformation properties by clicking the revert button.

28.What r the unsupported repository objects for a mapplet?COBOL source definition Joiner transformations Normalizer transformations Non reusable sequence generator transformations. Pre or post session stored procedures Target defintions Power mart 3.5 style Look Up functions XML source definitions IBM MQ source definitions

29. What r the mapping paramaters and maping variables? Mapping parameter represents a constant value that U can define before running a session. A mapping parameter retains the same value throughout the entire session. When u use the mapping parameter, U declare and use the parameter in a mapping or maplet.Then define the value of parameter in a parameter file for the session. Unlike a mapping parameter, a mapping variable represents a value that can change throughout the session. The informatica server saves the value of mapping variable to the repository at the end of session run and uses that value next time U run the session.

30.Can U use the maping parameters or variables created in one maping into another maping?N0.We can use mapping parameters or variables in any transformation of the same maping or mapplet in which U have created maping parameters or variables.

31. Can u use the maping parameters or variables created in one maping into any other reusable transformation?Yes.Because reusable tranformation is not contained with any maplet or maping.

32. How can U improve session performance in aggregator transformation?Use sorted input.

33.What is aggregate cache in aggregator transforamtion? The aggregator stores data in the aggregate cache until it completes aggregate calculations.When u run a session that uses an aggregator transformation,the informatica server creates index and data caches in memory to process the transformation.If the informatica server requires more space,it stores overflow values in cache files.

34.What r the diffrence between joiner transformation and source qualifier transformation? U can join hetrogenious data sources in joiner transformation which we can not achieve in source qualifier transformation. U need matching keys to join two relational sources in source qualifier transformation.Where as u doesn’t need matching keys to join two sources. Two relational sources should come from same datasource in sourcequalifier.U can join relatinal sources which r coming from diffrent sources also.

Page 15: Interview Questions for DWH Full Allignment

35. In which condtions we can not use joiner transformation(Limitaions of joiner transformation)? Both pipelines begin with the same original data source. Both input pipelines originate from the same Source Qualifier transformation. Both input pipelines originate from the same Normalizer transformation. Both input pipelines originate from the same Joiner transformation. Either input pipelines contains an Update Strategy transformation. Either input pipelines contains a connected or unconnected Sequence Generator transformation.

36.  what r the settiings that u use to cofigure the joiner transformation? Master and detail source Type of join Condition of the join

37. What r the join types in joiner transformation? Normal (Default) Master outer Detail outer Full outer

38.What r the joiner caches? When a Joiner transformation occurs in a session, the Informatica Server reads all the records from the master source and builds index and data caches based on the master rows. After building the caches, the Joiner transformation reads records from the detail source and perform joins.

39.what is the look up transformation?

Use lookup transformation in u’r mapping to lookup data in a relational table,view,synonym. Informatica server queries the look up table based on the lookup ports in the transformation.It compares the lookup transformation port values to lookup table column values based on the look up condition.

40. Why use the lookup transformation ?

To perform the following tasks.

Get a related value. For example, if your source table includes employee ID, but you want to include the employee name in your target table to make your summary data easier to read. Perform a calculation. Many normalized tables include values used in a calculation, such as gross sales per invoice or sales tax, but not the calculated value (such as net sales). Update slowly changing dimension tables. You can use a Lookup transformation to determine whether records already exist in the target.

41. What r the types of lookup?Connected and unconnected.

42. Differences between connected and unconnected lookup?

Connected lookup Unconnected lookup

Receives input values diectly from the pipe line.

Receives input values from the result of a  lkp expression in a another transformation.

U can use a dynamic or static cache U can use a static cache.

Cache includes all lookup columns used in the maping

Cache includes all lookup out put ports in the lookup condition and the lookup/return port.

Support user defined default values Does not support user defiend default43.What is meant by lookup caches? The informatica server builds a cache in memory when it processes the first row af a data in a cached look up transformation.It allocates memory for the cache based on the amount u configure in the transformation or session properties.The informatica server stores condition values in the index cache and output values in the data cache.

44. What r the types of lookup caches?

Page 16: Interview Questions for DWH Full Allignment

Persistent cache: U can save the lookup cache files and reuse them the next time the informatica server processes a lookup transformation configured to use the cache.

Recache from database: If the persistent cache is not synchronized with he lookup table,U can configure the lookup transformation to rebuild the lookup cache.

Static cache: U can configure a static or readonly cache for only lookup table.By default informatica server creates a static cache.It caches the lookup table and lookup values in the cache for each row that comes into the transformation.when the lookup condition is true,the informatica server does not update the cache while it prosesses the lookup transformation.

Dynamic cache: If u want to cache the target table and insert new rows into cache and the target,u can create a look up transformation to use dynamic cache.The informatica server dynamically inerts data to the target table.

shared cache: U can share the lookup cache between multiple transactions.U can share unnamed cache between transformations inthe same maping.

45.Difference between static cache and dynamic cache

Static cache Dynamic cache

U can not insert or update the cacheU can insert rows into the cache as u pass to the target

The informatic server returns a value from the lookup table or cache when the condition is true.When the condition is not true, informatica server returns the default value for connected transformations and null for unconnected transformations.

The informatic server inserts rows into cache when the condition is false.This indicates that the the row is not in the cache or target table. U can pass these rows to the target table

46.Which transformation should we use to normalize the COBOL and relational sources? Normalizer Transformation. When U drag the COBOL source in to the mapping Designer workspace, the normalizer transformation automatically appears, creating input and output ports for every column in the source.

47.How the informatica server sorts the string values in Ranktransformation? When the informatica server runs in the ASCII data movement mode it sorts session data using Binary sortorder.If U configure the seeion to use a binary sort order,the informatica server caluculates the binary value of each string and returns the specified number of rows with the higest binary values for the string.

48.What r the rank caches?

During the session, the informatica server compares an inout row with rows in the datacache.If the input row out-ranks a stored row, the informatica server replaces the stored row with the input row. The informatica server stores group information in an index cache and row data in a data cache.

49. What is the Rankindex in Ranktransformation?

The Designer automatically creates a RANKINDEX port for each Rank transformation. The Informatica Server uses the Rank Index port to store the ranking position for each record in a group. For example, if you create a Rank transformation that ranks the top 5 salespersons for each quarter, the rank index numbers the salespeople from 1 to 5:

50. What is the Router transformation? A Router transformation is similar to a Filter transformation because both transformations allow you to use a condition to test data. However, a Filter transformation tests data for one condition and drops the rows of data that do not meet the condition. A Router transformation tests data for one or more conditions and gives you the option to route rows of data that do not meet any of the conditions to a default output group. If you need to test the same input data based on multiple conditions, use a Router Transformation in a mapping instead of creating multiple Filter transformations to perform the same task.

Page 17: Interview Questions for DWH Full Allignment

51.What r the types of groups in Router transformation?Input group Output group

The designer copies property information from the input ports of the input group to create a set of output ports for each output group. Two types of output groups User defined groups Default group U can not modify or delete default groups.

52. Why we use stored procedure transformation?For populating and maintaining data bases.

53. What r the types of data that passes between informatica server and stored procedure? 3 types of data Input/Out put parameters Return Values Status code.

54. What is the status code?

Status code provides error handling for the informatica server during the session.The stored procedure issues a status code that notifies whether or not stored procedure completed sucessfully.This value can not seen by the user.It only used by the informatica server to determine whether to continue running the session or stop.

55. What is source qualifier transformation?

When U add a relational or a flat file source definition to a mapping need to connect it to a source qualifier transformation.The source qualifier transformation represents the records that the informatica server reads when it runs a session

56.What r the tasks that source qualifier performs? Join data originating from same source data base. Filter records when the informatica server reads source data. Specify an outer join rather than the default inner join specify sorted records. Select only distinct values from the source. Creating custom query to issue a special SELECT statement for the informatica server to read source data.

57.What is the target load order? U specify the target loadorder based on source qualifiers in a maping.If u have the multiple source qualifiers connected to the multiple targets,U can designatethe order in which informatica server loads data into the targets.

58. What is the default join that source qualifier provides? Inner equi join

59.What r the basic needs to join two sources in a source qualifier? Two sources should have primary and Foreign key relation ships. Two sources should have matching data types.

60.What is update strategy transformation ? This transformation is used to maintain the history data or just most recent changes in to target table.

Page 18: Interview Questions for DWH Full Allignment

61.Describe two levels in which update strategy transformation sets? Within a session. When you configure a session, you can instruct the Informatica Server to either treat all records in the same way (for example, treat all records as inserts), or use instructions coded into the session mapping to flag records for different database operations. Within a mapping. Within a mapping, you use the Update Strategy transformation to flag records for insert, delete, update, or reject.

62.What is the default source option for update stratgey transformation? Data driven

63.What is Datadriven? The informatica server follows instructions coded into update strategy transformations with in the session maping determine how to flag records for insert, update, delete or reject. If u do not choose data driven option setting,the informatica server ignores all update strategy transformations in the mapping.

64.What r the options in the target session of update strategy transsformatioin? Insert Delete Update Update as update Update as insert Update esle insert Truncate table

65.What r the types of mapping wizards that r to be provided in Informatica?

The Designer provides two mapping wizards to help you create mappings quickly and easily. Both wizards are designed to create mappings for loading and maintaining star schemas, a series of dimensions related to a central fact table.

Getting Started Wizard. Creates mappings to load static fact and dimension tables, as well as slowly growing dimension tables. Slowly Changing Dimensions Wizard. Creates mappings to load slowly changing dimension tables based on the amount of historical dimension data you want to keep and the method you choose to handle historical dimension data.

66.What r the types of maping in Getting Started Wizard? Simple Pass through mapping : Loads a static fact or dimension table by inserting all rows. Use this mapping when you want to drop all existing data from your table before loading new data. Slowly Growing target: Loads a slowly growing fact or dimension table by inserting new rows. Use this mapping to load new data when existing data does not require updates.

67.What r the mapings that we use for slowly changing dimension table? Type1: Rows containing changes to existing dimensions are updated in the target by overwriting the existing dimension. In the Type 1 Dimension mapping, all rows contain current dimension data. Use the Type 1 Dimension mapping to update a slowly changing dimension table when you do not need to keep any previous versions of dimensions in the table.

Type 2: The Type 2 Dimension Data mapping inserts both new and changed dimensions into the target. Changes are tracked in the target table by versioning the primary key and creating a version number for each dimension in the table. Use the Type 2 Dimension/Version Data mapping to update a slowly changing dimension table when you want to keep a full history of dimension data in the table. Version numbers and versioned primary keys track the order of changes to each dimension.

Type 3: The Type 3 Dimension mapping filters source rows based on user-defined comparisons and inserts only those found to be new dimensions to the target. Rows containing changes to existing dimensions are updated in the target. When updating an existing dimension, the Informatica Server saves existing data in different columns of the same row and replaces the existing data with the updates

Page 19: Interview Questions for DWH Full Allignment

68.What r the different types of Type2 dimension maping? Type2 Dimension/Version Data Maping: In this maping the updated dimension in the source will gets inserted in target along with a new version number.And newly added dimension in source will inserted into target with a primary key.

Type2 Dimension/Flag current Maping: This maping is also used for slowly changing dimensions.In addition it creates a flag value for changed or new dimension. Flag indiactes the dimension is new or newlyupdated.Recent dimensions will gets saved with cuurent flag value 1. And updated dimensions r saved with the value 0.

Type2 Dimension/Effective Date Range Mapping: This is also one flavor of Type2 mapping used for slowly changing dimensions. This mapping also inserts both new and changed dimensions in to the target. And changes r tracked by the effective date range for each version of each dimension.

69.How can u recognize whether or not the newly added rows in the source r gets insert in the target? In the Type2 mapping we have three options to recognize the newly added rows Version number Flag value Effective date Range

70.What r two types of processes that informatica runs the session?Load manager Process: Starts the session, creates the DTM process, and sends post-session email when the session completes. The DTM process. Creates threads to initialize the session, read, write, and transform data, and handle pre- and post-session operations.

71.What r the new features of the server manager in the informatica 5.0?U can use command line arguments for a session or batch. This allows U to change the values of session parameters, and mapping parameters and mapping variables. Parallel data processing: This feature is available for power center only. If we use the informatica server on a SMP system can use multiple CPU’s to process a session concurrently. Process session data using threads: Informatica server runs the session in two processes. Explained in previous question.

72.Can u generate reports in Informatcia?Yes. By using Metadata reporter we can generate reports in informatica.

73.What is metadata reporter?It is a web based application that enables you to run reports against repository metadata. With a Meta data reporter can access information about U’r repository with out having knowledge of sql, transformation language or underlying tables in the repository.

74. Define mapping and sessions?Mapping: It is a set of source and target definitions linked by transformation objects that define the rules for transformation. Session: It is a set of instructions that describe how and when to move data from source to targets.

75.Which tool U use to create and manage sessions and batches and to monitor and stop the informatica server?Informatica server manager.

76.Why we use partitioning the session in informatica?Performance can be improved by processing data in parallel in a single session by creating multiple partitions of the pipeline.Informatica server can achieve high performance by partitioning the pipeline and performing the extract , transformation, and load for each partition in parallel. 

77.To achieve the session partition what r the necessary tasks u have to do?Configure the session to partition source data. Install the informatica server on a machine with multiple CPU’s.

Page 20: Interview Questions for DWH Full Allignment

78.How the informatica server increases the session performance through partitioning the source?For a relational sources informatica server creates multiple connections for each parttion of a single source and extracts separate range of data for each connection.Informatica server reads multiple partitions of a single source concurently.Similarly for loading also informatica server creates multiple connections to the target and loads partitions of data concurently.

For XML and file sources,informatica server reads multiple files concurently.For loading the data informatica server creates a separate file for each partition(of a source file).U can choose to merge the targets.

79.Why u use repository connectivity?When u edit,schedule the sesion each time,informatica server directly communicates the repository to check whether or not the session and users r valid.All the metadata of sessions and mappings will be stored in repository.

80.What r the tasks that Loadmanger process will do?Manages the session and batch scheduling: When u start the informatica server the load maneger launches and queries the repository for a list of sessions configured to run on the informatica server.When u configure the session the load manager maintains list of list of sessions and session start times. When u sart a session load manger fetches the session information from the repository to perform the validations and verifications prior to starting DTM process. Locking and reading the session: When the informatica server starts a session load manager locks the session from the repository. Locking prevents U starting the session again and again. Reading the parameter file: If the session uses a parameter files, load manager reads the parameter file and verifies that the session level parameters are declared in the file Verifies permission and privileges: When the sesson starts load manger checks whether or not the user have privileges to run the session. Creating log files: Load manger creates log file contains the status of session.

81.What is DTM process?After the loadmanger performs validations for session,it creates the DTM process.DTM is to create and manage the threads that carry out the session tasks.I creates the master thread.Master thread creates and manges all the other threads.

82.What r the different threads in DTM process?Master thread: Creates and manages all other threads

Mapping thread: One mapping thread will be creates for each session.Fectches session and mapping information.

Pre and post session threads: This will be created to perform pre and post session operations.

Reader thread: One thread will be created for each partition of a source.It reads data from source.

Writer thread: It will be created to load data to the target.

83.What r the data movement modes in informatcia? Data movement modes determines how informatcia server handles the character data.U choose the data movement in the informatica server configuration settings. Two types of data movement modes available in Informatica. ASCIImode Uni code mode.

84.What r the out put files that the informatica server creates during the session running?Informatica server log: Informatica server(on unix) creates a log for all status and error messages(default name: pm.server.log).It also creates an error log for error messages.These files will be created in informatica home directory.

Session log file: Informatica server creates session log file for each session.It writes information about session into log files such as initialization process, creation of sql commands for reader and writer threads, errors encountered and load summary. The amount of detail in session log file depends on the tracing level that u set.

Page 21: Interview Questions for DWH Full Allignment

Session detail file: This file contains load statistics for each targets in mapping. Session detail include information such as table name, number of rows written or rejected. U can view this file by double clicking on the session in monitor window

Performance detail file: This file contains information known as session performance details which helps U where performance can be improved. To generate this file select the performance detail option in the session property sheet.

Reject file: This file contains the rows of data that the writer does not write to targets.

Control file: Informatica server creates control file and a target file when U runs a session that uses the external loader. The control file contains the information about the target flat file such as data format and loading instructions for the external loader.

Post session email: Post session email allows U to automatically communicate information about a session run to designated recipients can create two different messages. One if the session completed successfully the other if the session fails.

Indicator file: If u use the flat file as a target can configure the informatica server to create indicator file. For each target row, the indicator file contains a number to indicate whether the row was marked for insert, update, delete or reject.

Output file: If session writes to a target file, the informatica server creates the target file based on file properties entered in the session property sheet.

Cache files: When the informatica server creates memory cache it also creates cache files. For the following circumstances informatica server creates index and data cache files.

Aggregator transformation Joiner transformation Rank transformation Lookup transformation

85.In which circumstances that informatica server creates Reject files?When it encounters the DD_ Reject in update strategy transformation. Violates database constraint Filed in the rows was truncated or overflowed.

86.What is polling?It displays the updated information about the session in the monitor window. The monitor window displays the status of each session when U poll the informatica server

87.Can u copy the session to a different folder or repository?Yes. By using copy session wizard u can copy a session in a different folder or repository. But that target folder or repository should consists of mapping of that session. If target folder or repository is not having the mapping of copying session, u should have to copy that mapping first before u copy the session

88.What is batch and describe about types of batches?Grouping of session is known as batch. Batches r two types Sequential: Runs sessions one after the other Concurrent: Runs session at same time. If u has sessions with source-target dependencies u have to go for sequential batch to start the sessions one after another. If u have several independent session’s u can use concurrent batches. Which runs all the sessions at the same time.

89.Can u copy the batches?NO

90.How many number of sessions that u can create in a batch?Any number of sessions.

91.When the informatica server marks that a batch is failed?If one of session is configured to "run if previous completes" and that previous session fails.

Page 22: Interview Questions for DWH Full Allignment

92.What is a command that used to run a batch?pmcmd is used to start a batch.

93.What r the different options used to configure the sequential batches?Two options Run the session only if previous session completes successfully. Always runs the session.

94.In a sequential batch can u run the session if previous session fails?Yes.By setting the option always runs the session.

95.Can u start a batches with in a batch?U can not. If u wants to start batch that resides in a batch, create a new independent batch and copy the necessary sessions into the new batch.

96.Can u start a session inside a batch individually?We can start our required session only in case of sequential batch. in case of concurrent batch we cant do like this.

97.How can u stop a batch?By using server manager or pmcmd.

98.What r the session parameters?Session parameters r like mapping parameters, represent values U might want to change between sessions such as database connections or source files. Server manager also allows U to create user defined session parameters. Following r user defined session parameters.

Database connections Source file names: use this parameter when u wants to change the name or location of session source file between session runs

Target file name: Use this parameter when u want to change the name or location of session target file between session runs.

Reject file name: Use this parameter when u want to change the name or location of session reject files between session runs.

99.What is parameter file?Parameter file is to define the values for parameters and variables used in a session. A parameter file is a file created by text editor such as word pad or notepad. U can define the following values in parameter file

Mapping parameters Mapping variables session parameters

100.How can u access the remote source into U’r session?Relational source: To access relational source which is situated in a remote place, u need to configure database connection to the data source.

File Source: To access the remote source file U must configure the FTP connection to the host machine before u create the session.

Heterogeneous: When U’r mapping contains more than one source type, the server manager creates a heterogeneous session that displays source options for all types.

101.What r the transformations that restricts the partitioning of sessions?Advanced External procedure transformation and External procedure transformation: This transformation contains a check box on the properties tab to allow partitioning. Aggregator Transformation: If u use sorted ports u can not partition the associated source Joiner Transformation: U can not partition the master source for a joiner transformation Normalizer Transformation XML targets.

Page 23: Interview Questions for DWH Full Allignment

102.Performance tuning in Informatica?The goal of performance tuning is optimize session performance so sessions run during the available load window for the Informatica Server. Increase the session performance by following. The performance of the Informatica Server is related to network connections. Data generally moves across a network at less than 1 MB per second, whereas a local disk moves data five to twenty times faster. Thus network connections ofteny affect on session performance. So aviod netwrok connections. Flat files: If u’r flat files stored on a machine other than the informatca server, move those files to the machine that consists of informatica server. Relational data sources: Minimize the connections to sources, targets and informatica server to improve session performance. Moving target database into server system may improve session performance. Staging areas: If u use staging areas u force informatica server to perform multiple data passes. Removing of staging areas may improve session performance. U can run the multiple informatica servers against the same repository.Distibuting the session load to multiple informatica servers may improve session performance. Run the informatica server in ASCII data movement mode improves the session performance. Because ASCII data movement mode stores a character value in one byte. Unicode mode takes 2 bytes to store a character. If a session joins multiple source tables in one Source Qualifier, optimizing the query may improve performance. Also, single table select statements with an ORDER BY or GROUP BY clause may benefit from optimization such as adding indexes. We can improve the session performance by configuring the network packet size, which allows data to cross the network at one time.To do this go to server manger ,choose server configure database connections. If u r target consists key constraints and indexes u slow the loading of data. To improve the session performance in this case drop constraints and indexes before u run the session and rebuild them after completion of session. Running parallel sessions by using concurrent batches will also reduce the time of loading the data. So concurrent batches may also increase the session performance. Partitioning the session improves the session performance by creating multiple connections to sources and targets and loads data in paralel pipe lines. In some cases if a session contains a aggregator transformation ,u can use incremental aggregation to improve session performance. Avoid transformation errors to improve the session performance. If the sessioin contained lookup transformation u can improve the session performance by enabling the look up cache. If U’r session contains filter transformation, create that filter transformation nearer to the sources or u can use filter condition in source qualifier. Aggregator, Rank and joiner transformation may often decrease the session performance .Because they must group data before processing it. To improve session performance in this case use sorted ports option.

103.What is difference between maplet and reusable transformation?Mapplet consists of set of transformations that is reusable. A reusable transformation is a single transformation that can be reusable. If u creates a variables or parameters in maplet that can not be used in another mapping or maplet.Unlike the variables that r created in a reusable transformation can be useful in any other maping or maplet. We can not include source definitions in reusable transformations.But we can add sources to a maplet. Whole transformation logic will be hided in case of maplet.But it is transparent in case of reusable transformation. We cant use COBOL source qualifier, joiner, normalizer transformations in mapplet. Where as we can make them as a reusable transformations.

104.Define informatica repository?The Informatica repository is a relational database that stores information, or metadata, used by the Informatica Server and Client tools. Metadata can include information such as mappings describing how to transform source data, sessions indicating when you want the Informatica Server to perform the transformations, and connect strings for sources and targets. The repository also stores administrative information such as usernames and passwords, permissions and privileges, and product version. Use repository manager to create the repository. The Repository Manager connects to the repository database and runs the code needed to create the repository tables. These tables stores metadata in specific format the informatica server, client tools use.

105.What r the types of metadata that stores in repository?Following r the types of metadata that stores in the repository Database connections Global objects Mappings Mapplets Multidimensional metadata Reusable transformations Sessions and batches Short cuts

Page 24: Interview Questions for DWH Full Allignment

Source definitions Target definitions Transformations

106.What is power center repository?The Power Center repository allows you to share metadata across repositories to create a data mart domain. In a data mart domain, you can create a single global repository to store metadata used across an enterprise, and a number of local repositories to share the global metadata as needed.

107.What r the new features in Informatica 5.0?U can Debug U’r mapping in mapping designer U can view the work space over the entire screen The designer displays a new icon for a invalid mappings in the navigator window U can use a dynamic lookup cache in a lookup transformation Create mapping parameters or mapping variables in a mapping or maplet to make mappings more flexible U can export objects into repository and import objects from repository. when u export a repository object, the designer or server manager creates an XML file to describe the repository metadata. The designer allows u to use Router transformation to test data for multiple conditions. Router transformation allows u route groups of data to transformation or target. U can use XML data as a source or target.

Server Enhancements:

U can use the command line program pmcmd to specify a parameter file to run sessions or batches. This allows you to change the values of session parameters, and mapping parameters and variables at runtime.

If you run the Informatica Server on a symmetric multi-processing system, you can use multiple CPUs to process a session concurrently. You configure partitions in the session properties based on source qualifiers. The Informatica Server reads, transforms, and writes partitions of data in parallel for a single session. This is available for Power center only.

Informatica server creates two processes like load manager process, TM process to run the sessions.

Metadata Reporter: It is a web based application which is used to run reports against repository metadata.

U can copy the session across the folders and repositories using the copy session wizard in the informatica server manager

with new email variables, you can configure post-session email to include information, such as the mapping used during the session

108.What is incremental aggregation?When using incremental aggregation, you apply captured changes in the source to aggregate calculations in a session. If the source changes only incrementally and you can capture changes, you can configure the session to process only those changes. This allows the Informatica Server to update your target incrementally, rather than forcing it to process the entire source and recalculate the same calculations each time you run the session.

109.What r the scheduling options to run a session?U can schedule a session to run at a given time or interval, or u can manually run the session. Different options of scheduling Run only on demand: server runs the session only when user starts session explicitly Run once: Informatica server runs the session only once at a specified date and time. Run every: Informatica server runs the session at regular intervals as u configured. Customized repeat: Informatica server runs the session at the data and times specified in the repeat dialog box.

Page 25: Interview Questions for DWH Full Allignment

110.What is difference between stored procedure transformation and external procedure transformation?In case of stored procedure transformation procedure will be compiled and executed in a relational data source. U need data base connection to import the stored procedure in to u’r maping.Where as in external procedure transformation procedure or function will be executed out side of data source. I.e. u need to make it as a DLL to access in u r maping.No need to have data base connection in case of external procedure transformation.

111.Explain about Recovering sessions?If you stop a session or if an error causes a session to stop, refer to the session and error logs to determine the cause of failure. Correct the errors, and then complete the session. The method you use to complete the session depends on the properties of the mapping, session, and Informatica Server configuration. Use one of the following methods to complete the session: · Run the session again if the Informatica Server has not issued a commit. · Truncate the target tables and run the session again if the session is not recoverable. · Consider performing recovery if the Informatica Server has issued at least one commit.

112.If a session fails after loading of 10,000 records in to the target. How can u load the records from 10001 th record when u run the session next time?As explained above informatcia server has 3 methods to recovering the sessions. Use performing recovery to load the records from where the session fails.

113.How to recover the standalone session?A standalone session is a session that is not nested in a batch. If a standalone session fails, you can run recovery using a menu command or pmcmd. These options are not available for batched sessions.

To recover sessions using the menu:

1. In the Server Manager, highlight the session you want to recover. 2. Select Server Requests-Stop from the menu. 3. With the failed session highlighted, select Server Requests-Start Session in Recovery Mode from the menu.

To recover sessions using pmcmd: 1.From the command line, stop the session. 2. From the command line, start recovery.

114.If i done any modifications for my table in back end does it reflect in informatca warehouse or maping desginer or source analyzer?NO. Informatica is not at all concern with back end data base. It displays u all the information that is to be stored in repository. If want to reflect back end changes to informatica screens, again u have to import from back end to informatica by valid connection. And u have to replace the existing files with imported files.

115.After draging the ports of three sources(sql server,oracle,informix) to a single source qualifier, can u map these three ports directly to target?NO.Unless and until u join those three ports in source qualifier u cannot map them directly.

116.What is Data cleansing..?The process of finding and removing or correcting data that is incorrect, out-of-date, redundant, incomplete, or formatted incorrectly.

117.How can we partition a session in Informatica?The Informatica® Power Center® partitioning option optimizes parallel processing on multiprocessor hardware by providing a thread-based architecture and built-in data partitioning. GUI-based tools reduce the development effort necessary to create data partitions and streamline ongoing troubleshooting and performance tuning tasks, while ensuring data integrity throughout the execution process. As the amount of data within an organization expands and real-time demand for information grows, the Power Center Partitioning option enables hardware and applications to provide outstanding performance and jointly scale to handle large volumes of data and users.

Page 26: Interview Questions for DWH Full Allignment

118.What is a time dimension? give an example.

Time dimension is one of important in Datawarehouse. Whenever u genetated the report , that time u access all data from thro time dimension.  eg. Employee time dimension  Fields: Date key, full date, day of week, day , month, quarter, fiscal year.

119. Diff between informatica repository server & informatica serverInformatica Repository Server: It’s manages connections to the repository from client application. Informatica Server: Its extracts the source data, performs the data transformation, and loads the transformed data into the target

120.Explain the informatica Architecture in detail ?Informatica server connects source data and target data using native odbc drivers again it connect to the repository for running sessions and retrieving metadata information

Source------>informatica server--------->target                                       |                                       |                                REPOSITORY  

121.Discuss the advantages & Disadvantages of star & snowflake schema?In a star schema every dimension will have a primary key.  In a star schema, a dimension table will not have any parent table.  Whereas in a snow flake schema, a dimension table will have one or more parent tables.  Hierarchies for the dimensions are stored in the dimensional table itself in star schema.  Whereas hierarchies are broken into separate tables in snow flake schema. These hierarchies helps to drill down the data from topmost hierarchies to the lowermost hierarchies.star schema consists of single fact table surrounded by some dimensional table.In snowflake schema the dimension tables are connected with some sub dimension table.In star flake dimensional tables r denormalized,in snowflake dimension tables r normalized.star schema is used for report generation ,snowflake schema is used for cube.The advantage of snowflake schema is that the normalized table’s r easier to maintain. it also saves the storage space.The disadvantage of snowflake schema is that it reduces the effectiveness of navigation across the tables due to large no of joins between them.

122.What are main advantages and purpose of using Normalizer Transformation in Informatica?Normalizer Transformation is used mainly with COBOL sources where most of the time data is stored in de-normalized format. Also, Normalizer transformation can be used to create multiple rows from a single row of data.

123.How to read rejected data or bad data from bad file and reload it to target?Correction the rejected data and send to target relational tables using loadorder utility. Find out the rejected data by using column indicatior and row indicator.

124.How do you transfert the data from data warehouse to flatfile?You can write a mapping with the flat file as a target using a DUMMY_CONNECTION. A flat file target is built by pulling a source into target space using Warehouse Designer tool.

125.At the max how many tranformations can be us in a mapping?There is no such limitation to use this number of transformations. But in performance point of view using too many transformations will reduce the session performance. My idea is "if needed more transformations to use in a mapping its better to go for some stored procedure."  

Page 27: Interview Questions for DWH Full Allignment

126.What is the difference between Normal load and Bulk load?Normal Load: Normal load will write information to the database log file so that if any recorvery is needed it is will be helpful. When the source file is a text file and loading data to a table,in such cases we should you normal load only, else the session will be failed.Bulk Mode: Bulk load will not write information to the database log file so that if any recorvery is needed we can't do any thing in such cases. Comparatively Bulk load is pretty faster than normal load.

127.What is a junk dimensionA "junk" dimension is a collection of random transactional codes, flags and/or text attributes that are unrelated to any particular dimension. The junk dimension is simply a structure that provides a convenient place to store the junk attributes. A good example would be a trade fact in a company that brokers equity trades.

128.Can we lookup a table from a source qualifer transformation-unconnected lookup?No. we can't do. I will explain you why. 1) Unless you assign the output of the source qualifier to another transformation or to target no way it will include the field in the query.2) Source qualifier doesn’t have any variables fields to utilize as expression.

129.How to get the first 100 rows from the flat file into the target?Task ----->(link)   session (workflow manager)Double click on link and type $$source success rows(parameter in session variables) = 100it should automatically stops session.

130.Can we modify the data in flat file?

131.Difference between summary filter and details filter?Summary Filter --- we can apply records group by that contain common values.Detail Filter --- we can apply to each and every record in a database.

132.What are the difference between view and materialized view?Materialized views are schema objects that can be used to summarize, precompiled, replicate, and distribute data. E.g. to construct a data warehouse.A materialized view provides indirect access to table data by storing the results of a query in a separate schema object. Unlike an ordinary view, which does not take up any storage space or contain any data

133.Compare Data Warehousing Top-Down approach with Bottom-up approach?Top downODS-->ETL-->Datawarehouse-->Datamart-->OLAP

Bottom upODS-->ETL-->Datamart-->Datawarehouse-->OLAPBottom approach is the best because in 3 tier architecture data tier is the bottom one.

134.Discuss which is better among incremental load, Normal Load and Bulk load?Normal load is the best.

135.What is the difference between connected and unconnected stored procedures?Unconnected:The unconnected Stored Procedure transformation is not connected directly to the flow of the mapping. It either runs before or after the session, or is called by an expression in another transformation in the mapping.Connected:The flow of data through a mapping in connected mode also passes through the Stored Procedure transformation. All data entering the transformation through the input ports affects the stored procedure. You should use a connected Stored Procedure transformation when you need data from an input port sent as an input parameter to the stored procedure, or the results of a stored procedure sent as an output parameter to another transformation.

Page 28: Interview Questions for DWH Full Allignment

Run a stored procedure before or after your session. Unconnected

 Run a stored procedure once during your mapping, such as pre- or post-session.

Unconnected

 Run a stored procedure every time a row passes through the Stored Procedure transformation.

Connected or Unconnected

 Run a stored procedure based on data that passes through the mapping, such as when a specific port does not contain a null value.

Unconnected

 Pass parameters to the stored procedure and receive a single output parameter.

Connected or Unconnected

 Pass parameters to the stored procedure and receive multiple output parameters. Note: To get multiple output parameters from an unconnected Stored Procedure transformation, you must create variables for each output parameter. For details, see Calling a Stored Procedure From an Expression.

Connected or Unconnected

Run nested stored procedures. Unconnected

 Call multiple times within a mapping.

Unconnected

136.Differences between Informatica 6.2 and Informatica 7.0In 7.0 introduce custom transformation and union transformation and also flat file lookup condition.Features in 7.1 are:1. Union and custom transformation2. Lookup on flat file3. Grid servers working on different operating systems can coexist on same server4. We can use pmcmdrep5. We can export independent and dependent rep objects6. We ca move mapping in any web application7. Version controlling8. Data profiling.

137.What are the Differences between Informatica Power Center versions 6.2 and 7.1, also between Versions 6.2 and 5.1?VersioningLDAP authenticationSupport of 64 bit architectures.

138.What is the diff between Informatica powercenter server, repositoryserver and repository?Power center server contains the scheduled runs at which time data should load from source to targetRepository contains all the definitions of the mappings done in designer.

139.How to create the staging area in your database?A Staging area in a DW is used as a temporary space to hold all the records from the source system. So more or less it should be exact replica of the source systems except for the load strategy where we use truncate and reload options. So create using the same layout as in your source tables or using the Generate SQL option in the Warehouse Designer tab.

140.What does the expression n filter transformations do in Informatica Slowly growing target wizard?EXPESSION transformation detects and flags the rows from source.Filter transformation filters the rows that are not flagged and passes the flagged rows to the Update strategy transformation

Page 29: Interview Questions for DWH Full Allignment

141.In a filter expression we want to compare one date field with a db2 system field CURRENT DATE.Our Syntax: datefield = CURRENT DATE (we didn't define it by ports, its a system field ), but this is not valid (PMParser: Missing Operator)..The db2 date format is  "yyyymmdd"  where as sysdate in oracle will give "dd-mm-yy" so conversion of db2 date format to local database date format is compulsory. Other wise u will get that type of error.

142.Briefly explian the Versioning Concept in Power Center 7.1?In power center 7.1 use 9 Tem servers i.e. add in Look up. But in power center 6.x use only 8 tem servers. And add 5 transformations. In 6.x any 17 transformations but 7.x uses 22 transformations.

143.How to join two tables without using the Joiner Transformation?Its possible to join the two or more tables by using source qualifier. But provided the tables should have relationship. When u drag n drop the table’s u will getting the source qualifier for each table. Delete all the source qualifiers. Add a common source qualifier for all. Right click on the source qualifier u will find EDIT click on it. Click on the properties tab will find sql query in that u can write ur sql.144.Can Informatica be used as a Cleansing Tool? If Yes, give example of transformations that can implement a data cleansing routine.Yes, we can use Informatica for cleansing data. Some time we use stages to cleansing the data. It depends upon performance again else we can use expression to cleansing data.For example an field X have some values and other with Null values and assigned to target feild where target field is not null column, inside an expression we can assign space or some constant value to avoid session failure.The input data is in one format and target is in another format, we can change the format in expression.we can assign some default values to the target to represent complete set of data in the target.

145.How do you decide whether you need do aggregations at database level or at Informatica level?It depends upon our requirement only. If you have good processing database you can create aggregation table or view at database level else its better to use informatica. Here i'm explain why we need to use informatica. what ever it may be informatica is a third party tool, so it will take more time to process aggregation compared to the database, but in Informatica  an option we  called "Incremental aggregation" which will help you to update the current values with current  values +new values. No necessary to process entire values again and again. Unless this can be done if nobody deleted that cache files. If that happened total aggregation we need to execute on informatica also. In database we don't have Incremental aggregation facility.

146.How do we estimate the depth of the session scheduling queue? Where do we set the number of maximum concurrent sessions that Informatica can run at a given time?

147.How do we estimate the number of partitions that a mapping really requires? Is it dependent on the machine configuration?It depends upon the informatica version we r using. Suppose if we r using informatica 6 it supports only 32 partitions where as informatica 7 supports 64 partitions.

148.Suppose session is configured with commit interval of 10,000 rows and source has 50,000 rows. Explain the commit points for Source based commit and Target based commit. Assume appropriate value wherever required?Source based commit will commit the data into target based on commit interval. So, for every 10,000 rows it will commit into target.Target based commit will commit the data into target based on buffer size of the target. i.e., it commits the data into target when ever the buffer fills. Let us assume that the buffer size is 6, 000.So, for every 6,000 rows it commits the data.

149.We are using Update Strategy Transformation in mapping how can we know whether insert or update or reject or delete option has been selected during running of sessions in Informatica?In Designer while creating Update Strategy Transformation uncheck "forward to next transformation". If any rejected rows are there automatically it will be updated to the session log file. Update or insert files are known by checking the target file or table only. Operation           Constant               Numeric valueInsert                  DD_INSERT                      0Update                DD_UPDATE                    1Delete                 DD_DELETE                     2Reject                  DD_REJECT                    3

Page 30: Interview Questions for DWH Full Allignment

150.What is the procedure to write the query to list the highest salary of three employees?SELECT salFROM (SELECT sal FROM my_table ORDER BY sal DESC)WHERE ROWNUM < 4;

151.What is the limit to the number of sources and targets you can have in a mapping?There is one formula..no.of bloccks=0.9*( DTM buffer size/block size)*no.of partitions.here no.of blocks=(source+targets)*2the restriction is only on the database side. How many concurrent threads r u allowed to run on the db server

152.Which is better among connected lookup and unconnected lookup transformations in informatica or any other ETL tool?Its not a easy question to say which is better out of connected, unconnected lookups. Its depends upon our experience and upon the requirement.When you compared both basically connected lookup will return more values and unconnected returns one value. Conn lookup is in the same pipeline of source and it will accept dynamic caching. Unconn lookup don't have that facility but in some special cases we can use Unconnected. if o/p of one lookup is going as i/p of another lookup this unconnected lookups are favorable.

153.In Dimensional modeling fact table is normalized or denormalized?in case of star schema and incase of snow flake schema?In Dimensional modeling, Star Schema: A Single Fact table will be surrounded by a group of Dimensional tables comprise of de- normalized data Snowflake Schema: A Single Fact table will be surrounded by a group of Dimensional tables comprised of normalized dataThe Star Schema (sometimes referenced as star join schema) is the simplest data warehouse schema, consisting of a single "fact table" with a compound primary key, with one segment for each "dimension" and with additional columns of additive, numeric facts.The Star Schema makes multi-dimensional database (MDDB) functionality possible using a traditional relational database. Because relational databases are the most common data management system in organizations today, implementing multi-dimensional views of data using a relational database is very appealing. Even if you are using a specific MDDB solution, its sources likely are relational databases. Another reason for using star schema is its ease of understanding. Fact tables in star schema are mostly in third normal form (3NF), but dimensional tables in de-normalized second normal form (2NF). If you want to normalize dimensional tables, they look like snowflakes (see snowflake schema) and the same problems of relational databases arise - you need complex queries and business users cannot easily understand the meaning of data. Although query performance may be improved by advanced DBMS technology and hardware, highly normalized tables make reporting difficult and applications complex.The Snowflake Schema is a more complex data warehouse model than a star schema, and is a type of star schema. It is called a snowflake schema because the diagram of the schema resembles a snowflake.Snowflake schemas normalize dimensions to eliminate redundancy. That is, the dimension data has been grouped into multiple tables instead of one large table. For example, a product dimension table in a star schema might be normalized into a products table, a Product-category table, and a product-manufacturer table in a snowflake schema. While this saves space, it increases the number of dimension tables and requires more foreign key joins. The result is more complex queries and reduced query performance.Star schema--De-Normalized DimensionsSnow Flake Schema-- Normalized Dimensions

154.What is difference between IIF and DECODE function?You can use nested IIF statements to test multiple conditions. The following example tests for various conditions and returns 0 if sales is zero or negative:

IIF( SALES > 0, IIF( SALES < 50, SALARY1, IIF( SALES < 100, SALARY2, IIF( SALES < 200, SALARY3, BONUS))), 0 )

You can use DECODE instead of IIF in many cases. DECODE may improve readability. The following shows how you can use DECODE instead of IIF :   SALES > 0 and SALES < 50, SALARY1,

   SALES > 49 AND SALES < 100, SALARY2,   SALES > 99 AND SALES < 200, SALARY3,   SALES > 199, BONUS)

Page 31: Interview Questions for DWH Full Allignment

155.What are variable ports and list two situations when they can be used?We have mainly tree ports Inport, Outport, Variable port. Inport represents data is flowing into transformation. Outport is used when data is mapped to next transformation. Variable port is used when we mathematical caluculations are required. If any addition i will be more than happy if you can share.

156.How does the server recognise the source and target databases?By using ODBC connection.if it is relational.if is flat file FTP connection..see we can make sure with connection in the properties of session both sources && targets.

157.How to retrive the records from a rejected file. explane with syntax or example?There is one utility called "reject Loader" where we can findout the reject records.and able to refine and reload the rejected records..

158.How to lookup the data on multiple tabels?Using SQL override..we can lookup the Data on multiple tables.See in the properties.

159.What is the procedure to load the fact table.Give in detail?We use the 2 wizards (i.e) the getting started wizard and slowly changing dimension wizard to load the fact and dimension tables,by using these 2 wizards we can create different types of mappings according to the business requirements and load into the star schemas(fact and dimension tables).

160.What is the use of incremental aggregation? Explain me in brief with an example?It’s a session option. When the informatica server performs incremental aggregator. it passes new source data through the mapping and uses historical cache data to perform new aggregation calculations incrementally. For performance we will use it.

161.How to delete duplicate rows in flat files source is any option in informatica?Use a sorter transformation, in that u will have a "distinct" option make use of it.

162.How to use mapping parameters and what is their use?Mapping parameters and variables make the use of mappings more flexible. and also it avoids creating of multiple mappings. it helps in adding incremental data. Mapping parameters and variables has to create in the mapping designer by choosing the menu option as Mapping ----> parameters and variables and the enter the name for the variable or parameter but it has to be preceded by $$. and choose type as parameter/variable, data type once defined the variable/parameter is in the any expression for example in SQ transformation in the source filter properties tab. just enter filter condition and finally create a parameter file to assign the value for the variable / parameter and configure the session properties. however the final step is optional. if there parameter is npt present it uses the initial value which is assigned at the time of creating the variable

163.In the concept of mapping parameters and variables, the variable value will be saved to the repository after the completion of the session and the next time when u run the session, the server takes the saved variable value in the repository and starts assigning the next value of the saved value. for example i ran a session and in the end it stored a value of 50 to the repository.next time when i run the session, it should start with the value of 70. not with the value of 51.u can do one thing after running the mapping,, in workflow manager              Start-------->session. Right click on the session  u will get a menu, in that go for persistent values, there u will find the last value stored in the repository regarding to mapping variable. Then remove it and put ur desired one, run the session... i hope ur task will be done.

164.Significance of oracle 9i in informatica when compared to oracle 8 or 8i.i mean how is oracle 9i advantageous when compared to oracle 8 or 8i when used in informaticaIt’s very easy Actually oracle 8i not allowed user defined data types but 9i allows and then blob,clob allow only 9i not 8i and  more over list partition is there in 9i only.

165.Can we use aggregator/active transformation after update strategy transformation?We can use, but the update flag will not be remain. but we can use passive transformation

Page 32: Interview Questions for DWH Full Allignment

166.Why dimenstion tables are denormalized in nature ?Because in Data warehousing historical data should be maintained, to maintain historical data means suppose one employee details like where previously he worked, and now where he is working, all details should be maintain in one table, if u maintain primary key it won't allow the duplicate records with same employee id. so to maintain historical data we are all going for concept data warehousing by using surrogate keys we can achieve the historical data(using oracle sequence for critical column).so all the dimensions are marinating historical data, they are de normalized, because of duplicate entry means not exactly duplicate record with same employee number another record is maintaining in the table

167.Why dimenstion tables are denormalized in nature ?Because in Data warehousing historical data should be maintained, to maintain historical data means suppose one employee details like where previously he worked, and now where he is working, all details should be maintain in one table, if u maintain primary key it won't allow the duplicate records with same employee id. so to maintain historical data we are all going for concept data warehousing by using surrogate keys we can achieve the historical data(using oracle sequence for critical column).so all the dimensions are marinating historical data, they are de normalized, because of duplicate entry means not exactly duplicate record with same employee number another record is maintaining in the table.

168.In a sequential Batch how can we stop single session?We have a task called wait event using that we can stop. We start using raise event.This is as per my knowledge.

169.How do you handle decimal places while importing a flatfile into informatica?While importing the flat file, the flat file wizard helps in configuring the properties of the file so that select the numeric column and just enter the precision value and the scale. precision includes the scale for example if the number is 98888.654, enter precision as 8 and scale as 3 and width as 10 for fixed width flat file.

170.If you are workflow is running slow in informatica. Where do you start trouble shooting and what are the steps you follow?When the work flow is running slowly u has to find out the bottlenecks in this orderTargetSourceMappingSessionSystem

171.If you have four lookup tables in the workflow. How do you troubleshoot to improve performance?There r many ways to improve the mapping which has multiple lookups.1) We can create an index for the lookup table if we have permissions (staging area).2) Divide the lookup mapping into two (a) dedicate one for insert means: source - target,, these r new rows  . Only the new rows will come to mapping and the process will be fast. (b) Dedicate the second one to update: source=target, these r existing rows. Only the rows which exist already will come into the mapping.3) We can increase the cache size of the lookup.

172.Can anyone explain error handling in informatica with examples so that it will be easy to explain the same in the interview.You can create some generalized transformations to handle the errors and use them in your mapping. For example for data types create one generalized transformation and include in your mapping then you will know the errors where they are occurring.

173.How do I import VSAM files from source to target. Do I need a special plug-in?

174.Differences between Normalize and Normalizer transformation.?Normalizer: It is a transformation mainly using for Cobol sources,Its change the rows into columns and columns into rowsNormalization: To remove the redundancy and inconsistency.

Page 33: Interview Questions for DWH Full Allignment

175.What is IQD file?IQD file is nothing but Impromptu Query Definition, This file is maily used in Cognos Impromptu tool after creating a imr( report) we save the imr as IQD file which is used while creating a cube in power play transformer. In data source type we select Impromptu Query Definition.

176.What is data merging, data cleansing, sampling?Cleansing: ---TO identify and remove the retundacy and inconsistencySampling: just sample the data through send the data from source to target

177.How to import oracle sequence into Informatica?CREATE ONE PROCEDURE AND DECLARE THE SEQUENCE INSIDE THE PROCEDURE,FINALLY CALL THE PROCEDURE IN INFORMATICA WITH THE HELP OF STORED PROCEDURE TRANSFORMATION.

178. What are the steps required for type2 dimension/version data mapping. how can we implement it?Go to mapping designer in it go for mapping select wizard in it go for slowly changing dimension Here u will find a new window their u need to give the mapping name source table target table and type of scd then if select finish scd 2 mapping is created go to ware designer and generate the table then validate the mapping in mapping designer save it to repository run the session in workflow manager later update the source table and re run again u will find the difference in target table.

179.With out using Update strategy and sessions options, how we can do the update our target table?Using "update override" in target option.

180.Two relational tables are connected to SQ Trans,what are the possible errors it will be thrown?We can connect two relational tables in one sq Transformation. No errors will be performing.

181.What is the best way to show metadata(number of rows at source, target and each transformation level, error related data) in a report format?When your workflow get completed go to workflow monitor right click the session .then go to transformation statistics there we can see number of rows in source and target.if we go for session properties we can see errors related to data.

182.If u had to split the source level key going into two seperate tables. One as surrogate and other as primary. Since informatica does not gurantee keys are loaded properly(order!) into those tables. What are the different ways you could handle this type of situation?

183.How to append the records in flat file(Informatica) ? Where as in Datastage we have the options i) overwrite the existing fileii) Append existing fileThis is not there in Informatica v 7. but heard that its included in the latest version 8.0 where u can append to a flat file. Its about to be shipping in the market.

184.What are partition points?Partition points mark the thread boundaries in a source pipeline and divide the pipeline into stages.

185.What are cost based and rule based approaches and the difference?Cost based and rules based approaches are the optimization techniques which are used in related to databases, where we need to optimize a sql query. Basically Oracle provides Two types of Optimizers (indeed 3 but we use only these two techniques. bcz the third has some disadvantages.)When ever you process any sql query in Oracle, what oracle engine internally does is, it reads the query and decides which will the best possible way for executing the query. So in this process, Oracle follows these optimization techniques. 1. cost based Optimizer (CBO): If a sql query can be executed in 2 different ways (like may have path 1 and path2 for same query), then What CBO does is, it basically calculates the cost of each path and the analyses for which path the cost of execution is less and then executes that path so that it can optimize the query execution.2. Rule base optimizer (RBO): this basically follows the rules which are needed for executing a query. So depending on the number of rules which are to be applied, the optimizer runs the query.

Page 34: Interview Questions for DWH Full Allignment

Use: If the table you are trying to query is already analyzed, then oracle will go with CBO. If the table is not analyzed, the Oracle follows RBO. For the first time, if table is not analyzed, Oracle will go with full table scan.

186.What is mystery dimension?Using Mystery Dimension ur maintaining the mystery data in ur Project.

187.What is difference b/w Informatica 7.1 and Abinitio?In Informatica there is the concept of co-operating system, which makes the mapping in parallel fashionwhich  is not in Informatica.

188.Can i start and stop single session in concurrent batch?Just right click on the particular session and going to recovery option.

189.What is difference between lookup cache and uncached lookup?

190. Can i run the mapping with out starting the informatica server?The difference between cache and uncached lookup is when you configure the lookup transformation cache lookup it stores all the lookup table data in the cache when the first input record enter into the lookup transformation, in cache lookup the select statement executes only once and compares the values of the input record with the values in the cache but in uncache lookup the the select statement executes for each input record entering into the lookup transformation and it has to connect to database each time entering the new record.

191.What is the difference between stop and abort?Stop: If the session u want to stop is a part of batch you must stop the batch,if the batch is part of nested batch, Stop the outer most batchAbort: You can issue the abort command; it is similar to stop command except it has 60 second time out.If the server cannot finish processing and committing data with in 60 sec

192.Can we run a group of sessions without using workflow manager?Its Possible using pmcmd Command with out using the workflow Manager run the group of session.

193.How to perform a "Loop Scope / Loop condition" in an Informaticaprogram ? Give me few examples.

194.If a session fails after loading of 10,000 records in to the target.How can u load the records from 10001 th record when u run the session next time in informatica 6.1?Using performance recovery option.

195.I have an requirement where in the columns names in a table (Table A) should appear in rows of target table (Table B) i.e. converting columns to rows. Is it possible through Informatica? If so, how?If data in tables as follows Table AKey-1 char(3);table A values_______123

Table Bbkey-a char(3);bcode  char(1);table b values1 T1 A1 G2 A

Page 35: Interview Questions for DWH Full Allignment

2 T2 L3 Aand output required is as1, T, A2, A, T, L3, Athe SQL query in source qualifier should beselect key_1,          max(decode( bcode, 'T', bcode, null )) t_code,             max(decode( bcode, 'A', bcode, null )) a_code,          max(decode( bcode, 'L', bcode, null )) l_code    from a, b    where a.key_1 = b.bkey_a    group by key_1   /

196.What is meant by complex mapping?Complex mapping means involved in more logic and more business rules. Actually in my project complex mapping is In my bank project, I involved in construct a 1 data ware house Many customer is there in my bank project, They r after taking loans relocated in to another place that time I feel to difficult maintain both previous and current addresses in the sense i am using scd2.This is an simple example of complex mapping.

197.Explain use of update strategy transformation?Maintain the history data and maintain the most recent changes data.

198.What are mapping parameters and varibles in which situation we can use it?Mapping parameters have a constant value through out the session whereas in mapping variable the values change and the informatica server saves the values in the repository and uses next time when u run the session.

199.What is work let and what use of work let and in which situation we can use it?Worklet is a set of tasks. If a certain set of task has to be reused in many workflows then we use worklets. To execute a Worklet, it  has to be placed inside a workflow. The use of worklet in a workflow is similar to the use of mapplet in a mapping.

200.what is difference between dimention table and fact table and what are different dimention tables and fact tables?In the fact table contain measurable data and less columns and many rows, it’s contain primary keyDifferent types of fact tables:Additive, non additive, semi additiveIn the dimensions table contain textual description of data and also contain many columns, less rowsIts contain primary key.

201.How do you configure mapping in informatica?You should configure the mapping with the least number of transformations and expressions to do the most amount of work possible. You should minimize the amount of data moved by deleting unnecessary links between transformations. For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup transformations), limit connected input/output or output ports. Limiting the number of connected input/output or output ports reduces the amount of data the transformations store in the data cache. You can also perform the following tasks to optimize the mapping:

Configure single-pass reading. Optimize datatype conversions. Eliminate transformation errors. Optimize transformations. Optimize expressions. You should configure the mapping with the least number of transformations

and expressions to do the most amount of work possible. You should minimize the amount of data moved by deleting unnecessary links between transformations.

Page 36: Interview Questions for DWH Full Allignment

For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup transformations), limit connected input/output or output ports. Limiting the number of connected input/output or output ports reduces the amount of data the transformations store in the data cache. You can also perform the following tasks to optimize the mapping:

o Configure single-pass reading. o Optimize datatype conversions. o Eliminate transformation errors. o Optimize transformations. o Optimize expressions.

202.Can i use a session Bulk loading option that time can i make a recovery to the session?No, why because in bulk load u wont create redo log file, when u normal load we create redo log file, but in bulk load session performance increase.

203.What is lookup transformation and update strategy transformation and explain with an example?Look up transformation is used to lookup the data in a relation table,view,Synonym and Flat file.The informatica server queries the lookup table based on the lookup ports used in the transformation.It compares the lookup transformation port values to lookup table column values based on the lookup conditionBy using lookup we can get the realated value,Perform a caluclation and Update SCD.Two types of lookupsConnected UnconnectedUpdate strategy transformation This is used to control how the rows are flagged for insert, update ,delete or reject. To define a flagging of rows in a session it can be insert, Delete, Update or Data driven.

In Update we have three optionsUpdate as UpdateUpdate as insertUpdate else insert.

204.What is the difference between Power Centre and Power Mart? What is the procedure for creating Independent Data Marts from Informatica 7.1?Power center                         Power martNo. of repository                n No.                                        n No.Applicability                    high end WH                          low-minded range WHGlobal repository            supported                             not supportedLocal repository              supported                              supportedERP support                  available                               not available

205.In the source, if we also have duplicate records and we have 2 targets, T1- for unique values and T2- only for duplicate values. How do we pass the unique values to T1 and duplicate values to T2 from the source to these 2 different targets in a single mapping?This is not a right approach friends. There is a good practice of identifying duplicates. Normally when you ask someone how to identify a duplicate record in informatica, they say "Use aggregator transf". Well you can just get a count from this, but not really identify which record is a duplicate. If it is RDBMS, you can simply write a query "select ... from ...group by <key fields> having count (*) > 1. Great! But what if the source is a flat file? you can use an aggregate and get the count of it. Then you will filter and wanted to make sure it reached the T1 and T2 tgt's appropriately.This would be the easiest way.Use a sorter transformation. Sort on key fields by which u want to find the duplicates. then use an expression transformation.Example:Example:field1-->field2-->SORTER:field1 --ascending/descendingfield2 --ascending/descendingExpression:--> field1--> field2<--> v_field1_curr = field1<--> v_field2_curr = field2

Page 37: Interview Questions for DWH Full Allignment

v_dup_flag = IIF(v_field1_curr = v_field1_prev, true, false)o_dup_flag = IIF(v_dup_flag = true, 'Duplicate', 'Not Duplicate'<--> v_field1_prev = v_field1_curr<--> v_field2_prev = v_field2_curruse a Router transformation and put o_dup_flag = 'Duplicate' in T2 and 'Not Duplicate' in T1.Informatica evaluates row by row. So as we sort, all the rows come in order and it will evaluate based on the previous and current rows.

206.How do you create a mapping using multiple lookup transformation?

207Can any one explain real time complain mappings or complex transformations in Informatica.Specially in Sales Domain?Most complex logic we use is demoralization. We dont have any Demoralizer transformation in INformatica. So we will have to use an aggregator followed by an expression. Apart from this, we use most of the complex city in expression transformation involving lot of nested IIF's and Decode statements...another one is the union transformation and joiner.

208. About Informatica Power center 7:1) I want to Know which mapping properties can be overridden on a Session Task level.2) Know what types of permissions are needed to run and schedule Work flows.You can override any properties other than the source and targets. Make sure the source and targets exist in ur db if it is a relational db. If it is a flat file, you can override its properties. You can override sql if its a relational db, session log, DTM buffer size, cache sizes etc.

209.Can anyone explain about incremental aggregation with an example?When you use aggregator transformation to aggregate it creates index and data caches to store the data 1.Of group by columns 2. Of aggregate columns the incremental aggregation is used when we have historical data in place which will be used in aggregation incremental aggregation uses the cache which contains the historical data and for each group by column value already present in cache it add the data value to its corresponding data cache value and outputs the row, in case of a incoming value having no match in index cache the new values for group by and output ports are inserted into the cache.

210.Informatica Live Interview Questions?Explain grouped cross tab?Explain reference cursorWhat are parallel query's and query hintsWhat is meta data and system catalogWhat is factless fact schemaWhat is confirmed dimensionWhich kind of index is preferred in DWHWhy do we use DSS database for OLAP tools

211.Can any body write a session parameter file which will change the source and targets for every session. i.e. different source and targets for each session run.You are supposed to define a parameter file. And then in the Parameter file, you can define two parameters, one for source and one for target.Give like this for example:$Src_file = c:\program files\informatica\server\bin\abc_source.txt$tgt_file = c:\targets\abc_targets.txtThen go and define the parameter file:[folder_name.WF:workflow_name.ST:s_session_name]$Src_file =c:\program files\informatica\server\bin\abc_source.txt$tgt_file = c:\targets\abc_targets.txtIf its a relational db, you can even give an overridden sql at the session level...as a parameter. Make sure the sql is in a single line.

212.How do we remove the staging area?

213.What is polling?It displays update information about the session in the monitor window.

Page 38: Interview Questions for DWH Full Allignment

214.What is Transaction? A transaction can be defined as DML operation means it can be insertion, modification or deletion of data performed by users/ analysts/applicators.

215.What are the various test procedures used to check whether the data is loaded in the backend, performance of the mapping, and quality of the data loaded in INFORMATICA?

216.1)What are the various test procedures used to check whether the data is loaded in the backend, performance of the mapping, and quality of the data loaded in INFORMATICA.2) What are the common problems developers face while ETL development?

217.What happens if you try to create a shortcut to a non-shared folder?

It only creates a copy of it..

218.In a joiner transformation, you should specify the source with fewer rows as the master source. Why?

In joinner transformation informatica server reads all the records from master source builds index and data caches based on master table rows. after building the caches the joiner transformation reads records from the detail source and perform joins.

219.If you want to create indexes after the load process which transformation you choose?

a) Filter Tranformationb) Aggregator Tranformationc) Stored procedure Tranformationd) Expression Tranformation

Stored procedure transformation.

220.Where is the cache stored in informatica?Cache stored in informatica is in informatica server.

221.How to get two targets T1 containing distinct values and T2 containing duplicate values from one source S1?

222.What will happen if you are using Update Strategy Transformation and your session is configured for "insert"?

What are the types of External Loader available with Informatica?

If you have rank index for top 10. However if you pass only 5 records, what will be the output of such a Rank Transformation?

223.What are the real time problems generally come up while doing/running mapping/any transformation?can any body explain with example.

224.Can batches be copied/stopped from server manager?

225.What is rank transformation? where can we use this transformation?Rank transformation is used to find the status. Ex if we have one sales table and in this if we find more employees selling the same product and we are in need to find the first 5 0r 10 employee who is selling more products. we can go for rank transformation.

226.What is exact use of 'Online' and 'Offline' server connect Options while defining Work flow in Work flow monitor? . The system hangs when 'Online' server connect option. The Informatica is installed on a Personal laptop.

227.How can you delete duplicate rows with out using Dynamic Lookup? Tell me any other ways using lookup delete the duplicate rows?

Page 39: Interview Questions for DWH Full Allignment

Business ObjectsBO Designer

1.What is Cardinality?Expresses the minimum and the maximum number of instances of an entity B that can be associated with an instance of an entity A. The minimum and the maximum number of instances can be equal to 0, 1, or N.

2.What is Cartesian product?A situation in which a query includes two or more tables that are not linked by a join. If executed, this type of query retrieves all possible combinations between each table and may lead to inaccurate results.

3.What is Class?A class is a logical grouping of objects within a universe. In general, the name of a class reflects a business concept that conveys the category or type of objects. For example, in a universe pertaining to human resources, one class might be Employees.A class can be further divided into subclasses. In the human resources universe, a subclass of the Employees class could be Personal Information.As designer, you are free to define hierarchies of classes and subclasses in a model that best reflects the business concepts of your organization.

4. What is Condition?A component that controls the type and the amount of data returned by a specific object in a query. A condition created in the Designer module is referred to as a predefined condition.

5.What is Connection?Set of parameter that provides access to an RDBMS. These parameters include system information such as the data account, user identification, and path to the database. Designer provides three types of connections: secured, shared, and personal.

6.What is Context?A method by which Designer can decide which path to choose when more than one path is possible from one table to another in the universe.

7.What is Detail object?An object qualified as a detail provides descriptive data about a dimension object. A detail object cannot be used in drill down analysis.

8.What is Dimension object?An object being tracked in multidimensional analysis; the subject of the analysis. Dimensions are organized into hierarchies.

9.What is Document domain?The area of the repository that stores documents, templates, scripts, and lists of values.

10.What is Drill?There can be 3 types of Drill Analysis- Drill Down,Drill Up and Drill Through.Within the same universe one can Drill Up/Down like Country-State-City;of course with facts that relate to the same grain.A Drill through is possible when we can link different Data Marts like Profitability as defined by details of Asset, Liability, Income and Expense.

11.What is Equi-join?A join based on the equality between the values in the column of one table and the values in the column of another. Because the same column is present in both tables, the join synchronizes the two tables.

12.What is Enterprise mode?A work mode whereby a designer creates universes in an environment with a repository. The mode in which a universe is saved determines whether other designers are able to access it. By default, a universe is saved in the mode in which the designer is already working.

Page 40: Interview Questions for DWH Full Allignment

13.What is Hierarchy?An ordered series of related dimensions used for multidimensional analysis.

14.What is Join?A relational operation that causes two tables with a common column to be combined into a single table. Designer supports equi-joins, theta joins, outer joins, and shortcut joins.

15.What is List of values?A list of values contains the data values associated with an object. These data values can originate from a corporate database, or a flat file such as a text file or Excel file. In Designer you create a list of values by running a query from the Query Panel. You can then view, edit, purge, refresh and even export this file. A list of values is stored as an .lov file in a subfolder of the UserDocs folder.

16.What is Loop?A situation that occurs when more than one path exists from one table to another in the universe.

17.What is Measure object?An object that is derived from an aggregate function. It conveys numeric information by which a dimension object can be measured.

18.What is Object?A component that maps to data or a derivation of data in the database. For the purposes of multidimensional analysis, an object can be qualified as a dimension, detail, or measure. Objects are grouped into classes.

19.What is Offline mode?The work mode in which the designer works with universes stored locally.

20.What is Online mode?The work mode appropriate for a networked environment in which the general supervisor has set up a repository.

21.What is Outer join?A join that links two tables, one of which has rows that do not match those in the common column of the other table.

22.What is Personal connection?A personal connection is used to access resources such as universes or documents. It can be used only by the user who created it. Information about a personal connection is stored in both the PDAC.LSI and PDAC.SSI files; its definition is static and cannot be modified.

23.What is Qualification?A property of an object that determines how it can be used in multidimensional analysis. An object can be qualified as one of three types: a dimension, detail or measure.

24.What is Query?In Designer a query is a technique for creating or modifying a list of values associated with an object. From the Query Panel, a designer builds a query from the classes, objects, and conditions of a universe. In the BusinessObjects User module, a query is a type of data provider. An end user builds a query from a universe, and then runs the query to generate a BusinessObjects report.

25.What is Quick Design?A wizard in the Designer module that provides guided instructions for creating a basic universe. It lets a designer name a universe, set up a connection to a database, select strategies, create classes and objects, as well as generate joins with cardinalities.26.What is Repository?A centralized set of relational data structures stored in a database. It enables BusinessObjects users to share resources in a controlled and secured environment. The repository is made up of three domains: the security domain, the universe domain, and the document domain.

Page 41: Interview Questions for DWH Full Allignment

27.What is Secured connection?A secured connection is used to access universes or documents that may be restricted or confidential. It can be shared by several authorized users. Stored in the repository, the definition of a secured connection is updated dynamically.

28.What is Shared connection?A shared connection is used to access common resources such as universes or documents. It can be used by several users. Information about a shared connection is stored in a SDAC.LSI or SDAC.SSI file; its definition is updated dynamically.

29.What is Shortcut join?A join that links two tables by bypassing one or more other tables in the universe.

30.What is Strategy?Scripts that automatically extract structural information about tables, columns, joins, or cardinalities from a database. Designer provides default strategies but a designer can also create strategies. These are referred to as external strategies.

31.What is Structure pane?The graphical component within the main Designer window that shows the schema of the universe. It reflects the underlying database structure of the universe.

32.What is Subclass?A component within a class that groups objects. A subclass can itself contain other subclasses or objects.

33.What is Table Browser?The graphical component within the main Designer window that lets you create the classes and objects of the universe from the tables and columns of a database.

34.What is Theta join?A join that links tables based on a relationship other than equality between two columns.

35.What is Universe?A universe is a business-oriented mapping of the data structure found in databases: tables, columns, joins, etc. It can represent any specific application, system, or group of users. For example, a universe can relate to a department in a company such as marketing or accounting.

36.What is Universe domain?The area of the repository that holds exported universes. The universe domain makes it possible to store, distribute, and administrate universes. There may be multiple universe domains in a repository.

37.What is Universe pane?The graphical component within the main Designer window that displays a hierarchical view of the classes, objects, and conditions in a universe. Two radio buttons on the lower edge of the pane filter the display of the components. One button, the Classes/Objects filter, provides a view of the classes and objects. The other button, Classes/Conditions, provides a view of the classes and conditions.

38.What is User object?An object created by the end user from the Business Objects User module. A user object can be inserted into a universe by the designer.Insert a user object1 Select the User Objects command from the Insert menu.The Insert User Objects dialog box is displayed.2 Click a user object file.A user object file has a .udo extension3 Click the Open button.

39.What is Workgroup mode?A work mode whereby a designer creates universes in an environment without a repository. The mode in which a universe is saved determines whether other designers are able to access it. By default, a universe is saved in the mode in which the designer is already working.

Page 42: Interview Questions for DWH Full Allignment

40.What is Designer?Designer is a Business Objects IS module used by universe designers to create and maintain universes. Universes are the semantic layer that isolates end users from the technical issues of the database structure. Universe designers can distribute universes to end users by moving them as files through the file system, or by exporting them to the repository.

41.How do you design a universe?The design method consists of two major phases. During the first phase, you create the underlying database structure of your universe. This structure includes the tables and columns of a database and the joins by which they are linked. You may need to resolve loops which occur in the joins using aliases or contexts. You can conclude this phase by testing the integrity of the overall structure.During the second phase, you can proceed to enhance the components of your universe. You can also prepare certain objects for multidimensional analysis. As with the first phase, you should test the integrity of your universe structure. You may also wish to perform tests on the universes you create from the BusinessObjects User module. Finally, you can distribute your universes to users by exporting them to the repository or via your file system. For a universe based on a simple relational schema, Designer provides Quick Design, a wizard for creating a basic yet complete universe. You can use the resulting universe immediately, or you can modify the objects and create complex new ones. In this way, you can gradually refine the quality and structure of your universe.

42.What are the precautionary measures you will take in the project?

43.What is the drill up , drill down, drill by , drill trough ?Drill up: UP-one levelDrill down: DOWN-one levelDrill by: selection of level HierarchyDrill trough: Hierarchy to another Hierarchy

44.Explain the SQL Queries activated to data base from Data provider in BO ?BO automatically generates the SQL query when objects are selected into query panel. When you run the query, the query is processed to database based on your connectivity. For ex: If you run query using, Local Machine Full Client (BO reporter), the query directly connects to database through middlewareFull Client <---> DatabaseIf you run query using Web, the web browser will connects to Web Server and Web server will process request to Database.WEBI <--->Web Server <---> Database

45.What are steps to be taken care to create a good Universe?1)make the joins with optimization  2)reduce the creation user objects more in the universe 3)class should not be more than 60  4)try to use aggregate ware on measure objects

46.What are the steps to taken care to improve the Report performance?

In DESIGNER Level  1)eliminate the unnecessary joins  2)use conditions as much as at the database level  3)edit the SQL query in the Query Panel as per requirement  In REPORTER level  1)eliminate the filters as much as possible  2)try to reduce the user variables .

47.How can we achieve Correlated sub-query in Designer?can anyone help me in this regard.Right click on any object, go to the properties. Specify the query in select and put the next query in where clause, like select COLNAME from TABNAME1 where COLNAME IN (select colname2 from tab2)

Page 43: Interview Questions for DWH Full Allignment

48.What is broadcast agent?Broadcast agent is used for scheduling or bursting the reports. thru this, you can schedule your reports so that end-users can access reports thru internet. you have broadcast administration console here thru which, you can schedule reports and watch the tasks even. It is an administrator tool of Business Object.

49.How to connect one to one report?You can do this in different methods:In ASP environment..Universe level: create an object with Object format: HTML option enabled,SELECT :< A  href="Server:PortNo/wiasp/scripts/opendocument.asp?....etc..." >Report Level: use Hyperlink functionHyperlink("Server:PortNo/wiasp/scripts......etc..........." , "TaxtAsLink")For JSP environment, replace ASP with JSP.Any of the above methods will create a Hyperlink in the report pointing to another report. The arguments/prompts for this new report can be passes from the original report.

50.What are the diffrences between Business Objects 5.1 & 6.5 ?Main difference between BO 5.1 and 6.5 is Web Intelligence (info view).Using BO 5.1 info view. We can only view the reports that are created using USER module. We can't modify those reports. Using BO 6.5 info view. We can view and create the reports. We need not have BO installed on out systems to do this.

51.How do you connect the facts and dimensions in the tables?1. Smart Matching columns2. Manually you can link

52.I want to restrict top 26 rows ...in report is there any rownum object from which i can do this?Restrict Top 26 Rows: Depends on how you want to do it.In a report: Drag the Dimension and measures into report. Click on the Dimension object and use the RANK button on the tool bar. You can specify the Top 26 or Bottom (Depending on the requirement) values of a measure. It also provides some calculation like Sum/Percentage.If you want to restrict the Top 26 in the query itself, you might have to create an object in the universe. May have to use the analytical functions based on which you need to Rank the Data. For ex: use the Designer function : RANK() OVER (PARTITION BY () ORDER BY ())

53.What is the local filter and global filter in bo's?Local Filter >>Specific to a block (Table/cross tab/ chart etc.)Global Filter >> Specific to a TAB in a reportIf a report has multiple Tabs, each can have its own Local and Global filters.

54.What is pragma?

A PRAGMA statement appears in a PL/SQL block. It is not an executable statement, but it provides instructions to the compiler.

55.Could any one explain me how to create cascading prompts?

If you have source system > source location field. double click source system > edit properties menu will display, click properties tab and click edit you will get query pane there you pull down source location object in conditon pane and click inlist and click type new prompt type there "XXXXX" and save that.Now you click display button first source system lov will display and you select any one, based on that selection the source location will display. This is cascading prompt.

56.Which command using Query Analyzer will give you the version of SQL server and operating system?

57.Product date salesp1 20/Jan/1999 100p1 20/Feb/1999 200p1 20/apl/1999 100p1 20/DEC/1999 50p1 20/Jan/2000 100

Page 44: Interview Questions for DWH Full Allignment

p1 20/mar/2000 500now i want query like product ,sum of actual year ,sum of business yearp1 450 750 here actual year means 1 Jan 1999 to 31 st Jan 1999business year means 1 st APR 1999 to 31 st mar 200

58.What is slicing and dicing in business objects?Slice & Dice is facility in BO. We can enables change the positions of data in Report. Here in Bo we slice & dice panel by using this we can create cross tables and master details tables.

59.How to link two different universes ?For link 2 universes we hae 2 approaches..1. Go through Edit-->links2. Go through parameters of Universes there is one tab like Links, by using those we can link the 2 universes...

60.What's the Functional & Architectural Differences between Business Objects & Web Intelligence Reports?

61.How to Implement the the Built-in Strategy Script in BO Designer?

62.What's is the Guidelines to Build Universe with Better Performance? R Performance tuning Issues of Universes?

Page 45: Interview Questions for DWH Full Allignment

Business Objects1.What is a bo repository?Generally Repository is the Metadata BO 5.0 Repository create/maintain 50 tables, they are distributed as follows 25 tables for Security domain 24 Table for Universe domain 1 Table for Document Domain

2.Give the notes and functionalities of cascading prompts,@script in business objects?Syntax@Prompt ('message', ['type'], [lov], [MONO|MULTI], [FREE|CONSTRAINED])where message is the text of a message within single quotes. type can be one of the following: 'A' for alphanumeric, 'N' for number, or 'D' for date. can be either a list of values enclosed in brackets (each value must be within single quotes and separated by commas) or the name of a class and object separated by a backslash and within single quotes. MONO means that the prompt accepts only one value. MULTI means that the prompt can accept several values. FREE refers to free input as opposed to CONSTRAINED, which means that the end user must choose a value suggested by the prompt.DescriptionIs used to create an interactive object. In the Query Panel, this type of object causes a message to appear. This message prompts the end user to enter a specific value.NoteThe last four arguments are optional; however, if you omit an argument you must still enter the commas as separators.ExampleIn Where Clause:City.city IN @Prompt ('Choose City', 'A', {'Chicago', 'Boston', 'New York'}, MULTI, FREE)In the Query Panel, the object prompts the end user to choose a city.

3.When to use local filter and when to use global Filter?Local Filter is to single report to which it is create, but a global filter is to all the reports which consists of that column4.What are the user requirements in a universe?Database connections, key column, joins and check for loop if you need measures, metrics.

5. I have three predefined prompts. In a report it will come randomly. How they will come in a specified format?The Prompts will appear in the alphabetical order.To make them appear in the order of our requirement, need to prefix a numerical with the prompt.

5.What’s universal join in BOs?The level of join between two universes with a matching column.

6.Can we apply Rank and Sort at a time on a single report?No we can't apply rank and sort at a time on one object in one single report.If we try to apply , BO asks if you want to over write the previous condition.

7.What is difference between custom hierarchy and report based hierarchy?By default one class having one hierarchy ie called report hierarchy. Custom hierarchy we can create in designer according our req.

8.What is the multi value error ?Is there any types of Error in BO?You will get the Multi Value Error when you are trying to retrivr mutiple values into the cell.Ex: When u r tying to Insert cell into report and trying to assign a column to it which will have multiple values in it. (In A single cell you can’t show multiple values)

9.How many ways we test the universe & Report?By doing integrity check we can test universe & By coping report query and run in backend(oracle,sql server...) we can test the data by comparing both.

Page 46: Interview Questions for DWH Full Allignment

10.How achieve the many to many relation ship in Business ObjectsIf you have two tables with many to much relation ship, take the keys of those tables and form a 1-1 relationship table. Then you can connect both the existing tables to the newly formed table using the key.

11.How to filter repetitive values in the report to get distinct values.Example: In any database we can use distinct command But My Report source is flatfile.Therfore,I cann\'t manipulate source system.Is there any command or function to filter repeating records/values in the report.Using the eliminate duplicates option is one way. In a clightly different scenario, duplicates could come up due to the design structure of the report. e.g.In some specific cases, one may want to display a value in one field (say in a folded report) but multiple values come up (which could give you a COMPUTATION ERROR). To avoid such situation, one can define a variable which would e the max of the original field required. That ways the same value is fetched and also only one value is returned.

12.What is the difference between tabular report& cross tab report?Tabular reports are a 2D format and Cross tab reports is a 3D format. Most of the operational/formatted reports will be created in tabular form and most of the analytical reports are created in the cross tab report form (Product VS Year VS Sales amount)

13.Can you add the Cascading prompts in Business Objects? (Based on Country prompt selection you need to provide the option to select Sate/Province in the next State prompt. That means dynamically refresh the data load in the state list values)

14. My query takes 30 min to refresh ...there is any performance tunning is there to reduce refresh time?Make sure to have the right context - apply right indexes on the tables.

15.What is the difference between condition & filter? Conitions: these retrievals the data based on the condition (universe level) Filter: it brings the data and filter; performance will be slow (report level)

16.How can we do load testing for Webi Reports? (Example if I have have to test one web report for Concurrent users refreshing report at one time)

17.What is the diferrence between slice and dice & drill through?Slice and dice is you are going to analyze the data in different angle, but drill through is like using a single value you are going to grain the data.

18.What is the difference between slice and dice and cross tab report?Cross tab report is like at the intersection of 2 dimensions measure is displayed.like if we have two dimensions year and product and a measure revenu,revenue is displayed at the intersection of year and product.

19.How we drill up week to Month?In data Warehouse when u changes + symbol to - , u would c all the aggregation level in that + this is called rolling down. So if u again click over that - it would again compressed back to +. This is call drill up.Your Aggregation levels would b likeYear>Month>Week>Day>Hour>Minutes>Seconds.So by changing the symbol from - to + at month aggregation level u can drill up.

20.How to break a Date field into Year, month, week, day?(For example: I have a field INVOICE_DATE. Now i want to break it as Year, month..... I.e. in time dimension.) DO i need to have some changes in Universe?Using date functions separate the date, day, week, year, quarter, decade, holiday and more, and then you can get your answers.

21.What is the dense rank?