a final documentation

Upload: anish-gupta

Post on 07-Apr-2018

226 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/6/2019 a Final Documentation

    1/167

    1

    1.Informatica Product Overview1.1 Introduction1.2 Sources and Targets1.3 PowerCenter 8 Domain1.4 PowerCenter 8 Repository1.5 PowerCenter 8 Administration Console1.6 PowerCenter 8 Client1.7 Repository Manager1.8 Repository Objects1.9 Workflow Manager1.10 Workflow Monitor1.11 Repository Services1.12 Integration Services1.13 Web Services Hub1.14 Data Analyzer1.15 Metadata Manager1.16 PowerCenter Repository Reports

    2.Installation Steps

  • 8/6/2019 a Final Documentation

    2/167

    2

    3.Repository Manager3.1What is Repository?3.2Repository Connectivity3.3Repository Server3.4Repository Objects3.5Repository Metadata3.6Using Repository

    4.PowerCenter Designer4.1Designer Overview4.2About Transformation4.3Lookup Transformation4.4Expression Transformation4.5Router Transformation4.6Filter Transformation4.7Joiner Transformation4.8Sequence Generator Transformation4.9Source Qualifier Transformation4.10 Aggregator Transformation4.11

    Update Strategy

    4.12 Stored procedure Transformation4.13 Rank Transformation4.14 Java Transformation4.15 User Defined Functions4.16 Data Profiling

  • 8/6/2019 a Final Documentation

    3/167

    3

    4.17 Profile Manager4.18

    Debugger Overview

    5.PowerCenter Workflow Manager5.1Workflow Manager5.2Workflow Manager Tools5.3Workflow Structure5.4Workflow Tasks5.5Task Developer5.6Session Task5.7Event-Task5.8E-Mail Task5.9Worklet5.10 Workflow Scheduler5.11 Server Connections5.12 Relational Connections (Native)5.13 FTP Connection5.14 Workflows Design5.15 Workflow Monitor

    6.Transformations Overview

  • 8/6/2019 a Final Documentation

    4/167

    4

    Informatica Product Overview

    1.1 IntroductionInformatica Power Center is a single, unified enterprise data integration

    platform that allows companies and organizations of all sizes to access,

    discover, and integrate data from virtually any business system, in any

    format, and deliver that data throughout the enterprise at any speed.

    Power Center helps organizations derive business value from all their data

    so that they can reduce IT costs and complexity, streamline operations, and

    drive revenue growth.

    PowerCenter helps organizations derive business value from all their data

    so that they can

    Reduce IT costs and complexity Streamline business operations and processes Drive revenue growth

  • 8/6/2019 a Final Documentation

    5/167

    5

    1.2 Informatica Powercenter 8 can access the following data sources and loadthe data the data into the following targets

    Sources

    Targets

  • 8/6/2019 a Final Documentation

    6/167

    6

    Informatica PowerCenter 8 provides an environment which allows us to load data

    into a centralized location, such as a data warehouse or operational data store

    (ODS). We can extract data from multiple sources, transform the data according

    to business logic, and load the transformed data into file and relational targets.

    PowerCenter 8 also provides us the ability to view and analyze business

    information and browse and analyze metadata from disparate metadata

    repositories.

    PowerCenter 8 Components

    PowerCenter domain PowerCenter repository Administration Console PowerCenter Client Repository Service

  • 8/6/2019 a Final Documentation

    7/167

    7

    Integration Service Web Services Hub Data Analyzer Metadata Manager PowerCenter Repository Reports

  • 8/6/2019 a Final Documentation

    8/167

    8

    1.3 PowerCenter 8 DomainPowerCenter has a service-oriented architecture that provides the ability to scale

    services and share resources across multiple machines. It provides the

    PowerCenter domain to support the administration of the PowerCenter Services.

    Domain is the primary unit for management and administration of services in

    PowerCenter. It has the following components

    One or more nodesNode is a logical representation of a machine in a domain. Domain

    may contain more than one node. The node that hosts the domain is

    the master gateway for the domain. We can add the other machines

    as nodes in the domain and configure the nodes to run Integration

    Services and Repository Services. All service requests from other

    nodes in the domain go through the master gateway.

    Service ManagerService Manager is built in to the domain to support the domain and

    the application services. The Service Manager runs on each node in

    the domain. Service Manager performs the following functions

    Authentication Authorization Configuration Node configuration

  • 8/6/2019 a Final Documentation

    9/167

    9

    Licensing Logging

    Application servicesApplication services are a group of services that represent PowerCenter server-

    based functionality. The application services that run on each node in the domain

    depend on the way you configure the node and the application service. The

    following services are installed once you install PowerCenter services.

    Repository Service Integration Service Web Services Hub SAP BW Service

    1.4 PowerCenter 8 RepositoryPowerCenter repository resides in a relational database. Repository database

    tables contain the instructions required to extract, transform, and load data.

    PowerCenter Client applications access the repository database tables through

    the Repository Service. The repository consists of database tables that store

    metadata. Metadata describes different types of objects, such as mappings or

    transformations, that we can create or modify using the Client tools.

    Integration Service uses repository objects to extract, transform, and load data.

    The repository also stores administrative information such as user names,

    passwords, permissions, and privileges. We add metadata to the repository tables

    when you perform tasks in the PowerCenter Client application, such as creating

  • 8/6/2019 a Final Documentation

    10/167

    10

    users, analyzing sources, developing mappings or mapplets, or creating

    workflows. The Integration Service reads metadata created through the Clientapplication when you run a workflow. It also creates metadata, such as start and

    finish times of a session or session status.

    We can administer the repository using the Repository Manager Client tool and

    we can also develop global and local repositories to share metadata.

    Global repositoryGlobal repository is the hub of the repository domain. Use the global

    repository to store common objects that multiple developers can use through

    shortcuts. These objects may include operational or Application source

    definitions, reusable transformations, mapplets, and mappings.

    Local repositoriesLocal repository is any repository within the domain that is not the global

    repository. Use local repositories for development. From a local repository, we

    can create shortcuts to objects in shared folders in the global repository. Theseobjects include source definitions, common dimensions and lookups, and

    enterprise standard transformations. We can also create copies of objects in

    non-shared folders.

    1.5 PowerCenter 8 Administration ConsoleAdministration Console is a web application that we use to manage a

    PowerCenter domain. If you have a user login to the domain, we can access

    the Administration Consol to perform administrative tasks such as managing

    logs, user accounts, and domain objects. Domain objects include services,

    nodes, and licenses. Administration Console performs the following tasks in

    the domain.

  • 8/6/2019 a Final Documentation

    11/167

    11

    Manage application services Configure nodes Manage domain objects View and edit domain object properties View log events

    1.6 PowerCenter 8 ClientPowerCenter Client consists of the following applications that you use to manage

    the repository, design mappings, mapplets, and create sessions to load the data.

    Designer

  • 8/6/2019 a Final Documentation

    12/167

    12

    Designer is used to create mappings that contain transformation instructions for

    the Integration Service. The Designer has the following tools that we use toanalyze sources, design target schemas, and build source-to-target mappings

    Source AnalyzerIt imports or creates source definitions.

    Target DesignerIt imports or creates target definitions.

    Transformation DeveloperDevelop transformations to use in mappings. We can also

    develop user-defined functions to use in expressions.

    Mapplets DesignerIt Creates sets of transformations to use in mappings.

    Mapping Designer

    It Creates mappings that the Integration Service uses to extract,

    transform, and load data.

    The following things are displayed in Designer

    NavigatorIt connect to repositories, and open folders within the Navigator. We

    can also copy objects and create shortcuts within the Navigator.

    Workspace

  • 8/6/2019 a Final Documentation

    13/167

    13

    It opens different tools in this window to create and edit repository

    objects, such as sources, targets, mapplets, transformations, andmappings.

    OutputView details about tasks you perform, such as saving your work or

    validating a mapping.

    Status barIt Displays the status of the operation you perform.

    OverviewAn optional window to simplify viewing a workspace that

    contains a large mapping or multiple objects. Outlines the visible

    area in the workspace and highlights selected objects in color.

    Instance dataView transformation data while you run the Debugger to debug a

    mapping.

    Target dataView target data while you run the Debugger to debug a

    mapping.

  • 8/6/2019 a Final Documentation

    14/167

    14

    1.7 Repository ManagerRepository Manager is to create repository users and groups, assign privileges and

    permissions, and manage folders and locks. We can navigate through multiple

    folders and repositories, and complete the following tasks.

    Manage users and groupsCreate, edit, and delete repository users and user groups. We can

    assign and revoke repository privileges and folder permissions.

    Perform folder functions

  • 8/6/2019 a Final Documentation

    15/167

    15

    Create, edit, copy, and delete folders. Work you perform in the

    Designer and Workflow Manager is stored in folders. If you want toshare metadata, we can configure a folder to be shared.

    View metadataAnalyze sources, targets, mappings, and shortcut dependencies,

    search by keyword, and view the properties of repository objects.

    The following are displayed in Repository manager

    NavigatorIt displays all objects that you create in the Repository Manager, the

    Designer, and the Workflow Manager. It is organized first by

    repository, then by folder and folder version.

    MainIt provides properties of the object selected in the Navigator window.

    The columns in this window change depending on the object selected

    in the Navigator window.

    OutputIt provides the output of tasks executed within the Repository

    Manager, such as creating a repository.

  • 8/6/2019 a Final Documentation

    16/167

    16

    1.8 Repository ObjectsWe can create repository objects using the Designer and Workflow Manager

    Client tools. And we can view the following objects in the Navigator window of

    the Repository Manager.

    Source definitionsDefinitions of database objects (tables, views, synonyms) or files that

    provide source data.

  • 8/6/2019 a Final Documentation

    17/167

    17

    Target definitionsDefinitions of database objects or files that contain the target data.

    MappingsA set of source and target definitions along with transformations

    containing business logic that you build into the transformation.

    These are the instructions that the Integration Service uses to

    transform and move data.

    Reusable transformationsTransformations that you use in multiple mappings.

    MappletsA set of transformations that you use in multiple mappings.

    Sessions and workflowsSessions and workflows store information about how and when theIntegration Service moves data. A workflow is a set of instructions

    that describes how and when to run tasks related to extracting,

    transforming, and loading data. A session is a type of task that you

    can put in a workflow. Each session corresponds to a single mapping

    1.9 Workflow ManagerIn the Workflow Manager, we can define a set of instructions to

    execute tasks, such as sessions, emails, and shell commands. This setof instructions is called a workflow. It has the following tools to help

    you develop a workflow.

    Task DeveloperIt creates tasks that we want to accomplish in the workflow.

  • 8/6/2019 a Final Documentation

    18/167

    18

    Worklet DesignerIt creates a worklet in the Worklet Designer. A worklet is an object

    that groups a set of tasks. Worklet is similar to a workflow, but

    without scheduling information. We can nest worklets inside a

    workflow.

    Workflow DesignerIt creates a workflow by connecting tasks with links in the Workflow

    Designer. You can also create tasks in the Workflow Designer as we

    develop the workflow

  • 8/6/2019 a Final Documentation

    19/167

    19

    1.10Workflow MonitorWe can monitor workflows and tasks in the Workflow Monitor. View details about

    a workflow or task in Gantt chart view or Task view. We can run, stop, abort, and

    resume workflows from the Workflow Monitor. We can view sessions and

    workflow log events in the Workflow Monitor Log Viewer.

    The Workflow Monitor displays workflows that have run at least once. The

    Workflow Monitor continuously receives information from the Integration Service

    and Repository Service. It also fetches information from the repository to display

    historic information.

    The Workflow Monitor consists of the following windows.

    Navigator windowIt displays monitored repositories, servers, and repositories objects.

    Output window.It displays messages from the Integration Service and RepositoryService.

    Time windowIt displays progress of workflow runs.

    Task viewIt displays details about workflow runs in a report format.

    Gantt Chart viewIt displays details about workflow runs in chronological format.

  • 8/6/2019 a Final Documentation

    20/167

    20

    1.11Repository ServiceRepository Service manages connections to the PowerCenter repository from

    client applications. The Repository Service is a separate, multi-threaded processthat retrieves, inserts, and updates metadata in the repository database tables.

    Repository Service ensures the consistency of metadata in the repository.

    It accepts connection requests from the following PowerCenter applications:

    PowerCenter ClientUse the Designer and Workflow Manager to create and store

    mapping metadata and connection object information in the

    repository. Use the Workflow Monitor to retrieve workflow run

    status information and session logs written by the Integration

    Service. Use the Repository Manager to organize and secure

    metadata by creating folders, users, and groups.

  • 8/6/2019 a Final Documentation

    21/167

    21

    Command line programsUse command line programs to perform repository metadata

    administration tasks and service-related functions.

    Integration ServiceWhen you start the Integration Service, it connects to the repository

    to schedule workflows. When you run a workflow, the Integration

    Service retrieves workflow task and mapping metadata from therepository. The Integration Service writes workflow status to the

    repository.

    Web Services Hub.When you start the Web Services Hub, it connects to the repository

    to access web-enabled workflows. The Web Services Hub retrieves

    workflow task and mapping metadata from the repository and writes

    workflow status to the repository.

    1.12Integration ServicesThe Integration Service reads mapping and session information from the

    repository. It extracts data from the mapping sources and stores the data in

    memory while it applies the transformation rules that you configure in the

    mapping. Integration Service loads the transformed data into the mappingtargets. The Integration Service can combine data from different platforms and

    source types. It can also load data to different platforms and target types. The

    Integration Service connects to the repository through the Repository Service

    to fetch metadata from the repository.

  • 8/6/2019 a Final Documentation

    22/167

    22

    1.13Web Services HubThe Web Services Hub is a web service gateway for external clients. It

    processes SOAP requests from web service clients that want to access

    PowerCenter functionality through web services. Web service clients access

    the Integration Service and Repository Service through the Web Services Hub.Web Services Hub hosts the following web services.

    Batch web servicesRun and monitor web-enabled workflows.

    Real-time web servicesIt creates service workflows that allow you to read and write

    messages to a web service client through the Web Services Hub.

    1.14Data AnalyzerPowerCenter Data Analyzer provides a framework to perform business

    analytics on corporate data. With Data Analyzer, we can extract, filter, format,

    and analyze corporate information from data stored in a data warehouse,

    operational data store, or other data storage models. Data Analyzer uses a

    web browser interface to view and analyze business information at any level.

    It extracts, filters, and presents information in easy-to-understand reports. We

    can use Data Analyzer to design, develop, and deploy reports and set up

    dashboards and alerts to provide the latest information to users at the time

    and in the manner most useful to them. It works with a database repository to

    keep track of information about enterprise metrics, reports, and report

  • 8/6/2019 a Final Documentation

    23/167

    23

    delivery. Once we install Data Analyzer, users can connect to it from any

    computer that has a web browser and access to the Data Analyzer host.

    Data Analyzer can access information from databases, web services, or XML

    documents. You can set up reports to analyze information from multiple data

    sources. You can also set up reports to analyze real-time data from message

    streams.

    Data Analyzer Components

    With Data Analyzer, we can read data from a data source, create reports, and

    view the results on a web browser. It contains the following components.

    Data Analyzer repository.The repository stores the metadata necessary for Data Analyzer to

    track the objects and processes it requires to handle user requests.

    The metadata includes information on schemas, user profiles,

    personalization, reports and report delivery, and other objects and

    processes. We can use the metadata in the repository to createreports based on schemas without accessing the data warehouse

    directly. Data Analyzer connects to the repository through Java

    Database Connectivity (JDBC) drivers. The Data Analyzer repository is

    separate from the PowerCenter repository

  • 8/6/2019 a Final Documentation

    24/167

    24

    Application serverData Analyzer uses a third-party Java application server to manage

    processes. The Java application server provides services such as database

    access and server load balancing to Data Analyzer. The Java application

    server also provides an environment that uses Java technology to manage

    application, network, and system resources.

    Web serverData Analyzer uses an HTTP server to fetch and transmit Data Analyzer

    pages to web browsers.

    Data sourceFor analytic and operational schemas, Data Analyzer reads data from a

    relational database. It connects to the database through JDBC drivers. Forhierarchical schemas, Data Analyzer reads data from an XML document.

    The XML document may reside on a web server or be generated by a web

    service operation. Data Analyzer connects to the XML document or web

    service through an HTTP connection

  • 8/6/2019 a Final Documentation

    25/167

    25

    1.15Metadata ManagerPowerCenter Metadata Manager is a metadata management tool that we

    can use to browse and analyze metadata from disparate metadata

    repositories. Metadata Manager helps you understand and manage how

    information and processes are derived, the fundamental relationships

    between them, and how they are used. It provides the following tools

    Metadata Manager Console.It set ups, configure, and run XConnects, which load the source

    repository metadata into the Metadata Manager Warehouse. We can

    also use the Metadata Manager Console to set up connections to

    source repositories and other Metadata Manager components.

    Metadata Manager Custom Metadata Configurator.Create XConnects to load metadata from source repositories for

    which Metadata Manager does not package XConnects.

    Metadata Manager InterfaceBrowse source repository metadata and run reports to analyze the

    metadata. Also, use it to configure metamodels, set up source

    repositories, configure the reporting schema, and set up access and

    privileges for users and groups.

    Metadata Manager Components

    Application serverHelps the Metadata Manager Server manage its processes efficiently.

  • 8/6/2019 a Final Documentation

    26/167

    26

    Metadata Manager ServerIt manages the source repository metadata stored in the Metadata

    Manager Warehouse.

    Metadata Manager WarehouseIt stores the Metadata Manager Metadata, such as the Metadata

    Manager reporting schema, user profiles, and reports. It also stores

    source repository metadata and metamodels.

    PowerCenter repository.It stores the workflows, which are XConnect components that extract

    source metadata and load it into the Metadata Manager Warehouse.

    Web serverIt fetches and transmits Metadata Manager Pages to web browsers.

    Each supported application server contains an integrated web server.

  • 8/6/2019 a Final Documentation

    27/167

    27

    1.16PowerCenter Repository ReportsUse PowerCenter Repository Reports to browse and analyze PowerCenter

    metadata. PowerCenter Repository Reports provide the following types of

    reports to help us administer our PowerCenter environment.

    Configuration ManagementWith Configuration Management reports, we can analyze

    deployment groups and PowerCenter repository object labels.

  • 8/6/2019 a Final Documentation

    28/167

    28

    OperationsWith Operations reports, we can analyze operational statistics for

    workflows, Worklets, and sessions.

    PowerCenter ObjectsWith PowerCenter Object reports, we can identify PowerCenter

    objects, their properties, and their interdependencies with other

    repository objects.

    SecurityWith the Security report, we can analyze users, groups, and their

    association within the repository.

    Informatica power center 8 is having the following features which makes it

    more powerful, easy to use and manage when compared to previous

    versions.

    Supports Service oriented architecture Access to structured, unstructured and semi structured data Support for grid computing High availability Pushdown optimization Dynamic portioning Metadata exchange enhancements Team based Development

  • 8/6/2019 a Final Documentation

    29/167

    29

    Global Web-based Admin console New transformations

    23 New functions User defined functions Custom transformation enhancements

    Flat file enhancements New Data Federation option Enterprise GRID

  • 8/6/2019 a Final Documentation

    30/167

    30

    2 Installation Steps1. Verify that your environment meets the minimum system requirements and

    complete the pre-installation tasks.

    2. Log on to the machine with the user account you want to use to installPowerCenter.

    3. Close all other applications.4. To begin the installation on Windows from a DVD, insert the DVD into the

    DVD drive. Run install.bat from the DVD root directory.

    -Or-

    To begin the installation on Windows from a hard disk, run install.bat from

    the root directory in the location where you copied the installer.

    -Or-

    To begin the installation on UNIX, use a shell command line to run install.sh

    from the DVD root directory or the root directory in the location where you

    downloaded the installer.

    5. On UNIX, select the option for GUI mode installation.6. Select the language to use during installation and click OK.

    The Welcome window introduces the PowerCenter installation.

    7. Click Next.On UNIX, the Configure Environment Variables window appears. Verify that

    you have configured the required environment variables. The PowerCenter

    installer gives you the option to stop the installation and modify the

    environment variables.

  • 8/6/2019 a Final Documentation

    31/167

    31

    8. Click Next.The Choose Installation Type window appears.

    Choose Install PowerCenter 8.6

    9. Click Next.The PowerCenter License Key window appears.

    10. Enter the location and file name of the PowerCenter license key or clickBrowse to locate the license key file.

    11.Click Next.The Installation Prerequisites window displays the platforms and databases

    you can use and the disk space requirements. Verify that all PowerCenter

    installation requirements are met before you continue the installation.

    12.Click Next.The Installation Directory window appears.

    13.Enter an absolute path for the installation directory.Click Browse to find a directory or use the default directory.

    On Windows, the default directory is C:\Informatica\PowerCenter8.6.

    On UNIX, the default directory is $HOME/Informatica/PowerCenter8.6.

    Note: On Windows, the installation directory path must be on the currentmachine. On UNIX, HOME is the user home directory. The name of the

    installation directory cannot contain spaces.

    14.Click Next.The HTTPS Configuration window appears.

  • 8/6/2019 a Final Documentation

    32/167

    32

    15.. To use HTTP connection between the Administration Console and theService Manager, clear Enable

    HTTPS. Skip to step 18.

    To set up a secure connection between the Administration Console and the

    Service Manager, select Enable

    HTTPS and continue to the next step.

    17. Select the type of keystore to use and enter the following information based

    on your selection:

    Option

    Description

    Use a Keystore Generated by the Installer

    Select this option to use a self-signed keystore file generated by the

    PowerCenter installer. Specify the port number to use.

    Use an Existing Keystore File

    Select this option to use a keystore file you specify. The keystore file

    can be self signed or signed by a certification authority. Specify the

    port number and the location and password of the keystore.

    HTTPS Port Number

    Port used by the node to communicate between the Administration

    Console and the Service Manager.

    Keystore Password

  • 8/6/2019 a Final Documentation

    33/167

    33

    A plain-text password for the keystore file. Disabled when you use a

    keystore generated by the installer.

    Keystore File Location

    Path and file name of the keystore file. You can use a self-signed

    certificate or a certificate signed by a certification authority. Disabled

    when you use a keystore generated by the installer.

    If you use a generated keystore, the installer creates the keystore file in thefollowing directory:

    \server\tomcat\conf

    18. Click Next.

    19. On the Pre-Installation Summary window, review the installation information,

    and click Install to continue.

    The installer copies the files to the installation directory. When the file copy

    process completes, the Create or Join Domain window appears.

    20. Choose to create a domain if you are installing PowerCenter for the first time

    or you are installing PowerCenter on a single machine. Continue to the next step.

    -or-

    Choose to join a domain if you have created a PowerCenter domain on another

    machine and you want to

    Add the current machine as a node in the domain.

    On Windows, skip to step 27. On UNIX, skip to step 31.

  • 8/6/2019 a Final Documentation

    34/167

    34

    For more information about the available domain options, click the Help Me

    Select link.

    21. Click Next.

    The Configure Domain Database window appears.

    PowerCenter stores the PowerCenter domain configuration in a relational

    database. The domain configuration must be accessible to all gateway nodes in

    the domain.

    22.Enter the following information:

    Property

    Description

    Database Type

    Database for the domain configuration.

    Select Oracle, Microsoft SQL Server, Sybase ASE, or IBM DB2.

    Database URL

    Host name and port number for the database instance in the format

    :.

    Database User ID

    Domain configuration database user account.

    Database User Password

    Password for the domain configuration database user account.

  • 8/6/2019 a Final Documentation

    35/167

    35

    Database Service Name

    Service name for Oracle and IBM DB2 databases or database name for Microsoft

    SQL Server or Sybase ASE databases. Use the following guidelines:

    - If you want to use an Oracle SID, you can use the Custom String option.

    - If you want to use another database other than the default Sybase ASE database

    for the user acount, use the Custom String option.

    Custom String

    JDBC connect string.

    - To use an Oracle SID instead of an Oracle service name, use the following JDBC

    connect string:

    jdbc:informatica:oracle://host_name:port;SID=sid

    - To specify a non-default Sybase database, use the following JDBC connect string:

    DatabaseName=

    Tablespace (optional)

    Name of the tablespace in which to create the repository tables. If blank, the

    installation creates the repository tables in the default tablespace. Define the

    repository database in a single-node tablespace to optimize performance.

    Enabled if you select IBM DB2.

    23. Click Test Connection to verify that you can connect to the domain

    configuration database.

  • 8/6/2019 a Final Documentation

    36/167

    36

    24. Click Next.

    The Configure Domain window appears.

    25. Enter the following information:

    Property

    Description

    Domain Name

    Name of the PowerCenter domain to create. The domain name must be in 7-bit

    ASCII format and less than 79 characters. The domain name cannot contain

    spaces or the following characters: \ / : * ? > < " |

    Note: If you are upgrading from PowerCenter 7.x, the name of the PowerCenter

    domain cannot be the same as the name of the PowerCenter Server in 7.x.

    Domain Host Name

    Host name of the machine on which to create the PowerCenter domain. If you

    create a domain on a machine with a single network name, use the default host

    name. If you create a domain on a machine with multiple network names, you can

    modify the default host name to use an alternate network name. Optionally, you

    can use the IP address of the machine on which to create the domain.

    Node Name

    Node name for the current machine. This is the name of the gateway node for the

    domain. This is not the host name for the machine.

  • 8/6/2019 a Final Documentation

    37/167

    37

    Domain Port No.

    Port number for the current machine. The installer displays a default port number

    of 6001. If the port number is not available on the machine, the installer displays

    the next available port number.

    Domain User Name

    User name of the domain administrator. Use this name to log in to the

    PowerCenter

    Administration Console. The user name must be less than 79 alphanumeric

    characters and cannot contain special characters. Do not use Administrator or

    administrator as the domain user name. Default is admin.

    Domain Password

    Password for the domain administrator. The password must be between 3 and 16

    characters.

    Confirm Password

    Enter the password again.

    To set the range of port numbers for PowerCenter on the node, click Advanced

    Configuration.

    26. Enter the range of port numbers that the PowerCenter installer can use for

    PowerCenter on the node and

    Click OK.

    The default range for the port numbers is 6005 - 6105.

  • 8/6/2019 a Final Documentation

    38/167

    38

    Skip to step 29.

    27. Click Next.

    The Configure Domain window appears.

    Verify that the gateway node for the domain you want to join is available before

    you continue.

    28. Enter the following information:

    Property

    Description

    Domain Name

    Name of the domain you want to join.

    Domain Host Name

    Host name or IP address of the gateway node for the domain.

    Domain Port No.

    Port number for the gateway node.

    Domain User Name

    User name for a domain administrator in the PowerCenter domain you want to

    join.

    Domain Password

    Password for the domain administrator user account.

  • 8/6/2019 a Final Documentation

    39/167

    39

    29. Click Next.

    On Windows, the Configure Informatica Services window appears.

    Informatica Services is the Windows service that runs PowerCenter. You can

    specify a different user

    Account to run the service.

    30. Enter the following information:

    Property

    Description

    Run Informatica Services

    With a different user account

    Indicates whether the current Windows user account that installs Informatica

    Services also run Informatica Services. If selected, enter the user name and

    password of the user account to run Informatica Services.

    Use a different account to run Informatica Services if PowerCenter needs to

    access a network location not available to the current Windows user account. You

    must also use a different account to run Informatica Services to use a trusted

    connection for authentication with the PowerCenter repository database.

    If not selected, the current user account that installs Informatica Services also

    runs Informatica Services.

  • 8/6/2019 a Final Documentation

    40/167

    40

    User name

    User account used to run the Informatica Services service. Enter the Windows

    domain and user account in the format \. This

    user account must have the Act as operating system permission.

    Password

    Password for the Windows user account to run Informatica Services.

    31. Click Next.

    The PowerCenter Post-installation Summary window indicates whether the

    installation completed successfully. It also shows the configured properties for

    PowerCenter components and the status of installed components.

    32. Click Done.

    You can view the log files generated by the installer to get more information

    about the installation tasks performed by the installer and to view configuration

    properties for the installed components.

  • 8/6/2019 a Final Documentation

    41/167

    41

    3 Repository Manager3.1 What is Repository?

    The Informatica repository is a relational database managed by the Repository Server.

    The Repository stores information, or metadata, used by the Informatica Server and Client tools

    3.2Repository Connectivity. Through Repository Server Repository Client applications access the Repository database

    tables.

    Repository Server protects the metadata. Repository Server notifies you the objects you are working are modified or deleted by

    another user.

    Repository Server uses native drivers to communicate with Repository database. Informatica client tools and Server communicate with Repository Server over TCP/IP To manage Repository database Repository Server uses a process called Repository Agents. To manage multiple Repositories on different machines we use multiple Repository Agents.3.3Repository Server.Each Repository has independent architecture for the management of physical repository tables

    Components: One Repository Server and Repository Agent for each Repository

  • 8/6/2019 a Final Documentation

    42/167

    42

    The Repository Server starts the Repository Agent process for the repository database. The client application sends a repository connection request to the Repository Server. The Repository Server verifies connectivity information for the target repository.

    3.4Repository ObjectsFolders:

    Folders organize and store metadata in the Repository Folders share the metadata with other repository users. You must create a Folder in Repository before connect to Repository through Designer

    or Workflow manager.

    Folder Creation

  • 8/6/2019 a Final Documentation

    43/167

    43

  • 8/6/2019 a Final Documentation

    44/167

    44

    Folder Comparison

  • 8/6/2019 a Final Documentation

    45/167

    45

  • 8/6/2019 a Final Documentation

    46/167

    46

  • 8/6/2019 a Final Documentation

    47/167

    47

  • 8/6/2019 a Final Documentation

    48/167

    48

    Folder Sharing

  • 8/6/2019 a Final Documentation

    49/167

    49

    Shared Folder available to all other Folders inthe same Repository.

    Once you make a Folder shared you cannotreverse it.

    Shared Folder in Global Repository can beused any folder in the domain.

  • 8/6/2019 a Final Documentation

    50/167

    50

    Repository Manager has Four Types of Windows

    Navigator Window Contains:

    Repositories. Folders. Folder versions. Nodes. Repository objects.

  • 8/6/2019 a Final Documentation

    51/167

    51

    Main window displays details about the objects

  • 8/6/2019 a Final Documentation

    52/167

    52

    Dependency window appears when you configure the Repository Manager to

    display dependencies.

    Output Windows displays detailed information

  • 8/6/2019 a Final Documentation

    53/167

    53

    Folder Versions

    Folder Versions stores a copy of metadata indevelopment.

    Folder Versions can revert to previous version duringthe development process.

    Folder Versions can create at any time in the Designer.Users:

    Repository users have a username and password thatallow access to the repository.

    Repository user belongs to at least one user group.

    Create User

  • 8/6/2019 a Final Documentation

    54/167

    54

  • 8/6/2019 a Final Documentation

    55/167

    55

  • 8/6/2019 a Final Documentation

    56/167

    56

  • 8/6/2019 a Final Documentation

    57/167

    57

    User Groups:

    User groups organize individual repository users. Individual users inherit all privileges assigned to the user

    group.

    3.5Repository MetadataThe repository stores metadata that describes how to extract, transform,and load source and target data.

  • 8/6/2019 a Final Documentation

    58/167

    58

    Repository objects:

    Source definitions Target definitions Transformations Reusable transformations Mappings Mapplets Multi-Dimensional metadata

    Shortcuts Database connections Connection objects Sessions Workflows Workflow tasks Worklets Folders Folder versions Users User Groups

    Repository server Administration console contains two types of windows:

    Console tree Main window

    Console tree contains following nodes:

    Informatica Repository Server

  • 8/6/2019 a Final Documentation

    59/167

    59

    Repository Server name. Repositories. Repository name. Connections. Locks. Backups.

    Main window displays details of the node you select in the Console Tree.

  • 8/6/2019 a Final Documentation

    60/167

    60

    3.6Using RepositoryRegistering Repository

  • 8/6/2019 a Final Documentation

    61/167

    61

    Copying a Repository

    Provides a Metadata copy as a basis for New Repository. To Preserve Original Repository before Upgrading. When Moving a Repository from Development to

    Production.

    Repository Security

    Features to Implement Security:

    User groups.

    Repository users.

    Repository privileges

    Folder permissions.

    User connections.

    Locking.

    User Groups.

  • 8/6/2019 a Final Documentation

    62/167

    62

    o By default there are two default user groups

    Administrators Public

    o In Administrators group there are two default users. Administrator Database username used to create Repository.

    o To administer the user groups we must have one of thefollowing privileges.

    Administer Repository Super User

    o You can create custom groups and assign specificprivileges to that user groups.

    o Using the User groups you can manage users andRepository privileges efficiently.

    Stop the repository before Upgrading

  • 8/6/2019 a Final Documentation

    63/167

    63

    Change Password

  • 8/6/2019 a Final Documentation

    64/167

    64

    Managing privileges

    Repository Locks

    The repository uses locks to prevent users from duplicating or overriding work.

    Types of Locks

    In-use lock.

    Write-intent lock.

    Execute lock.

  • 8/6/2019 a Final Documentation

    65/167

    65

    In-Use Lock

    IN-Use Lock is created when:

    Viewing an object in a folder for which you do not havewrite permission.

    Viewing an object that is already write-locked. Exporting an object.

    We can create unlimited number of In-Use Locks per object.

    Write-Intent Lock

    Write-Intent Lock is created when:

    Viewing an object in a folder for which you have writepermission.

    Editing a repository object in a folder for which you havewrite permission.

    Importing an object.We can create Only One Write-Intent Lock per object.

    Execute Lock

    Execute Lock is created when:

    To start the Workflow that is already running.

    Drawback:

  • 8/6/2019 a Final Documentation

    66/167

    66

    Informatica Server load duplicate or inaccurate data.

    We can create Only One Execute Lock per object.

    4PowerCenter Designer4.1Designer Overview

    Designer is used to create mappings that contain transformation instructions for the Integration

    Service. The Designer has the following tools that we use to analyze sources, design target

    schemas, and build source-to-target mappings.

    Source AnalyzerIt imports or creates source definitions.

  • 8/6/2019 a Final Documentation

    67/167

    67

    Target DesignerIt imports or creates target definitions.

    Transformation DeveloperDevelop transformations to use in mappings. we can also develop user-defined

    functions to use in expressions.

    Mapplets DesignerIt creates sets of transformations to use in mappings.

    Mapping DesignerIt Creates mappings that the Integration Service uses to extract, transform, and load

    data.

    4.2About Transformation The transfer of data is called transformation. A transformation is a repository object that generates, modifies, or passes data. We configure logic in a transformation that the Integration Service uses to transform data. The Designer provides a set of transformations that perform specific functions. Transformations in a mapping represent the operations the Integration Service performs on

    the data.

    Data passes into and out of transformations through ports that we link in a mapping ormapplet.

    Transformations can be Active or Passive. An active transformation can change the number of rows that pass through it. A passive transformation does not change the number of rows that pass through it. Transformations can be connected to the data flow. An unconnected transformation is not connected to other transformations in the mapping. It is called within another transformation, and returns a value to that transformation.

    Tasks to incorporate a transformation into a mapping

    Create the transformation

  • 8/6/2019 a Final Documentation

    68/167

    68

    Configure the transformation Link the transformation to other transformations and target definitions Mapping Designer Transformation Developer Mapplet Designer

    Designer Transformations

    Aggregator - to do things like "group by". Expression - to use various expressions.

    Filter - to filter data with single condition. Joiner - to make joins between separate databases, file, ODBC sources. Lookup - to create local copy of the data. Normalizer - to transform denormalized data into normalized data. Rank - to select only top (or bottom) ranked data. Sequence Generator - to generate unique IDs for target tables. Source Qualifier - to filter sources (SQL, select distinct, join, etc.) Stored Procedure - to run stored procedures in the database - and capture their returned

    values.

    Update Strategy - to flag records in target for insert, delete, update (defined inside amapping).

    Router - same as filter but with multiple conditions

    Active Vs Passive Transformation

  • 8/6/2019 a Final Documentation

    69/167

    69

    4.3Lookup TransformationLookup Transformation Overview

    A Lookup transformation is a passive transformation. Use a Lookup transformation in a mapping to look up data in a flat file or a relational

    table, view, or synonym.

    We can import a lookup definition from any flat file or relational database to which boththe PowerCenter Client and Integration Service can connect.

    We can Use multiple Lookup transformations in a mapping. The Integration Service queries the lookup source based on the lookup ports in the

    transformation.

    It compares Lookup transformation port values to lookup source column values basedon the lookup condition.

  • 8/6/2019 a Final Documentation

    70/167

    70

    Tasks of Lookup Transformation

    Get a related value. Perform a calculation. Update slowly changing dimension tables. Connected or unconnected. Cached or uncached.

    Lookup Components

  • 8/6/2019 a Final Documentation

    71/167

    71

    We have to define the following components when we configure a Lookup transformation in a

    mapping.

    Lookup source Ports Properties Condition Metadata extensions

    Creating a Lookup Transformation

  • 8/6/2019 a Final Documentation

    72/167

    72

    In the Mapping Designer, click Transformation > Create. Select the Lookuptransformation. Enter a name for the transformation and Click OK. The naming

    convention for Lookup transformation is LKP_Transformation Name.

    In the Select Lookup Table dialog box, we can choose the following options. Choose an existing table or file definition. Choose to import a definition from a relational table or file. Skip to create a manual definition.

    If we want to manually define the lookup transformation, click the Skip button. Define input ports for each Lookup condition we want to define.

  • 8/6/2019 a Final Documentation

    73/167

    73

    For Lookup transformations that use a dynamic lookup cache, associate an input port orsequence ID with each lookup port.

    On the Properties tab, set the properties for the lookup. Click OK.Configuring Unconnected Lookup Transformations

    An unconnected Lookup transformation is separate from the pipeline in the mapping. Wewrite an expression using the: LKP reference qualifier to call the lookup within another

    transformation.

    Adding Input Ports. Adding the Lookup Condition.

    ITEM_ID = IN_ITEM_ID

    PRICE

  • 8/6/2019 a Final Documentation

    74/167

    74

    Setting the properties to port tab and properties tab

  • 8/6/2019 a Final Documentation

    75/167

    75

    Port Tab

    Lookup Transformation Tips

    Add an index to the columns used in a lookup condition Place conditions with an equality operator (=) first. Cache small lookup tables. Join tables in the database. Use a persistent lookup cache for static lookups. Call unconnected Lookup transformations with the: LKP reference qualifier.

    Properties Tab

    Lookup Caches

    The Integration Service builds a cache in memory when it processes the first row of data in acached Lookup transformation.

    It allocates memory for the cache based on the amount we configure in the transformationor session properties.

    The Integration Service stores condition values in the index cache and output values in thedata cache.

    The Integration Service queries the cache for each row that enters the transformation. The Integration Service also creates cache files by default in the $PMCacheDir.Types of lookup caches

    Persistent cache

  • 8/6/2019 a Final Documentation

    76/167

    76

    Recache from database Static cache Dynamic cache Shared cache

    4.4Expression Transformation

    Expression Transformation

    We can use the Expression transformation to calculate values in a single row before wewrite to the target

    We can use the Expression transformation to test conditional statements To perform calculations involving multiple rows, such as sums or averages we can use

    expression transformation

    We can use the Expression transformation to perform any non-aggregate calculations

    Setting Expression in Expression Transformation

    Enter the expression in the Expression Editor we have disable to in port. Check the expression syntax by clicking Validate.

  • 8/6/2019 a Final Documentation

    77/167

    77

    Connect to Next Transformation

    Connect the output ports to the next transformation or target

    Select a Tracing Level on the Properties Tab

    Select a tracing level on the Properties tab to determine the amount of transactiondetail reported in the session log file.

    Choose Repository-Save.

  • 8/6/2019 a Final Documentation

    78/167

    78

    4.5Router Transformation A Router transformation is an Active Transformation. A Router transformation is similar to a Filter transformation because both transformations

    allow us to use a condition to test data.

    A Filter transformation tests data for one condition and drops the rows of data that do notmeet the condition. However, a Router transformation tests data for one or more conditions

    and gives us the option to route rows of data that do not meet any of the conditions to a

    default output group.

    If we need to test the same input data based on multiple conditions, use a Routertransformation in a mapping instead of creating multiple Filter transformations to perform

    the same task.

    Creating a Router Transformation

    In the Mapping Designer, click Transformation > Create. Select the Routertransformation. Enter a name for the transformation and Click OK.

    The naming convention for router transformation is RTR_TransformationName.Input values in the Router Transformation

    Select and drag all the desired ports from a transformation to add them to the Routertransformation.

    Double-click the title bar of the Router transformation to edit transformationproperties.

  • 8/6/2019 a Final Documentation

    79/167

    79

    Setting the properties to port tab and properties tab

    Ports tab

    Properties tab

  • 8/6/2019 a Final Documentation

    80/167

    80

    Group tab in Router Transformation

    Click the Group Filter Condition field to open the Expression Editor. Enter a group filter condition. Click Validate to check the syntax of the conditions we entered. Click OK. Connect group output ports to transformations or targets. Choose Repository-Save.

    A Router transformation has the following types of groups.

    Input Output

    There are two types of output groups.

    User-defined groups

  • 8/6/2019 a Final Documentation

    81/167

    81

    Default group

    Router Transformation Components

    Working with Ports

    A Router transformation has input ports and output ports.

    Input ports reside in the input group, and output ports reside in the output groups. We can create input ports by copying them from another transformation or by manually

    creating them on the Ports tab.

    Port tab in Router Transformation

  • 8/6/2019 a Final Documentation

    82/167

    82

    Connecting Router Transformations in a Mapping

    When we connect transformations to a Router transformation in a mapping consider thefollowing rules.

    We can connect one group to one transformation or target.

    Connect one port to Multiple Target

    We can connect one output port in a group to multiple transformations or targets.

    Connect Multiple out ports to Multiple Target

    We can connect multiple output ports in one group to multiple transformations or targets.

  • 8/6/2019 a Final Documentation

    83/167

    83

    4.6Filter Transformation A Filter transformation is an Active Transformation. We can filter rows in a mapping with Filter transformation. We pass all the rows from a source transformation through the Filter transformation and

    then enter a filter condition for the transformation.

    All ports in a Filter transformation are input/output and only rows that meet the conditionpass through the Filter transformation.

    Creating a Filter Transformation

    In the Mapping Designer, click Transformation > Create. Select the Filter transformation.Enter a name, and click OK.

    The naming convention for Filter transformations is FIL_TransformationName. Select and drag all the ports from a source qualifier or other transformation to add them to

    the Filter transformation.

    After we select and drag ports, copies of these ports appear in the Filter transformation.Each column has both an input and an output port.

    Double-click the title bar of the Router transformation to edit transformation properties.

  • 8/6/2019 a Final Documentation

    84/167

    84

    A Filter transformation is an Active Transformation. We can filter rows in a mapping with Filter transformation. We pass all the rows from a source transformation through the Filter transformation and

    then enter a filter condition for the transformation.

    All ports in a Filter transformation are input/output and only rows that meet the conditionpass through the Filter transformation.

    Creating a Filter Transformation

    In the Mapping Designer, click Transformation > Create. Select the Filter transformation.Enter a name, and click OK.

    The naming convention for Filter transformations is FIL_TransformationName. Select and drag all the ports from a source qualifier or other transformation to add them to

    the Filter transformation.

  • 8/6/2019 a Final Documentation

    85/167

    85

    After we select and drag ports, copies of these ports appear in the Filter transformation.Each column has both an input and an output port.

    Double-click the title bar of the Router transformation to edit transformation properties.

    Click the Value section of the condition, and then click the Open button. The Expression Editor appears. Enter the filter condition we want to apply. Use values from one of the input ports in the transformation as part of this condition

    However, we can also use values from output ports in other transformations.

    We may have to fix syntax errors before continuing. Click OK. Select the Tracing Level, and click OK to return to the Mapping Designer. Choose Repository-Save.

  • 8/6/2019 a Final Documentation

    86/167

    86

    Filter Transformation Tips

    Use the Filter transformation early in the mapping. Use the Source Qualifier transformation to filter.

    4.7Joiner Transformation A Joiner transformation is an active transformation. Joiner transformation is used to join source data from two related heterogeneous sources

    residing in different locations or file systems.

    We can also join data from the same source. The Joiner transformation joins sources with at least one matching column. The Joiner transformation uses a condition that matches one or more pairs of columns

    between the two sources.

    We can use the following sources

    Two relational tables existing in separate databases. Two flat files in potentially different file systems. Two different ODBC sources. A relational table and an XML source. A relational table and a flat file source. Two instances of the same XML source.Creating a Joiner Transformation

    In the Mapping Designer, click Transformation > Create. Select the Joiner transformation. Entera name, and click OK.

    The naming convention for Joiner transformations is JNR_TransformationName. Drag all the input/output ports from the first source into the Joiner transformation.

  • 8/6/2019 a Final Documentation

    87/167

    87

    The Designer creates input/output ports for the source fields in the Joiner transformation asdetail fields by default. We can edit this property later.

    Select and drag all the input/output ports from the second source into the Joinertransformation.

    The Designer configures the second set of source fields and master fields by default.Edit Transformation

    Double-click the title bar of the Joiner transformation to open the Edit Transformations dialogbox.

    Select the port tab. Add default values for specific ports as necessary.

  • 8/6/2019 a Final Documentation

    88/167

    88

    Setting the Condition

    Select the Condition tab and set the condition. Click the Add button to add a condition. Click the Properties tab and configure properties for the transformation. Click OK.

    Defining the Join Type

    Join is a relational operator that combines data from multiple tables into a single result set. We define the join type on the Properties tab in the transformation. The Joiner transformation supports the following types of joins.

    Normal Master Outer Detail Outer Full Outer

    Joiner Transformation Tips

    Perform joins in a database when possible. Join sorted data when possible.

  • 8/6/2019 a Final Documentation

    89/167

    89

    For an unsorted Joiner transformation, designate the source with fewer rows as the mastersource.

    For a sorted Joiner transformation, designate the source with fewer duplicate key values as

    the master source.

    4.8Sequence Generator Transformation A Sequence Generator transformation is a passive transformation. The Sequence Generator transformation generates numeric values. We can use the Sequence Generator to create unique primary key values, cycle through a

    sequential range of numbers.

    The Sequence Generator transformation is a connected transformation. The Integration Service generates a value each time a row enters a connected

    transformation, even if that value is not used.

    When NEXTVAL is connected to the input port of another transformation, the IntegrationService generates a sequence of numbers.

    When CURRVAL is connected to the input port of another transformation, the IntegrationService generates the NEXTVAL value plus one.

    We can make a Sequence Generator reusable, and use it in multiple mappings. Web might reuse a Sequence Generator when we perform multiple loads to a single target. If we have a large input file we can separate into three sessions running in parallel, we can

    use a Sequence Generator to generate primary key values.

    If we use different Sequence Generators, the Integration Service might accidentallygenerate duplicate key values.

    Instead, we can use the reusable Sequence Generator for all three sessions to provide aunique value for each target row.

    Tasks with a Sequence Generator Transformation

    Create keys

  • 8/6/2019 a Final Documentation

    90/167

    90

    Replace missing values Cycle through a sequential range of numbers

    Creating a Sequence Generator Transformation

    In the Mapping Designer, select Transformation-Create. Select the Sequence Generatortransformation. The naming convention for Sequence Generator transformations is

    SEQ_TransformationName.

    Enter a name for the Sequence Generator, and click Create. Click Done. The Designer creates the Sequence Generator transformation.

    Edit Transformation

  • 8/6/2019 a Final Documentation

    91/167

    91

    Double-click the title bar of the transformation to open the Edit Transformations dialog box.

    Properties Tab

    Select the Properties tab. Enter settings as necessary. Click OK. To generate new sequences during a session, connect the NEXTVAL port to at least one

    transformation in the mapping.

    Choose Repository-Save.

  • 8/6/2019 a Final Documentation

    92/167

    92

    Sequence Generator Ports

    The Sequence Generator provides two output ports: NEXTVAL and CURRVAL. Use the NEXTVAL port to generate a sequence of numbers by connecting it to a

    transformation or target.

    We connect the NEXTVAL port to a downstream transformation to generate the sequencebased on the Current Value and Increment By properties.

    Connect NEXTVAL to multiple transformations to generate unique values for each row ineach transformation.

    We might connect NEXTVAL to two target tables in a mapping to generate unique primarykey values.

  • 8/6/2019 a Final Documentation

    93/167

    93

    NEXTVAL to Two Target Tables in a Mapping

    We configure the Sequence Generator transformation as follows: Current Value = 1,Increment By = 1.

    When we run the workflow, the Integration Service generates the following primary keyvalues for the T_ORDERS_PRIMARY and T_ORDERS_FOREIGN target tables.

    T_ORDERS_PRIMARY

    TABLE:

    PRIMARY KEY

    T_ORDERS_FOREIGN

    TABLE: PRIMARY

    KEY

    1 2

    3 4

    5 6

  • 8/6/2019 a Final Documentation

    94/167

    94

    7 8

    9 10

    Sequence Generator and Expression Transformation

    We configure the Sequence Generator transformation as follows: Current Value = 1,

    Increment By = 1

    Output

    key values for the T_ORDERS_PRIMARY and T_ORDERS_FOREIGN target tables

  • 8/6/2019 a Final Documentation

    95/167

    95

    T_ORDERS_PRIMARY

    TABLE:

    PRIMARY KEY

    T_ORDERS_FOREIGN

    TABLE:

    PRIMARY KEY

    1 1

    2 2

    3 3

    4 4

    5 5

    CURRVAL is the NEXTVAL value plus one or NEXTVAL plus the Increment By value. We typically only connect the CURRVAL port when the NEXTVAL port is already connected to a

    downstream transformation.

    When a row enters the transformation connected to the CURRVAL port, the Informatica Serverpasses the last-created NEXTVAL value plus one.

    Connecting CURRVAL and NEXTVAL Ports to a Target

    We configure the Sequence Generator transformation as follows: Current Value = 1, IncrementBy = 1.

    When we run the workflow, the Integration Service generates the following values for NEXTVALand CURRVAL.

  • 8/6/2019 a Final Documentation

    96/167

    96

    OUT PUT

    When we run the workflow, the Integration Service generates the following values for NEXTVALand CURRVAL.

    If we connect the CURRVAL port without connecting the NEXTVAL port, the Integration Servicepasses a constant value for each row.

    NEXTVALCURRVAL

    1 2

    2 3

    3 4

    4 5

    5 6

  • 8/6/2019 a Final Documentation

    97/167

    97

    Only the CURRVAL Port to a Target

    For example, we configure the Sequence Generator transformation as follows.

    OUTPUT

    Current Value = 1, Increment By = 1 When we run the workflow, the Integration Service generates the following constant values for

    CURRVAL.

    CURRVAL

    1

    1

    1

  • 8/6/2019 a Final Documentation

    98/167

    98

    1

    1

    4.9Source Qualifier Transformation A Source Qualifier is an active transformation. The Source Qualifier represents the rows that the Integration Service reads when it executes

    a session.

    When we add a relational or a flat file source definition to a mapping source Qualifiertransformation automatically comes.

    Task of Source Qualifier Transformation

    We can use the Source Qualifier to perform the following tasks.

    Join data originating from the same source database. Filter records when the Integration Service reads source data.

  • 8/6/2019 a Final Documentation

    99/167

    99

    Specify an outer join rather than the default inner join Specify sorted ports. Select only distinct values from the source. Create a custom query to issue a special SELECT statement for the Integration Service to read

    source data.

    Default Query of Source Qualifier

    For relational sources, the Integration Service generates a query for each Source Qualifier whenit runs a session.

    The default query is a SELECT statement for each source column used in the mapping.

    To view the Default Query

    Toview the default query. From the Properties tab, select SQL Query Click Generate SQL

    Click Cancel to exit

  • 8/6/2019 a Final Documentation

    100/167

    100

    Example of source Qualifier Transformation

    We might see all the orders for the month, including order number, order amount, andcustomer name.

    The ORDERS table includes the order number and amount of each order, but not the customername. To include the customer name, we need to join the ORDERS and CUSTOMERS tables.

    Setting the properties to Source Qualifier

    Double-click the title bar of the transformation to open the Edit Transformations dialog box. Select the Properties tab. Enter settings as necessary.

  • 8/6/2019 a Final Documentation

    101/167

    101

    SQL Query

    We can give query in the Source Qualifier transformation. From the Properties tab, select SQL Query the SQL Editor displays. Click Generate SQL.

    Joining Source Data

    We can use one Source Qualifier transformation to join data from multiple relational tables.These tables must be accessible from the same instance or database server.

  • 8/6/2019 a Final Documentation

    102/167

    102

    Use the Joiner transformation for heterogeneous sources and to join flat files.

    Sorted Ports

    In the Mapping Designer, open a Source Qualifier transformation, and click the Properties tab. Click in Number of Sorted Ports and enter the number of ports we want to sort. The Integration Service adds the configured number of columns to an ORDER BY clause, starting

    from the top of the Source Qualifier transformation.

    The source database sort order must correspond to the session.

  • 8/6/2019 a Final Documentation

    103/167

    103

    4.10

    Aggregator Transformation

    The Aggregator is an active transformation. The Aggregator transformation allows us to perform aggregate calculations, such as

    averages and sums.

    The Aggregator transformation is unlike the Expression transformation, in that we can usethe Aggregator transformation to perform calculations on groups.

    The Expression transformation permits us to perform calculations on a row-by-row basisonly.

    We can use conditional clauses to filter rows, providing more flexibility than SQL language. The Integration Services performs aggregate calculations as it reads, and stores necessary

    data group and row data in an aggregate cache.

    Components of the Aggregator Transformation

    Aggregate expression Group by port Sorted input Aggregate cache

    Aggregate Expression

    An aggregate expression can include conditional clauses and non-aggregate functions. Itcan also include one aggregate function nested within another aggregate function, such as.

    MAX( COUNT( ITEM )Aggregate Functions

    The aggregate functions can be used within an Aggregator transformation. You can nest one aggregate function within another aggregate function. AVG

  • 8/6/2019 a Final Documentation

    104/167

    104

    COUNT

    Aggregate Functions

    FIRST LAST MEDIAN MAX MIN STDDEV PERCENTILE SUM VARIANCEConditional Clauses

    We use conditional clauses in the aggregate expression to reduce the number of rows used inthe aggregation. The conditional clause can be any clause that evaluates to TRUE or FALSE.

    Null Values in Aggregate Functions

    When we configure the Integration Service, we can choose how we want the Integration

    Service to handle null values in aggregate functions. We can choose to treat null values in

    aggregate functions as NULL or zero. By default, the Integration Service treats null values as

    NULL in aggregate functions.

    Creating Aggregator Transformation

    In the Mapping Designer, click Transformation > Create. Select the Aggregatortransformation. Enter a name for the Aggregator, click Create. Then click done.

  • 8/6/2019 a Final Documentation

    105/167

    105

    The Designer creates the Aggregator transformation. Drag the ports to the Aggregator transformation. The Designer creates input/output ports for each port we include.

    Double-click the title bar of the transformation to open the Edit Transformations dialog box . Select the Ports tab. Click the group by option for each column you want the Aggregator to use in creating

    groups.

    Click Add and enter a name and data type for the aggregate expression port. Make the portan output port by clearing Input (I). Click in the right corner of the Expression field to open

    the Expression Editor. Enter the aggregate expression, click Validate, and click OK.

  • 8/6/2019 a Final Documentation

    106/167

    106

    Add default values for specific ports. Select the Properties tab. Enter settings as necessary.

    Click OK. Choose Repository-Save.

    4.11 Update Strategy An Update Strategy is an active transformation. When we design a data warehouse, we need to decide what type of information to store in

    targets. As part of the target table design, we need to determine whether to maintain all thehistoric data or just the most recent changes.

    The model we choose determines how we handle changes to existing rows. In PowerCenter,we set the update strategy at two different levels.

    Within a session

  • 8/6/2019 a Final Documentation

    107/167

    107

    Within a mapping

    Setting the Update Strategy

    We use the following steps to define an update strategy To control how rows are flagged for insert, update, delete, or reject within a mapping, add an

    Update Strategy transformation to the mapping. Update Strategy transformations are

    essential if we want to flag rows destined for the same target for different database

    operations, or if we want to reject rows.

    Define how to flag rows when we configure a session. We can flag all rows for insert, delete,or update, or we can select the data driven option, where the Integration Service follows

    instructions coded into Update Strategy transformations within the session mapping.

    Define insert, update, and delete options for each target when we configure a session. On atarget-by-target basis, we can allow or disallow inserts and deletes.

    Creating an Update Transformation

    In the Mapping Designer, select Transformation-Create. Select the Update transformation.The naming convention for Update transformations is UPD_TransformationName.

    Enter a name for the Update transformation, and click Create. Click Done. The Designer creates the Update transformation.

    Drag all ports from another transformation representing data we want to pass through theUpdate Strategy transformation.

  • 8/6/2019 a Final Documentation

    108/167

    108

    In the Update Strategy transformation, the Designer creates a copy of each port we drag. TheDesigner also connects the new port to the original port. Each port in the Update Strategy

    transformation is a combination of input/output port.

    Normally, we would select all of the columns destined for a particular target. After they passthrough the Update Strategy transformation, this information is flagged for update, insert,

    delete, or reject.

    Double-click the title bar of the transformation to open the Edit Transformations dialog box. Click the Properties tab.

    Click the button in the Update Strategy Expression field. The Expression Editor appears. Enter an update strategy expression to flag rows as inserts, deletes, updates, or rejects. Validate the expression and click OK. Click OK to save the changes. Connect the ports in the Update Strategy transformation to another transformation or a

    target instance.

    Click Repository > Save

  • 8/6/2019 a Final Documentation

    109/167

    109

    Setting the Update Strategy for a Session

    When we configure a session, we have several options for handling specific databaseoperations, including updates.

    Specifying an Operation for All Rows

    When we configure a session, we can select a single database operation for all rowsusing the Treat Source Rows As setting.

    Configure the Treat Source Rows As session property. Treat Source Rows displays the options like. Insert Delete Update

  • 8/6/2019 a Final Documentation

    110/167

    110

    Data Driven

    Specifying Operations for Individual Target Tables

    Once we determine how to treat all rows in the session, we also need to set update strategyoptions for individual targets. Define the update strategy options in the Transformations view

    on mapping tab of the session properties.

    We can set the following update strategy options for Individual Target Tables. Insert. Select this option to insert a row into a target table. Delete. Select this option to delete a row from a table.. Update. You have the following options in this situation.

    Update as Update. Update each row flagged for update if it exists in the target table.

    Update as Insert. Inset each row flagged for update. Update else Insert. Update the row if it exists. Otherwise, insert it.

    Truncate table. Select this option to truncate the target table before loading data.

  • 8/6/2019 a Final Documentation

    111/167

    111

    Specifying Operations for Individual Target Tables

    Once we determine how to treat all rows in the session, we also need to set update strategyoptions for individual targets. Define the update strategy options in the Transformations view

    on mapping tab of the session properties.

    We can set the following update strategy options for Individual Target Tables. Insert. Select this option to insert a row into a target table. Delete. Select this option to delete a row from a table. Update. You have the following options in this situation.

    Update as Update. Update each row flagged for update if it exists in the target table. Update as Insert. Inset each row flagged for update. Update else Insert. Update the row if it exists. Otherwise, insert it.

    Truncate table. Select this option to truncate the target table before loading data.

  • 8/6/2019 a Final Documentation

    112/167

    112

    4.12 Stored procedure Transformation A Stored Procedure is a passive transformation A Stored Procedure transformation is an important tool for populating and maintaining

    databases. Database administrators create stored procedures to automate tasks that are

    too complicated for standard SQL statements.

    Stored procedures run in either connected or unconnected mode. The mode we use dependson what the stored procedure does and how we plan to use it in a session. we can configureconnected and unconnected Stored Procedure transformations in a mapping.

    Connected: The flow of data through a mapping in connected mode also passesthrough the Stored Procedure transformation. All data entering the transformation

    through the input ports affects the stored procedure. We should use a connected

    Stored Procedure transformation when we need data from an input port sent as an

  • 8/6/2019 a Final Documentation

    113/167

    113

    input parameter to the stored procedure, or the results of a stored procedure sent

    as an output parameter to another transformation.

    Unconnected: The unconnected Stored Procedure transformation is not connecteddirectly to the flow of the mapping. It either runs before or after the session, or is

    called by an expression in another transformation in the mapping.

    Creating a Stored Procedure Transformation

    After we configure and test a stored procedure in the database, we must create the Stored

    Procedure transformation in the Mapping Designer

    To import a stored procedure

    In the Mapping Designer, click Transformation >Import Stored Procedure. Select the database that contains the stored procedure from the list of ODBC sources.

    Enter the user name, owner name, and password to connect to the database and click

    Connect

  • 8/6/2019 a Final Documentation

    114/167

    114

    Select the procedure to import and click OK..

    The Stored Procedure transformation appears in the mapping. The Stored Proceduretransformation name is the same as the stored procedure we selected.

    Open the transformation, and click the Properties tab Select the database where the stored procedure exists from the Connection

    Information row. If we changed the name of the Stored Procedure transformation to

    something other than the name of the stored procedure, enter the Stored Procedure

    Name.

    Click OK. Click Repository > Save to save changes to the mapping.

  • 8/6/2019 a Final Documentation

    115/167

    115

    4.13 Rank Transformation The Rank transformation is Active Transformation The Rank transformation allows us to select only the top or bottom rank of data. The Rank transformation differs from the transformation functions MAX and MIN, to

    select a group of top or bottom values, not just one value.

  • 8/6/2019 a Final Documentation

    116/167

    116

    Creating Rank Transformation In the Mapping Designer, click Transformation > Create. Select the Rank

    transformation. Enter a name for the Rank. The naming convention for Rank

    transformations is RNK_TransformationName.

    Enter a description for the transformation. This description appears in theRepository Manager.

    Click Create, and then click done. The Designer creates the Rank transformation. Link columns from an input transformation to the Rank transformation. Click the Ports tab, and then select the Rank (R) option for the port used to

    measure ranks.

    If we want to create groups for ranked rows, select Group By for the port thatdefines the group.

  • 8/6/2019 a Final Documentation

    117/167

    117

    Click the Properties tab and select whether we want the top or bottom rank For the Number of Ranks option, enter the number of rows we want to select for

    the rank.

    Change the other Rank transformation properties, if necessary. Click OK. Click Repository > Save.

    Properties Tab

  • 8/6/2019 a Final Documentation

    118/167

    118

    4.14 Java Transformation

    The Java transformation is a Active/Passive Connected transformation that provides asimple native programming interface to define transformation functionality with the Java

    programming language.

    You create Java transformations by writing Java code snippets that define transformationlogic.

    The Power Center Client uses the Java Development Kit (JDK) to compile the Java code andgenerate byte code for the transformation. The Integration Service uses the Java Runtime

    Environment (JRE) to execute generated byte code at run time.

    Steps To Define Java Transformation

    Create the transformation in the Transformation Developer or Mapping Designer.

  • 8/6/2019 a Final Documentation

    119/167

    119

    Configure input and output ports and groups for the transformation. Use port names asvariables in Java code snippets.

    Configure the transformation properties. Use the code entry tabs in the transformation to write and compile the Java code for the

    transformation.

    Locate and fix compilation errors in the Java code for the transformation. Enter the ports and use that ports as identifier in java code. Go to java code and enter the java code and click compile and check the output in the output

    window.

    Create session and workflow and run the session.Functions

    Some functions used in designer are

    AVGSyntax: AVG( numeric_value [,filter_condition ] )

    MAXSyntax: MAX( value [,filter_condition ] )

    MINSyntax :MIN( value [,filter_condition ] )

    INSTRSyntax: INSTR (string, search_value [, start[, occurrence ] ] )

    SUBSTRSyntax: SUBSTR (string, start[, length])

    IS_DATESyntax: IS_DATE (value)

  • 8/6/2019 a Final Documentation

    120/167

    120

    4.15 User Defined Functions We can create user-defined functions using the PowerCenter transformation language.

    Create user-defined functions to reuse expression logic and build complex expressions.

    User-defined functions are available to other users in a repository.

    Once you create user-defined functions, we can manage them from the User-DefinedFunction Browser dialog box. We can also use them as functions in the Expression Editor.

    They display on the User-Defined Functions tab of the Expression Editor.

    We create a user-defined function in the Transformation Developer. Configure thefollowing information when we create a user-defined function.

    Name Type Description Arguments Syntax

  • 8/6/2019 a Final Documentation

    121/167

    121

    Steps to Create User-Defined Functions

    In the Transformation Developer, click Tools > User-Defined Functions. Click New

    The Edit User-Defined Function dialog box appears

    Enter a function name Select a function type

    If we create a public user-defined function, we cannot change the function to private when weedit the function

    Optionally, enter a description of the user-defined function.We can enter up to 2,000 characters.

    Create arguments for the user-defined function.When we create arguments, configure the argument name, data type, precision, and

    scale. We can select transformation data types.

    Click Launch Editor to create an expression that contains the arguments we defined. Click OK

    The Designer assigns the data type of the data the expression returns. The data types

    have the precision and scale of transformation data types.

  • 8/6/2019 a Final Documentation

    122/167

    122

    Click OKThe expression displays in the User-Defined Function Browser dialog box.

    4.16 Data Profiling Data profiling is a technique used to analyze source data. PowerCenter Data Profiling can

    help us to evaluate source data and detect patterns and exceptions. We can profile

    source data to suggest candidate keys, detect data patterns and evaluate join criteria.

    Use Data Profiling to analyze source data in the following situations. During mapping development. During production to maintain data quality. To profile source data, we create a data profile. We can create a data profile based on a

    source or mapplet in the repository. Data profiles contain functions that perform

    calculations on the source data.

    The repository stores the data profile as an object. We can apply profile functions to acolumn within a source, to a single source, or to multiple sources.

    We can create the following types of data profiles. Auto profile

    Contains a predefined set of functions for profiling source data. Use an auto profile

    during mapping development.

    Custom profileUse a custom profile during mapping development to validate documented business

    rules about the source data. We can also use a custom profile to monitor data

    quality or validate the results of BI reports.

    Steps To Create Auto Profile

    When we create an auto profile, we can profile groups or columns in the source. Or, we canprofile the entire source.

  • 8/6/2019 a Final Documentation

    123/167

    123

    To create an auto profile.Select the source definition in the Source Analyzer or mapplet in the Mapplet Designer you

    want to profile.

    Launch the Profile Wizard from the following Designer tools.Source Analyzer. Click Sources > Profiling > Create Auto Profile.Mapplet Designer. Click Mapplets > Profiling > Create Auto Profile.You set the default data profile options to open the Auto Profile Column Selection

    dialog box when you create an auto profile.

    The source definition contains 25 or more columns. Optionally, click Description to add a description for the data profile. Click OK.

    Enter a description up to 200 characters.

    Optionally, select the groups or columns in the source that you want to profile.By default, all columns or groups are selected

    Select Load Verbose Data if you want the Integration Service to write verbose data to theData Profiling warehouse during the profile session. By default, Load Verbose Data option

    is disabled.

    Click Next. Select additional functions to include in the auto profile. We can also clear functions we

    do not want to include.

  • 8/6/2019 a Final Documentation

    124/167

    124

    Optionally, click Save As Default to create new default functions based on the functionsselected here.

    Optionally, click Profile Settings to enter settings for domain inference and structureinference tuning.

    Optionally, modify the default profile settings and click OK. Click Configure Session to configure the session properties after you create the data

    profile.

    Click Next if you selected Configure Session, or click Finish if you disabled ConfigureSession.

    The Designer generates a data profile and profile mapping based on the profile functions. Configure the Profile Run options and click next. Configure the Session Setup options. Click Finish.

    We can create a custom profile from the following Designer tools. Source Analyzer. Click Sources > Profili