Oracle® Business Intelligence: Data Warehouse Administration Console User's Guide
Oracle® Business Intelligence: Data Warehouse Administration Console User's Guide
Oracle® Business Intelligence: Data Warehouse Administration Console User's Guide
August 2008
Oracle Business Intelligence Data Warehouse Administration Console Users Guide Version 10.1.3.4 E12652-01 Copyright 2003, 2008, Oracle. All rights reserved. The Programs (which include both the software and documentation) contain proprietary information; they are provided under a license agreement containing restrictions on use and disclosure and are also protected by copyright, patent, and other intellectual and industrial property laws. Reverse engineering, disassembly, or decompilation of the Programs, except to the extent required to obtain interoperability with other independently created software or as specified by law, is prohibited. The information contained in this document is subject to change without notice. If you find any problems in the documentation, please report them to us in writing. This document is not warranted to be error-free. Except as may be expressly permitted in your license agreement for these Programs, no part of these Programs may be reproduced or transmitted in any form or by any means, electronic or mechanical, for any purpose. If the Programs are delivered to the United States Government or anyone licensing or using the Programs on behalf of the United States Government, the following notice is applicable: U.S. GOVERNMENT RIGHTS Programs, software, databases, and related documentation and technical data delivered to U.S. Government customers are "commercial computer software" or "commercial technical data" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the Programs, including documentation and technical data, shall be subject to the licensing restrictions set forth in the applicable Oracle license agreement, and, to the extent applicable, the additional rights set forth in FAR 52.227-19, Commercial Computer Software--Restricted Rights (June 1987). Oracle USA, Inc., 500 Oracle Parkway, Redwood City, CA 94065. The Programs are not intended for use in any nuclear, aviation, mass transit, medical, or other inherently dangerous applications. It shall be the licensee's responsibility to take all appropriate fail-safe, backup, redundancy and other measures to ensure the safe use of such applications if the Programs are used for such purposes, and we disclaim liability for any damages caused by such use of the Programs. Oracle, JD Edwards, PeopleSoft, and Siebel are registered trademarks of Oracle Corporation and/or its affiliates. Other names may be trademarks of their respective owners. The Programs may provide links to Web sites and access to content, products, and services from third parties. Oracle is not responsible for the availability of, or any content provided on, third-party Web sites. You bear all risks associated with the use of such content. If you choose to purchase any products or services from a third party, the relationship is directly between you and the third party. Oracle is not responsible for: (a) the quality of third-party products or services; or (b) fulfilling any of the terms of the agreement with the third party, including delivery of products or services and warranty obligations related to purchased products or services. Oracle is not responsible for any loss or damage of any sort that you may incur from dealing with any third party.
Contents
Preface ................................................................................................................................................................. ix
Audience....................................................................................................................................................... Documentation Accessibility ..................................................................................................................... Related Documents ..................................................................................................................................... Conventions ................................................................................................................................................. ix ix x x
Best Practices for Tables .......................................................................................................................... Best Practices for Indexes........................................................................................................................ Best Practices for Columns ..................................................................................................................... Best Practices for Configuration Tags................................................................................................... Best Practices for Source System Parameters...................................................................................... Best Practices for Subject Areas............................................................................................................. Best Practices for Execution Plans.........................................................................................................
iv
About Parameter Management.............................................................................................................. 8-8 Overview of Parameters.................................................................................................................... 8-8 Parameter Data Types ................................................................................................................ 8-8 Preconfigured Parameters ................................................................................................................ 8-9 How DAC Handles Parameters at Runtime .................................................................................. 8-9 Nesting Parameters within Other Parameters ............................................................................ 8-10 Defining a Text Type Parameter ................................................................................................... 8-10 Defining a Database Specific Text Type Parameter ................................................................... 8-11 Defining a Timestamp Type Parameter....................................................................................... 8-11 Defining a SQL Type Parameter ................................................................................................... 8-12 Specifying Tablespaces for Indexes by Table Type ........................................................................ 8-13 Working with Configuration Tags ..................................................................................................... 8-13 Overview of Subject Areas.................................................................................................................. 8-17 Designing a Subject Area ............................................................................................................... 8-17 Previewing and Pruning Subject Areas ................................................................................ 8-17 How DAC Determines Tasks Required for Subject Areas........................................................ 8-17 Creating a Subject Area........................................................................................................................ 8-18
Repository Upgrade (784): High-Level Process Flow................................................................ Repository Upgrade (784): Procedure for Upgrading ............................................................... About the Refresh Base Option.......................................................................................................... Refresh Base: High-Level Process Flow ....................................................................................... Refresh Base: Procedure for Upgrading ...................................................................................... About the Simplified Refresh From Base Option ........................................................................ About the Replace Base Option........................................................................................................ Replace Base: High-Level Process Flow..................................................................................... Replace Base: Procedure for Upgrading .................................................................................... About the Peer to Peer Merge Option ............................................................................................. Peer to Peer Merge: High-Level Process Flow.......................................................................... Peer to Peer Merge: Procedure for Merging.............................................................................. Resolving Object Differences in the View Difference Report ................................................... Overview of View Difference Report......................................................................................... View Difference Report Interface ............................................................................................... Possible Repository Merge Outcomes Based on Your Decisions...........................................
10-4 10-4 10-7 10-8 10-9 10-10 10-11 10-12 10-13 10-14 10-15 10-15 10-17 10-17 10-18 10-19
11
vi
Change Capture Using Tables ............................................................................................. Primary and Auxiliary Tables....................................................................................... Example: Building S_ETL_I_IMG_ Table for Loading Account Dimension ......... Change Capture Using the Date Column........................................................................... Using the Change Capture Filter...................................................................................................... Tracking Deleted Records.................................................................................................................. Pointing Multiple Informatica PowerCenter Services to a Single Informatica Repository . Handling ETL Failures with the DAC............................................................................................. When the Execution of an Execution Plan Fails ....................................................................... In Case of Abnormal Termination of the DAC Server............................................................. Discarding the Current Run Execution Plan............................................................................. Failure of Aggregator Transformation Tasks with Sorted Input ...........................................
11-13 11-13 11-14 11-16 11-16 11-16 11-18 11-18 11-18 11-19 11-19 11-19
vii
Tasks Tab: Configuration Tags Subtab ............................................................................... Tasks Tab: Parameters Subtab.............................................................................................. Tasks Tab: Phase Dependency Subtab................................................................................ Tasks Tab: Refresh Date Tables Subtab .............................................................................. Tasks Tab: Source Tables Subtab ......................................................................................... Tasks Tab: Target Tables Subtab.......................................................................................... Setup View Tabs .................................................................................................................................. DAC System Properties Tab ........................................................................................................ Email Recipients Tab .................................................................................................................... Informatica Servers Tab ............................................................................................................... Physical Data Sources Tab ........................................................................................................... Physical Data Sources Tab: Index Spaces Subtab.............................................................. Physical Data Sources Tab: Analyze Frequencies Subtab................................................ Physical Data Sources Tab: Refresh Dates Subtab ............................................................ Execute View Tabs............................................................................................................................... Current Run Tab............................................................................................................................ Current Run Tab: Audit Trail (RO) Subtab ........................................................................ Current Run Tab: Phase Summary (RO) Subtab ............................................................... Current Run Tab: Run Type Summary (RO) ..................................................................... Current Run Tab: Tasks Subtab ........................................................................................... Current Run Tab: Task Details Subtab ............................................................................... Execution Plans Tab...................................................................................................................... Execution Plans Tab: All Dependencies Subtab ................................................................ Execution Plans Tab: Following Tasks Subtab .................................................................. Execution Plans Tab: Immediate Dependencies Subtab .................................................. Execution Plans Tab: Ordered Tasks Subtab ..................................................................... Execution Plans Tab: Parameters Subtab ........................................................................... Execution Plans Tab: Preceding Tasks Subtab................................................................... Execution Plans Tab: Refresh Dates Subtab ....................................................................... Execution Plans Tab: Subject Areas Subtab ....................................................................... Execution Plans Tab: Tables (RO) Subtab .......................................................................... Run History Tab ............................................................................................................................ Run History Tab: Audit Trail (RO) Subtab ........................................................................ Run History Tab: Phase Summary (RO) Subtab ............................................................... Run History Tab: Run Type Summary (RO) Subtab ........................................................ Scheduler Tab ................................................................................................................................
12-19 12-19 12-20 12-21 12-22 12-22 12-24 12-25 12-29 12-30 12-32 12-33 12-34 12-34 12-35 12-36 12-37 12-37 12-38 12-38 12-38 12-39 12-40 12-40 12-40 12-41 12-42 12-42 12-43 12-43 12-43 12-44 12-44 12-44 12-44 12-45
Index
viii
Preface
The Oracle Business Intelligence Data Warehouse Administration Console Users Guide contains information about using the Data Warehouse Administration Console (DAC), a centralized console for management, configuration, administration, loading, and monitoring of the Oracle Business Analytics Warehouse. Oracle recommends reading the Oracle Business Intelligence Data Warehouse Administration Console Release Notes before installing, using, or upgrading DAC. The Oracle Business Intelligence Data Warehouse Administration Console Release Notes are available:
On the Oracle Business Intelligence Data Warehouse Administration Console CD-ROM. On the Oracle Technology Network at http://www.oracle.com/technology/documentation/bi_dac.html (to register for a free account on the Oracle Technology Network, go to http://www.oracle.com/technology/about/index.html).
Audience
This document is intended for data warehouse administrators and ETL developers and operators.
Documentation Accessibility
Our goal is to make Oracle products, services, and supporting documentation accessible, with good usability, to the disabled community. To that end, our documentation includes features that make information available to users of assistive technology. This documentation is available in HTML format, and contains markup to facilitate access by the disabled community. Accessibility standards will continue to evolve over time, and Oracle is actively engaged with other market-leading technology vendors to address technical obstacles so that our documentation can be accessible to all of our customers. For more information, visit the Oracle Accessibility Program Web site at http://www.oracle.com/accessibility Accessibility of Code Examples in Documentation Screen readers may not always correctly read the code examples in this document. The conventions for writing code require that closing braces should appear on an otherwise empty line; however, some screen readers may not always read a line of text that consists solely of a bracket or brace.
ix
Accessibility of Links to External Web Sites in Documentation This documentation may contain links to Web sites of other companies or organizations that Oracle does not own or control. Oracle neither evaluates nor makes any representations regarding the accessibility of these Web sites. TTY Access to Oracle Support Services Oracle provides dedicated Text Telephone (TTY) access to Oracle Support Services within the United States of America 24 hours a day, seven days a week. For TTY support, call 800.446.2398.
Related Documents
For more information, see the following documents in the Oracle Business Intelligence Applications Release 7.9.4 documentation set (available at http://www.oracle.com/technology/documentation/bi_apps.html):
Oracle Business Intelligence Data Warehouse Administration Console Release Notes System Requirements and Supported Platforms for Oracle Business Intelligence Data Warehouse Administration Console Oracle Business Intelligence Data Warehouse Administration Console Installation, Configuration, and Upgrade Guide
Conventions
The following text conventions are used in this document:
Convention boldface italic monospace Meaning Boldface type indicates graphical user interface elements associated with an action, or terms defined in text or the glossary. Italic type indicates book titles, emphasis, or placeholder variables for which you supply particular values. Monospace type indicates commands within a paragraph, URLs, code in examples, text that appears on the screen, or text that you enter.
1
1
This section lists changes described in this version of the documentation to support version 10.1.3.4 of the software.
What's New in Oracle Business Intelligence Data Warehouse Administration Console Users Guide, Version 10.1.3.4
The Oracle Business Intelligence Data Warehouse Administration Console Users Guide includes the following new and changed topics:
User Management. The User Management feature enables you to create and manage user accounts. See "About User Account Management". DAC Repository Database Authentication File. Upon logging into DAC for the first time, the DAC Repository Database Authentication File authenticates the database in which the repository resides. See "DAC Repository Database Authentication File". Best Practices for Defining Repository Objects. See "Best Practices for Defining Repository Objects". Enhanced Subject Area Design. See "Customizing DAC Objects and Designing Subject Areas" Enhanced Execution Plan Design. See "Building, Running and Monitoring Execution Plans". Multi-Source Execution Plans. See "About Multi-Source Execution Plans". Repository Audit Trail. See "Tools Menu Options". Actions Feature. Provides a flexible way to drop and create indexes, truncate and analyze tables, and define SQL scripts to enhance task behaviors. See "About Index, Table and Task Actions". Upgrade/Merge Wizard. Enables you to upgrade and merge DAC repositories by comparing repositories and creating a Difference Report. See "Upgrading, Comparing and Merging DAC Repositories".
What's New in Oracle Business Intelligence Data Warehouse Administration Console Users Guide, Version 10.1.3.4
1-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
2
Overview of Oracle Business Analytics Warehouse
This chapter provides an overview of the Oracle Business Analytics Warehouse and the Data Warehouse Administration Console (DAC). It includes the following topics:
Oracle Business Analytics Warehouse Overview Oracle Business Analytics Warehouse Architecture
A data integration engine that combines data from multiple source systems to build a data warehouse. An open architecture to allow organizations to use third-party analytical tools in conjunction with the Oracle Business Analytics Warehouse using the Oracle Business Intelligence Server. Prebuilt data extractors to incorporate data from external applications into the Oracle Business Analytics Warehouse. A set of ETL (extract-transform-load) processes that takes data from multiple source systems and creates the Oracle Business Analytics Warehouse tables. The DAC, a centralized console for schema management as well as configuration, administration, loading, and monitoring of the Oracle Business Analytics Warehouse.
transform, and load data from transactional databases into the Oracle Business Analytics Warehouse. Figure 21 illustrates how the Oracle Business Analytics Warehouse interacts with the other components of Oracle BI Applications.
Figure 21 Oracle Business Intelligence Applications Architecture
DAC Client. A command and control interface for the data warehouse to allow for schema management, and configuration, administration, and monitoring of data warehouse processes. It also enables you to design subject areas and build execution plans.
2-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
DAC Server. Executes the instructions from the DAC Client. The DAC Server manages data warehouse processes, including loading of the ETL and scheduling execution plans. It dynamically adjusts its actions based on information in the DAC repository. Depending on your business needs, you might incrementally refresh the Oracle Business Analytics Warehouse once a day, once a week, once a month, or on another similar schedule. DAC repository. Stores the metadata (semantics of the Oracle Business Analytics Warehouse) that represents the data warehouse processes. Informatica Server. Also referred to as PowerCenter Services. Loads and refreshes the Oracle Business Analytics Warehouse. Informatica Repository Server. Also referred to as Repository Services. Manages the Informatica repository. Informatica Repository. Stores the metadata related to Informatica workflows. Informatica client utilities. Tools that enable you to create and manage the Informatica repository.
2-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
3
3
Introduction to DAC About the DAC Process Life Cycle About Source System Containers About DAC Repository Objects Held in Source System Containers
Introduction to DAC
DAC provides a framework for the entire life cycle of data warehouse implementations. It enables you to create, configure, execute, and monitor modular data warehouse applications in a parallel, high-performing environment. For information about the DAC process life cycle, see "About the DAC Process Life Cycle". DAC complements the Informatica ETL platform. It provides application-specific capabilities that are not prebuilt into ETL platforms. For example, ETL platforms are not aware of the semantics of the subject areas being populated in the data warehouse nor the method in which they are populated. DAC provides the following application capabilities at a layer of abstraction above the ETL execution platform:
Dynamic generation of subject areas and execution plans Dynamic settings for parallelism and load balancing Intelligent task queue engine based on user- defined and computed scores Automatic full and incremental mode aware Index management for ETL and query performance Embedded high performance Siebel OLTP change capture techniques Ability to restart at any point of failure Phase-based analysis tools for isolating ETL bottlenecks
Create a physical data model in the data warehouse Set language, currency, and other settings Design subject areas and build execution plans
Basic Concepts About DAC 3-1
Introduction to DAC
Generate custom ETL execution plans Automate change capture for the Siebel transactional database Capture deleted records Assist in index management Perform dry runs and test runs of execution plans
Perform error monitoring and email alerting Perform structured ETL analysis and reporting
Automate full and incremental mode optimization rules Set the level of Informatica session concurrency Load balance across multiple Informatica servers Restart from point of failure Queue execution tasks for performance (See Figure 31.) DAC manages the task execution queue based on metadata driven priorities and scores computed at runtime. This combination allows for flexible and optimized execution. Tasks are dynamically assigned a priority based on their number of dependents, number of sources, and average duration.
3-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
The phases of the process and the actions associated with them are as follows:
Setup Set up database connections Set up ETL processes (Informatica) Set up email recipients
Execute Define scheduling parameters to run execution plans Access runtime controls to restart or stop currently running schedules
Monitor Monitor runtime execution of data warehouse applications Monitor users, DAC repository, and application maintenance jobs
the DAC client to track the customizations, such as newly created objects, modified objects, and those that are used as is.
Caution: You should not modify objects in the preconfigured source system containers either through the DAC client or directly through SQL statements to the DAC repository. You must make a copy of a preconfigured container in order to make any changes to it.
For instructions on creating a new source system container or copying an existing container, see "Creating or Copying a Source System Container".
Subject areas. A logical grouping of tables related to a particular subject or application context. It also includes the tasks that are associated with the tables, as well as the tasks required to load the tables. Subject areas are assigned to execution plans, which can be scheduled for full or incremental loads. Tables. Physical database tables defined in the database schema. Can be transactional database tables or data warehouse tables. Table types can be fact, dimension, hierarchy, aggregate, and so on, as well as flat files that can be sources or targets. Indexes. Physical database indexes to be defined in the database schema either to better the performance of the ETL processes or the queries for reporting purposes. Tasks. A unit of work for loading one or more tables. A task comprises the following: source and target tables, phase, execution type, truncate properties, and commands for full or incremental loads. When you assemble a subject area, DAC automatically assigns tasks to it. Tasks that are automatically assigned to the subject area by DAC are indicated by the Autogenerated flag in the Tasks subtab of the Subject Areas tab. Task groups. A group of tasks that you define because you want to impose a specific order of execution. A task group is considered to be a "special task." Execution plans. A data transformation plan defined on subject areas that needs to be transformed at certain frequencies of time. An execution plan is defined based on business requirements for when the data warehouse needs to be loaded. An execution plan comprises the following: ordered tasks, indexes, tags, parameters, source system folders, and phases. Schedules. A schedule specifies when and how often an execution plan runs. An execution plan can be scheduled for different frequencies or recurrences by defining multiple schedules.
3-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
3-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
4
DAC Quick Start
This section includes the following topics:
About User Account Management Creating, Deleting and Inactivating User Accounts Logging into DAC for the First Time About the DACs Extract and Load Processes Performing an Initial Full Load of Data into the Data Warehouse
A user with the Administrator role must distribute the DAC Repository database authentication file to user accounts that need to access the DAC Repository. For information about the authentication file, see "DAC Repository Database Authentication File".
Table 41 Role
User Account Roles and Permissions Permissions Read and write permission on all DAC tabs and dialog boxes.
Administrator
Table 41 (Cont.) User Account Roles and Permissions Role Developer Permissions Read and write permission on the following:
All Design view tabs All Execute view tabs Export dialog box New Source System Container dialog box Rename Source System Container dialog box Delete Source System Container dialog box Purge Run Details All functionality in the Seed Data menu
Operator
From the toolbar, select File, then User Management. In the User Management dialog box, click New. In the new record field, do the following:
a. b.
Enter a unique Name and Password. Click in the Roles field, and then select the roles you want to associate with this user account.
4. 5. 6.
Click Save. Click Close to exit the User Management dialog box. Distribute the authentication file for the database where the DAC Repository resides to the user account. For more information about authentication files, see "DAC Repository Database Authentication File".
From the toolbar, select File, then User Management. In the User Management dialog box, select the user account you want to delete. Click Delete. Click Close to exit the User Management dialog box.
From the toolbar, select File, then User Management. In the User Management dialog box, select the user account you want to inactivate. Click the Inactive check box. Click Save.
4-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
5.
The administrator generates the connection information and password files and distributes them to the users. Using the encrypted authentication file, the users will be able to use into the DAC Client using their own user name and password.
Start the DAC Client by navigating to the $ORACLE_HOME\bifoundation\dac directory and double-clicking the startclient.bat file. The Login ... dialog box appears.
2. 3. 4.
Click Configure. In the Configuring ... dialog box, select Create Connection, and then click Next. Enter the appropriate connection information:
Required Value Enter a unique name for the connection to the DAC Repository. Select the type of database in which the DAC Repository will be stored.
Required Value Select the database name or database account name of the DAC Repository. If you are using:
Oracle (OCI8), use the tnsnames entry. Oracle (Thin), use the instance name. SQL Server, use the database name. DB2-UDB, use the connect string as defined in the DB2 configuration.
Enter the name of the machine where the DAC Repository will reside. Enter the port number on which the database listens. For example, for an Oracle database the default port is 1521, or for a SQL Server database the default port is 1433. Can be used to override the standard URL for this connection. Can be used to override the standard driver for this connection. Click in this field to do one of the following:
Select an existing authentication file: Navigate to the appropriate location, select the authentication file , and click OK.
Create a new authentication file: Navigate to the folder where you want to save the authentication file, and click OK.
Click in the Authentication File field of the Configuring... dialog box. In the Authentication File dialog box, select Choose existing authentication file. Navigate to the appropriate folder, and select the authentication file. Click OK. In the Configuring... dialog box, click Test Connection to confirm the connection works. Click Apply, and then click Finish.
Note:
You must distribute this authentication file to all user accounts that need to access this DAC Repository.
6.
Click in the Authentication File field of the Configuring... dialog box. In the Authentication File dialog box, select Create authentication file. Navigate to the folder where you want to save the new authentication file, and click OK.
4-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
d. e. f. g.
In the Create Authentication File dialog box, enter a unique name for the authentication file, and click OK. Enter the Table Owner Name and Password for the database where the repository will reside. In the Configuring... dialog box, click Test Connection to confirm the connection works. Click Apply, and then click Finish.
Note:
You must distribute this authentication file to all user accounts that need to access this DAC Repository.
7.
Select the appropriate Connection from the drop-down list. Enter Administrator as the User Name. Enter Administrator as the Password. Click Login.
8.
If asked whether you want to create or upgrade the DAC Repository schema, click Yes.
Full extract and full load This extract and load combination is used for the very first extract and load. All data is extracted from the source system and loaded into the data warehouse. The DAC performs a full extract for a task if the source and staging tables have null refresh dates. The DAC performs a full load for a task if the staging and target tables have null refresh dates.
Full extract and incremental load This extract and load combination loads existing data warehouse tables with data from new sources. Data is extracted from the source system through a full extract
command. When the source or staging table is null, the DAC invokes the full extract workflow. Data is loaded from the staging table into the target table through an incremental load command. When the staging and target tables have refresh dates, the DAC invokes an incremental load command. This situation arises when data is loaded into an existing data warehouse from a new source connection. The incremental load process requires additional logic to determine whether a record should be inserted or updated. Therefore, if you add a new source connection to populate an existing data warehouse, you should expect the incremental load to be slower than when running a full load.
Incremental extract and incremental load This extract and load combination is used for regular nightly or weekly ETL processes. New or changed records are extracted from the source system and loaded into the data warehouse. The DAC performs an incremental extract for a task if the source and staging tables have refresh dates and performs an incremental load for a task if the staging and target table have refresh dates.
Create a copy of your source system container. For instructions, see "Creating or Copying a Source System Container".
2.
Create a subject area. For instructions, see "Creating a Subject Area". See also "Overview of Subject Areas".
3.
Assign a subject area to an execution plan and run the execution plan. For instructions, see "Building and Running Single-Source and Multi-Source Execution Plans".
4-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
5
5
Best Practices for Containers Best Practices for Tasks Best Practices for Task Groups Best Practices for Tables Best Practices for Indexes Best Practices for Columns Best Practices for Configuration Tags Best Practices for Source System Parameters Best Practices for Subject Areas Best Practices for Execution Plans
When changes are made to objects in the container that owns them, the change is instantaneous. Changes made to parent objects in the owner container are automatically pushed to the parent referenced objects. When you add child objects to a parent object, you must you the Push to References right-click command (Design view) to push the changes to the child referenced objects. For example, if you add a column to a table registered in DAC, the new column is not automatically added to the references in the other containers referencing the parent object. When you delete a referenced object, only the referenced object is deleted, not the original object. If you delete an object from the owner container, the object is deleted and all references are deleted from the containers that may be referencing this object. This is referred to as a deep delete. For example, if you delete a table from the owner container, the table and columns are deleted and subsequently from all the containers that may be referencing this object.
If you delete a column from the owner table, the column is deleted in all the referenced objects. If you delete child objects from the owner object, the referenced child objects are automatically deleted.
Start your work with tasks in Informatica. After you create a workflow, do the following in the DAC Task tab: Create a new task and assign it a logical (readable) name. Enter the command for a full load or incremental load. The commands can be the same. If the Command for Incremental Load field is left blank, no action occurs for this task while in incremental mode. If the Command for Full Load field is left blank, no action occurs for this task while in full mode. Make sure all the source and target tables are defined for the task. You can use the task synchronize functionality to import data from Informatica. You can also manually assign the source or target tables.
Select at least one primary table because the incremental and full mode properties are determined based on the refresh dates of the primary table. Design tasks so that they load only one table at a time. Define granular tasks rather than tasks that include bigger blocks of processes. Granular tasks are more efficient and have better restartability. Do not truncate a table on the source system tables (for example, Oracle, Siebel or Peoplesoft). Make sure the truncate property for the target tables is set properly. For tables that need to get truncated regardless of the mode of the run (Full or Incremental), set the Truncate Always property to True. For tables that need to get incrementally loaded, set the Truncate for Full Load property to True. Select the Analyze Table option if the task should analyze the table. The default value for this option is True if either of the Truncate options are selected. Do not truncate a table more than once within the single life span of an ETL. If a task that writes to a target table is contingent upon another table being loaded, use conditional tables. This ensures that the task qualifies only if the conditional table is part of the subject area design. Assign an appropriate phase to the task. An understanding of task phases is essential to understanding ETL processes. If you want to force a relationship where none exists, consider using phase dependencies. For example, if you have an ETL process in which the extract facts and extract dimensions do not share any common source or target tables, but your design requires that the extract facts should run before extracting dimensions, then, for the task that extracts facts, add extract dimension as the phase that waits.
5-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
For more information about phase dependencies, see "Tasks Tab: Phase Dependency Subtab".
Make sure you do not introduce conflicting phase dependencies. This can cause the DAC Server to hang. If the source qualifier needs to use a data parameter, always use the DAC date parameter that can be formatted to the database-specific syntax.
Do not create task groups unnecessarily, because this can impact a better way of ordering tasks. Create task groups only if you want the following scenarios and the auto-dependency (auto ordering of tasks for an execution plan) cannot solve the order in a proper way. Avoid circular relationships among tasks. For example, avoid situations in which Task 1 reads from Table A and writes to Table B, and Task 2 reads from Table B and writes to Table A. You can use task groups to avoid these situations. If the tasks belong to different phases, however, this situation is acceptable. If you have many tasks belonging to the same phase that write to the same table, you can use task groups to run the tasks in parallel. If the target tables need to be truncated before the tasks are run, select the properties Truncate Always and Truncate Full Load in the Task Group tab. Do not mix extracts and loads under a single table group. Do not make Task Groups for obvious ordering needs. DAC handles ordering in such cases. If a source system container uses a task group, make other containers that reference the task also include the task group.
Always use all upper case characters for table names. Make sure you set the Table Type property correctly in the Tables tab of the Design view. For Teradata databases, pay attention to the Set/Multiset property. If you anticipate that the data will contain duplicate rows, choose Multiset as the value for this property. DAC automatically associates foreign key tables with the referenced table. You can also define which other tables need to be loaded, such as aggregate tables, by associating these tables with the referenced table using the Related Tables subtab of the Tables tab.
If you have a foreign key column, associate the foreign key table and the join column. DAC uses this information to identify all the related tables to be loaded when a certain table needs to be loaded in the data warehouse. For Teradata databases, pay attention to the Teradata Primary Index property. For Teradata databases: Pay attention to the Teradata Primary Index property. Pay attention to which columns need to gather statistics. Note that column statistics are somewhat equivalent to indexes. If you would have had indexes that span multiple columns for other databases, consider defining multi-column statistics for Teradata.
Do not register any columns for source system container tables. Make sure you add all the appropriate system columns. For example, all tables should have the following: ROW_WID in number format. INTEGRATION_ID in varchar format. DATASOURCE_NUM_ID in number format.
Always use all upper case characters for table names. Make sure you set the Table Type property correctly in the Tables tab of the Design view. Always use all upper case characters for column names. If you have a foreign key column, associate the foreign key table with the join column. DAC uses this information to identify all the related tables to be loaded when a certain table needs to be loaded in the data warehouse. For Teradata databases: Pay attention to which columns need to gather statistics. Note that column statistics are somewhat equivalent to indexes. If you would have had indexes that span multiple columns for other databases, consider defining multi-column statistics for Teradata. Pay attention to the Teradata Primary Index property.
Do not register any columns for source system container tables. Make sure you add all the appropriate system columns. For example, all tables should have the following: ROW_WID in number format. INTEGRATION_ID in varchar format. DATASOURCE_NUM_ID in number format. ETL_PROC_WID in number format.
5-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Use configuration tags to tag tasks that you do not want to be part of all the defined subject areas. A tagged task can be re-associated with a subject area by assigning the configuration tag to the subject area.
Use source system parameters when tasks in the source system container need a particular value.
To define a subject area, associate only fact tables with it. DAC automatically computes which additional aggregate tables and dimension tables to associate with the subject area based on the related tables you define and foreign key relationships. If you delete a task from a subject area using the Delete button on the Task tab, the next time you assemble the subject area the task may be included. However, if you inactivate the task by selecting Inactive in the Task tab, the task will remain inactive when you re-assemble the subject area. Avoid adding tasks or inactivating tasks manually.
If many tasks with the same name across source system containers read and write to the same data source, DAC will consider them to be the same task. If the logical source to physical mapping yields multiple records, DAC will produce as many runtime instances of the task. DAC orders tasks in the most efficient manner possible based on the following: Phase of task Source and target tables Truncate table properties Data source priority
5-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
6
Overview of the DAC Interface
This chapter contains the following topics:
Navigating the DAC Interface The DAC Menu Bar The DAC Views The DAC Top Pane Toolbar The DAC Right-Click Menus The DAC Server Monitor Icons The DAC Navigation Tree The DAC Editable Lists Using the DAC Query Functionality
User Management Enables users with the Administrator role to manage user accounts. New Source System Container Enables you to create a new source system container or make a copy of an existing container. For instructions see, "Creating or Copying a Source System Container".
Rename Source System Container Enables you to rename an existing source system container. Delete Source System Container Enables you to delete an existing source system container. Close Closes the DAC client.
Views
The Views menu enables you to navigate to the various tabs in the top pane window.
6-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Table 61 (Cont.) DAC Menu Bar Options Menu Names Tools Description The Tools menu provides access to functionality related to the DAC and Informatica repositories. Table 62 provides a description of the Tools menu commands. The Help menu provides access to the following:
Help
Login Details Lists database connection information. System Information DAC Help About DAC Lists information about the DAC version.
Logical. Exports all information contained in the Design view and database connection information. Run Time. Exports all information contained in the Execute view. System. Exports all information contained in the Setup view, except passwords for servers and database connections.
Enables you to import the DAC metadata for the source system containers you specify. In the Import dialog, you can specify the following:
Import/Export folder. A directory from which to import the data. The default directory is DAC\export. Truncate repository tables. Indicates whether you want to truncate the repository tables. If you select this option, the existing metadata is overwritten. Enable batch mode. Indicates whether batch mode is enabled. In Batch mode the imported metadata is inserted into the repository as an array insert.
In the Import dialog, you can select the following category options:
Logical. Imports all information contained in the Design view and database connection information. Run Time. Imports all information contained in the Execute view. System. Imports all information contained in the Setup view, except passwords for servers and database connections.
Table 62 (Cont.) DAC Tools Menu Commands Tools Menu Command DAC Repository Management > Create Repository Report Description Enables you to generate a DAC repository report based on the following criteria:
Table Row Counts Object References by Entity Ownerless Objects Unreferenced Objects Dead References
The Clean Up command removes unused referenced objects. DAC Repository > Upgrade/Merge Wizard See "Upgrading, Comparing and Merging DAC Repositories"
DAC Repository Enables you to purge completed runs from the run history. You Management > Purge Run can purge all runs (except the last run) or specify particular runs to Details be purged. The last run cannot be purged. In the Purging Runs... dialog, the following options are available:
All. Purges all completed runs except for the last run. By Execution Plan. Enables you to select an execution plan whose associated runs you want to purge. By Run Name. Enables you to select an individual run for purging. Before Specified Date. Enables you to select a date before which all runs except the last run will be purged. Details Only. Purges all related information about a run but leaves the run header information.
DAC Repository Management > Analyze Repository Tables DAC Repository Management > Default Index Properties
Enables you to run analyze table commands for all the DAC repository tables. Enables you to specify which databases will be associated with newly created indexes.
Enables you to access the Repository Audit Trail, which stores DAC Repository Management > Repository information about users, machine IP addresses, and repository timestamps. Audit Trail DAC Repository Enables you to drop all the DAC repository tables. This action Management > Drop DAC deletes all data in the repository. Repository DAC Server When the DAC Server is running an ETL process, this command Management > Get Server opens a text box that displays streaming data related to the Log process. DAC Server Management > DAC Server Setup ETL Management > Configure ETL Management > Reset Data Sources Seed Data > Task Phases Enables you to configure the DAC Server connections and server email settings. This action should be performed on the machine where the DAC Server is running. Opens the Data Warehouse Configuration wizard, which enables you to create and drop data warehouse tables and to create delete triggers. Clears the refresh dates for all source and target tables. This action forces a full load to occur. Enables you to add, edit, or delete task phases.
6-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Table 62 (Cont.) DAC Tools Menu Commands Tools Menu Command Seed Data > Task Logical Folders Description Enables you to add, edit, or delete Task Logical folders.
Seed Data > Task Physical Enables you to add, edit, or delete Task Physical folders. Folders Seed Data > Logical Data Sources UI Styles > Windows (MFC) Seed Data > Actions > Index Enables you to add, edit, or delete logical data sources. Changes the user interface to the Windows style. Enables you to set up index actions in order to trigger SQL scripts to create or drop indexes.
Execute
DAC Top Pane Toolbar Description Creates a placeholder for a new record in the selected list. Saves the current record. Undoes changes made to the current record after the last save. Deletes the selected record. If you delete a parent record, the child records are also deleted. When you delete a column from a table, the column is not automatically deleted from the index. The DAC does not display deleted objects. You must look at the database to figure out what objects were deleted.
Opens a blank query. Retrieves the data from the repository with the last used query. Design view only. Opens the Reference dialog, which enables you to copy objects from one container to another. For more information about referencing objects, see "About Object Ownership in DAC". Design view only. Assembles a subject area, with dimension and related tables as well as tasks. Design view only. Enables you to filter the source system container objects that appear in the top pane list. Execute view, Execution Plans tab only. Starts a new ETL process. Execute view, Current Run and Run History tabs only. Restarts the selected ETL, after the ETL has failed, stopped, or been aborted. Execute view, Current Run and Run History tabs only. Stops an ETL in progress. All currently running tasks will complete, and queued tasks will stop. The status of the ETL changes to Stopped. Execute view, Current Run and Run History tabs only. Causes an ETL in progress to abort. All currently running tasks will be aborted. The status of queued tasks and the ETL itself will change to Stopped. Execute view, Current Run tab only. Enables you to turn on and off the automatic screen refresh functionality and set the refresh interval.
Abort
Auto Refresh
Common Right-Click Menu Commands Design View Right-Click Menu Commands Setup View Right-Click Menu Commands Execute View Right-Click Menu Commands
6-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Table 65 (Cont.) Common Right-Click Menu Commands Command Paste String Copy Record Description Pastes a string from the clipboard into a selected cell that supports a string data type. Creates a copy of the selected record, with a unique record ID. The new record is committed to the database when you click the Save button or click outside the cell. In the Design view tabs (except for the Indices tab), Copy Record copies the selected record and the record's child records. When you copy a subject area, the tables are also copied but the tasks are not copied. You need to use the Assemble command to reassemble the subject area and add tasks to it. In the Design view Indices tab and Setup and Execute views, Copy Record copies only the selected record. Delete Deletes the selected record. If you delete a parent record, the child records are also deleted. When you delete a column from a table, the column is not automatically deleted from the index. You must manually delete columns from indexes that were deleted from a table or else the ETL process will fail. The DAC does not display deleted objects. You must look at the database to figure out what objects were deleted. Output to File Record Info Outputs to a text file in the DAC root folder the contents of the current tab's record list. Displays the record's unique ID, object type, current source system, owner source system, and the timestamp for when it was last updated. It also displays the source system lineage and the source systems that reference the object.
Update Records For some columns, enables you to update the column value for each row to a single value.
Reference. Opens the Reference dialog, which enables you to reference objects from one container to another. The reference function works like a symbolic link or shortcut. Re-Reference. If an object is a referenced object, that is, a reference to an object in another container and a change is made to the original object's child objects, you can use this command to import the changes to the referenced object. Push to References. If an original object is changed, you can use this command to export the changes to all referenced objects' child objects. De-Clone. When you make changes to a referenced object, the new object is called a clone. This command enables you to revert a cloned object back to its state as a reference. Re-Assign Record. This command enables you to reassign an objects ownership.
For more information about the ownership of objects, see "About Object Ownership in DAC". Assemble (Subject Areas tab) Assembles a subject area, with dimension and related tables as well as tasks.
Table 66 (Cont.) Design View Right-Click Menu Commands Command Generate Index Scripts Description (Tables and Indices tabs) Generates drop index, create index, and analyze table scripts for all tables that participate in the ETL process. The results are stored in the log\scripts directory.
Change Capture (Tables tab) For Siebel sources only. Scripts Image and Trigger Scripts. Generates change capture scripts for tables with defined image suffixes. The scripts may include delete triggers, create and drop statements for delete triggers, and image tables and their indexes.
View Scripts. Generates change capture view scripts for full or incremental mode for tables that participate in the change capture process. This command can be used for unit testing. Change Capture SQL. Generates change capture SQL scripts for full or incremental mode for tables that participate in the change capture process. This command can be used for unit testing.
(Tables tab)
Import Database Tables. Enables you to import table definitions from a selected database. This action does not import columns. Import Indices. Enables you to import index definitions from a selected database for one or more tables as listed in the result of the query. Import Database Columns. Enables you to import column definitions from a selected database.
Add Actions Add Actions Output Task Description Synchronize Tasks Flat Views
(Tables tab) See "About Index, Table and Task Actions". (Indices tab) See "About Index, Table and Task Actions". (Tasks tab) Saves to an HTML file the description for a selected task or for all tasks. (Tasks tab) Synchronizes the information the DAC has for a task's source and target tables with the information in the Informatica repository. Opens a dialog that enables you to query for various objects, modify data, and do mass updates. You can query for the following objects: Tables tab:
Task Source Tables Task Target Tables Task Conditional Tables Task Phase Dependencies Task Parameters Audit Trail
6-8 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Mark as Completed
Get Run Information > Get Log File Get Run Information > Analyze Run Get Run Information > Get Chart Get Run Information > Get Graph Auto Refresh
Figure 64. In addition, clicking on the icon when there is a connection to the server opens a text box that displays data related to the ETL process.
Figure 62 DAC Server Down Icon
6-10 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Edit the data in place and save the record by either clicking another record in the list or clicking the Save button. Reorder the columns. Sort the data in the list by clicking on the column name. Select predefined values from picklists. For fields that refer to values from other entities, use the query functionality in pop-up dialogs. Use Ctrl+C to copy an editable string to the clipboard (not available for read-only strings). Ctrl+V to paste a string from the clipboard into a selected cell that supports a string data type.
DAC Query Commands and Operators Common DAC Query Procedures Common DAC Query Procedures Using Flat Views Querying
() NULL AND OR
Extract* lists all tasks whose name starts with Extract. *Extract* lists all tasks whose name contains the word Extract. !Extract* lists all tasks whose name does not start with the word Extract. !null lists all tasks whose name is not null. Extract* or Aggregate* lists all tasks whose name starts with Extract or Aggregate. Load* and *Aggregate* lists all tasks whose name starts with Load and also contains the word Aggregate.
6-12 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
"Extract for Wave Dimension" or "Load into Wave Dimension" lists tasks whose name is either Extract for Wave Dimension or Load into Wave Dimension.
Note:
When using spaces within strings, you need to surround the string with quotes ("").
In the top or bottom pane of the DAC, click Query on the toolbar or in right-click menu. A blank row in a list appears.
2. 3.
Enter the query criteria in the appropriate fields. Click Run Query on the toolbar. The query is executed and the records appear.
In the date field, click the calendar icon on the right side of the cell. The Date dialog appears.
2.
Enter the date and time for which you want to search, and select the appropriate query condition.
In the DAC, right-click in the Tables, Tasks or Indices tab. Select Flat Views, and then select a context-sensitive column on which you want to query. In the query dialog, enter search criteria, and click Go. In the query results dialog, right-click and select Update Records. In the Update Record Set dialog, select the column you want to update, and then click Set Value. Enter a value for the column. To update records that are referenced objects, select Update Referenced Records. If you select this check box, referenced objects as well as original and cloned objects will be updated. The referenced objects will become clones, and the ownership column for these records will be updated to reflect the new ownership.
If you do not select this check box, only the columns in records that are original or cloned objects (objects owned by the source system container) will be modified.
8. 9.
Click OK. Click Yes when asked if you want to proceed. An informational message tells you which records were updated.
6-14 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
7
7
Overview of Index, Table and Task Actions Defining a SQL Script for an Action Assigning an Action to a Repository Object Functions for Use with Actions Example of How to Use a DAC Source System Parameter in an Action
Index actions can override the default behavior for dropping and creating indexes by mode type (full load, incremental load, or both). The Index action type are Create Index and Drop Index. For example you can define an index action to create indexes for tasks with the commands defined for full loads, or incremental loads, or both types. Index actions override all other index properties.
Table actions can override the default behavior for truncating and analyzing tables by mode type. The Table action types are Truncate Table and Analyze Table. For example you can define a table action to truncate tables with the commands defined for full loads, or incremental loads, or both . Table actions override all other index properties.
Task actions can add new functionality based on various task behaviors. The following task action types are available:
Preceding Action Use this type to execute a SQL script before a task runs.
Success Action Use this type to execute a SQL script after a task runs successfully.
Failure Action Use this type to execute a SQL script if a task fails during its execution.
Use this type to execute a SQL script when a task that previously failed is restarted. You can also use the Actions Template feature to:
Combine SQL templates to do synchronized actions, such as create and analyze indexes. Combine object level properties with user-defined parameters in SQL statements and stored procedures.
To define an action and assign it to a repository object, you need to complete the following procedures:
From the Tools menu, select Seed Data, then select one of the following:
In the toolbar, click New. In the new record field, enter a name for the action, and then click Save. Double-click in the Value field. The Value dialog box appears.
5.
Flat view displays the SQL entries in a list format in their order of execution. Category view displays the entries by the categories SQL and Stored Procedure. You can reorder the entries in the tree by dragging and dropping them.
6. 7.
Click Add. In the new record field, enter or select the appropriate information.
Description Logical name of the SQL block. SQL or Stored procedure
Description Should be used only for SQL types (not stored procedures). Defines which database the SQL statement will run against. Possible values are: Source - SQL runs against the source connection defined for the task. Target - SQL runs against the source connection defined for the task. Both - SQL runs against both the source and target connection. Table Connection - SQL runs against the table-specific connection if a separate table connection is available.
Table Type
Specifies the table type against which the SQL will run. Possible values are: All Source - SQL runs against all source tables defined for the task. All Target - SQL runs against all target tables defined for the task. Source Lookup - SQL runs against all the source lookup tables defined for the task. Source Primary - SQL runs against all the source primary tables defined for the task. Source Auxiliary - SQL runs against all the source auxiliary tables defined for the task.
Specifies whether an execution should proceed if a given SQL block fails. Specifies how many retries are allowed. If the number is not positive, a default number of one (1) will be used.
Valid Database Platforms Specifies the valid database platforms against which the SQL will run. If this field is left empty, the SQL can be run against any database. 8.
In the lower-right side text box, enter a SQL statement. The SQL Statement tab to the left of the text box lists all the supported SQL functions and DAC source system parameters that you can use in constructing custom SQLs. Double-click a function or source system parameter to move it into the text box. For a description of the available functions, see "Functions for Use with Actions". The source systems parameters list contains the names of all source system parameters defined in the DAC Repository, with the prefix @DAC_. During runtime, the DAC Server resolves the source system parameter and replaces its name with the runtime value. For an example of how to use a source system parameter in a SQL statement, see "Example of How to Use a DAC Source System Parameter in an Action".
9.
Note:
You can add multiple SQL statements and stored procedures to a single action.
Repository Object".
In the Design view, navigate to one of the following tabs, depending on the object type for which you want to assign an action:
2. 3. 4. 5.
Select or query for the object for which you want to assign an action. With the appropriate object selected in the top window, select the Actions subtab. Click New in the subtab toolbar. In the new record field, do the following:
a.
Select an Action Type. For a description of the available Action Types, see the following: "Indices Tab: Actions Subtab", "Tables Tab: Actions Subtab", and "Tasks Tab: Actions Subtab".
b. c.
Select the Full, Incremental, or Both load type. Double-click the Action field to open the Choose Action dialog box, and select an action.
Note:
For instructions on defining an action, see "Defining a SQL Script for an Action".
d. e.
Click OK to close the Choose Action dialog box. Click Save in the subtab toolbar.
Table 71, " Functions for Index Actions" Table 72, " Functions for Table Actions"
Table 71 Function
getAdditionalColumns()
getAnalyzeStatement() getAnalyzeTableStatement() getBitMapString getClusteredString getCreateIndexStatement() getDBType() getDropIndexStatement() getHashString() getImageSuffix() getIndexColumns() getIndexName() getIndexTableSpace() getNamedSource() getRvrsScanString() getTableName() getTableOwner() getTableSpace() getTruncateTableStatement() getUniqueColumns()
getUniqueString()
Table 72 Function
Functions for Table Actions Description Returns the default DAC table Analyze statement. Returns the physical data source connection type (Oracle OCI8, Oracle Thin, DB2, DB2-390, MSSQL, or Teradata).
getAnalyzeTableStatement() getDBType()
Table 72 (Cont.) Functions for Table Actions Function getImageSuffix() getCreateIndexStatement() getNamedSource() getDropIndexStatement() getTableName() getTableOwnerName() getTableSpace() getTruncateTableStatement() Description Returns the table image suffix if one exists; otherwise the string is empty. Returns the default DAC index creation statement. Returns the DAC physical connection name. Returns the default DAC index drop statement. Returns the table name. Returns the table owner. Returns the table space name if one exists; otherwise the string is empty. Returns the default DAC table truncate statement.
Table 73 Function
Functions for Task Actions Description Returns the default DAC analyze table statement. Returns the physical data source connection type (Oracle OCI8, Oracle Thin, DB2, DB2-390, MSSQL, or Teradata). Returns the table image suffix if one exists; otherwise the string is empty. Returns the physical connection name in DAC. Returns the table name. Returns the table owner. Returns the table space name if one exists; otherwise the string is empty. Returns the default DAC truncate table statement.
Note:
Table-related task action functions should be used only for SQL blocks with a specified table type. For these blocks, DAC will loop through the tables of the type you specify and execute custom SQL for each table. Functions will be substituted with table-specific values for each iteration. For example, if you wanted to gather statistics in a particular way after creating specific indexes, you would need to create index actions with two SQL blocks:
1. 2.
getCreateIndexStatement() DBMS_STATS.GATHER_TABLE_STATS(ownname => getTAbleOwner(), tabname => getTableName(), estimate_percent => 70, method_opt +> FOR ALL COLUMNS SIZE AUTO, cascade => false)
8
Customizing DAC Objects and Designing Subject Areas
This chapter provides information about customizing, designing, executing, and monitoring ETL processes. This section includes the following topics:
Creating or Copying a Source System Container About Customizing the Data Warehouse Adding a New Table and Columns to the Data Warehouse Adding an Index to the Data Warehouse Importing New Data Warehouse Objects into the Informatica Repository Creating Informatica Mappings and Workflows Creating Tasks in the DAC for New or Modified Informatica Workflows Setting a Task Phase Dependency Creating a Task Group About Parameter Management Specifying Tablespaces for Indexes by Table Type Working with Configuration Tags Overview of Subject Areas Creating a Subject Area
In DAC menu bar, select File, then select New Source System Container. Enter an ID and a name for the container.
Customizing DAC Objects and Designing Subject Areas 8-1
The ID and Name fields are alphanumeric. The Name can contain spaces but the ID cannot. The Name field must be at least five characters long.
3.
Create Empty New Source System Container Create as a Copy of Source System Container
4. 5.
If you are creating an empty, new container, click OK. If you are making a copy of an existing container, select the existing container from the drop-down list, and then click OK.
8-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Tables. Create the new table.
a. b.
In the Tables tab, click New. In the Edit child tab, enter the appropriate information about the table, and click Save.
Customizing DAC Objects and Designing Subject Areas 8-3
In the Columns child tab, click New. Enter the appropriate column information for each column you want to add to the table, and click Save. Enter the appropriate foreign key table and column information.
Note:
For performance purposes, it is recommended that you do not enter more than 254 columns to a dimension of fact table.
5.
Create the new tables and columns in the data warehouse database.
a. b. c. d.
Select Tools, then select ETL Management, then select Configure. Select the appropriate Source and Target database platforms, and then click OK. In the Data Warehouse Configuration Wizard, select Create Data Warehouse Tables, and then click Next. Enter the required information, and then click Start. An informational message reports whether the process was successful. For information about the process, you can review the createwtables.log file in the OracleBI\DAC\log\config folder.
To add a new table and columns using the DAC's Import command
1. 2. 3. 4.
Add the new table and column definitions into the data warehouse database. In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Tables. Import the new table definition.
a. b. c.
Right-click and select Import from Database, then select Import Database Tables. In the Import Tables dialog, select DataWarehouse. Optionally, enter filter criteria to identity the table name you entered in Step 1. See "DAC Query Commands and Operators" for available filter commands and operators.
d. e. f.
Click Read Tables. In the list of tables displayed, select the Import check box for the tables you want to import. Click Import Tables. An informational message indicates whether the process was successful.
5.
In the Tables tab, query for the table you imported in Step 4. With the table highlighted, right-click and select Import from Database, then select Import Database Columns.
8-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
c. d.
In the Importing Columns... dialog, select Selected Record Only, and then click OK. In the Import Table Columns dialog, click Read Columns. The Changes column displays a description of column changes, which are explained below:
Change The object was added to the database. The object was added to the repository. The object was modified.
Explanation The column is in the database but not the DAC repository. Importing it will add the column to the DAC repository. The column is in the DAC repository but not in the database. Importing it will delete it from the DAC repository. The column definition in the database does not match the definition in the DAC repository.
e. f.
In the list of columns displayed, select the Import check box for the columns you want to import. Click Import Columns. An informational message indicates whether the process was successful.
Add the new index definition into the data warehouse database. In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Tables. Query for the table for which you want to import index definitions. Right-click and select Import from Database, then select Import Indices. Choose to import indexes for a selected table or for all the records retrieved in your query, and click OK. In the Import Indices dialog, select DataWarehouse from the Data Sources drop-down list. Click Read Indices.
a. b.
In the list of indexes displayed, select the Import check box for the indexes you want to import. Click Import Indices. An informational message indicates whether the process was successful.
warehouse, see Oracle Business Intelligence Applications Fusion Edition Installation and Configuration Guide.
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. In the DAC, create custom logical and physical task folders for the custom folder you created in the Informatica repository.
a. b. c. d. e. f. g.
In the DAC, navigate to Tools, then select Seed Data, then select Task Folders. To create a custom logical folder, click New. In the Name field, enter a name for the custom logical folder, for example, Custom Logical. In the Type field, select Logical. To create a custom physical folder, click New. In the Name field, enter a name for the custom physical folder, for example, Custom Physical. In the Type field, select Physical.
3.
Register the folders you created in Step 2 in the Source System Folders tab.
a. b. c. d.
Navigate to Design, then select Source System Folders. Click New. In the Edit child tab, enter the name of the custom logical folder in the Logical Folder field. Enter the name of the custom physical folder in the Physical Folder field, and click Save.
4.
Navigate to Design, then select Tasks, and click New in the top pane toolbar. In the Edit child tab, enter the workflow name as it appears in Informatica Workflow Manager. Right-click and select Synchronize Tasks. Select Selected Record Only, and click OK. Click OK in the informational message box. This command synchronizes the source and target table information between the DAC and Informatica.
8-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
e.
In the Tasks tab, enter the remaining information required for the task. For a description of the fields in this tab, see "Tasks Tab".
The new table is now ready to be associated with a subject area. For information about creating a subject area, see "Creating a Subject Area".
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Tasks. Query for the task for which you want to add a phase dependency, and make sure it is highlighted. Click the Phase Dependency child tab. Click Add/Remove. In the Choose Phases dialog, select a phase and click Add. Click OK in the message box that states the phase was added. Select Phase, Grain, and Scope and click OK. For information about these fields, see "Tasks Tab: Phase Dependency Subtab". The task phase dependency appears in the Phase Dependency child tab.
9.
In the Subject Areas tab, query for the appropriate subject area. Click Assemble.
Navigate to the Execution Plans tab in the Execute view. Query for the appropriate execution plan. In the Parameters child tab, click Generate. In the top pane toolbar, click Build.
In the DAC, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Task Groups. Create a new task group.
a. b.
Click New in the top pane toolbar. In the Edit child tab, enter a name and select the appropriate properties.
4. 5. 6. 7. 8.
Click the Child Tasks child tab, and click Add/Remove in the toolbar. In the left-hand window of the Choose Child Tasks dialog, query for the tasks you want to add to the task group. Select the tasks, and click Add. In the right-hand window, enter an execution order. Click Save, and then click OK to close the window.
Overview of Parameters Preconfigured Parameters How DAC Handles Parameters at Runtime Nesting Parameters within Other Parameters Defining a Text Type Parameter Defining a Database Specific Text Type Parameter Defining a Timestamp Type Parameter Defining a SQL Type Parameter
Overview of Parameters
The ETL logic in Oracle Business Intelligence Applications uses parameters in the Informatica mappings and sessions. You define and manage parameters using the DAC parameter management feature. A parameter can apply to all tasks under a source system container (referred to as a source system parameter) or it can apply to a particular task (referred to as a task level parameter). Parameters set up at the task level have priority over parameters set up at the source system level. In the DAC, there are two types of parameters: static and runtime. The value of static parameters remains constant for all ETL runs. Examples of static parameters include language codes and currencies. The value of runtime parameters is dynamic, and the DAC updates this value for each ETL run. Examples of dynamic parameters include last refresh dates and last WID sequence numbers.
8-8 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Text The value for the parameter is defined as text. You can use the Text data type for both static and runtime parameters. DB Specific Text The value for the parameter is defined as database-specific text. This parameter should be used only if you have a heterogeneous database environment, and the parameter value needs to be different for the different database types. The DAC evaluates the string based on the source or target database type. If you do not specify database-specific text, the DAC returns the default value. Timestamp The value for the parameter is defined as a timestamp. You can use the Timestamp data type for both static and runtime parameters. A static timestamp can be any time that is constant. A runtime timestamp parameter is a variable for which the value is supplied by the DAC at runtime. You can define the timestamp in one of multiple formats or define a custom format. You can also use SQL to fetch any value that can be fired against the specified logical database connection. The DAC executes the SQL against a data source that maps to the specified logical connection and then formats the resulting value in the specified format. A SQL specified for a given timestamp parameter can include nested DAC parameters. For information about nested parameters, see "Nesting Parameters within Other Parameters". SQL The DAC fetches the value for the parameter from a database using SQL.
Preconfigured Parameters
Oracle Business Intelligence Applications ships with preconfigured parameters. Some of these preconfigured parameters are held in text files named parameterfileDW.txt and parameterfileOLTP.txt, which are stored in the folder \OracleBI\DAC\Informatica\parameters\input. Other preconfigured parameters are held in the DAC. The parameters held in the DAC are specific to the different source system containers. You can add new parameters in the DAC or change the existing parameters held in the DAC. However, Oracle recommends that you do not change the parameters held in the parameter text files. You can override the parameters in the text files by creating new parameters in the DAC. If you do make changes to the parameter text files, however, make sure the files remain in the folder \OracleBI\DAC\Informatica\parameters\input.
Note:
The Informatica Server must be configured to read parameter files from the location specified in the DAC system property InformaticaParameterFileLocation. For instructions on setting this property, see Oracle Business Intelligence Applications Fusion Edition Installation and Configuration Guide.
Avoid circular nesting, such as parameter A nested within parameter B, which is nested within parameter A. In such situations, the DAC randomly picks one of the parameters in the circle and evaluates it as an empty string.
To define a source system parameter, from the Views menu, select Design, and then select Source System Parameters. To define a task level parameter, from the Views menu, select Design, then select Tasks, and then click the Parameters subtab.
2. 3. 4. 5. 6.
Click New. Enter a parameter name. Select the Text data type. Click in the Value field to open the Enter Parameter Value dialog. Select one of the following options:
Static. This option specifies a value that remains constant for all ETL runs. Runtime. This option specifies a value will be updated by the DAC before each ETL run.
7. 8.
If you selected the Static option, enter a text value in the text window, and click OK. If you selected the Runtime option, select a DAC Variable from the list, and click OK.
8-10 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
9.
To define a source system parameter, from the Views menu, select Design, and then select Source System Parameters. To define a task level parameter, from the Views menu, select Design, then select Tasks, and then click the Parameters subtab.
2. 3. 4. 5. 6.
Click New. Enter a parameter name. Select the DB Specific Text data type. Click in the Value field to open the Enter Parameter Value dialog. Select one of the following Connection Type options:
@DAC_SOURCE_DBTYPE. This option specifies a source database connection. @DAC_TARGET_DBTYPE. This option specifies a target database connection.
7.
Click in the Default field to open the Default text box. Enter the parameter definition, and click OK.
8.
Click in the appropriate database type field to open the text box. Enter the parameter definition, and click OK.
9.
10. (Optional) To inactivate the parameter, select Inactive. 11. Click Save.
To define a source system parameter, from the Views menu, select Design, and then select Source System Parameters. To define a task level parameter, from the Views menu, select Design, then select Tasks, and then click the Parameters subtab.
2.
Click New.
3. 4. 5. 6.
Enter a parameter name. Select the Timestamp data type. Click in the Value field to open the Enter Parameter Value dialog. Select one of the following options:
Static. This option specifies a value that remains constant for all ETL runs. Runtime. This option specifies the value will be updated by the DAC before each ETL run. SQL. This option
7.
Click in the Date field to open the Date dialog. Enter a data and time, click OK.
8.
Click in the Value field to open the Enter Parameter Value dialog. Select a Variable from the list. From the Function list, select a format to which the DAC will convert the date. If you select Custom, enter a custom date format. If you select SQL Syntax or SQL Syntax (Date Only), select a Connection Type.
9.
Click in the SQL field to open the Enter Parameter Value dialog. Select a Logical Data Source from the list. Enter the parameter definition and click OK.
10. Click OK to close the Enter Parameter Value dialog. 11. (Optional) To inactivate the parameter, select Inactive. 12. Click Save.
To define a source system parameter, from the Views menu, select Design, and then select Source System Parameters. To define a task level parameter, from the Views menu, select Design, then select Tasks, and then click the Parameters subtab.
2. 3. 4. 5. 6.
Click New. Enter a parameter name. Select the SQL data type. Click in the Value field to open the Enter Parameter Value dialog. Select a Logical Data Source.
8-12 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
7. 8. 9.
Enter a SQL statement, and click OK. (Optional) To inactivate the parameter, select Inactive. Click Save.
You must create the tablespace in the database before you can specify tablespaces for indexes.
In the DAC toolbar, click Setup, then click the Physical Data Sources tab, then click the Index Spaces subtab. In the Index Spaces toolbar, click Generate. A list of available table types appears in the Table Type list.
3.
In the Index Space column, enter an index space for each table type, and click Save. If the Index Space property is left empty for a table type, the default index space for the default database connection will be used. The default index space is specified on the Edit subtab of the Physical Data Sources tab in the Setup view. If the Default Index Space property is also empty, the default tablespace assigned for the table owner will be used.
Remove tasks from all subject areas If you assign a task to a configuration tag, the task will not be eligible to participate in any subject area. For instructions, see "To remove tasks from all subject areas".
Reassign autogenerated tasks to a specific subject area An autogenerated task is a task that the DAC automatically assigns to a subject area when the subject area is assembled. For autogenerated tasks that were removed from participating in a subject area, you can set up the configuration tag to reassign a task to participate in specific subject areas. You do this by associating the configuration tag with the desired subject area. This method only applies to tasks that are autogenerated tasks of a subject area. For instructions, see "To reassign autogenerated tasks to a subject area".
You can set up a configuration tag to add non-autogenerated tasks to a subject area. The non-autogenerated tasks will participate in the subject area along with the subject area's autogenerated tasks. For instructions, see "To add non-autogenerated tasks to a subject area".
Assign only configuration tag tasks to a subject area (excludes the subject area's autogenerated tasks) You can also set up a configuration tag so that only tasks that were assigned to the configuration tag participate in a specific subject area. In this case, the subject area's autogenerated tasks do not participate. For instructions, see "To assign only configuration tag tasks to a subject area (excludes the subject area's autogenerated tasks)"
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Configuration Tags. Create a new configuration tag.
a. b. c. d.
Click New in the top pane toolbar. In the Edit child tab, enter a name. Make sure the Include Tasks check box is not selected. Click Save.
4.
With the new configuration tag highlighted in the top pane, click the Tasks child tab. In the bottom pane toolbar, click Add/Remove. In the Tasks dialog, query for the tasks you want to add to the configuration tag. Highlight the tasks, and then click Add. The tasks appear in the right-hand window.
e.
Click Save, and then click OK to close the window. These tasks will not be eligible to participate in any subject area.
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then Design, then select Configuration Tags. Query for the configuration tag that contains the tasks you want to reassign to a subject area. Verify the configuration tag contains the appropriate tasks by clicking the Tasks child tab and reviewing the list of tasks associated with this configuration tag.
Note:
Only a subject areas autogenerated tasks will be reassigned. If non-autogenerated tasks appear in the list, the DAC will ignore them.
8-14 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
5.
Associate the configuration tag with the subject areas to which you want to reassign the tasks.
a. b. c. d. e.
With the configuration tag highlighted in the top pane, click the Subject Areas child tab. Click Add/Remove in the bottom pane toolbar. In the Subject Areas dialog, query for one or more subject areas to which you want to reassign the task or tasks. Highlight the appropriate subject areas, and click Add. Click Save, and then click OK to close the window.
6.
In the Subject Area tab, query for all the subjects areas you added to the configuration tag. Highlight the subject areas, and click Reassemble.
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Configuration Tags. Create a new configuration tag.
a. b. c. d.
Click New in the top pane toolbar. In the Edit child tab, enter a name. Select the Include Tasks check box. Click Save.
4.
With the new configuration tag highlighted in the top pane, click the Tasks child tab. In the bottom pane toolbar, click Add/Remove. In the Tasks dialog, query for the extraneous tasks you want to add to the configuration tag. Highlight the tasks, and then click Add. Click Save, and then click OK to close the window.
5.
Associate the configuration tag with the subject areas to which you want to add the non-autogenerated tasks.
a. b. c. d. e.
With the configuration tag highlighted in the top pane, click the Subject Areas child tab. Click Add/Remove in the bottom pane toolbar. In the Subject Areas dialog, query for one or more subject areas to which you want to add the non-autogenerated tasks. Highlight the appropriate subject areas, and click Add. Click Save, and then click OK to close the window.
6.
a. b.
In the Subject Area tab, query for all the subjects areas you added to the configuration tag. Highlight the subject areas, and click Reassemble.
To assign only configuration tag tasks to a subject area (excludes the subject area's autogenerated tasks)
1. 2. 3.
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Subject Areas. Query for the subject area to which you want to add configuration tag tasks.
Note:
4. 5.
Select the Configuration Tag Tasks Only check box, and click Save. Create a configuration tag.
a. b. c. d. e.
Navigate to the Configuration Tags tab. Click New in the top pane toolbar. In the Edit child tab, enter a name. Select the Include Tasks check box. Click Save.
6.
With the new configuration tag highlighted in the top pane, click the Tasks child tab. In the bottom pane toolbar, click Edit. In the Tasks dialog, query for the tasks you want to add to the configuration tag. Highlight the tasks, and then click Add. Click Save, and then click OK to close the window.
7.
With the configuration tag highlighted in the top pane, click the Subject Areas child tab. Click Add/Remove in the bottom pane toolbar. In the Subject Areas dialog, query for the appropriate subject area. Highlight the subject area, and click Add. Click Save, and then click OK to close the window.
8.
In the Subject Area tab, query for all the subjects areas you added to the configuration tag. Highlight the subject areas, and click Reassemble.
8-16 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
To change a preconfigured subject area or to create a new subject area, you must first make a copy of an existing source system container or create a new container. For instructions, see "Creating or Copying a Source System Container".
Tables. Which tables need to be populated for the data warehouse? From which tables does your organization source data? What tables will create the star schemas. Subject areas. Do the subject areas cover all the relevant tables? Tasks. Are the tasks that load this table defined? Indexes. Do the target tables have the correct indexes defined?
Initial selection of tables. Find all the fact tables that belong to the subject areas.
2.
Recursive selection of related tables. Recursively find all the tables directly related through foreign keys and all other logically related tables.
3.
Initial selection of tasks. Find all the tasks that load into the tables selected above, that is, tasks whose target tables are one of the tables identified above. Tasks that DAC automatically assigns to a subject area are indicated with the Autogenerated flag (in the Tasks subtab of the Subject Areas tab). You can inactivate a task from participating in a subject area by selecting the Inactive check box (in the Tasks subtab of the Subject Areas tab). When the
Customizing DAC Objects and Designing Subject Areas 8-17
Inactive check box is selected, the task remains inactive even if you reassemble the subject area. You can also remove a task from a subject area using the Add/Remove command in the Tasks subtab of the subject Areas tab, but when you remove a task it is only removed from the subject area until you reassemble the subject area.
4.
Recursive selection of all tasks. Depending on the source and target table relationships, recursively figure out the prerequisite tasks.
In DAC, select the appropriate source system container from the drop-down list in the toolbar.
Note:
You cannot modify objects in the preconfigured source system containers. You must make a copy of a preconfigured container in order to make any changes to it. For instructions on how to make a copy of a preconfigured source system container, see "Creating or Copying a Source System Container".
2. 3. 4. 5. 6.
From the Menu bar, select Views, then select Design, and then select Subject Areas. In the top pane toolbar, click New. In the Edit child tab, enter a name for the subject area, and click Save. Make sure the new subject area name is highlighted in the top pane, and click the Tables child tab. Click Add/Remove in the child tab toolbar. The Choose Tables dialog opens. The left-hand window lists all the tables held in the selected container.
7. 8.
Query for one or more fact tables. Select the fact table (use Shift+click to select more than one table), and click Add. The tables are added to the right-hand window, which represents the subject area.
9.
10. In the top pane toolbar, click Assemble. 11. In the Assembling... dialog, select Selected Record Only.
If you select the option All Records in the List, DAC will reassemble all the subject areas listed in the top pane. DAC assembles the selected subject area by determining what dimensions and other related tables are required.
8-18 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
The Subject Area Assembly dialog box appears, displaying on the left a tree view that shows the fact tables that belong to the subject area. You can expand the fact table node to view its related tables.
12. Click Calculate Task List to assemble the tasks needed to load the tables displayed
in the tree view. If configuration tags have been assigned to any of the tasks, an information message will appear. A list of tasks appears in the Task List tab on the right side of the window. Also, the Table List tab displays the tables included in the subject area.
13. Click Accept to complete the subject area assembly process. 14. (Optional) Prune the subject area to better suit your business needs by removing
In the tree view, remove one or more tables from the subject area by deselecting the check box to the left of the table name. Click Calculate List. Review the changes to the subject area: The Task List tab displays the new set of tasks required to load the pruned subject area. The Task Difference Report tab displays the tasks that were added and the tasks that were deleted during the last assembly process. The Table List tab displays the tables that are included in the pruned subject area as well as the tables that were excluded during the pruning process.
d. e.
To accept the changes made to the subject area, click Accept. Click OK in the message box stating the subject area was successfully assembled.
15. (Optional) Click the Tasks tab to view which tasks DAC has determined are
required for this subject area. Tasks that are automatically assigned to the subject area by DAC are indicated with the Autogenerated check mark. You can inactivate a task from participating in the subject area by selecting the Inactive check box. When the Inactive check box is selected, the task remains inactive even if you reassemble the subject area.
You can also remove a task from the subject area using the Add/Remove command, but when you remove a task it is only removed from the subject area until you reassemble the subject area.
8-20 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
9
9
Types of Execution Plans Common Extract and Load Scenarios Best Practices for Multi-Source Execution Plans How DAC Determines the Order of Task Execution within an Execution Plan Building and Running Single-Source and Multi-Source Execution Plans Building and Running Micro ETL Execution Plans Unit Testing Execution Plan Tasks Scheduling an Execution Plan About Refresh Dates Monitoring Execution Plan Processes
Homogeneous
Building, Running and Monitoring Execution Plans 9-1
This type of execution plan extracts data from multiple instances of the same source system. For example, a business might have an instance of Oracle EBS 11i in one location and time zone and another instance of Oracle EBS 11i in another location and time zone. In such cases, the timing of data extraction from the different instances can be staggered to meet your business requirements.
Heterogeneous This type of execution plan extracts data from one or more instances of dissimilar source systems. For example, a business might have an instance of Siebel 7.8 in one location, an instance of Oracle EBS 11i in another location, and a second instance of Oracle EBS 11i in yet a third location. You can also stagger the timing of data extraction when you use this type of execution plan.
For information about data extraction and loads for multi-source execution plans, see "Common Extract and Load Scenarios".
Delay
As shown in Figure 91, the Delay property is located in the Parameters subtab in the Execution Plans tab of the Execute view. You set this property to specify how many minutes an extract of a data source will be delayed after the first extract of a multiple source extract process started.
Figure 91 Multi-Source Delay Property
9-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Single Extract and Single Load Scenario Multiple Extract and Single Load Scenario Multiple Extract and Multiple Load Scenario
The multiple extract scenarios apply to both homogeneous and heterogeneous multi-source execution plan types.
3.
Truncate table. Drop indexes. Run Informatica mapping. Recreate indexes. Analyze Table.
9-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
drops the indices. The last task writing to the table from the data source with the highest priority creates the indexes. In a multiple extract scenario, the truncation process for a table without indexes is as follows:
1. 2. 3. 4.
Truncate table (first extract task). Run Informatica mapping (first extract task). Run Informatica mapping (second extract task). Analyze table.
Truncate table (first extract task). Drop indexes (first extract task). Run Informatica mapping (first extract task). Run Informatica mapping (second extract task). Recreate indexes (second extract task). Analyze table.
Both Indicates the blocking action is active for tasks that have the same source and target physical data source connections.
Source Indicates the blocking action is active for tasks that have the same source physical data source connection.
Target Indicates the blocking action is active for tasks that have the same target physical data source connection.
None Indicates the blocking action is active for all tasks regardless of the source and target physical data source connections.
Delay Property The Delay property is located in the Parameters subtab in the Execution Plans tab of the Execute view. You set this property to specify how many minutes an extract of a data source will be delayed after the first extract of a multiple source extract process started. Source Priority For multi-source execution plans, you must assign each physical data source a priority (in the Physical Data Sources tab of the Setup view). The priority ranks the data sources and specifies the order in which DAC will load the data from the different sources. This property ensures that tasks attempting to write to the same target table will not be in conflict. Truncate Table Behavior If a target table is shared across different sources, it will be truncated only once. The priority of the data source determines which of the extracts truncates the tables. The task reading from the data source with the highest priority truncates the tables and drops the indexes. The last task writing to the table from the data source with the highest priority creates the indexes. Truncate Options Truncate options should be the same across the source system containers. Task Groups Tasks should belong to the same task groups across the source system containers.
9-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
How DAC Determines the Order of Task Execution within an Execution Plan
Customized Tasks for Vertical Applications In the case of Vertical Applications in which a task could have been customized in a Vertical workflow folder, the task will pick up the Vertical task folder when the source and target connections are the same based on the task folder priority. DATASOURCE_NUM_ID All tables should contain the column DATASOURCE_NUM_ID.
Unique indexes should always include the column DATASOURCE_NUM_ID. All the extract mappings populate the DATASOURCE_NUM_ID column from the parameter file produced by DAC. All the load mappings extract the value of the DATASOURCE_NUM_ID column from the staging tables. Level 1 aggregate tables should always contain the column DATASOURCE_ NUM_ID.
Load Strategies You should use the same load strategy for loading a table across the source system containers; that is, you should always use full loads or always use incremental loads. Avoid using a combination of load types. Failure Restart Strategies The Upon Failure Restart option of the Task Actions feature can be useful to restart tasks upon failure. For information about using task actions, see "About Index, Table and Task Actions".
How DAC Determines the Order of Task Execution within an Execution Plan
An execution plan is a collection of subject areas and a unique collection of tasks. A task can have prerequisite tasks that need to be executed before its own execution. DAC determines the order of tasks based on the following considerations:
A task's source and target table The dependency algorithm first looks at a task's source and target table. For example, suppose table A is populated by task T1 by reading from table B, and table B is populated by task T2 by reading from table C. The algorithm would determine task T2 should be executed before T1. The dependency algorithm next considers the following:
Task phase An ETL process typically goes through several phases. An example of a typical order in which phases are executed is as follows:
1. 2. 3. 4. 5. 6.
Extract Dimension Extract Fact Load Dimension Load Fact and Load Hierarchy (executed in parallel) Load Aggregate tables Update Dimensions
Building, Running and Monitoring Execution Plans 9-7
A table's Truncate Always properties The order of execution based on Truncate Always properties is as follows:
1. 2. 3. 4.
DAC randomly organizes tasks that have the same property values. If some tasks need to be executed in a particular order, you can create a task group that allows you to specify an execution order.
Set database connections to the transactional and data warehouse databases (in the Physical Data Sources tab). Registered the Informatica PowerCenter Services and Integration Service (in the Informatica Servers tab).
Before you attempt to run a multi-source execution plan, you must first define the priority for each source. The priority specifies the order in which DAC will load the data from the different sources. For more information about the source priority property, see "Physical Data Sources Tab". To define a source priority
1. 2.
Navigate to the Setup view, and then select the Physical Data Sources tab. For each of the physical data sources that will participate in the multi-source execution plan, enter a numerical value in the Priority field. The lower the numerical value, the higher the priority. For example, if you enter a value of 1, data from this source will be loaded first.
Navigate to the Execute view, then select the Execution Plans tab. Create a new execution plan.
a. b.
In the top pane toolbar, click New. In the top pane window or in the Edit subtab, enter a name for the execution plan and other appropriate information. For a description of the fields in this tab, see "Execution Plans Tab".
c. 3.
Click Save.
Click the Subject Areas child tab. Click Add/Remove in the bottom pane toolbar. In the Choose Subject Areas dialog, select the appropriate source system container from the drop-down list. Query for the subject area you want to associate with the execution plan.
9-8 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
e.
Select the subject area and click Add. You can repeat this process to associate multiple subject areas from any available source system container with an execution plan.
f. 4.
Click the Parameters subtab, and then click Generate in the bottom pane toolbar. In the Generating Parameters dialog box, enter the number of copies for each source system container, and then click OK. Click OK in the informational message. DAC automatically generates the parameters required for each copy. Not all copies require all of the possible parameters.
d.
On the Parameters subtab, edit the parameters for each copy of the source system container as follows: For each data source type, select the appropriate value from the Value drop-down list.
Note:
For the data source type of FlatFileConnection, make sure you have copied all files into the directory specified in the DAC system property InformaticaParameterFileLocation.
For each Informatica SIL and SDE folder, select the appropriate value in the Value drop-down list.
e. f.
For each data source type, enter the appropriate name in the Value field. (Optional) If you are extracting data from more than one source system container and want to stagger the data extracts, in the Delay field for the appropriate data source, enter a value for the number of minutes you want to delay the extract.
5.
Click the Ordered Tasks child tab and verify the following:
a.
Click Details in the toolbar, and review each tasks predecessor and successor tasks to confirm tasks common to multiple sources are ordered in a manner consistent with the priority of the source connection. Confirm that load tasks appear only once even if there are multiple extracts for tables common to multiple sources. For tasks common to multiple sources, click Preview Run Details in the toolbar, and confirm the following: The first common task truncates the common target table and the following tasks do not. The first common task truncates the common target table and the following tasks do not. For instructions on unit testing a task, see "Unit Testing Execution Plan Tasks".
b. c.
6.
In the top pane of the Execution Plans tab, make sure the new execution plan is highlighted, and click Build.
7. 8.
In the Building... dialog box, select the option Selected Record Only, to build only the selected execution plan. To run the execution plan, select the execution plan in the top pane, and click Run Now. Once the ETL process starts running you can monitor its progress in the Current Run tab. For information about how refresh dates are tracked, see "About Refresh Dates". To schedule an execution plan, see "Scheduling an Execution Plan".
Navigate to the Execute view, then select the Execution Plans tab. Select the execution plan to which the task you want to test belongs. Click the Ordered Tasks child tab. Select a task from the list. On the toolbar, select Preview Run Details. In the Preview Run Details dialog, click Execute. DAC executes the task and displays the results in the Preview Run Details dialog.
7.
9-10 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Caution: Micro ETL processes can cause issues with data inconsistencies, data availability, and additional load on the transactional database. Therefore, you should consider the following factors before implementing a micro ETL process:
For related star schemas, if one schema is omitted from a micro ETL execution plan, the cross-star reports may be inaccurate. For example, if the Person fact table is refreshed more frequently than the Revenue fact table, a report that spans the Person and Revenue dimensional schemas may produce inconsistent results. If you omit dimension tables from a micro ETL execution plan, the foreign keys for the fact tables will point to Unspecified rows for the new dimension records. The foreign key references will be resolved when the Complete ETL execution plan is run, but users of the reports should be aware of such inconsistencies. If you do not include aggregate tables in micro ETL execution plans, the reports that use data from these tables will be inconsistent with the reports that use data from the detailed fact tables. However, if aggregate tables are included in the micro ETL execution plan, the aggregate calculations are performed for each ETL process, which will take a constant amount of time and may be inefficient to perform at such frequent intervals. Hierarchy tables are rebuilt during every ETL execution plan by querying the base dimension tables. This operation takes a constant amount of time. If the base tables are big, this operation may take a long time and may be inefficient if the micro ETL execution plan runs several times a day. However, if you avoid populating the hierarchy tables during micro ETL processes, data inconsistencies will occur. With micro ETL execution plans, caching will occur more frequently, which may have performance implications. Micro ETL execution plans will put more load on the transactional database because of the frequent extracts.
In the DAC toolbar, select the appropriate source system container from the drop-down list in the toolbar. From the Menu bar, select Views, then select Design, then select Subject Areas. In the Subject Areas tab, assemble a small subject area. In the Tasks child tab, inactivate all tasks that are not required for the execution plan. Create a new execution plan.
a. b. c. d.
Navigate to the Execute view, then select the Execution Plans tab. Enter a name for the execution plan Select the Keep Separate Refresh Dates check box. Click Save.
6.
Associate one or more subject areas with the execution plan. The subject areas can belong to one or more source systems.
a. b.
Click the Subject Areas child tab. Click Add/Remove in the bottom pane toolbar.
c. d. e.
In the Choose Subject Areas dialog, select the appropriate source system container. Query for the subject area you want to associate with the execution plan. Select the subject area and click Add. You can associate multiple subject areas with an execution plan, but all the subject areas must be from the same source system container.
f. 7.
Click the Parameters subtab, and then click Generate in the bottom pane toolbar. In the Generating Parameters dialog box, enter the number of copies for each source system container, and then click OK. Click OK in the informational message. DAC automatically generates the parameters required for each copy. Not all copies require all of the possible parameters.
d.
On the Parameters subtab, edit the parameters for each copy of the source system container as follows: For each data source type, select the appropriate value from the Value drop-down list.
Note:
For the data source type of FlatFileConnection, make sure you have copied all files into the directory specified in the DAC system property InformaticaParameterFileLocation.
For each Informatica SIL and SDE folder, select the appropriate value in the Value drop-down list.
e. f.
For each data source type, enter the appropriate name in the Value field. (Optional) If you are extracting data from more than one source system container and want to stagger the data extracts, in the Delay field for the appropriate data source, enter a value for the number of minutes you want to delay the extract.
8.
In the top pane of the Execution Plans tab, make sure the new execution plan is highlighted, and click Build. DAC builds the execution plan.
9.
Click the Ordered Tasks child tab and verify the following:
a.
Click Details in the toolbar, and review each tasks predecessor and successor tasks to confirm tasks common to multiple sources are ordered in a manner consistent with the priority of the source connection. Confirm that load tasks appear only once even if there are multiple extracts for tables common to multiple sources. For tasks common to multiple sources, click Preview Run Details in the toolbar, and confirm the following: The first common task truncates the common target table and the following tasks do not.
b. c.
9-12 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
The first common task truncates the common target table and the following tasks do not. The execution plan is now ready to run as a micro ETL execution plan.
10. Create a schedule for the micro ETL execution plan. For instructions, see
In DAC, navigate to the Scheduler tab. The current list of schedules appears in the top pane.
2.
Click New in the top pane toolbar. The Edit tab in the bottom pane becomes active.
3. 4. 5. 6. 7.
Enter a name for the schedule. Select an execution plan. If you want the schedule to run once, select the Run Only Once check box, and then select a start and end date and time. To create a periodic schedule, select a recurrence pattern, and enter the appropriate date and time parameters. Click Save.
Table 91 (Cont.) Refresh Date Scenarios Scenario 3 (See note below) 3 4 4 Table Type (in Tasks child tabs) Primary Source Primary Target Primary Source Primary Target Refresh Date Not Null Null Not Null Not Null Command DAC Truncate Will Use Target Table? Full Load Not applicable Incremental Load Not applicable Yes Not applicable No Not applicable
Scenario 2. When two or more source tables load into the same target table as separate tasks, the source table in the second task may have refresh date as null while the target may have a refresh date. Scenario 3. When a source loads into more than one target table in separate tasks, the refresh date may be null for the second target table while the source table may have refresh dates.
In DAC, navigate to the Current Run tab. Right-click and select Get Run Information. The following options are available:
Get log file Analyze run Get chart Get phase chart Get graph
9-14 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
10
10
Overview of Upgrade/Merge Wizard Overview of Upgrade and Merge Options About the Repository Upgrade (DAC 784) Option About the Refresh Base Option About the Simplified Refresh From Base Option About the Replace Base Option About the Peer to Peer Merge Option Resolving Object Differences in the View Difference Report
Major Stages of the Upgrade/Merge Wizard Resetting the Upgrade or Merge Process
The Upgrade/Merge Wizard enables you to suspend an upgrade or merge while in process and return to it at a later time. You can also repeat a stage that you just completed or reset the process to start from the beginning. Table 101 summarizes the possible actions you can take at each stage of the upgrade or merge process.
Table 101 Stage Possible Upgrade and Merge Actions Possible Actions
After you have completed You can perform one of the following actions: Stage 1: Import repository data Repeat Stage 1: Import repository data
Repeat Stage 2: Create Difference Report Perform Stage 3: View Difference Report Reset the process
After you have completed You can perform one of the following actions: Stage 3: View Difference Report Repeat Stage 3: View Difference Report
Open the Upgrade/Merge Wizard by selecting from the DAC toolbar Tools, then DAC Repository Management, and then Upgrade/Merge Wizard. The Upgrade/Merge Wizard dialog box displays the type of upgrade or merge process you previously started.
2. 3.
From the Perform drop-down list, select Reset and click OK. In the Reset Upgrade Process dialog box, re-type the text in the text box to confirm you want to proceed, and click Yes. An informational message tells you the reset process was successful.
10-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
4.
Click OK.
Repository Upgrade (DAC 784) Use this option to upgrade a DAC Repository in the release 7.8.4 format, which has a non-partitioned DAC Repository structure, to the new release, which has a partitioned DAC Repository structure. For more information, see "About the Repository Upgrade (DAC 784) Option".
Refresh Base Use this option to upgrade the DAC Repository when you are upgrading from an older release of Oracle BI Applications to a new release. For more information, see "About the Refresh Base Option".
Simplified Refresh From Base This option is similar to the Refresh Base option. It allows you to upgrade the DAC Repository from an older release of Oracle BI Applications to a new release without comparing repositories and creating a Difference Report. For more information, see "About the Simplified Refresh From Base Option".
Replace Base Use this option to upgrade the DAC Repository when you are phasing out an older release of a transactional application and moving to a newer release, for example, phasing out Siebel 7.5.3 and moving to Siebel 7.8. For more information, see "About the Replace Base Option".
Peer to Peer Merge Use this option to merge DAC repositories of different instances of the same release. For example, in a development environment you may have two instances of a DAC Repository used with Oracle BI Applications release 7.9.5 that you want to merge. For more information see, "About the Peer to Peer Merge Option".
Repository Upgrade (784): High-Level Process Flow Repository Upgrade (784): Procedure for Upgrading
DAC versions before Oracle BI Applications 7.9 had a non-partitioned DAC Repository structure and held metadata for a single application. DAC versions released with Oracle BI Applications 7.9 and higher and with the DAC 10.1.3.4 release have partitioned DAC Repository structures. The partitioned structure, also known as a container, can hold metadata for multiple applications. The Repository Upgrade (DAC 784) option enables you to upgrade from the DAC Repository release 7.8.4 (non-partitioned) to the new partitioned structure.
Note:
If you want to upgrade a pre-7.8.4 release of a DAC Repository, you must first upgrade the repository to the 7.8.4 release before you can use the Repository Upgrade (DAC 784) option to upgrade to the new release.
In Step 1 of the high-level process flow, you import the existing 7.8.4 DAC Repository into the repository temporary tables. The existing 7.8.4 repository is referred to as the source in the Upgrade/Merge Wizard. In Step 2, you create a Difference Report that compares the existing repository with the new base container, which is referred to as the target. (The new base container is the version to which you are upgrading and is the version you imported from the file system.) In Step 3, you accept or reject the objects that the Difference Report shows as being either present or changed in the source but not the target. See Section 10.8, "Resolving Object Differences in the View Difference Report" for a description of how objects are merged based on the action you take. In Step 4, after you have resolved the differences, you then execute the merge. In Step 5, the Upgrade/Merge Wizard references the customizations in the newly merged custom container with the new base container.
10-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Note:
You cannot use the Repository Upgrade (784) option with pre-7.8.4 releases of the DAC. To confirm your existing repository is in the release 7.8.4 format, open the 7.8.4 DAC client and in the toolbar, select Help, and then select About DAC. The About DAC dialog box displays the version of DAC you are running.
Before you begin this procedure, you need to have already installed the new release of Oracle BI Applications and imported the metadata that is the same version as the metadata in the existing 7.8.4 DAC Repository. For example, if the existing 7.8.4 repository contains Siebel 6.3 metadata, you need to have imported the Siebel 6.3 source system container metadata from the new Oracle BI Applications release. If you are using Oracle BI Applications 7.9.x with the DAC 7.9.x, then you should have the correct Siebel source system container in the 7.9.x DAC Repository. You should also review the section "Resolving Object Differences in the View Difference Report" to gain an understanding of your options for resolving object differences. To upgrade a DAC Repository in the release 7.8.4 format to the new release
1.
Navigate to the folder DAC\conf\upgrade Open the datamapping.xml file for editing. Edit the entries for folder, database connection, and subject area to reflect the upgrade environment. See the comments in the datamapping.xml file for instructions.
2.
Configure the connection between the release 7.8.4 DAC Repository and the new DAC client.
a. b. c.
In the Setup view, click Physical Data Sources. Click New to create a new record for the 7.8.4 DAC Repository. Enter the appropriate information to connect to the release 7.8.4 DAC Repository database. For information about the required fields, see the section "Physical Data Sources Tab".
Note:
The value for the Type field must be set to DAC Repository.
d. e. 3. 4.
Navigate to the Upgrade/Merge Wizard by selecting Tools, then DAC Repository Management, and then Upgrade/Merge Wizard. From the drop-down list, select Repository Upgrade (DAC 784), and then click OK. The Import 7.8.4 Repository dialog box appears.
5.
From the 784 Repository drop-down list, select the repository you configured in Step 1.
Note:
In order for a repository to appear in the drop-down list, it must be configured in the Physical Data Sources tab of the Setup view.
6.
From the Source System Container drop-down list, select the new base container to which you are upgrading. This is the container you will compare against the existing release 7.8.4 DAC Repository.
Caution: Make sure you select the appropriate container. If you select a container different from that contained in the existing release 7.8.4 DAC Repository, the Difference Report will compare dissimilar source system containers and will be inaccurate.
7.
Select the categories of metadata from the existing release 7.8.4 DAC Repository you want to import into the repository temporary tables for comparison with the new base container. The categories are as follows:
Description Imports all information contained in the DAC Design view and the execution plan information for the DAC Execute view. Imports ETL Run History and the last refresh date information. Imports all information contained in the DAC Setup view, except passwords for servers and database connections.
The release 7.8.4 DAC Repository tables are imported into the temporary tables.
8.
Click OK in the Importing Tables dialog box when the process is finished. The Create Difference dialog box appears.
9.
Create the Difference Report to compare the differences between the source and target DAC repositories.
a. b. c.
Enter a name for the Difference Report, or leave the default name. From the Source System Container drop-down list, select the new base container. In the New Custom Container ID/Name fields, enter an ID and a name for the custom container that will be created during the upgrade process. The ID and Name fields are alphanumeric. The Name field can contain spaces and must be at least five characters long. The ID field cannot contain spaces.
d. e.
(Optional) Enter a description for the Difference Report. Click OK. When the Difference Report is complete, the Creating Difference Report dialog box tells you how long the process took.
f.
Click OK. The View Difference Report dialog box displays the differences between the existing and new DAC repositories.
10. In the View Difference Report dialog box, resolve the differences between the
existing (source) and new (target) containers, that is, between the release 7.8.4
10-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
DAC Repository and the repository that contains the new base container. For detailed information about the View Difference Report, see "Resolving Object Differences in the View Difference Report". To resolve the differences, you either accept or reject the objects that appear in the release 7.8.4 DAC Repository but do not appear in the new base container to which you are upgrading.
a.
In the navigation tree, select the repository object for which you want to view the differences between the existing and new repositories. If the object selected in the navigation tree is a hierarchical object, the subtabs for the child objects appear in the bottom pane of the Object Difference window.
b. c.
(Optional) Filter the objects that appear in the Object Difference window by selecting one of the options from the drop-down list in the toolbar. For parent objects in the top pane and any child objects in the bottom pane, accept or reject the object in the difference list by selecting or deselecting the Accept Source check box. For detailed information about these options and the merge outcome, see "Possible Repository Merge Outcomes Based on Your Decisions".
Note:
If a child object has been changed but not the parent object, the parent object will still appear in the Object Difference window even though it has not been changed.
d.
(Optional) Once you have made a decision about whether to accept or reject the difference, select the Resolved check box to indicate you have resolved the object. Repeat Steps a, b, and c, until you have resolved all object differences. Click Merge. The Merge dialog box appears and lists the details of the merge.
e. f.
11. Click Merge to begin the merge process. 12. Click OK in the Merging Repositories dialog box when the merge process is
complete.
Refresh Base: High-Level Process Flow Refresh Base: Procedure for Upgrading
The Refresh Base option enables you to upgrade an existing customized DAC Repository. You should use this option if you are upgrading from a DAC release higher than 7.8.4. If you are upgrading a DAC Repository in the release 7.8.4 format or lower-numbered releases, see "About the Repository Upgrade (DAC 784) Option". The Refresh Base option enables you to compare a new base container with the existing customized repository and to create a Difference Report. If you want to upgrade an existing customized DAC Repository without comparing repositories and
creating a Difference Report, you can use the Simplified Refresh From Base option. See "About the Simplified Refresh From Base Option" for more information.
In Step 1 of the high-level process flow, you import the repository data for the new base container from the file system into the repository temporary tables. This repository is referred to as the source in the Upgrade/Merge Wizard. In Step 2, you create a Difference Report that compares the new base container with the existing repository (including customizations). The existing customized repository is referred to as the target. In Step 3, you accept or reject the objects that the Difference Report shows as being present or changed in the source but not the target. See Section 10.8.3, "Possible Repository Merge Outcomes Based on Your Decisions" for a description of how objects are merged based on the action you take. In Step 4, after you have resolved the differences, you then execute the merge. In Step 5, the DAC references the customizations in the newly merged repository with the new base container.
10-8 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Note:
The repository data is imported from a file system and should be in the DAC 10.1.3.4 format. If it is in the 7.9.x format, you should do the following:
Restore the repository data into a database using the 7.9.x DAC (using the regular DAC import process). Install and configure DAC 10.1.3.4 to the 7.9.x repository. Export the relevant source system container to a file folder. The metadata in the folder in Step 3 above will become the source for this upgrade.
1. 2. 3.
Navigate to the Upgrade/Merge Wizard by selecting Tools, then DAC Repository Management, and then Upgrade/Merge Wizard. From the drop-down list, select Refresh Base, and then click OK. The Import Source System Container dialog box appears.
3. 4. 5.
Click Change import/export folder to navigate to the directory that holds the metadata files for the new base container to which you are upgrading. Select the container to which you are upgrading from the Source System Container drop-down list, and click OK. In the Importing Tables dialog box, re-type the text in the text box to confirm you want to proceed, and click Yes. When the import process is complete, the Importing Tables dialog box tells you how long the process took.
6.
7.
Create the Difference Report to view the differences between the new and existing DAC repositories.
a. b. c. d.
Enter a name for the Difference Report, or leave the default name. Select the appropriate existing container. (Optional) Enter a description for the Difference Report. Click OK. When the Difference Report is complete, the Creating Difference Report dialog box tells you how long the process took.
e.
Click OK. The View Difference Report dialog box displays the differences between the new and existing DAC repositories.
8.
In the View Difference Report dialog box, resolve the differences between the new repository (source) and existing repository (target). For detailed information about the View Difference Report, see "Resolving Object Differences in the View Difference Report". To resolve the differences, you either accept or reject the objects that are listed as new or changed in the new repository (the version to which you are upgrading).
a.
In the navigation tree, select the repository object for which you want to view the differences between the new and existing repositories. If the object selected in the navigation tree is a hierarchical object, the subtabs for the child objects appear in the bottom pane of the Object Difference window.
b. c.
(Optional) Filter the objects that appear in the top, right window by selecting one of the options from the drop-down list in the toolbar. For parent objects in the top pane and any child objects in the bottom pane, accept or reject the object in the difference list by selecting or deselecting the Accept Source check box. For detailed information about these options and the merge outcome, see "Possible Repository Merge Outcomes Based on Your Decisions".
Note:
If a child object has been changed but not the parent object, the parent object will still appear in the Object Difference window even though it has not been changed.
d.
(Optional) Once you have made a decision about whether to accept or reject the difference, select the Resolved check box to indicate you have resolved the object. Repeat Steps a, b, and c, until you have resolved all object differences. Click Merge. The Merge dialog box appears and lists the details of the merge.
e. f.
9.
10. Click OK in the Merging Repositories dialog box when the merge process is
10-10 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Make sure you have renamed and backed up your existing DAC Repository into a different database. When you backup the DAC Repository, you export the DAC metadata, in XML format (using the DAC's Export tool), into a folder other than the standard DAC export folder where backups are stored (DAC\export). For instructions on exporting DAC metadata, see "Exporting DAC Metadata".
Upgrade your existing DAC Repository tables to be compatible with the new DAC Repository release.
a. b.
Configure the new DAC client version to read the DAC metadata from your existing DAC Repository. Log in to the DAC and select Yes if prompted to upgrade the repository tables.
2.
Export the custom applications from your existing custom repository to a folder other than the standard DAC export folder (DAC\export) or the folder into which you backed up metadata in the previous format. Import the new DAC metadata for your application from the standard DAC export folder (DAC\export). Select the Truncate Repository Tables check box. For instructions on importing DAC metadata, see "Importing DAC Metadata".
3.
4.
Import the customized DAC metadata that you exported in Step 2, and deselect the Truncate Repository Tables check box. This will append all the custom data to the repository, thus bringing in the customizations, which include the following:
All modified data. All newly created custom data. All deleted data.
5.
Refresh the source system container to locate any missing objects in your customized application. The missing objects are any new objects that the preconfigured applications may have that are not referenced in the custom applications.
a. b. c. d.
Navigate to the Upgrade/Merge Wizard by selecting Tools, then DAC Repository Management, and then Upgrade/Merge Wizard. From the drop-down list, select Simplified Refresh From Base, and then click OK. Select the appropriate container from the Source System Container drop-down list, and click OK. Confirm that you want to refresh the source system container.
6.
Rebuild all the subject areas and execution plans in your customized application to include any new changes in the object dependency. For information about building subject areas and execution plans, see "Customizing DAC Objects and Designing Subject Areas" and "Building, Running and Monitoring Execution Plans".
The Replace Base option enables you to upgrade the DAC Repository when you are phasing out an older release of a transactional application and moving to a newer release, for example, phasing out Siebel 7.5.3 and moving to Siebel 7.8.
In Step 1 of the high-level process flow, you import the repository data for the new base container into the repository temporary tables. This repository is referred to as the source in the Upgrade/Merge Wizard In Step 2, you create a Difference Report that compares the new base container (source repository) with the existing base container (including customizations). The existing base container is referred to as the target In Step 3, you accept or reject the objects that the Difference Report shows as being present in the source but not the target or changed in the source but not the target. See Section 10.8.3, "Possible Repository Merge Outcomes Based on Your Decisions" for a description of how objects are merged based on the action you take. In Step 4, after you have resolved the differences, you then execute the merge. In Step 5, the DAC references the customizations in the newly merged repository with the new base container.
10-12 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Note:
The repository data is imported from a file system and should be in the DAC 10.1.3.4 format. If it is in the 7.9.x format, you should do the following:
Restore the repository data into a database using the 7.9.x DAC (using the regular DAC import process). Install and configure DAC 10.1.3.4 to the 7.9.x repository. Export the relevant source system container to a file folder. The metadata in the folder in Step 3 above will become the source for this upgrade.
1. 2. 3.
Navigate to the Upgrade/Merge Wizard by selecting Tools, then DAC Repository Management, and then Upgrade/Merge Wizard. From the drop-down list, select Replace Base, and then click OK. The Import Source System Container dialog box appears.
3. 4. 5.
Click Change import/export folder to navigate to the directory that holds the metadata files for the new base container to which you are upgrading. Select the appropriate container from the Source System Container drop-down list, and click OK. In the Importing Tables dialog box, re-type the text in the text box to confirm you want to proceed, and click Yes. When the import process is complete, the Importing Tables dialog box tells you how long the process took.
6.
7.
Create the Difference Report to view the differences between the new and existing DAC repositories.
a. b. c. d.
Enter a name for the Difference Report, or leave the default name. Select the appropriate existing container. (Optional) Enter a description for the Difference Report. Click OK. When the Difference Report is complete, the Creating Difference Report dialog box tells you how long the process took.
e.
Click OK.
The View Difference Report dialog box displays the differences between the new and existing DAC repositories.
8.
In the View Difference Report dialog box, resolve the differences between the new repository (source) and existing repository (target). For detailed information about the View Difference Report, see "Resolving Object Differences in the View Difference Report". To resolve the differences, you either accept or reject the objects that are listed as new or changed in the new repository (the version to which you are upgrading).
a.
In the navigation tree, select the repository object for which you want to view the differences between the new and existing repositories. If the object selected in the navigation tree is a hierarchical object, the subtabs for the child objects appear in the bottom pane of the Object Difference window.
b. c.
(Optional) Filter the objects that appear in the top, right window by selecting one of the options from the drop-down list in the toolbar. For parent objects in the top pane and any child objects in the bottom pane, accept or reject the object in the difference list by selecting or deselecting the Accept Source check box. For detailed information about these options and the merge outcome, see "Possible Repository Merge Outcomes Based on Your Decisions".
Note:
If a child object has been changed but not the parent object, the parent object will still appear in the Object Difference window even though it has not been changed.
d.
(Optional) Once you have made a decision about whether to accept or reject the difference, select the Resolved check box to indicate you have resolved the object. Repeat Steps a, b, and c, until you have resolved all object differences. Click Merge. The Merge dialog box appears and lists the details of the merge.
e. f.
9.
10. Click OK in the Merging Repositories dialog box when the merge process is
Peer to Peer Merge: High-Level Process Flow Peer to Peer Merge: Procedure for Merging
The Peer to Peer Merge option enables you to merge DAC repositories of different instances of the same release. For example, in a development environment you may have two instances of a DAC Repository used with Oracle BI Applications release 7.9.5 that you want to merge.
10-14 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
In Step 1 of the high-level process flow, you export one instance of the existing customized source system container to the file system and then import this container into the repository temporary tables. This repository is referred to as the source in the Upgrade/Merge Wizard. In Step 2, you create a Difference Report that compares the instance 1 container (including customizations) with the instance 2 container (including customizations). The instance 2 container is referred to as the target. In Step 3, you accept or reject the objects that the Difference Report shows as being present or changed in the source but not the target. See Section 10.8.3, "Possible Repository Merge Outcomes Based on Your Decisions" for a description of how objects are merged based on the action you take. In Step 4, after you have resolved the differences, you then execute the merge. In Step 5, the Upgrade/Merge Wizard references the customizations in the newly merged container with the instance 1 base container.
Navigate to the Upgrade/Merge Wizard by selecting Tools, then DAC Repository Management, and then Upgrade/Merge Wizard. From the drop-down list, select Replace Base, and then click OK. The Import Source System Container dialog box appears.
3. 4. 5.
Click Change import/export folder to navigate to the directory that holds the metadata files for instance 1 of the source system container you want to merge. Select the appropriate container from the Source System Container drop-down list, and click OK. In the Importing Tables dialog box, re-type the text in the text box to confirm you want to proceed, and click Yes. When the import process is complete, the Importing Tables dialog box tells you how long the process took.
6.
7.
Create the Difference Report to view the differences between the instance 1 container and the instance 2 container.
a. b. c. d.
Enter a name for the Difference Report, or leave the default name. In the Existing Container drop-down list, select the instance 2 container. (Optional) Enter a description for the Difference Report. Click OK. When the Difference Report is complete, the Creating Difference Report dialog box tells you how long the process took.
e.
Click OK. The View Difference Report dialog box displays the differences between the instance 1 and instance 2 containers.
8.
In the View Difference Report dialog box, resolve the differences between the instance 1 and instance 2 DAC repositories. The instance 1 repository is referred to as the source or existing container, and instance 2 as the target or new container. For detailed information about the View Difference Report, see "Resolving Object Differences in the View Difference Report". To resolve the differences, you either accept or reject the objects that appear as new or changed in the instance 1 container but do not appear in the instance 2 container.
a.
In the navigation tree, select the repository object for which you want to view the differences between the instance 1 and instance 2 containers. If the object selected in the navigation tree is a hierarchical object, the subtabs for the child objects appear in the bottom pane of the Object Difference window.
b. c.
(Optional) Filter the objects that appear in the top, right window by selecting one of the options from the drop-down list in the toolbar. For parent objects in the top pane and any child objects in the bottom pane, accept or reject the object in the difference list by selecting or deselecting the Accept Source check box.
10-16 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
For detailed information about these options and the merge outcome, see "Possible Repository Merge Outcomes Based on Your Decisions".
Note:
If a child object has been changed but not the parent object, the parent object will still appear in the Object Difference window even though it has not been changed.
d.
(Optional) Once you have made a decision about whether to accept or reject the difference, select the Resolved check box to indicate you have resolved the object. Repeat Steps a, b, and c, until you have resolved all object differences. Click Merge. The Merge dialog box appears and lists the details of the merge.
e. f.
9.
10. Click OK in the Merging Repositories dialog box when the merge process is
Overview of View Difference Report View Difference Report Interface Possible Repository Merge Outcomes Based on Your Decisions
The Upgrade/Merge Wizard generates a View Difference Report for the following upgrade and merge options:
Repository Upgrade (DAC 784) Refresh Base Replace Base Peer to Peer Merge
A View Difference Report is not available if you select to upgrade using the Simplified Refresh From Base option. For more information about the upgrade and merge options, see "Overview of Upgrade and Merge Options".
Navigation Tree The navigation tree displays the repository object types that are compared in the existing and new repositories. When you select a repository object in the navigation tree, the differences are listed in the object difference window.
Object Difference Window The Object Difference window lists the differences found between the existing and new containers for the object type selected in the navigation tree. If the object selected in the navigation tree is a hierarchical object, the subtabs for the child objects appear below the Object Difference window. The following columns appear in the Object Difference window and in the subtabs: User Key Unique identifier for the object. Change Description Indicates the type of difference that exists between the containers. Container The new or target container. Accept Source
10-18 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Selecting the Accept Source check box indicates you want to accept the change that appears in the source container. For detailed information about the possible merge outcomes, see "Possible Repository Merge Outcomes Based on Your Decisions". Resolved Optional field that indicates the object difference has been resolved. Note: This field is for user reference only. It does not have an affect on the merge process.
Subtabs for Child Objects If the object selected in the navigation tree is a hierarchical object, the subtabs for the related child objects appear in the bottom pane of the Object Difference window.
Text Fields The text fields in at the bottom of the dialog, labeled "New Source System Container "and "Existing Source System Container," display a textual and hierarchical representation of the object difference selected in the Object Difference window.
Query Functionality You can query for objects in the Object Difference window and in the subtabs. For more information about the DAC query functionality, see "Using the DAC Query Functionality".
Difference Type Filter The drop-down list in the Object Difference window toolbar enables you to filter the list of differences based on the Change Description. The following filter options are available: All. Displays all changed objects. Added-Source. Displays objects that were added in the source container. Added-Target. Displays objects that were added in the target container. Cloned-Source. Displays objects that were cloned in the source container. Cloned-Target. Displays objects that were cloned in the target container. Deleted-Source. Displays objects that were deleted from the source container. Deleted-Target. Displays objects that were deleted from the target container. Modified. Displays objects that were modified differently in the source and target containers.
The following tables list the merge outcomes based on the decision you make about the object differences listed in the View Difference Report.
Upgrading, Comparing and Merging DAC Repositories 10-19
Table 102, " Merge Outcomes for Repository Upgrade (DAC 784) Option" Table 103, " Merge Outcomes for Replace Base and Refresh Base Options" Table 104, " Merge Outcomes for Peer to Peer Option"
Merge Outcomes for Repository Upgrade (DAC 784) Option Merge Action with User Choice: Accept Source Object will be merged into the custom container as an original object. Merge Action with User Choice: Reject Source Object will be deleted from the custom container. Not applicable
Table 102
Change Added-Source
Object Ownership Type Description Original An original object has been added in the source container.
Added-Source
Clone
A cloned object has been Not applicable added in the source container. A referenced object has been added in the source container. Not applicable
Added-Source
Reference
Not applicable
Added-Target
Original
An original object has Object will be deleted been added in the target from the custom container. container.
Object will be merged into the custom container as a referenced object to the base container. Object will be merged into the custom container as a referenced object to the base container. Object will be merged into the custom container as a referenced object to the base container. Object will be retained in the custom container as a reference to the base container. Object will be retained in the custom container as a reference to the base container. Object will be retained in the custom container as a reference to the base container. Object will be deleted from the custom container. Object will be deleted from the custom container.
Added-Target
Clone
A cloned object has been Object will be deleted added in the target from the custom container. container.
Added-Target
Reference
A referenced object has Object will be deleted been added in the target from the custom container. container.
Deleted-Source
Original
Deleted-Source
Clone
A cloned object has been Object will be deleted deleted from the source from the custom container. container. A referenced object has been deleted from the source container. An original object has been deleted from the target container. Object will be deleted from the custom container. Object will be merged into the custom container as an original object.
Deleted-Source
Reference
Deleted-Target
Original
Deleted-Target
Clone
A cloned object has been Object will be merged into the custom deleted from the target container as an original container. object.
10-20 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Table 102 (Cont.) Merge Outcomes for Repository Upgrade (DAC 784) Option Object Ownership Type Description Reference A referenced object has been deleted from the target container. An object has been modified differently in the source and target containers. Merge Action with User Choice: Accept Source Object will be merged into the custom container as an original object. Object will be merged into the custom container as a cloned object. Merge Action with User Choice: Reject Source Object will be deleted from the custom container. Object will be merged into the custom container as a referenced object to the base container. Not applicable
Change Deleted-Target
Modified
Not applicable
Modified-Cloned
Clone
Not applicable The parent of a cloned object has been modified in the source container. The parent of a referenced object has been modified in the source container. A referenced object has been cloned in the source container. Not applicable
Modified-Referenced Reference
Not applicable
Cloned-Source
Not applicable
Not applicable
Cloned-Target
A referenced object has Not applicable been cloned in the target container.
Not applicable
Table 103
Merge Outcomes for Replace Base and Refresh Base Options Merge Action with User Choice: Accept Source Object will be added to the target custom container as an original object. Merge Action with User Choice: Reject Source Not applicable
Change Added-Source
Object Ownership Type Description Original An original object has been added in the source container.
Added-Source
Clone
A cloned object has been Object will be added to the target custom added in the source container as a cloned container. object. Automatically adjusts the lineage for objects in child containers if the object was referenced in them.
Not applicable
Added-Source
Reference
Not applicable
Added-Target
Original
Object will be deleted An original object has been added in the target from the target custom container. container. Automatically adjusts the lineage for objects in child containers.
Table 103 (Cont.) Merge Outcomes for Replace Base and Refresh Base Options Object Ownership Type Description Clone Merge Action with User Choice: Accept Source Merge Action with User Choice: Reject Source Object will be retained in the target custom container.
Change Added-Target
A cloned object has been Object will be deleted added in the target from the target custom container. container. Automatically adjusts the lineage for objects in child containers.
Added-Target
Reference
A referenced object has Object will be deleted been added in the target from the target custom container. container. Automatically adjusts the lineage for objects in child containers.
Deleted-Source
Original
Object will be deleted from the target custom container. Automatically adjusts the lineage for objects in child containers if the object was referenced in them.
Deleted-Source
Clone
A cloned object has been Object will be deleted deleted from the source from the target custom container. container. Automatically adjusts the lineage for objects in child containers if the object was referenced in them.
Deleted-Source
Reference
Object will be deleted from the target custom container. Automatically adjusts the lineage for objects in child containers if the object was referenced in them.
Deleted-Target
Original
Object will be added as Not applicable an original object to the target custom container .
Deleted-Target
Clone
A cloned object has been Object will be added as Not applicable deleted from the target a cloned object to the container. target custom container . A referenced object has been deleted from the target container. An object has been modified differently in the source and target containers. The right-click command Record Info shows the objects lineage. Object will be added as Not applicable a referenced object to the target custom container . Not applicable Not applicable
Deleted-Target
Reference
Modified
Not applicable
10-22 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Table 103 (Cont.) Merge Outcomes for Replace Base and Refresh Base Options Object Ownership Type Description Clone Merge Action with User Choice: Accept Source Merge Action with User Choice: Reject Source
Change Modified-Cloned
The parent of a cloned De-clone the object and Not applicable object has been modified re-reference it. in the source container. Automatically adjusts the lineage for objects in child containers if the object was referenced in them. The parent of a referenced object has been modified in the source container. Re-reference the object. Automatically adjusts the lineage for objects in child containers if the object was referenced in them. Not applicable Not applicable Not applicable
Modified-Referenced Reference
Cloned-Source
Cloned-Target
A referenced object has Not applicable been cloned in the target container.
Not applicable
Table 104
Merge Outcomes for Peer to Peer Option Merge Action with User Choice: Accept Source Object will be added to the target custom container as an original object. Merge Action with User Choice: Reject Source Not applicable
Change Added-Source
Object Ownership Type Description Original An original object has been added in the source container.
Added-Source
Clone
A cloned object has been Object will be added to added in the source the target custom container. container as a cloned object. Automatically adjusts the lineage for objects in child containers.
Not applicable
Added-Source
Reference
Object will be added to Not applicable the target custom container as a referenced object. Object will be retained in the target custom container.
Added-Target
Original
An original object has Object will be deleted been added in the target from the target custom container. container. Automatically adjusts the lineage for objects in child containers.
Table 104 (Cont.) Merge Outcomes for Peer to Peer Option Object Ownership Type Description Clone Merge Action with User Choice: Accept Source Merge Action with User Choice: Reject Source Object will be retained in the target custom container.
Change Added-Target
A cloned object has been Object will be deleted added in the target from the target custom container. container. Automatically adjusts the lineage for objects in child containers.
Added-Target
Reference
A referenced object has Object will be deleted been added in the target from the target custom container. container. Automatically adjusts the lineage for objects in child containers.
Deleted-Source
Original
Object will be deleted from the target custom container. Automatically adjusts the lineage for objects in child containers if the objects were referenced in them.
Deleted-Source
Clone
A cloned object has been Object will be deleted deleted from the source from the target custom container. container. Automatically adjusts the lineage for objects in child containers if the objects were referenced in them.
Deleted-Source
Reference
Object will be deleted from the target custom container. Automatically adjusts the lineage for objects in child containers if the objects were referenced in them.
Deleted-Target
Original
Not applicable
Deleted-Target
Clone
A cloned object has been Object will be added to the target custom deleted from the target container as a cloned container. object. A referenced object has been deleted from the target container.
Not applicable
Deleted-Target
Reference
Object will be added to Not applicable the target custom container as a referenced object.
10-24 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Table 104 (Cont.) Merge Outcomes for Peer to Peer Option Object Ownership Type Description Not applicable An object has been modified differently in the source and target containers. The right-click command Record Info shows the objects lineage. Merge Action with User Choice: Accept Source Not applicable Merge Action with User Choice: Reject Source Not applicable
Change Modified
Modified-Cloned
Clone
De-clone the object and Not applicable The parent of a cloned object has been modified re-reference it. in the source container. Automatically adjusts the lineage for objects in child containers if the objects were referenced in them. The parent of a referenced object has been modified in the source container. Re-reference the object. Automatically adjusts the lineage for objects in child containers if the objects were referenced in them. Clone object in target custom container. Automatically adjusts the lineage for objects in child containers. Not applicable Not applicable
Modified-Referenced Reference
Cloned-Source
Not applicable
Cloned-Target
Not applicable
A referenced object has De-clone object in target Not applicable been cloned in the target custom container. container. Automatically adjusts the lineage for objects in child containers.
10-26 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
11
11
Importing DAC Metadata Exporting DAC Metadata Distributing DAC Metadata Running the DAC Server Automatically Command Line Access to the DAC Server DAC Repository Command Line Options Replacing an Informatica Workflow with a Custom SQL File Determining the Number of Transactional and Data Warehouse Database Connections Running Two DAC Servers on the Same Machine Customizing Index and Analyze Table Syntaxes Overview of Change Capture Process (Siebel Sources Only) Using the Change Capture Filter Tracking Deleted Records Pointing Multiple Informatica PowerCenter Services to a Single Informatica Repository Handling ETL Failures with the DAC
In the DAC menu bar, select Tools, then select DAC Repository Management, then select Import. Select the directory from which you want to import DAC metadata, or accept the default directory. Select the appropriate source system containers.
4.
Logical. Imports all information contained in the Design view and database connection information. System. Imports all information contained in the Setup view, except passwords for servers and database connections. Run Time. Imports information about ETL runs (contained in the Execute view)
5.
If you are importing metadata into a blank repository or to completely replace the current metadata in the repository, select Truncate Repository Tables. This action overwrites the content in the current repository. Selecting the Truncate Repository Tables option greatly increases the speed of the import process.
6.
(Optional) Select Enable Batch Mode to insert the imported metadata into the repository as an array insert. This action increases the speed of the import process.
7. 8.
Click OK. Verify the import process by reviewing the log file \OracleBI\DAC\log\import.log.
In the DAC menu bar, select Tools, then select DAC Repository Management, then select Export. Select the directory to which you want to export DAC metadata, or accept the default directory. Select the appropriate source system containers. Select the appropriate categories of metadata you want to export:
Logical. Exports all information contained in the Design view and database connection information. System. Exports all information contained in the Setup view, except passwords for servers and database connections. Run Time. Exports information about ETL runs (contained in the Execute view).
5. 6.
Click OK. Verify the export process by reviewing the log file \OracleBI\DAC\log\export.log.
11-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
test it, and then deliver it, exporting the whole environment and distributing it to the other environments. The data is exported as XML files, which are stored in the DAC\export directory on the client machine where the export is done. To apply changes from the development environment to any other, you copy all of the XML files into the DAC\export folder and then import the data. To export the DAC metadata, follow the instructions in the procedure, "Exporting DAC Metadata". To import the DAC metadata, follow the instructions in the procedure, "Importing DAC Metadata".
Navigate to Programs, then select Accessories, then select System Tools, then select Scheduled Tasks. Double-click Add Scheduled Task. In the Scheduled Task Wizard, browse to the startserver.bat file, and click Open. Select the option "When my computer starts," and click Next. Enter the domain user account to start the DAC server and a password, and click Finish. The startserver task appears in the Scheduled Task window.
6. 7.
Right-click the task and select Properties. In the Settings tab, remove the check from the "Stop the task if it runs for 72 hours" check box.
Navigate to Programs, then select Accessories, then select System Tools, then select Scheduled Tasks. Right-click startserver, and then click Run.
Navigate to Programs, then select Accessories, then select System Tools, then select Scheduled Tasks. Right-click startserver, and then click End Task.
Navigate to Programs, then select Accessories, then select System Tools, then select Scheduled Tasks. Select the startserver task. In the Windows menu bar, select View, then select Details.
You can access the DAC server through a command line to start and stop execution plans and to get status information for servers, databases, and execution plans. This feature enables you to access the DAC server using third-party administration or management tools, without using the DAC client. Refer to the file daccommandline.bat/.sh for usage.
The Generic Task Concurrency Limit value in the DAC System Properties tab is not a positive number. There are no active Informatica PowerCenter Services or Integration Service registered in the Informatica Servers tab. One or more Informatica PowerCenter Services or Integration Service do not have the passwords defined in the Informatica Servers tab. One or more Informatica PowerCenter Services or Integration Service do not have a Maximum Sessions number properly defined in the Informatica Servers tab. One or more data sources do not have the Table Owner or Table Owner Password values properly defined in the Physical Data Sources tab. One or more data sources do not have a maximum number of connections (Max Num Connections) value properly defined in the Physical Data Sources tab. One or more data sources do not have a Data Source Number defined in the Physical Data Sources tab.
The name of the execution plan that is running is different from the name in the request. There is no execution plan currently running.
11-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Summary of the requested execution plan. If there are multiple instances of the same execution plan, a summary of the instance that last ran is returned. Below is an example of the information contained in the summary.
(c) 2003 Siebel Systems, Inc. Siebel DAC Server comprising the etl execution-management, scheduler, logger, and network server. ETL details for the last run: ETL Process Id : 255 ETL Name : Complete ETL Run Name : DRY RUN OF Complete ETL: ETL Run - 2004-06-17 18:30:13.201 DAC Server : (aqamarD510) DAC Port : 3141 Status: Stopped Log File Name: Complete_ETL.255.log Database Connection(s) Used :
OLTP jdbc:microsoft:sqlserver://vranganaw8:1433;DatabaseName=OLTP Data Warehouse jdbc:microsoft:sqlserver://vranganaw8:1433;DatabaseName=olap Informatica Server(s) Used : InformaticaServer2-vranganaw8:(2) InformaticaServer1-vranganaw8:(10)
InformaticaServer4-vranganaw8:(4) InformaticaServer3-vranganaw8:(3)
Start Time: 2004-06-17 19:00:06.885 Message: ETL was interrupted Actual Start Time: 2004-06-17 18:30:13.357 End Time: 2004-06-17 19:05:56.781 Total Time Taken: 35 Minutes Start Time For This Run: 2004-06-17 19:00:06.885 Total Time Taken For This Run: 5 Minutes Total steps: 212 Running steps: 0 Complete steps: 142 Failed/Stopped steps:70
Make sure you have installed the supported version of the Java SDK. Copy the following files from the OracleBI\DAC directory to a local directory:
3.
Edit the JAVA_HOME variable to point to the directory where the Java SDK is installed.
Edit the DAC_HOME variable to point to the directory where the DAC is installed.
Value Host name of the DAC server. Port of the DAC server. The default is 3141. Repository stamp that appears in the DAC client Login Details screen. To find this value, in the DAC client navigate to Help, then select Login Details.
DatabaseStatus InformaticaStatus
Note:
The method names are case insensitive. Execution plan names are case sensitive. Also, if the execution plan name contains spaces, place beginning and ending double quotes around the name.
For example:
Command Line dacCmdLine EtlStatus dacCmdLine EtlStatus Forecast Description Returns the status of the execution plan that last ran. Returns the status of the last instance of the Forecast execution plan.
11-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Description If the execution plan currently running is Forecast, the operation will be terminated. Otherwise, the request is ignored. Returns the health status of all the database connections as defined in the DAC repository from the DAC server.
dacCmdLine databasestatus
dacCmdLine InformaticaStatus Returns the health status of all the Informatica PowerCenter Services connections as defined in the DAC client on the Informatica Services tab.
where:
Parameter folderName contName Description Full path to the root of the import file structure. (Optional) Name of the source system container for which you want to import DAC metadata. If no container is named, all containers that are found in the file structure will be imported.
where:
Parameter folderName contName Description Full path to the root of the export file structure. (Optional) Name of the source system container for which you want to export DAC metadata. If no container is named, all containers that are found in the file structure will be exported.
Syntax:
IMPORTCATEGORY <folderName> <logical> <runtime> <system>
where:
Parameter folderName logical runtime system Description Full path to the root of the import file structure. Imports all data categorized as logical (information contained in the DAC Design view). Imports all data categorized as run time (information contained in the DAC Execute view). Imports all data categorized as run time (information contained in the DAC Setup view).
where:
Parameter folderName logical runtime system Description Full path to the root of the import file structure. Exports all data categorized as logical (information contained in the DAC Design view). Exports all data categorized as run time (information contained in the DAC Execute view). Exports all data categorized as run time (information contained in the DAC Setup view).
Create Schema
The CREATESCHEMA option creates the schema of a new DAC repository. Syntax:
CREATESCHEMA <unicodeFlag> <workSpace name>
where:
Parameter unicodeFlag workSpace name Description If the value of this parameter is true, the schema is created as unicode. If the value is false, it is not created as unicode. The name of the workspace in which the schema is created.
Drop Schema
The DROPSCHEMA option drops the schema of the DAC repository. Syntax:
DROPSCHEMA
11-8 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Analyze
The ANALYZE option analyzes the DAC repository tables. Syntax:
ANALYZE
Upgrade
The UPGRADE option upgrades the DAC repository. Syntax:
UPGRADE
Set Password
The SETPASSWORD option sets the passwords for the PowerCenter Services, Integration Service, and physical data sources in the DAC repository. Syntax:
SETPASSWORD <type> <logicalName> <password>
where:
Parameter type logicalName Description Possible values are server or dbconn. Logical name of the server or data source record in the DAC.
Note: : If the logical name or password contains spaces, quotes are required.
Create a SQL action to be used to load the table, and unit test it. You can create one action for a full load and one for an incremental load, or you can use the same file for both full and incremental loads.
2. 3. 4.
In the Tasks tab of the DAC Design view, query for the task for which you want to replace the Informatica workflow. Change the Execution Type to SQL. Replace the workflow name in the Command for Incremental Load or Command for Full Load fields with the SQL action.
PowerCenter Services. If the number of sessions is zero or is not specified, the DAC server assigns the default value of 10. You should consider the following factors when determining the Maximum Sessions parameter value:
How powerful the machine is that hosts the PowerCenter Services. How many instances of PowerCenter Services are available. The number of Runnable tasks in the queue. A Runnable task is a task for which the Depends On tasks have completed and is ready to be run but is waiting for an Informatica slot to be available. For information about the different task run statuses, see "Current Run Tab". For an optimal run, the runnable queue should be at zero or should reach zero in a short time. For example, Figure 111 shows an ideal performance run when 15 sessions were run in parallel. There were many runnable tasks before the process began, but the queue soon reached zero. You can generate a run analysis such as Figure 111 from the right-click menu (select Get Run Information, then select Get Graph) on the DAC Current Run and Run History tabs. If you find that the runnable curve does not reach zero in a short time, you should increase the Maximum Sessions parameter value to make more Informatica slots available.
11-10 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
For the transactional database, the DAC server uses these connections to perform change capture. The number of connections you set for this connection pool determines how many change capture processes can run concurrently. If you have a powerful transactional database server and are going to run ETL processes during off-peak times, you can increase the Max Num Connections setting to 15 or 20 (10 is the default). If you have a less powerful transactional database server, you should not overload the operational system with ETL processes. Therefore, you should set the value below 10. For the data warehouse database, the DAC server uses these connections to perform processes such as truncate tables, drop and create indexes, and analyze tables. You should not set the Max Num Connections value higher than the Maximum Sessions parameter value (the maximum number of workflows that can be executed in parallel in the PowerCenter Services) because these values have a one to one relationship.
Copy the OracleBI\DAC folder to a different folder on the same machine. For example, you might copy the C:\OracleBI\DAC folder to C:\DAC_ SERVER2\DAC.
2.
Edit the config.bat file to set the DAC_HOME variable appropriately for each instance. For example if you copy the C:\OracleBI\DAC folder to C:\DAC_ SERVER2\DAC, make sure that the C:\DAC_SERVER2\DAC\config.bat file is configured correctly.
3. 4.
Launch each of the DAC clients by navigating to the DAC directories and double-clicking the startclient.bat file. For each instance, configure the DAC repository connection.
a.
Navigate to Tools, then select DAC Server Management, then select DAC Server Setup. An informational dialog states this operation should be performed on the machine running the DAC server. It asks whether you want to continue.
b. c.
Click Yes. In the Repository Connection Information tab, enter the appropriate information for each instance. The Database Host should be the same for each instance, and the Database Port should be different.
5.
Navigate to Setup, then select DAC System Properties. Set the DAC Server Host, OS, and Port properties.
6.
Open the customsql.xml file located in the OracleBI\DAC\CustomSQLs directory. Locate the Analyze Table syntax for the appropriate database type. For example, the syntax for an Oracle database is as follows:
<SqlQuery name = "ORACLE_ANALYZE_TABLE" STORED_PROCEDURE = "TRUE"> DBMS_ STATS.GATHER_TABLE_STATS(ownname => '@TABLEOWNER', tabname => '%1', estimate_ percent => 30, method_opt => 'FOR ALL COLUMNS SIZE AUTO',cascade => true ) </SqlQuery>
3.
Edit the syntax. For example, to gather statistics for only the indexed columns, edit the syntax as follows:
<SqlQuery name = "ORACLE_ANALYZE_TABLE" STORED_PROCEDURE = "TRUE"> DBMS_ STATS.GATHER_TABLE_STATS(ownname => '@TABLEOWNER', tabname => '%1', estimate_ percent => 30, method_opt => 'FOR ALL INDEXED COLUMNS',cascade => true ) </SqlQuery>
Note:
The variables @TABLEOWNER, %1, %2, and so on, will be substituted appropriately by the DAC when the statement is executed.
Open the customsql.xml file located in the OracleBI\DAC\CustomSQLs directory. Locate the Create Index syntax for the appropriate database type, and edit the syntax.
11-12 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
This is the most common method. It uses S_ETL_I_IMG_ and S_ETL_R_IMG_ table types and the LAST_UPD column of the source tables.
Change capture using the date column In some cases, a predefined date column is used to enable change capture (without use of S_ETL_I_IMG_ and S_ETL_R_IMG_ tables).
SDE process). This situation occurs when child table rows have changed and the header/master rows need to be extracted so that data warehouse tables are loaded with a consistent set. When the S_CONTACT_X row is changed, the corresponding S_CONTACT also needs to be extracted. In this case, rows in S_CONTACT are also marked as changed by inserting rows in the change capture row image table. When the S_ORDERITEM row is changed, the corresponding S_DOC_ORDER also needs to be extracted. In this case, rows in S_DOC_ORDER are also marked as changed (by inserting rows in the change capture row image table). These auxiliary changes capture processes are heavily dependent on the data warehouse data model and are required to support the ETL logic. Example: Building S_ETL_I_IMG_ Table for Loading Account Dimension This section gives an extended example of the process of change capture using tables.
1.
Load image tables for all relevant source tables. The content of this entity comes from the S_ORG_EXT and S_ORG_EXT_X tables. Whenever any of the rows change in either of these tables, the record is marked as changed. The image table for S_ORG_EXT is S_ETL_I_IMG_26. The image table prefix can be found using the DAC to view any source table. This table is truncated before loading with fresh data during every refresh. During the ETL, process rows are inserted into S_ETL_I_IMG_26 by selecting ROW_ID information from S_ORG_EXT, for rows (combined ROW_ID and MODIFICATION_NUM) that do not exist in the S_ETL_R_IMG_26 and for which LAST_UPD is more recent than LAST_REFRESH_DATE minus the Prune Days setting. This is done during the ETL execution by the DAC's internal image building tasks. Similarly, the image table S_ETL_I_IMG_27 for S_ORG_EXT_X is loaded.
2.
Load the image table for auxiliary table-based changes. In addition to the basic change capture, extra processing might be required due to special ETL requirements. In this example, it happens that S_ORG_EXT needs to be extracted for processing even if only S_ORG_EXT_X changes. This is because both the tables are joined to form W_ORG_D, and the extract process of W_ORG_ D (a SDE mapping) looks for a changed ROW_ID in the change capture row image table for the primary table S_ORG_EXT only. Therefore, the extract happens only when the ROW_ID for S_ORG_EXT exists in the row image table. In this case, the SDEINC_FindAux_ mapping is needed to insert corresponding rows of S_ORG_EXT.ROW_ID in the change capture row image table whenever S_ ORG_EXT_X changes. The following logical statement shows the method: Identify the records that have changed in the S_ORG_EXT_X (rows in S_ETLI_ IMG_27) table and then find the corresponding rows in S_ORG_EXT. Insert the ROW_ID and MODIFICATION_NUM of those corresponding rows from S_ORG_ EXT into S_ETL_I_IMG_26 table. Using Informatica, the auxiliary mapping SDEINC_FindAux_ has to be written for each primary table that requires it, depending on data warehouse extract logic. Using the DAC, this auxiliary task has to be linked as a parent to the extract mapping for the base table (S_ORG_EXT in this case). This is the SQL override for the SDEINC_FindAux Informatica mapping:
11-14 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
SELECT S_ORG_EXT.ROW_ID, S_ORG_EXT.MODIFICATION_NUM, S_ORG_EXT.LAST_UPD FROM S_ORG_EXT, S_ORG_EXT_X, S_ETL_I_IMG_27 IMG WHERE ( IMG.ROW_ID = S_ORG_EXT_X.ROW_ID AND S_ORG_EXT_X.PAR_ROW_ID = S_ORG_EXT.ROW_ID ) AND NOT EXISTS ( SELECT 'X' FROM S_ETL_I_IMG_26 IMG1 WHERE IMG1.ROW_ID = S_ORG_EXT.ROW_ID ) 3.
Extract source table information using change capture image information. After the records that are new or modified are identified, those rows are loaded into the staging tables. The Informatica mappings that load the staging tables use the ROW_ID information captured in the image tables. This example shows the loading of staging table W_ORG_DS. The main logic of populating this table lies in the SQL override of the mapping SDE_ OrganizationDimension. The DAC creates views on tables that are being extracted. The views are different, depending on whether a table is extracted the first time or is a change capture extract.
If extracting for the first time, the view is created as SELECT * FROM S_ ORG_EXT. If it is a change capture extract, the view is created as SELECT * FROM S_ ORG_EXT, S_ETL_I_IMG_26 IMG WHERE S_ORG_EXT.ROW_ID = IMG.ROW_ID.
The SQL override in the mapping uses the view to extract the data.
SELECT S_ORG_EXT.ROW_ID, S_ORG_EXT.NAME, .. . . FROM V_ORG_EXT, S_ORG_EXT_X, .. WHERE { V_ORG_EXT S_ORG_EXT LEFT OUTER JOIN S_ORG_EXT_X ON S_ORG_EXT.ROW_ID = S_ORG_EXT_X.PAR_ROW_ID .. } AND
Forecasts in the Oracle Business Analytics Warehouse are never updated. Once loaded, they are frozen.
SELECT .. FROM S_FCSTSER_DATE, S_FCSTSER, S_ETL_CURR_RUN, .. WHERE S_FCSTSER_DATE.FCSTSER_ID = S_FCSTSER.ROW_ID AND S_FCSTSER_DATE.ARCHIVE_TS > S_ETL_CURR_RUN.PREV_LOAD_DT AND S_FCSTSER_DATE.ARCHIVE_TS <= S_ETL_CURR_RUN.LOAD_DT
11-16 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
For vertical applications, the preconfigured ETL process captures deleted records for W_FUND_F and W_ALIGNMT_DH. You need to create delete triggers in the transactional database for the following additional tables: S_MDF_TXN, S_ASGN_ GRP_POSTN, S_ASGN_RULE_ITEM. In the Oracle Business Analytics Warehouse, preconfigured visibility tables are inactivated. If you activate visibility tables, you should also create delete triggers on the optional tables. The preconfigured SIA Account and Contact visibility tables are activated by default for vertical applications. If your organization is not going to use any of the visibility tables, you need to inactivate them in the DAC. On the target tables for which deleted records are tracked, a D appears in the INACTIVE_FLG column to show the records as deleted when the source records are deleted. This method of flagging a record as deleted is known as a soft delete, as compared to a hard delete when the record is physically deleted. When deleted records are tracked on visibility-related data warehouse tables, the records are physically deleted. The general rule is that soft deletes should be used for tables that are referenced by other tables. If a table is not referenced by any other table, then you can use hard deletes. Aggregate tables are rebuilt during each ETL process. Therefore, records can be physically deleted from the base tables without consequence. If you want to use the soft delete method, you should consider changing the aggregate building mappings so that the deleted records are omitted.
Note:
The Oracle BI Server does not recognize soft deletes. Therefore, you have to modify the .rpd file so that it does not pick up soft-deleted records for reporting.
From the DAC menu bar, select Tools, then select ETL Management, then select Configure. In the Sources dialog, select the database platform for the target and transactional databases, and click OK. In the Data Warehouse Configuration Wizard, select the Create Delete Triggers in Transaction Database check box, and click Next. The Delete Triggers tab is active.
4.
5. 6. 7. 8.
Select the database type as defined in the DAC. For DB2 zSeries databases, enter the base table owner. (Optional) Select the Include Optional Triggers check box to create triggers for the optional tables. Click Start.
Common Tasks Performed in the DAC 11-17
In the DAC, navigate to the Design view, then select Tables. Select the table for which you want to track deleted records. Make sure the table has an image suffix.
3. 4. 5. 6.
Right-click the table and select Change Capture Scripts, then select Generate Image and Trigger Scripts. In the Triggers and Image Tables dialog, select the database type of the source database. Make sure the Generate Image Table Scripts and Generate Trigger Script(s) options are selected. Execute the script on the database.
Make sure the delete triggers are enabled for the appropriate tables. Write custom Informatica workflows with a clause WHERE operation = 'D' to the appropriate I_IMG table to take them across to the dimension and fact tables. In the DAC, register the workflows as tasks. Define the appropriate dependencies. For an example of such a workflow, see the preconfigured task SDE_ OrganizationDimension_LoadDeletedRows.
When the Execution of an Execution Plan Fails Discarding the Current Run Execution Plan Failure of Aggregator Transformation Tasks with Sorted Input
11-18 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
task, change its status to Completed, and then restart the ETL. Tasks with a Completed status are skipped.
Caution:
The DAC server does not validate tasks that have been run
manually. To restart a failed ETL, click Run Now from the Current Run tab of the Execute View.
11-20 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
12
12
This chapter describes the functionality available in the Data Warehouse Administration Console (DAC) tabs. It contains the following topics:
Common Elements of Interface Tabs Design View Tabs Setup View Tabs Execute View Tabs
12-2 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Configuration Tags Tab Indices Tab Source System Folders Tab Source System Parameters Tab Subject Areas Tab Tables Tab Task Groups Tab Tasks Tab
Remove tasks from all subject areas When you assign a task to a configuration tag, the task will not be eligible to participate in any subject area.
Reassign autogenerated tasks to a specific subject area An autogenerated task is a task that the DAC automatically assigns to a subject area when the subject area is assembled. For autogenerated tasks that were removed from participating in a subject area, you can set up the configuration tag to reassign a task to participate in specific subject areas. You do this by associating the configuration tag with the desired subject area. This method only applies to tasks that are autogenerated tasks of a subject area.
Add non-autogenerated tasks to a subject area You can set up a configuration tag to add non-autogenerated tasks to a subject area. The non-autogenerated tasks will participate in the subject area along with the subject areas autogenerated tasks.
Assign only configuration tag tasks to a subject area (excludes the subject areas autogenerated tasks) You can also set up a configuration tag so that only tasks that were assigned to the configuration tag participate in a specific subject area. In this case, the subject areas autogenerated tasks do not participate.
For instructions on creating configuration tags, see "Working with Configuration Tags". Include Tasks If this check box is selected, the tasks that are assigned to a configuration tag will participate in the ETL process for the subject area to which this configuration tag is assigned. For example, suppose Configuration Tag 1 is made up of Task 1 and Task 2, and Configuration Tag 1 is assigned to Subject Area 1. Task 1 and Task 2 will be executed when the execution plan for Subject Area 1 is executed, whether or not Task 1 and Task 2 relate to the tables that make up the subject area.
12-4 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Indices Tab
Indices Tab
The Indices tab lists all the indexes associated with the selected source system container. It is recommended that you do not register any indexes for source tables. During the ETL process, when a table is going to be truncated, all the indexes as defined in the repository will be dropped before the data is loaded and will be created after the data is loaded automatically. While this improves the ETL performance, the preconfigured workflows have the bulk load option turned on. The bulk load will fail if there are indexes on the table. Therefore, it is important to keep the index definitions in sync with the database. For example, if you create an index on the database, and it is not registered in the repository, the index will not be dropped and the load will fail. For Teradata databases, only secondary indexes should be registered in the DAC. You should not register primary indexes or the more complex indexes, such as single- and multi-table indexes, because they cannot be dropped and recreated. You can use SQL commands to drop and create such tasks in the DAC. For information using the Index Action feature to override default index behavior, see "About Index, Table and Task Actions". Table Name The table for which an index is created. Index Usage Specifies the index usage: ETL or Query. An ETL index is typically used during the ETL process. A Query index is an index used only during the reporting process. It is recommended that you have a clear understanding of when and where the index will be used at the time of registering the index in the DAC. Databases Lists the databases associated with the selected index. # Unique Columns For unique indexes, the number of columns that will be unique. Is Unique Indicates whether the index is unique. Is Clustered Indicates whether the index is clustered. There can be only one clustered index per table. Is Bitmap Indicates whether the index is of the bitmap type. Allow Reverse Scan Applicable only for DB2-UDB databases. The index will be created with the Allow Reverse Scan option. Always Drop & Create Indicates whether the index will be dropped and created regardless of whether the table is being loaded using a full load or incremental load.
12-6 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Always Drop & Create Bitmap Indicates whether indexes of the bitmap type will be dropped and created regardless of whether the table is being loaded using a full load or incremental load.
About Index, Table and Task Actions Defining a SQL Script for an Action Assigning an Action to a Repository Object
12-8 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
About Parameter Management Defining a Text Type Parameter Defining a Database Specific Text Type Parameter Defining a Timestamp Type Parameter Defining a SQL Type Parameter
Data Type Parameter data type. For more information, see "Overview of Parameters". Possible values are the following:
Data Type Option Text DB Specific Text Description The value for the parameter is defined as text. Enables you to add database specific hints in Informatica mappings. When you select this option, in the Value field, you specify the logical connection where the parameter applies, and you specify a default value for the parameter. The DAC evaluates the parameter to this default value for all databases. If you enter text that is applicable to a particular database, you can specify a value for the database type, and the DAC will evaluate the parameter to this value at runtime as long as the logical connection matches the specified database. Timestamp SQL The value for the parameter is a timestamp and can be static, runtime or SQL. The value for the parameter is a SQL statement.
Name The name of the table associated with the selected subject area.
Tables Tab
Tables Tab
The Tables tab lists the physical database tables defined in the database schema that are associated with the selected source system container. It enables you to view and edit existing tables and to create new ones. For information about adding new tables, see "Adding a New Table and Columns to the Data Warehouse". Table Type The table type. Warehouse Indicates whether the table is a warehouse table. If this option is not selected, the schema creation process will not include this table. Image Suffix Suffix for image tables. Applicable only to Siebel source tables. For more information about image tables, see the description of the Change Capture Scripts command in the section "Design View Right-Click Menu Commands". Is MultiSet Indicates whether the table is a MultiSet table. Applicable only to Teradata databases. Has Unique Primary Index Indicates whether the table has a Unique Primary Index. Applicable only to Teradata databases.
About Index, Table and Task Actions Defining a SQL Script for an Action Assigning an Action to a Repository Object
The table action types enable you to trigger SQL scripts to analyze or truncate tables. Table actions for analyzing or truncating tables override all other table properties. Action Type The default actions on a table during ETL execution are truncating a table and analyzing a table. If you want to override any of these syntaxes, you can define actions per mode (full or incremental). Once an action is defined, it overrides the default behavior. Note: You can associate an action to multiple objects at the same time. First, identify the objects that you want to associate an action with by using the query functionality. Then, right-click on the results displayed, and select Associate Actions. Next, select an action and the mode you want to associate it with. The table action types are the following:
Analyze Table
12-12 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Load Type The load type specifies whether the SQL script is to be called for incremental runs, full runs, or both. Action Actions are the SQL statement or stored procedure that you define. You can define one or more SQL statements or stored procedures for each action.
Tables Tab
Allow Reverse Scan Applicable only for DB2-UDB databases. The index will be created with the Allow Reverse Scan option. Table Space Name Name of the table space.
12-14 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Task Task that extracts data from the table. Type Source table type. If a table is marked as Primary or Auxiliary and the Build Image property of the task is selected, the change capture process is invoked. There are special tasks that force the base table data to be extracted when data in auxiliary tables changes. A table can be neither Primary nor Auxiliary but still be used for getting some attributes to populate a dimension or fact table. The changes in these kinds of source tables are not reflected in the dimension or fact table once the data is populated.
12-16 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Tasks Tab
The Tasks tab lists all the tasks associated with the selected source system container. Parent Group If the task is a member of a group, this field lists the task group name. Group Order The order in which the task is defined to execute in a certain group. Command for Incremental Load A table can be loaded in Full Mode or Incremental Mode. Full Mode refers to data loaded for the first time or data that is truncated and then loaded. Incremental Mode refers to new or changed data being added to the existing data. The DAC maintains a last refresh timestamp whenever a table is changed during the ETL process. (You can view this timestamp by selecting Setup, then selecting Physical Data Sources, and then selecting Refresh Dates.) If a table has a timestamp, the command appearing in this column is executed. If a table does not have a timestamp, the command for a full load is executed. If the execution type is Informatica, the workflow name is used as the command Command for Full Load If a table has no last refresh timestamp, this command is executed. Folder Name Only for execution type of Informatica. The folder in which the workflow resides. Note: The name cannot contain spaces. Primary Source Logical database connection for the primary source database. Primary Target Logical database connection for the primary target database. Task Phase Task phase of the ETL process. This information is primarily used for dependency generation. Certain phases, such as Change Capture and Query Index Creation, are not available for you to assign to a task. The DAC server uses the task phase to prioritize tasks and to generate a summary of the time taken for each of the phases. Execution Type Tasks are executed based on their execution type. The following types are supported:
External Program Task is an operable program on the operating system where the DAC server is running. This program can be a batch file, shell script, or any other program that can be run like a bulk loader.
Tasks Tab
Stored Procedures Task is a stored procedure that has been defined on the databases.
In addition, there are several internal execution types that you will not be able to select when creating new tasks. These tasks are categorized as either internal change capture tasks or internal data warehouse tasks; all of these tasks are color-coded in pink in the Tasks tab.
IMG_BUILD Used for internal change capture. If you are using multiple Siebel transactional sources, you cannot change the behavior of the change capture process. This task requires change capture tables to be created on the other sources also. When adding additional Siebel sources, go to Design > Tables, right-click and select Change Capture Tasks. This action generates change capture tasks. Use this same action to disable or delete change capture tasks.
IMG_SYNC Used for internal change capture. If you are using multiple Siebel transactional sources, you can create this task for the additional tasks for doing similar change capture sync processes. You cannot change the behavior of the change capture sync process. This task requires change capture tables to be created on the other sources also. This task should be used with discretion for Siebel sources only.
Execution Priority Indicates the order in which the task is executed. If two or more similar tasks (tasks having same phase, similar truncate properties, same number of successors, same number of source tables) have the same priority, the order occurs randomly. Build Image Applicable for Siebel transactional sources only. Indicates the change capture for the primary/auxiliary source tables will be executed. Continue on Error When this check box is selected, if the command fails, the dependent tasks are not stopped. However, if any autogenerated tasks fail, the dependent tasks are stopped.
About Index, Table and Task Actions Defining a SQL Script for an Action Assigning an Action to a Repository Object
Action Type Action types are predefined categories of task behaviors that trigger the execution of a SQL script. The following types are available:
Preceding Action Use this type to execute a SQL script before a task runs.
12-18 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Success Action Use this type to execute a SQL script after a task runs successfully.
Failure Action Use this type to execute a SQL script if a task fails during its execution.
Restart Action Use this type to execute a SQL script when a task that previously failed is restarted.
Upon Failure Restart Action Use this type to execute a SQL script to restart a task that fails.
Load Type The load type specifies whether the SQL script is to be called for incremental runs, full runs, or both. Action You define the task action in the form of a SQL statement or stored procedure. You can define one or more SQL statements for each action. Double-click in the Action field to open the Choose Action dialog box, where you can select the appropriate action. You define the SQL statement or stored procedure for an action in the Task Actions dialog box, which you access by selecting Tools, then Seed Data, then Actions, and then Task Actions. For instructions, see "Defining a SQL Script for an Action".
Tasks Tab
Defining a Text Type Parameter Defining a Database Specific Text Type Parameter Defining a Timestamp Type Parameter Defining a SQL Type Parameter Name Name of the parameter. Data Type Parameter data type. For more information, see "Overview of Parameters". Possible values are the following:
DB Specific Text Enables you to add database specific hints in Informatica mappings. When you select this option, in the Value field, you specify the logical connection where the parameter applies, and you specify a default value for the parameter. The DAC evaluates the parameter to this default value for all databases. If you enter text that is applicable to a particular database, you can specify a value for the database type, and the DAC will evaluate the parameter to this value at runtime as long as the logical connection matches the specified database.
Timestamp The value for the parameter is a timestamp and can be static, runtime or SQL.
Data Type The parameter data type. Possible values are Text, Timestamp, and SQL. Value The parameter value.
Wait Indicates the selected task will wait to be executed until the tasks of a specified phase have been executed.
Block
12-20 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Indicates the selected task will block all tasks of the specified phase from being executed until is has been executed. Grain Applicable only for blocks. Enables you to specify whether the action you choose affects all tasks of a specified phase or related tasks. Possible values are the following:
Related Indicates the action will affect only related tasks. You can view a tasks related tasks by navigating to the Execution Plans tab, All Dependencies subtab and viewing the specified tasks predecessor tasks.
Scope For multi-source execution plans only. Specifies how the Block action of the phase dependency behaves in relation to multi-source execution plans. Possible values are the following:
Both Indicates the blocking action is active for tasks that have the same source and target physical data source connections.
Source Indicates the blocking action is active for tasks that have the same source physical data source connection.
Target Indicates the blocking action is active for tasks that have the same target physical data source connection.
None Indicates the blocking action is active for all tasks regardless of the source and target physical data source connections.
Phase The ETL phase that will apply to the Action and Grain properties.
Tasks Tab
aggregates more efficient. Hence, the decision to load the table in Full mode should depend on the ultimate source table, which is qualified as OLTP.S_REVN, rather than based on the immediate source table Datawarehouse.W_REVENUE_F.
If a table is marked as Primary or Auxiliary and the Build Image property of the task is selected, the change capture process is invoked. There are special tasks that force the base table data to be extracted when data in auxiliary tables change. A table can be neither Primary nor Auxiliary but still be used for getting some attributes to populate a dimension or fact table. The changes in these kinds of source tables are not reflected in the dimension or fact table once the data is populated.
Data Source Data source for the table. When a data source is not specified, the default is the tasks primary source.
12-22 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Truncate for Full Load Indicates the target tables will be truncated only when a full load is occurring. Any indexes registered for this table are dropped before the command is executed and are recreated after the command completes successfully. When indexes are dropped and created, the table is analyzed so that the index statistics are up-to-date. When the Truncate Always option is selected, this option is unnecessary.
DAC System Properties Tab Email Recipients Tab Informatica Servers Tab Physical Data Sources Tab
12-24 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Drop and Create Change Capture Views Always Possible values are True and False. When set to True (the default value), the DAC server drops and creates change capture views every time it performs a change capture process, including for both full and incremental loads. Setting this property to True can create system catalog lock up for DB2-UDB and DB2-390 databases. Therefore, by setting the property to False, the DAC server will drop and create views selectively, using the following rules:
In full mode: During the change capture phase, views will be dropped and created as full views. During the change capture sync process, incremental views will be generated.
In incremental mode: If the view exists, it will not be dropped and created. If the view does not exist, the incremental view will be created.
Dryrun Possible values are True and False. Indicates whether tasks are executed without invoking Informatica workflows. The following processes are executed: change capture, truncation of tables, drop and creation of indexes, and analyze statements. This option should be used for debugging purposes only and not used in a production environment. Generic Task Concurrency Limit Determines how many tasks with execution types other than Informatica can be run concurrently. The value must be numerical. To set this value, you should consider what the external tasks do. For example, if the tasks open connections to a database, you should consider how this would affect the preconfigured tasks. HeartBeatInterval Frequency (in seconds) the DAC server checks on the health of the database connections. The value must be numerical. For example, a value of 300 (the default value) indicates the system will perform subsystem diagnostics and recovery procedures every 300 seconds. InformaticaFileParameterLocation Directory where the Informatica parameter file is stored. No Run Generates tasks in the Task Details subtab of the Current Run tab but does not execute them. Output Redirect Indicates whether logging information and standard output and errors are redirected to files in the log directory (when property is set to True). The file containing standard
12-26 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
output starts with out_ and ends with the .log extension. The standard error messages are in the file starting with err_ and ending with the .log extension. If this property is set to False, the logging information is directed to the machines standard output and error files, which typically defaults to the console from which the DAC server was launched if the server was launched in a visible console mode. If the server is launched as a Windows service, the logging information is directed to the service log. If the server is launched with the command shell not visible, all logging information is deleted Repository DB Pool Size Indicates the maximum number of connections to the DAC repository that the server will maintain. Repository Name Unique name for the DAC repository. Scheduler.Poll.Interval Frequency (in seconds) the DAC server polls for changes in the schedule configuration. Script After Every ETL The name of the script or executable to be run after every execution plan. For more information, see the description of the property Script Before Every ETL. Script Before Every ETL The name of the script or executable to be run before every execution plan. For example, before running an execution plan, you might want to run a process or perform certain tasks. These can be contained in a script or executable. This file should be placed in the scripts subdirectory of the DAC server. The execution plan runs only after the external process has finished. Therefore, it is important that the script or executable does not fail. Server Log Level Output logging level. Possible values are Finest, Finer, Fine, Config, Info, Warning, and Severe. The Severe value produces minimal log details, and Finest produces the most extensive amount of reporting. SQL Trace Possible values are True and False. Indicates whether the SQL statements to the DAC repository and database connections are added to the log file. Possible values are True and False. The True value sends a hint to the database connectivity layer of the DAC server to enable SQL tracing; thus, every SQL statement that is run by the DAC server is spooled to the appropriate output log file. It is recommended that you set this property to False. Test Run Possible values are True and False. When set to True, the execution plan will not stop on errors.
Verify and Create Non-Existing Indices Possible values are True and False. Indicates whether indexes defined in the DAC repository will be automatically created in the data warehouse database during an incremental load. When this system property is set to True, the DAC server verifies whether indexes defined in the DAC repository are also defined in the data warehouse database. This verification process can delay the execution of an execution plan.
12-28 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
You can install multiple Informatica servers and point them to a single Informatica Repository. You need to register each Informatica Server in the DAC and specify a unique machine name and server name. For instructions on registering an Informatica Server in the DAC, see the Oracle Business Intelligence Applications Installation and Configuration Guide.
Name Name of Informatica Server or Informatica Repository Server. Type Type of server.
Service For Informatica 7.x installations, indicates the host machine name where the Informatica Server is installed. For Informatica 8.x installations, indicates the Integration Service name. Server Port Port number used by the Informatica Server or Informatica Repository Server to listen to requests. Domain For Informatica 8.x installations only. Indicates the domain file location. Login Informatica Repository user login. Password Informatica Repository password. Maximum Sessions The maximum number of workflows that can be executed in parallel on the Informatica Server. Repository Name Informatica Repository name.
12-30 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
You deploy only one Informatica Repository Server, but you can deploy multiple Informatica Servers.
Connection Type Type of database connection. Possible values are the following:
Flat File
Oracle (OCI8), use the tnsnames entry. Oracle (Thin), use the instance name. SQL Server, use the database name. DB2-UDB/DB2-390, use the connect string as defined in the DB2 configuration. Teradata, use the database name.
12-32 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Table Owner Name of the table owner. Max Num Connections Maximum number of database connections this connection pool can contain. DBHost Host machine where the database resides. This field is mandatory if you are using Oracle (Thin), MSSQL, or Teradata, but is not required if you are using Oracle (OCI8), DB2, or DB2-390. Port Port where the database receives requests. Required for Oracle (Thin) and MSSQL databases. Not required for Oracle (OCI8), DB2, or DB2-390, or Teradata databases. Dependency Priority User-defined priority of the data source. Data Source Number User-defined number of the data source. Default Index Space Applicable to Oracle databases only. The default index space for the physical data source. When indexes are dropped and created, they are created in this index space. Num Parallel Indexes Use this field to specify how many indexes are to be created in parallel. For example, if a table with thirty indexes is the only task running in an execution plan at a given time, and you specify that 10 indexes are to be created in parallel, 10 of the 30 indexes will be created in parallel. The number of indexes that can be created in parallel is limited by the value you set in the Max Num Connections property. Parallel Index Creation You must select the Parallel Index Creation check box in order to specify a number in the Num Parallel Indexes field.
You must create the index space on the database before you specify an index space in the DAC.
Refresh dates for micro ETL processes are captured in the Refresh Dates subtab of the Execution Plans tab.
Name Name of source or target table. Execution Plan The name of the execution plan to which the source or target table belongs. Refresh Date The refresh date for the source or target table. Analyze Date Indicates when the table was analyzed. Number of Rows Valid for target tables only. Indicates the total number of rows in the table after the table has been loaded.
12-34 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Current Run Tab Execution Plans Tab Run History Tab Scheduler Tab
Start Timestamp Start time of the ETL process. Reflects the start time of every ETL attempt. For example, if the ETL fails and is run again, it gets a new start timestamp. The history of attempted runs is maintained in the audit trail for the run. This field is read only. End Timestamp End time of the ETL process. Reflects the end time of every ETL attempt. For example, if the ETL fails and is run again, it gets a new start timestamp. The history of attempted runs is maintained in the audit trail for the run. This field is read only. Duration A calculated field that shows the difference between start and end time stamps. Status Description Displays messages generated during run time. You can add notes to this field for Completed runs. Process ID ID for the process. This value is an integer that is incremented by 1 for every run. This value is stored as ETL_PROC_WID in all the data warehouse tables. This field is read-only. Total Number of Tasks The total number of tasks for this run. This field is read only.
12-36 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Number of Failed Tasks The sum total of tasks that have failed and that have stopped. This field is read only. Number of Successful Tasks The number of tasks whose status is Completed. This field is read only. Number of Tasks Still in Queue The number of tasks whose prerequisite tasks have not completed, and the number of tasks whose prerequisite tasks are completed and are waiting for resources. This field is read only. Schedule Name The name of the scheduled ETL process.
Tasks with no dependencies are executed first. If a task has failed and has been requeued, it gets the maximum priority. Tasks with greater phase priorities are executed next. When several tasks of the same phase are eligible to run, the tasks with greater task priorities are executed next. The prioritization is also based on the number of dependent tasks, the number of source tables, and the average time taken by a task.
12-38 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Drop/Create Indices Indicates indexes of the tables associated with this execution plan will be dropped and created. Run Now Button The Run Now button submits a request to the DAC server to execute the execution plan. Build Button The Build button does the following:
Collects all the tasks from all the subject areas. Substitutes the logical values with the physical values as defined in the parameters subtab (primary source/target connection and folder information). Removes any redundancies based on the task name and primary source/target connection information. Creates multiple instances of the same task to accommodate the number of copies of the parameters. Comes up with the ordered list of tasks with the dependencies computed among them.
For more information about building an execution plan, see "Building and Running Single-Source and Multi-Source Execution Plans".
Depth The level of the tasks dependency. Tasks that have no dependencies are depth 0. Tasks that depend on other tasks of depth 0 are depth 1, and so on. Task Name Name of immediate dependent task. Source Source table from which the task extracts data. Target Target table into which the task loads data.
Pre-ETL actions Upon Failure Restart actions Truncate table information List of indexes to be dropped and created Full or incremental command based on the current situation
DAC Functional Reference 12-41
About Parameter Management Defining a Text Type Parameter Defining a Database Specific Text Type Parameter Defining a Timestamp Type Parameter Defining a SQL Type Parameter
Copy Number Number of the data source. Type Possible values are the following:
Name Logical name of the folder or database connection. Value Physical name of the folder or database connection. Delay Indicates how many minutes an extract of a data source will be delayed after the start of a multiple source ETL. Source System Name of the source system associated with the parameter.
12-42 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Execution Priority Indicates the order in which the task is executed. If two or more tasks have the same priority, the DAC will execute them in parallel. Command for Incremental Load Command associated with the task. Source System Source system container from which the task extracts data.
12-44 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Scheduler Tab
The Scheduler tab enables you to schedule ETL processes to be executed either once at a later time or periodically. When you schedule an ETL or make changes to a schedule, the DAC server picks up the information from the DAC client. The DAC server polls the DAC repository for changes periodically at a frequency set in the DAC system properties. The top window of the Scheduler tab lists ETL runs that have been scheduled. The bottom window enables you to schedule an ETL run. Execution Plan The name of the scheduled execution plan. Last Schedule Status The last run status of the scheduled ETL process. Possible values are Running, Completed or Stopped. Next Trigger Time the scheduled ETL run will next be executed. Status Description Description of the last ETL run. Possible values are Running, Completed, or the reason the process stopped. Recurrence Indicates how often the schedule will be executed.
Scheduler Tab
12-46 Oracle Business Intelligence Data Warehouse Administration Console Users Guide
Index
A
actions about, 7-1 assigning to repository object, 7-5 defining SQL script for, 7-2 example, 7-8 functions, 7-5 overview of, 7-1 types, 7-1 assigning actions to repository object, 7-5 authentication file DAC, 4-3 introduction, 3-1 object ownership, 3-5 process life cycle, 3-3 quick start, 4-1 DAC repository command line options, 11-7 DAC server command line access, 11-3 handling failures, 11-18 running automatically, 11-3 running two on same machine, 11-11 data flow Online Analytical Processing (OLAP) database, about and diagram, 2-2 data warehouse architecture, 2-1 customizing, 8-2 overview, 2-1 Data Warehouse Administration Console (DAC) DAC features, 3-1 DAC window, 6-1 editable lists, 6-11 exporting metadata, 11-2 importing metadata, 11-1 menu bar, 6-2 navigation tree, 6-10 object ownership, 3-5 top pane toolbar, 6-5 user interface, 6-1 deleted records tracking, 11-16
B
best practices columns, 5-4 configuration tags, 5-5 execution plans, 5-5 indexes, 5-3 repository objects, 5-1 source system containers, subject areas, 5-5 tables, 5-3 task group, 5-3 tasks, 5-2
5-1
C
change capture about, 11-12 filter, 11-16 columns best practices, 5-4 configuration tags best practices, 5-5 working with, 8-13
E
ETL extract and load processes, 4-5 full load, 4-5 incremental load, 4-5 execution plan micro ETL, 9-10 monitoring processes, 9-14 scheduling, 9-13 execution plans about multi-source, 9-1 about single-source, 9-1 best practices, 5-5
D
DAC authentication file, 4-3 basic concepts, 3-1 first time logging in, 4-3 important features, 3-1
Index-1
building and running, 9-8 heterogeneous, about, 9-1 homogeneous, about, 9-1 micro ETL, 9-10 monitoring, 9-14 scheduling, 9-13 types of, 9-1 unit testing, 9-10 extract and load processes common scenarios, 9-3
P
Parameter Management about, 8-8 parameters at runtime, 8-9 data types, 8-8 defining database specific text type parameters, 8-11 defining SQL type parameters, 8-12 defining text type parameter, 8-10 defining timestamp type parameters, 8-11 nesting, 8-10 overview, 8-8 preconfigured, 8-9
F
flat views querying, 6-13 full load, 4-5
Q I
indexes adding to data warehouse, 8-5 best practices, 5-3 Informatica mappings, creating, 8-6 replacing workflow with SQL file, 11-9 server sessions, 11-9 Informatica repository importing objects, 8-5 Informatica server pointing multiple servers to single repository, 11-18 PowerCenter Services, 2-3 query functionality flat views querying, 6-13 query commands, 6-12 query operators, 6-12 query procedures, 6-13
R
refresh dates about, 9-13 repository objects best practices, 5-1 right-click menu common commands, 6-6 Design view commands, 6-7 Execute view commands, 6-9 Setup view commands, 6-9
L
log in DAC, 4-3
S
source system container about, 3-3 best practices, 5-1 copying, 8-1 creating, 8-1 DAC Repository objects, 3-4 SQL script defining, 7-2 subject area customizing, 8-1 designing, 8-17 subject areas best practices, 5-5
M
multi-source execution plans best practices multi-source execution plans, 9-6
O
object ownership, 3-5 Online Analytical Processing database Data Warehouse, data flow into, 2-2 Oracle Business Analytics Data Warehouse overview, 2-1 Oracle Business Analytics Warehouse adding columns, 8-3 adding indices, 8-5 adding new table, 8-3 architecture, 2-1 architecture components, 2-2 customizing, 8-2 overview, 2-1
T
tables adding new tables, 8-3 best practices, 5-3 truncate behavior, multi-source ETL, 9-4 truncate behavior, single-source ETL, 9-3 tablespaces specifying for indexes by table type, 8-13
Index-2
task group best practices, 5-3 creating, 8-7 task phase dependency setting, 8-7 Tasks creating tasks for workflows, 8-6 tasks best practices, 5-2
U
upgrade Upgrade/Merge Wizard, 10-1 Upgrade/Merge Wizard about, 10-1 Difference Report, 10-17 major stages of upgrade/merge, 10-1 Peer to Peer Merge option, 10-14 Refresh Base option, 10-10 Refresh Base upgrade option, 10-7 Replace Base option, 10-11 Repository Upgrade (DAC 784) option, 10-3 resetting, 10-2 resolving object differences, 10-17 upgrade/merge options, 10-3 user account about, 4-1 creating, deleting, inactivation, 4-2 management, 4-1
Index-3
Index-4