Oracle® Essbase Integration Services: System Administrator'S Guide
Oracle® Essbase Integration Services: System Administrator'S Guide
Oracle® Essbase Integration Services: System Administrator'S Guide
RELEASE 11.1.1
Integration Services System Administrators Guide, 11.1.1 Copyright 1998, 2008, Oracle and/or its affiliates. All rights reserved. Authors: EPM Information Development Team This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this software or related documentation is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, the following notice is applicable: U.S. GOVERNMENT RIGHTS: Programs, software, databases, and related documentation and technical data delivered to U.S. Government customers are "commercial computer software" or "commercial technical data" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, the use, duplication, disclosure, modification, and adaptation shall be subject to the restrictions and license terms set forth in the applicable Government contract, and, to the extent applicable by the terms of the Government contract, the additional rights set forth in FAR 52.227-19, Commercial Computer Software License (December 2007). Oracle USA, Inc., 500 Oracle Parkway, Redwood City, CA 94065. This software is developed for general use in a variety of information management applications. It is not developed or intended for use in any inherently dangerous applications, including applications which may create a risk of personal injury. If you use this software in dangerous applications, then you shall be responsible to take all appropriate fail-safe, backup, redundancy and other measures to ensure the safe use of this software. Oracle Corporation and its affiliates disclaim any liability for any damages caused by use of this software in dangerous applications. This software and documentation may provide access to or information on content, products and services from third parties. Oracle Corporation and its affiliates are not responsible for and expressly disclaim all warranties of any kind with respect to third party content, products and services. Oracle Corporation and its affiliates will not be responsible for any loss, costs, or damages incurred due to your access to or use of third party content, products or services.
Contents
Chapter 1. Using Essbase Integration Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 Integration Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 Starting Integration Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 Running Integration Server as a Background Process on UNIX Systems . . . . . . . . . . . 14 Integration Server Startup Switches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 -? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 -A . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 -B . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 -C . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 -D . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 -E . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 -F . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 -I . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 -L . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 -M . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 -N . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 -P . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 -Q . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 -R . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 -S . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 -T . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 -U . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 Storing Startup Information in the Configuration File . . . . . . . . . . . . . . . . . . . . . . . . 26 Storing Startup Information in the Startup File (Windows Only) . . . . . . . . . . . . . . . . 27 Additional Configuration Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 Windows Configuration Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 Windows and UNIX Configuration Parameters . . . . . . . . . . . . . . . . . . . . . . . . . 29 Reconnecting Integration Server and OLAP Metadata Catalog . . . . . . . . . . . . . . . . . . . . . 30 Shutting Down Integration Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 Integration Server Log File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 Items Logged in Integration Server File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
Contents
iii
Integration Server Log File Size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 Data Load Error File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 Unicode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 Unicode and Non-Unicode Application Modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 When to Use Unicode-Mode Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 Unicode-Enabled Administration Tools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 Enabling Regional Options for Multiple Languages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 Enabling Regional Options in Windows 2000 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 Enabling Regional Options in Windows XP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 Enabling Regional Options in Windows 2003 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 Creating Database User Aliases and Synonyms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 Scheduling Jobs in UNIX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 Starting the Windows Task Scheduler Service . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 Chapter 2. Configuring Data Sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 About Data Source Configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 Server and Client Software and ODBC Connections . . . . . . . . . . . . . . . . . . . . . . . . . 42 Supported ODBC Drivers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42 Configuring Data Source Names on Windows Systems . . . . . . . . . . . . . . . . . . . . . . . . . . 43 Accessing ODBC Data Source Administrator in Windows . . . . . . . . . . . . . . . . . . . . . 44 Configuring DataDirect Wire Protocol Drivers in Windows . . . . . . . . . . . . . . . . . . . 44 IBM DB2 UDB or DB2 OS/390 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 Informix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 Oracle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 Sybase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 Testing DataDirect Wire Protocol Drivers on Windows . . . . . . . . . . . . . . . . . . . . . . . 47 Adding a Configuration File Entry for IBM DB2 OS/390 in Windows . . . . . . . . . . . . . 48 Configuring the DataDirect Text Driver on Windows . . . . . . . . . . . . . . . . . . . . . . . . 49 Configuring the SQL Server ODBC Driver for Microsoft SQL Server . . . . . . . . . . . . . 51 Configuring the Teradata ODBC Driver for Teradata . . . . . . . . . . . . . . . . . . . . . . . . 53 Editing a Data Source Name . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54 Configuring Data Source Names on UNIX Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54 Configuring Environment Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 Editing the odbc.ini File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 Examples of ODBC Settings for IBM DB2 UDB . . . . . . . . . . . . . . . . . . . . . . . . . 57 Example of ODBC Settings for IBM DB2 OS/390 . . . . . . . . . . . . . . . . . . . . . . . . 58 Example of ODBC Settings for Informix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58 Example of ODBC Settings for Oracle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 Example of ODBC Settings for Flat File Data Source . . . . . . . . . . . . . . . . . . . . . . 59
iv
Contents
Example of ODBC Settings for Microsoft SQL Server . . . . . . . . . . . . . . . . . . . . . 61 Example of ODBC Settings for Sybase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 Examples of ODBC Settings for Teradata . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 Creating IBM DB2 Bind Packages for DataDirect Wire Protocol Driver Connections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 Adding a Configuration File Entry for IBM DB2 OS/390 . . . . . . . . . . . . . . . . . . . . . . 64 After Configuration of Data Source Names . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65 Connecting to Server Components and Data Sources . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 Connecting to Integration Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 Connecting to Essbase Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 Setting Connection Defaults . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68 Connecting to Data Sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69 Mapping JDBC Data Sources for Windows (Teradata Only) . . . . . . . . . . . . . . . . . . . . . . 72 Mapping JDBC Data Sources for UNIX (Teradata Users Only) . . . . . . . . . . . . . . . . . . . . 73 Manually Configuring the Environment for UNIX Systems . . . . . . . . . . . . . . . . . . . . . . . 75 Chapter 3. Creating, Upgrading, and Deleting OLAP Metadata Catalogs . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 About OLAP Metadata Catalogs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 Creating or Upgrading OLAP Metadata Catalogs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78 Creating Databases for OLAP Metadata Catalog Tables . . . . . . . . . . . . . . . . . . . . . . . 78 Creating or Upgrading OLAP Metadata Catalogs Automatically . . . . . . . . . . . . . . . . . 79 Creating OLAP Metadata Catalogs Manually . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80 Upgrading OLAP Metadata Catalogs Manually . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81 SQL Scripts Used to Create and Upgrade Tables . . . . . . . . . . . . . . . . . . . . . . . . . 82 Proper Order of SQL Scripts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84 Upgrading Tables in the OLAP Metadata Catalog . . . . . . . . . . . . . . . . . . . . . . . . 85 Mapping OLAP Metadata Catalogs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85 Deleting OLAP Metadata Catalogs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86 Chapter 4. Tables in OLAP Metadata Catalog . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 OLAP Metadata Catalog . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 Tables Relating to the OLAP Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 Tables Relating to the Metaoutline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 Tables Relating to Drill-Through . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 Tables Relating to Hybrid Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 Miscellaneous Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92 Chapter 5. Setting Up the Sample Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 Overview of Sample Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 Preparing to Set Up the Standard Sample Application . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
Contents
Setting Up the Standard Sample Application Automatically . . . . . . . . . . . . . . . . . . . . . . . 95 Setting Up the Standard Sample Application Manually . . . . . . . . . . . . . . . . . . . . . . . . . . 97 Setting Up the TBC Relational Data Source . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98 Creating the TBC Database . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98 Creating Tables for the TBC Database . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99 Loading Data into the TBC Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100 Setting Up the TBC_MD OLAP Metadata Catalog . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 Creating the TBC_MD Database . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 Creating Tables for the TBC_MD Database . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 Using the XML Import Utility to Load Data into the TBC_MD Tables . . . . . . . . . . . 104 Using XML Import to Load Metadata into TBC_MD . . . . . . . . . . . . . . . . . . . . . . . 104 Setting Up the Hybrid Analysis Sample Application . . . . . . . . . . . . . . . . . . . . . . . . . . . 106 Setting Up the Hybrid Analysis Sample Application on IBM DB2 . . . . . . . . . . . . . . . 106 Running the Batch File on IBM DB2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 Importing the Sample OLAP Model and Metaoutline . . . . . . . . . . . . . . . . . . . . 108 Setting Up Hybrid Analysis Sample Application on Oracle . . . . . . . . . . . . . . . . . . . . 108 Running the Batch File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 Importing the Sample OLAP Model and Metaoutline . . . . . . . . . . . . . . . . . . . . 110 Setting Up Hybrid Analysis Sample Application on Microsoft SQL Server . . . . . . . . . 111 Running the Batch File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111 Importing the Sample OLAP Model and Metaoutline . . . . . . . . . . . . . . . . . . . . 112 Setting Up Hybrid Analysis Sample Application on Teradata . . . . . . . . . . . . . . . . . . 113 Running the Batch File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113 Importing the Sample OLAP Model and Metaoutline . . . . . . . . . . . . . . . . . . . . 115 Setting Up the Unicode Sample Application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115 Creating the Unicode TBC_U Database . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116 Creating the Unicode TBC_MD_U OLAP Metadata Catalog . . . . . . . . . . . . . . . . . . 116 Setting Up the Unicode Sample Application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116 Running the SQL File . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117 Importing the Unicode Sample Model and Metaoutline . . . . . . . . . . . . . . . . . . 117 After Setting Up the Sample Application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118 Viewing TBC Tables and Columns . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118 Viewing OLAP Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119 Viewing Metaoutlines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119 Chapter 6. Working with Users, Locks, and Permissions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121 Working with Users . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121 Working with Locks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122 Using Standard Access Mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
vi
Contents
Using Exclusive Access Mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122 Correcting Problems with Locks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 Viewing Integration Services Users with Locks . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 Deleting Locks for Integration Services Users . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124 Working with Permissions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125 Viewing Permissions for OLAP Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125 Viewing Permissions for Metaoutlines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126 Chapter 7. Troubleshooting ODBC and Connections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129 Common Connection Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129 Database-Specific Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131 ODBC and Connection Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131 Isolating Connection Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132 Testing the Database Client Connection to the Data Source . . . . . . . . . . . . . . . . . . . 134 Testing the ODBC Connection to the Data Source . . . . . . . . . . . . . . . . . . . . . . . . . 135 Testing ODBC on Windows Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135 Testing ODBC on UNIX Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136 Testing the Integration Server Connection to the Data Source . . . . . . . . . . . . . . . . . 139 Testing the Integration Services Console Connection to the Data Source . . . . . . . . . 141 Testing the Integration Server Connection to Essbase Server . . . . . . . . . . . . . . . . . . 142 Testing the Integration Services Console Connection to Essbase Server . . . . . . . . . . . 143 Correcting Connection Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144 User Name and Password Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145 Correcting Data Source Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146 Correcting Database Client Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146 Correcting ODBC Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148 Correcting Integration Server Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149 Correcting Integration Services Console Problems . . . . . . . . . . . . . . . . . . . . . . . . . 150 Using ODBC Tracing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151 Using ODBC Tracing on Windows Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151 Using ODBC Tracing on UNIX Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151 Chapter 8. Using Integration Services Shell . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153 Starting Integration Services Shell . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153 Integration Services Shell . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154 SET and LOAD Commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 155 Informational Commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156 Command Syntax and Execution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156 Recording Member and Data Load Scripts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157 Running Integration Services Shell Script and Batch Files . . . . . . . . . . . . . . . . . . . . 157
Contents
vii
Running OLAP Command Script Files . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 158 Integration Services Shell Commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159 LOGIN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159 STATUS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160 VERSION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160 SETLOCALE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161 SETSOURCE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161 SETTARGET . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162 SETCATALOG . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162 LOADMEMBER . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163 LOADDATA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166 LOADALL . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 168 LOGOUT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172 SHUTDOWN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172 EXIT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172 Peforming Member Loads Using Integration Services Shell . . . . . . . . . . . . . . . . . . . . . . 172 Loading Data Using Integration Services Shell . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 173 Chapter 9. Naming Restrictions for Essbase Applications, Databases, and Members . . . . . . . . . . . . . . . . . . 175 Naming Restrictions for Applications and Databases . . . . . . . . . . . . . . . . . . . . . . . . . . . 175 Naming Restrictions for Dimensions, Members, and Aliases . . . . . . . . . . . . . . . . . . . . . 176 Using Dimension and Member Names in Calculation Scripts, Report Scripts, Formulas, and Filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 178 Appendix A. Return Codes and Error Messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 179 OLAPICMD Session Return Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 179 Member Load Error Messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180 Data Load Error Messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183 Drill-Through Report Error Messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184 Miscellaneous Error Messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 185 Essbase Error Messages Generated During Data Loads . . . . . . . . . . . . . . . . . . . . . . . . . . 187 Appendix B. Integration Services Limits and Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189 Artifact Limits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189 Source Database Artifact Naming Limits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192 Setting the Compression Dimension . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193 Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193 Hybrid Analysis Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 194 Data Source Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 194 Dimensions Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 194
viii
Contents
Members Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 194 Spreadsheet Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195 Operations Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195 Limitations on Transformation Rules with Hybrid Analysis . . . . . . . . . . . . . . . . . . . 195 Transformations Not Supported by Hybrid Analysis . . . . . . . . . . . . . . . . . . . . . . . . 196 Limitations on Using Formulas with Hybrid Analysis . . . . . . . . . . . . . . . . . . . . . . . 196 Unsupported Essbase Functions in Hybrid Analysis . . . . . . . . . . . . . . . . . . . . . 196 Using Advanced Relational Access . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197 Advanced Relational Access Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 198 General Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 198 Data Source Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199 Dimension Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199 Member Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199 Unsupported Data Types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200 Essbase Databases and Advanced Relational Access . . . . . . . . . . . . . . . . . . . . . . . . . 200 Different Values Loaded in Ragged Hierarchies . . . . . . . . . . . . . . . . . . . . . . . . . 200 Drill-through Report Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201 Unicode Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202 Duplicate Member Names Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202 Dimensions in Duplicate Outlines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202 Members in Duplicate Outlines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203 Other Guidelines for Duplicate Outlines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203 Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205 Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 229
Contents
ix
Contents
1
In This Chapter
Integration Server ..................................................................................................................11 Starting Integration Server ........................................................................................................12 Reconnecting Integration Server and OLAP Metadata Catalog ................................................................30 Shutting Down Integration Server ................................................................................................30 Integration Server Log File.........................................................................................................31 Data Load Error File................................................................................................................32 Unicode .............................................................................................................................33 Enabling Regional Options for Multiple Languages .............................................................................35 Creating Database User Aliases and Synonyms ................................................................................37 Scheduling Jobs in UNIX ..........................................................................................................38 Starting the Windows Task Scheduler Service...................................................................................38
This chapter describes starting and stopping Essbase Integration Server and how to view its log file.
Integration Server
Integration Server is multithreaded server software that is the bridge between the data source, OLAP Metadata Catalog, and Essbase Server. Oracle Essbase Integration Services performs several tasks:
Retrieves OLAP model and metaoutline information from OLAP Metadata Catalog Generates SQL statements Retrieves data from external sources Loads members and data into the Oracle Essbase database
Integration Server must be running if you want to use Integration Services Console (the graphical user interface) or Integration Services Shell (the command line interface).
Integration Server
11
The relational database management system (RDBMS) with the relational data source containing the OLAP Metadata Catalog where the metadata is stored One or more RDBMSs containing the relational data sources you want to use to create or modify an OLAP model or metaoutline
Essbase Server must be running if you create, change, or load data into an Essbase database or if you want to preview an Essbase database outline. You can start Integration Server using several methods:
From a DOS command prompt From a UNIX command prompt From the Windows Desktop Start menu As a Windows service
Note:
To start Integration Server as a Windows service, you must have configured it as a Windows service during the configuration process. See the Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide.
2 If you want to change a default startup switch, add the switch to the ais.cfg file or the
startup.bat file.
3 If you want to change a default configuration parameter, add the parameter to the ais.cfg file.
See Storing Startup Information in the Configuration File on page 26 and Storing Startup Information in the Startup File (Windows Only) on page 27 for information on adding startup switches and configuration parameters to files.
12
Note:
If Integration Server does not start when olapisvr is executed from the command line, the operating system path may not be updated correctly. See the Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide.
3 If you want to change a default startup switch, type the command syntax after the olapisvr command or
add the switches to the ais.cfg file.
See Storing Startup Information in the Configuration File on page 26 and Storing Startup Information in the Startup File (Windows Only) on page 27 for information on adding startup switches to these files. Values for startup switches entered at the command line override startup switch values stored in the ais.cfg file. When entering parameters for startup switches at the command line, enclose path names that contain spaces in quotation marks (" "). The following example illustrates the syntax for passing the -E and -M options:
olapisvr -E"c:\my logs\eislog" -M"c:\ess\bin\essbase.mdb"
On Windows 2000 and Windows 2003, select Start, then Settings, then Control Panel, double-click Administrative Tools, and then double-click Services. On Windows XP, select Start, then Control Panel. In the Control Panel, select Performance & Maintenance, and then Administrative Tools. In the Administrative Tools window, double-click Services. The Services dialog box is displayed with Essbase Integration Server in the list of services.
2 Select, but do not double-click, Essbase Integration Server. 3 If you want to change any default startup switches or configuration parameters, add the switches to the
ais.cfg file or the startup.bat file.
On Windows 2000 and 2003, select Action, then Start. On Windows XP, click Start. The Service Control message box is displayed. After a few seconds it is removed, and the Status field entry for Integration Server is changed to Started.
13
Note:
If you choose to run Integration Server by using the above commands, ensure you have configured your environment. The nohup command keeps the Integration Server program running even if you log off. The ampersand (&) suffix makes the program a background process.
Caution!
To allow the Integration Services Console client to create OLAP models and metaoutlines, you must run Integration Server and keep it running while any clients are connected to it.
When you start Integration Server, it begins writing to the log, olapisvr.log, in the log directory. To write to a log on another directory, specify the log name in ais_start or type the following command:
olapisvr -Emydir/mylog
Note:
Do not type the .log file extension. The .log extension is automatically appended to the name that you type for the log.
-? -A -C -D -E
14
-F -I [K] -L -M -N -P -R -S -T -U [V]
For information on storing startup switches in startup.bat or ais.cfg, see Storing Startup Information in the Configuration File on page 26 and Storing Startup Information in the Startup File (Windows Only) on page 27. The following topics describe the default switches.
-?
To view a list of available switches for starting Integration Server.
-A
To ignore the parent of a given member when updating OLAP intersections during a drillthrough operation. To ignore parents when updating OLAP intersections, do one of the following:
In Windows, add -AY to the startup.bat file. In UNIX, type -AY when you start Integration Server. In Windows or UNIX, add [A]=Y to the ais.cfg file.
By default, Integration Server sets the value at N so that parents are not ignored when updating OLAP intersections. If a parent is ignored during an update of OLAP intersections, subsequent drill-through operations cannot be performed on any shared members related to that parent. In Windows: In the startup.bat file, to ignore the parent of a given member when updating OLAP intersections during a drill-through operation, add -AY after "C:\Hyperion\products \Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -AY
In UNIX:
15
To ignore the parent of a given member when updating OLAP intersections during a drillthrough operation, type:
olapisvr -AY
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[A]=Y
-B
Note:
-B is no longer available.
-C
To set the number of records that Integration Server commits to Essbase:
In Windows, add -Cnumber_of_records to the startup.bat file. In UNIX, type -Cnumber of records when you start Integration Server. In Windows or UNIX, add [C]=number_of_records to the ais.cfg file.
The default setting is all records. -C can be used with the -N switch.
Note:
-C is often used in testing environments. The recommended setting for testing purposes is 5,000 to 10,000 records. After you have completed testing, shut down Integration Server to revert to the default setting, or reset to the limit allowed by DATAERRORLIMIT. When -C is used with the DATAERRORLIMIT setting in the essbase.cfg file, you can maximize the number of records written to dataload.txt, the data load error file. For example, if you set DATAERRORLIMIT to 65,000, the maximum number of error records allowed in the Essbaselog, and then set -C to 65,000 or less, you will be able to view all error records allowed by Essbase.
Note:
If you have included a setting for testing purposes for -C in ais.cfg, be sure to edit the file to clear the ais.cfg file so that it will change the setting back to all records.
16
Setting a smaller number of records during testing enables you to see data load errors incrementally, providing you an opportunity to correct problems before continuing. Resetting the number of records to the default of all records results in faster performance. There is no limit to the number of records that Integration Server can transfer to Essbase during a data load. If, however, you experience memory problems during a data load, set -C to a smaller value.
Note:
In Windows: For example, in the startup.bat file, to set the number of records that Integration Server incrementally transfers to Essbase during a data load to 5,000, add -C5000 after "C: \Hyperion\products\ Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -C5000
In UNIX: For example, to set the number of records that Integration Server incrementally transfers to Essbase during a data load to 5,000, type:
olapisvr -C5000
In Windows or UNIX: For example, to set the number of records that Integration Services incrementally transfers to Essbase during a data load to 5,000, in the ais.cfg file, type:
[C]=5000
-D
To specify whether or not to add the DISTINCT clause to the SELECT statement of drill-through SQL. The default behavior adds the DISTINCT clause. To not add the DISTINCT clause:
In Windows, add -DN to the startup.bat file. In UNIX, type -DN when you start Integration Server. In Windows or UNIX, add [D]=N to the ais.cfg file.
In Windows: In the startup.bat file, to specify that the DISTINCT clause should not be added to drillthrough queries, add -DN after "C:\Hyperion\products\Essbase\eis\server\bin \olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -DN
17
Any value other than -DN or no value specifies that the DISTINCT clause should always be added to drill-through SQL. In UNIX: To specify that the DISTINCT clause should not be added to drill-through queries, type:
olapisvr -DN
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[D]=N
-E
To give the log file a different name:
In Windows, add -Elog_file_name to the startup.bat file. In UNIX, type -Elog_file_name when you start Integration Server. In Windows or UNIX, add [E]=log_file_name to the ais.cfg file.
The .log extension is added automatically. The default name and location is HYPERION_HOME \logs\eis\olapisvr.log. In Windows: In the startup.bat file, to name the log file myserver.log and put it in the temp directory on Windows, add -Ec:\temp\myserver after "C:\Hyperion\products\Essbase\eis \server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe -E"c:\temp \myserver
Do not type the .log file extension. The .log extension is automatically appended to the name you type for the log. Always enclose path names that contain spaces with quotation marks (" "); for example, type:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -I -E"c: \program files\eis\myserver"
In UNIX: To name the log file myserver.log and put it in the temp directory on Windows, type
olapisvr -Ec:\temp\myserver
Enclose path names that contain spaces in quotation marks (" "); for example, type:
olapisvr -E"c:\program files\eis\myserver"
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[E]=c:\temp\myserver
or
18
[E]=c:\program files\eis\myserver
Note:
In the ais.cfg file, you do not need to type quotation marks around path names that contain spaces.
-F
To specify whether to add parentheses around user-defined drill-through filters. See Essbase Spreadsheet Add-in help. The default is not to add parentheses around user-defined drill-through filters. To add parentheses:
In Windows, add -FY to the startup.bat file. In UNIX, type -FY when you start Integration Server. In Windows or UNIX, add [F]=Y to the ais.cfg file.
In Windows: In the startup.bat file, to specify that parentheses should be added around user-defined drillthrough filters, add -FY after "C:\Hyperion\products\Essbase\eis\server\bin \olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -FY
Any other value or no value specifies that parentheses should not be added around user-defined drill-through filters. In UNIX: To specify that parentheses should be added around user-defined drill-through filters, type:
olapisvr -FY
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[F]=Y
-I
To prevent Integration Server from shutting down as a service on Windows after you log off, add -I to the startup.bat file. When you run Integration Server as a service on Windows, Oracle recommends you use -I.
Note:
19
Note: -I cannot be entered in the ais.cfg file. It can only be specified in the startup.bat file.
In Windows: In the startup.bat file, to prevent Integration Server from shutting down as a service, add -I after "C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -I
-L
To set the level of detail of the messages that Integration Server logs:
In Windows, add -Llevel to the startup.bat file. In UNIX, type -Llevel when you start Integration Server. In Windows or UNIX, add [L]=I to the ais.cfg file.
0 To log all debug messages and the ODBC SQL generated by Integration Server; generate the dataloadrecord.txt file, listing the first 1,000 records, which may include successfully loaded records and rejected records; and create the hisdld.rul file consisting of the names of the dimensions loaded followed by the leaf node number of each dimension or member
2 To log all informational messages and some of the ODBC SQL generated by Integration Server
7 To log all critical errors where Integration Server terminates the command
The recommended level setting is 2 or 3. In Windows: In the startup.bat file, to use the highest level of logging, add -L0 after "C:\Hyperion \products\Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -L0
20
In Windows or UNIX: In the ais.cfg file, this switch is specified in the following format:
[L]=0
-M
To specify the location of the Essbase message database file if the file is moved to a new location:
In Windows, add -Mpath\essbase.mdb to the startup.bat file. In UNIX, type -Mpath\essbase.mdb when you start Integration Server. In Windows, add [M]=path\essbase.mdb to the ais.cfg file.
The default location is HYPERION_HOME\products\Essbase\eis\server\bin \essbase.mdb. In Windows: In the startup.bat file, to specify the new location of the Essbase message database file as essbasemsg\bin, add -Mc:\essbasemsg\bin\essbase.mdb after "C:\Hyperion \products\Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -Mc:\essbasemsg \bin\essbase.mdb
Always enclose path names that contain spaces with quotation marks (" "); for example, type:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -M"c:\eis svr \bin\essbase.mdb"
Enclose path names that contain spaces with quotation marks (" "); for example:
olapisvr -M"c:\eis svr\bin\essbase.mdb"
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[M]=c:\eissvr\bin\essbase.mdb
or
[M]=c:\ess svr\bin\essbase.mdb
21
Note:
In the ais.cfg file, you do not need to add quotation marks around path names that contain spaces.
-N
To specify the number of threads Integration Server uses when sending data to Essbase during a data load, do one of the following:
In Windows, add -N to the startup.bat file. In UNIX, type -Nnumber_of_threads when you start Integration Server. In Windows or UNIX, add [N]=number_of_threads to the ais.cfg file.
The default setting is 1 thread. -N may be used with -C. -N controls the number of threads allocated to data load optimization. Usually, users start with 2 threads and adjust according to the environment. In Windows: In the startup.bat file, to set the number of threads to 2, add -N2 after "C:\Hyperion \products\Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -N2
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[N]=2
-P
To set the TCP port number with which Integration Server communicates with its clients to a port different from the default 3388:
In Windows, add -Pportnumber to the startup.bat file In UNIX, type -Pportnumber when you start Integration Server In Windows or UNIX, add [P]=portnumber to the ais.cfg file
In Windows: In the startup.bat file, add -Pportnumber after "C:\Hyperion\products\Essbase\eis \server\bin \olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -P8850
22
If you change the default port number, in Integration Services Console you must log in to the Integration Server by typing the server name or IP address and the non-standard port number, separated by a colon; for example:
aspen:8850
When starting Integration Services Shell, to change the TCP port number, at the command line, type -Pportnumber. For example:
olapicmd -P8850
In UNIX: When starting Integration Server or Integration Services Shell, to change the TCP port number, type:
olapisvr -P8850
or
olapicmd -P8850
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[P]=8850
-Q
Note:
-Q is no longer available.
-R
To specify that users are restricted from browsing table records when customizing a drill-through report.
When -R is not set, users have unrestricted access to all table records. When -R is set to Y, users have no access to any table records. When -R is set to F, users have no access to fact table records but have unrestricted access to all other table records.
In Windows, add -Raccess_parameter to the startup.bat file. In UNIX, type -Raccess_parameter when you start Integration Server. In Windows or UNIX, add [R]=access_parameter to the ais.cfg file.
By default, -R is not set, giving users unrestricted access to all table records. In Windows:
23
In the startup.bat file, to prevent users from accessing any table records, add -RY after "C: \Hyperion\products\Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -RY
In the startup.bat file, to prevent users from accessing fact table records, add -RF after "C: \Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -R. For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -RF
In UNIX: To prevent users from having any access to any table records, type:
olapisvr -RY
To prevent users from having any access to fact table records, type:
olapisvr -RF
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[R]=Y
or
[R]=F
-S
To specify the stack size in megabytes of each thread that Integration Server creates::
In Windows, add -Sstacksize_in_megabytes to the startup.bat file. In UNIX, type -Sstacksize_in_megabytes when you start Integration Server. In Windows or UNIX, add [S]=stacksize_in_megabytes to the ais.cfg file.
For Windows and Solaris, no stack size setting is required. For AIX, the default stack size setting is 3 MB; for HP-UX, the default stack size setting is 8 MB. In Windows: In the startup.bat file, to set the stack size that Integration Server creates to 1 MB, add -S1 after "C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -S1
In UNIX: To set the stack size that Integration Server creates to 2 MB, type:
olapisvr -S2
In Windows or UNIX: In the ais.cfg file, this parameter is specified in the following format:
[S]=2
24
-T
To set the number of network listeners that Integration Server starts:
In Windows, add -Tnumber to the startup.bat file In UNIX, type -Tnumber when you start Integration Server In Windows or UNIX, add [T]=number to the ais.cfg file
The default (and recommended) number is 10. Network listeners receive requests from Integration Services Console. Integration Server automatically adds and subtracts listeners as needed, so the number of listeners set at startup does not limit the number of users that can connect to an Integration Server. In Windows: In the startup.bat file, to set the number of network listeners that Integration Server starts to 17, add -T17 after "C:\Hyperion\products\Essbase\eis\server\bin \olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -T17
In UNIX: To set the number of network listeners Integration Server starts to 17, type:
olapisvr -T17
-U
To specify whether the Essbase database and application should be unloaded from memory after a load is completed. Integration Server can successfully unload the Essbase database from memory only if there are no users accessing it. To specify whether the Essbase database and application should be unloaded from memory after a load is completed, do one of the following:
In Windows, add -UY to the startup.bat file. In UNIX, type -UY when you start Integration Server. In Windows or UNIX, add [U]=Y to the ais.cfg file.
By default, Integration Server does not unload the Essbase database from memory after a load is completed. In Windows:
25
In the startup.bat file, to specify whether the Essbase database and application should be unloaded from memory after a load is completed, add -UY after "C:\Hyperion\ products \Essbase\eis\server\bin\olapisvr.exe". For example:
"C:\Hyperion\products\Essbase\eis\server\bin\olapisvr.exe" -UY
In UNIX: To specify whether the Essbase database and application should be unloaded from memory after a load is completed, type
olapisvr -UY
As an environment for Teradata and Oracle users to configure the required JDBC drivers, as discussed in Mapping JDBC Data Sources for Windows (Teradata Only) on page 72 and Mapping JDBC Data Sources for UNIX (Teradata Users Only) on page 73. To pass startup switches and configuration parameters, other than the defaults, when launching Integration Server
The ais.cfg file is a text file residing in the \bin directory on Windows, UNIX, and Linux platforms. When you start Integration Server from the command line using the startup.bat command (in DOS or UNIX), the startup routine checks the contents of ais.cfg for any startup switch or parameter information. Then one of the following applies:
Startup switch or configuration parameter syntax added to the ais.cfg file override default settings (see Integration Server Startup Switches on page 14 and Storing Startup Information in the Configuration File on page 26). If you did not add startup switch or parameter information to ais.cfg, the default settings are used to start Integration Server. If ais.cfg contains startup switch syntax and you enter startup switch overrides in the startup.bat file (Windows only) any startup switches in startup.bat override both the startup switch settings contained in the ais.cfg file and the default settings.
3 Add the new information for startup switches in the following format:
[L]=0 [E]=c:\temp\myserver
26
[N]=40 [M]=c:\ess\bin\essbase.mdb
As an environment to configure the common Oracle Hyperion Enterprise Performance Management System components, such as Java, ODBC, and JDBC drivers, for Teradata and Oracle users. See the Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide. As a means to pass startup switches and configuration parameters, other than the defaults, when launching Integration Server
The startup.bat file is a text file residing in the eis\server\bin directory. When you start Integration Server as a Windows service, the startup routine checks the contents of startup.bat for any startup switch information. Then the startup routine checks ais.cfg for any startup switch and configuration parameter information. Then one of the following applies:
Startup switch syntax added to the startup.bat file (Windows only), override default settings (see Integration Server Startup Switches on page 14). If you added startup switch or configuration parameter syntax only to the ais.cfg file, those settings that you added override the default settings for those switches and parameters (see Integration Server Startup Switches on page 14 and Storing Startup Information in the Configuration File on page 26). If you did not add startup switch or parameter information to startup.bat or ais.cfg, the default settings (see Integration Server Startup Switches on page 14 and Storing Startup Information in the Configuration File on page 26) are used to start Integration Server.
The values for startup switches that you enter into startup.bat override any startup switch values you have stored in the ais.cfg file.
27
When entering parameters for startup switches in startup.bat, enclose path names that contain spaces in quotation marks (" "). The following example illustrates the syntax for passing the -L, -E, -M, -N options:
"C:\Hyperion\eis serv\bin\olapisvr.exe" -I -L0 -Ec:\temp\myserver -N40 -M"c:\ess\bin\essbase.mdb"
Note:
In the startup.bat file, the above syntax should be entered all on one line. You can only enter startup switch information into startup.bat. Configuration parameters must be entered in ais.cfg.
[K]
To specify whether to display the primary keys in an OLAP model. The default behavior is not to display the primary keys in an OLAP model. If you want to display the primary keys in an OLAP model, add [K]=Y to the ais.cfg file. In Windows: In the ais.cfg file, this parameter is specified in the following format:
[K]=Y
[V]
To specify whether Integration Server automatically validates an OLAP model or metaoutline when a Save or Save As operation is performed. The default behavior is for Integration Server to validate an OLAP model or metaoutline before saving it. If you want to disable automatic OLAP model and metaoutline validation, add [V]=0 to the ais.cfg file.
28
In Windows: In the ais.cfg file, this parameter is specified in the following format:
[V]=0
[ADDDUPLICATESFORASO]
To allow addition of duplicate data records when using an aggregate storage database. The default behavior is to set the data load option in aggregate storage to not allow duplicate records. To allow duplicate data records when using an aggregate storage database, add [ADDDUPLICATESFORASO]=Y to the ais.cfg file. In Windows and UNIX: In the ais.cfg file, this parameter is specified in the following format:
[ADDDUPLICATESFORASO]=Y
[UNIFORMVALUEFORMAT]
To obtain correct results when an aggregate storage database contains duplicate records with different values. In the following example, duplicate records at the intersection of SKU, MONTH, and STATE show different results for SALES:
SKU 10010 10010 MONTH JAN JAN STATE CA CA SALES 389.99999999999994 390
When a database contains duplicate records with different values, data loads are terminated with an error; for example:
IS Error(EssLoadBufferTerm): Data load failed: input contains different values for the same cell [(Jan, Sales, 100-20, Florida): 389. 99999999999994 / 390] 1270089
When you add the [UNIFORMVALUEFORMAT] parameter to the ais.cfg file, data load results for measures are rounded up or down as appropriate to six decimal places (0.000000), ensuring that the duplicate records will have the same value. Using the same data from the example above, the results would be:
SKU 10010 10010 MONTH JAN JAN STATE CA CA SALES 390.000000 390.000000
To enable uniform value formatting, add [UNIFORMVALUEFORMAT]=Y to the ais.cfg file In Windows and UNIX:
29
Always close Integration Services Console before shutting down Integration Server.
To shut down Integration Server from the command line, either close the Integration Server
window, or use Integration Services Shell:
On Windows platforms, close the Integration Server window. Issue the SHUTDOWN command in Integration Services Shell.
On Windows 2000 and 2003, double-click Administrative Tools, and then double-click Services.
30
On Windows XP, select Performance & Maintenance, and then Administrative Tools. In the Administrative Tools window, double-click Services.
3 Locate Essbase Integration Server in the list of services. 4 Stop the service.
On Windows 2000 and 2003, select Action, then Stop. On Windows XP, right-click Essbase Integration Server and select Stop.
Processing messages Member load error messages Settings for switches used to start Integration Server from the command line or as a Windows service Two copies of each SQL statement generated to access the data source as follows:
The first statement is in the SQL dialect understood by the ODBC driver and is generated by Integration Server. The second statement is generated by the ODBC driver and is translated by the driver into the dialect of SQL understood by the data source being accessed.
Integration Services Shell commands used to perform an action Connection information Load status Names of any metaoutline running Other informational messages
When a user accesses a drill-through report using Essbase Spreadsheet Add-in, items related to the drill-through report are also logged:
Drill-through report name Metaoutline with which this drill-through report is associated Essbase Server computer being accessed
31
See Integration Server Startup Switches on page 14. Integration Server error message text is located in HYPERION_HOME\products\Essbase\eis \server\bin\error.txt on Windows systems; $HYPERION_HOME/products/Essbase/ eis/server/bin/error.txt on UNIX. Essbase error message text is located in HYPERION_HOME\products\Essbase\eis\server \bin\message.txt on Windows systems; $HYPERION_HOME/products/Essbase/eis/ server/bin/message.txt on UNIX.
If you performed a data load for the MyTBC application and MyTBC_DB database at 10 P.M. on May 1, 2003, the folder is named:
MyTBC_MyTBC_DB_2003_May_1_10_00_pm_<sessionnumber>
In the case of data load failure, open the dataload.txt file located in the folder described previously and review the error codes. These are the most common error codes:
32
Message Bad data value supplied. Record rejected because of duplicate member names. Member/Data Unknown. Record rejected because of dimension conflicts with Header Name.
Figure 1 shows three records from the dataload.txt file with the error code 3303.
Figure 1
Unicode
Sharing data across national and language boundaries is a challenge for multi-national businesses. Traditionally, each computer stores and renders text based on its locale specification. A locale identifies the local language and cultural conventions such as the formatting of currency and dates, sort order of the data, and the character set encoding to be used on the computer. The encoding of a character set refers to the specific set of bit combinations used to store the character text as data, as defined by a code page or an encoding format. In Essbase, code pages map characters to bit combinations for non-Unicode encodings. Because different encodings can map the same bit combination to different characters, a file created on one computer can be misinterpreted by another computer that has a different locale. The Unicode Standard was developed to enable computers with different locales to share character data. Unicode provides encoding forms with thousands of bit combinations, enough to support the character sets of multiple languages simultaneously. By combining all character mappings into a single encoding form, Unicode enables users to correctly view character data created on computers with different locale settings. Users whose computers are set up in different languages can work with the same database. For example, using alias tables in their respective languages, users in Taiwan can view database reports displaying Chinese characters while users in France can view the same reports in French characters. User-defined character sets (UDC) are not supported and the Chinese National Standard GB 18030-2000 is not supported.
Note:
For information on using Unicode in Essbase, see the Oracle Essbase Database Administrator's Guide.
Unicode
33
You need to enable users with different languages to view, in their own languages and character sets, information from a common database. For example, using alias tables in Japanese and German, users in Japan and Germany could view, in their own languages, information about a common product set. You need to handle artifact names longer than non-Unicode-mode applications support. For example, application and database names need to be larger than eight characters or, if you are working with a multi-byte character set, you need to handle more characters in artifact names. You have experienced what is called the round-trip problem. The round-trip problem can occur in communications between multi-byte operating systems and application programs where two different bit values can map to the same character. As Java applications, Oracle Essbase Administration Services and Oracle Hyperion Provider Services always work in Unicode. No encoding conversions occur when these clients work with Unicode-mode applications and UTF-8-encoded text files; hence no round-trip conversion errors.
When deciding on using Unicode-mode applications, you should also consider the following points:
Using non-Unicode text files with Unicode-mode applications requires an understanding of locales and care in managing to them. To prevent errors that could cause database corruption, using UTF-8-encoded files is recommended. For details, see the Oracle Essbase Database Administrator's Guide.
34
To work with Unicode-mode applications, custom client applications that were written to support non-Unicode-mode applications must be built to use the longer string lengths used by Unicode-mode applications. This may be a simple re-build or may involve reprogramming, depending on the design of the applications. Also, depending on how they are coded, the new client applications may require more memory.
2 In the Your locale (location) drop-down list, select the appropriate locale.
In most cases, you set the locale to match your location.
3 In the Language settings for the system frame, select the languages that you want to display on the
Integration Services Console computer.
For example, you may select Cyrillic to display Russian characters or Simplified Chinese and Traditional Chinese to display Chinese characters.
4 If the languages that you need do not appear in the Language settings for the system frame, or if you want
to attach a specific code page to a language, click the Advanced button and perform the following tasks:
a.
In the Advanced Regional Options dialog box, select the appropriate code page conversion tables. Select the code page tables that correspond to the languages that you want to display on the Integration Services Console computer.
b.
5 In the Regional Options dialog box, click Apply to apply the changes you made.
35
The Regional Options tab of the Regional and Language Options dialog box is displayed.
2 In the Standards and formats frame, from the drop-down list, select the appropriate language to use for
formatting items such as numbers, currencies, time, and dates.
4 In the Installed services frame, click Add. 5 In the Input Languages list, select the language that you want to add and select the check box for the type
of text service you want to install.
For example, you may select Cyrillic to display Russian characters or Simplified Chinese and Traditional Chinese to display Chinese characters.
6 If the languages that you need do not appear in the Input Languages frame, or if you want to add a specific
code page to a language, click the Advanced tab and perform the following tasks:
a.
In the Code page conversion tables frame, select the appropriate code page conversion tables. Select the code page tables that correspond to the languages that you want to display on the Integration Services Console computer.
b.
Click Apply.
The Regional Options tab of the Regional and Language Options dialog box is displayed.
36
2 In the Standards and formats frame, from the drop-down list, select the appropriate language to use for
formatting items such as numbers, currencies, time, and dates.
4 In the Default input language drop-down list, select the installed input language to use. 5 If the language you want does not appear in the Default input language drop-down list, click Add in the
Installed services frame.
6 In the Input languages list, select the language that you want to add. 7 In the Keyboard layout/IME list, select the check box for the type of text service you want to install and
click OK.
For example, you may select Cyrillic to display Russian characters or Simplified Chinese and Traditional Chinese to display Chinese characters.
8 Optional: If the languages that you need do not appear in the Add input languages dialog box, or if you
want to add a specific code page to a language, perform the following steps:
a. b. c. d.
Click Cancel to close both the Add input languages dialog box. Click Cancel to close the Text Services and Languages dialog box. In the Regional and Language Options dialog box, click the Advanced tab. In the Code page conversion tables frame, select the appropriate code page conversion tables. Select the code page tables that correspond to the languages that you want to display on the Integration Services Console computer.
e.
Click Apply.
On Microsoft SQL Server and Sybase, make your login user name an alias of TBC. On IBM DB2 and Oracle, log in with your user name and password, and create synonyms that map to the tables in TBC_MD.
This user alias or table synonym enables you to access a table that is qualified by TBC because it was created by the TBC user; for example, TBC.MO_INFO in the TBC_MD OLAP Metadata Catalog.
37
For more information, see the documentation for the RDBMS you are using.
2 Give authorization to the Integration Services user to schedule jobs using the cron scheduling daemon.
After the configurations are complete, you can access the relational data source to create an OLAP Metadata Catalog andif you chooseto set up the sample application. For more information about creating the OLAP Metadata Catalog, see Chapter 3, Creating, Upgrading, and Deleting OLAP Metadata Catalogs For more information about setting up the sample application, see Chapter 5, Setting Up the Sample Applications.
The Task Scheduler service must be started on the computer that runs Essbase Integration Server. The Task Scheduler service is enabled to start automatically each time the computer that runs Essbase Integration Server is started.
4 In the list of services, locate Task Scheduler and verify that the entry in the Status column is Started and
the entry in the Startup type column is Automatic.
If you must change one or both entries, this is the process: a. b. c. d. Double-click Task Scheduler. In the Task Scheduler Properties dialog box, select the General tab. Under Startup type, select Automatic. To start the service, in the Task Scheduler Properties dialog box, click Start.
38
If you must change one or both entries, this is the process: a. b. c. d. Double-click Task Scheduler. In the Task Scheduler Properties dialog box, select the General tab. Under Startup type, select Automatic. To start the service, click Start in the Task Scheduler Properties dialog box.
If you must change one or both entries, this is the process: a. b. c. d. Double-click Task Scheduler. In the Task Scheduler Properties dialog box, select the General tab. Under Startup type, select Automatic. To start the service, click Start in the Task Scheduler Properties dialog box.
39
40
2
In This Chapter
About Data Source Configuration ................................................................................................41 Configuring Data Source Names on Windows Systems ........................................................................43 Configuring Data Source Names on UNIX Systems .............................................................................54 After Configuration of Data Source Names ......................................................................................65 Connecting to Server Components and Data Sources .........................................................................66 Mapping JDBC Data Sources for Windows (Teradata Only)....................................................................72 Mapping JDBC Data Sources for UNIX (Teradata Users Only) .................................................................73 Manually Configuring the Environment for UNIX Systems ......................................................................75
To create OLAP models or build metaoutlines, you must connect Integration Services to a database that contains metadata (the OLAP Metadata Catalog database) and to a database in which needed users data is stored (the Data Source database). To make these connections, you must configure a data source name for an OLAP Metadata Catalog database and a Data Source database. You do this by mapping a supported Open Database Connectivity (ODBC) driver to each database.
Note:
You are not required to use two different databases for your Data Source database and OLAP Metadata Catalog database. Additionally, you are not required to use two different data source names to connect to the Data Source database and OLAP Metadata Catalog database if the two databases reside in the same database.
41
If database client software is required, the relational database management system (RDBMS) should include one of the database clients listed in the Oracle Hyperion Enterprise Performance Management System Installation Start Here. If your configuration requires database client software, ensure that the versions of the database client and the RDBMS are compatible. This chapter provides information on how to configure ODBC drivers. For information on how to install and configure database client software, see the RDBMS documentation.
Note:
Integration Services does not support data source table names and column names that contain spaces or special characters, such as a period (.). See Appendix B, Integration Services Limits and Guidelines for a complete listing of the unsupported characters.
On Windows, use the ODBC Administrator to configure ODBC data sources. All ODBC data source names are configured only on the computer that runs Essbase Integration Server.
Note:
If a supported ODBC driver is already mapped to a Data Source database, you do not need to map it again. In this case, you need to map a supported ODBC driver only to the OLAP Metadata Catalog database.
42
Integration Services does not support data source table names and column names that contain spaces or special characters, such as a period (.). See Appendix B, Integration Services Limits and Guidelines for a complete list of the unsupported characters.
The procedures in the following sections show you how to configure data source names to create connections to databases. The examples in the procedures use a database named TBC as the sample Data Source database and a database named TBC_MD as the sample OLAP Metadata Catalog database. For each site-specific data source name that you configure, obtain the following information from your database administrator:
The name of the relational database or flat file data source for which you want to configure a data source name For example, in the sample application, TBC is the name of the Data Source database; TBC_MD is the name of the OLAP Metadata Catalog database.
IP address or name of the computer on which the database tables are stored You can enter either an alphabetic computer name (for example, sequoia), or an IP address (for example, 172.0.0.125). In the sample application, if you are configuring a data source name for the TBC database, use the IP address of the computer where the TBC database tables are stored. If you are configuring a data source name for the TBC_MD database (the sample OLAP Metadata Catalog database), use the IP address of the computer where the TBC_MD tables are stored.
Port number on which your database listens For clarity in presenting the sample application in Integration Services Console, the sample Data Source database and sample OLAP Metadata Catalog database are two separate databases that use two different data source names. If your data is stored in a relational database, you are not required to use two different databases for your Data Source database and OLAP Metadata Catalog database. Additionally, you are not required to use two different data source names to connect to your Data Source database and OLAP Metadata Catalog database if all the tables reside in the same database. If your data is stored in one or more flat files, you must still use a relational database when setting up your OLAP Metadata Catalog database.
43
Note:
If you plan to use flat files as data sources, you must configure their Data Source Name (DSN). See Configuring the DataDirect Text Driver on Windows on page 49 for detailed instructions.
3 In the ODBC Data Source Administrator dialog box, click the System DSN tab.
You may repeat the configuration procedure to create any number of data source names to relational Data Source databases or OLAP Metadata Catalog databases.
The Wire Protocol Driver Setup dialog box for your RDBMS is displayed.
3 In the Data Source Name text box, type the name that you want to use for this data source connection.
In the sample application, TBC is the data source name used to connect to the Data Source database called TBC; TBC_MD is the data source name used to connect to the sample OLAP Metadata Catalog database called TBC_MD.
4 Optional: If you want to record a description, in the Description text box, type a description that indicates
how you use this data source name.
44
For example, you might type the following to describe the My Business database:
Customers, products, markets
You might type the following to describe the sample application database:
Sample relational data source
You might type the following to describe the sample OLAP Metadata Catalog database:
Sample OLAP Metadata Catalog
5 Make the necessary entries to configure data source names by completing the set of steps appropriate to
your RDBMS.
Refer to the following sections for the specific configuration instructions for your RDBMS:
IBM DB2 UDB or DB2 OS/390 on page 45 Informix on page 46 Oracle on page 47 Sybase on page 47
6 Test the connection by following the procedures in Testing DataDirect Wire Protocol Drivers on Windows
on page 47.
When you finish testing, repeat Step 2 through Step 5 to configure a data source name for an OLAP Metadata Catalog database.
1 In the IP Address text box, type the address (either the computer name or the numeric IP address) of the
computer where the database tables are stored.
2 In the Tcp Port text box, type the port number on which the IBM DB2 database server listens.
If you are using IBM DB2 UDB, the default port number is 50000. If you are using IBM DB2 OS/390, the default port number is 446.
3 If you are using IBM DB2 UDB, in the Database Name text box, type the name of the database for which
you want to create this data source name.
4 If you are using IBM DB2 OS/390, perform the following actions:
a. b. In the Location text box, type the IBM DB2 location name as defined during the IBM DB2 OS/390 installation. In the Collection text box, type the name that identifies a logical group of database objects (the default is DATADIRECT00).
45
Note:
For IBM DB2 OS/390, you configure a data source name for the Data Source database only. You cannot configure a data source name for the OLAP Metadata Catalog database. OLAP Metadata Catalog functionality is not supported in IBM DB2 OS/390.
To use Integration Services with IBM DB2 OS/390, you must add an entry to the ais.cfg file after you configure the data source name for the source database. See Adding a Configuration File Entry for IBM DB2 OS/390 in Windows on page 48 for more information.
5 Click the Bind tab. 6 Accept the default in the Bind tab and click the Create Package button.
For more information about the bind process, see DataDirect Technologies Connect ODBC Reference. The DataDirect documentation is located in HYPERION_HOME/common/ ODBC/ Merant/5.2/books. You can also access the Product Documentation on the DataDirect Web site at http://www.datadirect-technologies.com.
7 In the Login dialog box, type the user name and password associated with the database for which this ODBC
data source name is being created, and click OK.
A message is displayed informing you that the package was created successfully.
8 Return to Step 6 in Configuring DataDirect Wire Protocol Drivers in Windows on page 44 to complete the
configuration process.
Informix
Perform the following steps to configure the Wire Protocol driver for Informix:
1 In the Host Name text box, type the address (either the computer name or the numeric IP address) of the
computer where the database tables are stored.
2 In the Port Number text box, type the port number on which the Informix database server listens.
The default port number for Informix is 1526.
3 In the Server Name text box, type the address, either numerically or by name, of the Informix server as it
appears in the sqlhosts file.
If you do not know this information, contact your Informix database administrator.
4 In the Database Name text box, type the name of the database for which you want to create this data source
name.
5 Return to Step 6 in Configuring DataDirect Wire Protocol Drivers in Windows on page 44 to complete the
configuration process.
46
Oracle
Perform the following steps to configure the Wire Protocol driver for Oracle:
1 In the Host text box, type the address (either the computer name or the numeric IP address) of the computer
where the database tables are stored.
2 In the Port Number text box, type the port number on which the Oracle database server listens.
The default port number for Oracle is 1521.
3 In the SID text box, type the name of the database for which you want to create this data source name. 4 Return to Step 6 in Configuring DataDirect Wire Protocol Drivers in Windows on page 44 to complete the
configuration process.
Sybase
Perform the following steps to configure the Wire Protocol driver for Sybase:
1 In the Network Address text box, type the address (either the computer name or the numeric IP address)
of the computer where the database tables are stored and the port number on which the Sybase database server listens.
Separate the computer name or IP address and the port number with a comma as shown in the following examples: By name: sequoia,4100 Numerically: 172.0.0.125,4100
2 In the Database Name text box, type the name of the database for which you want to create this data source
name.
3 If you are using a DataDirect Wire Protocol driver for a Sybase relational database server, or a Microsoft SQL
server, you must perform the following tasks:
a. b. c.
Open the ODBC Sybase Wire Protocol Driver Setup dialog box. Select the Advanced tab. Select the Enable Quoted Identifiers option.
4 Return to Step 6 in Configuring DataDirect Wire Protocol Drivers in Windows on page 44 to complete the
configuration process.
47
2 In the User Name and Password text boxes, enter a valid user name and password for this data source
connection, and click OK.
If an error message is displayed, make the necessary corrections, and click Test Connect again. When no errors occur, a dialog box displays the message, Connection established!
4 In the ODBC Wire Protocol Driver Setup dialog box, click OK.
You are returned to the System DSN tab of the ODBC Data Source Administrator dialog box. The data source name that you entered and the driver that you mapped to the data source name are displayed in the System Data Sources list box.
5 Repeat Step 1 through Step 5 in Configuring DataDirect Wire Protocol Drivers in Windows on page 44 to
configure a data source name for an OLAP Metadata Catalog database.
6 When you have completed configuration of all needed data source names, click OK to close the ODBC Data
Source Administrator dialog box.
For IBM DB2 OS/390, you configure a data source name for the Data Source database only. You cannot configure a data source name for the OLAP Metadata Catalog database. OLAP Metadata Catalog functionality is not supported in IBM DB2 OS/390.
For example, using the sample application database, TBC, the following is the entry to the ais.cfg file:
[DS:TBC:390]
48
Integration Services does not support data source table names and column names that contain spaces or special characters, such as a period (.). See Appendix B, Integration Services Limits and Guidelines for a complete listing of the unsupported characters.
For more information on configuring the ODBC driver settings for text files, see the DataDirect Connect ODBC Reference PDF file in the $HYPERION_HOME/common/ODBC/Merant/5.2/ books directory.
Note:
You must complete the steps in Configuring DataDirect Wire Protocol Drivers in Windows on page 44 to configure a data source name for the relational database that you want to use as your OLAP Metadata Catalog database.
Before beginning the configuration procedure, perform the procedure in Accessing ODBC Data Source Administrator in Windows on page 44.
The General tab of the ODBC Text Driver Setup dialog box for flat file configuration is displayed.
3 In the Data Source Name text box, type a name for the flat file data source. 4 Optional: If you want to record a description, in the Description text box, type a description that indicates
how you use this data source name.
For example, you might type the following to describe the My Business database:
Customers, products, markets
5 In the Database Directory text box, type the path to the folder where the flat files are located. 6 In the Default Table Type drop-down list, select whether the text file is comma-separated, tab-separated,
character-separated, fixed length, or stream.
Note:
If your text files are character-separated, type the character that is used as a delimiter in the Delimiter Character text box.
49
7 If the first line of your text files contains the column names of the database table, select the Column Names
in First Line check box; otherwise, proceed to Step 8.
8 Select the Advanced tab. 9 In the Rows to Scan text box, type the number of rows that should be scanned for the driver to determine
the data types in the file.
The default is 25 rows. If 0 is entered, all rows in the file are scanned.
10 In the Action for Undefined Tables group, choose one of the following actions that the driver should take
when it encounters a file that has not been defined.
Prompt for Definition. The text driver prompts you when it encounters a file that has not been defined. Guess Definition. The text driver analyzes the file and guesses the files format.
Note:
The options in the Return Additional Tables group are not supported.
11 Complete the following steps to define the structure of your data source text files:
a. b. c. d. Click the Define button to display the Define File dialog box. Navigate to the folder that contains your data source text files. Select a text file to define and click Open to display the Define Table dialog box. In the Table text box, type a table name to associate with this text file. The table name you define is the table name that will be displayed in the left frame of the OLAP Model main window. e. If the first line of text in the file contains the column names, select the Column Names in the First Line check box. If the first line of the text file does not contain the column names, then do not select this check box. f. g. From the Table Type drop-down list, select the delimiter that is used in the text file. If you select Character in the Table Type drop-down list in Step f, specify the character in the Delimiter Character text box. You can specify any printable character except single and double quotation marks. h. In the Decimal Symbol text box, specify either a comma or period as a decimal separator to use when data is stored.
If you specified in Step 11f that your text files were comma-separated, tab-separated, or character-separated, click the Guess button to have the text driver guess at the column names and display them in the list box under Column Information. If you specified in Step 11f that your text files were fixed-length or stream type, click the Parse button to display the Parse Table dialog box, where you define the table column names.
50
For more information on using the Parse Table dialog box, see the DataDirect Connect ODBC Reference. You can also view the DataDirect online help by clicking the Help button in the Parse Table dialog box.
If you want to define the values and fields for each column manually, click Add to add the column name to the list box under Column Information. For each table that you add manually, you must specify the data type, data mask (where appropriate), column precision, column scale, length, and offset. For more information on these settings, see the DataDirect Connect ODBC Reference. You can also view the DataDirect online help by clicking the Help button in the Define Table dialog box.
13 Optional: Modify or delete any of the existing column definitions by selecting a column name in the list box
under Column Definition and clicking the Modify or Remove button.
14 Click OK to define the table. 15 Complete Step 11 through Step 14 for each data source table that you want to define. 16 When you have finished defining each data source table, click Cancel in the Define File dialog box to return
to the Advanced tab of the ODBC Text Driver Setup dialog box.
Configuring the SQL Server ODBC Driver for Microsoft SQL Server
Use the following procedure to configure a data source name for your Data Source database using the SQL Server driver. Then repeat the procedure to configure a data source name for your OLAP Metadata Catalog database.
Note:
You may repeat the configuration procedure to create any number of data source names to relational Data Source databases or OLAP Metadata Catalog databases.
The procedures that follow provide information on basic configuration. If you are using more advanced options, refer to the online help for each wizard screen for assistance in completing the wizards.
2 Click Add to open the Create New Data Source dialog box.
51
3 In the driver list box of the Create New Data Source dialog box, select the SQL Server driver, and click
Finish.
The Create a New Data Source to SQL Server dialog box is displayed.
4 In the Name text box, type the data source name that you want to use for this data source connection.
In the sample application, TBC is the data source name used to connect to the Data Source database called TBC; TBC_MD is the data source name used to connect to the sample OLAP Metadata Catalog database called TBC_MD.
5 Optional: In the Description text box, type a description that indicates how you use this data source name.
For example, you might type the following names to describe the My Business database:
Customers, products, markets
You might type the following statement to describe the sample application database:
Sample relational data source
You might type the following statement to describe the sample OLAP Metadata Catalog database:
Sample OLAP Metadata Catalog
6 In the Server text box, type the address (either the computer name or the numeric IP address) of the computer
where the database tables are stored.
7 Click Next.
The second wizard screen is displayed.
8 Select the appropriate option button for the method to use for verifying login IDs:
With Window NT authentication using the network login ID. With SQL Server authentication using a login ID and password entered by the user.
9 Optional: If you are using a network library other than TCP/IP (the default) to connect to the database server,
perform the following steps:
a. b.
Click the Client Configuration button. In the Network libraries list, select the appropriate option button for the library that you are using to connect to the database server computer, and click OK. You are returned to the second wizard screen.
10 Optional: Select the Connect to SQL Server to obtain default settings for the additional
configuration check box.
When this check box is selected, the driver obtains default settings from Microsoft SQL Server that it uses to complete additional setup screens in the wizard. When this check box is clear, the driver uses standard defaults to complete the additional setup screens in the wizard.
11 Click Next.
The third wizard screen is displayed.
12 Select the Change the default database to check box and then type or select the name of the database
for which you want to create this data source name.
52
For example, in the sample application, TBC is the name of the Data Source database; TBC_MD is the name of the OLAP Metadata Catalog database.
13 Click Next.
The fourth wizard screen is displayed.
14 Make any entries appropriate to your system configuration and click Finish. 15 Repeat Step 2 through Step 14 to configure a data source name for an OLAP Metadata Catalog. 16 When you have completed configuring all needed data sources, click OK to close the ODBC Data Source
Administrator dialog box.
You may repeat the configuration procedure to create any number of data source names to Data Source databases or OLAP Metadata Catalog databases.
2 Click Add to open the Create New Data Source dialog box. 3 In the driver list box of the Create New Data Source dialog box, select Teradata, and click Finish.
The ODBC Driver Setup for Teradata RDBMS dialog box is displayed.
4 In the Name text box, type the data source name that you want to use for this data source connection.
In the sample application, TBC is the data source name used to connect to the Data Source database called TBC; TBC_MD is the data source name used to connect to the sample OLAP Metadata Catalog database called TBC_MD.
5 Optional: In the Description text box, type an description that indicates how you use this data source name.
For example, you might type the following to describe the My Business database:
Customers, products, markets
You might type the following to describe the sample application database:
Sample relational data source
You might type the following to describe the sample OLAP Metadata Catalog database:
Sample OLAP Metadata Catalog
53
6 In the Teradata Server Info text box, type the address (either the computer name or the numeric IP address)
of the computer where the database tables are stored.
7 In the Default Database text box, type the name of the database for which you want to create this data
source name.
For example, in the sample application, TBC is the name of the Data Source database; TBC_MD is the name of the OLAP Metadata Catalog database.
9 Click OK to return to the System DSN tab of the ODBC Data Source Administrator dialog box.
The data source name that you entered and the driver that you mapped to it are displayed in the System Data Sources list box.
10 Repeat Step 2 through Step 9 above to configure a data source name for an OLAP Metadata Catalog.
Note:
You may repeat the above procedure to create any number of data source connections to relational data sources or OLAP Metadata Catalogs.
11 When you have completed configuring all needed data sources, click OK to close the ODBC Data Source
Administrator dialog box.
2 Select the data source name and click Configure to open the driver setup dialog box or wizard specific to
your RDBMS.
54
odbc.ini file (using a text editor such as vi) to configure, add, or change data sources names,
Integration Services does not support data source table names and column names that contain spaces or special characters, such as a period (.). See Appendix B, Integration Services Limits and Guidelines for a complete listing of the unsupported characters.
Flat File Data Sources: If you are using a flat file data source, you must define the structure of the text files in your data source in a QETXT.INI file. The process for creating the QETXT.INI file for UNIX operating systems is described in the DataDirect Connect ODBC Reference in the PDF file in $HYPERION_HOME/common/ODBC/Merant/5.2/books.
55
Teradata: When configuring a data source name, the data source description must be set to tdata.sl (HP-UX) or tdata.so (AIX and Solaris). Specific examples for AIX, HP-UX, and Solaris operating systems are shown in Examples of ODBC Settings for Teradata on page 62.
Tip:
The procedure in this section shows you how to configure a data source name by manually editing the odbc.ini file.
Note:
If you use a file other than the odbc.ini file (located by default in $HYPERION_HOME/ common/ODBC/Merant/5.2), be sure to set the ODBCINI environment variable to the name of the file that you use.
To minimize confusion, the data source name can match the name of the database in the RDBMS.
3 Add a new section to the file by creating a new line with the new data source name enclosed in brackets;
for example:
[mydata]
4 On the lines following the data source name, add the full path and file name for the ODBC driver required
for this data source and any other required ODBC driver information.
Use the examples shown in the following sections as guidelines for specific RDBMSs.
Note:
Ensure that the ODBC driver file actually exists in the location that you specify for the Driver= setting.
5 Add a new section to the file by creating a new line with ODBC enclosed in brackets; for example:
[ODBC]
6 Under the [ODBC] heading, add the full path to the location where the /lib and /messages directories
are contained.
Use the examples shown in the following sections as guidelines for specific RDBMSs.
7 When you finish editing odbc.ini, save the file and exit the text editor.
For information about the odbc.ini file and the ODBC driver settings for each RDBMS or flat file data source, see the DataDirect Connect ODBC Reference in the PDF file and the Readme files
56
in the $HYPERION_HOME/common/ODBC/Merant/5.2/books directory. For information about vendor-supplied ODBC driver settings, refer to the installation documentation for the vendor-supplied ODBC drivers. Flat File Data Sources: If you are using a flat file data source, you must define the structure of the text files in your data source in a QETXT.INI file. The process for creating the QETXT.INI file for UNIX operating systems is described in the DataDirect Connect ODBC Reference in the PDF file in $HYPERION_HOME/common/ODBC/Merant/5.2/books.
57
Note:
To use Integration Services with IBM DB2 OS/390, you must add an entry to the ais.cfg file after you configure the data source name for the source database.
58
Tip:
Run ivtestlib to verify that the environment is set to run the correct ODBC driver file. For example, run ivtestlib and paste the path and file name that follow Driver= in the odbc.ini file that you edited.
Integration Services does not support data source table names and column names that contain spaces or special characters, such as a period (.). See Appendix B, Integration Services Limits and Guidelines for a complete listing of the unsupported characters.
Configuring Data Source Names on UNIX Systems
59
Note:
You must use a relational database as your OLAP Metadata Catalog database and configure it as described in Editing the odbc.ini File on page 55.
[ODBC Data Sources] tbc_ff=TBC flat file data source ... [tbc_ff] Driver=/home/hyperion/Hyperion/common/ODBC/Merant/5.2/Drivers/ARtxt22.so (location of a driver) Description=DataDirect 5.2 TextFile(*.*) AllowUpdateAndDelete=0 (a variable edited by user) ApplicationUsingThreads=1 (a variable edited by user) CacheSize=64 (a variable edited by user) CenturyBoundary=20 (a variable edited by user) Database=/home/hyperion/flat_files/stream (a data file location) DataFileExtension=TXT (a variable edited by user, default is TXT) DecimalSymbol=. (a variable edited by user, default is dot(.)) Delimiter=~ (a variable edited by user) FileOpenCache=0 (a variable edited by user) FirstLineNames=1 (a variable edited by user) IntlSort=1 (a variable edited by user) ScanRows=5 (a variable edited by user) TableType=Character (a variable edited by user) UndefinedTable=GUESS (a variable edited by user) [ODBC] IANAAppCodePage=4 InstallDir=$HYPERION_HOME/common/ODBC/Merant/5.2 Trace=0 TraceDll=$HYPERION_HOME/common/ODBC/Merant/5.2/Drivers/odbctrac.so TraceFile=odbctrace.out UseCursorLib=0
For more information on editing the ODBC.INI file in UNIX to configure a flat file data source see the DataDirect Connect ODBC Reference in PDF format in $HYPERION_HOME/common/ ODBC/Merant/5.2/books/odbcref.
60
DS=. FIELD1=PRODUCTID,NUMERIC,9,0,9,0, FIELD2=FAMILYID,NUMERIC,8,0,8,0, FIELD3=SKU,VARCHAR,6,0,6,0, FIELD4=SKUNAME,VARCHAR,18,0,18,0, FIELD5=CAFFIENATED,VARCHAR,11,0,11,0, FIELD6=OUNCES,NUMERIC,6,0,6,0, FIELD7=PKGTYPE,VARCHAR,7,0,7,0, FIELD8=INTRODATE,DATE,10,0,10,0,yyyy/m/d [REGION] FILE=REGION.txt FLN=0 TT=Comma Charset=ANSI DS= [SalesFactShort] FILE=SALESFACTSHORT.txt FLN=1 TT=Comma Charset=ANSI DS=. FIELD1=STATEID,NUMERIC,7,0,7,0, FIELD2=PRODUCTID,NUMERIC,9,0,9,0, FIELD3=SCENARIOID,NUMERIC,10,0,10,0, FIELD4=SUPPLIERID,NUMERIC,10,0,10,0, FIELD5=TRANSDATE,DATE,10,0,10,0,yyyy/m/d FIELD6=SALES,NUMERIC,7,2,6,0, FIELD7=COGS,NUMERIC,6,2,6,0, FIELD8=MARKETING,NUMERIC,11,2,9,0, FIELD9=PAYROLL,NUMERIC,9,2,7,0, FIELD10=MISC,NUMERIC,6,2,4,0, FIELD11=OPENINGINVENTORY,NUMERIC,18,2,16,0, FIELD12=ADDITIONS,NUMERIC,11,2,9,0, [Market] FILE=MARKET.txt FLN=1 TT=Comma Charset=ANSI DS=. FIELD1=STATEID,NUMERIC,7,0,7,0, FIELD2=REGIONID,NUMERIC,8,0,8,0, FIELD3=STATE,VARCHAR,13,0,13,0, FIELD4=POPULATIONID,NUMERIC,12,0,12,0,
For more information on creating the QETXT.INI file in UNIX to configure tables for a flat file data source see the DataDirect Connect ODBC Reference in PDF format in $HYPERION_HOME/ common/ODBC/Merant/5.2/books/odbcref.
61
... [sqldata] Driver=/home/hyperion/Hyperion/common/ODBC/Merant/5.2/Drivers/ARmsss22.so Description=SQL Server 2000 Driver Database=sqldata Address=isqa17,1433 QuotedId=No AnsiNPW=No [ODBC] IANAAppCodePage=4 InstallDir=$HYPERION_HOME/common/ODBC/Merant/5.2 Trace=0 TraceDll=$HYPERION_HOME/common/ODBC/Merant/5.2/lib/odbctrac.so TraceFile=odbctrace.out UseCursorLib=0
62
HP-UX The following example illustrates how you might edit odbc.ini to connect to the same Data Source database, tbc, on Teradata (on HP-UX), using the Teradata ODBC driver. You can make a separate entry to connect to the OLAP Metadata Catalog database. The data source description in [ODBC Data Sources] must be set to tdata.sl.
[ODBC Data Sources] tbc=tdata.sl ... [tbc] Driver=/usr/odbc/drivers/tdata.sl Description=NCR 3600 running Teradata V2R5 DBCName=139.64.140.45 DefaultDatabase=tbc UseXViews=Yes [ODBC] IANAAppCodePage=4 InstallDir=$HYPERION_HOME/common/ODBC/Merant/5.0 Trace=0 TraceDll=$HYPERION_HOME/common/ODBC/Merant/5.0/lib/odbctrac.so TraceFile=odbctrace.out UseCursorLib=0
Solaris: The following example illustrates how you might edit odbc.ini to connect to the same Data Source database, tbc, on Teradata (on Solaris), using the Teradata ODBC driver. You can make a separate entry to connect to the OLAP Metadata Catalog database. The data source description in [ODBC Data Sources] must be set to tdata.so.
[ODBC Data Sources] tbc=tdata.so ... [tbc] Driver=/usr/odbc/drivers/tdata.so Description=NCR 3600 running Teradata V2R5 DBCName=139.64.140.45 DefaultDatabase=tbc UseXViews=Yes [ODBC] IANAAppCodePage=4 InstallDir=$HYPERION_HOME/common/ODBC/Merant/5.0 Trace=0 TraceDll=$HYPERION_HOME/common/ODBC/Merant/5.0/lib/odbctrac.so TraceFile=odbctrace.out UseCursorLib=0
63
Note:
The UseXViews parameter is optional on all UNIX platforms. The UseXViews parameter limits the number of data sources displayed in the left frame of an OLAP Model main window based on the security of the Teradata user ID that was used to connect.
Creating IBM DB2 Bind Packages for DataDirect Wire Protocol Driver Connections
In the UNIX environment, you must create a bind package for each connection between Integration Server and an IBM DB2 database that you make using the DataDirect (formerly MERANT) Wire Protocol driver. Without a bind package for each connection, the Wire Protocol driver will not work properly. DataDirect supplies a command line program to create the bind packages.
To create a bind package for a connection using the DataDirect Wire Protocol Driver for IBM
DB2:
1 On the computer running Integration Server, open a command shell window. 2 In the command shell window type:
bind19 dsn
Replace dsn with the data source name for which you are creating this bind package. For example, to create a bind package for the data source name TBC, type:
bind19 TBC
Note:
You are prompted for a user name and password if this information is not stored in the system information file.
If successful, a message is displayed stating that the package was created and bound. For troubleshooting information, access the DataDirect Web site at http://www.datadirecttechnologies.com and select the Product Documentation link.
3 Repeat Step 2 for all bind packages you want to create. 4 When all bind packages are created, close the command shell window.
64
Note:
For IBM DB2 OS/390, you configure data source name for the Data Source database only. You cannot configure a data source name for the OLAP Metadata Catalog database. OLAP Metadata Catalog functionality is not supported in IBM DB2 OS/390.
For example, using the sample application database, TBC, the following is the entry to the ais.cfg file:
[DS:TBC:390]
The SQL Server ODBC driver may time out during a call to an SQL Server database, particularly during a data load. If a timeout occurs, try again when the database is not busy. Increasing the driver time-out period may avoid this problem. For more information, see the ODBC documentation for the driver that you are using. For more information on ODBC connection problems and solutions, see Chapter 7, Troubleshooting ODBC and Connections.
65
An OLAP Metadata Catalog must be created and configured before you can connect to it. For information on configuring a relational data source for an OLAP Metadata Catalog, see Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide. For information on creating OLAP Metadata Catalogs, see Chapter 3, Creating, Upgrading, and Deleting OLAP Metadata Catalogs.
After the process of creating or upgrading an OLAP Metadata Catalog is complete, you can begin to create OLAP models and metaoutlines. You must connect to Integration Server, to an external data source, and to the OLAP Metadata Catalog where you want to store the OLAP models and metaoutlines that you create. To use Integration Services Console to perform data loads, member loads, or member and data loads, you must connect the client software to an additional server component, Essbase Server. You are prompted by the Login dialog box to connect to an Integration Server and an OLAP Metadata Catalog. Optionally, in the Login dialog box, you can also connect to an instance of Essbase Server and create default connection settings for both server components. For information on troubleshooting connections to server components, see Chapter 7, Troubleshooting ODBC and Connections.
The Integration Server system administrator must provide a user name and password to an OLAP Metadata Catalog before you can connect to the server.
Note:
It is highly recommended that you do not connect the Integration Server and Integration Services Console via a wide area network (WAN). Doing so may cause severe performance degradation.
66
The Login dialog box is displayed automatically when you start Integration Services Console. You can redisplay the Login dialog box by using the procedures described in this topic.
2 In the Server drop-down list under Integration Services, enter or select the name or IP address of an
Integration Server computer; for example, cypress.
If the port number on which Essbase Integration Server communicates with the console has been set to a value other than the default, you must type the server name or IP address and type the nonstandard port number, separated by a colon; for example:
cypress:3390
3 In the OLAP Metadata Catalog drop-down list, select the data source name of an OLAP Metadata Catalog;
for example, TBC_MD for the sample TBC (The Beverage Company) catalog database.
Oracle:TBC_MD
4 In the Code Page drop-down list box, select the code page of the language you want to use during the current
Integration Services Console session.
The code page is a portion of the locale which identifies the local language and cultural conventions such as the formatting of currency and dates and the sort order of data.
5 In the User Name drop-down list box under Integration Services, type or select your user name.
It is necessary to type a user name the first time it is used. After a user is successfully connected to Integration Services, the user name is then displayed in the User Name drop-down list.
6 In the Password text box under Essbase Integration Server, type your password.
Note:
Use the same user name and password that you use to connect directly to the database server and to access the database that contains the OLAP Metadata Catalog.
67
If you plan to use Essbase Integration Services Console on a client computer that is outside the firewall for your network, and the console requires that Integration Server and Essbase Server be located inside the firewall for your network, you must log on to Essbase Server with a name or an IP address that connects from both sides of the firewall. The system administrator provides you with this name or IP address.
Note:
The system administrator for Essbase Server must provide you with a user name and password before you can connect.
The Login dialog box is displayed automatically when you start Integration Services Console. You can redisplay the Login dialog box at any time while working in the OLAP Metaoutline main window by using the procedures described in this topic.
2 In the Server text box under Essbase Server, type or select the name of a computer that is running Essbase
Server; for example, sequoia.
Note:
If you are using Integration Services Console on a client computer that is outside the firewall for your network, and you require access to instances of Integration Server and Essbase Server located inside the firewall for your network, be sure to use a name or an IP address for the instance of Essbase Server that is accessible from both sides of the firewall.
3 In the User Name drop-down list box under Essbase Server, type or select your Essbase user name.
It is necessary to type a user name the first time it is used. After a user is successfully connected to Essbase Server, the user name is then displayed in the User Name drop-down list.
4 In the Password text box under Essbase Server, type your Essbase password. 5 Click OK.
68
3 In the Server text box, type or select the name of an Integration Server computer; for example, cypress. 4 In the OLAP Metadata Catalog text box, type or select the name of an OLAP Metadata Catalog; for example,
TBC_MD for the sample TBC database.
5 In the Code Page drop-down list box, select the code page of the language you want to use for this Integration
Services Console session.
The code page is a portion of the locale which identifies the local language and cultural conventions such as the formatting of currency and dates and the sort order of data.
6 In the User Name text box, type or select a user name for the OLAP Metadata Catalog. 7 Take one of the following actions:
To set the connection defaults, click Set Default. To save the connection information without setting the information as the default, click Add to List.
8 In the Default Server text box, type or select the name of an Essbase Server computer; for example, cypress. 9 In the Default User Name text box, type or select a user name for the instance of Essbase Server; for example,
TBC.
You can connect to more than one relational data source to create OLAP models.
69
2 In the Data Source drop-down list, select the data source to be used; for example, TBC in the sample
application.
An ODBC data source must be created on the computer that is running Integration Server for any external data sources that you want to use. If the data source that you need is not visible in the scroll list, contact the Integration Services system administrator. For more information about troubleshooting server and data source connections, see Chapter 7, Troubleshooting ODBC and Connections. The following example is a Net Service Name stanza that defines TBC in the tnsnames.ora file:
TBC = (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = labmachine2)(PORT = 1521)) ) (CONNECT_DATA = (SERVICE_NAME = orasid) ) )
For Oracle using onames, in the example above, TBC is the Schema Name used to connect to an Oracle database. This is the database identifier that you use when you are using SQL *Plus to connect to a database. DataDirect Drivers: If you are using DataDirect drivers with Oracle, pick a data source name from the Data Source drop-down list.
3 In the Code Page drop-down list box, select the code page of the language you want to use during the
current Integration Services Console session, and click OK.
The code page is a portion of the locale which identifies the local language and cultural conventions such as the formatting of currency and dates and the sort order of data.
4 In the User Name drop-down list box, type or select your user name.
Note:
If you are using a flat file data source, no user name is required.
6 Click OK.
The left frame of the OLAP Model main window initially displays information about the first data source to which you connected. Data source information is displayed hierarchically by data source name and owner name. For each owner name, data is further sorted and organized by tables, views, and synonyms. You can use the listed source tables to create an OLAP model or use additional source tables by connecting to other data sources.
70
Configuring Data Sources
Information about subsequent data sources to which you connect is displayed in the same manner as information about the first connected data source was displayed; that is, by data source name and owner name, then tables, views, and synonyms. Expand the plus sign, , to display tables, views, and synonyms contained in the data source.
If you want to connect to additional data sources, complete the steps in the following procedure:
2 In the Data Source drop-down list, select the additional data source to which you want to connect.
If you are using DataDirect drivers with Oracle, pick a data source name from the Data Source drop-down list.
3 In the Code Page drop-down list box, select the code page of the language you want to use during the
current Integration Services Console session, and click OK.
4 In the User Name drop-down list box, type or select your user name.
It is necessary to type a user name the first time it is used. After a user is successfully connected to a data source, the user name is then displayed in the User Name drop-down list.
Note:
If you are using a flat file data source, no user name is required.
5 In the Password text box, type your password and click Connect.
Note:
The left frame of the OLAP Model main window displays information about the data source to which you just connected along with information about the first data source to which you connected.
6 Repeat Step 2 through Step 3 for each data source to which you want to connect.
As you connect to additional data sources, the left frame of the OLAP Model main window displays information about all data sources to which you are connected.
7 When you finish connecting to all appropriate data sources, click Close.
71
The name of the Teradata data source used as the OLAP Metadata Catalog in Integration Services The host on which the Teradata Gateway program is running Port number of the Gateway host The host on which the Teradata server resides
The name of the Oracle Net Service Name (found in $ORACLE_HOME/network/ admin/ tnsnames.ora) used as the OLAP Metadata Catalog in Integration Services The Schema Name used to connect to an Oracle database if you are using onames instead of tnsnames This is the database identifier that you use when you connect to the database using SQL*Plus.
oname
The host on which the Oracle server is installed or running Port number of the host on which the Oracle server is configured The server identifier for Oracle on host <host name>
In the following examples for mapping a JDBC data source for Teradata and Oracle, note that the pound sign (#) at the beginning of a line indicates that the line has been commented out.
72
Configuring Data Sources
When using Oracle with tnsnames, if the definitions for tnsname, host name, port #, and SID elements do not match the corresponding elements of a Net Service Name stanza in the tnsnames.ora file, a Java environment error will be the result. The preceding JDBC settings example is built using the elements in the following tnsnames.ora stanza:
ora_tbc_md = (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = labmachine2)(PORT = 1521)) ) (CONNECT_DATA = (SERVICE_NAME = orasid) ) )
When using Oracle with onames, in the example above, ora_tbc_md is the Schema Name used to connect to an Oracle database. This is the database identifier you use when you are connecting to the database using SQL*Plus.
73
The name of the Teradata data source used as the OLAP Metadata Catalog in Integration Services The host on which the Teradata JDBC Type 3 driver is configured and the Jserver is running Port number of the Gateway host The host on which the Teradata server resides
The name of the Oracle Net Service Name (found in $ORACLE_HOME/network/ admin/ tnsnames.ora) used as the OLAP Metadata Catalog in Integration Services The Schema Name used to connect to an Oracle database if you are using onames instead of tnsnames This is the database identifier that you use when you are using SQL *Plus to connect to the database.
oname
The host on which the Oracle server is installed and running Port number of the host on which the Oracle server is configured The server identifier for Oracle on host <host name>
In the following examples for mapping a JDBC data source for Teradata and Oracle, note that the pound sign (#) at the beginning of a line indicates that the line has been commented out. Example of JDBC Type 3 Settings for Teradata
[JDBC DSN] td_catalog_dsn:labmachine1:6063/teradataServer #tbc_md:labmachine1:6063/tera1 td_tbc_md:labmachine1:6063:tera1/
When using Oracle with tnsnames, if the definitions for tnsname, host name, port #, and SID elements do not match the corresponding elements of a Net Service Name stanza in the tnsnames.ora file, a Java environment error results. The preceding JDBC settings example is built using the elements in the following tnsnames.ora stanza:
ora_tbc_md = (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = labmachine2)(PORT = 1521))
74
When using Oracle with onames, in the example above, ora_tbc_md is the Schema Name used to connect to an Oracle database. This is the database identifier that you use when you are using SQL*Plus to connect to a database.
Running ais_start automatically configures your environment. The steps to configure your environment include configuring the hyperion user login scripts to run the is.sh (or) is.csh file and update the environment variables for Essbase Integration Server.
75
76
3
In This Chapter
About OLAP Metadata Catalogs ..................................................................................................77 Creating or Upgrading OLAP Metadata Catalogs................................................................................78 Mapping OLAP Metadata Catalogs ...............................................................................................85 Deleting OLAP Metadata Catalogs ...............................................................................................86
Before you run Integration Services, you must create an OLAP Metadata Catalog to store the OLAP models that you create and the metaoutlines that you build from the OLAP models.
Note:
If you intend to use the Sample application to acquaint yourself with Integration Services, skip to Chapter 5, Setting Up the Sample Applications. The Sample application has its own relational data source, OLAP Metadata Catalog, OLAP model, and metaoutline. Return to this chapter when you are ready to use a relational data source to build OLAP models and metaoutlines. It is recommended that you use the automatic installation process to install the Sample applications. See Setting Up the Standard Sample Application Automatically on page 95.
This chapter tells you what to consider before you create an OLAP Metadata Catalog in a relational database. You can create the catalog either automatically, through the use of Integration Services Console, or manually. You can also manually create a Unicode-enabled OLAP Metadata Catalog. For manual creation of an OLAP Metadata Catalog, this chapter lists the SQL scripts that you must run to create tables for the catalog. The procedures in this chapter assume that you know how to create tables by running SQL scripts.
77
You can run the RDBMS for an OLAP Metadata Catalog on any supported platform as long as you have the ODBC driver and database client software required to connect to the catalog from the computer that runs Essbase Integration Server. You can also run the RDBMS for the relational data source that you use to create OLAP models and build metaoutlines on any supported platform, if you have the ODBC driver and database client software required to access the data source. For a detailed matrix of specifically supported databases and ODBC drivers, see Oracle Hyperion Enterprise Performance Management System Installation Start Here. The RDBMS for the OLAP Metadata Catalog can be different from the RDBMS for the relational data source, and the platforms for the two RDBMSs do not need to be the same. You can have more than one OLAP Metadata Catalog and, by using the XML Import/Export utility, you can move OLAP models and metaoutlines from one OLAP Metadata Catalog to another OLAP Metadata Catalog.
Microsoft SQL ServerThe tables in a catalog are not accessible if they are created by a user that does not have db_owner privileges.
Create an OLAP Metadata Catalog automatically by using Integration Services Console (see Creating or Upgrading OLAP Metadata Catalogs Automatically on page 79). Create tables for the OLAP Metadata Catalog database by running SQL scripts (see Creating OLAP Metadata Catalogs Manually on page 80).
Create a database device or tablespace Allot 30 MB for storage Create user names and passwords, if needed Grant user privileges or permissions
78
One OLAP Metadata Catalog can store all OLAP models and metaoutlines from numerous relational data sources. You can also create separate OLAP Metadata Catalogs to store additional OLAP models and metaoutlines for different projects.
Figure 2
If you prefer to create the OLAP Metadata Catalog manually, see Upgrading OLAP Metadata Catalogs Manually on page 81.
It is necessary to type a server name the first time it is used. After the OLAP Metadata Catalog is successfully created, the server name is then displayed in the Server Name drop-down list box.
2 From the Catalog ODBC DSN drop-down list, select the Data Source Name for the OLAP Metadata Catalog
that you are creating.
79
Note:
You must create the database in which you will store the OLAP Metadata Catalog, assign the appropriate user permissions, and configure the ODBC connection before you can create the catalog.
3 In the Code Page drop-down list box, select the code page that corresponds to the code page of the relational
database that will be used as the OLAP Metadata Catalog.
The code page is a portion of the locale which identifies the local language and cultural conventions such as the formatting of currency and dates and the sort order of data.
4 In the User Name drop-down list, select or type the user name to which you have assigned permission to
access the database in which the OLAP Metadata Catalog will be stored.
5 In the Password text box, type the password for the user name to which you have assigned permission to
access the database in which the OLAP Metadata Catalog will be stored.
6 If you want the OLAP Metadata Catalog Setup dialog box to be displayed automatically each time you start
Integration Services Console, leave the Show this dialog at Startup check box selected.
If you do not select the Show this dialog at Startup check box, you can access the OLAP Metadata Catalog Setup dialog box from the console by selecting Tools, then Create Catalog. If an OLAP model or metaoutline is open when you select Tools, then Create Catalog, you are prompted to disconnect from the current catalog (Connections, then OLAP Metadata Catalog, then Disconnect). Note that disconnecting from an OLAP Metadata Catalog causes the open OLAP model or metaoutline to close.
7 Click Create.
Integration Services Console creates the OLAP Metadata Catalog automatically.
Note:
If you are migrating from a previous release of Integration Services, the OLAP Metadata Catalog auto-creation process upgrades the existing OLAP Metadata Catalog.
The scripts that you use to create the tables for a Unicode OLAP Metadata Catalog are named:
oc_create_database_name_unicode.sql
These scripts are listed in Table 1 on page 83, along with the utility programs with which they have been tested.
80
Note:
Hyperion strongly recommends that you use the automatic installation process to create the OLAP Metadata Catalog (see Creating or Upgrading OLAP Metadata Catalogs Automatically on page 79).
5 Verify that you have created the OLAP Metadata Catalog tables.
For example, type a command such as:
SELECT * FROM JOIN_HINTS
or, start the applicable RDBMS utility program and verify that the OLAP Metadata Catalog has the new tables.
If you use Integration Services Console to create an OLAP Metadata Catalog automatically, the system upgrades existing OLAP Metadata Catalogs and you do not need to upgrade them manually (see Creating or Upgrading OLAP Metadata Catalogs Automatically on page 79). Hyperion recommends that you use the automatic process to create the OLAP Metadata Catalog.
You manually upgrade the tables of an OLAP Metadata Catalog by running the SQL scripts that use the same database utility program that you typically use to create tables. The SQL scripts to upgrade tables for the OLAP Metadata Catalog are in the ocscript directory where you installed Integration Services. The upgrade scripts are named oc_upgrade*_ database_name.sql and are listed in Table 1 on page 83, along with the utility programs with which they have been tested.
81
Note:
If you manually rebuild an OLAP Metadata Catalog, you must drop (delete) the catalog tables by using oc_drop*database_name.sql. Then you create the OLAP Metadata Catalog tables by using oc_create*database_name.sql. If you choose to rebuild a catalog manually, it is not necessary to run any upgrade scripts.
Caution!
If you drop (delete) an OLAP Metadata Catalog, you also delete the OLAP models and metaoutlines that it contains.
The following topics discuss various aspects of upgrading OLAP Metadata Catalogs manually:
SQL Scripts Used to Create and Upgrade Tables on page 82 Proper Order of SQL Scripts on page 84 Upgrading Tables in the OLAP Metadata Catalog on page 85
oc_create_database_name.sql to build tables oc_drop_database_name.sql to drop tables oc_upgrade20_database_name.sql to upgrade tables from Integration Services Release
6.1 to 6.5
If you need to rebuild an OLAP Metadata Catalog, you must drop (delete) the catalog tables before you build them. Remember that if you drop an OLAP Metadata Catalog, you also delete the OLAP models and metaoutlines that it contains. The utilities listed in Table 1 have been tested to work with the SQL scripts.
82
Table 1
SQL Scripts for Creating an OLAP Metadata Catalog SQL Script oc_create_db2.sql oc_drop_db2.sql oc_upgrade20_db2.sql oc_upgrade61_db2.sql oc_upgrade65_db2.sql oc_upgrade651_db2.sql Utility Program IBM DB2 Command Center or >DB2 -tvf
Informix
DBAccess
Oracle
SQL*Plus
Sybase
ISQL
83
Database
Utility Program
Teradata
BTEQ
Upgrading from Integration Services Release 2.0 If you are upgrading from Release 2.0 to the current release, you run the following upgrade scripts in the order listed:
oc_upgrade61_*.sql oc_upgrade65_*.sql oc_upgrade651_*.sql
Upgrading from Integration Services Release 6.1 If you are upgrading from Release 6.1 to the current release, you run the following scripts in succession:
oc_upgrade65_*.sql oc_upgrade651_*.sql
Upgrading from Integration Services Release 6.5 If you are upgrading from Release 6.5 to the current release, you run the following script:
84
Creating, Upgrading, and Deleting OLAP Metadata Catalogs
oc_upgrade651_*.sql
Caution!
After you update an OLAP Metadata Catalog, you cannot roll back to a previous version. The new version of OLAP Metadata Catalog is not compatible with previous releases of Integration Services. In addition, do not attempt to use the new catalog with previous releases of Integration Services software. Such an attempt can result in corrupted OLAP Metadata Catalog data.
4 Run the SQL script (or scripts, if applicable) to upgrade the tables.
See Proper Order of SQL Scripts on page 84. On Microsoft SQL Server, you receive a message that you did not create data or rows. This message is normal because you created only tables and columns.
5 Verify that the new tables have been added to the existing OLAP Metadata Catalog.
You can verify this by entering a command, for example:
SELECT * FROM OM_DESCRIPTION
or, you can start the applicable RDBMS utility program and verify that the OLAP Metadata Catalog has the new tables.
85
Note:
If you attempt to access a Microsoft SQL Server database with the Microsoft native ODBC driver and you do not have access permission, SQL Server connects you to the default database without notifying you.
Note:
To use Integration Services on Sybase, you must have execute permission for sp-fkeys in Sybsystemprocs.
Remember that if you delete an OLAP Metadata Catalog, you also delete the OLAP models and metaoutlines that it contains.
2 From the Server Name drop-down list, select the appropriate Essbase Integration Server computer. 3 From the Catalog ODBC DSN drop-down list, select the Data Source Name for the OLAP Metadata Catalog
that you want to delete.
4 In the User Name drop-down list, select or type your user name. 5 In the Password text box, type your password. 6 Click Yes to the confirmation prompt to delete the catalog.
If the OLAP Metadata Catalog that you are attempting to delete is currently in use, you are prompted to disconnect from the current catalog (Connections, then OLAP Metadata Catalog, then Disconnect). You must disconnect before you can delete the catalog.
Note:
Disconnecting from an OLAP Metadata Catalog causes the open OLAP model or metaoutline to close.
86
4
In This Chapter
OLAP Metadata Catalog ...........................................................................................................87 Tables Relating to the OLAP Model...............................................................................................88 Tables Relating to the Metaoutline ...............................................................................................89 Tables Relating to Drill-Through...................................................................................................91 Tables Relating to Hybrid Analysis ...............................................................................................91 Miscellaneous Tables ..............................................................................................................92
OLAP Metadata Catalog is a relational database that contains OLAP models, metaoutlines, and the information necessary to retrieve the required data from the data source. You can install OLAP Metadata Catalog in any supported data source. See the Oracle Hyperion Enterprise Performance Management System Installation Start Here for a complete list of supported data sources. This chapter describes the tables in OLAP Metadata Catalog.
What data to retrieve from the data source What operations to perform on the data How to load the data into the Essbase database
Regularly back up the database that contains OLAP Metadata Catalog so that you can restore OLAP Metadata Catalog from backups if it becomes corrupted. Use one of the procedures in Table 2, depending on what OLAP Metadata Catalog component you want to restore.
Caution!
It is strongly recommended that you do not alter the table names or data in OLAP Metadata Catalog with tools other than Integration Services.
87
Table 2
OLAP Metadata Catalog Restore Procedures Procedure Run a script to create the database tables. See Chapter 3, Creating, Upgrading, and Deleting OLAP Metadata Catalogs. Restore the database that contains OLAP Metadata Catalog from backups. See the documentation for the data source.
Component To Restore OLAP Metadata Catalog as you installed it, without any OLAP models or metaoutlines that you might have created OLAP Metadata Catalog, including the OLAP models and metaoutlines that you have created
Tables Relating to the Metaoutline on page 89 Tables Relating to Drill-Through on page 91 Tables Relating to Hybrid Analysis on page 91 Miscellaneous Tables on page 92
Table 3
OLAP Model Tables Description Contains information about filters in the OLAP model. This information includes the hierarchy filter and system-generated ID numbers for parts of the OLAP model. Contains information about hierarchies in the OLAP model. This information includes transformations to perform on the hierarchy and system-generated ID numbers for parts of the OLAP model. Contains information about columns in a dimension table or a fact table. This information includes the column name, the column data type, the OLAP model it belongs to, the dimension or fact table it belongs to, the source table and column on which it is based, whether it is hidden in OLAP models and metaoutlines, whether it is used for Drill-Through, whether it is enabled for attributes, and how to create it if it is a user-defined column. Contains information about dimensions in the OLAP model. This information includes system-generated ID numbers for parts of the OLAP model. Contains information about dimensions in the OLAP model. This information includes the dimension name, the OLAP model the dimension belongs to, and whether the dimension is a measures dimension. Contains high-level information about an OLAP model: the name of the OLAP model, the description, the name of the data source on which the OLAP model is based, the owner, and access privileges. Contains property information for the OLAP model.
MB_TX_RULE
OA_INFO
OD_DETAILS
OD_INFO
OM_INFO
OM_PROPERTIES
88
Description Contains information about hierarchies in the OLAP model. This information includes the position of each column in the hierarchy, the order in which to build the hierarchy, and system-generated ID numbers for parts of the OLAP model. Contains information about hierarchies in the OLAP model. This information includes the hierarchy name, the OLAP model that it belongs to, and the dimension to which it belongs. Contains information about the dimension tables and the fact table in an OLAP model. This information includes the dimension or fact table name, the OLAP model it belongs to, its position in the right frame of the OLAP Model main window, and whether it is a dimension table or a fact table. Contains information about logical joins between dimension tables and fact tables. This information includes the name of the first column and the name of the column to which the first column is joined. Contains information about logical joins between dimension tables and fact tables. This information includes the name of the first dimension or fact table and the name of the dimension or fact table to which the first dimension is joined. Contains information about physical joins in the source tables. This information includes the name of the first source column and the name of the source column to which the first source column is joined. Contains information about physical joins in the source tables. This information includes the name of the first source table and the name of the source table to which the first source table is joined. Contains descriptions for a model, its dimensions, its members, and its hierarchies. This information includes the system-generated ID number for each model, dimension, member, or hierarchy that is described. Contains information about the database hints defined for a join. This information includes the type of join, the physical table to join, and the text of the join hint.
OMB_INFO
OV_INFO
OVL_REL_DETAILS
OVL_REL_INFO
OVP_REL_DETAILS
OVP_REL_INFO
OM_DESCRIPTIONS
JOIN_HINTS
Tables Relating to the OLAP Model on page 88 Tables Relating to Drill-Through on page 91 Tables Relating to Hybrid Analysis on page 91 Miscellaneous Tables on page 92
Table 4
Metaoutline Tables Description Contains information about metaoutline filters. This information includes the filter name and the actions the filter will perform.
89
Description Contains high-level information about a metaoutline: the name of the metaoutline, its description, the date it was last changed, the OLAP model on which the metaoutline is based, its owner, its access privileges, and whether the spreadsheet user of Drill-Through reports must enter the Integration Server name and a user name and password for the external data source. Contains information about metaoutline properties. This information includes the name and value of metaoutline properties, the name of the source table and column that Integration Services uses to assign the properties, the numeric range that represents a range specified by the user, Global Settings properties, and sort order. Sort order is based on a column in the data source, including columns not present in the OLAP model or metaoutline. Sort order can be ascending or descending. Also contains information on association of attributes (by level or by name) to base dimensions.
MO_PROPERTIES
MOD_ATTRIBUTES
Contains information about attribute dimensions, such as attribute dimension type, alias names, and transformation rules. Contains information about dimensions in a metaoutline. This information includes the dimension name, the OLAP dimension with which it is associated (unless it is a userdefined dimension), the type of dimension (a user-defined dimension, a measures dimension, a standard dimension, or an attribute dimension), and how members are positioned within the dimension. Contains information about members in a metaoutline. This information includes the name of the member, the dimension to which it belongs, whether it is a user-defined member or a standard member, the OLAP table with which it is associated, the OLAP model column with which it is associated, the name of its parent, how it is positioned with its siblings, and the database measure with which it is associated (if it is a user-defined member). Contains information about members in a metaoutline. This information includes member filters. Contains information about members in a metaoutline. This information includes member transformation rules. Contains information about database measures in a metaoutline. This information includes the name of the measure, the OLAP model table with which it is associated, the OLAP model column with which it is associated, and how it is positioned with its siblings. Contains information about database measures in a metaoutline. This information includes measure filters. Contains information about database measures in a metaoutline. This information includes the name and definition of measure transformation rules. Contains descriptions for a metaoutline, its filters, its dimensions, and its members. This information includes the system-generated ID number for each metaoutline, filter, dimension, or member that is described.
MOD_INFO
MS_INFO
MS_PREDICATES
MS_TX_RULE
MSR_INFO
MSR_PREDICATES
MSR_TX_RULE
MO_DESCRIPTIONS
90
Tables Relating to the OLAP Model on page 88 Tables Relating to the Metaoutline on page 89 Tables Relating to Hybrid Analysis on page 91 Miscellaneous Tables on page 92
Table 5
Drill-Through Tables Description Contains information about drill-through columns. This information includes the order in which Essbase Spreadsheet Add-in sorts and displays drill-through information. Contains information about drill-through columns. This information includes the name of the drill-through column, whether the spreadsheet wizard is displayed, and the time or row query governor specified. Contains information about drill-through columns. This information includes systemgenerated ID numbers for the OLAP model, the drill-through column, and the intersection level on which the column is defined. Contains information about drill-through filters. This information includes the Essbase user name and password, drill-through report name, and metaoutline name.
DTR_INFO
MO_INTERSECTION
MO_PROPERTIES
Tables Relating to the OLAP Model on page 88 Tables Relating to the Metaoutline on page 89 Tables Relating to Drill-Through on page 91 Miscellaneous Tables on page 92
Tables Relating to Drill-Through
91
Table 6
Hybrid Analysis Tables Description Contains information about Hybrid Analysis columns. This information includes systemgenerated ID numbers for the OLAP model; the Hybrid Analysis column; and column precision, scale, and nullability.
Miscellaneous Tables
Table 7 describes tables in OLAP Metadata Catalog that store miscellaneous information about OLAP models and metaoutlines and about Integration Services. For information about the other tables in OLAP Metadata Catalog, see:
Tables Relating to the OLAP Model on page 88 Tables Relating to the Metaoutline on page 89 Tables Relating to Drill-Through on page 91 Miscellaneous Tables on page 92
Table 7
Miscellaneous Tables Description Contains information about users accessing an OLAP model or metaoutline. This information includes the user name, the user session ID, the OLAP model or metaoutline that the user is accessing, whether the user is accessing an OLAP model or a metaoutline, whether the user has a read or read/write lock, and the date and time the user acquired the lock. Contains information about the Integration Services software you have installed. This information includes the release number of Integration Services.
CB_VERSION
92
5
In This Chapter
Overview of Sample Applications.................................................................................................93 Preparing to Set Up the Standard Sample Application.........................................................................95 Setting Up the Standard Sample Application Automatically...................................................................95 Setting Up the Standard Sample Application Manually ........................................................................97 Setting Up the TBC Relational Data Source .....................................................................................98 Setting Up the TBC_MD OLAP Metadata Catalog ............................................................................. 101 Setting Up the Hybrid Analysis Sample Application .......................................................................... 106 Setting Up the Unicode Sample Application .................................................................................. 115 After Setting Up the Sample Application ...................................................................................... 118
This chapter tells you what to consider before and after you set up the standard sample application, whether automatically or manually. For manual setup, the topics in this chapter detail which scripts and batch files you must run to create and load the TBC database and the TBC_MD OLAP Metadata Catalog. The procedures in this chapter assume that you know how to create a database using a relational database management system (RDBMS). For manual setup, the procedures assume that you know how to create tables and load them with data running SQL scripts, batch files, and shell scripts. For information on these topics, see the documentation for the RDBMS that you are using.
A database for a fictitious company called The Beverage Company (TBC) A sample OLAP model and metaoutline created from the TBC database A Hybrid Analysis-enabled OLAP model and metaoutline A special OLAP Metadata Catalog to store the sample OLAP models and metaoutlines and any additional OLAP models and metaoutlines that you create using the sample data
93
Optional: Setting up the Hybrid Analysis sample application Optional: Setting up the Unicode sample application
You can set up the standard sample application automatically using Integration Services Console (see Setting Up the Standard Sample Application Automatically on page 95), or you can set up the standard sample application manually (Setting Up the Standard Sample Application Manually on page 97. Using either method, setting up the standard sample application involves two main tasks:
You create two relational databases: one for the data in the TBC sample application and one for the OLAP Metadata Catalog database (TBC_MD).
Note:
Creating the relational databases is a manual task that you must perform prior to setting up the standard sample application, regardless of whether you choose the automatic or manual method.
You then create tables in each of these databases and load data into them, using scripts provided with Integration Services. You can create tables either automatically, through the use of Integration Services Console, or manually.
When you finish setting up the standard sample application, the OLAP Metadata Catalog database (TBC_MD) will contain an OLAP model (TBC Model) and a metaoutline (TBC Metaoutline) based on data in the sample TBC relational data source. You create the Hybrid Analysis sample application manually by completing the tasks in Setting Up the Hybrid Analysis Sample Application on page 106. Upon completion of these tasks, the sample TBC relational data source will contain Hybrid Analysis data. The OLAP Metadata Catalog will contain a Hybrid Analysis-enabled OLAP model (HA TBC Model) and metaoutline (HA TBC Metaoutline). You create the Unicode sample application manually by completing the tasks in Setting Up the Unicode Sample Application on page 115. Upon completion of these tasks, the sample TBC_U relational data source will contain Unicode enabled data. The Unicode TBC_MD_U OLAP Metadata Catalog will contain Unicode-enabled OLAP model (TBC_Model_Unicode) and metaoutline (TBC_Metaoutline_Unicode).
Caution!
If you have installed the sample application from a previous release of Integration Services, you should back up and preserve your existing sample database, OLAP Metadata Catalog, and the OLAP models and metaoutlines that OLAP Metadata Catalog contains. You can then upgrade your existing catalog to be compatible with the current release of the software (see Upgrading OLAP Metadata Catalogs Manually on page 81). You cannot, however, store newly created OLAP models and metaoutlines in your previous catalog.
94
Hyperion strongly recommends that you use the automatic installation process to install the standard sample applications (see Setting Up the Standard Sample Application Automatically on page 95).
If you choose to set up the standard sample application manually, refer to Table 8 and also verify that the appropriate client utility program is installed.
Table 8
Required Utilities for Setting Up the Standard Sample Application Utility Programs IBM DB2 Command Window or >DB2 -tvf DBAccess Query Analyzer SQL*Plus
Database IBM DB2 UDB Informix Microsoft SQL Server Oracle Sybase
Teradata
BTEQ
The remainder of this chapter assumes that you have installed the required database client software.
95
Metadata Catalog. You can create the standard sample application quickly and easily using the automatic feature. You are encouraged to use this convenient feature, but if you choose manual setup, see Setting Up the Standard Sample Application Manually on page 97 for detailed procedures. Hybrid Analysis: If you plan to set up the sample application for Hybrid Analysis, you must follow the procedures in Setting Up the Hybrid Analysis Sample Application on page 106.
Note:
ODBC connections must be configured before you begin to set up the standard sample application. If you have not yet configured ODBC connections, see Chapter 2, Configuring Data Sources for information on setting up ODBC data sources, including an OLAP Metadata Catalog and relational data sources.
Figure 3
2 In the Sample ODBC DSN text box, select or type the data source name for creating or dropping (deleting)
tables in the source database.
The User Name text box displays the name, TBC (lowercase tbc for Informix). You must use this name to create the TBC standard sample application database.
3 In the Password text box, type the password for the user to which you have assigned privileges to access
the TBC database for the standard sample application.
96
If Integration Services detects that you are not currently connected to the OLAP Metadata Catalog in which to load the sample OLAP model and metaoutline data, the Login dialog box is displayed so that you can enter the information necessary to connect. After you close the Login dialog box and are returned to the Sample Application Setup dialog box, click Create again to initiate setup of the standard sample application. If, after clicking Create, Integration Services detects either an existing sample data source or an existing sample OLAP model and metaoutline in the OLAP Metadata Catalog, you are prompted with the message that tables already exist in the database. Click OK to return to the Sample Application Setup dialog box to reenter information, or click Delete to delete the existing standard sample application data. If you click Delete, your existing OLAP Metadata Catalog sample OLAP models and metaoutlines will be deleted and replaced with the sample OLAP model and metaoutline for the current release. After you have completed creation of the Integration Services standard sample application, you are ready to begin working with Integration Services Console to create your own sample OLAP models and metaoutlines.
Hyperion strongly recommends that you use the automatic installation process to install the standard sample applications (see Setting Up the Standard Sample Application Automatically on page 95).
After you create the TBC and TBC_MD databases in a supported RDBMS, you use several scripts and files to create the standard sample application:
A script for creating the TBC sample database tables This script is located in the samples\tbcdbase directory.
A script for loading the TBC sample database data This script is located in the samples\tbcdbase directory.
A script for creating the TBC_MD OLAP Metadata Catalog database tables This script is located in the ocscript directory.
XML files for importing a sample OLAP model and metaoutline in the TBC_MD OLAP Metadata Catalog database
97
These XML files are located in the samples\tbcmodel directory. The scripts for creating the standard sample application and OLAP Metadata Catalog are installed with the server software during Integration Services installation. Different versions of the scripts are provided for each of the supported RDBMSs.
Note:
Some of the standard sample application scripts require slightly different procedures, depending on the RDBMS that you are using.
To create a database, you must have database administrator or similar access privileges required by the RDBMS that you are using.
Create a database device or tablespace named TBC. Allot 20 MB for storage. Create a user TBC who can drop and create tables. Grant user privileges or permissions to create and drop tables.
Caution!
The user TBC must create the tables for the TBC database, or portions of the standard sample application will not work.
98
Tested Utilities for Creating TBC Tables SQL Script tbc_create_db2.sql tbc_drop_db2.sql Utility Program IBM DB2 Command Window or >DB2 -tvf
Informix
tbc_create_informix.sql tbc_drop_informix.sql
DBAccess
tbc_create_sqlsrv.sql tbc_drop_sqlsrv.sql
Query Analyzer
Oracle
tbc_create_oracle.sql tbc_drop_oracle.sql
Sybase
tbc_create_sybase.sql tbc_drop_sybase.sql
ISQL
Teradata
tbc_create_teradata.sql tbc_drop_teradata.sql
BTEQ
Each RDBMS has two scriptsone to build tables and another to drop tables.
Caution!
You must create the tables for the TBC database as user TBC, or portions of the standard sample application will not work.
Use lowercase tbc for Informix; use uppercase TBC for Sybase and Microsoft SQL Server.
Setting Up the TBC Relational Data Source
99
3 In the samples\tbcdbase directory, open the appropriate SQL script file. 4 Run the SQL script.
Microsoft SQL Server: You should receive a message that you did not create data or rows. This message is normal because you created only tables and columns. Informix: If you run the Informix SQL script (tbc_create_informix.sql) more than once, you must first run tbc_drop_informix.sql to drop tables before you build them again.
5 Verify that you have created the TBC tables; for example, type
SELECT * FROM PRODUCTDIM
or start the RDBMS and verify that the TBC database has the new tables.
Tested Utilities for Loading Data into TBC Tables SQL Script sampledata.sql sampledata.sql sampledata.sql sampledata.sql sampledata.sql sampledata.sql Utility Program IBM DB2 Command Window or >DB2 -tvf DBAccess Query Analyzer SQL*Plus ISQL BTEQ
Database IBM DB2 Informix Microsoft SQL Server Oracle Sybase Teradata
3 Load data into the TBC_MD database using the XML import utility.
For instructions, see Using the XML Import Utility to Load Data into the TBC_MD Tables on page 104.
Note:
If you have a previous release of Integration Services and are upgrading your OLAP Metadata Catalog, you cannot roll back to the previous version. The new version of the OLAP Metadata Catalog is not compatible with earlier releases of Integration Services.
Create a database device or tablespace named TBC_MD. Allot 20 MB for storage. Create a user TBC who can drop and create tables. Grant user privileges or permissions.
Caution!
You must create the tables for the TBC_MD database as user TBC or portions of the standard sample application will not work.
The standard sample application SQL scripts used to create tables for the TBC_MD database are in the ocscript directory where you installed Integration Server. The SQL scripts in the ocscript directory are the same scripts that you use to create any OLAP Metadata Catalog. For information on OLAP Metadata Catalogs, see Chapter 3, Creating, Upgrading, and Deleting OLAP Metadata Catalogs. The utilities listed in Table 11 have been tested to work with the SQL scripts:
Table 11
Tested Utilities for Creating TBC_MD Tables SQL Script oc_create_db2.sql oc_drop_db2.sql oc_upgrade20_db2.sql oc_upgrade61_db2.sql oc_upgrade65_db2.sql oc_upgrade651_db2.sql Utility Program IBM DB2 Command Center or >DB2 -tvf
Informix
DBAccess
Query Analyzer
Oracle
SQL*Plus
Database
Utility Program
Sybase
ISQL
Teradata
BTEQ
Note:
The scripts used to upgrade OLAP Metadata Catalog manually from Release 2.0 to Release 6.2 are identical to those used to upgrade from Release 2.0 to Release 6.1. Therefore, the names have not been changed from the names used in Integration Services Release 6.1. If you are upgrading OLAP Metadata Catalog from Release 2.0 to Release 6.2, use the scripts named *61*.sql that are appropriate for your RDBMS. Integration Services provides five SQL scripts for each RDBMS (with the exception of Teradata):
oc_create_database_name.sql to build tables oc_drop_database_name.sql to drop tables oc_upgrade20_database_name.sql to upgrade tables from Integration Services Release
Release 6.5 to 6.5.1 If you need to rebuild tables, you must first drop the tables before you build them again.
Caution!
You must create the tables for the TBC_MD database as user TBC or portions of the standard sample application will not work.
5 Verify that you have created the TBC_MD tables; for example, type the following command:
SELECT * FROM TBC.MS_INFO
or start the RDBMS and verify that the TBC_MD database has the new tables.
Using the XML Import Utility to Load Data into the TBC_MD Tables
Use the XML Import utility to load OLAP model and metaoutline data into the TBC_MD tables. The XML files listed in Table 12 have been tested to work with their respective RDBMS. These files are located in the samples\tbcmodel directory.
Table 12
XML Files for Loading Data into TBC_MD Tables XML File in sample\tbcmodel sample_model.xml sample_metaoutline.xml
Database
Informix
sample_model_informix.xml sample_metaoutline.xml
2 On the Windows desktop, click Start, and select All Programs, then Oracle EPM System, then Essbase,
then Integration Services, then Console.
3 If necessary, click Close to clear the OLAP Metadata Catalog Setup dialog box.
Integration Services Console automatically displays the Login dialog box. a. In the Server text box, under Integration Services, select or type the name of an Integration Server computer; for example, cypress. If the port number on which Essbase Integration Server communicates with the console has been set to a value other than the default, you must type the server name or IP address and type the non-standard port number, separated by a colon; for example:
cypress:3390
b. c.
In the OLAP Metadata Catalog text box, select or type the name of the OLAP Metadata Catalog, TBC_MD, for the sample TBC (The Beverage Company) database. In the User Name and Password text boxes, type the user name and password for the standard sample application user, TBC, and click OK.
To load sample OLAP model metadata into the TBC_MD tables using XML Import:
1 In Integration Services Console, select File, then XML Import/Export. 2 Select the Import tab. 3 In the XML File Path text box, enter the file path for the sample OLAP model or click the Open XML File
button and navigate to the following folder:
\eis\server\samples\tbcmodel
4 Select the XML file to import for the sample OLAP model, sample_model.xml.
Informix: Select the XML file to import for the sample OLAP model for Informix, sample_model_informix.xml.
5 Click OK.
Integration Services Console displays the XML Import/Export dialog box with the XML file path and name in the XML File Path text box. The name of the sample OLAP model is displayed in the OLAP Model Name text box.
6 Click the Import to Catalog button to load the selected XML file into the sample OLAP Metadata Catalog.
To load sample OLAP metaoutline metadata into the TBC_MD tables using XML Import:
1 In Integration Services Console, select File > XML Import/Export. 2 Select the Import tab. 3 In the XML File Path text box, enter the file path for the sample OLAP metaoutline or click the Open XML
File button and navigate to the following file:
\eis\server\samples\tbcmodel
4 Select the XML file to import for the sample OLAP metaoutline, sample_metaoutline.xml. 5 Click OK.
Integration Services Console displays the XML Import/Export dialog box with the XML file path and name in the XML File Path text box, the name of the sample OLAP metaoutline in the OLAP Metaoutline Name text box, and the name of the OLAP model upon which this metaoutline is based in the Based on OLAP Model text box.
6 Click the Import to Catalog button to load the selected XML file into the sample OLAP Metadata Catalog.
The Hybrid Analysis application scripts require slightly different procedures, depending on the RDBMS that you are using. The utilities listed in Table 9 are required for running the various batch files and SQL scripts used in creating the Hybrid Analysis sample application.
Table 13
Required Utilities for Setting Up the Hybrid Analysis Sample Application Required Utilities DB2CMD command line interface
SQL*Plus SQL*Loader command line utility ISQL BCP command line utility BTEQ FastLoad
Teradata
An XML file that you use to import a sample OLAP model into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Model.xml
An XML file that you use to import a sample metaoutline into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Metaoutline.xml
Setting up the Hybrid Analysis sample application on IBM DB2 consists of performing the following tasks:
Running the batch file install_db2.bat Importing the sample OLAP model and metaoutline
Note:
Before setting up the Hybrid Analysis Sample Application on IBM DB2, verify that the DB2CMD command line interface is installed.
This batch file sends SQL commands to the DB2CMD command line interface. Pass the following three parameters to install_db2.bat:
The user ID that you use to connect with IBM DB2 Your password The IBM DB2 database name
The install_db2.bat file runs the DB2CMD command line interface, which executes the ha_create_db2.sql load script file. This script file performs the following tasks:
Deletes any existing tables in the sample TBC database Creates a new set of tables in the TBC database Inserts records into the dimension tables
Creates the ha_results_createtables.txt file Loads the sample data into the fact table
The ha_results_createtables.txt file contains the results of the table-creating and loading operations.
Note:
After the member and data load is finished, close the IBM DB2 command window.
3 Select the Import tab. 4 In the XML File Path text box, enter the file path for the sample OLAP model or click the Open XML File
button and navigate to the following file:
\eis\server\samples\tbcmodel\HA_TBC_Model.xml
2 Select the Import tab. 3 In the XML File Path text box, enter the file path for the sample OLAP metaoutline or click the Open XML
File button and navigate to the following file:
\eis\server\samples\tbcmodel\HA_TBC_Metaoutline.xml
ha_create_oracle.sql
An XML file that you use to import a sample OLAP model into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Model.xml
An XML file you use to import a sample metaoutline into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Metaoutline.xml
Setting up the Hybrid Analysis sample application on Oracle consists of performing the following tasks:
Running the batch file install_oracle.bat Importing the sample OLAP model and metaoutline
Note:
Before setting up the Hybrid Analysis Sample Application on Oracle, verify that SQL*PLUS and SQLLDR command line interfaces are installed.
This batch file sends SQL commands to the SQL*PLUS command line interface. Pass the following three parameters to install_oracle.bat :
The user ID that you use to connect with Oracle Your password The Oracle service name as defined in the TNSNAMES.ORA file
Deletes any existing tables in the sample TBC database Creates a new set of tables in the TBC database
Inserts records into the dimension tables Creates the ha_results_createtables.txt file
Loads the sample data into the fact table Creates the ha_results_loadfact.txt file
3 Select the Import tab. 4 In the XML File Path text box, enter the file path for the sample OLAP model or click the Open XML File
button and navigate to the following file:
\eis\server\samples\tbcmodel\HA_TBC_Model.xml
2 Select the Import tab. 3 In the XML File Path text box, enter the file path for the sample OLAP metaoutline or click the Open XML
File button and navigate to the following file:
\eis\server\Samples\tbcmodel\HA_TBC_Metaoutline.xml
An XML file that you use to import a sample OLAP model into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Model.xml
An XML file that you use to import a sample metaoutline into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Metaoutline.xml
Setting up the Hybrid Analysis sample application on Microsoft SQL Server consists of performing the following tasks:
Running the install_sqlsrv.bat batch file Importing the sample OLAP model and metaoutline
Note:
Before setting up the Hybrid Analysis Sample Application on Microsoft SQL Server, verify that you have ISQL and BCP (Bulk Copy) installed. These utilities were most likely included when you installed SQL Server Client.
This file sends commands to the Interface SQL (ISQL) command line interface. These commands install the sample application to the sample database TBC. Pass the following three parameters to install_sqlsrv.bat:
The user ID that you use to connect with the Microsoft SQL Server (TBC) Your password The Microsoft SQL Server name (name of machine where the SQL Server is installed)
Deletes any existing tables in the sample TBC database Creates a new set of tables in the TBC database Inserts records into the dimension tables Creates the ha_results_createtables.txt file
Loads the sample data into the fact table Creates the ha_results_loadfact.txt file
3 Select the Import tab. 4 In the XML File Path text box, enter the file path for the sample OLAP model or click the Open XML File
button and navigate to the following file:
\eis\server\samples\tbcmodel\HA_TBC_Model.xml
2 Select the Import tab. 3 In the XML File Path text box, enter the file path for the sample OLAP metaoutline or click the Open XML
File button and navigate to the following file:
\eis\server\Samples\tbcmodel\HA_TBC_Metaoutline.xml
An SQL file to add constraints to the fact table after the data load has been completed:
ha_altertable_teradata.sql
An XML file that you use to import a sample OLAP model into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Model.xml
An XML file that you use to import a sample metaoutline into the TBC_MD OLAP Metadata Catalog database:
HA_TBC_Metaoutline.xml
Setting up the Hybrid Analysis sample application on Microsoft SQL Server consists of performing the following tasks:
Running the batch file install_teradata.bat Importing the sample OLAP model and metaoutline
Note:
Before setting up the Hybrid Analysis Sample Application on Teradata, verify that the BTEQ and FastLoad utilities are installed.
This batch file sends SQL commands to the BTEQ command line interface. Pass the following three required parameters to install_teradata.bat:
The user ID that you use to connect with Teradata Your password The Teradata database name (DBC)
You can also pass, as an optional parameter, the database name where you want to install the sample application. (The default database is TBC.)
Deletes any existing tables in the sample TBC database Creates a new set of tables in the TBC database Inserts records into the dimension tables Creates the ha_results_createtables.txt file
Loads the sample data into the fact table using the ha_fastload_teradata.txt file Creates the ha_results_loadfact.txt file
Deletes a few tables no longer needed in the sample TBC database Creates indexes and places constraints on the fact table Creates the ha_results_altertables.txt file
3 Select the Import tab. 4 In the XML File Path text box, enter the file path for the sample OLAP model or click the Open XML File
button and navigate to the following file:
\eis\server\Samples\tbcmodel\HA_TBC_Model.xml
2 Select the Import tab. 3 In the XML File Path text box, enter the file path for the sample OLAP metaoutline or click the Open XML
File button and navigate to the following file:
\eis\server\Samples\tbcmodel\HA_TBC_Metaoutline.xml
The Unicode sample application in Integration Services is supported only on Oracle relational database management systems (RDBMSs).
The utilities listed in Table 14 are required for running the various batch files and SQL scripts used in creating the Unicode sample application.
Table 14
Required Utilities for Setting Up the Unicode Sample Application Required Utilities
Database Oracle
Create a database device or tablespace named TBC_U. Allot 20 MB for storage. Create a user TBC_U who can drop and create tables. Grant user privileges or permissions to create and drop tables.
Caution!
You should create the tables for the Unicode-enabled database as user TBC_U, or portions of the Unicode sample application may not work.
The Unicode sample application in Integration Services is supported only on Oracle relational database management systems (RDBMSs).
A table creation and data load script for the Unicode sample database:
tbc_create_database_name_unicode.sql
An XML file that you use to import a sample OLAP model into the TBC_MD_Unicode OLAP Metadata Catalog database:
TBC_Model_Unicode.xml
An XML file that you use to import a sample metaoutline into the TBC_MD_Unicode OLAP Metadata Catalog database:
TBC_Metaoutline_Unicode.xml
Setting up the Unicode sample application consists of performing the following tasks:
Running the SQL file tbc_create_database_name_unicode.sql Importing the Unicode sample model Importing the Unicode sample metaoutline
Ensure you have created a Unicode-enabled database for your sample application as described in Creating the Unicode TBC_U Database on page 116.
Deletes any existing tables in the Unicode sample database Creates a new set of tables in the Unicode sample database Loads Unicode sample metadata into the dimension tables Loads Unicode sample data into the fact table
3 Select the Import tab. 4 In the XML File Path text box, enter the file path for the Unicode sample OLAP model or click the Open XML
File button and navigate to the following file:
eis\server\samples\tbcmodel\TBC_Model_Unicode.xml
6 Click OK.
2 Select the Import tab. 3 In the XML File Path text box, enter the file path for the Unicode sample metaoutline or click the Open XML
File button and navigate to the following file:
\eis\server\Samples\tbcmodel\TBC_Metaoutline_Unicode.xml
5 Click OK. 6 Click Close to close the XML Import/Export dialog box.
You must connect to TBC and TBC_MD as user TBC unless you create user name aliases or synonyms in the RDBMS.
To see TBC tables and columns in the left frame of Integration Services Console:
1 Start Integration Server. 2 Start Integration Services Console. 3 In a blank Integration Services Console window, select Connections, then OLAP Metadata Catalog, then
Connect, and connect to the OLAP Metadata Catalog, TBC_MD.
4 In the New tab, double-click the OLAP Model icon. 5 In the Data Source dialog box, connect to the TBC sample database.
To see the OLAP model (TBC Model) in the right frame of Integration Services Console:
1 Start Integration Server. 2 Start Integration Services Console. 3 If the Login dialog box is not already displayed, in a blank Integration Services Console window, select
Connections, then OLAP Metadata Catalog, then Connect.
4 Connect to the OLAP Metadata Catalog TBC_MD. 5 Click the Existing tab, select TBC Model, and click Open.
Hybrid Analysis: To view the Hybrid Analysis-enabled OLAP model, select HA TBC Model. Duplicate Member (Oracle only): To view the Duplicate Member-enabled OLAP model, select Duplicate Member TBC Model.
6 In the Data Source dialog box, connect to the TBC sample database.
Viewing Metaoutlines
After you set up the sample application and configure TBC and TBC_MD by mapping them to supported ODBC drivers, you can view metaoutlines in Integration Services Console.
To see the metaoutline (TBC Metaoutline) in the right frame of Integration Services Console:
1 Start Integration Server. 2 Start Integration Services Console. 3 If the Login dialog box is not already displayed, in a blank Integration Services Console window, select
Connections, then OLAP Metadata Catalog, then Connect.
4 Connect to the OLAP Metadata Catalog TBC_MD. 5 Click the Existing tab, then click the plus symbol (+) to the left of TBC Model to expand the view, and select
TBC Metaoutline.
Hybrid Analysis: To view the Hybrid Analysis-enabled metaoutline, click the plus symbol (+) to the left of HA TBC Model to expand the view, and select HA TBC Metaoutline. Duplicate Member (Oracle only): To view the Duplicate Member-enabled metaoutline, click the plus symbol (+) to the left of Sample_duplicate member to expand the view, and select Duplicate Member TBC Metaoutline.
6 Click Open.
The Data Source dialog box is displayed.
7 In the Data Source dialog box, connect to the TBC sample database.
6
In This Chapter
Working with Users............................................................................................................... 121 Working with Locks .............................................................................................................. 122 Working with Permissions ....................................................................................................... 125
This chapter describes how to view information about users, OLAP models, and metaoutlines in OLAP Metadata Catalog. It also describes how to clear locks and change permissions on OLAP models and metaoutlines. See Chapter 4, Tables in OLAP Metadata Catalog.
Integration Services users who can access OLAP models and metaoutlines stored in OLAP Metadata Catalogs. (Integration Services users are the database users defined in the data source that contains the OLAP Metadata Catalog that you are using.) Essbase users who can access the Essbase database that you create from a metaoutline Source database and data warehouse users who can access the tables and views in the specified source database or data warehouse
Use the tools provided with each data source to manage each set of users. Consult the data source documentation if you are not sure how to perform any of the following tasks:
Create new users View a list of available users Change the permissions for users Delete current users Disconnect users
If you plan only to view an OLAP model or metaoutline, use standard access mode. If you plan to edit an OLAP model or metaoutline, use exclusive access mode.
If you opened and edited an OLAP model or metaoutline in standard access mode, and are
concerned that other users may overwrite your changes, perform one of the following actions:
Save the OLAP model or metaoutline using a different name. See Integration Services Console Help. If you have not yet started editing the document or have entered only a few changes, close the OLAP model or metaoutline and then reopen it using exclusive access mode and reapply the changes. Contact the users with read locks and ask them to close the OLAP model or metaoutline. To view a list of users with read locks, see Viewing Integration Services Users with Locks on page 123.
Integration Services gives you a write lock on that OLAP model that remains in effect until you close the model. When you open a metaoutline in exclusive access mode, Integration Services gives you a write lock on that metaoutline and a read lock on the OLAP model on which the metaoutline is based that remains in effect until you close the metaoutline. While you have an OLAP model or metaoutline open in exclusive access mode, other users can open and view the documents, but they cannot save them.
CB_USER_NAME: the name of the user accessing the OLAP model or metaoutline; for example, sa. CB_USER_SESSION_ID: a system-generated reference number that uniquely identifies an editing session CB_OBJECT_ID: an OLAP model or metaoutline number CB_OBJECT_TYPE: a type reference that indicates whether the user is accessing an OLAP model or a metaoutline
CB_OBJECT_LOCK: a lock reference number that indicates whether the user has a read or write lock
The number 1 represents a read lock. The number 2 represents a write lock.
CB_LOCK_TIMESTAMP: the date and time that the user acquired the lock
You can determine the following information from the sample CB_CONCURRENCY table shown in the preceding list:
The first row of the results tells you that the sa user (CB_USER_NAME = sa) has a read lock (CB_OBJECT_LOCK = 1) on an OLAP model (CB_OBJECT_TYPE = 1) with an ID of 889844639. The second row of the results tells you that the sa user (CB_USER_NAME = sa) has a read lock (CB_OBJECT_LOCK = 1) on a metaoutline (CB_OBJECT_TYPE = 2) with an ID of 889845263. The third row of the results tells you that the sa user (CB_USER_NAME = sa) has a read lock (CB_OBJECT_LOCK = 1) on an OLAP model (CB_OBJECT_TYPE = 1) with an ID of 889844639. The fourth row of the results tells you that the sa user (CB_USER_NAME = sa) has a read lock (CB_OBJECT_LOCK = 1) on a metaoutline (CB_OBJECT_TYPE = 2) with an ID of 892167813.
When you open a metaoutline, you receive a read lock on the metaoutline and on the OLAP model on which it is based; therefore, you can assume that the sa user is working on two different metaoutlines based on the same OLAP model.
Make sure the users do not need their locks before you delete the locks.
Allow read/write access for other users. This setting permits all other users to read and write to the OLAP model or metaoutline. This setting is the default. Allow read access for other users. This setting permits all other users to read but not write to the OLAP model or metaoutline; that is, other users cannot save changes to the OLAP model or metaoutline. Disallow all access for other users. This setting denies all other users read or write permission to the OLAP model or metaoutline. You are the only user who can read or write to it.
To change the permissions of an OLAP model or metaoutline, take one of the following actions:
Change the OLAP model properties in Integration Services Console. See the Integration Services Console Help. Save the metaoutline with a different name by selecting File, then Save As. See the Integration Services Console Help. To edit tables containing the permissions for the OLAP model or metaoutline, use the tools provided with the data source that contains OLAP Metadata Catalog. See Viewing Permissions for OLAP Models on page 125 or Viewing Permissions for Metaoutlines on page 126.
MODEL_ID: a system-generated reference number. MODEL_NAME: the name of the OLAP model; for example, TBC Model.
MODEL_DESCRIPTION: a description of the OLAP model. If you do not enter a description when you save the model, this column is blank. MODEL_DATA_SOURCE: the name of the Open Database Connectivity (ODBC) data source on which the OLAP model is based; for example, TBC. MODEL_OWNER: the login name of the OLAP model owner; for example, sa. The login name is specified in the data source that contains the OLAP Metadata Catalog. MODEL_ACCESS_CODE: a reference number that indicates what level of access users, other than the owner, have to the OLAP model.
The number 0 represents no permissionsother users can neither read nor write to the OLAP model. The number 1 represents read accessother users can read the OLAP model but cannot write to it. The number 2 represents read and write accessother users can both read and write to the OLAP model; this is the default.
To view access permissions for all OLAP models in OLAP Metadata Catalog:
1 Issue the following SELECT statement in the OLAP Metadata Catalog database, using the tools for the data
source.
MODEL_ID: a system-generated reference number for the OLAP model on which the metaoutline is based. MO_NAME: the name of the metaoutline; for example, TBC Metaoutline. MO_DESC: a description of the metaoutline. If you do not enter a description when you save the metaoutline, this column is blank. MO_CHANGE_DATE: the date on which changes were last made to the metaoutline. MO_OWNER: the login name of the metaoutline owner; for example, sa. The login name is specified in the database that contains the associated OLAP Metadata Catalog. MO_ACCESS_CODE: a reference number that indicates what level of access users, other than the owner, have to the metaoutline.
The number 0 represents no permissionsother users can neither read nor write to the metaoutline. The number 1 represents read accessother users can read the metaoutline but cannot write to it. The number 2 represents read and write accessother users can both read and write to the metaoutline; this setting is the default.
To determine the access permissions for all metaoutlines in OLAP Metadata Catalog:
1 Issue the following SELECT statement in the OLAP Metadata Catalog database, using the tools for the data
source.
SELECT MO_NAME,MO_OWNER,MO_ACCESS_CODE FROM MO_INFO
7
In This Chapter
Common Connection Problems................................................................................................. 129 Database-Specific Problems.................................................................................................... 131 ODBC and Connection Problems ............................................................................................... 131 Isolating Connection Problems ................................................................................................. 132 Correcting Connection Problems ............................................................................................... 144 Using ODBC Tracing.............................................................................................................. 151
Integration Services requires several layers of software components to connect to data sources and load data into Essbase databases. Each of the components must be configured properly to run and communicate with each other. Consequently, diagnosing and correcting connection problems in the system can be challenging. This chapter assists you in isolating and correcting connection problems that can occur during the operation of Integration Services. The first topic of this chapter lists common causes for connection problems. The subsequent topics outline a strategy for isolating and correcting connection problems not addressed by the common problems list. This chapter assumes that you know how to perform basic database administration tasks, such as using a database client utility program to connect to a database server, navigating directories using a command prompt, and editing configuration files.
Does the user name you used have the correct privileges to access both OLAP Metadata Catalog and the data source at the database level? If you are trying to connect to OLAP Metadata Catalog, did you use the same user name and password as the user who created the tables in OLAP Metadata Catalog?
If you create an OLAP Metadata Catalog when logged in as one user name, you cannot access the tables in OLAP Metadata Catalog using a different user name unless you create an alias for the user name (for Microsoft SQL Server) or synonyms for the tables (for IBM DB2 and Oracle).
Essbase Integration Server The database servers that manage OLAP Metadata Catalog and the data source databases The data source database listener for OLAP Metadata Catalog and the data source
Is the database client software installed and configured on the computer where Integration Server is installed?
Do the database client software utility programs, such as the command-line SQL utility, run? Can you connect to databases with them? Does the user who starts Integration Server have the correct environment variables set and the necessary access privileges to run the database client software? Are the required environment variables for the database client software, such as the path (and the library path on UNIX), configured on the computer where Integration Server is installed?
On UNIX, use ivtestlib located in the eis/server/bin directory to confirm that the library path contains all files required by the open database connectivity (ODBC) driver.
If necessary, is the bin directory for each database in the operating system path of the user who is running Integration Server? Does the ODBC data source name (DSN) configuration identify the name of the data source client?
On UNIX, use the odbcconfig utility to check the path and environment variable, and to test connectivity to ODBC.
Are OLAP Metadata Catalog and the data source configured as ODBC data sources on the Integration Server computer? For information on configuring ODBC data sources, see Chapter 2, Configuring Data Sources. Are you using the ODBC data source names for OLAP Metadata Catalog and the data source databases as configured on the Integration Server computer? Do not assume that the data source database name is the same as the ODBC data source name for that database. Are you experiencing time-out problems when connecting to the external data source?
Wait and retry the action when the data source is less busy. To avoid the problem, increase the ODBC driver time-out limit using ODBC Administrator on Windows systems. See the ODBC documentation.
Did you create tables in OLAP Metadata Catalog? For information on creating tables for the catalog, see Chapter 3, Creating, Upgrading, and Deleting OLAP Metadata Catalogs. Are you having problems related to ODBC while you are creating an Essbase outline? If so, turn on tracing, perform the action again, and analyze the results in the tracing log file. See Using ODBC Tracing on Windows Systems on page 151.
If you experience problems connecting to the data source, view the log for the data source for information about why the connection failed. See your data source documentation. If none of the above steps help, review the list of database-specific problems in Database-Specific Problems on page 131.
Database-Specific Problems
If you have problems connecting to OLAP Metadata Catalog or to the external data source with Integration Services Console, review the following list of common problems for specific databases:
IBM DB2: You may be required to bind the database driver to a IBM DB2 server that is managing the databases you want. See the IBM DB2 installation documentation. Microsoft SQL Server: If you try to access a Microsoft SQL Server database with the Microsoft-supplied ODBC driver when you do not have access permission, SQL Server connects you to the default database and does not notify you. Configure the ODBC DSN to use the database that you want to access. See the Microsoft SQL Server documentation.
If none of the previous suggestions help, proceed to ODBC and Connection Problems on page 131 for an overview of connection problems, then see Isolating Connection Problems on page 132.
The components of Integration Services or the data source are not installed or are installed on the wrong computer. The components do not run because they are not set up properly. The components stopped running or were never started. The components are not compatible with the operating system. The versions of the different components are not compatible with each other.
The software components are not properly configured to communicate with each other. User names, passwords, and permissions are configured or used inconsistently across the components and computers. The communication interfaces of the software components are incompatible with each other.
Changes to components, databases, users, or permissions have put the software components out of sync with each other.
A physical network path does not exist between the computers that must communicate with each other. One or both computers attempting to communicate do not have the TCP/IP network protocol installed or configured. The computer attempting to contact another computer does not have access to addressing information, such as a domain name server (DNS). Security systems prevent all communication or certain types of communication between the computers that must talk with each other. For example, this is the case when client and server computers are separated by a firewall.
Figure 4
The numbers in Figure 4 identify the general diagnostic steps that you should take to isolate connection problems. Follow these steps as indicated, starting with the connection between the database client software and the data source (1) and working backwards through the system to isolate connection problems. After you isolate a problem, refer to the appropriate components topic in Correcting Connection Problems on page 144.
2 Attempt to connect to the data source databases using ODBC. See Testing the ODBC Connection to the
Data Source on page 135.
3 Attempt to connect to the data source databases using Integration Server. See Testing the Integration Server
Connection to the Data Source on page 139.
4 Attempt to connect to the data source databases using Integration Services Console. See Testing the
Integration Services Console Connection to Essbase Server on page 143.
5 Attempt to connect to Essbase Server using Integration Server. See Testing the Integration Server Connection
to Essbase Server on page 142.
6 Attempt to connect to Essbase Server using Integration Services Console. See Testing the Integration Server
Connection to Essbase Server on page 142.
Logging on as a different user, such as an administrative user, may fail to reveal problems in the configuration.
2 Start a database client utility program with which you can use SQL statements to connect to and query
databases, such as a command-line SQL utility.
If the database client utility does not start, check to make sure that the client software is installed and the required environment variables are set. See Correcting Database Client Problems on page 146.
3 Connect to the OLAP Metadata Catalog database in the data source using a valid database user account.
If the connection fails, try to connect as a user who you know has access permission, such as an administrative user. If the administrative connection succeeds, check the other user accounts and permissions in the data source for the OLAP Metadata Catalog database, and carefully check user accounts with connection problems. See, Correcting Data Source Problems on page 146.
4 Execute an SQL select statement against a table in the OLAP Metadata Catalog database; for example, if
you are connected to the sample OLAP Metadata Catalog, type
SELECT * FROM TBC.OV_INFO
If the select statement fails, check the permissions for the user name you are using to connect to the data source. See Correcting Data Source Problems on page 146. Also, check that the version
of the database client software you are using is compatible with the version of the data source. See Correcting Database Client Problems on page 146.
5 After you have successfully completed the preceding steps, try connecting to OLAP Metadata Catalog from
Integration Services Console.
See Testing the Integration Services Console Connection to the Data Source on page 141. If the connection fails, proceed to Testing the ODBC Connection to the Data Source on page 135. If you still cannot complete a connection and select statement with the user who runs Integration Server, contact technical support for the data source.
3 In the ODBC Data Source Administrator dialog box, click the System DSN tab. 4 In the System Data Sources list, select the data source created for OLAP Metadata Catalog or the data
source database, and click Configure.
A data source configuration dialog box is displayed. Depending on the data source and the driver that you are using, you may be asked to log on to the database immediately, or you may be asked to review all the settings before testing the connection. In either case, after attempting to log on to the database, observe the resulting message boxes to determine if you connected to the data source successfully.
Note:
To be accessible from other computers, ODBC data sources must be configured as System DSNs, not User DSNs.
5 If you cannot connect to the database, check the Integration Server name, database name, user name, and
password information for accuracy, and make sure that you are using the correct ODBC driver for the data source.
6 After you have successfully completed the preceding steps, try connecting to the data source from Integration
Services Console.
See Testing the Integration Services Console Connection to the Data Source on page 141. If the connection from Integration Services Console fails, proceed to Testing the Integration Server Connection to the Data Source on page 139.
4 If the required DSN does not exist, add it by completing the following steps:
Type 2 and press Enter. At the prompts, type a database type, data source name, and a description. You may also be prompted for additional information depending on the data source platform of the DSN that you are adding, such as server name, host name, or database name. To verify the addition of the DSN, type 1 and press Enter. Verify that the DSN you just added is included in the list.
You can delete a DSN by typing 3 at the starting odbcconfig utility window. Use the List option (1) to verify that the DSN has been deleted.
3 To test the ODBC connection, type 4 and press Enter. 4 At the prompts, type a data source name, user name, and password.
If no errors are returned, the connection is valid. In the event of errors, proceed to Step 5 to add corrected DSN information.
6 Type 1 and press Enter to list DSNs and verify that the DSN you just added is included in the list. 7 Type 4 and press Enter to test the new ODBC connection.
If no errors are returned, the connection is valid.
8 Repeat the steps in this section until the connection tests return no errors. 9 Type 5 and press Enter to exit the utility.
Note:
The primary login script file may be named differently, depending on the UNIX operating system and the system administrator.
3 Check the main login script file for the inclusion of the Integration Services script (is.sh or is.csh).
If one of these scripts is included in the main login script, make sure that the inclusion of the script file name follows the syntax for executing a script file within another script, and that the script file is specified with a complete directory path. See Correcting ODBC Problems on page 148.
Note:
If you make a change to the login scripts, log out and then log back on to reset the environment variables.
4 After completing the preceding steps, test the connection by following the steps listed in Testing the
Integration Services Console Connection to Essbase Server on page 143.
If you cannot make a connection, proceed to Testing the Integration Server Connection to the Data Source on page 139.
If this command does not display a file name and directory location, then you have not included the Integration Services script in the login script for the user. See Correcting Integration Server Problems on page 149.
3 Move to the directory specified by $ODBCINI and open the odbc.ini file with a text editor.
Ensure that you are using the exact file name and directory that you have set. For example, type vi $ODBCINI to display the file name and path of the $ODBCINI environment variable.
4 Check that the name of the data sources you are using with Integration Services are listed in the ODBC Data
Sources section; for example:
[ODBC Data Sources] TBC_MD=Essbase Integration Services sample catalog TBC=Essbase Integration Services sample database
5 For each data source, check that there is a section starting with the name listed in the ODBC Data Sources
section, enclosed in brackets; for example:
For Oracle:
[myoracle] Driver= $HYPERION_Home/common/ODBC/Merant/5.2/Drivers/ARor819.so Description=my oracle source ServerName=mytnsServerName
6 Within the section for each data source, verify that an ODBC driver file is specified.
Make sure that the driver file exists by exiting the text editor and attempting to get a directory listing for that file; for example
ls /home/db2inst1/sqllib/lib/db2.so
If the file is listed, use ivtestlib to check the actual name and location of the driver, then make the appropriate changes to the odbc.ini file.
Tip:
Copy the directory and file name for the driver from the odbc.ini file and paste it into an ls command.
7 Within the section for each data source, verify that the database name, computer name, and other required
information are specified.
The following procedure uses the sample OLAP Metadata Catalog and sample database. For your tests, substitute the OLAP Metadata Catalog and data source information you are using.
This test uses the LOADMEMBER command to test connections. When the loadmember starts a load, it attempts to make connections in the following order: 1. Data source, using ODBC 2. OLAP Metadata Catalog, using ODBC 3. Essbase Server, using TCP/IP In this test, you set up these connections one at a time and execute a LOADMEMBER command each time. By examining the errors each time, you can determine which connections are successful.
2 On the same computer that is running Integration Server, start Integration Services Shell at a command
prompt by typing
olapicmd
If Integration Services Shell does not start, make sure that the software is installed and that the Integration Services environment variables are set. See Correcting Integration Server Problems on page 149.
3 At the Integration Services Shell command prompt, use the login command to connect to the Integration
Server computer; for example, type
login cypress
Make sure that the version of the product you are using is compatible with the ODBC drivers and the version ofEssbase you are using. For information on component compatibility, see Oracle Hyperion Enterprise Performance Management System Installation Start Here.
The DSN parameter requires the ODBC data source name, which may be different from the database name in the data source.
6 Set OLAP Metadata Catalog to the sample OLAP Metadata Catalog by typing
setcatalog "DSN=TBC_MD;CODEPAGE=English_UnitedStates.Latin1@Binary; UID=TBC;PWD=password"
The DSN parameter requires the ODBC data source name, which may differ from the OLAP Metadata Catalog database name in the data source.
7 Set the instance of Essbase Server to which you want to connect by typing
settarget "DSN=sequoia;UID=TBC;PWD=password;"
8 Test the connection to the data source database. Attempt to perform a member load by typing
loadmember "APP=OLAP_TBC;DBN=Basic;OTL=TBC Metaoutline;"
During a member load, Integration Services Shell connects first to the data source and then to OLAP Metadata Catalog. If the connection to the catalog is unsuccessful, the message IS Error: Unable to log in to Catalog is displayed. If the connection to the data source fails, the message IS Error: Unable to login to Datasource is displayed.
If the connection to the data source failed, verify the ODBC data source name, user name, and password. If the connection continues to fail, see Correcting Integration Server Problems on page 149 and Correcting ODBC Problems on page 148. Repeat Step 4 through Step 8 of this procedure to retest the connection. When you successfully connect to the data source, continue to the next step.
10 Test the connection to the OLAP Metadata Catalog database by attempting to perform a member load. Type:
loadmember "APP=OLAP_TBC;DBN=Basic;OTL=TBC Metaoutline;"
During a member load, Integration Services Shell connects to OLAP Metadata Catalog and then to the Essbase Server. If the connection to OLAP Metadata Catalog is successful, the message IS Error: Unable to login to Essbase Server is displayed. If the connection to OLAP Metadata Catalog fails, the message IS Error: Unable to login to Catalog is displayed. If the connection to OLAP Metadata Catalog failed, verify the ODBC data source name, user name, and password. If the connection continues to fail, see Correcting Integration Server Problems on page 149 and Correcting ODBC Problems on page 148. Repeat Step 9 and Step 10 of this procedure to retest the connection. When you successfully connect to OLAP Metadata Catalog, continue to the next step.
12 Test the connection from Integration Services Console to the data source by proceeding to Testing the
Integration Services Console Connection to the Data Source on page 141.
2 On a computer other than the one running Integration Server, start Integration Services Console. 3 Connect to the Integration Server computer and OLAP Metadata Catalog; for example, cypress and TBC_MD.
If the connection fails, verify that you have specified the correct ODBC data source name for the OLAP Metadata Catalog database and the correct Integration Server computer name. Also verify that you specified the correct user name and password for the OLAP Metadata Catalog
database on the data source. See Correcting Integration Services Console Problems on page 150.
4 After connecting to OLAP Metadata Catalog, open an OLAP model and connect to the data source for the
OLAP model.
If the connection fails, verify that you have specified the correct ODBC data source name for the data source and the correct Integration Server computer name. Also verify that you specified the correct user name and password for the data source database on the data source. See Correcting Integration Services Console Problems on page 150.
2 Verify that the Essbase Server is running. 3 At a command prompt, start Integration Services Shell by typing
olapicmd
If Integration Services Shell does not start, make sure that the software is installed and that the Integration Services environment variables are set. See Correcting Integration Server Problems on page 149.
4 At the Integration Services Shell command prompt, use the login command to connect to Integration Server;
for example, type
login cypress
The DSN parameter requires the ODBC data source name, which may differ from the database name in the data source.
6 Set OLAP Metadata Catalog to the sample OLAP Metadata Catalog by typing
setcatalog "DSN=TBC_MD;CODEPAGE=English_UnitedStates.Latin1@Binary; UID=TBC;PWD=password"
The DSN parameter requires the ODBC data source name, which may differ from the OLAP Metadata Catalog database name in the data source.
7 Set the instance of Essbase Server to which you want to connect by typing
settarget "DSN=sequoia;UID=TBC;PWD=password;"
Note:
If you are using Integration Services Console on a client computer that is outside the firewall for your network, and you require access to an Integration Server and Essbase Server located inside the firewall for your network, be sure to use a name or an IP address for the Essbase Server that is accessible from both sides of the firewall.
8 Test the connection to the Essbase Server by attempting to perform a member load by typing
loadmember "APP=OLAP_TBC;DBN=Basic;OTL=TBC Metaoutline;"
If the connection failed, the message IS Error: Unable to login to Essbase Server is displayed. Make sure that the Essbase Server is running. Verify the name of the metaoutline in Essbase. Also verify the name of the test application and database in Essbase. Repeat Step 7 and Step 8 of this procedure to retest the connection. After you can connect successfully to the Essbase Server, proceed to Testing the Integration Services Console Connection to Essbase Server on page 143.
The following procedure uses the sample application, including the sample database, OLAP Metadata Catalog, OLAP model, and metaoutline. For your tests, substitute the application and databases you are using.
5 After connecting to OLAP Metadata Catalog, open a metaoutline and connect to the data source for the
metaoutline; for example, TBC Metaoutline and TBC.
7 If prompted, provide the name of the Essbase Server, user name, and password.
If the Essbase Application and Database dialog box is displayed, the connection is successful. If the connection failed, verify that you have entered the correct computer name for the Essbase Server and the correct user name and password. See the Essbase documentation. If the Essbase Server is located on a large network, you may need to provide a more specific server name address, such as cypress.hyperion.com rather than simply cypress.
Note:
If you are using Integration Services Console on a client computer that is outside the firewall for your network, and you require access to an Integration Server and Essbase Server located inside the firewall for your network, be sure to use a name or an IP address for the Essbase Server that is accessible from both sides of the firewall.
Are all user names and passwords being used consistently across all the components and computer user accounts? See User Name and Password Problems on page 145. Do all user accounts used in the system have the appropriate permissions and privileges? System administrators typically use highly privileged administrative accounts when setting up and testing a system. After the system is set up, users with fewer privileges may be unable to use the system successfully. Check permissions on the following components:
User account on the client computer User account on the Integration Server computer OLAP Metadata Catalog database Data source database User account for the Essbase Server
Are the software components using adequate addressing to identify the computers to which they are connecting? Depending on the size and configuration of the computer network, it may be necessary to use a fully-qualified host address to connect to a particular computer. For example, instead of the host name cypress, the complete host name cypress.mydivision.mycompany.com may be required. Alternatively, you might need to use the IP address number for the server computer; for example, 127.0.0.1.
Is the user who starts Integration Server (for example, hyperion) configured to run all the required software?
The user (for example, hyperion) must have all required environment variables set and must have permissions to run all the following software components:
Integration Server ODBC software Database client software Data source server (this may not be required)
During installation, configuration, and testing, keep a list of all user names and passwords that you use to create or to modify the system, including the applications with which the user names and passwords are associated and the purpose of each application. In a smaller, single-server implementation of Integration Services, where Integration Server, Essbase Server, and the data source are all run on a single computer, you must manage five user name and password sets and use them consistently:
User account on the Integration Services client computer User account on the computer that runs all the server software OLAP Metadata Catalog database in the data source Data source database in the data source Essbase Server software
In a fully distributed, three-server implementation, where Integration Server, Essbase Server, and the data source are all run on separate computers, you must manage seven user name and password sets and use them consistently:
User account on the Integration Services client computer User account on the computer that runs Integration Server User account on the computer that runs the data source OLAP Metadata Catalog database in the data source Data source database User account on the computer that runs Essbase Server Essbase Server software
Adding aliases or synonyms to the data source databases adds to the number of user name and password sets you must manage.
Tip:
During installation and connection testing, use a single set of user names and passwords to avoid confusion. Add user aliases or synonyms after the system is configured and running.
Is the data source server computer and software running? Has the server locked up, stopped, or failed? Is the data source software installed? Are all components required for ODBC connections installed? Some data sources may require additional software components beyond the basic server software to enable ODBC connections. The database client software is usually required on the computer that connects to the data source server using ODBC and also may be required on the data source server computer.
Do the OLAP Metadata Catalog and data source databases to which you are trying to connect exist? Have the tables for OLAP Metadata Catalog been created? You can create the OLAP Metadata Catalog tables either automatically or by using the creation scripts provided with Integration Services. See Chapter 3, Creating, Upgrading, and Deleting OLAP Metadata Catalogs.
What are the exact names of the OLAP Metadata Catalog and data source databases? Are the names case-sensitive? Does the user name that you are using to connect to the data source have adequate permissions to access the database? Does the user have read and write permissions? For the OLAP Metadata Catalog database, users need full read and write permissions. For the data source database, users need read-only permissions.
If the data source server is on a computer by itself, is there a corresponding user account on the computer for the user names you are using to connect to the databases? To allow connections to the database server from other computers, you may need to set up user accounts on the data source server computer, separate from the database users you create.
Does the data source require any addressing beyond the name of the database if you are using the database client software or ODBC to connect? Some data sources may have specific naming requirements for connecting to their databases, including different naming conventions for third-party applications (such as Integration Services). See the documentation for the data source.
Is the database client software installed? Are all database client components required for the ODBC connections installed? To enable ODBC connections, some database client software may require additional software components beyond the basic server software. The database client software usually is required on the computer that uses ODBC to connect to the data source server and may also be required on the data source server computer.
Is the database client software compatible with the data source? Earlier or later versions of the database client software may not work with the data source version that you are running. See your data source documentation.
Do the database client software utilities, such as the command-line SQL interface, start? Do they start when you are logged on as the user who runs Integration Server? When you run the database client software, the user account must have the appropriate file access permissions and must have set the path and other environment variables. Verify that the user who runs Integration Server (typically, hyperion) is configured to run the database client software. On UNIX systems, a script file is usually provided by the data source vendor to set the required environment variables and should be included in the main user login script. If you include one of these files in the main login script, make sure that you specify a complete directory path and that you follow the syntax for executing a script file within a script.
Are the data source server names, database names, user names, and passwords specified correctly? Make sure you are using the correct names, including the exact spelling, uppercase or lowercase letters, and any required data source-specific identification syntax. See Correcting Data Source Problems on page 146.
Is the database client software configured correctly? File access permissions, path, and other environment variables must be set for the user who runs the database client software. Additional configuration steps also may be required for specific database clients. See the installation documentation for the database client software. IBM DB2: You may be required to bind the database client software utilities to the IBM DB2 databases. See the IBM DB2 installation documentation. Oracle: The database name for Oracle databases can be specified in a tnsnames.ora file, which must be configured by the database administrator. See the Oracle installation documentation. Informix: Ensure that the sqlhosts environment variable is properly set. In the sqlhosts file, you must specify a TCP protocol for your operating system as follows:
AIX and HP-UX: onsoctcp Solaris: ontlitcp Sybase: Use the dsedit utility (type $SYBASE/bin/dsedit) to view and edit server entries.
Is the ODBC software installed? Are all required ODBC components installed? On Windows systems, make sure the ODBC core components and any required drivers are installed. On UNIX systems, the Integration Services setup program installs the required ODBC core components if you choose to install the DataDirect ODBC drivers.
Is the ODBC driver compatible with the data source? Is the driver compatible with the operating system? Verify that the ODBC driver you are using is supported by Integration Services by referring to the compatibility matrix provided in the Oracle Hyperion Enterprise Performance Management System Installation Start Here.
Are the data source server names, database names, user names, and passwords specified correctly in the ODBC data source? Make sure you are using the correct names, including the exact spelling, uppercase or lowercase letters, and any required data source-specific identification syntax. On Windows systems, configure ODBC data sources by using ODBC Administrator. On UNIX systems, configure ODBC data sources either by using the odbcconfig utility or by editing the odbc.ini file manually. IBM DB2: On Windows systems, the IBM DB2 Client Configuration Assistant shows the ODBC DSN. See Correcting Data Source Problems on page 146.
On UNIX systems, have you specified the full path and file name of the ODBC driver for the data source in the odbc.ini file? Does the ODBC driver actually exist in the specified directory? Can ivtestlib load it? Try copying the driver path and file name from the odbc.ini file and pasting it into an ls command. If the file is not listed when you execute the command, check the accuracy of the driver path and file name in the odbc.ini file.
Are any parameters missing in the ODBC data source configuration? On Windows systems, make sure you have filled in all required parameters for the data source in ODBC Administrator. On UNIX systems, the Integration Services installation provides a sample odbc.ini file, which may not include data source parameters required for the data source that you are using. For information on required data source-specific ODBC driver parameters, see the ODBC documentation provided with Integration Services. For examples of odbc.ini file configurations for the supported data source platforms, see the Oracle Hyperion Enterprise Performance Management System Installation Start Here.
Can other applications connect to the data source by using ODBC? If you can identify another application that is successfully using ODBC to connect to databases on the data source, analyze the configuration of the computer from which the connection is made and use it as a model for the computer connections you are trying to fix.
Note:
Most data source vendors provide utilities to test ODBC connections. For information, see the documentation for the data source. If the preceding checks do not enable you to correct the problems connecting from ODBC to the data source, try using ODBC tracing to isolate the problem. See Using ODBC Tracing on page 151.
Is Integration Server installed? Are all the required Integration Server components installed? The following components must be installed to have a functional instance of Integration Server:
Integration Server software Related base components (on UNIX, these components are included as part of Integration Server software) DataDirect ODBC drivers (unless ODBC and drivers are already installed)
Is Integration Server running? Is it configured properly? Start Integration Server by opening a command prompt window and typing olapisvr . If Integration Server does not start, review the following possible problems:
Are the required environment variables set? Are they set for the correct user name? On Windows systems, if you decide not to enable the setup program to update environment variables automatically, you must update the variables manually. On UNIX systems, a script file (is.sh or is.csh) that sets environment variables must be included in the login script of the user who starts Integration Server.
What release of Essbase is installed? If an incompatible version of Essbase is installed on the same computer as Integration Server and the \essbase\bin directory is in the current path, Integration Server does not start. For information on compatible versions of Essbase, see the Oracle Hyperion Enterprise Performance Management System Installation Start Here.
Is port 3388 in use? Integration Server uses port 3388 to communicate with the Integration Services client software. If this port is being used by another program, Integration Server does not start.
On UNIX, is more than one copy of olapisvr running? If so, all servers except one will hang indefinitely. To correct the problem, run the following command to display all programs that are currently running:
ps -fe | grep olapisvr
Many of the problems related to Integration Server not working properly may be traced to configuration steps that you may have missed. See the Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide.
Is Integration Services Console installed? Are all required components installed? The following components must be installed to have a functional Integration Services Console client:
Is the Integration Services Console configured properly? Make sure the required environment variables are set. Run the Oracle's Hyperion Enterprise Performance Management System Configurator as described in the Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide.
Are the Integration Server computer name, ODBC data source names, user names, and passwords specified correctly? For the Integration Server computer name, verify that you are using the correct spelling and uppercase and lowercase letters for the server name. For OLAP Metadata Catalog and the data source, make sure you are using an ODBC data source name, which may be different from the data source database name.
Note:
The user names and passwords for the OLAP Metadata Catalog database may be different from the data source. Verify that you are using the correct login information for each database. See Correcting Data Source Problems on page 146.
Does the user account with which you are running Integration Services Console have permission to connect to the computer running Integration Server? You must have a user account and appropriate access permissions on the server computer running Integration Server.
If you have Integration Server installed on more than one computer, have you specified the correct server name? A particular Integration Server may not be set up to access the same OLAP Metadata Catalogs and data sources used by other servers.
Are you using adequate addressing to identify the Integration Server computer?
Depending on the size and configuration of the computer network, it may be necessary to use a fully-qualified host address to connect to a particular computer. For example, instead of the host name cypress, the complete host name cypress.mydivision.mycompany.com may be required. Alternatively, you might need to use the IP address number for the server computer; for example, 127.0.0.1.
Tracing quickly creates a large log file. Disable tracing when you are finished: On the Tracing tab, click Stop Tracing Now.
3 Start Integration Server and Integration Services Console. 4 From Integration Services Console, selectConnections , then OLAP Metadata Catalog, then Connect, and
try to connect to OLAP Metadata Catalog.
5 If you can connect to OLAP Metadata Catalog, open an OLAP model and try to connect to a data source. 6 If you cannot connect to OLAP Metadata Catalog (step 4) or to a data source (step 5), see the file that you
specified as the log file in the root directory.
If necessary, share the information in the log file with Hyperion Technical Support to help solve the connectivity problem.
Tracing quickly creates a large log file. Disable tracing when you are finished by setting the Trace parameter to 0.
4 Start Integration Server and Integration Services Console. 5 From Integration Services Console, select Connections , then OLAP Metadata Catalog, then Connect, and
try to connect to OLAP Metadata Catalog.
6 If you can connect to OLAP Metadata Catalog, open an OLAP model and try to connect to a data source. 7 If you cannot connect to OLAP Metadata Catalog (step 5) or to a data source (step 6), read the
odbctrace.out file.
If necessary, share the information in the log file with Hyperion Technical Support to help solve the connectivity problem.
8
In This Chapter
Starting Integration Services Shell ............................................................................................. 153 Integration Services Shell ....................................................................................................... 154 Integration Services Shell Commands ......................................................................................... 159 Peforming Member Loads Using Integration Services Shell.................................................................. 172 Loading Data Using Integration Services Shell................................................................................ 173
This chapter describes the commands supported by Essbase Integration Services Shell, the command-line tool provided with Integration Services. The majority of these commands provide the same functionality available through Essbase Integration Services Console. You use Integration Services Shell commands and scripts to create Essbase outlines and loadEssbase databases. If you schedule member or data loads using Integration Services Console, Integration Services automatically creates an Integration Services Shell batch file and a corresponding .cbs file in the Batch directory. You can modify and reuse the.cbs file for future member and data loads.
A database that contains the OLAP Metadata Catalog where the metadata is stored One or more data sources that you want to use to create OLAP models and metaoutlines Essbase Integration Server See Starting Integration Server on page 12.
Essbase Server See the Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide.
From the Windows Start menu, select All Programs, then Oracle EPM System, then Essbase, then Integration Services, then Shell. Enter the executable name on the command line; for example, type
olapicmd
To set the TCP port number with which Essbase Integration Server communicates with its clients to a port different from the default 3388, enter the executable name and Pportnumber when you start Integration Services Shell; for example, type
olapicmd -P3390
Note:
If Integration Services Shell does not start when olapicmd is executed from the command line, the operating system path may not be updated correctly. See the Oracle Hyperion Enterprise Performance Management System Installation and Configuration Guide. The Integration Services Shell window opens, as shown in Figure 5:
Figure 5
In the Integration Services Shell window, enter commands at the command prompt. For example, type the following commands:
Type login machine_name to log on to a server running on the same machine. Type shutdown to stop the server. Type exit to stop Integration Services Shell. Type help to get a list of Integration Services Shell commands.
Batch mode. To automate routine server maintenance and diagnostic tasks, write a script or batch file and run it from the command line.
Batch mode is convenient if you use a particular series of commands frequently or if the task requires many commands.
Interactive mode. To perform tasks interactively, enter OLAP commands at the command line. Interactive mode is convenient for short operations that require few commands, such as checking for information on the fly and error checking.
Tip:
It can be difficult to enter commands correctly when using the command line. The easiest method is to use the Integration Services Console OLAP Metaoutline standard user interface to perform a data or member load, and then choose to create a script file. You can then edit the script file as needed. Run all the commands in a script file by redirecting the standard input to use your command file. See Running OLAP Command Script Files on page 158.
To perform a successful member and data load, the SET commands must be executed before the LOAD commands. SET commands provide Integration Server with the location of the source databases, Essbase Server, and OLAP Metadata Catalog. The commands can be issued in any order, but you must issue all three SET commands before executing a LOAD command. Use SET commands as follows:
Command SETSOURCE SETTARGET Description Specifies the data source databases from which an Essbase outline is built Specifies the name of the Essbase Server computer on which an outline for an Essbase application and database is built Specifies the OLAP Metadata Catalog that you created to store OLAP models and metaoutlines Specifies the locale for OLAPICMD.
SETCATALOG SETLOCALE
The STATUS command (see STATUS on page 160) does not check or validate the syntax for SET commands. LOAD commands use the information from the SET commands to load members, data, or both. Use the LOAD commands as follows:
Integration Services Shell 155
Description Loads members into an Essbase outline. You must load members before you load data. Loads data into an Essbase database. You cannot load data without loading members. Loads both members and data into an Essbase database.
LOAD commands for small outlines (fewer than 1000 members) take a few seconds to execute. Larger outlines (1000 or more members) take a few minutes to a few hours, depending on the amount of data being loaded, the configuration of the metaoutline, and the computer system capabilities. After executing LOAD commands, wait a few seconds until the command prompt (->) is displayed and then use the STATUS command to check the success or failure of the command execution. See Peforming Member Loads Using Integration Services Shell on page 172 for a sample sequence of commands for loading members. See Loading Data Using Integration Services Shell on page 173 for a sample sequence of commands for loading data.
Informational Commands
The STATUS and VERSION commands provide processing status and system release version information for you and can be used at any time in the sequence of command execution.
You must separate the statements with a semicolon. The final semicolon is optional. Do not use semicolons as part of data source names, because Integration Services Shell recognizes semicolons as field separators. It is recommended that you use double quotation marks around the parameters, as shown in the preceding example. In this guide, all optional syntax parameters are indicated by their enclosure in brackets [ ]; however, do not type the brackets when you enter the commands. Press Enter to execute the commands.
Note:
If you are unsure of a specific syntax, use Integration Services Console to schedule member or data loads, or to create load scripts. An Integration Services Shell batch file and a corresponding .cbs file are created automatically in the eis\server\Batch directory. You can reference, modify, and reuse the .cbs file for future member and data loads. In the Schedule Essbase Load dialog box, be sure to check the Save Script Only option instead of the Schedule option. See Recording Member and Data Load Scripts on page 157.
3 In the Essbase Application and Database dialog box, type or select the appropriate application and
database names.
4 Click Next to display the Schedule Essbase Load dialog box. 5 Save the load to a script file by clicking Only Save Load Script and then click Save Scripts. 6 In the Save Script As dialog box, type a name for the script that you want to save, and then click OK. 7 Click Finish to start the load or to complete scheduling of the load.
A script file (.cbs extension) contains Integration Services Shell commands. You can run a script file from the operating system command line or from within an operating system batch file. On Windows platforms, a batch file (.bat extension) is an operating system file that calls multiple Integration Services Shell scripts and is used to run multiple sessions of Integration Services Shell commands. You can run a batch file on the server from the operating system prompt. On UNIX, shell scripts are used in place of batch or script files.
When you run a script file, Integration Services Shell executes the commands in the order specified in the script until the program reaches the end of the file.
Replace scriptFileName with the name of the script file you are using. Replace logFileName with the name of the file where you want the feedback from the script to be recorded. For example, the following sample script file, olap_tbc.cbs, was created in a text editor. This script connects to Essbase from the Integration Server computer and generates outlines for a sample database. In the following example, the status command (see STATUS on page 160) is used to check the success or failure of each command. The status command returns the execution status in the Integration Services Shell window.
login labmachine1 status settarget "DSN=labmachine1;UID=hyperion;PWD=password" status setcatalog "DSN=TBC_MD;CODEPAGE= English_UnitedStates.Latin1@Binary;UID=tbc;PWD=password" status setsource "DSN=TBC;CODEPAGE= English_UnitedStates.Latin1@Binary;UID=tbc;PWD=password" status loadall "OTL=TBC Metaoutline;APP=olaptbc;DBN=Tbf1;FLT_ID=1; CALC_SCRIPT=#DEFAULT#;" status loadall "OTL=TBC Metaoutline;APP=olaptbc;DBN=Tbf2;FLT_ID=2; OTL_CLEAR=Y;" status loadall "OTL=TBC Metaoutline;APP=olaptbc;DBN=Tbf3a;FLT_ID=3; OTL_CLEAR=N;CALC_SCRIPT=#DEFAULT#;" status loadall "OTL=TBC Metaoutline;APP=olaptbc;DBN=Tbf3ec;FLT_ID=3; OTL_CLEAR=N;ESSC_SCRIPT=mytest1" status exit
The following sample batch file, olap_tbc.bat, uses input from a script file named olap_tbc.cbs and saves the feedback in a file named olap_tbc.log:
olapicmd -folap_tbc.cbs > olap_tbc.log
LOGIN on page 159 STATUS on page 160 VERSION on page 160 SETLOCALE on page 161 SETSOURCE on page 161 SETTARGET on page 162 SETCATALOG on page 162 LOADMEMBER on page 163 LOADDATA on page 166 LOADALL on page 168 LOGOUT on page 172 SHUTDOWN on page 172 EXIT on page 172
Unless otherwise noted, you must be logged in to Integration Server to execute these commands. The load commands in this topic are shown with an abbreviated syntax, which includes the syntax required for incremental loads. To see the full syntax for standard or incremental loads, use the Integration Services Console to record a script for a load. See Recording Member and Data Load Scripts on page 157.
LOGIN
The LOGIN command connects you to the Integration Server computer. On a successful connection, the Integration Services Shell prompt changes to show the name of the Integration Server computer to which Integration Services Shell is connected. You can also use the LOGIN command to connect to an Integration Server computer that is running on a non-default port. Whether you are logging in to an Integration Server computer that is running on the default port or on a non-default port, this command returns an error if you already are connected to the Integration Server computer. Logging in to Integration Server Running on the Default Port Syntax
LOGIN IShost
Parameter Description IShost TCP/IP name or number of the computer where Integration Server is running
Example
LOGIN cypress
Parameters IShost
Description TCP/IP name or number of the computer where Integration Server is running
Example
LOGIN cypress:3390
STATUS
The STATUS command displays the processing status of the most recently issued command. Use it to check the success or failure of all commands. After issuing a command, wait a few seconds until the command prompt (->) is displayed, and then enter the STATUS command. The STATUS command does not check or validate the syntax for SET commands. Syntax
STATUS
Example Typing STATUS after issuing the LOADMEMBER command returns the following message in the Integration Services Shell window:
LOADMEMBER executed successfully, elapsed time is '22' seconds.
VERSION
The VERSION command returns the release number of the Integration Services software you are running. Syntax
VERSION
Example VERSION returns the release number for Integration Services software.
160 Using Integration Services Shell
SETLOCALE
The SETLOCALE command enables you to override the default locale for OLAPICMD. The default locale for OLAPICMD is UTF-8 (Unicode). Although a script file saved from Integration Services Console does not have the UTF-8 signature, OLAPICMD nevertheless reads the script as UTF-8. If you create a script file in Microsoft Notepad and save the script file as a UTF-8 file, you do not need to use the SETLOCALE command to select a locale. You can use the default locale. If you choose to select a locale, you must select one that is UTF-8. A native locale causes load operations to fail. If you create a script file in Microsoft Notepad and save the script file as an ANSI file, you must select a native locale. A UTF-8 locale causes load operations to fail. If you intend to enter Integration Services Shell commands manually (rather than executing them through a script) on a native computer, you must select the appropriate native locale using the SETLOCALE command. Failure to do so will cause load operations to fail. Syntax
SETLOCALE <LANGUAGE_TERRITORY.CODEPAGE@SORT> SETLOCALE .UTF8@default
Examples
SETLOCALE Japanese_Japan.MS932@Binary SETLOCALE .UTF8@default
SETSOURCE
The SETSOURCE command enables you to identify one or more source databases for Integration Server to use with a load command. Syntax
SETSOURCE "DSN=PrimaryODBCdatasource;CODEPAGE=Codepage;UID=username; PWD=password[;][DSN=SecondaryODBCdatasource;CODEPAGE=Codepage; UID=username;PWD=password][;]...[DSN=SecondaryODBCdatasource; CODEPAGE=Codepage;UID=username;PWD=password]"
Description The primary ODBC data source name configured on the computer where Integration Server runscase-sensitive The code page of the language you want to use during the current Integration Services Console session The name for logging on to a primary or secondary ODBC data sourcecase-sensitive The password for logging on to a primary or secondary ODBC data sourcecase-sensitive
Codepage
username password
Parameters
Description
SecondaryODBC datasource The secondary ODBC data source name configured on the computer where Integration Server runscase-sensitive
Example
SETSOURCE "DSN=TBC;CODEPAGE=English_UnitedStates.Latin1@Binary;UID=TBC; PWD=Password;DSN=TBC2;CODEPAGE=English_UnitedStates.Latin1@Binary; UID=TBC;PWD=Password;DSN=TBC3; CODEPAGE=English_UnitedStates.Latin1@Binary;UID=TBC;PWD=Password"
SETTARGET
The SETTARGET command enables you to identify a target Essbase Server computer for Integration Server to use with a load command. Syntax
SETTARGET DSN=EssbaseServicesdataservername;UID=username;PWD=password[;] Parameters Description
EssbaseServices dataservername The TCP/IP name or number of the computer where Essbase Server is running username password The name for logging on to Essbase Server The password for logging on to Essbase Servercase-sensitive
Example
SETTARGET "DSN=FIG;UID=TBC;PWD=Password"
SETCATALOG
The SETCATALOG command enables you to identify an OLAP Metadata Catalog database for Integration Services to use with a load command. Syntax
SETCATALOG DSN=ODBCdatasource;CODEPAGE=Codepage;UID=username; PWD=password[;] Parameters Description
ODBCdatasource The ODBC data source name configured on the computer where Integration Server runscasesensitive Codepage username password The code page of the language you want to use during the current Integration Services Console session The name for logging on to the ODBC data sourcecase-sensitive The password for logging on to the ODBC data sourcecase-sensitive
Example
162 Using Integration Services Shell
LOADMEMBER
The LOADMEMBER command initiates a member load operation on Integration Server. The load environment must be set up properly before you invoke the LOADMEMBER command. To set up the load environment, issue the SET commands in any order. See SET and LOAD Commands on page 155. To learn the status of the member load invoked by the LOADMEMBER command, wait until you see the command prompt (->), then use the STATUS command. Syntax
LOADMEMBER "OTL=Metaoutline;APP=Essbase Application; DBN=Essbase Database;[FLT_ID_MEMBER=Member_Load_Filter_ID;] [DELETE_DB=Delete and Restore Database Y|N;] [OTL_CLEAR=Delete All Members First Y|N;] [INCUPD=DimID-MbrID,DimID-MbrID,...DimID-MbrID;] [INCUPD_MEMBER=Dynamic Restructuring Member Load Options;] [ATTR_UPDATE_LEVEL=Attribute Update Level;] [@@USERS=Username List;][ESSCMD_SCRIPT=Esscmd Script File;] [UNICODE=Create Unicode Application] [UPDATE_DTDATA_ONLY=Update Drill-Through Data Only Y/N]" Parameters Metaoutline Description The name of the metaoutlinecase-sensitive. This metaoutline is used to extract data and create an Essbase outline. The name of the Essbase application on the target Essbase Server computer where the Essbase outline is created. The name of the Essbase database on the target Essbase Server computer. Integration Server applies this name to the Essbase outline. Optional. You can specify a member load filter ID to use when loading members. If you do not specify a filter ID, the default name (*DEFAULT) is used. The default filter ID is 1; any additional filters are numbered sequentially, starting with the number 2. For information about creating filters, see the Integration Services Console Help. Optional. The Delete and Restore Database parameter directs whether Integration Server should delete all members in the Essbase database before performing a member load. The default is N (No). This is the format: DELETE_DB=[Y|N] When this option is selected (DELETE_DB=Y), the options Delete All Members First, Modify Essbase Users, and Incremental Update cannot be used. Values are as follows:
Essbase Application
Essbase Database
Parameters
Description Y Delete and restore (re-create) theEssbase database during the member load process. N Do not delete and restore the Essbase database (this is the default).
Optional. When the Delete All Members First parameter is selected (OTL_CLEAR=Y), Integration Server removes all members from the existing database and then re-creates the database by using the members in the metaoutline. This process is slower than creating or updating an Essbase outline without deleting the members; therefore, do not delete all members first unless it is necessary. You should delete all members if, for example, you know that some members have been removed from the OLAP metaoutline and you want to build an Essbase outline containing a smaller set of members. The default is N (No). This is the format: OTL_CLEAR=[Y|N] Values are as follows: Y Remove all members from the existing database and then re-create the database by using the members in the metaoutline. N Do not remove members from the existing database; only update the Essbase database (this is the default).
Optional. Incremental Upate. The level at which dimensions are updated during an incremental member load. The format is: INCUPD=DimID-MbrID,DimID-MbrID,...DimIDMbrID;[;] Values for incremental update are as follows:
DimID represents the dimension ID of the dimension to which the incremental update applies. MbrID represents the member ID of the level zero member of the hierarchy to which the incremental update applies.
Dynamic Restructuring Member Load Options Optional. Specify 1, 2, 3, or 4. Dynamic restructuring member load options set the parameters for restructuring the database during a member load. The option that you select affects any subsequent data loads. Values are as follows: 1 Preserve all data. Select to preserve all existing data that applies to the changed outline when restructuring occurs. This is the default (INCUPD_MEMBER=1). 2 Discard all data. Select to clear all data from the database. 3 Preserve level 0 data. Select to preserve data only for level zero members. This is the optimal restructure option if you change the source database and need to recalculate the data, and if all data required for the calculation is in level zero members. Selecting this option deletes all upper-level blocks before restructuring. This reduces the disk space for restructuring and improves calculation
Parameters
Description time when the database is recalculated. The upper-level blocks are recreated when you calculate the database. 4 Preserve input data. Select to preserve only those blocks containing data that is loaded. Many applications contain data that is entered at parent levels. Selecting this option prevents deletion of any blocks that are created by data load, whether they are non-level zero or level zero (leaf member) blocks.
Optional. Use with the Incremental Update parameter. The level at which attribute dimensions are updated when a member load is performed. Values are as follows: 0 Delete all attribute dimensions after the member load is performed. 1 Do not update existing attribute dimensions while updating other types of dimensions (for example: standard, time, or measures dimensions) during a member load. 2 Update all existing attribute dimensions and add new attribute dimensions during a member load.
Username List
Optional. Lists all user IDs that can access the Essbase application and database into which you are loading the members. Commas separate user IDs. For example, @@USERS="admin","ljones","ksmith". Optional. The name of an ESSCMD script. The script that you define instructs Essbase to perform specified actions on the Essbase database. These actions might include changing a member tag or loading data from a worksheet into user-defined members after you load the data from the data source. The ESSCMD script must be in the eis/ server/esscript directory. Optional. You can specify that a Unicode application is created on Essbase Server. The default is that no Unicode application is created. This parameter cannot be overwritten. For example, if a previous application exists and you are overwriting that application with the current member load, you cannot overwrite the original Unicode or nonUnicode setting. This is the format: UNICODE=[Y|N] For example: UNICODE=Y
Optional. Update drill-through information or hybrid analysis information. There is no separate option for hybrid analysis. Use this command for both drill-through information and hybrid analysis information. This is the format: UPDATE_DTDATA_ONLY=[Y|N] For example: UPDATE_DTDATA_ONLY=Y
Example
LOADMEMBER "OTL=ProductsAnalysis;APP=Products; DBN=Analysis;FLT_ID_MEMBER=1;OTL_CLEAR=Y; INCUPD=3-2,4-2,5-1,5-2,5-3;INCUPD_MEMBER=3; @@USERS="TBC","CFO";UNICODE=Y;"
LOADDATA
The LOADDATA command initiates a data load operation on Integration Server. The load environment must be set up properly before you invoke this command. To set up the load environment, issue the SET commands in any order. See SET and LOAD Commands on page 155. LOADDATA takes a few seconds to execute. After executing LOADDATA, use the STATUS command to learn the status of the command execution. Syntax
LOADDATA "OTL=Metaoutline;APP=Essbase Application; DBN=Essbase Database;[FLT_ID_DATA=Data Load Filter ID;] [REPLACE_ZEROS=Replace Zeros with #MISSING;] [INCUPD=DimID-MbrID,DimID-MbrID,...DimID-MbrID;] [INCUPD_DATA=Dynamic Restructuring Data Load Options;] [@@USERS=Username List;][CALC_SCRIPT=Calc Script Name;] [ESSCMD_SCRIPT=Esscmd Script File;] [FT_COLUMN=FactTable Column for Incremental Update;]" Parameters Metaoutline Description The name of the metaoutline. Integration Server uses the specified metaoutline to extract data from the data source to create an Essbase outline. The name of the Essbase application on the target Essbase Server computer where the Essbase outline is created. The name of the Essbase database on the target Essbase Server computer. Integration Server applies this name to the Essbase outline. Optional. You can specify a data load filter ID to use when loading data. If you do not specify a filter ID, the default filter (*DEFAULT) is used. The default filter ID is 1; any additional filters are numbered sequentially, starting with the number 2. For information about creating filters, see the Integration Services Console Help. Optional. The Replace Zeros with #MISSING parameter replaces Essbase account dimension values of zeros with #MISSING. The default is N (No). This is the format: REPLACE_ZEROS=[Y|N] Here are the reasons that you may want to replace zeros with #MISSING:
Essbase Application
Essbase Database
Parameters
Description
The calculation time required in the Essbase database is decreased significantly. Because of differences in calculator behavior depending on whether a value is zero or missing, faulty data is not generated.
DimID-MbrID, DimID-MbrID,... DimID-MbrID Optional. Incremental Update. The level at which dimensions are updated during an incremental data load. This is the format: INCUPD=DimID-MbrID,DimID-MbrID,...DimID-MbrID; [;] Values for incremental update are as follows:
DimID represents the dimension ID of the dimension to which the incremental update applies. MbrID represents the member ID of the level zero member of the hierarchy to which the incremental update applies.
Dynamic Restructuring Data Load Options Optional. Dynamic restructuring data load options determine how Essbase loads values from a data source to the database. Specify 1, 2, or 3. Values are as follows: 1 Overwrite. Select to replace the values in the database with the values in the data source. This is the default (INCUPD_DATA=1). 2 Add. Select to add values in the data source to the existing values in the database. 3 Subtract. Select to subtract the values in the data source from the existing values in the database. Username List Optional. Lists all user IDs that can access the Essbase application and database into which you are loading the members. Commas separate user IDs. For example, @@USERS="admin","ljones","ksmith". Optional. The name of the calculation script. The calculation script determines how Essbase calculates the data values in the database. If you do not select a calculation script, Integration Server does not perform a calculation. Note: If the data storage model property of the metaoutline is aggregate storage, this parameter is not applicable. Esscmd Script File Optional. The name of an ESSCMD script. The script that you define instructs Essbase Server to perform specified actions on the Essbase database. These actions might include changing a member tag or loading data from a worksheet into user-defined members after you load the data from the data source. The ESSCMD script must be in the eis/server/ esscript directory.
FactTable Column for Incremental Update Optional. You can specify that a time-based incremental data load be performed based on the date the last data load was performed. The DateTimeColumn Name parameter is the name of the datetime column in the fact table that contains each records create date or modified date. You can perform time-based incremental loads only if the fact table on which a metaoutline is based has a datetime column that contains each record's creation date or modified date. This is the format:
Parameters
Example
LOADDATA "OTL=ProductsAnalysis;APP=Products;DBN=Analysis; FLT_ID_DATA=2;REPLACE_ZEROS=Y;CALC_SCRIPT=#DEFAULT#; INCUPD=1-2,2-1,3-2,4-2,5-1,5-2,5-3;INCUPD_DATA=3;@@USERS="TBC","CFO";"
LOADALL
The LOADALL command initiates a load of both members and data on Integration Server. The load environment must be set up properly before you invoke this command. To set up the load environment, issue the SET commands in any order. See SET and LOAD Commands on page 155. LOADALL takes a few seconds to execute. After executing LOADALL, use the STATUS command to learn the status of the command execution.
Note:
If you want an ESSCMD script to execute between a member load and a data load, use the LOADMEMBER and LOADDATA commands instead of the LOADALL command.
Syntax
LOADALL "OTL=Metaoutline;APP=Essbase Application;DBN=Essbase Database; [FLT_ID_MEMBER=Member Load Filter ID;] [FLT_ID_DATA=Data Load Filter ID;] [DELETE_DB=Delete and Restore Database;] [REPLACE_ZEROS=Replace Zeros with #MISSING;] [OTL_CLEAR=Delete All Members First;] [INCUPD=DimID-MbrID,DimID-MbrID,...DimID-MbrID;] [INCUPD_MEMBER=Dynamic Restructuring Member Load Options;] [INCUPD_DATA=Dynamic Restructuring Data Load Options;] [ATTR_UPDATE_LEVEL=Attribute Update Level;][@@USERS=Username List;] [CALC_SCRIPT=Calc Script Name;][ESSCMD_SCRIPT=Esscmd Script File;] [UNICODE=Create Unicode Application]"
Parameters Metaoutline
Description The name of the metaoutline. Integration Server uses the specified metaoutline to extract data from the data source to create an Essbase outline. The name of the Essbase application on the target Essbase Server computer where the Essbase outline is created.
Essbase Application
Description The name of the Essbase database on the target Essbase Server computer. Integration Server applies this name to the Essbase outline. Optional. You can specify a member load filter ID to use when loading members. If you do not specify a filter ID, the default name (*DEFAULT) is used. The default filter ID is 1; any additional filters are numbered sequentially, starting with the number 2. For information about creating filters, see the Integration Services Console Help. Optional. You can specify a data load filter ID to use when loading data. If you do not specify a filter ID, the default filter (*DEFAULT) is used. The default filter ID is 1; any additional filters are numbered sequentially, starting with the number 2. For information about creating filters, see the Integration Services Console Help. Optional. The Delete and Restore Database parameter directs whether Integration Server should delete all members in the Essbase database before performing a member load. The default is N (No). This is the format: DELETE_DB=[Y|N] When this option is selected (DELETE_DB=Y), the options Delete All Members First, Modify Essbase Users, and Incremental Update cannot be used. Values are as follows: Y Delete and restore (re-create) the Essbase database during the member load process. N Do not delete and restore the Essbase database (this is the default).
Optional. The Replace Zeros with #MISSING parameter replaces Essbase account dimension values of zeros with #MISSING. The default is N (No). This is the format: REPLACE_ZEROS=[Y|N] Here are the reasons that you may want to replace zeros with #MISSING:
Fewer input blocks are created during the data load. The calculation time required in the Essbase database is decreased significantly. Because of differences in calculator behavior depending on whether a value is zero or missing, faulty data is not generated.
Optional. When the Delete All Members First parameter is selected (OTL_CLEAR=Y), Integration Server removes all members from the existing database and then re-creates the database by using the members in the metaoutline. This process is slower than creating or updating an Essbase outline without deleting the members; therefore, do not delete all members first unless it is necessary. You should delete all members if, for example, you know that some members have been removed from the OLAP
Parameters
Description metaoutline and you want to build an Essbase outline containing a smaller set of members. The default is N (No). This is the format: OTL_CLEAR=[Y|N] Values are as follows: Y Remove all members from the existing database and then re-create the database by using the members in the metaoutline. N Do not remove members from the existing database; only update the Essbase database (this is the default).
Optional. Incremental Update. The level at which dimensions are updated during an incremental member and data load. This is the format: INCUPD=DimID-MbrID,DimID-MbrID,...DimIDMbrID;[;] Values for incremental update are as follows:
DimID represents the dimension ID of the dimension to which the incremental update applies. MbrID represents the member ID of the level zero member of the hierarchy to which the incremental update applies.
Dynamic Restructuring Member Load Options Optional. Specify 1, 2, 3, or 4. Dynamic restructuring member load options set the parameters for restructuring the database during a member load. The option that you select affects any subsequent data loads. Values are as follows: 1 Preserve all data. Select to preserve all existing data that applies to the changed outline when restructuring occurs. This is the default (INCUPD_MEMBER=1). 2 Discard all data. Select to clear all data from the database. 3 Preserve level 0 data. Select to preserve data only for level zero members. This is the optimal restructure option if you change the source database and need to recalculate the data, and if all data required for the calculation is in level zero members. Selecting this option deletes all upper-level blocks before restructuring. This reduces the disk space for restructuring and improves calculation time when the database is recalculated. The upper-level blocks are recreated when you calculate the database. 4 Preserve input data. Select to preserve only those blocks containing data that is loaded. Many applications contain data that is entered at parent levels. Selecting this option prevents deletion of any blocks that are created by data load, whether they are non-level zero or level zero (leaf member) blocks. Dynamic Restructuring Data Load Options Optional. Dynamic restructuring data load options determine how Essbase loads values from a data source to the database. Specify 1, 2, or 3. This is the default (INCUPD_DATA=1).
Parameters
Description Values are as follows: 1 Overwrite. Select to replace the values in the database with the values in the data source. 2 Add. Select to add values in the data source to the existing values in the database. 3 Subtract. Select to subtract the values in the data source from the existing values in the database.
Optional. Use with the Incremental Update parameter. The level at which attribute dimensions are updated when a member load is performed. Values are as follows: 0 Delete all attribute dimensions after the member load is performed. 1 Do not update existing attribute dimensions while updating other types of dimensions (for example: standard, time, or measures dimensions) during a member load. 2 Update all existing attribute dimensions and add new attribute dimensions during a member load.
Essbase Users
Optional. Lists all user IDs that can access the Essbase application and database into which you are loading members. Commas separate user IDs. For example, @@USERS="admin","ljones","ksmith". Optional. The name of the calculation script. The calculation script determines how Essbase calculates the data values in the database. If you do not select a calculation script, Integration Server does not perform a calculation. If the data storage model property of the metaoutline is aggregate storage, this parameter is not applicable. Optional. The name of an ESSCMD script. The script that you define instructs Essbase Server to perform specified actions on the Essbase database. These actions might include changing a member tag or loading data from a worksheet into user-defined members after you load the data from the data source. The ESSCMD script must be in the eis \server\esscript directory. Optional. You can specify that a Unicode application is created on Essbase Server. The default is that no Unicode application is created. This parameter cannot be overwritten. For example, if a previous application exists and you are overwriting that application with the current member load, you cannot overwrite the original Unicode or nonUnicode setting. This is the format: UNICODE=[Y|N] For example: UNICODE=Y
Example
LOADALL OTL=ProductsAnalysis;APP=Products;DBN=Analysis;FLT_ID_MEMBER=2; FLT_ID_DATA=2;REPLACE_ZEROS=Y;OTL_CLEAR=Y;CALC_SCRIPT=#DEFAULT#; INCUPD=1-2,2-1,3-2,4-2,5-1,5-2,5-3;INCUPD_MEMBER=3;INCUPD_DATA=3; @@USERS="TBC","CFO";UNICODE=Y;
LOGOUT
The LOGOUT command logs you out of Integration Server. On a successful logout, the Integration Services Shell prompt changes to LOCAL. Syntax
LOGOUT
If you log out of Integration Server, you no longer have an active session. To execute another command, you must log on to Integration Server again.
SHUTDOWN
The SHUTDOWN command shuts down Integration Server. You must be logged on to Integration Server before you can use this command. Syntax
SHUTDOWN
EXIT
The EXIT command exits Integration Services Shell and closes the Integration Services Shell window. Before you exit from Integration Services Shell, Integration Server automatically executes a LOGOUT command. Syntax
EXIT
Example EXIT logs you out of Integration Server and closes the Integration Services Shell window.
To load members and dimensions with Integration Services Shell, follow the same steps for
loading members that you use with Integration Services Console:
1 Log on to the Integration Server computer by using the LOGIN command; for example
LOGIN cypress
2 Connect to the external data source by using the SETSOURCE command; for example
SETSOURCE DSN=TBC;CODEPAGE=English_UnitedStates.Latin1@Binary;UID=TBC; PWD=password
3 Connect to the Essbase Server computer by using the SETTARGET command; for example
SETTARGET DSN=sequoia;UID=sys;PWD=password
4 Connect to OLAP Metadata Catalog by using the SETCATALOG command; for example
SETCATALOG DSN=TBC_MD;CODEPAGE=English_UnitedStates.Latin1@Binary; UID=TBC;PWD=password
5 Start the member load by using the LOADMEMBER command; for example
LOADMEMBER "OTL=TBC Metaoutline;APP=OLAP_TBC;DBN=Basic;FLT_NAME=Filter1"
To load data with Integration Services Shell, follow the same steps for loading data that you use
with Integration Services Console:
2 Connect to the external data source by using the SETSOURCE command; for example
SETSOURCE DSN=TBC;CODEPAGE=English_UnitedStates.Latin1@Binary;UID=TBC; PWD=password
3 Connect to the Essbase Server computer by using the SETTARGET command; for example
SETTARGET DSN=sequoia;UID=sys;PWD=password
4 Connect to OLAP Metadata Catalog by using the SETCATALOG command; for example
SETCATALOG DSN=TBC_MD;CODEPAGE=English_UnitedStates.Latin1@Binary; UID=TBC;PWD=password
5 Start the data load by using the LOADDATA command; for example
LOADDATA "OTL=TBC Metaoutline;APP=OLAP_TBC;DBN=Basic;FLT_NAME=Filter1"
9
In This Chapter
Naming Restrictions for Applications and Databases ........................................................................ 175 Naming Restrictions for Dimensions, Members, and Aliases ................................................................ 176 Using Dimension and Member Names in Calculation Scripts, Report Scripts, Formulas, and Filters.................... 178
This chapter describes the rules for naming applications, databases, dimensions, members, and aliases in Essbase. For detailed information on creating Essbase applications and databases, see the Essbase product documentation.
Use no more than 8 bytes when naming non-Unicode-mode applications and databases; use no more than 30 characters when naming Unicode-mode applications and databases. Do not use spaces anywhere in the name. Do not use the following special characters anywhere in the name:
* asterisk \ back slash [] brackets : colon , comma = equals > greater than < less than . period
+ plus ? question mark quotation marks ; semicolon ' apostrophe / forward slash tabs | vertical bar
For aggregate storage databases only, do not use any of the following words as application or database names:
Naming Restrictions for Applications and Databases 175
Enter the name in the case in which you want the word displayed. The application or database name is created exactly as you enter it. If you enter the name as all capital letters (for instance, NEWAPP), Essbase does not automatically convert it to upper- and lowercase (for instance, Newapp).
Use no more than 80 characters when naming dimensions, members, or aliases. Names are not case-sensitive unless case-sensitivity is enabled. See Setting Outline Properties in Essbase Administration Services Online Help. Do not use quotation marks (""), brackets ([]), or tabs anywhere in a name.
Note:
Brackets ([]) are supported in block storage outlines, but are not recommended because their use causes errors when converting to an aggregate storage outline.
At the beginning of a dimension or member name, do not use the following characters:
@ at \ back slash {} braces , comma - dash, hyphen, or minus = equals < less than
Do not place spaces at the beginning or end of a name. Essbase ignores spaces at the beginning or end of a name. Do not use the following words as dimension or member names:
Calculation script commands, operators, and keywords. For a list of commands, see the Essbase Technical Reference. Report writer commands. For a list of commands, see the Essbase Technical Reference.
Function names and function arguments. For a list of functions, see the Essbase Technical Reference. Names of other dimensions and members (unless the member is shared), and generation names, level names, and aliases in the database. Any of the following words:
ALL AND ASSIGN CALC CALCMBR COPYFORWARD CROSSDIM CURMBRNAME DIM DIMNAME DIV DYNAMIC EMPTYPARM EQ EQOP EXCEPT EXP EXPERROR FLOAT FUNCTION GE GEN
GENRANGE GROUP GT ID IDERROR INTEGER LE LEVELRANGE LOOPBLOCK LOOPARMS LT MBR MBRNAME MBRONLY MINUS MISSING MUL MULOP NE NON NONINPUT NOT
OR PAREN PARENPARM PERCENT PLUS RELOP SET SKIPBOTH SKIPMISSING SKIPNONE SKIPZERO TO TOLOCALRATE TRAILMISSING TRAILSUM UMINUS UPPER VARORXMBR XMBRONLY $$$UNIVERSE$$$ #MISSING #MI
Note:
If you enable Dynamic Time Series members, do not use the associated generation names, including History, Year, Season, Period, Quarter, Month, Week, or Day.
Using Dimension and Member Names in Calculation Scripts, Report Scripts, Formulas, and Filters
In calculation scripts, report scripts, filter definitions, partition definitions, or formulas, you must enclose member names in quotation marks (" ") for block storage databases, and in brackets ([]) for aggregate storage databases, in the following situations:
The name starts with one or more numerals (for example, 100). The name contains spaces or any of the following characters:
& ampersand * asterisk @ at \ back slash {} braces : colon , comma - dash, hyphen, or minus = equals ! exclamation point ~ tilde > greater than < less than () parentheses % percent . period + plus ; semicolon / forward slash
In calculation scripts and formulas, you must enclose the following member names in quotation marks (" ") for block storage databases, and in brackets ([]) for aggregate storage databases:
BEGIN DOUBLE END MACRO MEMBER RANGE STRING THEN
A
In This Appendix
OLAPICMD Session Return Codes .............................................................................................. 179 Member Load Error Messages .................................................................................................. 180 Data Load Error Messages ...................................................................................................... 183 Drill-Through Report Error Messages ........................................................................................... 184 Miscellaneous Error Messages.................................................................................................. 185 Essbase Error Messages Generated During Data Loads ..................................................................... 187
This appendix provides a partial list of the return codes and error messages generated by Essbase Integration Server and Essbase Server. Return codes and error messages are grouped into two categories: return codes and errors generated by Integration Server and errors generated by Essbase Server for member and data loads. The complete lists of OLAPICMD return codes (unixrtcd.txt and winrtcd.txt) are located in eis/server/docs on UNIX and eis\server\docs on Windows. The complete list of Integration Server error messages is located in eis/server/bin/ error.txt on UNIX and eis\server\bin\error.txt on Windows. For errors generated by Essbase for member and data loads, see Essbase Error Messages Generated During Data Loads on page 187.
The unixrtcd.txt file lists the return codes for UNIX, the corresponding Windows return code, and the associated message text. Because of a limitation in the number of return code values available for use on UNIX, some code numbers are used more than once. When you look up a return code number and find that it is used multiple times, you need to examine the messages for all instances of that return code number. Then you need to determine which
message applies to your situation by understanding the context in which the problem occurred.
The winrtcd.txt file lists the return codes for Windows and the associated message text. The return code values in Windows are unique.
Table 15 shows examples of return codes on UNIX that are used more than once. Italicized words represent variable names inserted in the message text. Refer to the return code value and make a note of it in case you need to contact Hyperion Technical Support with a problem. For the complete list of return codes, refer to the unixrtcd.txt (located in the eis/server/ docs directory on UNIX) and the winrtcd.txt file (located in the eis\server\docs directory on Windows).
Table 15
Examples of Multiple Instances of Integration Services Shell (OLAPICMD) Return Code Values for UNIX Windows Return Code 1195069 Message Contents Failed to add Essbase member member_name (DUPLICATE, ignored). There must be an active session to process this command. Data source error. Could not locate the metadata attributes. Failed to create a local context. ODBC Error. Encountered unknown ODBC exception while opening database. Failed to get user attributes. Cube builder error. Client request error. Invalid number of parameters.
094 094
1195358 2002014
Essbase Integration Server Member Load Error Messages Message Content Detected metadata join problem (table.column < - > table.column). Encountered unknown ODBC exception while opening database. Aborting the command. Failed to add database Essbase database. Failed to add dimension dimension.
1195064 1195065
Windows Message Number 1195066 1195067 1195068 1195069 1195070 1195071 1195072 1195073 1195074 1195075 1195076 1195077 1195078 1195079 1195080 1195083 1195085 1195088 1195089 1195093 1195095 1195097 1195098 1195115 1195126 1195129 1195130 1195132
Message Content Failed to add Essbase member member (%d). Failed to add Essbase member member (child of shared parent). Failed to add Essbase member member (DUPLICATE). Failed to add Essbase member member (DUPLICATE, ignored). Failed to add Essbase member member (no parent or sibling members found). Failed to add Essbase member member, %d. Failed to add Essbase member member; the member exists. Failed to add formula formula. Failed to add member member. Failed to add metaoutline member member. Failed to add user attribute user-defined attribute. Failed to create a local context. Failed to create a local outline. Failed to create application Essbase application. Failed to create dimension dimension. Failed to get information for parent of recursive hierarchy member. Failed to get metadata for dimension.member. Failed to initialize Essbase API. Failed to open outline Essbase outline. Failed to restructure Essbase outline. Failed to update Essbase server (NULL HEADER). Filter metaoutline filter doesnt exist for metaoutline metaoutline. Filter expression length exceeded the limit of maximum bytes bytes. Incremental update specification incremental update specification string is incorrect. Invalid aggregation function aggregate function for table.column. Invalid esscmd script name esscmd script. Invalid parameter parameter. Invalid parameter string.
Message Content Metadata error: property ID is not a valid property for member or dimension. Metaoutline validation error. Recursive hierarchy member, a recursive member cannot have aggregation filter. Metaoutline validation error. Filter = member filter, is not associated with any member. Unknown exception encountered while opening dimension dimension. Unknown exception encountered while opening member member. Unknown exception encountered while opening metaoutline metaoutline. Failed to Login to Essbase Server. Failed to Get Essbase Settings. Failed to unload database. Failed to delete database. Failed to close outline. Failed to read outline. Failed to unlock Database database. Failed to save outline. Unable to process state for Essbase command. Failed to get access permissions for application. Failed to set access permissions for application. Failed to get access permissions for database. Failed to set access permissions for database. No member found. Essbase Error: error definition. Failed to get user attributes. Failed to add member to outline. Failed to calculate the data. Failed to get parent. Failed to get member. Failed to move member. Failed to get shared member.
1195144 1195188 1195189 1195190 1195341 1195342 1195344 1195345 1195346 1195347 1195349 1195350 1195351 1195352 1195353 1195354 1195355 1195356 1195357 1195358 1195360 1195361 1195362 1195363 1195364 1195365
Windows Message Number 1195366 1195367 1195377 1195378 1195379 1195380 1195381 1195382 1195383 1195384 1195385 1195386 1195387 1195389
Message Content Failed to get member information. Failed to sort outline. IS Error: Unable to delete outline. IS Error: SQL Generation for member load failed. IS Error: Database close operation failed after load. IS Error: Catalog close operation failed after load. IS Error: Essbase close operation failed after load. IS Error: No OLAP model name specified. IS Error: Failed to open OLAP model OLAP model. IS Error: Failed to close OLAP Metadata Catalog OLAP Metadata Catalog. Failed to create virtual model. Member load completed with errors. Member load terminated due to error. Failed to get dimbuild SQL.
Essbase Integration Server Data Load Error Messages Message Content Cell load exceptions encountered. Data load exceptions encountered. ? , amount. Failed to activate Essbase application.Essbase.database. Failed to execute esscmd script esscmd script. Data load completed with errors. Data load terminated with errors. Failed to get data load SQL.
Windows Message Number 1195034 1195047 1195063 1195082 1195369 1195370 1195388
Message Content Number of data load SQLs not equal to number of hierarchies.
Essbase Integration Server Drill-Through Report Error Messages Message Content Failed to enable DTS Member member. Failed to get the universal member handle. Intersection element member does not exist in table.column. Invalid parameter count for the drill-through request. Aborting... Received NULL pointer in DTAttributes( ) for Report = Drill-Through report. Received NULL pointer in GetDTData ( ) for Report = Drill-Through report. Received NULL pointer in GetDTDomain ( ). Received NULL pointer in GetDTReport ( ). Received NULL pointer in SetDTReport ( ). Too many Drill-Through reports defined (exceeded Essbase metadata size limit). DrillThrough is disabled for this outline. Unknown exception in GetDTReport ( ). Essbase Error: Invalid Drill-Through Metadata. IS Error: Data load completed with errors. IS Error: Data load terminated due to errors. IS Error: Unable to log in to data source. IS Error: Unable to log in to OLAP Metadata Catalog. IS Error: Unable to log in to Essbase Server.
Windows Message Number 1195081 1195087 1195125 1195131 1195153 1195154 1195155 1195156 1195157 1195183
Message Content IS Error: Unable to read Metaoutline information. IS Error: Data calculation failed. IS Error: Esscmd script execution failed.
Essbase Integration Server Miscellaneous Error Messages Message Contents Metaoutline failed validation. Database measures not specified. Add job failed. Syntax Syntax error at character location in the filter expression filter. Member cannot have aggregation filter. Only leaf members can have aggregation filters. Error message unavailable for this error. Essbase server on Essbase server computer is not enabled with Integration Services option. Please inform your system administrator. Failed to get job info (OS error Job-ID). Failed to get the ODBC message. Failed to remove temporary batch file (OS error operating system error). Failed to remove job (OS error operating system error). Failed to remove old job (OS error operating system error). Failed to schedule job (OS error operating system error). Fatal error. Internal system error. Please contact Technical Support with the error number 1999999. Invalid command line option OLAP Integration Server switch. Invalid datatype specification data type.
1195127 1195128
Windows Message Number 1195133 1195159 1195185 1195187 1195206 1195207 1195208 1195333 1195334 1195336 1195338 1195343
Message Contents Invalid session identifier; please log in to Integration Services again. Replace Failed member transformation. Trying to access invalid memory. Contact Technical Support. Unknown ODBC exception encountered while closing database. Syntax Syntax error at character location in the filter expression filter. Integration Services is already running on this machine. Failed to release locks. ODBC Error: Encountered unknown ODBC exception while opening database. ODBC Error: Encountered unknown ODBC exception while closing database. ODBC Error: Failed to get ODBC Error message for error number. IS Error: Error during parsing Load Parameters. Essbase Error: Essbase Server on Essbase server computer is not enabled with Integration Services option. Failed to unload database. Failed to get information for Database database. IS Error: Unable to get tables for pattern name SQL regular expression pattern from data source. IS Error: Unable to get columns for table data source table from data source. IS Error: Unable to extract foreign keys for tables data source table, data source table from data source. IS Error: Failed to bind columns. IS Error: Failed to fetch next row. IS Error: Failed to get number of columns. IS Error: Failed to get column attributes. IS Error: Failed to get statement handle. IS Error: Failed to release statement handle. IS Error: Failed to get number of rows. IS Error: Unable to delete the model model name.
1195392 1195393
Locate the subfolder for the data load that you want to review and open the dataload.txt file to view the errors.
Table 20
Essbase Messages Generated During Loads Message Contents Member not found in database. Insufficient access to store data. Bad data value supplied. Record rejected because of duplicate member names. Member/Data unknown. Record rejected because of dimension conflicts with Header Name.
B
In This Appendix
Artifact Limits..................................................................................................................... 189 Source Database Artifact Naming Limits ...................................................................................... 192 Setting the Compression Dimension ........................................................................................... 193 Hybrid Analysis Guidelines ...................................................................................................... 194 Using Advanced Relational Access............................................................................................. 197 Drill-through Report Guidelines ................................................................................................. 201 Unicode Guidelines .............................................................................................................. 202 Duplicate Member Names Guidelines ......................................................................................... 202
This appendix describes Integration Services artifacts and database naming limits; Hybrid Analysis, Drill-through, and Unicode guidelines. You should be familiar with these limits and guidelines before you begin using Integration Services. For restrictions and guidelines on naming artifacts, see Chapter 9, Naming Restrictions for Essbase Applications, Databases, and Members. See also Integration Services Console online help.
Artifact Limits
Table 21 contains a list of limits that you may encounter when creating or manipulating Integration Services artifacts.
Table 21
Limit for Applications Non-Unicode application limit: 29 bytes Unicode-mode application limit: 50 characters
Application name
Application description
Custom-defined function name Custom-defined macro name Custom-defined function specification Custom-defined macro specification
Non-Unicode application limit: 127 bytes. MaxL and the API truncate characters after 127 bytes. Unicode-mode application limit: 128 characters. MaxL and the API truncate characters after 128 characters. In either case, no truncation on server. No error is displayed if truncation occurs.
Non-Unicode application limit: 255 bytes. After 255 bytes, characters are truncated by MaxL and API. Unicode-mode application limit: 256 characters. MaxL and the API truncate characters after 256 characters. In either case, no truncation on server. No error is displayed if truncation occurs.
Database description
Directory path For example: eis/server/bin File names for calculation scripts, report scripts, and rules files
Non-Unicode application limit: 256 bytes Unicode-mode application limit: 1024 bytes Non-Unicode application limit: 8 bytes Unicode-mode application limit: If included within a path, the smaller of the following two values:
1024 bytes The limit established by the operating system If not included within a path, as in some MaxL statements, 1024 bytes.
Filter name
Group name
Artifact Linked reporting artifact cell note Linked reporting artifact URL Member comment field
Limit for Applications 599 bytes 512 characters (always single-byte characters) Non-Unicode application limit: 255 bytes Unicode-mode application limit: 256 characters
8192 bytes Non-Unicode application limit: 80 bytes Unicode-mode application limit: 80 characters
80 characters 80 characters Non-Unicode application limit: 100 bytes Unicode-mode application limit: 100 characters
Substitution variable name Substitution variable value Trigger name User-defined query User names
80 bytes 255 bytes 30 bytes No limitations Non-Unicode application limit: 30 bytes Unicode-mode application limit: 30 characters
Variable names Data Load and Dimension Building Limits Data load query Member load query Number of alias tables associated with a metaoutline Number of error messages written to a data load or dimension build error log (DATAERRORLIMIT in essbase.cfg) Selection and rejection criteria
32 bytes
No limitations No limitations 9, including the Default alias table Default 1000, minimum 1, maximum 65000
Number of characters that describe selection and rejection criteria: combination of all criteria limited to 32 KB
Created in Formula Editor: 64 KB. Formulas in calculation scripts are not subject to this limit. Created in MaxL, using multi-byte characters: 40 KB.
Artifact
Limit for Applications Formulas in calculation scripts are not subject to these limits.
Number of users
30,000. Errors can occur if you create more than 30,000 users. Approximately 1,000,000 explicitly defined in an Essbase outline for block storage Approximately 20,000,000 explicitly defined in an Essbase outline for aggregate storage Hybrid Analysis and some uses of partitions enable access to many more members than are explicitly listed in an outline, the actual number of members accessible through the database is much higher. Longer names, which often occur if multi-byte characters are used, decrease the number of members that are allowed.
Table names may contain blank spaces. Column names may contain blank spaces and question marks.
Note:
Table 22
Unsupported Characters in Table and Column Names Character Description () . | ' \ parenthesis period vertical pipe single quotation mark backslash
Character Description & , = @ quotation mark ampersand comma equal sign at sign
Character Description # \t pound sign Tab character blank space + {} plus sign dash, minus sign, or hyphen braces / : * ? < >
Character Description forward slash colon asterisk question mark less than sign more than sign
Guidelines
The compression dimension option is ignored if the outline is being built with Essbase versions earlier than 9.3. In versions less than 9.3, the Accounts dimension is tagged as the compression dimension internally. The following dimensions cannot be specified as the compression dimension:
Multiple hierarchy dimensions Attribute dimensions A base dimension with an attribute association
If you tag a compression dimension with an Outline Hierarchy Information option other than Dynamic at Dimension Level, that dimension will be tagged as a dynamic dimension when it is loaded into Essbase. The compressions dimension option applies to metaoutlines specified for aggregate storage only. To facilitate switching between aggregate and block data storage models, the option is always available, whether the metaoutline is specified for block storage or aggregate storage. If the metaoutline is specified for block storage, this option is ignored.
A single Essbase database can be associated with only one hybrid analysis relational data source. A hybrid analysis data source can consist of only one relational database. Hybrid Analysis supports data that is stored using either block storage or aggregate storage. Hybrid Analysis supports Unicode-enabled data sources.
Dimensions Guidelines
Hybrid Analysis is not supported on accounts dimensions. If the time dimension contains hybrid analysis-enabled members, the time dimension does not support Dynamic Time Series. Hybrid Analysis is not supported on user-defined dimensions. In an outline that is hybrid analysis-enabled, you can perform operations and analyses on dimensions that have attributes attached to one or more levels. The attribute dimension should be fully loaded into Essbase. Only the first hierarchy of a dimension with alternate hierarchies can have members enabled for hybrid analysis on its lowest levels. When building a dimension that is enabled for hybrid analysis, you must ensure that the column in the data source table that contributes to the leaf level of the Essbase portion of the dimension is non-nullable.
Members Guidelines
Only the lowest level members of a dimension can be enabled for hybrid analysis. You should not rename a hybrid analysis-enabled member. If you rename a member, the member may not be retrieved the next time you perform a drill-through operation. Hybrid Analysis supports only parent-child prefixing on member names. Essbase does not support aliases for members that are enabled for hybrid analysis. Hybrid Analysis does not support scaling of measures dimension members using any of the operators + (addition), - (subtraction), * (multiplication), and / (division). If you use the scaling operators, drill-through queries into hybrid analysis data may show a mismatch between aggregated level-0 values in the Essbase database and the corresponding detail values in your data source. Essbase ignores all member properties, such as formulas, UDAs, and aliases for members that are enabled for hybrid analysis.
You can associate an attribute member with a member enabled for hybrid analysis but you must make the association by metaoutline member name and not by level. Essbase supports drill-through operations defined on members that are enabled for Hybrid Analysis. You cannot apply properties of the Account Info tab to Hybrid Analysis members. Specifically, Time Balance, Skip, Variance Reporting, or Currency Conversion properties are not supported for Hybrid Analysis-enabled members.
Spreadsheet Guidelines
Hybrid Analysis does not return numeric data in a spreadsheet if the member from the Accounts dimension is part of a ragged hierarchy. Hybrid Analysis is not supported with the Member Selection feature. You cannot select Hybrid Analysis members from the Member Selection dialog box. If you have multiple levels of hybrid analysis members in your outline, performing a zoom out operation on the bottom level hybrid analysis member takes you directly to the Essbase parent member, bypassing all other hybrid analysis levels.
Operations Guidelines
Hybrid Analysis supports Dynamic Time Series. Essbase requires the OLAP Metadata Catalog created in Integration Services in order to drill down in a hybrid analysis data source. Hybrid Analysis does not support transparent, replicated, or linked partitions. Hybrid Analysis supports recursive hierarchies.
A database value cannot have a separator character that is the same as the one used for the prefix or suffix. A member name cannot be more than 80 characters (excluding blanks). A prefix or suffix must always have a separator associated with it. The data source database value cannot have trailing blanks. If spaces are converted to underscores during a transformation, then the Hybrid Analysis Manager assumes there are no underscores present in the database value. The use of all ancestors as a prefix or as a suffix is not supported.
Dropping spaces from around a member name Applying a prefix without a separator Applying names prefixes for all parents with or without a separator Applying a suffix without a separator Applying all suffixes of parent names with or without a separator Applying scaling to measures
Formulas are supported only on a measures dimension. Formulas cannot be attached to relational members. Formulas cannot reference a relational member by name. Member set functions (such as @CHILDREN and @DESCENDANTS), which generate member lists in a formula, execute only in the Essbase portion of the outline.
If a formula contains one or more functions that are not supported by Hybrid Analysis, Essbase returns the following error message:
Error executing formula for member [member-name-to-which-formula-is-attached] (line [line# where the offending function appears inside the formula): function [Name of the offending function] cannot be used in Hybrid Analysis.
Relationship Functions
Hybrid Analysis does not support functions that look up specific values in the database based on current cell location and a series of parameters. Examples:
Range Functions
Hybrid Analysis does not support functions that take a range of members as arguments. Rather than return a single value, these functions calculate a series of values internally based on the range specified. Examples:
@PRIOR @SHIFT @PRIORS @SHIFTS @NEXT @MDSHIFT @MOVSUM @MOVAVG @ALLOCATE @MDALLOCATE @VAR @VARPER @MEDIAN @RANK
Attribute Functions
Hybrid Analysis does not support any Essbase functions that deal with attributes. Examples:
@ATTRIBUTEVAL @ATTRIBUTESVAL @WITHATTR
Integration Server to build to a specified member level (Hybrid Analysis) or build only to the dimension level (Advanced Relational Access). Building down to a specified level produces a smaller Essbase outline and a smaller multidimensional database. Smaller databases can be useful for users with limited disk space who do not need to see the lowest level of detail. Integration Services uses Advanced Relational Access to give Essbase end users direct access to data from relational databases or data warehouses. This feature enables users to perform online analytical processing (OLAP) on very large data sets. In Integration Services Console, Advanced Relational Storage is enabled at the metaoutline level. When the Relational Storage option is selected, all members of all non-accounts dimensions are automatically enabled for relational storage. Alternatively, you can enable relational storage on selected non-accounts dimensions. When a metaoutline is enabled for Advanced Relational Access, end users are able to query directly on relationally-stored members. Essbase Server issues SQl queries to retrieve data from the database or data warehouse. All members of the dimension are accessed directly from the relational data source.
Note:
For detailed information on enabling Advanced Relational Access, see Integration Services Console online help.
General Guidelines
Here are general guidelines to keep in mind when using Advanced Relational Access:
Only outlines that are valid for aggregate storage can be Advanced Relational Access outlines. Block storage is not supported. Dynamic Time Series members are not supported. Advanced Relational Access requires MDX formulas.
Note:
An MDX query made against a ragged hierarchy in an Essbase cube returns data results that can differ significantly from the results obtained when the same MDX query is made against a ragged hierarchy in an Advanced Relational Access cube. See the section Different Values Loaded in Ragged Hierarchies on page 200.
Members enabled for Advanced Relational Access are shown in Administration Services Console Outline Viewer but are not shown in the Outline Editor. A Time dimension from the Fact table is not supported. Hyperion strongly recommends that you create a separate Time dimension table with time data only.
When there are multiple measures defined in a metaoutline, consolidation at the highest level is based on the first measure only.
Advanced Relational Access cannot be enabled in metaoutlines that are connected to multiple data sources. Star and snowflake schemas are supported. Teradata RDBMS provides different types of SQL date types, but Advanced Relational Access only supports the DATE SQL date type.
Dimension Guidelines
Here are guidelines to keep in mind when working with dimensions in Advanced Relational Access:
Advanced Relational Access does not support recursive dimensions/hierarchies. Alternate hierarchies in standard dimensions are not allowed. Attribute dimensions are ignored. User-defined dimensions are not supported. When a dimension is enabled for relational storage, Integration Server builds only to the dimension level. All members of the dimension are accessed directly from the relational data source. Accounts dimension must be created from the fact table. For the accounts dimension, no hierarchy is supported. All members should be of generation 2. Aliases are not supported.
Member Guidelines
Here are guidelines to keep in mind when working with members in Advanced Relational Access:
The first member of the accounts dimension must be a base or stored measure. It cannot be a user-defined member. Aliases are not supported. Duplicate members are not supported. Shared members are not supported. Relational members are not automatically expanded when part of an outline is expanded. To view the relational members, you must manually click the member node containing the relational members.
Relational members are not automatically collapsed when part of an outline is collapsed. To collapse the relational members, you must manually click the member node containing the relational members. In Integration Services Console, you can specify multiple sort order columns for each generation, selecting ascending or descending order. In Integration Services Console, you can associate a key column with each generation name and tag the key column as unique.
IBM DB2: REAL Oracle: NCHAR and NVARCHAR2 Microsoft SQL Server: TINYINT and SMALLINIT
Note:
The data types TINYINT and SMALLINT cannot be used as a measure; however, they can be used as members.
Teradata: FLOAT
Note:
Your member set column names should not be based on columns of FLOAT data types.
The accounts dimension completely resides in Essbase. For all non-accounts dimensions, only the root members (the dimension level) reside in the Essbase database. All other non-accounts members are accessed directly from the relational database.
Drill-through operations can be performed on a Unicode Essbase database. Drill-through reports may be directed to an alternate Integration Server. Select the alternate server in the OLAP Metaoutline Properties dialog box in Integration Services Console. Drill-through operations can be performed with Dynamic Time Series. Drill-through operations can be performed on an alternate data source. An alternate data source is a source other than the primary or secondary data source used to create a metaoutline. The alternate data source must contain the same data structure, including column names and data types, as the primary or secondary data source originally used to create the report. Drill-through operations can be performed on an alternate data source that is in a second language. Integration Services is not capable of processing drill-through report requests from more than one instance of Essbase running on one server. In order for Integration Services to successfully process requests for drill-through reports from instances of Essbase running on non-standard Agent ports, an essbase.cfg file with the non-standard AGENTPORT setting must be present in eis/server/bin at the time Integration Services is started. If you create drill-through reports for a metaoutline based on multiple data sources, the SQL used for each drill-through report cannot include a join across a data source boundary; that is, each drill-through report can be based on only one data source. A metaoutline can contain multiple drill-through reports but all must be based on a single data source. Multi-cell drill-through is supported under the following conditions:
All members selected for multi-cell drill-through come from same physical table and column in the database. All members selected for multi-cell drill-through come from the same level in the metaoutline. The selected members cannot come from more than one hierarchy.
The metaoutline in which you are designing a drill-through report cannot contain alternate recursive hierarchies. For example, you cannot insert the same member from a recursive hierarchy into a dimension more than once. You cannot use the $$ substitution variable in the template SQL (user-defined SQL) for a drill-through report when the intersection level of the dimension is defined at Generation 1 and the dimension is built from a parent/child table. Oracle does not recommend using Drill-Through Template SQL for OLAP metaoutlines that support duplicate member names. Integration Server generates special tags internally to uniquely identify the duplicate members. If you still choose to use Drill-through Template SQL with OLAP metaoutlines that support duplicate member names, refer to the Essbase Integration Services Online Help for guidelines.
There is a limit to the amount of information that can be stored in the Universal Member Comments (UMC) of an Essbase outline, meaning that users can typically store information for about 40 to 50 drill-through reports only. Because of this limitation, Oracle recommends that users design their drill-through reports so that the total number of reports is kept to under 40.
Unicode Guidelines
You must use the manual catalog creation procedures to create a Unicode-enabled OLAP Metadata Catalog, as described in Chapter 3, Creating, Upgrading, and Deleting OLAP Metadata Catalogs. Drill-through operations are not supported on a Unicode Essbase database. Integration Services does not support Unicode databases in Microsoft SQL Server, Sybase and Informix RDBMSs. If the database is UTF-8 but you are using only one language, select the UTF-8 option from the Code Page drop-down list whenever you access the database with the Login, OLAP Metadata Catalog Setup, Set Login Defaults, and Data Source dialog boxes.
Dimension names must be unique. In a dimension, a user-defined member cannot have the same name as a measure.
Duplicate member names are not allowed under the same parent. This is also true for userdefined members. A member cannot have children with duplicate names. Duplicate members may be shared or non-shared members. Duplicate members are supported in both aggregate storage and block storage. Duplicate member names are not supported if there are multiple data sources. Duplicate member names support duplicate aliases.
Drill-through operations containing duplicate member names are supported. Oracle does not recommend using user-defined data load SQL or Drill-Through Template SQL for metaoutlines that support duplicate member names. Integration Server generates special tags internally to uniquely identify the duplicate members. If you still choose to override data load SQL commands, exercise caution. You may have records rejected in error. It is especially recommended that you not change column names in SQL. If you still choose to use Drill-through Template SQL with OLAP metaoutlines that support duplicate member names, refer to the Essbase Integration Services Online Help for guidelines.
For detailed information on creating an Oracle Essbase outline with duplicate member names, see Integration Services Console online help.
Glossary
! See bang character (!). #MISSING See missing data (#MISSING). access permissions A set of operations that a user can
permission.
adjustment See journal entry (JE). Advanced Relational Access The integration of a relational database with an Essbase multidimensional database so that all data remains in the relational database and is mapped to summary-level data residing in the Essbase database. agent An Essbase server process that starts and stops
perform on a resource.
accessor Input and output data specifications for data
mining algorithms.
account blocking The process by which accounts accept input data in the consolidated file. Blocked accounts do not receive their value through the additive consolidation process. account eliminations Accounts which have their values set to zero in the consolidated file during consolidation. account type How an account's value flows over time, and its
applications and databases, manages connections from users, and handles user-access security. The agent is referred to as ESSBASE.EXE.
aggregate cell A cell comprising several cells. For example, a
sign behavior. Account type options can include expense, income, asset, liability, and equity.
accountability map A visual, hierarchical representation of the responsibility, reporting, and dependency structure of the accountability teams (also known as critical business areas) in an organization. accounts dimension A dimension type that makes accounting
data cell that uses Children(Year) expands to four cells containing Quarter 1, Quarter 2, Quarter 3, and Quarter 4 data.
aggregate function A type of function, such as sum or
than Hold.
activity-level authorization Defines user access to applications
and the types of activities they can perform on applications, independent of the data that will be operated on.
ad hoc report An online analytical query created on-the-fly
designed to support large-scale, sparsely distributed data which is categorized into many, potentially large dimensions. Upper level members and formulas are dynamically calculated, and selected data values are aggregated and stored, typically with improvements in overall aggregation time.
aggregate view A collection of aggregate cells based on the
by an end user.
adapter Software that enables a program to integrate with
levels of the members within each dimension. To reduce calculation time, values are pre-aggregated and stored as aggregate views. Retrievals start from aggregate view totals and add up from there.
Glossary 205
the allocation model that controls the direction of allocated costs or revenue flow within Profitability and Cost Management.
attribute Characteristic of a dimension member. For
identifiable column descriptor you can display the alias instead of the member name.
alias table A table that contains alternate names for
example, Employee dimension members may have attributes of Name, Age, or Address. Product dimension members can have several attributes, such as a size and flavor.
attribute association A relationship in a database outline
members.
alternate hierarchy A hierarchy of shared members. An alternate hierarchy is based upon an existing hierarchy in a database outline, but has alternate levels in the dimension. An alternate hierarchy allows the same data to be seen from different points of view. ancestor A branch member that has members below it. For
whereby a member in an attribute dimension describes a characteristic of a member of its base dimension. For example, if product 100-10 has a grape flavor, the product 100-10 has the Flavor attribute association of grape. Thus, the 100-10 member of the Product dimension is associated with the Grape member of the Flavor attribute dimension.
Attribute Calculations dimension A system-defined dimension that performs these calculation operations on groups of members: Sum, Count, Avg, Min, and Max. This dimension is calculated dynamically and is not visible in the database outline. For example, using the Avg member, you can calculate the average sales value for Red products in New York in January. attribute dimension A type of dimension that enables analysis based on the attributes or qualities of dimension members. attribute reporting A reporting process based on the attributes
example, the members Qtr2 and 2006 are ancestors of the member April.
appender A Log4j term for destination. application (1) A software program designed to run a specific task or group of tasks such as a spreadsheet program or database management system. (2) A related set of dimensions and dimension members that are used to meet a specific set of analytical and/or reporting requirements. application currency The default reporting currency for the
application.
area A predefined set of members and values that makes up
a partition.
arithmetic data load A data load that performs operations on
attribute type that enables different functions for grouping, selecting, or calculating data. For example, because the Ounces attribute dimension has the type numeric, the number of ounces specified as the attribute of each product can be used to calculate the profit per ounce for that product.
authentication Verification of identity as a security measure. Authentication is typically based on a user name and password. Passwords and digital signatures are forms of authentication. authentication service A core service that manages one
example, scripts, forms, rules files, Interactive Reporting documents, and financial reports. Also known as an object.
assemblies Installation files for EPM System products or
components.
asset account An account type that stores values that
authentication system.
auto-reversing journal A journal for entering adjustments that
206 Glossary
every report and book in the batch. While scheduling the batch, you can set the members selected on the batch POV.
batch processing mode A method of using ESSCMD to write a batch or script file that can be used to automate routine server maintenance and diagnostic tasks. ESSCMD script files can execute multiple commands and can be run from the operating system command line or from within operating system batch files. Batch files can be used to call multiple ESSCMD scripts or run multiple instances of ESSCMD. block The primary storage unit which is a multidimensional
for measurement and categorization. (2) A report aspect used to arrange and relate multidimensional data, such as filters, pages, rows, and columns. For example, for a data query in Simple Basic, an axis can define columns for values for Qtr1, Qtr2, Qtr3, and Qtr4. Row data would be retrieved with totals in the following hierarchy: Market, Product.
backup A duplicate copy of an application instance. balance account An account type that stores unsigned values that relate to a particular point in time. balanced journal A journal in which the total debits equal the
total credits.
bang character (!) A character that terminates a series of
report commands and requests information from the database. A report script must be terminated with a bang character; several bang characters can be used within a report script.
bar chart A chart that can consist of one to 50 data sets, with
categorizing and storing data based on the sparsity of data values defined in sparse dimensions. Data values are stored in blocks, which exist only for sparse dimension members for which there are values.
Blocked Account An account that you do not want calculated
any number of values assigned to each data set. Data sets are displayed as groups of corresponding bars, stacked bars, or individual bars in separate rows.
base currency The currency in which daily business
one or more attribute dimensions. For example, assuming products have flavors, the Product dimension is the base dimension for the Flavors attribute dimension.
base entity An entity at the bottom of the organization
displayed on a personal page of a user. The two types of bookmarks are My Bookmarks and image bookmarks.
bounding rectangle The required perimeter that encapsulates
the Interactive Reporting document content when embedding Interactive Reporting document sections in a personal page, specified in pixels for height and width or row per page.
broadcast message A simple text message sent by an administrator to a user who is logged on to a Planning application. The message displays information to the user such as system availability, notification of application refresh, or application backups. budget administrator A person responsible for setting up, configuring, maintaining, and controlling an application. Has all application privileges and data access permissions.
in batch; for example, a calculation script or a full database calculation. Dynamic calculations are not considered to be batch calculations.
batch file An operating system file that can call multiple
ESSCMD scripts and run multiple sessions of ESSCMD. On Windows-based systems, batch files have BAT file extensions. On UNIX, batch files are written as a shell script.
batch loader An FDM component that enables the processing of multiple files.
Glossary 207
Choice of a build method is based on the format of data in data source files.
business process A set of activities that collectively accomplish a business objective. business rules Logical expressions or formulas that are
example, Month.
cause and effect map Depicts how the elements that form
your corporate strategy relate and how they work together to meet your organization's strategic goals. A Cause and Effect map tab is automatically created for each Strategy map.
CDF See custom-defined function (CDF). CDM See custom-defined macro (CDM). cell (1) The data value at the intersection of dimensions in a multidimensional database; the intersection of a row and a column in a worksheet. (2) A logical group of nodes belonging to one administrative domain. cell note A text annotation for a cell in an Essbase database.
consolidated or aggregated. A calculation script may also contain commands that specify allocation and other calculation rules separate from the consolidation process.
calculated member in MaxL DML A member designed for analytical purposes and defined in the optional WITH section of a MaxL DML query. calculated member in MaxL DML A member designed for analytical purposes and defined in the optional WITH section of a MaxL DML query. calculation The process of aggregating data, or of running a
visual nature expedites analysis, color-coding, and visual cues that aid comparisons.
chart template A template that defines the metrics to display
in Workspace charts.
child A member with a parent above it in the database
outline.
choice list A list of members that a report designer can specify for each dimension when defining the report's point of view. A user who wants to change the point of view for a dimension that uses a choice list can select only the members specified in that defined member list or those members that meet the criteria defined in the function for the dynamic list. clean block A data block that where the database is fully
some values or formula calculations have changed. You must reconsolidate to get the correct values for the affected entity.
calendar User-defined time periods and their relationship
to each other. Q1, Q2, Q3, and Q4 comprise a calendar or fiscal year.
cascade The process of creating multiple reports for a subset of member values. Catalog pane Displays a list of elements available to the active
calculated, if a calculation script calculates all dimensions at once, or if the SET CLEARUPDATESTATUS command is used in a calculation script.
cluster An array of servers or databases that behave as a
section. If Query is the active section, a list of database tables is displayed. If Pivot is the active section, a list of results columns is displayed. If Dashboard is the active section, a list of embeddable sections, graphic tools, and control tools are displayed.
single resource which share task loads and provide failover support; eliminates one server or database as a single point of failure in a system.
clustered bar charts Charts in which categories are viewed
side-by-side; useful for side-by-side category analysis; used only with vertical bar charts.
208 Glossary
characters. Different code pages support different sets of characters. Each computer contains a code page setting for the character set requirements of the language of the computer user. In the context of this document, code pages map characters to bit combinations for non-Unicode encodings. See also encoding.
column A vertical display of information in a grid or table. A column can contain data from one field, derived data from a calculation, or textual information. committed access An Essbase Kernel Isolation Level setting
the consolidation of the node of the hierarchy. This rule can contain customer specific formulas appropriate for the correct consolidation of parent balances. Elimination processing can be controlled within these rules.
content Information stored in the repository for any type of
file.
content browser A Component that allows users to Browse
that affects how Essbase handles transactions. Under committed access, concurrent transactions hold long-term write locks and yield predictable results.
computed item A virtual column (as opposed to a column that is physically stored in the database or cube) that can be calculated by the database during a query, or by Interactive Reporting Studio in the Results section. Computed items are calculations of data based on functions, data items, and operators provided in the dialog box and can be included in reports or reused to calculate other data. configuration file The security platform relies on XML documents to be configured by the product administrator or software installer. The XML document must be modified to indicate meaningful values for properties, specifying locations and attributes pertaining to the corporate authentication scenario. connection file See Interactive Reporting connection file
certification and assessment information, especially helpful for meeting Sarbanes-Oxley requirements.
conversion rate See exchange rate. cookie A segment of data placed on your computer by a Web
site.
correlated subqueries Subqueries that are evaluated once for every row in the parent query; created by joining a topic item in the subquery with a topic in the parent query. critical business area (CBA) An individual or a group
organized into a division, region, plant, cost center, profit center, project team, or process; also called accountability team or business area.
critical success factor (CSF) A capability that must be established and sustained to achieve a strategic objective; owned by a strategic objective or a critical process and is a parent to one or more actions. crosstab reporting Categorizes and summarizes data in table format. The table cells contain summaries of the data that fit within the intersecting categories. For example, a crosstab report of product sales information could show size attributes, such as Small and Large, as column headings and color attributes, such as Blue and Yellow, as row headings. The cell in the table where Large and Blue intersect could contain the total sales of all Blue products that are sized Large. cube A block of data that contains three or more dimensions. An Essbase database is a cube.
(.oce).
consolidated file (Parent) A file into which all of the business
dependent entities to parent entities. For example, if the dimension Year consists of the members Qtr1, Qtr2, Qtr3, and Qtr4, its consolidation is Year.
consolidation file (*.cns) The consolidation file is a graphical
interface that enables you to add, delete or move Strategic Finance files in the consolidation process using either a Chart or Tree view. It also enables you to define and modify the consolidation.
Glossary 209
options for a model to build an outline and load data into an Essbase application and database.
cube schema In Essbase Studio, the metadata elements, such as measures and hierarchies, representing the logical model of a cube. currency conversion A process that converts currency values in a database from one currency into another. For example, to convert one U. S. dollar into the European euro, the exchange rate (for example, 0.923702) is multiplied with the dollar (1* 0.923702). After conversion, the European euro amount is .92. Currency Overrides In any input period, the selected input
provide an interactive summary of your business. Dashboards enable you to build and deploy analytic applications.
data cache A buffer in memory that holds uncompressed
data blocks.
data cell See cell. data file cache A buffer in memory that holds compressed
method can be overridden to enable input of that period's value as Default Currency/Items. To override the input method, enter a pound sign (#) either before or after the number.
currency partition A dimension type that separates local
averages, maximums, counts, and other statistics, that summarize groupings of data.
data load The process of loading data to dimensions and
currency members from a base currency, as defined in an application. Identifies currency types, such as Actual, Budget, and Forecast.
custom calendar Any calendar created by an administrator. custom dimension A dimension created and defined by users. Channel, product, department, project, or region could be custom dimensions. custom property A property of a dimension or dimension
submitting source data into the target system. Typically, there is one FDM data load location for each source file loaded to the target system.
data load rules A set of criteria that determines how to load data from a text-based file, a spreadsheet, or a relational data set into a database. data lock Prevents changes to data according to specified
developed in Java and added to the standard Essbase calculation scripting language using MaxL. See also customdefined macro (CDM).
custom-defined macro (CDM) Essbase macros written with
Essbase calculator functions and special macro functions. Custom-defined macros use an internal Essbase macro language that enables the combination of calculation functions and they operate on multiple input parameters. See also custom-defined function (CDF).
cycle through To perform multiple passes through a database
210 Glossary
dimension where measures are represented. The cell values are displayed as formatted dates. Dates as measures can be useful for types of analysis that are difficult to represent using the Time dimension. For example, an application may need to track acquisition dates for a series of capital assets, but the acquisition dates span too large a period to allow for feasible Time dimension modeling. See also typed measure.
Default Currency Units Define the unit scale of data. For example, if you select to define your analysis in Thousands, and enter 10, this is interpreted as 10,000. dense dimension In block storage databases, a dimension
that you see in a Summary chart. Detail charts appear in the Investigate Section in columns below the Summary charts. If the Summary chart shows a Pie chart, then the Detail charts below represent each piece of the pie.
dimension A data category used to organize business data for retrieval and preservation of values. Dimensions usually contain hierarchies of related members grouped within them. For example, a Year dimension often includes members for each time period, such as quarters and months. dimension build The process of adding dimensions and
likely to contain data for every combination of dimension members. For example, time dimensions are often dense because they can contain all combinations of all members. Contrast with sparse dimension.
dependent entity An entity that is owned by another entity in
the organization.
derived text measure In Essbase Studio, a text measure whose
values are governed by a predefined rule expressed as a range. For example, a derived text measure, called "Sales Performance Index," based on a measure Sales, could consist of the values "High," "Medium," and "Low." This derived text measure is defined to display "High," "Medium," and "Low," depending on the range in which the corresponding sales values fall. See also text measure.
descendant Any member below a parent in the database outline. In a dimension that includes years, quarters, and months, the members Qtr2 and April are descendants of the member Year. Design Report An interface in Web Analysis Studio for
(and the order in which they are represented) in a set. For example, the following set consists of two tuples of the same dimensionality because they both reflect the dimensions (Region, Year): { (West, Feb), (East, Mar) }
direct rate A currency rate that you enter in the exchange
rate table. The direct rate is used for currency conversion. For example, to convert balances from JPY to USD, In the exchange rate table, enter a rate for the period/scenario where the source currency is JPY and the destination currency is USD.
dirty block A data block containing cells that have been
converted. You enter exchange rates and convert from the source currency to the destination currency. For example, when you convert from EUR to USD, the destination currency is USD.
changed since the last calculation. Upper level blocks are marked as dirty if their child blocks are dirty (that is, they have been updated).
display type One of three Web Analysis formats saved to the repository: spreadsheet, chart, and pinboard.
Glossary 211
that is based on user-defined criteria. The list is refreshed automatically whenever it is referenced in the application. As dimension members are added and deleted, the list automatically reapplies the criteria to reflect the changes.
dynamic reference A pointer in the rules file to header records
in a data source.
dynamic report A report containing data that is updated when
mathematical relationship between the sources that utilize the driver, and the destinations to which those sources allocate cost or revenue.
duplicate alias name A name that occurs more than once in an alias table and that can be associated with more than one member in a database outline. Duplicate alias names can be used with duplicate member outlines only. duplicate member name The multiple occurrence of a member
consolidated file.
elimination The process of zeroing out (eliminating) transactions between entities within an organization. employee A user responsible for, or associated with, specific business objects. Employees need not work for an organization; for example, they can be consultants. Employees must be associated with user accounts for authorization purposes. encoding A method for mapping bit combinations to
name in a database, with each occurrence representing a different member. For example, a database has two members named New York. One member represents New York state and the other member represents New York city.
duplicate member outline A database outline containing
outline that Essbase calculates only upon the first retrieval of the value. Essbase then stores the calculated value in the database. Subsequent retrievals do not require calculating.
Dynamic Calc members A member in a block storage outline
characters for creating, storing, and displaying text. Each encoding has a name; for example, UTF-8. Within an encoding, each character maps to a specific bit combination; for example, in UTF-8, uppercase A maps to HEX41. See also code page and locale.
ending period A period enabling you to adjust the date range in a chart. For example, an ending period of month, produces a chart showing information through the end of the current month. Enterprise View An Administration Services feature that
that Essbase calculates only at retrieval time. Essbase discards calculated values after completing the retrieval request.
dynamic calculation In Essbase, a calculation that occurs only when you retrieve data on a member that is tagged as Dynamic Calc or Dynamic Calc and Store. The member's values are calculated at retrieval time instead of being precalculated during batch calculation. dynamic hierarchy In aggregate storage database outlines
enables management of the Essbase environment from a graphical tree view. From Enterprise View, you can operate directly on Essbase artifacts.
entity A dimension representing organizational units.
Examples: divisions, subsidiaries, plants, regions, products, or other financial reporting units.
212 Glossary
variance between its return and the market return, indicated by an index called beta. For example, if a stock's return normally moves up or down 1.2% when the market moves up or down 1%, the stock has a beta of 1.2.
essbase.cfg An optional configuration file for Essbase. Administrators may edit this file to customize Essbase Server functionality. Some configuration settings may also be used with Essbase clients to override Essbase Server settings. EssCell A function entered into a cell in Essbase Spreadsheet
handles the selection, orientation, grouping, and ordering of raw data extracted from a database; begins with the less than (<) character.
fact table The central table in a star join schema,
characterized by a foreign key and elements drawn from a dimension table. This table typically contains numeric data that can be related to all other tables in the schema.
Favorites gadget Contains links to Reporting and Analysis documents and URLs. field An item in a data source file to be loaded into an Essbase database. file delimiter Characters, such as commas or tabs, that
can define formatting indicators or notify subscribing users when exceptions are generated.
exchange rate A numeric value for converting one currency to another. For example, to convert 1 USD into EUR, the exchange rate of 0.8936 is multiplied with the U.S. dollar. The European euro equivalent of $1 is 0.8936. exchange rate type An identifier for an exchange rate.
criteria; for example, to exclude certain tables, metadata, or values, or to control access.
flow account An unsigned account that stores periodic and
year-to-date values.
folder A file containing other files for the purpose of
structuring a hierarchy.
footer Text or images at the bottom of report pages, containing dynamic functions or static text such as page numbers, dates, logos, titles or file names, and author names. format Visual characteristics of documents or report
Different rate types are used because there may be multiple rates for a period and year. Users traditionally define rates at period end for the average rate of the period and for the end of the period. Additional rate types are historical rates, budget rates, forecast rates, and so on. A rate type applies to one point in time.
expense account An account that stores periodic and year-
objects.
format string In Essbase, a method for transforming the way
of tags used to assign attributes to data that can be interpreted between applications according to a schema.
external authentication Logging on to Oracle's Hyperion applications with user information stored outside the applications, typically in a corporate directory such as MSAD or NTLM.
Glossary 213
installed.
host properties Properties pertaining to a host, or if the host
provide easy viewing of EPM content and enable access to core Reporting and Analysis functionality.
genealogy data Additional data that is optionally generated after allocation calculations. This data enables reporting on all cost or revenue flows from start to finish through all allocation steps. generation A layer in a hierarchical tree structure that defines
has multiple Install_Homes, to an Install_Home. The host properties are configured from the CMC.
Hybrid Analysis An analysis mapping low-level data stored in a relational database to summary-level data stored in Essbase, combining the mass scalability of relational systems with multidimensional data. hyperlink A link to a file, Web page, or an intranet HTML
page.
Hypertext Markup Language (HTML) A programming language
member relationships in a database. Generations are ordered incrementally from the top member of the dimension (generation 1) down to the child members. Use the unique generation name to identify a layer in the hierarchical tree structure.
generic jobs Non-SQR Production Reporting or non-
items.
IMPACTED status Indicates changes in child entities
that is effective until replaced by another global command or the file ends.
grid POV A means for specifying dimension members on a
grid without placing dimensions in rows, columns, or page intersections. A report designer can set POV values at the grid level, preventing user POVs from affecting the grid. If a dimension has one grid value, you put the dimension into the grid POV instead of the row, column, or page.
group A container for assigning similar access permissions
to multiple users.
GUI Graphical user interface head up display A mode that shows your loaded Smart Space
ZoomChart details may be highlighted, indicating value status: red (bad), yellow (warning), or green (good).
Historical Average An average for an account over a number
of historical periods.
combinations to retrieve data in block storage databases. (2) The index file.
214 Glossary
index cache A buffer containing index pages. index entry A pointer to an intersection of sparse dimensions.
Index entries point to data blocks on disk and use offsets to locate cells.
index file An Essbase file storing block storage data retrieval
to data blocks.
input data Data loaded from a source rather than calculated. Install_Home A variable for the directory where EPM System
hierarchies based on the inherent relationships in the database. Contrast with scraping.
Investigation See drill-through. isolation level An Essbase Kernel setting that determines the lock and commit behavior of database operations. Choices are: committed access and uncommitted access. iteration A pass of the budget or planning cycle in which
products are installed. Refers to one instance of an EPM System product when multiple applications are installed on the same computer.
integration Process that is run to move data between EPM
System products using Shared Services. Data integration definitions specify the data moving between a source application and a destination application, and enable the data movements to be grouped, ordered, and scheduled.
intelligent calculation A calculation method tracking updated data blocks since the last calculation. Interactive Reporting connection file (.oce) Files encapsulating
database connection information, including: the database API (ODBC, SQL*Net, etc.), database software, the database server network address, and database user name. Administrators create and publish Interactive Reporting connection files (.oce).
intercompany elimination See elimination. intercompany matching The process of comparing balances
for pairs of intercompany accounts within an application. Intercompany receivables are compared to intercompany payables for matches. Matching accounts are used to eliminate intercompany transactions from an organization's consolidated totals.
intercompany matching report A report that compares
based on common content in a column or row. A join typically occurs between identical or similar items within different tables or topics. For example, a record in the Customer table is joined to a record in the Orders table because the Customer ID value is the same in each table.
journal entry (JE) A set of debit/credit adjustments to account
intercompany account balances and indicates if the accounts are in, or out, of balance.
interdimensional irrelevance A situation in which a dimension
and provides access to Smart Space Collaborator. For example, you can have a KeyContacts gadget for your marketing team and another for your development team.
latest A Spreadsheet key word used to extract data values
does not intersect with other dimensions. Because the data in the dimension cannot be accessed from the nonintersecting dimensions, the non-intersecting dimensions are not relevant to that dimension.
Glossary 215
hierarchical structure, specified by generation (top down) or level (bottom up). (2) Position of objects relative to other objects. For example, in the Sample Basic database, Qtr1 and Qtr4 are in the same layer, so they are also in the same generation, but in a database with a ragged hierarchy, Qtr1 and Qtr4 might not be in same layer, though they are in the same generation.
layout area Used to designate an area on a Workspace Page
reference folders, files, shortcuts, and other links. (2) In a task flow, the point where the activity in one stage ends and another begins.
link condition A logical expression evaluated by the taskflow
in a repository.
linked partition A shared partition that enables you to use a
categories of a dimension.
level A layer in a hierarchical tree structure that defines database member relationships. Levels are ordered from the bottom dimension member (level 0) up to the parent members. level 0 block A data block for combinations of sparse, level 0
data cell to link two databases. When a user clicks a linked cell in a worksheet, Essbase opens a new sheet displaying the dimensions in the linked database. The user can then drill down those dimensions.
linked reporting object (LRO) A cell-based link to an external
members.
level 0 member A member that has no children. liability account An account type that stores point in time
file such as cell notes, URLs, or files with text, audio, video, or pictures. (Only cell notes are supported for Essbase LROs in Financial Reporting.) Contrast with local report object.
local currency An input currency type. When an input
balances of a company's liabilities. Examples of liability accounts include accrued expenses, accounts payable, and long term debt.
life cycle management The process of managing application
currency type is not specified, the local currency matches the entity's base currency.
local report object A report object that is not linked to a
Financial Reporting report object in Explorer. Contrast with linked reporting object (LRO).
local results A data model's query results. Results can be used
in local joins by dragging them into the data model. Local results are displayed in the catalog when requested.
locale A computer setting that specifies a location's language, currency and date formatting, data sort order, and the character set encoding used on the computer. Essbase uses only the encoding portion. See also encoding and ESSLANG. locale header record A text record at the beginning of some
represented by a line. A line chart can display each line stacked on the preceding ones, as represented by an absolute value or a percent.
line item detail The lowest level of detail in an account. lineage The relationship between different metadata elements showing how one metadata element is derived from one or more other metadata elements, ultimately tracing the metadata element to its physical source. In Essbase Studio, a lineage viewer displays the relationships graphically. See also traceability.
non-Unicode-encoded text files, such as scripts, that identifies the encoding locale.
location alias A descriptor that identifies a data source. The
location alias specifies a server, application, database, user name, and password. Location aliases are set by DBAs at the database level using Administration Services Console, ESSCMD, or the API.
locked A user-invoked process that prevents users and
216 Glossary
user.
LOCKED status A consolidation status indicating that an
calculated in formulas and outlines. Can be any of the standard mathematical or Boolean operators; for example, +, -, *, /, and %.
MaxL The multidimensional database access language for
Essbase, consisting of a data definition language (MaxL DDL) and a data manipulation language (MaxL DML). See also MaxL DDL, MaxL DML, and MaxL Shell.
MaxL DDL Data definition language used by Essbase for batch or interactive system-administration tasks. MaxL DML Data manipulation language used in Essbase for
complete.
Map File Used to store the definition for sending data to or
Essbase MaxL DDL. This module can be added to the Perl package to provide access to Essbase databases from Perl programs.
MaxL Script Editor A script-development environment in
retrieving data from an external database. Map files have different extensions (.mps to send data; .mpr to retrieve data).
Map Navigator A feature that displays your current position on a Strategy, Accountability, or Cause and Effect map, indicated by a red outline. Marginal Tax Rate Used to calculate the after-tax cost of debt.
Administration Services Console. MaxL Script Editor is an alternative to using a text editor and the MaxL Shell for administering Essbase with MaxL scripts.
MaxL Shell An interface for passing MaxL statements to
Essbase Server. The MaxL Shell executable file is located in the Essbase bin directory (UNIX: essmsh, Windows: essmsh.exe).
MDX (multidimensional expression) The language that give instructions to OLE DB for OLAP- compliant databases, as SQL is used for relational databases. When you build the OLAPQuery section's Outliner, Interactive Reporting Clients translate requests into MDX instructions. When you process the query, MDX is sent to the database server, which returns records that answer your query. See also SQL spreadsheet. measures Numeric values in an OLAP database cube that are
Represents the tax rate applied to the last earned income dollar (the rate from the highest tax bracket into which income falls) and includes federal, state and local taxes. Based on current level of taxable income and tax bracket, you can predict marginal tax rate.
Market Risk Premium The additional rate of return paid over
the risk-free rate to persuade investors to hold riskier investments than government securities. Calculated by subtracting the risk-free rate from the expected market return. These figures should closely model future market conditions.
master data model An independent data model that is referenced as a source by multiple queries. When used, Locked Data Model is displayed in the Query section's Content pane; the data model is linked to the master data model displayed in the Data Model section, which an administrator may hide.
available for analysis. Measures are margin, cost of goods sold, unit sales, budget amount, and so on. See also fact table.
member A discrete component within a dimension. A
member identifies and differentiates the organization of similar units. For example, a time dimension might include such members as Jan, Feb, and Qtr1.
Glossary 217
command that selects member ranges based on outline relationships, such as sibling, generation, and level.
member-specific report command A type of Report Writer formatting command that is executed as it is encountered in a report script. The command affects only its associated member and executes the format command before processing the member. merge A data load option that clears values only from the accounts specified in the data load file and replaces them with values in the data load file. metadata A set of data that defines and describes the
attribute that describes the data format of an item, so that the system knows which application should open the object. A file's mime type is determined by the file extension or HTTP header. Plug-ins tell browsers what mime types they support and what file extensions correspond to each mime type.
mining attribute In data mining, a class of values used as a
properties and attributes of the data stored in a database or used by an application. Examples of metadata are dimension names, member names, properties, time periods, and security.
metadata elements Metadata derived from data sources and
content, hyperlinks, and the query or queries to load the report. Each report can include one or more minireports.
minischema A graphical representation of a subset of tables
other metadata that is stored and cataloged for Essbase Studio use.
metadata sampling The process of retrieving a sample of
labeled location does not exist, contains no value, or was never entered or loaded. For example, missing data exists when an account contains data for a previous or future period but not for the current period.
model (1) In data mining, a collection of an algorithm's
the structure and rules for creating an Essbase outline from an OLAP model.
metric A numeric measurement computed from business
findings about examined data. A model can be applied against a wider data set to generate useful information about that data. (2) A file or content string containing an application-specific representation of data. Models are the basic data managed by Shared Services, of two major types: dimensional and non-dimensional application objects. (3) In Business Modeling, a network of boxes connected to represent and calculate the operational and financial flow through the area being examined.
monetary A money-related value. multidimensional database A method of organizing, storing,
and referencing data through three or more dimensions. An individual value is the intersection point for a set of dimensions. Contrast with relational database.
218 Glossary
multiple sources including documents, URL, and other content types. Enables a user to aggregate content from Oracle and non-Oracle sources.
named set In MaxL DML, a set with its logic defined in the
method used from any application regardless of how the database management system (DBMS) processes the information.
OK status A consolidation status indicating that an entity has
already been consolidated, and that data has not changed below it in the organization structure.
OLAP Metadata Catalog In Integration Services, a relational
optional WITH section of a MaxL DML query. The named set can be referenced multiple times in the query.
native authentication The process of authenticating a user
database containing metadata describing the nature, source, location, and type of data that is pulled from the relational data source.
OLAP model In Integration Services, a logical model (star
displays data from multiple dimensions. For example, a column heading that contains Year and Scenario members is a nested column. The nested column heading shows Q1 (from the Year dimension) in the top line of the heading, qualified by Actual and Budget (from the Scenario dimension) in the bottom line of the heading.
NO DATA status A consolidation status indicating that this
schema) that is created from tables and columns in a relational database. The OLAP model is then used to generate the structure of a multidimensional database.
online analytical processing (OLAP) A multidimensional, multiuser, client-server computing environment for users who analyze consolidated enterprise data in real time. OLAP systems feature drill-down, data pivoting, complex calculations, trend analysis, and modeling. Open Database Connectivity (ODBC) Standardized application
includes application objects such as security files, member lists, calculation scripts, and Web forms.
non-unique member name See duplicate member name. note Additional information associated with a box,
programming interface (API) technology that allows applications to access multiple third-party databases.
organization An entity hierarchy that defines each entity and
database, including all dimensions, members, tags, types, consolidations, and mathematical relationships. Data is stored in the database according to the structure defined in the outline.
outline synchronization For partitioned databases, the process of propagating outline changes from one database to another database. P&L accounts (P&L) Profit and loss accounts. Refers to a typical grouping of expense and income accounts that comprise a company's income statement. page A display of information in a grid or table often
represented by the Z-axis. A page can contain data from one field, derived data from a calculation, or text.
page file Essbase data file.
Glossary 219
represented on the current page of the report. All data values on the page have the members in the page heading as a common attribute.
page member A member that determines the page axis. palette A JASC compliant file with a .PAL extension. Each
palette contains 16 colors that complement each other and can be used to set the dashboard color elements.
parallel calculation A calculation option. Essbase divides a calculation into tasks and calculates some tasks simultaneously. parallel data load In Essbase, the concurrent execution of
files. This may be faster than exporting to a single file, and it may resolve problems caused by a single data file becoming too large for the operating system to handle.
parent adjustments The journal entries that are posted to a
Essbase operation or setting. For example, an Essbase administrator may limit the persistence of user name and password validity.
personal pages A personal window to repository information. You select what information to display and its layout and colors. personal recurring time events Reusable time events that are accessible only to the user who created them. personal variable A named selection statement of complex member selections. perspective A category used to group measures on a
entities that report directly to them. Because parents are both entities and associated with at least one node, they have entity, node, and parent information associated with them.
partition area A sub cube within a database. A partition is composed of one or more areas of cells from a portion of the database. For replicated and transparent partitions, the number of cells within an area must be the same for the data source and target to ensure that the two partitions have the same shape. If the data source area contains 18 cells, the data target area must also contain 18 cells to accommodate the number of values. partitioning The process of defining areas of data that are shared or linked between data models. Partitioning can affect the performance and scalability of Essbase applications. pattern matching The ability to match a value with any or all
scorecard or strategic objectives within an application. A perspective can represent a key stakeholder (such as a customer, employee, or shareholder/financial) or a key competency area (such as time, cost, or quality).
pie chart A chart that shows one data set segmented in a pie formation. pinboard One of the three data object display types.
characters of an item entered as a criterion. Missing characters may be represented by wild card values such as a question mark (?) or an asterisk (*). For example, Find all instances of apple returns apple, but Find all instances of apple* returns apple, applesauce, applecranberry, and so on.
Pinboards are graphics, composed of backgrounds and interactive icons called pins. Pinboards require traffic lighting definitions.
220 Glossary
pinboards. Pins are dynamic. They can change images and traffic lighting color based on the underlying data values and analysis tools criteria.
pivot The ability to alter the perspective of retrieved data.
When Essbase first retrieves a dimension, it expands data into rows. You can then pivot or rearrange the data to obtain a different viewpoint.
planner Planners, who comprise the majority of users, can
input and submit data, use reports that others create, execute business rules, use task lists, enable e-mail notification for themselves, and use Smart View.
planning unit A data slice at the intersection of a scenario, version, and entity; the basic unit for preparing, reviewing, annotating, and approving plan data. plot area The area bounded by X, Y, and Z axes; for pie
together in an implementation. For example, a Planning project may consist of a Planning application, an Essbase cube, and a Financial Reporting server instance.
property A characteristic of an artifact, such as size, type, or
processing instructions.
provisioning The process of granting users and groups
of balance differences between intercompany account pairs during the elimination process.
post stage assignment Assignments in the allocation model
retrieval.
precision Number of decimal places displayed in numbers. predefined drill paths Paths used to drill to the next level of
differentiates duplicate member names in a duplicate member outline. For example, [Market].[East].[State]. [New York] or [Market].[East].[City].[New York]
query Information requests from data providers. For example, used to access relational data sources. query governor An Essbase Integration server parameter or Essbase server configuration setting that controls the duration and size of queries made to data sources.
reports to be grouped, organized, ordered, distributed, and reviewed. Includes pointers referencing reports in the repository.
preserve formulas User-created formulas kept within a
Glossary 221
properties defining behavior or appearance, such as text boxes, grids, images, and charts.
report script A text file containing Essbase Report Writer commands that generate one or more production reports. Report Viewer An Essbase component that displays complete
information after a database has been damaged or destroyed, typically performed after shutting down and restarting the database.
restructure An operation to regenerate or rebuild the
representing the process unit level, such as Not Started, First Pass, Submitted, Approved, and Published.
Risk Free Rate The rate of return expected from safer
all accounts for periods specified in the data load file, and loads values from the data load file. If an account is not specified in the load file, its values for the specified periods are cleared.
replicated partition A portion of a database, defined through
Partition Manager, used to propagate an update to data mastered at one site to a copy of data stored at another site. Users can access the data as though it were part of their local database.
Report Extractor An Essbase component that retrieves report
Service Configurator, including Repository Service, Service Broker, Name Service, Event Service, and Job Service.
runtime prompt A variable that users enter or select before a business rule is run. sampling The process of selecting a representative portion
data from the Essbase database when report scripts are run.
222 Glossary
drive key business calculations (for example, the cost per square foot of office floor space).
scaling Scaling determines the display of values in whole
installed servlets.
shared member A member that shares storage space with another member of the same name, preventing duplicate calculation of members that occur multiple times in an Essbase outline. Shared Services Registry Part of the Shared Services database,
operation or setting; for example, the area of data affected by a security setting. Most commonly, scope refers to three levels of granularity, where higher levels encompass lower levels. From highest to lowest, these levels are as follows: the entire system (Essbase Server), applications on Essbase servers, or databases within Essbase server applications. See also persistence.
score The level at which targets are achieved, usually expressed as a percentage of the target. scorecard Business object that represents the progress of an employee, strategy element, or accountability element toward goals. Scorecards ascertain this progress based on data collected for each measure and child scorecard added to the scorecard. scraping An inspection of a data source to derive the most
the Shared Services Registry stores and reuses information for most installed EPM System products, including installation directories, database settings, deployment settings, computer names, ports, servers, URLs, and dependent service data.
Shared Workspace Page Workspace Pages shared across an
organization which are stored in a special System folder and can be accessed by authorized users from the Shared Workspace Pages Navigate menu.
sibling A child member at the same generation as another
child member and having the same immediate parent. For example, the members Florida and New York are children of East and each other's siblings.
single sign-on Ability to access multiple EPM System
repository. The Search gadget looks for a match in the document keywords and description, which are set when you import a document.
secondary measure A low-priority measure, less important than primary measures. Secondary measures do not have Performance reports but can be used on scorecards and to create dimension measure templates. security agent A Web access management provider (for example, Netegrity SiteMinder) that protects corporate Web resources. security platform A framework enabling EPM System products to use external authentication and single sign-on. serial calculation The default calculation setting. Divides a
systems to send messages and share Reporting and Analysis repository content. The message can take many forms, including instant message style discussions, meetings, and toast messages.
Glossary 223
are associated with predefined actions available from the Smart Tag menu. In EPM System products, smart tags can also be used to import Reporting and Analysis content, and access Financial Management and Essbase functions.
SmartBook gadget Contains documents from the Reporting and Analysis repository or URLs. All documents are loaded when the SmartBook is opened so you can access all content immediately. SmartCut A link to a repository item, in URL form. snapshot Read-only data from a specific time. source currency The currency from which values originate
adjustments that have common adjustment information for each period. For example, you can create a standard template that contains the common account IDs, entity IDs, or amounts, then use the template as the basis for many regular journals.
Status bar The status bar at the bottom of the screen displays
helpful information about commands, accounts, and the current status of your data file.
stored hierarchy In aggregate storage databases outlines only.
A hierarchy in which the members are aggregated according to the outline structure. Stored hierarchy members have certain restrictions, for example, they cannot contain formulas.
strategic objective (SO) A long-term goal defined by measurable results. Each strategic objective is associated with one perspective in the application, has one parent, the entity, and is a parent to critical success factors or other strategic objectives. Strategy map Represents how the organization implements high-level mission and vision statements into lower-level, constituent strategic goals and objectives. structure view Displays a topic as a simple list of component
unlikely to contain data for all member combinations when compared to other dimensions. For example, not all customers have data for all products. Contrast with dense dimension.
SPF files Printer-independent files created by an SQR
Production Reporting server, containing a representation of the actual formatted report output, including fonts, spacing, headers, footers, and so on.
Spotlighter A tool that enables color coding based on selected
data items.
Structured Query Language A language used to process instructions to relational databases. Subaccount Numbering A system for numbering subaccounts using non-sequential, whole numbers. subscribe Flags an item or folder to receive automatic
conditions.
SQL spreadsheet A data object that displays the result set of
a SQL query.
SQR Production Reporting A specialized programming language for data access, data manipulation, and creating SQR Production Reporting documents. stage A task description that forms one logical step within
charts shown below in the same column, plotting metrics at the summary level at the top of each chart column.
super service A special service used by the
a specific application. A staging area is a snapshot or restructured version of one or more RDBMSs.
standard dimension A dimension that is not an attribute
dimension.
224 Glossary
data consistently.
text list In Essbase, an object that stores text values mapped
enables multiprocessing and multithreading. Performance is not significantly degraded when a large number of users connect to an single instance simultaneously.
sync Synchronizes Shared Services and application models. synchronized The condition that exists when the latest
version of a model resides in both the application and in Shared Services. See also model.
system extract Transfers data from an application's metadata
expressed as text. In Essbase, a member tagged as Text in the dimension where measures are represented. The cell values are displayed as predefined text. For example, the text measure "Satisfaction Index" may have the values Low, Medium, and High. See also typed measure, text list, derived text measure.
time dimension Defines the time period that the data
Finance.
target Expected results of a measure for a specified period of time (day, quarter, and so on). task list A detailed status list of tasks for a particular user. taskflow The automation of a business process in which tasks are passed from one taskflow participant to another according to procedural rules. taskflow definition Represents business processes in the
dates and times of completed process flow steps for specific locations.
time scale Displays metrics by a specific period in time, such as monthly or quarterly. time series reporting A process for reporting data based on a calendar date (for example, year, quarter, month, or week). Title bar Displays the Strategic Finance name, the file name,
taskflow management system. Consists of a network of stages and their relationships; criteria indicating the start and end of the taskflow; and information about individual stages, such as participants, associated applications, associated activities, and so on.
taskflow instance Represents a single instance of a taskflow including its state and associated data. taskflow management system Defines, creates, and manages the execution of a taskflow including: definitions, user or application interactions, and application executables. taskflow participant The resource who performs the task associated with the taskflow stage instance for both manual and automated stages. Taxes - Initial Balances Strategic Finance assumes that the
Initial Loss Balance, Initial Gain Balance and the Initial Balance of Taxes Paid entries have taken place in the period before the first Strategic Finance time period.
TCP/IP See Transmission Control Protocol/Internet Protocol
in a dimension outline hierarchy, or the first member of the dimension in sort order if there is no hierarchical relationship among dimension members. The top-level member name is generally the same as the dimension name if a hierarchical relationship exists.
trace allocations A feature of Profitability and Cost
(TCP/IP).
Management that enables you to visually follow the flow of financial data, either forwards or backwards, from a single intersection throughout the model.
Glossary 225
trace level Defines the level of detail captured in the log file. traceability The ability to track a metadata element to its
physical source. For example, in Essbase Studio, a cube schema can be traced from its hierarchies and measure hierarchies, to its dimension elements, date/time elements, and measures, and ultimately, to its physical source elements.
traffic lighting Color-coding of report cells, or pins based on
intersection of a member from each dimension. If a dimension is omitted, its top member is implied. Examples: (Jan); (Jan, Sales); ( [Jan], [Sales], [Cola], [Texas], [Actual] )
two-pass An Essbase property that is used to recalculate
members that are dependent on the calculated values of other members. Two-pass members are calculated during a second pass through the outline.
typed measure In Essbase, a member tagged as Text or Date in the dimension where measures are represented. The cell values are displayed as predefined text or dates. unary operator A mathematical indicator (+, -, *, /, %)
properly in the destination environment after application migration. (2) In data mining, modifies data (bidirectionally) flowing between the cells in the cube and the algorithm.
translation See currency conversion. Transmission Control Protocol/Internet Protocol (TCP/IP) A
associated with an outline member. The unary operator defines how the member is calculated during a database rollup.
Unicode-mode application An Essbase application wherein character text is encoded in UTF-8, enabling users with computers set up for different languages to share application data. Uniform Resource Locator The address of a resource on the
standard set of communication protocols linking computers with different operating systems and internal architectures. TCP/IP utilities are used to exchange files, send mail, and store data to various computers that are connected to local and wide area networks.
transparent login Logs in authenticated users without
Internet or an intranet.
unique member name A non-shared member name that exists
access and change data in a remote database as though it is part of a local database
triangulation A means of converting balances from one
currency to another via a third common currency. In Europe, this is the euro for member countries. For example, to convert from French franc to Italian lira, the common currency is defined as European euro. Therefore, in order to convert balances from French franc to Italian lira, balances are converted from French franc to European euro and from European euro to Italian lira.
triggers An Essbase feature whereby data is monitored
according to user-specified criteria which when met cause Essbase to alert the user or system administrator.
trusted password A password that enables users
authenticated for one product to access other products without reentering their passwords.
trusted user Authenticated user.
associated with members of an outline to describe a characteristic of the members. Users can use UDAs to return lists of members that have the specified UDA associated with them.
226 Glossary
script, or partition definition against the outline to make sure that the object being checked is valid. For example, in FDM, validation rules ensure that certain conditions are met after data is loaded from FDM to the target application.
value dimension Used to define input value, translated value,
used for creating re-usable logic to define sets or members. Sets or custom members can be defined once in the WITH section, and then referenced multiple times during a query.
work flow The steps required to process data from start to finish in FDM. The workflow consists of Import (loading data from the GL file), Validate (ensures all members are mapped to a valid account), Export (loads the mapped members to the target application), and Check (verifies accuracy of data by processing data with user-defined validation rules). workbook An entire spreadsheet file with many worksheets. Workspace Page A page created with content from multiple
one or more dimensions. It can be used to track a value in relation to these dimensions; for example, the varying attribute Sales Representative, associated with the Product dimension, can be used to track the value Customer Sales of several different sales representatives in relation to the Time dimension. Varying attributes can also be used for member selection, such as finding the Products that a Sales Representative was responsible for in May.
version Possible outcome used within the context of a
sources including documents, URL, and other content types. Enables a user to aggregate content from Oracle and non-Oracle sources.
write-back The ability for a retrieval client, such as a
scenario of data. For example, Budget - Best Case and Budget - Worst Case where Budget is scenario and Best Case and Worst Case are versions.
view Representation of either a year-to-date or periodic
display of data.
visual cue A formatted style, such as a font or a color, that highlights specific types of data values. Data values may be dimension members; parent, child, or shared members; dynamic calculations; members containing a formula; read only data cells; read and write data cells; or linked objects. Web server Software or hardware hosting intranet or
indicates the relative importance of that item in the calculation of the overall scorecard score. The weighting of all items on a scorecard accumulates to 100%. For example, to recognize the importance of developing new features for a product, the measure for New Features Coded on a developer's scorecard would be assigned a higher weighting than a measure for Number of Minor Defect Fixes.
magnify a report to fit whole page, page width, or percentage of magnification based on 100%.
Glossary 227
a chart. Enables you to see detailed numeric information on the metric that is displayed in the chart.
228 Glossary
Index
Symbols
& (ampersand), 14 /eis/server/samples/tbcdbase, 117
A
access codes, for metaoutlines, 127 codes, for OLAP models, 126 exclusive access mode, 122 multi-user, 122 privileges, common problems, 130 problems, correcting, 144 setting, 125 standard access mode, 122 accessing data sources, 129 OLAP Metadata Catalog, 129, 131 servers and databases, troubleshooting, 129 accessing, servers and databases, 66 accounts problems in the RDBMS, 134 user, administrative, 134 user, problems, 145 ADDDUPLICATESFORASO (ais.cfg parameter), 29 adding data source to odbc.ini, 48, 56, 65 administration tools, Unicode-enabled, 35 administrative accounts, using, 134 aggregate storage, 194 ais.cfg IBM DB2 OS/390 configuration parameter (for UNIX), 64 IBM DB2 OS/390 configuration parameter (for Windows), 48 ais.cfg file, 12, 13 described, 26 JDBC data source mapping, 72, 73 storing startup switch settings, 26
ais.cfg file parameters, 29 ADDDUPLICATESFORASO, 29 UNIFORMVALUEFORMAT, 29 aliases, 37 ampersand, 14 Anaytic Services, testing connection from Integration Server to Essbase Server, 142 application, non-Unicode mode, 34 applications, Unicode-mode, defined, 34 artifact limits for source databases, 192 artifact limits in Integration Services, 189 AT service, 172, 173
B
backtracking connection problems, 132 backups, OLAP Metadata Catalog, 87 bat extension files, 157 batch files for loading TBC database, 100 Integration Services Shell and, 154 running, 158 running Integration Services Shell scripts with, 157 scheduling manually, 172, 173 batch mode, Integration Services Shell, 155 BCP utility, 95 binding, requirement for IBM DB2 databases, 147 block storage, 194 brackets, in Integration Services Shell syntax, 156 building OLAP Metadata Catalog, 77 sample application, 93
C
CB_CONCURRENCY table column descriptions, 123 defined, 92
Index 229
CB_OBJECT_TYPE column, data defined, 123 CB_VERSION table, 92 cbs extension files, 157 checking common connection problems, 129 release number of Integration Services, 140 script syntax for loads, 157 client programs, custom, 35 client software port number, 149 problems, correcting, 150 RDBMS, common problems, 130 RDBMS, correcting problems, 146 RDBMS, required, 42 troubleshooting, 129 code page, defined, 33 codes, access for metaoutlines, 127 for OLAP models, 126 columns, viewing in sample application, 118 Command Center, 83 command line Integration Services Shell, 154 loading data from, 173 loading members from, 172 command syntax, Integration Services Shell, 156 Command Window, IBM DB2, 95 command-line interface, using, 153 commands, Integration Services Shell, 159 commands, status of, 160 committing records during data load, 16 common problems areas of failure, 131 database-specific, 131 with connections, 129 communication problems, 131 compatibility between RDBMS and RDBMS client, 147 ODBC drivers, 148 problems with product versions, 140 components, software, common problems, 130, 131 compression dimension, 193 concurrent users, 123 configuration parameters, 29 ADDDUPLICATESFORASO, 29 automatic validation, 28 display primary keys, 28
UNIFORMVALUEFORMAT, 29 configuration problems (ODBC), 131 configuring data sources on UNIX, 54 data sources on Windows, 43 ODBC (Open Database Connectivity) data sources, 41 ODBC computer location, 42 ODBC data sources on UNIX, 54 ODBC data sources on Windows, 43 odbc.ini file, 54, 55 connecting database client to RDBMS, 134 problems with data source, 129, 131 problems with OLAP Metadata Catalog, 129, 131 setting defaults, 68 to Essbase Server, 67 to Integration Server, 66 to OLAP Metadata Catalog, 66 to relational data sources, 66, 69 to servers and databases, 66 connections areas of failure, 131 common problems, 129 isolating problems, 132 LOADMEMBER command, 139 OLAP Metadata Catalog, 129, 131 path through components, 133 problems, correcting, 144 problems, diagnosing, 131 problems, isolating, 132 testing, 135, 141 troubleshooting, 129 Control Panel dialog box, 38 correcting connection problems, 129, 144 data source problems, 148 database client problems, 146 Integration Server problems, 149 Integration Services Console problems, 150 ODBC problems, 129, 148 password problems, 145 RDBMS problems, 146 user name problems, 145 creating database for TBC, 98, 116 database for TBC_MD, 101
230 Index
ODBC server software data sources, 41 OLAP Metadata Catalog automatically, 79 OLAP Metadata Catalog manually, 80, 84 OLAP Metadata Catalog tables, 78 OLAP Metadata Catalogs, 77, 78 sample application, 93 tables for TBC database, 99 TBC database, 98 critical server error messages, 20 cron scheduling daemon, 38, 172
D
data loads checking script syntax, 157 command line and, 173 enabling scheduled loads, 38 error messages, 183 incremental, script syntax, 159 Integration Services Shell and, 172, 173 LOADALL command, 168 LOADDATA command, 166 OLAP Metadata Catalog and, 11, 87 scripts, recording, 157 time required, 156 with TBC sample application tables, 100 data migration OLAP Metadata Catalog, 81 data sources accessing, 129 adding to odbc.ini (UNIX), 48, 56, 65 configuring, 41 configuring on UNIX systems, 54 configuring on Windows systems, 43 connecting to, 66, 69, 173 connection problems, 129, 131 connections to, diagram, 133 identifying before loading, 161 problems, correcting, 148 setting up TBC, 98 data warehouse, 121 data, retrieving, 11 database (source), artifact limits, 192 database client software environment problems, 134, 147 information, 42 problems, correcting, 146 testing connection to RDBMS, 134
database client software, common problems, 130 database scripts for OLAP Metadata Catalog, 82 for TBC database, 99 for TBC_MD database, 102 database servers common problems, 130 problems, correcting, 146 database users, source, 121 databases configuring for ODBC, 41 connecting to, 66 connections to, diagram, 133 OLAP Metadata Catalog, requirements, 78 TBC, creating, 98, 116 TBC, requirements, 98, 101, 116 TBC_MD, creating, 101 TBC_MD, requirements, 101 user aliases and synonyms, 37 vendor-specific problems, 131 DataDirect Text driver configuring on UNIX., 5961 configuring on Windows, 4951 DataDirect, ODBC driver problems, correcting, 148 dataload.txt file, 32 DB2. See IBM DB2 DBAccess, 83, 95 debug messages, 20 defaults for connections, setting, 68 login, setting, 69 Delete Locks command, 123 deleting Integration Server users, 121 locks, 123, 124 OLAP Metadata Catalog, 86 OLAP Metadata Catalog tables, 82 read and write locks, 124 TBC database tables, 99 TBC_MD database tables, 102 users, 121 DemoODBC program, 135 diagnosing problems (connections, ODBC), 129 diagram, connections path, 133 dialog boxes Control Panel, 38 Login, 67, 105
Index 231
OLAP Metadata Catalog Setup, 79 Sample Application Setup, 96 Set Login Defaults, 69 directories sample application scripts, 97, 108 TBC data load script, 97, 107, 108, 111, 113, 116 TBC table creation script, 97, 107, 108, 111, 113 TBC_MD data load script, 97 TBC_MD table creation script, 97, 107, 109, 111, 113, 116 disconnecting, users, 121 displaying primary keys in OLAP models, 28 DISTINCT clause in drill-through SQL, 17 DNS, problems, 132 domain name server, problems, 132 drill-through reports error messages, 184 guidelines, 201 limits, 201 tables relating to, 91 driver file, ODBC, specifying on UNIX, 139 drivers, ODBC, problems, 148 dropping tables OLAP Metadata Catalog, 82 TBC database, 99 TBC_MD database, 102 DTR_ATTRIBUTES table, 91 DTR_INFO table, 91
setting logging level, 20 Essbase Essbase Server, connecting to, 67, 173 message file, 32 server name problems, 144 server, message database, 21 unsupported functions in Hybrid Analysis, 196 users, defined, 121 Essbase Integration Server running in background, 14 Essbase message.txt file, 32 Essbase Server connecting to, 67 connection defaults, setting, 68 setting connection information for, 67 Essbase Server, connecting to, 173 Essbase Server, target data server, specifying, 162 Essbase Services relationship to Integration Server, 11 essbase.mdb file, 21 examples, odbc.ini files, 5764, 58, 62, 63, 64 exclusive access mode, 122 executable name for Integration Server, 12 for Integration Services Shell, 154 EXIT command, 172 extra locks deleting, 124 viewing, 123
E
editing configuration for a data source, 54 odbc.ini file, 55 enabling scheduled loads on UNIX, 38 scheduled loads on Windows, 38 encoding, defined, 33 environment variables common problems, 130 for database client, problems, 134, 147 for ODBC on UNIX, 55 ODBC problems, 135 error messages data load, 183 drill-through reports, 184 member load, 180 miscellaneous server messages, 185
F
file name, Essbase messages file, 18 file names, 21 fixing.. See troubleshooting
G
guidelines for drill-through reports, 201 guidelines for Hybrid Analysis, 194
H
host names, problems, 144 Hybrid Analysis enabling in manual sample application setup, 93 guidelines, 194 limits on transformation rules, 195 required utilities for sample application setup, 106
232 Index
sample application for IBM DB2, 106 sample application for Microsoft SQL Server, 111 sample application for Oracle, 108 sample application for Teradata, 113 sample application setup, 106 transformations not supported, 196 unsupported Essbase functions, 196 viewing the metaoutline, 119 viewing the OLAP model, 119 Hybrid Analysis, defined, 91 Hybrid Analysis, tables relating to, 91 hyperion user name, permissions required, 38
I
IBM DB2 -tvf utility, 83, 95 Command Center utility, 83 Command Window, 95 common problems, 131 configuring the Wire Protocol driver on Windows, 45 database binding requirement, 147 odbc.ini file example (UNIX), 57 IBM DB2 OS/390 ais.cfg entry for Windows, 48 configuration parameter in ais.cfg (for UNIX), 64 configuration parameter in ais.cfg (for Windows), 48 default port number, 45 odbc.ini file example (UNIX), 58 ignore parent during drill-through, 15 incompatibility, problems with product versions, 140 incremental loads LOADALL command, 168 LOADDATA command, 166 LOADMEMBER command, 163 script syntax, 159 informational messages, 20 Informix configuring the Wire Protocol driver on Windows, 46 DBAccess utility, 83, 95 example odbc.ini file, 58 installation manual steps, UNIX, 75 installing
OLAP Metadata Catalog, 77 Integration Server common problems, 129 connecting to, 66 connection defaults, setting, 68 defined, 11 determining user access, 124 error message file, 32 error messages, 179 executable name, 12 executable name in UNIX, 12 log files, 31 logging on to, 159, 173 logging out of, 172 message file, 32 OLAP Metadata Catalog, information in, 87 olapisvr file, 12 port number, 22, 149, 154 prerequisites for, 12 problems, correcting, 149 reconnecting to, 30 server problems, 144 shutting down, 30, 172 starting, 12, 30 starting as a Windows service, 13 starting from the command line, 12 starting from the Windows Desktop, 13 startup switches, 14 startup.bat file, 12 stopping, 30 stopping as a Windows Service, 30 stopping from command line, 30 stopping from Windows Desktop, 30 switches, 14 testing connection to Essbase Server, 142 testing connection to RDBMS, 139 users, defined, 121 Integration Server users deleting, 121 disconnecting, 121 multiple, 123 viewing locks, 123 Integration Services artifact limits, 189 connection path, diagram, 133 release number, checking, 140 Integration Services Console
Index 233
common problems, 129 port number, 22, 149, 154 problems, correcting, 150 testing connection to Essbase Server, 143 testing connection to RDBMS, 141 Integration Services Shell asynchronous commands, 160 batch mode, 155 command syntax, 156 commands, list, 159 description, 154 executable name, 154 executing commands, 156 interactive mode, 155 loading data with, 173 loading members with, 172 running scripts with, 157 starting, 153 using, 153 using to test connections, 139 Integration Services Shell commands EXIT, 172 LOADALL, 168 LOADDATA, 166 LOADMEMBER, 163 LOGIN, 159 LOGOUT, 172 SETCATALOG, 162 SETSOURCE, 161 SETTARGET, 162 SHUTDOWN, 172 STATUS, 160 syntax, 156 VERSION, 160 Integration Services users creating, 121 viewing, 121 interactive mode, Integration Services Shell, 155 is.csh file, using, 138 is.sh file, using, 138 isolating, connection problems, overview, 132 ISQL utility, 95 ivtestlib program, 135
JOIN_HINTS table, 89
L
lc*.sql files, 104 ld*.sql files, 100 levels of access permissions, 126 library path, common problems, 130 limits drill-through report guidelines, 201 Hybrid Analysis guidelines, 194 Integration Services artifacts, 189 source database artifacts, 192 transformation rules with Hybrid Analysis, 195 transformations not supported by Hybrid Analysis, 196 unsupported Essbase functions in Hybrid Analysis, 196 listeners, setting, 25 LOAD commands, 155, 156 LOADALL command, described, 156, 168 LOADDATA command described, 156, 166 example of, 173 loading data, time required, 156 from the command line, 173 Integration Services Shell and, 172, 173 OLAP Metadata Catalog and, 11, 87 test, members into Essbase database, 142 loading data into TBC database, 100 into TBC_MD database using XML Import utility, 104 LOADMEMBER command connections made by, 139 described, 156, 163 example of, 173 using to test connections, 139 loads, scheduled, enabling, 38 locale, defined, 33 locks deleting, 123, 124 exclusive access mode, 122 metaoutlines, 122 OLAP models, 122 read, deleting, 124 read, viewing, 124
J
JDBC driver mapping a data source for Teradata, 72, 73
234 Index
standard access mode, 122 storage location in the OLAP Metadata Catalog, 123 timestamps and, 123 unreleased, 123 viewing, 123 working with, 121, 122 write, deleting, 124 write, viewing, 124 log files changing name and location, 18 contents, 31 Integration Server, viewing, 31 messages recording level, 20 logging level, 20 logging on, to servers and databases, 66 LOGIN command described, 159 example of, 173 login defaults, setting, 69 Login dialog box, 67, 105 login problems, correcting, 145 login script diagnosing problems, 137 problems, 134 LOGOUT command, 172
M
managing locks and permissions, 121 user names and passwords, 145 manual steps, UNIX installation, 75 MB_FILTER table, 88 MB_TX_RULE table, 88 member loads checking script syntax, 157 command line, 172, 173 enabling scheduled loads, 38 error messages, 180 LOADMEMBER command, 163 OLAP Metadata Catalog and, 11, 87 scripts, recording, 157 Merant.. See DataDirect, ODBC driver problems, correcting message file, Essbase, 32 message file, Integration Server, 32 messages, debug, 20
messages, error, 179 Metadata Catalog, creating Unicode, 80 metaoutlines access codes, 127 access permissions, 125 locks, 122 opening in exclusive access mode, 122 opening in standard access mode, 122 overwriting, 122 permissions, viewing, 126 restoring, 88 tables in OLAP Metadata Catalog, 89 users, 123 validating automatically, 28 Microsoft SQL Server example odbc.ini file on UNIX, 61 Query Analyzer utility, 83, 95 Microsoft SQL Server driver, configuring on Windows, 5153 Microsoft SQL Server, common problems, 131 migration OLAP Metadata Catalog, 81 MO_ACCESS_CODE column, data defined, 127 MO_DESCRIPTIONS table, 90 MO_FILTERS table, 89 MO_INFO table, 90, 126 MO_INTERSECTION table, 91 MO_PROPERTIES table, 90, 91, 92 MOD_ATTRIBUTES table, 90 MOD_INFO table, 90 MODEL_ACCESS_CODE column, data defined, 126 MS_INFO table, 90 MS_PREDICATES table, 90 MS_TX_RULE table, 90 MSR_INFO table, 90 MSR_PREDICATES table, 90 MSR_TX_RULE table, 90 multi-threaded server, 11 multi-user access, 122
N
names, server problems, 144 network listeners, setting, 25 network problems, 132 non-critical error messages, 20 non-Unicode-mode application, defined, 34
Index 235
O
OA_INFO table, 88 oc*_upgrd20.sql files, 80, 81 oc_create_*.sql files, 83 oc_create_*_unicode.sql files, 102, 103 oc_drop_*.sql files, 83 oc_upgrade20_*.sql files, 83 oc_upgrade61_*.sql files, 83 oc_upgrade651_*.sql files, 83 oc_upgrade65_*.sql files, 83 ocscript directory, contents of, 82 OD_DETAILS table, 88 OD_INFO table, 88 ODBC common problems, 130 driver compatibility, 148 driver file, specifying on UNIX, 139 environment problems, 134 problems, correcting, 148 testing connection to RDBMS, 135 testing on UNIX, 136 testing on Windows, 135 tracing on UNIX, 151 tracing utility, using, 151 troubleshooting, 129 ODBC (Open Database Connectivity) configuring data sources, 41 configuring, which computer, 42 data sources, odbc.ini file heading, 56 environment variables on UNIX, 55 required database client software, 42 ODBC Administrator, 43, 44, 135 ODBC Data Sources, odbc.ini file section, 138 ODBC drivers adding data source to odbc.ini file on UNIX, 56 configuring DataDirect Text driver on UNIX, 59 61 configuring DataDirect Text driver on Windows, 4951 configuring IBM DB2 OS/390 on UNIX, 58 configuring IBM DB2 OS/390 on Windows, 45 configuring IBM DB2 UDB Wire Protocol on UNIX, 57 configuring IBM DB2 UDB Wire Protocol on Windows, 45 configuring Informix Wire Protocol on UNIX, 58
configuring Informix Wire Protocol on Windows, 46 configuring Microsoft SQL Server driver on UNIX, 61 configuring Microsoft SQL Server driver on Windows, 51, 53 configuring on UNIX, 5464 configuring on Windows, 4354 configuring Oracle Wire Protocol on UNIX, 59 configuring Oracle Wire Protocol on Windows, 47 configuring SQL Server driver on Windows, 51 53 configuring Sybase Wire Protocol on UNIX, 62 configuring Sybase Wire Protocol on Windows, 47 configuring Teradata driver on AIX, 62 configuring Teradata driver on HP-UX, 63 configuring Teradata driver on Solaris, 63 configuring Teradata driver on Windows, 5354 configuring the odbc.ini file on UNIX, 55 configuring Wire Protocol drivers on Windows, 4445 editing a data source name on Windows, 54 testing Wire Protocol drivers on Windows, 47 odbc.ini file adding a data source to, 48, 56, 65 bracketed headings, 56 common problems, 135 configuring data sources, 55 editing, 54 IBM DB2 example, 57 IBM DB2 OS/390 example, 58 Informix example, 58 inspecting for problems, 138 Microsoft SQL Server example, 61 Oracle example, 59 Sybase example, 62 Teradata example, 62, 63 Teradata example on AIX, 62 Teradata example on HP-UX, 63 Teradata example on Solaris, 63 Trace parameter, 152 odbcconfig utility, 136 ODBCINI variable, 138 OLAP Command Script files, 158 OLAP Metadata Catalog accessing, 121 backups, 87
236 Index
connecting to, 66, 173 connection defaults, setting, 68 connection problems, 129, 131 creating, 77 creating automatically, 79 creating tables for manually, 80, 84 data migration, 81 database platforms supported, 77 database requirements, 78 deleting, 86 deleting locks, 124 deleting tables, 80, 84 drill-through tables in, 91 Hybrid Analysis tables in, 91 identifying before loading, 162 information, 77 Integration Server and, 11 manual editing, 87 metaoutline tables, 89 OLAP model tables, 88 post-setup steps, 85 problems, database level, 146 recovery, 87 upgrading, 81 viewing permissions, 126 OLAP Metadata Catalog Setup dialog box, 79 OLAP metaoutlines, viewing in sample application, 118, 119 OLAP models access codes, 127 access permissions, 125 accessing, 121 displaying primary keys, 28 locks, 122 opening in exclusive access mode, 122 opening in standard access mode, 122 overwriting, 122 permissions, viewing, 125 restoring, 88 tables in OLAP Metadata Catalog, 88 users, 123, 124 OLAP models, viewing in sample application, 118, 119 OLAPICMD program problems, 140 starting, 154 olapisrv.log file, 14
olapisvr file, using to start Integration Server, 12 OLAPISVR program command-line switches, 14 starting in UNIX, 12 starting using startup.bat file, 12 olapisvr.log file, specifying location, 18 OM_DESCRIPTION table, 89 OM_INFO table, 88, 125 OM_PROPERTIES table, 88 OMB_DETAILS table, 89 OMB_INFO table, 89 open database connectivity. See ODBC (Open Database Connectivity) opening metaoutlines, options, 122 opening OLAP models, options, 122 Oracle configuring the Wire Protocol driver on Windows, 47 odbc.ini file example, 59 SQL*Loader, 99 SQL*Plus, 83, 95 tnsnames.ora file, 147 orphaned locks, deleting, 123, 124 OV_INFO table, 89 overwriting, OLAP models and metaoutlines, problem, 122 OVL_REL_DETAILS table, 89 OVL_REL_INFO table, 89 OVP_REL_DETAILS table, 89 OVP_REL_INFO table, 89
P
parentheses, adding to drill-through filters, 19 password problems, 129, 145 permissions changing, 125 for metaoutlines, 125, 126 for OLAP models, 125 problems in the RDBMS, 134 required for hyperion user, 38 shutting down Integration Server, 172 viewing, 125 working with, 121 port 3388, 22, 149, 154 prerequisites for Integration Server, 12 Integration Services Shell, 153
Index 237
primary data source, 161 privileges, common problems, 130 problems starting Integration Server, 12 system-level, 144 TCP/IP, 132 with connections, correcting, 129, 131 with ODBC, correcting, 129, 131, 146 profile file, 137
Q
Query Analyzer utility (Microsoft SQL Server), 83, 95
R
RDBMS client problems, correcting, 146 common problems, 130 problems, correcting, 146 server name problems, 144 testing connection from database client, 134 testing connection from Integration Server, 139 testing connection from Integration Services Console, 141 testing connection to ODBC, 135 testing ODBC connection on UNIX, 136 testing ODBC connection on Windows, 135 user account problems, 134 read access, setting, 125 read locks deleting, 124 viewing, 124 rebuilding tables, 82 reconnecting to Integration Server, 30 recording data load scripts, 157 recording scripts, 157 records, committing during data load, 16 recovery, OLAP Metadata Catalog, 87 relational database management system.. See RDBMS relational databases client software required, 42 connections to, diagram, 133 problems, correcting, 146 user aliases and synonyms, 37 releases compatibility problems, 140
determining for Integration Services, 160 removing OLAP Metadata Catalog, 86 OLAP Metadata Catalog tables, 82 TBC database tables, 99 TBC_MD database tables, 102 restoring metaoutlines, 88 OLAP Metadata Catalogs, 87 OLAP models, 88 restricting access to table records, 23 rights, problems in the RDBMS, 134 round-trip problem, Unicode as solution, 34 running Integration Server, 12, 30 Integration Services Shell, 153 scripts with Integration Services Shell, 157
S
sample application creating, 93 creating for Hybrid Analysis, 106 creating for Unicode, 115 preparation steps, 95 setting up automatically, 95 setting up manually, 97 Unicode (Oracle only), 115 viewing, 118, 119 Sample Application Setup dialog box, 96 sample database, TBC, 94 samples\\tbcdbase directory, contents of, 99 scheduled loads, enabling on UNIX, 38 scheduling, batch files for loads, 172, 173 scripts for loading Hybrid Analysis data, 100 for loading TBC database, 100 for loading TBC_MD database, 104 for OLAP Metadata Catalog, 82 for TBC database, 99 for TBC_MD database, 102 incremental loads, syntax, 159 location, sample application, 97, 108 member and data loads, recording, 157 running with Integration Services Shell, 157 secondary data source, 162 security access permissions, 125
238 Index
locks, 122 problems, 132 semicolons, in Integration Services Shell syntax, 156 server problems, 144 server software port number, 149 problems, correcting, 149 troubleshooting, 129 Servers menu, OLAP Metadata Catalog, Delete Locks command, 123 servers, connecting to, 66 SET commands, 155 Set Login Defaults dialog box, 69 SETCATALOG command described, 155, 162 example of, 173 SETLOCALE command, described, 155 SETSOURCE command described, 155, 161 example of, 173 SETTARGET command described, 155, 162 example of, 173 setting connection defaults, 68 connection information for Essbase Server, 67 data source, 173 Essbase Server, 173 network listeners, 25 OLAP Metadata Catalog, 162, 173 primary data source, 161 secondary data source, 161 target Essbase Server, 162 shell scripts for loading TBC database, 100 SHUTDOWN command, 30, 172 shutting down Integration Server, 30, 172 software components common problem areas, 131 connection problems, 130 source database users, 121 source database, artifact limits, 192 sp-fkeys, 86 SQL scripts for loading TBC database, 100 for OLAP Metadata Catalog, 82, 83 for TBC database, 99
for TBC_MD database, 102 location, sample application, 97, 108 SQL Server. See Microsoft SQL Server SQL Server driver, configuring on Windows, 5153 SQL*Loader, 99 SQL*Plus, 83, 95 stack size of thread, specifying, 24 standard access mode, 122 starting Integration Server, 12, 30 Integration Server, methods, 12 Integration Server, programs necessary, 12 Integration Services Shell, 153 schedule services on Windows, 38 startup switches for Integration Server, 14 startup.bat file, 12, 13 described, 27 storing startup switch settings, 27 using to start Integration Server, 12 STATUS command, described, 156, 160 stopping Integration Server, 30 storage metaoutline, 87 OLAP model, 87 switches, for OLAPISVR, 14 Sybase BCP utility, 95 configuring Wire Protocol driver on Windows, 47 example odbc.ini file, 62 execute permission required, 86 ISQL utility, 83, 95 sp-fkeys, 86 Sybsystemprocs, 86 Sybsystemprocs, 86 synonyms, 37 syntax, Integration Services Shell commands, 156 system-level problems, 144
T
tables creating for OLAP Metadata Catalog manually, 80, 84 creating for TBC database, 99 creating for TBC_MD database, 101 drill-through, 91 Hybrid Analysis, 91 metaoutline, 89
Index 239
miscellaneous, 92 OLAP model, 88 read-only, 87 rebuilding tables, 82 restricting access, 23 upgrading for OLAP Metadata Catalog, 81 tables, viewing in sample application, 118 Task Scheduler service Windows 2000, starting, 38 Windows 2003, starting, 39 Windows XP, starting, 39 TBC data load script location, 97, 107, 108, 111, 113, 116 database requirements, 98, 101, 116 database, creating tables for, 99 database, viewing, 118, 119 loading data into, 100 sample database, 94 setting up relational data source, 98 table creation script location, 97, 107, 108, 111, 113 viewing columns in, 118 viewing metaoutline in, 119 viewing OLAP model in, 119 viewing tables in, 118 TBC Metaoutline information, 94 viewing, 118, 119 TBC Model information, 94 viewing, 118, 119 tbc_create_database_name_unicode.sql file, 117 TBC_MD data load script location, 97 database requirements, 101 database, creating, 101 database, creating tables for, 101 loading data into using XML Import, 104 sample OLAP Metadata Catalog, 94 table creation script location, 97, 107, 109, 111, 113, 116 TBC_Model_Unicode.xml, 117 TCP/IP problems, 132 Teradata example odbc.ini file for AIX, 62 example odbc.ini file for HP-UX, 63
example odbc.ini file for Solaris, 63 mapping a JDBC driver, 72, 73 odbc.ini file example, 62, 63 Teradata ODBC driver, configuring on Windows, 5354 testing connections, overview, 132 database client, 134 Integration Server connection to Essbase Server, 142 Integration Server connection to RDBMS, 139 Integration Services Console connection to RDBMS, 141 Integration Services Console to Essbase Server, 143 ODBC connection to RDBMS, 135 ODBC on UNIX, 136 ODBC on Windows, 135 Text driver configuring on UNIX, 5961 configuring on Windows, 4951 threads specifying number during data load, 22 specifying stack size, 24 timestamp, locks and, 123 tnsnames.ora file, 147 Trace parameter, odbc.ini file, 152 tracing utility ODBC for UNIX, 151 ODBC for Windows, 151 transformation rules, limits with Hybrid Analysis, 195 transformations not supported by Hybrid Analysis, 196 troubleshooting connections, 129 ODBC, 129 system-level problems, 144 TCP/IP, 132 turning off Integration Server, 30
U
Unicode administration tools, 35 application modes, 34 creating OLAP Metadata Catalog, 80 features not supported, 33
240 Index
parameter in LOADALL command, 168, 171 parameter in LOADMEMBER command, 163, 165 required utilities for sample application setup, 115 sample application for IBM DB2, 116 sample application setup, 115 viewing the metaoutline, 119 viewing the OLAP model, 119 Unicode-enabled administration tools, 35 parameter in LOADALL command, 168, 171 parameter in LOADMEMBER command, 163, 165 Unicode-mode application, defined, 34 UNIFORMVALUEFORMAT (ais.cfg parameter), 29 UNIX environment variables for ODBC, 55 environment variables, configuring, 55 manual installation steps, 75 ODBC data sources, configuring, 54 ODBC environment variables, 55 ODBC tracing, 151 odbcconfig utility, 136 scheduled loads, enabling, 38 testing ODBC, 136 unload database and application from memory, 25 unreleased locks, deleting, 123, 124 upgrading OLAP Metadata Catalog, 81 user accounts, administrative, 134 user aliases, 37 user names locks and, 123 problems with, 129, 145 user synonyms, 37 users deleting, 121 disconnecting, 121 Essbase, defined, 121 Integration Server, defined, 121 managing permissions and locks, 121 multiple, 123 source database users, 121 UTF-8 encoding, 34
variables.. See environment variables VERSION command, described, 156, 160 versions, compatibility problems, 140 vi program, 56 viewing log files, 31 permissions for metaoutlines, 126 permissions for OLAP models, 125 viewing sample application, 118, 119
W
warning messages, 20 wide area network (WAN), 66 Windows ODBC testing, 135 ODBC tracing, 151 testing ODBC, 135 Windows 2000 ODBC Administrator, 43 Task Scheduler service, starting, 38 Windows 2003 Task Scheduler service, starting, 39 Windows service Task Scheduler, starting, 38, 39 using -I parameter, 19 using to start Integration Server, 13 using to stop Integration Server, 30 Windows XP ODBC Administrator, 43 Task Scheduler service, starting, 39 write access, setting, 125 write locks deleting, 124 viewing, 124
X
XML files for loading TBC_MD database, 104 XML Import for loading TBC_MD database, 104
V
validation of metaoutlines, 28
Index 241
242 Index