(SAP) Connector en
(SAP) Connector en
(SAP) Connector en
SAP Connector
Informatica Cloud Data Integration SAP Connector
October 2023
© Copyright Informatica LLC 2015, 2023
This software and documentation are provided only under a separate license agreement containing restrictions on use and disclosure. No part of this document may be
reproduced or transmitted in any form, by any means (electronic, photocopying, recording or otherwise) without prior consent of Informatica LLC.
U.S. GOVERNMENT RIGHTS Programs, software, databases, and related documentation and technical data delivered to U.S. Government customers are "commercial
computer software" or "commercial technical data" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such,
the use, duplication, disclosure, modification, and adaptation is subject to the restrictions and license terms set forth in the applicable Government contract, and, to the
extent applicable by the terms of the Government contract, the additional rights set forth in FAR 52.227-19, Commercial Computer Software License.
Informatica, the Informatica logo, Informatica Cloud, and PowerCenter are trademarks or registered trademarks of Informatica LLC in the United States and many
jurisdictions throughout the world. A current list of Informatica trademarks is available on the web at https://www.informatica.com/trademarks.html. Other company
and product names may be trade names or trademarks of their respective owners.
Portions of this software and/or documentation are subject to copyright held by third parties. Required third party notices are included with the product.
DISCLAIMER: Informatica LLC provides this documentation "as is" without warranty of any kind, either express or implied, including, but not limited to, the implied
warranties of noninfringement, merchantability, or use for a particular purpose. Informatica LLC does not warrant that this software or documentation is error free. The
information provided in this software or documentation may include technical inaccuracies or typographical errors. The information in this software and documentation
is subject to change at any time without notice.
NOTICES
This Informatica product (the "Software") includes certain drivers (the "DataDirect Drivers") from DataDirect Technologies, an operating company of Progress Software
Corporation ("DataDirect") which are subject to the following terms and conditions:
1. THE DATADIRECT DRIVERS ARE PROVIDED "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING BUT NOT LIMITED TO,
THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
2. IN NO EVENT WILL DATADIRECT OR ITS THIRD PARTY SUPPLIERS BE LIABLE TO THE END-USER CUSTOMER FOR ANY DIRECT, INDIRECT, INCIDENTAL,
SPECIAL, CONSEQUENTIAL OR OTHER DAMAGES ARISING OUT OF THE USE OF THE ODBC DRIVERS, WHETHER OR NOT INFORMED OF THE POSSIBILITIES
OF DAMAGES IN ADVANCE. THESE LIMITATIONS APPLY TO ALL CAUSES OF ACTION, INCLUDING, WITHOUT LIMITATION, BREACH OF CONTRACT, BREACH
OF WARRANTY, NEGLIGENCE, STRICT LIABILITY, MISREPRESENTATION AND OTHER TORTS.
The information in this documentation is subject to change without notice. If you find any problems in this documentation, report them to us at
[email protected].
Informatica products are warranted according to the terms and conditions of the agreements under which they are provided. INFORMATICA PROVIDES THE
INFORMATION IN THIS DOCUMENT "AS IS" WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING WITHOUT ANY WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND ANY WARRANTY OR CONDITION OF NON-INFRINGEMENT.
Table of Contents 3
Part III: Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
Chapter 5: Troubleshooting. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
4 Table of Contents
Chapter 10: IDoc mapplets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
Segments and groups. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
IDocs properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
Generating IDoc metadata to local file. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
Outbound mapplet. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
Outbound mapplet ports. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
Target object for outbound mapplet error output. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
Inbound mapplet. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
Key fields and control record fields. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
IDoc primary and foreign keys. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
Configuring IDoc mapplets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
Table of Contents 5
Part VII: Data integration for SAP ADSO. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
Chapter 16: Data Integration for SAP ADSO sources using SAP Table
Reader. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
Importing an SAP ADSO object. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
Chapter 17: Data Integration for SAP ADSO targets using SAP ADSO Writer 102
Rules and guidelines for SAP ADSO targets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
Chapter 18: Mappings and mapping tasks with SAP ADSO Writer. . . . . . . . . . 104
SAP ADSO Extractor targets in mappings. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
Activate option mode for SAP ADSO Writer mappings. . . . . . . . . . . . . . . . . . . . . . . . . . . 106
Fixed partitioning for SAP ADSO targets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106
Configuring a mapping with an SAP ADSO target. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
Creating a mapping task. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
Index. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
6 Table of Contents
Preface
Use SAP Connector to learn how to read from or write to SAP by using Cloud Data Integration. Learn to create
an SAP connection, develop and run synchronization tasks, mappings, mapping tasks, and data transfer tasks
in Cloud Data Integration.
Informatica Resources
Informatica provides you with a range of product resources through the Informatica Network and other online
portals. Use the resources to get the most from your Informatica products and solutions and to learn from
other Informatica users and subject matter experts.
Informatica Documentation
Use the Informatica Documentation Portal to explore an extensive library of documentation for current and
recent product releases. To explore the Documentation Portal, visit https://docs.informatica.com.
If you have questions, comments, or ideas about the product documentation, contact the Informatica
Documentation team at [email protected].
https://network.informatica.com/community/informatica-network/products/cloud-integration
Developers can learn more and share tips at the Cloud Developer community:
https://network.informatica.com/community/informatica-network/products/cloud-integration/cloud-
developers
7
https://marketplace.informatica.com/
To search the Knowledge Base, visit https://search.informatica.com. If you have questions, comments, or
ideas about the Knowledge Base, contact the Informatica Knowledge Base team at
[email protected].
Subscribe to the Informatica Intelligent Cloud Services Trust Center to receive upgrade, maintenance, and
incident notifications. The Informatica Intelligent Cloud Services Status page displays the production status
of all the Informatica cloud products. All maintenance updates are posted to this page, and during an outage,
it will have the most current information. To ensure you are notified of updates and outages, you can
subscribe to receive updates for a single component or all Informatica Intelligent Cloud Services
components. Subscribing to all components is the best way to be certain you never miss an update.
To subscribe, on the Informatica Intelligent Cloud Services Status page, click SUBSCRIBE TO UPDATES. You
can choose to receive notifications sent as emails, SMS text messages, webhooks, RSS feeds, or any
combination of the four.
To find online support resources on the Informatica Network, click Contact Support in the Informatica
Intelligent Cloud Services Help menu to go to the Cloud Support page. The Cloud Support page includes
system status information and community discussions. Log in to Informatica Network and click Need Help to
find additional resources and to contact Informatica Global Customer Support through email.
The telephone numbers for Informatica Global Customer Support are available from the Informatica web site
at https://www.informatica.com/services-and-training/support-services/contact-us.html.
8 Preface
Part I: Introduction to SAP
Connector
This part contains the following chapter:
9
Chapter 1
SAP integrates multiple business applications and solutions, such as Customer Relationship Management
(CRM), Advanced Planner and Optimizer (APO), and Bank Analyzer. Developers can add business logic within
SAP using Java 2 Enterprise Edition (J2EE) or Advanced Business Application Programming-Fourth
Generation (ABAP/4 or ABAP), a language proprietary to SAP.
Data Integration supports ABAP, IDoc read, IDoc write, BAPI/RFC, ADSO write, and ODP read functions to
integrate with SAP systems. You can choose an SAP connection type to connect to SAP systems based on
the interface requirements.
You can use the SAP connection in synchronization tasks, mappings, mapping tasks, and data transfer tasks.
You can switch mappings to advanced mode to include transformations and functions that enable advanced
functionality. Create a synchronization task to synchronize data between a source and target. Create a
mapping task to process data based on the data flow logic defined in a mapping or integration template.
For more information about configuring assets and transformations, see Mappings, Transformations, and
Tasks in the Data Integration documentation.
The following table lists the integration method corresponding to the different levels and connections:
10
SAP Integration Level of Connections
Method Integration
SAP ADSO Object-level SAP Table and SAP ADSO Writer connections
integration The SAP Table connection information is available in the SAP Table
Connector documentation. For more information about the SAP Table
connection, see the SAP Table Connector documentation.
Communication interfaces
SAP uses TCP/IP as the native communication interface to communicate with Data Integration.
SAP also uses the Remote Function Call (RFC) communication protocol to communicate with Data
Integration. To execute remote calls from Data Integration, SAP requires connection information, and the
service name and gateway on the application server. The service and gateway parameters, and connection
information is stored in a configuration file named sapnwrfc.ini on the Secure Agent machine.
Communication interfaces 11
Part II: SAP Connector
administration
This part contains the following chapters:
12
Chapter 2
To process data through SAP, the administrator must verify the required licenses are enabled for the SAP
systems.
13
Downloading and configuring SAP libraries for IDoc and BAPI/RFC
Download and configure the SAP NetWeaver RFC SDK libraries. Contact SAP Customer Support if you
encounter any issues when you download the libraries.
Linux 64 - libicudata.so.50
- libicui18n.so.50
- libicuuc.so.50
- libsapnwrfc.so
- libsapucum.so
Windows 64 - icudt50.dll
- icuin50.dll
- icuuc50.dll
- libsapucum.dll
- sapnwrfc.dll
Verify that you download the most recent version of the libraries.
3. Copy the SAP NetWeaver RFC SDK 7.50 libraries to the following directory:
<Informatica Secure Agent installation directory>\apps\Data_Integration_Server\ext
\deploy_to_main\bin\rdtm\
Create the deploy_to_main\bin\rdtm directory if it does not already exist.
Note: If you upgrade from a 32-bit operating system, the Secure Agent copies the 32-bit SAP NetWeaver
RFC SDK 7.50 libraries to the directory. You must replace the 32-bit libraries with 64-bit libraries. If you
upgrade from a 64-bit operating system, you do not need to perform this step. The Secure Agent copies
the 64-bit SAP NetWeaver RFC SDK 7.50 libraries to the directory.
4. Set the following permissions for each NetWeaver RFC SDK library:
• Read, write, and execute permissions for the current user.
• Read and execute permissions for all other users.
5. Download the 64-bit SAP JCo libraries based on the operating system on which the Secure Agent runs:
Windows sapjco3.jar
sapjco3.dll
Linux sapjco3.jar
libsapjco3.so
Verify that you download the most recent version of the libraries.
Windows ../bin/rdtm-extra/tpl/sap/sapjco3.jar;../bin/rdtm/javalib/sap/sap-adapter-
common.jar
Linux ../bin/rdtm-extra/tpl/sap/sapjco3.jar:../bin/rdtm/javalib/sap/sap-adapter-
common.jar
Note: If you copy the value directly from the table, the hyphens (-) in the value are incorrectly copied.
Copy the value to a text editor and make sure that the value you copied is not corrupted.
7. Click OK to save the changes.
8. Repeat steps 2 through 7 on every machine where you installed the Secure Agent.
8. Restart the Secure Agent.
Configuring sapnwrfc.ini
SAP uses the communications protocol, Remote Function Call (RFC), to communicate with other systems.
SAP stores RFC-specific parameters and connection information in a file named sapnwrfc.ini. To enable the
Secure Agent to connect to the SAP system as an RFC client, create and configure the sapnwrfc.ini file on
the machines that host the Secure Agent.
When you read data from SAP, if you define the path and file name of the sapnwrfc.ini file in the SAP
connection, the Secure Agent uses the sapnwrfc.ini file. However, if you define only the path of the
sapnwrfc.ini file in the connection, the Secure Agent first verifies if an sapnwrfc.ini file exists in the
specified path. If the sapnwrfc.ini file exists, the Secure Agent uses the sapnwrfc.ini file. Else, an
exception occurs.
To process data through RFC/BAPIs, read IDocs, and write IDocs, you cannot use the sapnwrfc.ini file.
Use a DOS editor or WordPad to configure the sapnwrfc.ini file. Notepad can introduce errors to the
sapnwrfc.ini file.
After you create the sapnwrfc.ini file, copy the file to the following directory and restart the Secure Agent:
Note: If you are upgrading from an earlier version, you do not need to perform this step. The Secure Agent
copies the sapnwrfc.ini file to the directory.
For the serverless runtime environment, the sapnwrfc.ini file is copied from the AWS location to the
following serverless agent directory:
/data2/home/cldagnt/SystemAgent/apps/Data_Integration_Server/ext/deploy_to_main/bin/rdtm
You can configure the following types of connections in the sapnwrfc.ini file:
Create this connection to enable communication between an RFC client and an SAP system. Each
connection entry specifies one application server and one SAP system.
The following sample shows a connection entry for a specific SAP application server in the
sapnwrfc.ini file:
DEST=sapr3
ASHOST=sapr3
SYSNR=00
Connection to use SAP load balancing
Create this connection to enable SAP to create an RFC connection to the application server with the least
load at run time. Use this connection when you want to use SAP load balancing.
The following sample shows a connection entry for SAP load balancing in the sapnwrfc.ini file:
DEST=sapr3
R3NAME=ABV
MSHOST=infamessageserver.informatica.com
GROUP=INFADEV
Connection to an RFC server program registered at an SAP gateway
Create this connection to connect to an SAP system from which you want to receive outbound IDocs.
The following sample shows a connection entry for an RFC server program registered at an SAP gateway
in the sapnwrfc.ini file:
DEST=sapr346CLSQA
PROGRAM_ID=PID_LSRECEIVE
GWHOST=sapr346c
GWSERV=sapgw00
DEST Logical name of the SAP system for the Use this parameter for the following types of
connection. connections:
All DEST entries must be unique. You must - Connection to a specific SAP application server
have only one DEST entry for each SAP - Connection to use load balancing
system. - Connection to an RFC server program registered
at an SAP gateway
For SAP versions 4.6C and later, use up to 32
characters. For earlier versions, use up to
eight characters.
ASHOST Host name or IP address of the SAP Use this parameter to create a connection to a
application. The Secure Agent uses this entry specific SAP application server.
to attach to the application server.
R3NAME Name of the SAP system. Use this parameter to create a connection to use
SAP load balancing.
MSHOST Host name of the SAP message server. Use this parameter to create a connection to use
SAP load balancing.
GROUP Group name of the SAP application server. Use this parameter to create a connection to use
SAP load balancing.
PROGRAM_ID Program ID. The Program ID must be the Use this parameter to create a connection to an
same as the Program ID for the logical RFC server program registered at an SAP gateway.
system that you define in the SAP system to
send or receive IDocs.
GWHOST Host name of the SAP gateway. Use this parameter to create a connection to an
RFC server program registered at an SAP gateway.
GWSERV Server name of the SAP gateway. Use this parameter to create a connection to an
RFC server program registered at an SAP gateway.
TRACE Debugs RFC connection-related problems. Use this parameter for the following types of
Set one of the following values based on the connections:
level of detail that you want in the trace: - Connection to a specific SAP application server
- 0. Off - Connection to use load balancing
- 1. Brief - Connection to an RFC server program registered
- 2. Verbose at an SAP gateway
- 3. Full
Create a single logical system in SAP for IDoc ALE integration with SAP Connector. When you define SAP
Connector as a logical system, SAP acknowledges SAP Connector as an external system that can receive
outbound IDocs from SAP and send inbound IDocs to SAP.
Note: These steps are based on SAP version 4.6C. The steps may differ if you use a different version. For
complete instructions on creating a logical system in SAP, see the SAP documentation.
1. Go to transaction SALE.
The Display IMG window appears.
2. Expand the tree to navigate to the Application Link Enabling > Sending and Receiving Systems > Logical
Systems > Define Logical System operation.
3. Click the IMG - Activity icon to run the Define Logical System operation.
An informational dialog box appears.
4. Click Enter.
The Change View Logical Systems window appears.
5. Click New Entries.
The New Entries window appears.
6. Enter a name and description for the logical system entry for SAP Connector.
1. Go to transaction SM59.
The Display and Maintain RFC Destinations window appears.
2. Click Create.
The RFC Destination window appears.
3. Enter the name of the logical system you created as the RFC destination.
4. To create a TCP/IP connection, enter T as the connection type.
5. Enter a description for the RFC destination.
6. Click Save.
7. For Activation Type, click Registration.
8. For Program ID, enter the same name as the RFC destination name.
Use the Program ID as the value for the PROGRAM_ID parameter in the sapnwrfc.ini file.
1. Go to transaction WE21.
2. Click Ports > Transactional RFC.
3. Click Create.
The Ports in IDoc Processing dialog box appears.
4. Click Generate Port Name or Own Port Name and enter a name.
5. Click Enter.
6. Enter a description for the port.
7. Select the IDoc record version type.
8. Enter the name of the RFC destination you created.
1. Go to transaction WE20.
2. Click Create.
3. Enter the following properties:
Partner number Name of the logical system you created for SAP Connector.
Partner type Partner profile type. Enter LS for logical system for ALE distribution systems.
Lang Language code that corresponds to the SAP language. Enter EN for English.
Partner status Indicates the status of communication with the partner. To communicate with the
partner, enter A for active.
Step 5. Create outbound and inbound parameters for the partner profile
Outbound parameters define the IDoc message type, IDoc basic type, and port number for outbound IDocs.
Inbound parameters define the IDoc message type for inbound IDocs.
SAP uses outbound parameters when it sends IDocs to SAP Connector. Create an outbound parameter for
each IDoc message type that SAP sends to SAP Connector. SAP uses inbound parameters when it receives
IDocs from SAP Connector. Create an inbound parameter for each IDoc message type that SAP receives from
SAP Connector.
Message Type The IDoc message type the SAP system sends to SAP Connector.
IDoc Type The IDoc basic type of the IDocs the SAP system sends to SAP Connector.
3. Click Save.
The Packet Size property appears.
4. Enter a value between 10 and 200 IDocs as the packet size.
The packet size determines the number of IDocs that SAP sends in one packet to SAP Connector.
5. Click Enter.
6. Repeat steps from 1 to 5 to create an outbound parameter for each IDoc message type that SAP sends
to SAP Connector.
7. Click Create Inbound Parameter.
The Partner Profiles: Inbound Parameters window appears.
Message Type The IDoc message type the SAP system receives from SAP Connector.
Process Code The process code. The SAP system uses the process code to call the appropriate
function module to process the IDocs it receives.
9. Click Enter.
10. Repeat steps 7 through 9 to create an inbound parameter for each IDoc message type that SAP receives
from SAP Connector.
The setup of the user and profiles is done within SAP using the SAP GUI. This activity is external to Data
Integration.
BAPI/RFC
The following table describes the authorization an SAP user requires to run tasks using the BAPI/RFC
functions:
Note: You must add the S_RFC authorization object to run tasks with the BAPI/RFC functions at runtime.
In addition to the above authorization, you must have access to any BAPI/RFC function that you want to
run.
IDoc
The following table describes the authorization an SAP user requires to run tasks with IDoc messages:
Note: You must add the S_RFC authorization object to run tasks with the IDoc messages at runtime. In
addition to the above authorization, you must have access to specific IDoc and underlying transactions
that you want to run.
Contact SAP Customer Support if you encounter any issues with downloading the libraries.
Windows sapjco3.jar
sapjco3.dll
Linux sapjco3.jar
libsapjco3.so
Verify that you download the most recent version of the libraries.
Windows ../bin/rdtm-extra/tpl/sap/sapjco3.jar;../bin/rdtm/javalib/sap/sap-adapter-
common.jar
Linux ../bin/rdtm-extra/tpl/sap/sapjco3.jar:../bin/rdtm/javalib/sap/sap-adapter-
common.jar
Note: If you copy the value directly from the table, the hyphens (-) in the value are incorrectly copied.
Copy the value to a text editor and make sure that the value you copied is not corrupted.
g. From the Type list, select DTM.
h. Enter the following JVMClassPath value based on the operating system on which the Secure Agent
runs:
Windows pmserversdk.jar;../../bin/rdtm-extra/tpl/sap/
sapjco3.jar;../../bin/rdtm/javalib/sap/sap-adapter-common.jar
Linux pmserversdk.jar:../../bin/rdtm-extra/tpl/sap/
sapjco3.jar:../../bin/rdtm/javalib/sap/sap-adapter-common.jar
Note: If you copy the value directly from the table, the hyphens (-) in the value are incorrectly copied.
Copy the value to a text editor and make sure that the value you copied is not corrupted.
i. Click OK to save the changes.
j. Repeat steps 2 through 4 on every machine where you installed the Secure Agent.
5. Restart the Secure Agent.
1. Find the transport files in the following directory on the Secure Agent machine:
<Informatica Secure Agent installation directory>\downloads\package-bwreader.<Latest
version>\package\rdtm\sap-transport\SAPBWReader
2. Copy the cofile transport file to the Cofile directory in the SAP transport management directory on each
SAP machine that you want to access.
The cofile transport file uses the following naming convention: RUN_BWRDR_K<number>.g00
3. Remove "RUN_BWRDR_" from the file name to rename the cofile. For example, for a cofile transport file
named RUN_BWRDR_K900723.g00, rename the file to K900723.g00.
Configuring HTTPS
To connect to SAP through HTTPS and read data from SAP BW sources, you must configure the machine that
hosts the Secure Agent and the machine that hosts the SAP system. You must also enable HTTPS when you
configure an SAP BW Reader connection in Data Integration.
Perform the following configuration tasks on the Secure Agent and SAP systems:
To configure HTTPS on the machine that hosts the Secure Agent, perform the following tasks:
To configure HTTPS on the machine that hosts the SAP system, perform the following tasks:
The SAP administrator must create a profile in the development, test, and production SAP systems so that
you can use the integration features. The profile name must include authorization for the objects and related
activities. The profile on the test system must be the same as the profile on the production system.
ACTVT 16
JOBGROUP *
S_RS_ADMWB RSADMWBOBJ Provide the Administrator Workbench Object name from which you want to
read data based on the requirements.
For more information about Administrator Workbench Objects, see the SAP
documentation.
ACTVT 3
S_RS_ICUBE RSINFOAREA Provide the InfoArea names that you want to access based on the
requirements.
For more information about InfoAreas, see the SAP documentation.
RSINFOCUBE Provide the InfoCube object names that you want to access based on the
requirements.
For more information about InfoCubes, see the SAP documentation.
ACTVT 3
Note: You can also add the following authorization objects based on the requirements of the SAP BW tasks
that you run from Data Integration:
• The S_RFC authorization object with the BAPI_ODSO_GETLIST and BAPI_ISET_GETLIST optional RFC
objects.
• The S_RS_ISET and S_RS_ODSO authorization objects with the activity value ACTVT=3.
After the administrator has performed the configuration, you can set up and use an SAP ADSO Writer
connection in mapping tasks.
The minimum version of the BW/4HANA system must be 2.0, and is identified through the SAP component
DW4CORE, Release 200.
Verify that you install the latest transport files to write to an SAP ADSO object.
Install the following data file and cofile to load data to an SAP BW/4HANA ADSO object.
1. Find the transport files in the following directory on the Secure Agent machine:
<Informatica Secure Agent installation directory>\downloads\package-SAPADSO.<Latest
version>\package\SAPADSO\sap-transport
2. Copy the cofile transport file to the Cofile directory in the SAP transport management directory on each
SAP machine that you want to access.
The cofile transport file uses the following naming convention: <TR Number>.<Sap system>.
3. Copy the data transport file to the Data directory in the SAP transport management directory on each
SAP machine that you want to access.
The data transport file uses the following naming convention: ADSO_WRITER_R<TR Number>.<Sap
system>.
4. To import the transports to SAP, in the STMS, click Extras > Other Requests > Add and add the transport
request to the system queue.
5. In the Add Transport Request to Import Queue dialog box, enter the request number for the cofile
transport.
The request number inverts the order of the renamed cofile as follows: <Sap system><number>.
6. In the Request area of the import queue, select the transport request number that you added, and click
Import.
The following table describes the required authorization to read from an SAP BW/4HANA ADSO:
Windows sapjco3.jar
sapjco3.dll
Linux sapjco3.jar
libsapjco3.so
Verify that you download the most recent version of the libraries.
3. Copy the JCo libraries to the following directory: <Informatica Secure Agent installation
directory>\apps\Data_Integration_Server\ext\deploy_to_main\bin\rdtm-extra\tpl\sap
Create the deploy_to_main\bin\rdtm-extra\tpl\sap directory if it does not already exist.
4. Configure the JAVA_LIBS property in Data Integration:
a. Log in to Data Integration.
b. Click Runtime Environments to access the Runtime Environments page.
c. To the left of the agent name, click Edit Secure Agent.
d. From the Service list, select Data Integration Server.
e. From the Type list, select Tomcat JRE.
Windows ../bin/rdtm-extra/tpl/sap/sapjco3.jar;../bin/rdtm/javalib/sap/
sap-adapter-common.jar
Linux ../bin/rdtm-extra/tpl/sap/sapjco3.jar:../bin/rdtm/javalib/sap/
sap-adapter-common.jar
Note: If you copy the value directly from the table, the hyphens (-) in the value are incorrectly copied.
Copy the value to a text editor and make sure that the value you copied is not corrupted.
g. From the Type list, select DTM.
h. Enter the JVMClassPath value based on the operating system on which the Secure Agent runs.
Windows pmserversdk.jar;../../bin/rdtm-extra/tpl/sap/
sapjco3.jar;../../bin/rdtm/javalib/sap/sap-adapter-common.jar
Linux pmserversdk.jar:../../bin/rdtm-extra/tpl/sap/
sapjco3.jar:../../bin/rdtm/javalib/sap/sap-adapter-common.jar
Note: If you copy the value directly from the table, the hyphens (-) in the value are incorrectly copied.
Copy the value to a text editor and make sure that the value you copied is not corrupted.
i. Click OK to save the changes.
j. Repeat steps 2 through 7 on every machine where you installed the Secure Agent.
5. Restart the Secure Agent.
When you use the serverless runtime environment, you do not need to download, install, configure, or
maintain the Secure Agent machine.
Use the serverless runtime environment for the following SAP connections:
For more information about serverless runtime environment properties, see Runtime Environments in the
Administrator documentation.
1. Create the following structure for the serverless agent configuration in AWS: <Supplementary file
location>/serverless_agent_config
2. Add the libraries in the Amazon S3 bucket in the following location in your AWS account:
<Supplementary file location>/serverless_agent_config/sap
3. Copy the following code snippet to a text editor:
version: 1
agent:
dataIntegrationServer:
autoDeploy:
sap:
jcos:
- fileCopy:
sourcePath: sap/jco/<sapjco_libary_filename>
- fileCopy:
sourcePath: sap/jco/<sapjco_libary_filename>
nwrfcs:
30
- fileCopy:
sourcePath: sap/nwrfc/<rfc_libary_filename>
- fileCopy:
sourcePath: sap/nwrfc/<sapnwrfc_filename>
where the source path is the directory path of the library files in AWS.
4. Ensure that the syntax and indentations are valid, and then save the file as
serverlessUserAgentConfig.yml in the following AWS location: <Supplementary file location>/
serverless_agent_config
When the .yml file runs, the libraries are copied from the AWS location to the serverless agent directory.
5. To configure the JAVA_LIBS or JVMClassPath property for the serverless runtime environment on Linux,
perform the following tasks in Administrator:
a. On the Serverless Environments tab, select the serverless runtime environment.
b. Click Edit.
c. On the Runtime Configuration Properties tab, select Data Integration Server as the service.
d. Select the following Type and Name, and then enter the JAVA_LIBS or JVMClassPath value based
on the connection type you want to use for the serverless runtime environment:
For more information about serverless runtime environment properties, see Runtime Environments in the
Administrator documentation.
Note: You cannot create an SNC connection when you use the serverless runtime environment.
• SAP connections, 33
• Troubleshooting, 48
32
Chapter 4
SAP connections
Use an SAP connection to read from and write data to SAP systems.
Data Integration supports ABAP, IDoc read, IDoc write, BAPI/RFC, ADSO write, and ODP read functions to
integrate with SAP systems. You can choose an SAP connection type to connect to SAP systems based on
the interface requirements.
SAP Table Connector Read data from SAP tables and write to any target. You can also write data from any source
to custom tables in SAP. Contact Global Customer Support for information about using the
SAP Table connection to write data to SAP systems.
The SAP Table connection information is available in the SAP Table Connector
documentation. For more information about the SAP Table connection, see the SAP Table
Connector documentation.
SAP BW Reader Read data from SAP BW objects such as InfoCubes, InfoSets, MultiProviders, and DataStore
objects.
SAP ODP Extractor Read data from SAP ODP objects from S/4HANA.
The SAP ODP Extractor connection information is available in the SAP ODP Extractor
documentation. For more information about the SAP ODP Extractor connection, see the SAP
ODP Extractor Connector documentation.
33
SAP IDoc and BAPI/RFC connections
SAP connections enable you to access SAP data through the IDoc or BAPI/RFC interfaces.
When you select SAP as the type, you can configure the following connections from the SAP Connection
Type list:
• IDoc Reader
• IDoc Writer
• SAP RFC/BAPI Interface
The following table describes the SAP RFC/BAPI Interface connection properties:
Connection Description
property
User Name Required. SAP user name with authorization on S_DATASET, S_TABU_DIS, S_PROGRAM, and
B_BTCH_JOB objects.
Connection String Required. DEST entry that you specified in the sapnwrfc.ini file for the SAP application
server.
Code Page The code page compatible with the SAP target. Select one of the following code pages:
- MS Windows Latin 1. Select for ISO 8859-1 Western European data.
- UTF-8. Select for Unicode data.
- Shift-JIS. Select for double-byte character data.
- ISO 8859-15 Latin 9 (Western European).
- ISO 8859-2 Eastern European.
- ISO 8859-3 Southeast European.
- ISO 8859-5 Cyrillic.
- ISO 8859-9 Latin 5 (Turkish).
- IBM EBCDIC International Latin-1.
Language Code Required. Language code that corresponds to the SAP language.
The following table describes the SAP IDoc Reader connection properties:
Connection Description
property
Destination Required. DEST entry that you specified in the sapnwrfc.ini file for the RFC server program
Entry registered at an SAP gateway. The Program ID for this destination entry must be the same as the
Program ID for the logical system you defined in SAP to receive IDocs.
Code Page Required. The code page compatible with the SAP source. Select one of the following code pages:
- MS Windows Latin 1. Select for ISO 8859-1 Western European data.
- UTF-8. Select for Unicode data.
- Shift-JIS. Select for double-byte character data.
- ISO 8859-15 Latin 9 (Western European).
- ISO 8859-2 Eastern European.
- ISO 8859-3 Southeast European.
- ISO 8859-5 Cyrillic.
- ISO 8859-9 Latin 5 (Turkish).
- IBM EBCDIC International Latin-1.
The following table describes the SAP IDoc Writer connection properties:
Connection Description
property
User Name Required. SAP user name with authorization on S_DATASET, S_TABU_DIS, S_PROGRAM, and
B_BTCH_JOB objects.
Connection String Required. DEST entry that you specified in the sapnwrfc.ini file for the SAP application
server.
Code Page Required. The code page compatible with the SAP target. Select one of the following code
pages:
- MS Windows Latin 1. Select for ISO 8859-1 Western European data.
- UTF-8. Select for Unicode data.
- Shift-JIS. Select for double-byte character data.
- ISO 8859-15 Latin 9 (Western European).
- ISO 8859-2 Eastern European.
- ISO 8859-3 Southeast European.
- ISO 8859-5 Cyrillic.
- ISO 8859-9 Latin 5 (Turkish).
- IBM EBCDIC International Latin-1.
Language Code Required. Language code that corresponds to the SAP language.
To enable the Secure Agent to connect to the SAP BW system through HTTPS, you must enable HTTPS and
specify the keystore details in the SAP BW Reader connection. To read data from SAP BW objects through
HTTPS, specify an SAP BW Reader connection configured for HTTPS when you create synchronization tasks,
mappings, or mapping tasks.
Property Description
Runtime Required. Runtime environment that contains the Secure Agent that you want to use to read data
Environment from SAP BW objects.
Username Required. SAP user name with the appropriate user authorization.
Message host Required when you create an SAP load balancing connection.
name Host name of the SAP message server.
Language Language code that corresponds to the language used in the SAP system.
Trace Use this option to track the JCo calls that the SAP system makes.
Specify one of the following values:
- 0. Off
- 1. Full
Default is 0.
SAP stores information about the JCo calls in a trace file.
You can access the trace files from the following directories:
- Design-time information: <Informatica Secure Agent installation directory>
\apps\Data_Integration_Server\<Latest version>\ICS\main\tomcat
- Run-time information: <Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<Latest version>\ICS\main\bin\rdtm
Port Range HTTP port range that the Secure Agent must use to read data from the SAP BW server in
streaming mode.
Enter the minimum and maximum port numbers with a hyphen as the separator. The minimum and
maximum port number can range between 10000 and 65535.
Default is 10000-65535.
SAP Additional Additional SAP parameters that the Secure Agent uses to connect to the SAP system as an RFC
Parameters client.
Specify the required RFC-specific parameters and connection information to enable
communication between Data Integration and SAP.
You can specify the Secure Network Communication (SNC) parameters as additional arguments
to securely connect to SAP as shown in the following format:
You can use SAP ADSO Writer Connector to create an application server connection or a load balancing
connection to connect using the SAP user name and password. You can also use create a Secured Network
Application (SNC) application server connection or a load balancing connection to connect to SAP BW/
4HANA using the SNC parameters.
You must create a connection to SAP BW/4 HANA on-premise systems before designing the mapping or
mapping task.
The following table describes the SAP ADSO Writer connection properties:
Connection Description
property
Runtime Runtime environment that contains the Secure Agent that you want to use to access SAP BW/
Environment 4HANA.
The following table describes the properties that must configure when you select Application Server
Connection as the connection type:
Connection Description
property
SAP System Number The system number of the SAP Server to connect.
SAP Username The SAP user name with the appropriate user authorization.
Additional Additional SAP parameters that the Secure Agent uses to connect to the SAP system.
Parameters For example, to generate SAP JCo and SAP CPIC trace, specify the following properties:
jco.client.trace="1";
jco.client.cpic_trace="3";
During the runtime, the JCo and CPIC traces file are generated in the following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>\ICS\main\bin\rdtm
During the design time, the CPIC traces are generated in the tomcat.out files at the
following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>tomcat.out
The following table describes the properties that must configure when you select Load Balancing Server
Connection as the connection type:
SAP Message Server The IP address or the host name of the SAP Message Server.
SAP Username The SAP user name with the appropriate user authorization.
Additional Additional SAP parameters that the Secure Agent uses to connect to the SAP system.
Parameters For example, to generate SAP JCo and SAP CPIC trace, specify the following properties:
jco.client.trace="1";
jco.client.cpic_trace="3";
During the runtime, the JCo and CPIC traces file are generated in the following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>\ICS\main\bin\rdtm
During the design time, the CPIC traces are generated in the tomcat.out files at the
following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>tomcat.out
SAP System Number The system number of the SAP Server to connect.
SNC My Name Optional. The Informatica client Personal Security Environment (PSE) or certificate name.
Default length is 256.
Use X509 Certificate Specifies the quality of protection. Select to use X509 Certificate based SNC connection.
Additional Parameters Additional SAP parameters that the Secure Agent uses to connect to the SAP system.
For example, to generate SAP JCo and SAP CPIC trace, specify the following properties:
jco.client.trace="1";
jco.client.cpic_trace="3";
During the runtime, the JCo and CPIC traces file are generated in the following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>\ICS\main\bin\rdtm
During the design time, the CPIC traces are generated in the tomcat.out files at the
following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>tomcat.out
SAP Message Server The IP address or the host name of the SAP Message Server.
SNC My Name Optional. The Informatica client PSE or certificate name generated on the Secure Agent
machine.
Default length is 256.
SNC Partner Name The Informatica client PSE or certificate name generated on the SAP Server.
Default length is 256.
Use X509 Certificate Specifies the quality of protection. Select to use X509 Certificate based SNC connection.
Additional Additional SAP parameters that the Secure Agent uses to connect to the SAP system.
Parameters For example, to generate SAP JCo and SAP CPIC trace, specify the following properties:
jco.client.trace="1";
jco.client.cpic_trace="3";
During the runtime, the JCo and CPIC traces file are generated in the following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>\ICS\main\bin\rdtm
During the design time, the CPIC traces are generated in the tomcat.out files at the
following location:
<Informatica Secure Agent installation directory>\apps
\Data_Integration_Server\<DIS version>tomcat.out
c. SAP Cryptographic Library Path. Specify the path of the SAP Cryptographic library.
You must specify sapcrypto.dll for Windows or libsapcrypto.so for Linux.
d. Use X509 Certificate. Select this option to use the X.509 certificate to log in.
e. X509 Certificate Path or SAP Username. Specify the path of the X509 certificate file with .crt
extension if you selected the X509 certificate. Otherwise, specify the SAP user name for which SNC
is configured in the SAP server.
f. Additional Parameters. Enter any additional connection parameters that you want to use.
9. Click Test Connection to determine if the connection to the SAP BW4/HANA system is successful.
10. Click Save to save the connection.
Troubleshooting
Use the following sections to troubleshoot errors in Data Integration.
Note: To get support for Data Integration, you might need to give your organization ID to Informatica Global
Customer Support. You can find your organization ID through the Organization menu in the upper right
corner.
To copy the organization ID, click the Copy option that appears when you hover the cursor to the right of the
Org ID field.
You can also find your organization ID on the Organization page in Administrator.
48
Part IV: Data integration using
BAPI/RFC functions
This part contains the following chapters:
49
Chapter 6
You can also use the BAPI/RFC Interface connection to read data from or write data to an SAP BAPI object
that contains a hierarchical structure.
When you run a mapping or mapping task, Data Integration makes the RFC function calls to SAP to process
data synchronously.
Business Application Programming Interfaces (BAPI) provide a way for third-party applications to
synchronously integrate with SAP at the object-level. You use BAPIs to read, create, change, or delete data in
SAP.
BAPIs allow access to the SAP system objects through methods for the business object types. Together with
the business object types, BAPIs define and document the interface standard at the business level.
You define BAPIs in the SAP Business Objects Repository. You can call BAPIs as an ABAP program within
SAP or from any external application. SAP Connector uses RFC protocol to call BAPI/RFC functions outside
of SAP.
You can view and test the BAPI interface definitions in SAP using transaction SE37.
50
Chapter 7
BAPI/RFC mapplets
You can generate SAP BAPI/RFC mapplets in Data Integration.
The BAPI/RFC mapplet includes a BAPI/RFC transformation. The BAPI/RFC transformation makes BAPI/RFC
calls in SAP. BAPI/RFC calls include requests to the SAP system, such as creating, changing, or deleting data
in SAP applications. To perform these tasks, BAPI/RFC functions use function parameter values.
For more information about configuring SAP BAPI Connector as a business service within a mapping or a
mapping task, see the How-To Library article,
How to Configure SAP BAPI Connector as a Business Service in Cloud Data Integration.
Function Parameters
BAPI/RFC functions can have the following parameters:
Scalar input Scalar input values. Some BAPI functions require scalar input values to perform tasks such
parameters. as changing data.
Scalar output Scalar output values that a BAPI function returns after performing a task.
parameters.
Table parameters. SAP structures with more than one row. Table parameters can be input, output, or both.
Input table parameters pass table input values to a BAPI/RFC function. For example, some
BAPI/RFC functions require table inputs to change data.
51
Input and Output Groups
The BAPI/RFC transformation can contain the following groups:
Group Description
Name
Scalar input Input group for scalar parameters. Contains a field for each scalar input parameter. The group name is
SCALAR_INPUT and the field names are SI_<FieldName>.
Table input One group for each table parameter. The fields represent import structures. The group name is
TABLE_INPUT_<structure name> and field names are TI_<StructureName>_<FieldName>.
Scalar Output group for scalar parameters. Contains a field for each scalar output parameter. The group
output name is SCALAR_OUTPUT and field names are SO_<FieldName>.
Table output One group for each table parameter. The fields represent output structures. The group name is
TABLE_OUTPUT_<structure name> and field names are TO_<StructureName>_<FieldName>.
Error output Passes data from data conversion errors and invalid BAPI/RFC calls. Map the error output field to a
target to see error messages about data conversion and BAPI/RFC calls.
You can also change the direction of the BAPI/RFC table parameters. The direction of the table parameters
determine which groups in the transformation are input groups and which are output groups.
If the BAPI return structure is custom, you can edit the return structure properties that you selected during
import for reusable BAPI/RFC transformations. Otherwise, you can view the BAPI/RFC return structure
parameters.
Parameter Description
Name Name of the export, import, and table parameters and columns.
Short Description Short description of the export, import, and table parameters and columns.
Optional Indicates if the Secure Agent should pass a value to the parameter when it calls the BAPI.
Direction Indicates if the parameter requires input or provides output. Values are: Input, Output, Both, or
None.
Default Default value of the parameter in SAP, if any. SAP uses the default value when the Secure Agent
does not pass a value to SAP.
Return Structure Return parameter name to determine the status of function calls. Value is RETURN if the BAPI
contains a default return structure. If the BAPI does not contain a default return structure,
select any table output parameter or scalar output parameter of type STRUCTURE. Default is
None.
Status Field Required if you select a value for the return structure. Select a field from the structure for
status. If you select None for the return structure, this parameter is blank.
Text Field Required if you select a value for the return structure. Select a field from the structure for
status messages. If you select None for the return structure, this parameter is blank.
Status Indicator For Enter an indicator message for warning. If you select None for the return structure, the value is
Warning W. Default is W.
Status Indicator for Enter an indicator message for error. If you select None for the return structure, the value is E.
Error Default is E.
Status Indicator for Enter an indicator message for abort. If you select None for the return structure, the value is A.
Abort Default is A.
A BAPI/RFC transformation includes multiple groups. When a BAPI function contains a nested structure,
ports for the input and output groups in a BAPI/RFC transformation use the following naming convention:
<group_name>_<parameter_name>_<field name>
For example:
SCALAR_INPUT_PARAM1_FIELD1
If there are multiple input or output structures, the BAPI/RFC transformation includes each structure
parameter name in the port names. For example, BAPI Z_TST2 has the parameter INPUT1, which is of the
type ZTYPE1. ZTYPE1 has several components such as FIELD1 and FIELD2. FIELD2 is a component of the
type structure. It contains field F1. The naming convention in the BAPI/RFC transformation for FIELD 1 is:
SCALAR_INPUT_INPUT1_FIELD1
The naming convention in the BAPI/RFC transformation for the field F1 is:
SCALAR_INPUT_INPUT1_FIELD2_F1
System variables
SAP uses system variables to set default values for some BAPI import parameters. The variables provide
information, such as current date and time for the operating system on which SAP runs. System variables
start with “SY-”. For example, SY-DATLO represents the local date of the SAP system.
The Secure Agent provides values for some system variables to define default input values for BAPI/RFC
parameters. The Secure Agent uses the values as default input values for some ports of BAPI/RFC
transformations. The Secure Agent uses the default values when there is no input for a port or when the port
is not connected to an upstream transformation or source.
SY-DATUM Local date on the Security Agent machine processing the data.
SY-UZEIT Local time on the Security Agent machine processing the data.
SY-HOST SAP host name from the SAP application connection properties.
When you run a mapping with a BAPI/RFC mapplet, the Secure Agent makes a BAPI/RFC call to SAP to
process the data. By default, the BAPI/RFC mapplet commits every 1000 rows when you run a mapping.
The BAPI/RFC call is based on the input data of the Integration ID ports. The Secure Agent makes one call to
SAP for each Integration ID. Pass a value to the Integration ID ports in the scalar input group and all
mandatory table input groups of the BAPI/RFC mapplet.
Note: You must map the Integration ID input field even when a BAPI or RFC does not require other input
fields.
If the BAPI/RFC call fails or if there is a data conversion error, SAP passes the data for the integration ID in
comma-separated format to an error output group. If the mapping contains a target instance that is
connected to the error output group, the Secure Agent writes the data to the target.
Create a target transformation with a column of the String data type and precision of 65535. Connect the
column in the target object to the Error Output Group in the BAPI/RFC mapplet. The Secure Agent writes the
error output data up to 65,535 characters to the target in comma-delimited format. If the error output data is
longer than 65,535 characters, the Secure Agent truncates the data.
• Pass a value to the Integration ID ports in the scalar input group and all mandatory table input groups of
the BAPI/RFC transformation.
• Add a target object if you want to receive BAPI/RFC function call errors from the BAPI error group.
Use the following guidelines when passing data to BAPI/RFC function input parameters:
• When the function input parameter data type is INT1 or NUMC, provide positive values for the function
input.
• When the source input data for a BAPI/RFC function is of the integer data type, do not use string data in
the source transformation. Otherwise, the mapping fails.
• If the input data for a BAPI/RFC function mapping contains a higher scale than the SAP metadata
specification, the Secure Agent rounds the data to comply with the SAP metadata. When you run a
mapping in high precision mode, the mapping can fail due to overflow if the round-off data cascades to
the precision digits. For example, the data type and precision for a BAPI/RFC function parameter is DEC
(6,5). The input data that you pass to the function parameter is 9.99999. When the Secure Agent
processes the input data, it rounds the input data to 10, which is not compatible with the SAP metadata.
The mapping fails.
• When you run an SAP BAPI mapping, even though the mapping runs successfully, the system log of the
SAP system displays the CPIC error.
To avoid the CPIC error at the SAP system, see KB article 000176711.
For example, to update sales order data in SAP, generate a BAPI/RFC mapplet from the
BAPI_SALESORDER_CHANGE function and configure a mapping using the mapplet.
You can configure a mapping with a BAPI/RFC mapplet to pass input data to BAPI/RFC function input
parameters.
In advanced mode, the Mapping Designer updates the mapping canvas to include transformations and
functions that enable advanced functionality.
To access the BAPI/RFC functionality through Data Integration, perform the following tasks:
For more information about configuring SAP BAPI Connector as a business service within a mapping or a
mapping task, see the How-To Library article,
How to Configure SAP BAPI Connector as a Business Service in Cloud Data Integration.
1. To create a mapplet, click Data Integration > New > Mapplets > Mapplet - SAP BAPI and then click
Create.
To edit a mapplet, on the Explore page, navigate to the mapplet. In the row that contains the mapplet,
click Actions and select Edit.
2. Enter a unique name for the BAPI/RFC mapplet.
3. Optionally, enter a description for the BAPI/RFC mapplet that you want to generate.
56
4. Select the transformation type as BAPI.
5. Select the BAPI/RFC connection that you want to use.
6. Select the scope of the transformation.
For real-time processing, select the transformation scope as Transaction. For batch processing and
validation in a non-production environment, select the transformation scope as All Input.
7. Enter the name of the BAPI/RFC function.
You can also search for a BAPI/RFC function by entering % as a wildcard character along with the search
string. For example, to search for all BAPI/RFC functions whose names start with the term Customer,
enter the search string as Customer%. To search for all BAPI/RFC functions whose names end with the
term Customer, enter the search string as %Customer.
8. Click Get Objects.
The Select Source Object dialog box appears displaying the search results. You can drill down the
search results further by entering a search string in the Select Source Object dialog box.
9. Click the required BAPI/RFC function name and click Select.
The import, export, and table parameter details of the BAPI/RFC function appear.
10. Specify the direction for the tables to indicate if the table parameters in the BAPI are input, ouput, or
both.
11. Click Save.
The mapplet for the specified BAPI/RFC function is created. You can create a mapping for the generated
mapplet.
Step1: Generate a BAPI/RFC mapplet in Data Integration for the bapi_salesorder_createfromdat1 BAPI function.
Log in to Data Integration and generate a BAPI/RFC mapplet that contains the metadata of the
bapi_salesorder_createfromdat1 BAPI function.
1. Specify source objects to enter the order header data from the ORDER_HEADER_IN structure, the
partner data using the ORDER_PARTNERS table, and item data using the ORDER_ITEMS_IN table as
input parameters.
1. To create a mapplet, click Data Integration > New > Mapplets > Mapplet - SAP and then click Create.
To edit a mapplet, on the Explore page, navigate to the mapplet. In the row that contains the mapplet,
click Actions and select Edit.
The New SAP Mapplet page appears. The following image shows the New SAP Mapplet page:
6. Add order header, order items, and order partners as flat file source objects for the mapping.
7. To add a BAPI/RFC mapplet transformation, on the Transformation palette, click Mapplet.
a. On the General tab, enter a name and description for the mapplet.
b. Draw a link to connect the flat file Source transformation to the Mapplet transformation.
c. On the Mapplet tab, click Select.
The Select Mapplet dialog box appears.
d. Select the bapi_salesorder_createfromdat1 mapplet and click OK.
e. Click Connection to specify an SAP RFC/BAPI Interface connection.
The following image shows the bapi_salesorder_createfromdat1 mapplet details:
f. To preview fields, configure the field rules, or rename fields, click Incoming Fields.
64
Chapter 9
IDocs electronically exchange data between SAP applications or between SAP applications and external
programs. IDoc is a message-based integration interface that processes data asynchronously.
IDoc is a component of Application Link Enabling (ALE) module within SAP that can send and receive
Intermediate Documents (IDocs) over RFC protocol.
ALE Layers
The message-based architecture of ALE comprises three layers:
• Application layer that provides ALE an interface to SAP to send or receive messages from external
systems.
• Distribution layer that filters and converts messages to ensure that they are compatible between different
SAP releases.
• Communications layer that enables ALE to support synchronous and asynchronous communication. You
use IDocs for asynchronous communication.
The architecture of ALE provides a way to send IDocs as text files without connecting to a central database.
Applications can communicate with each other without converting between formats to accommodate
hardware or platform differences.
ALE Components
ALE has the following components:
Component Description
Logical System All systems that need to communicate using ALE/IDoc must be setup as a Logical System within
SAP. An SAP administrator can setup logical systems in transaction BD54.
Distribution Model Defines an agreement between two logical systems on the messages that can be exchanged and
identifies the sender and the receiver. An SAP administrator can setup distribution models in
transaction BD64.
65
Component Description
Partner Profile Stores the IDoc type and processing logic related to the distribution model. An SAP administrator
can setup partner profiles in transaction WE20.
RFC Destination Defines the protocol and access to the logical system. An SAP administrator can setup RFC
destinations in transaction SM59.
IDoc Type Representation of a message type. SAP uses IDoc types to support backward compatibility
across various SAP releases.
IDoc mapplets
You can import an IDoc as a mapplet using the Mapplet - SAP IDoc asset. An IDoc contains a hierarchical
structure consisting of segments. Each segment is an SAP structure defined in the SAP system.
An IDoc has header and data record components. The header component contains control information, such
as creation date and status. The control information is in an SAP structure called EDIDC. The data records
are in an SAP structure called EDIDD.
IDoc segments are organized into groups. The following rules determine the group to which a segment
belongs:
• A parent segment starts a new group. For example, in the MATMAS04 IDoc, the E1MARCM segment
contains a child and therefore starts a group.
• A child segment that is not a parent belongs to the group that is started by its immediate parent. For
example, in the MATMAS04 IDoc, the E1MARA1 segment does not contain a child and therefore belongs
to the group of its parent E1MARAM.
• A group can also be a parent or a child.
67
IDocs properties
When you fetch an IDoc in Mapplet - SAP IDoc, you can view all the IDoc segments. Select a segment to view
the fields in the segment.
IDocs Properties
The following table describes the IDocs properties that you can view and specify in the Mapplet - SAP IDoc
asset:
Property Description
Transformation Indicates the type of transformation you want to use to generate the IDoc mapplet.
Type Select one of the following options:
- IDoc Prepare. Select to generate a mapplet that writes source data as an IDoc message.
- IDoc Interpreter. Select to generate a mapplet that reads IDoc messages.
Transformation Indicates how the Secure Agent applies the transformation logic to incoming data.
Scope Select one of the following options:
- Transaction. Select Transaction to apply the transformation logic to all rows in a
transaction. Select Transaction when the results of the transformation depend on all rows in
the same transaction, but not on rows in other transactions. When you select Transaction,
associated mappings can run in real time.
- All Input. Select All Input to apply the transformation logic to all incoming data, and to drop
the incoming transaction boundaries. Select All Input when the results of the transformation
depend on all rows of data in the source.
Default is All Input.
Message Type The type of application messages that classify categories of data.
For example, ORDERS and MATMAS (Material Master).
Basic IDoc Type The type of data structure associated with the message type.
For example, MATMAS01, MATMAS02 for MATMAS. IDocs contain the data associated with the
message type.
IDoc Document Displays the IDoc segments associated with the message type and IDoc type.
Control Record Displays the control record segments in the IDoc, their values, and precision.
The Designer provides values for some of the segments when you create a mapplet using the
IDoc prepare transformation. You can also provide values for other segments.
Segment Definition
You can view the segment field details from the segment definition section. To view the field names of the
segment, select a segment name.
Field Description
To import IDoc metadata from a local file, run the RSEIDoc3 program from the SAP client to generate the
metadata. When you run the program, select the IDoc type and range for the IDoc metadata you want to
generate. The program exports the metadata it generates into a metadata file. For example, you can export
the metadata to a file with the .idc extension. You can then use the metadata file to import the metadata into
the Designer for use in the SAP/ALE IDoc transformation.
Outbound mapplet
You can capture changes to the master data or transactional data in the SAP application database in real
time.
When data in the application database changes, the SAP system creates IDocs to capture the changes and
sends the IDocs to Data Integration. You can use the IDoc Reader connection to read the IDoc message in
real time as they are generated by the SAP system.
If the Secure Agent is not running when the SAP system sends outbound IDocs, the Secure Agent does not
receive the IDocs. However, the SAP system stores the outbound IDocs in EDI tables, which are a staging
area for guaranteed message delivery. You can configure the SAP system to resend the IDocs by configuring
the tRFC port used to communicate with the Secure Agent. When you configure the port, you can enable
background processes in SAP that try to resend the IDocs to the Secure Agent a set number of times.
To generate the outbound mapplet to read IDoc messages from SAP system, use the IDoc Interpreter when
you import the IDoc metadata. Import the outbound mapplet to Data Integration and configure an outbound
mapping.
Use an inbound SAP IDoc mapping to send the transactional data from the legacy application database to the
SAP system. Data Integration extracts the data from the legacy application data source, prepares the data in
SAP IDoc format, and sends the data to the SAP system as inbound IDocs using ALE. You can use the IDoc
Writer connection to write inbound SAP IDoc messages to SAP systems.
To generate the inbound mapplet to write IDocs to SAP systems, use the Prepare transformations when you
import the IDoc metadata. Import the inbound mapplet to Data Integration and configure an inbound
mapping.
Note: You can enable control record fields when you import the IDoc metadata.
The Prepare transformation in the SAP IDoc Writer mapplet can have primary key and foreign key fields and
other input fields, including control record fields.
When you import IDoc metadata, you can add fields to an IDoc Prepare transformation. In Mapplet - SAP
IDoc, on the Control Record tab, you can add the control record keys as key-value pairs. For example, you can
add the following key-value pairs for the sender partner type and the sender partner number:
In an inbound mapping, you can pass the sender partner number to SAP. You can pass a value to the
CONTROL_SNDPRN port in the control input group of the Prepare transformation. If you do not connect this
port to an upstream transformation, the Secure Agent uses the partner number value of SNDPRN key you
specify in the Mapplet - SAP IDoc asset.
To maintain the structure of the IDoc data, the Prepare transformation in the SAP IDoc Writer mapplet uses
primary and foreign keys. The top-level parent segment has a primary key. Each child segment has a primary
key and a foreign key. The foreign key of each child segment references the primary key of its parent
segment. For example, the foreign key of a second-level child segment references the primary key of the top-
level parent segment. Similarly, the foreign key of a third-level child segment references the primary key of
the second-level child segment.
The Prepare transformation groups incoming IDoc data based on the values in the primary and foreign key
fields. The Control Input group of the Prepare transformation represents the parent segment. All other groups
of the Prepare transformation except the ErrorIDocData group represent second-level or third-level child
segments. The ErrorIDocData group is used for processing invalid IDocs.
Inbound mapplet 71
The following table shows the groups of the Prepare transformation and the fields used for the primary and
foreign keys:
Each value for the GPK_<name> field needs to be unique. Each GFK_<parent_name>_<group_name> field
needs to reference the primary key of its parent segment.
For example, the following table shows the relationship of primary and foreign keys in an IDoc message
named ABSEN1 with four child segments:
CONTROL_INPUT_ABSEN1 GPK_DOCNUM P1
E2ABSE1 GPK_E2ABSE1 C1
- GFK_DOCNUM_E2ABSE1 P1
E2ABSE2 GPK_E2ABSE2 C2
- GFK_DOCNUM_E2ABSE2 P1
- GFK_E2ABSE2_E2ABSE2A C2
E2ABSE3 GPK_E2ABSE3 C3
- GFK_DOCNUM_E2ABSE3 P1
- GFK_E2ABSE2_E2ABSE2A C3
E2ABSE4 GPK_E2ABSE4 C4
- GFK_DOCNUM_E2ABSE4 P1
The Prepare transformation uses these primary and foreign key relationships to maintain the structure of the
IDoc data. Any foreign key field that does not match the primary key of its parent segment results in an
orphan row. Any primary key field that is not unique results in a duplicate row.
Verify that each IDoc message has a unique primary key for the top-level parent segment, each child
segment, and that each foreign key matches the primary key of its parent.
1. In Data Integration, click New > Mapplets > Mapplet - SAP IDoc, and then click Create.
2. Enter a unique name for the mapplet.
3. Specify the location where you save the mapplet.
4. Optionally, enter a description for the mapplet.
5. Select the transformation type. You can select one of the following options:
• Select IDoc Interpreter to generate outbound mappings to read IDocs from the SAP system.
• Select IDoc Prepare to generate inbound mappings to write IDocs to the SAP system.
6. Select the source by which you fetch the IDoc. You can select one of the following options:
• Select Connection to fetch the IDoc using an SAP Connection.
• Select Local File to fetch the IDoc using a file from your local drive.
7. Select the scope of the transformation. You can select one of the following options:
• Select Transaction to apply the transformation logic to all rows in a transaction. Select Transaction
when the results of the transformation depend on all rows in the same transaction, but not on rows in
other transactions. When you select Transaction, associated mappings can run in real time. For
outbound mappings, select Transaction.
• Select All Input to apply the transformation logic to all incoming data, and to drop the incoming
transaction boundaries. Select All Input when the results of the transformation depend on all rows of
data in the source.
8. Enter the message type and the basic IDoc type, and then click Fetch IDoc.
The segment details of the IDoc appears.
9. On the IDoc Document tab, in the Select IDoc Segments section, select the segments you want to
include in the mapplet.
ALE is an SAP proprietary technology that enables data communication between SAP systems. ALE also
enables data communication between SAP and external systems.
You can configure outbound mappings to read IDocs from SAP and inbound mappings to write IDocs to SAP.
In advanced mode, the Mapping Designer updates the mapping canvas to include transformations and
functions that enable advanced functionality.
To configure an outbound mapping to read IDocs from SAP, perform the following tasks:
1. Configure an IDoc mapplet using the Mapplet - SAP IDoc asset from Data Integration. Verify that you
selected the segments and groups you want to include in the IDoc. In addition, select the IDoc
Interpreter transformation.
2. Configure an outbound mapping using the generated IDoc mapplet. Add a Source transformation to read
data from the SAP system, add the generated IDoc mapplet, and add a Target transformation to write the
IDoc in the target object.
To configure an inbound mapping to write IDocs to SAP, perform the following tasks:
1. Create an IDoc mapplet using the Mapplet - SAP IDoc asset from Data Integration. Verify that you
selected the segments and groups you want to include in the IDoc. In addition, select the IDoc Prepare
transformation.
2. Configure an inbound mapping using the generated IDoc mapplet. Add a Source transformation to read
data from the source system, add the generated IDoc mapplet, and add a Target transformation to write
the IDoc to the SAP system.
Specify the name and description of the IDoc Reader source. Configure the source and advanced properties
for the source object.
75
The following table describes the IDoc Reader source properties that you can configure in a Source
transformation:
Property Description
Source Type Source type. Select Single for a single source object.
When you select an SAP IDoc Reader as connection, the source type can be a single object and source
object is the IDoc Reader Object. The source object has the generic structure of an IDoc message.
The following table describes the SAP IDoc Reader advanced source properties:
Property Description
Idle Time Indicates the number of seconds the Secure Agent waits for IDocs to arrive before it stops
reading from the SAP source. For example, if you enter 30 seconds for idle time, the Secure
Agent waits 30 seconds after reading from the SAP source. If no new IDocs arrive within 30
seconds, the Secure Agent stops reading from the SAP source. Default is 300.
Packet Count Controls the number of packets the Secure Agent reads from SAP before stopping. For example,
if you enter 10 for Packet Count, the Secure Agent reads the first 10 packets from the SAP
source and then stops. The packet Size property in the ALE configuration determines the number
of IDocs the Secure Agent receives in a packet.
If you enter packet count as -1, you can read infinite number of packets. Default is -1.
Realtime Flush Determines, in seconds, how often the Secure Agent flushes data from the source.
Latency
Reader Time Limit Sets a period of time, in seconds, during which the Secure Agent reads IDocs from the SAP
source. For example, if you specify 10 as the reader time limit, the Secure Agent stops reading
from the SAP source after 10 seconds.
If you enter reader time limit as 0, the Security Agent continues to read IDocs from SAP for an
infinite period of time. Default is 0.
Number of Retries Determines the number of times Cloud Data Integration to attempt to connect to the SAP system.
for Connection Default is 1.
Resiliency
Delay between Determines the time interval in seconds between the connection retries to connect to the SAP
Retries for system if it could not connect on a previous attempt.
Connection Default is 5.
Resiliency
Tracing Level Sets the amount of detail that appears in the log file. You can choose terse, normal, verbose
initialization or verbose data. Default is normal.
Specify the name and description of the IDoc Writer target. Configure the target and advanced properties for
the target object in a Target transformation.
The following table describes the IDoc Writer target properties that you can configure in a Target
transformation:
Property Description
The following table describes the SAP IDoc Writer advanced target properties:
Property Description
Packet Size Number of IDocs that you want to send in a packet to SAP.
Send IDocs Based On The commit method of IDocs to SAP based on the value you set for the Packet Size property
or commit properties at every commit point.
Select one of the following options:
- Packet Size. Commits IDocs to SAP based on the value you set for the Packet Size
property. Data Integration collects IDoc messages until the total count reaches the packet
size. It then sends messages as a packet to SAP.
- Commit Call. Commits IDocs to SAP based on the commit properties at every commit point.
Forward Rejected Determines whether the transformation passes rejected rows to the next transformation or
Rows drops rejected rows. By default, the agent forwards rejected rows to the next transformation.
1. In Data Integration, click New > Mapplets > Mapplet - SAP IDoc, and then click Create.
2. Enter a unique name for the mapplet.
3. Specify the location where you save the mapplet.
1. To create a mapping, click Data Integration > New > Mappings. Select Mapping and click Create.
2. Enter a name and description for the mapping, and click OK.
You can use alphanumeric characters and underscores (_) in the mapping name.
3. To configure an SAP source, on the Transformation palette, click Source.
4. In the Properties panel, on the General tab, enter a name and description.
5. Click the Source tab and select an SAP IDoc Reader connection.
When you select an SAP IDoc Reader as connection, the source type is a single object and source object
is the IDoc Reader Object.
You can also parameterize the SAP IDoc Reader connection.
6. If required, configure the advanced source properties.
1. To create a mapping, click Data Integration > New > Mappings. Select Mapping and click Create.
2. Enter a name and description for the mapping, and click OK.
You can use alphanumeric characters and underscores (_) in the mapping name.
3. To configure a source, on the Transformation palette, click Source.
4. In the Properties panel, on the General tab, enter a name and description.
5. Click the Source tab and configure the source details.
In this example, to read the MATMAS IDocs, perform the following steps:
1. In Data Integration, click New > Mapplets > Mapplet - SAP IDoc, and then click Create.
2. Enter a unique name for the mapplet.
3. Specify the location where you save the mapplet.
4. Optionally, enter a description for the mapplet.
5. Select IDoc Interpreter as the transformation type.
6. From the Connection list, select the source connection by which you fetch the IDoc.
7. From the Transformation Scope list, select Transaction.
8. In the Message Type field, enter MATMAS.
9. In the Basic IDoc Type field, enter MATMAS04.
10. Click Fetch IDoc.
The segment details of the IDoc appears in the Select IDoc Segments section.
11. On the IDoc Document tab, in the Select IDoc Segments section, select the segments you want to
include in the mapplet.
12. Click Save.
1. To create a mapping, click Data Integration > New > Mappings. Select Mapping and click Create.
2. Enter a name and description for the mapping, and click OK.
3. To configure an SAP source, on the Transformation palette, click Source.
4. In the Properties panel, on the General tab, enter a name and description.
5. Click the Source tab and select an SAP IDoc Reader connection.
When you select an SAP IDoc Reader as connection, the source type is a single object and source object
is the IDoc Reader Object.
13. Click Field Mappings to map the incoming IDoc Record field with the IDocData field in the mapplet.
14. To add a Target transformation, on the Transformation palette, click Target.
15. On the General tab, enter a name and description.
16. Click the Target tab, and configure the required target database details.
Note: When you use IDoc Listener as a source, you can specify only those targets that honor commit.
Contact Global Customer Support for the list of supported targets.
17. Draw a link to connect the mapplet groups to the target database object.
18. Save and run the mapping.
In this example, to write MATMAS IDoc to SAP, perform the following steps:
1. In Data Integration, click New > Mapplets > Mapplet - SAP IDoc, and then click Create.
2. Enter a unique name for the mapplet.
3. Specify the location where you save the mapplet.
4. Optionally, enter a description for the mapplet.
5. Select IDoc Prepare as the transformation type.
6. From the Connection list, select the source connection by which you fetch the IDoc.
7. From the Transformation Scope list, select All Input.
8. In the Message Type field, enter MATMAS.
9. In the Basic IDoc Type field, enter MATMAS03.
11. On the IDoc Document tab, in the Select IDoc Segments section, select the segments you want to
include in the mapplet.
12. Optionally, to add a control record field to the mapplet, on the Control Record tab, select the checkbox
for the field you want to add. This enables you to map the selected control record fields when you
configure a mapping, integration template, or mapping task.
13. Click Save.
1. To create a mapping, click Data Integration > New > Mappings. Select Mapping and click Create.
2. Enter a name and description for the mapping, and click OK.
3. To configure a flat file source, on the Transformation palette, click Source.
4. In the Properties panel, on the General tab, enter a name and description.
5. Click the Source tab and select an flat file connection.
6. Add separate flat file sources for the control_input group and segments in the MATMAS IDoc.
7. To add an IDoc Mapplet transformation, on the Transformation palette, click Mapplet.
8. On the General tab, enter a name and description for the mapplet.
9. On the Mapplet tab, click Select.
The Select Mapplet dialog box appears.
10. Specify an IDoc mapplet that you imported using the Prepare transformation and click OK.
11. Draw a link to connect the Source transformations to the Mapplet transformation. For example, connect
the Control_Input source object to the Control_Input_Group for the MATMAS IDoc.
13. Click Field Mappings to map the incoming fields with the mapplet input fields.
The following image shows the field mapping in the mapplet:
14. To add a Target transformation for writing to SAP, on the Transformation palette, click Target.
a. On the General tab, enter a name and description.
b. Click the Target tab and configure SAP target details.
You can also parameterize the SAP target details.
15. To add a flat file Target transformation for the error output, on the Transformation palette, click Target.
a. On the General tab, enter a name and description.
b. Click the Target tab and configure flat file target details.
87
Chapter 12
Use the SAP BW Connector connection to read data from SAP BW objects such as InfoCubes, InfoSets,
MultiProviders, and DataStore objects.
You can use the SAP BW objects as sources in synchronization tasks, mappings, and mapping tasks.
88
Chapter 13
You can use SAP BW objects as sources. You can read data from InfoCubes, InfoSets, MultiProviders, and
DataStore objects. You can use expressions to transform the data according to your business logic and use
data filters to filter data before writing it to targets.
Note: You cannot perform a lookup on an SAP BW object.
When you create a task, you can associate it with a schedule to run it at specified times or at regular
intervals. You can also manually run a task. You can monitor tasks and view logs about completed tasks.
Property Description
Connection SAP BW Reader connection that you want to use to read data.
Source Type Source type. You can read data from a single SAP BW object.
You cannot edit this field.
Source Object SAP BW source object from which you want to read data.
Display technical field When selected, displays technical names instead of business names of the fields in
names instead of labels the specified source object.
Display source fields in When selected, displays source fields in alphabetic order. By default, fields appear in
alphabetical order the order returned by the source system.
Data Preview Displays the first 10 rows of the first five columns in the object, and the total number
of columns in the object.
89
You can also configure advanced source properties when you schedule the synchronization task. Advanced
source properties appear on the Schedule page of the Synchronization Task wizard.
Property Description
Packet Size in MB Size of the HTTP packet that SAP sends to the Secure Agent. The unit is MB.
Default is 10 MB.
Package Size in ABAP in Number of rows that are read and buffered in SAP at a time.
Rows Default is 1000 rows.
Enable Compression When selected, the ABAP program compresses the data in the gzip format before it sends
the data to the Secure Agent.
If the Secure Agent and the SAP system are not on the same network, you might want to
enable the compression option to optimize performance.
Default is not selected.
In Monitor, you can monitor the status of the logs after you run the task.
You can also monitor the progress of the task by calling Transaction SM37 from SAP. You can view the
actual job duration in SAP. The job duration listed in the Data Integration activity log is a higher value
because it also includes time required to complete processing in Data Integration.
You can view the HTTP and HTTPS log files in the SMICM transaction. Optionally, you can increase trace
level to 3 to view the detailed logs.
You can create a mapping to read data from an SAP BW source. You cannot perform a lookup on an SAP BW
object. In advanced mode, the Mapping Designer updates the mapping canvas to include transformations
and functions that enable advanced functionality.
When you configure a mapping to describe the flow of data from a source to a target, you can also add
transformations to transform data. A transformation includes field rules to define incoming fields. Links
visually represent how data moves through the data flow.
After you create a mapping, you can run the mapping or you can deploy the mapping in a mapping task. You
can use the Mapping application to process data based on the data flow logic that you defined in the
mapping.
Use the Mapping Task wizard to create a mapping task. When you create a mapping task, you can select the
mapping based on which you want to create the task.
If you configured parameters, which are placeholders for information, in a mapping, you can define the
parameters in the mapping task. You can define parameters for additional flexibility and to use the same
mapping in multiple mapping tasks. For example, you can use a parameter for a source connection in a
mapping, and then define the source connection when you configure the mapping task.
When you create a mapping task, you can associate the task with a schedule to run it at specified times or at
regular intervals. Or, you can run it manually. You can also configure advanced session properties. You can
monitor tasks and view details about completed tasks.
Specify the name and description of the SAP BW source. Configure the source and advanced properties for
the source object.
93
The following table describes the SAP BW source properties that you can configure in a Source
transformation:
Property Description
Connection SAP BW Reader connection that you want to use to read data.
Object SAP BW source object from which you want to read data.
Property Description
Packet Size in MB Size of the HTTP packet that SAP sends to the Secure Agent. The unit is MB.
Default is 10 MB.
Package Size in ABAP Number of rows that are read and buffered in SAP at a time.
in Rows Default is 1000 rows.
Enable Compression When selected, the ABAP program compresses the data in the gzip format before it sends
the data to the Secure Agent.
If the Secure Agent and the SAP system are not on the same network, you might want to
enable the compression option to optimize performance.
Default is not selected.
Tracing Level Amount of detail that appears in the log file. You can select one of the following options:
- Terse
- Normal
- Verbose Initialization
- Verbose Data
Default is normal.
1. To create a mapping, click Data Integration > New > Mappings. Select Mapping and click Create.
2. Enter a name and description for the mapping, and click OK.
You can use alphanumeric characters and underscores (_) in the mapping name.
3. To configure an SAP BW source, on the Transformation palette, click Source.
4. In the Properties panel, on the General tab, enter a name and description for the SAP BW source.
5. Click the Source tab and configure the SAP BW source details.
1. To create a mapping task, click Data Integration > New > Tasks and then complete one of the following
steps:
• To create a mapping task based on a mapping, select Mapping Task and click Create.
• To create a mapping task using a template, expand the appropriate template category and select the
template you want to use, and then click Create.
To edit a mapping task, on the Explore page, navigate to the mapping task. In the row that contains the
task, click Actions and select Edit.
2. Enter a name for the task.
Task names must be unique within the organization. Task names are not case sensitive. Task names
can contain alphanumeric characters, spaces, and the following special characters: _ . + -
3. Select the runtime environment that contains the Secure Agent that you want to use to access the SAP
BW objects.
4. Select Mapping as the task based on which you want to create the mapping task.
5. Click Select to specify a mapping.
The Select a Mapping dialog box appears.
6. Select a mapping. You can also search for a mapping based on its name and description.
7. Click OK.
The image of the selected mapping appears.
8. Click Next.
If you specified any parameters in the source or target details in the mapping, the Source or Target page
appears. If not, the Schedule page appears.
9. Configure a schedule and advanced options. Perform any of the following steps based on your
requirements:
a. Click Run this task on schedule and specify the schedule that you want to use.
In Monitor, you can monitor the status of the logs after you run the task.
You can also monitor the progress of the task by calling Transaction SM37 from SAP. You can view the
actual job duration in SAP. The job duration listed in the Data Integration activity log is a higher value
because it also includes time required to complete processing in Data Integration.
You can view the HTTP and HTTPS log files in the SMICM transaction. Optionally, you can increase trace
level to 3 to view the detailed logs.
• When you establish a communication from the SAP system with the Secure Agent using the IP address of
the NAT gateway, you must include the DTM property named SapTableReaderNatIpAddress for the Secure
Agent and specify the NAT IP address as the value.
• When you create a mapping that contains an InfoObject with the Attribute Only option selected, the
mapping fails.
98
Chapter 15
ADSO is the central object for data storage and consolidation in the SAP BW system. ADSO is the primary
data persistency object in BW/4HANA that simplifies data modeling and the support environment. ADSOs are
commonly stored in 2-dimensional transparent database tables.
99
Chapter 16
SAP Advanced DataStore Object (ADSO) is the central object for data storage and data consolidation in the
SAP BW system. ADSO is the primary persistent object in BW/4HANA and simplifies the modeling and
support environment. ADSOs are commonly stored in 2-dimensional transparent database tables.
You can use SAP Table Reader Connector in Data Integration to read large volumes of data from the SAP
BW/4HANA application.
• ADSO name
• Description
• Type
• Info area
• Transfer structure
• Field information such as the name, type, description, precision, scale, SAP type, SAP length, SAP output
length, and SAP decimals.
The naming convention of an ADSO is: /BIC/A<technical_name of ADSO>*, where * represents the table
number associated with an ADSO.
When you import an ADSO into Cloud Data Integration, the Secure Agent imports the following five tables
associated with each ADSO:
100
• /BIC/A<technical_name of ADSO>7. View for reporting for datastore <technical_name of ADSO>.
• /BIC/A<technical_name of ADSO>8. View for external access (introduced in BW/4HANA 2.0
<technical_name of ADSO>.
The following image shows the five tables of an ADSO with technical name ZDELTA that you can import from
BW/4HANA 1.0 into Cloud Data Integration:
SAP Advanced DataStore Object (ADSO) is the central object for data storage and data consolidation in the
SAP BW system. ADSO is the primary persistent object in BW/4HANA and simplifies the modeling and
support environment. ADSOs are commonly stored in 2-dimensional transparent database tables.
You can use SAP ADSO Writer Connector in Data Integration to load large volumes of data to the target ADSO
objects in the SAP BW/4HANA application.
• ADSO name
• Description
• Type
• Info area
• Transfer structure
• Field information such as the name, type, description, precision, scale, SAP type, SAP length, SAP output
length, and SAP decimals.
• Only the activated ADSOs appear in the browser list of the SAP ADSO Writer Connector. Ensure that the
required ADSO is activated in the SAP system before you can use in Data Integration.
• If the ADSO name has namespace (“/”) characters at the beginning, the Secure Agent replaces those
characters with T__. Every subsequent namespace character in the ADSO name is replaced by “__”. For
example, if the ADSO name is /IMO/CMSD17, it appears in the browser list as T__IMO__CMSD17.
• Do not edit the field metadata as it might corrupt the extracted data.
• If the field name of the selected ADSO has a namespace character (“/”), the Secure Agent replaces it with
an underscore “_” character.
• When you define a mapping, you must select the operation as Insert. While activating, SAP BW/4HANA
uses its algorithm to handle loaded data. You do not need to set any additional property in Data
102
Integration to perform update or delete operation. Operations such as update and delete are handled at
the SAP BW/4HANA layer itself.
• Ensure that data with the different data types that you load to BW/4HANA is correct. The activation of the
loaded packet might fail if the inserted data is incorrect.
• The Secure Agent does check the validity of the data according to the supported transformation data type.
However, if the data is incorrect, the Secure Agent loads the data to SAP BW/4HANA, but fails during the
data activation. Ensure that the data is valid according to the data type of the fields.
• The default packet size to write data is 8 MB. When you set the packet size value equal to greater than 8
MB and you encounter out-of-memory issues, you can increase the Java heap memory for the Secure
Agent. In the System Configuration Details section of the Secure Agent properties, select DTM and set the
JVMOption1 property to the following value: Xmx512m. Click OK to save the change and then restart the
Secure Agent. Adjust the value for the JVMOption1 property based on the amount of records you want to
write and the available memory on the Secure Agent machine.
• The default precision for variable length data types such as STRING and RAWSTRING is set to a higher
value of 1333. To manage the optimal record length, you must, however, set the length for these data
types in the Edit Metadata Section of the mapping based on the expected length of the actual data.
When you configure a mapping to describe the flow of data from source and target, you can also add
transformations to transform data. A transformation includes field rules to define incoming fields. Links
visually represent how data moves through the data flow.
After you create a mapping, you can run the mapping or you can deploy the mapping in a mapping task. The
Mapping Configuration application allows you to process data based on the data flow logic defined in a
mapping or integration template.
Use the Mapping task wizard to create a mapping task. When you create a mapping task, you select the
mapping or integration template for the task to use.
If you configured parameters, which are placeholders for information, in a mapping, you can define the
parameters in the mapping task. Defining parameters provides additional flexibility and allows you to use the
same mapping in multiple mapping tasks. For example, you can use a parameter for a target connection in a
mapping, and then define the target connection when you configure the mapping task.
When you create a mapping task, you can associate the task with a schedule to run it at specified times or on
regular intervals. Or, you can run it manually. You can also configure advanced session properties. You can
monitor tasks that are currently running and view details about completed tasks.
Specify the name and description of the SAP ADSO target. Configure the target and advanced properties for
the target object.
104
The following table describes the SAP ADSO Writer target properties that you can configure in a Target
transformation:
Property Description
Object Target object for the task. Click Select and then select the target object.
The following table describes the SAP ADSO Writer advanced target properties:
Property Description
Activate Option Loads the data into the target ADSO object and then sends data activation requests to SAP BW/
Mode 4HANA based on the activate options you select.
You can select from the following activate options:
- 0-Do Not Activate. Does not send an activate request.
- 1-Activate. Creates a single activation request for all load requests in the queue.
- 2-Activate Request by Request. Creates a separate activation request for each of the load
requests in the queue.
- Parameter. Use this option to parameterize the activate options. Specify a valid activate option
value in the Parameter Name for Activate Option Mode field.
Default is Do Not Activate.
Parameter Name The parameter name that you defined for the Activate Option Mode in the parameter file.
for Activate You can specify the following values for the data fetch mode:
Option Mode - 0 for Do Not Activate
- 1 for Activate
- 2 for Activate Request by Request
If you pass any other value, the jobs fail when you run the mapping,
Packet size in The packet size of each SAP data packet while writing data to SAP BW4/HANA.
MB Default value is 10 MB.
Tracing Sets the amount of detail that appears in the log file.
You can choose terse, normal, verbose initialization, or verbose data.
Default is normal
When you specify the activate mode, the Secure Agent first loads the data into the target ADSO object and
then sends the data activation requests to SAP BW/4HANA based on the following activate options you
select:
• 0-Do Not Activate. Does not send an activate request and the activation precessing does not occur.
• 1-Activate. Creates a single activation request for all load requests in the queue.
• 2-Activate Request by Request. Creates a separate activation request for each of the load requests in the
queue.
The activation request activates all the previous load requests in the queue that are pending for activation.
Activation occurs per load request in a chronological sequence, starting from the earliest pending request to
the most recent load request. The Secure Agent waits for the activation to be fully processed in SAP BW/
4HANA system. The wait time depends upon the number of load requests in the queue that were available
earlier to the load request associated with the current mapping.
During the activation process, if the activation of a specific load request fails, the whole activation process
fails and the rest of the load requests are also not activated. In such cases, the mapping fails even though
the Secure Agent loaded the data correctly based on the enabled activate mode.
Enable partitioning when you configure the Source transformation in the Mapping Designer. You do not have
to enable fixed partitioning in the Target transformation. The Secure Agent considers the same number of
partitions you specified for the source as the target partition number.
On the Partitions tab for the Source transformation, you select fixed partitioning and enter the number of
partitions, as shown in the following image:
When you run the mapping, the Secure Agent distributes rows of data for the target based on the number of
partitions that you specify in the Source transformation. You can specify up to 64 partitions.
106 Chapter 18: Mappings and mapping tasks with SAP ADSO Writer
Configuring a mapping with an SAP ADSO target
Use the Data Integration Mapping Designer to configure a mapping.
1. To create a mapping, click Data Integration > New > Mappings. Select Mapping and click Create.
2. Enter a name and description for the mapping, and click OK.
You can use alphanumeric characters and underscores (_) in the mapping name.
3. To configure a source, on the Transformation palette, click Source.
4. In the Properties panel, on the General tab, enter a name and description.
5. Click the Source tab and configure source details.
6. Specify the source type. You can choose one of the following options:
• Select Single Object to select a single SAP object.
• Select Parameter to configure the source objects in a mapping task associated with this mapping.
7. Click Query Options in the Source tab to specify any filter for the SAP ADSO object.
8. Click Advanced to specify the advanced source properties.
Note: You cannot add or remove source fields or edit the field metadata.
9. To add a transformation, on the Transformation palette, click the transformation name. Or, drag the
transformation onto the mapping canvas.
a. On the General tab, you can enter a name and description for the transformation.
b. Draw a link to connect the previous transformation to the transformation.
When you link transformations, the downstream transformation inherits the incoming fields from
the previous transformation.
For a Joiner transformation, draw a master link and a detail link.
c. To preview fields, configure the field rules, or rename fields, click Incoming Fields.
d. Configure additional transformation properties, as needed.
The properties that you configure vary based on the type of transformation you create.
e. To add another transformation, repeat these steps.
10. To add an ODSO object to the Target transformation, on the Transformation palette, click Target.
a. On the General tab, you can enter a name and description.
b. Draw a link to connect the previous transformation to the Target transformation.
c. Click the Target tab, configure the target details. Specify the target type:
• Select Single Object to select an ADSO object.
• Select Parameter to configure the target objects in a mapping task associated with this
mapping.
• Select Operation as Insert.
• Click Advanced and specify the advanced target properties.
d. Click Field Mapping and map the fields that you want to write to the target.
e. To add another Target transformation, repeat these steps.
11. Click the Field Mapping to map the Source and Target fields.
You can use the Automatic option if the fields are similar between the source and target.
12. Save and run the mapping or save and create a mapping task.
1. To create a mapping task, click Data Integration > New > Tasks.
2. Enter a name for the task.
Task names must be unique within the organization. Task names can contain alphanumeric characters,
spaces, and the following special characters:_ . + -Task names are not case sensitive.
3. Select the runtime environment that contains the Secure Agent that you want to use to access the SAP
ADSO Writer connection.
4. Select Mapping as the task based on which you want to create the mapping task.
5. Click Select to specify a mapping.
The Select a Mapping dialog box appears.
6. Select a mapping or search for the required mapping and select OK.
The image of the selected mapping appears.
7. Click Next.
If you specified any parameters in the source or target details in the mapping, the Source or Target page
appears. If not, the Schedule page appears.
8. Click Next to configure a schedule and advanced options. Perform any of the following steps based on
your requirements.
a. Click Run this task on schedule and specify the schedule you want to use.
b. Configure the email notification options.
c. Configure advanced options for the task.
d. Configure the advanced target properties.
e. Specify the execution mode.
9. Optionally, add the advanced session properties.
a. Click Add.
b. Select a session property.
c. Configure the value of the session property.
10. Save and run the mapping task.
108 Chapter 18: Mappings and mapping tasks with SAP ADSO Writer
Appendix A
Native data types are data types specific to the source and target databases or flat files. They appear in
non-SAP sources and targets in the mapping.
SAP data types appear in the Fields tab for Source and Target transformations when you choose to edit
metadata for the fields. SAP performs any necessary conversion between the SAP data types and the
native data types of the underlying source database tables.
Set of data types that appear in the remaining transformations. They are internal data types based on
ANSI SQL-92 generic data types, which Data Integration uses to move data across platforms.
Transformation data types appear in all remaining transformations in a mapping, synchronization task,
or mapping task.
When Data Integration reads source data, it converts the native data types to the comparable transformation
data types before transforming the data. When Data Integration writes to a target, it converts the
transformation data types to the comparable native data types.
109
SAP Data Transformation Range for Transformation Data Type
Type Data Type
DATS Date/time Jan 1, 0001 A.D. to Dec 31, 9999 A.D. Precision to the nanosecond.
DF16_DEC Decfloat16 Range of 1-15 and scaling of maximum 14. Decimal floating point number
stored in BCD format.
You can use the DF16_DEC data type in the following scenarios:
- When you configure SAP BAPI Connector as a business service within a
mapping or a mapping task to integrate with SAP
- When you use SAP BAPI/RFC mapplets to integrate with SAP
DF34_DEC Decfloat34 Range of 1-31 and scaling of maximum 30. Decimal floating point number
stored in BCD format.
You can use the DF34_DEC data type when you configure SAP BAPI
Connector as a business service within a mapping or a mapping task to
integrate with SAP.
DF16_RAW Double Maximum of 16 positions with floating decimal. Decimal floating point
number stored in binary format.
You can use the DF16_RAW data type in the following scenarios:
- When you configure SAP BAPI Connector as a business service within a
mapping or a mapping task to integrate with SAP
- When you use SAP BAPI/RFC mapplets to integrate with SAP
DF34_RAW Double Maximum of 34 positions with floating decimal. Decimal floating point
number stored in binary format.
You can use the DF34_RAW data type in the following scenarios:
- When you configure SAP BAPI Connector as a business service within a
mapping or a mapping task to integrate with SAP
- When you use SAP BAPI/RFC mapplets to integrate with SAP
LRAW Binary Uninterrupted sequence of bytes with a maximum length of 255 positions.
PREC Binary Uninterrupted sequence of bytes with a maximum length of 255 positions.
RAW Binary Uninterrupted sequence of bytes with a maximum length of 255 positions.
TIMS Date/time Jan 1, 0001 A.D. to Dec 31, 9999 A.D. Precision to the nanosecond.
SAP Data Type Transformation Data Range for Transformation Data Type
Type
DATS Date/time Jan 1, 0001 A.D. to Dec 31, 9999 A.D. Precision to the nanosecond.
TIMS Date/time Jan 1, 0001 A.D. to Dec 31, 9999 A.D. Precision to the nanosecond.
A I
ADSO targets IDoc
rules and guidelines in mapping tasks 102 generating metadata file 69
IDoc mappings
overview 75
C
Cloud Application Integration community
L
URL 7 libraries
Cloud Developer community to read from SAP ADSO 28
URL 7 load balancing connection
connections entry in sapnwrfc.ini 16
SAP ADSO Writer 38, 39 logical system
configuration for SAP ADSO Writer connections 25 creating for SAP Connector 18
configuration for SAP IDoc and RFC/BAPI connections 13 defining SAP Connector 18
SAP ADSO Writer 38, 39 SAP, program ID 19
SAP ADSO Writer configuration requirements 26
SAP BW Reader 36
SAP configuration requirements 26
SAP IDoc Reader 35
M
SAP IDoc Writer 35 maintenance outages 8
SAP RFC/BAPI Interface 34 mapping tasks
creating 96
rules and guidelines for SAP ADSO targets 102
D Mapping tasks
creating 108
Data Integration community overview 104
URL 7 mappings
data types IDoc Reader sources 75
SAP 109 IDoc Writer targets 77
defining logical system overview 104
configuration for SAP IDoc connections 18 SAP ADSO targets 104
113
mapplet SAP BW Reader
outbound 70 connection properties 36
mapplets SAP BW Reader connection
BAPI/RFC 51 creating 44
inbound 71 SAP connections
metadata IDoc and BAPI/RFC 34
generating file for IDoc 69 Serverless runtime environment 30
SAP Connector
administration 30
O communication interfaces 11
integration methods 10
organization ID overview 10
finding 48 user authorizations 21, 24
organizations SAP Connector administration 13
finding your organization ID 48 SAP data extraction
outbound mappings using SAP BW Connector 88
configuring to read IDocs 78 SAP IDoc Reader
configuring to read IDocs example 81 connection properties 35
outbound mapplet SAP IDoc Reader connection
map error output 70 creating 43
ports 70 SAP IDoc Writer
connection properties 35
SAP iDoc Writer connection
P creating 43
SAP integration methods
Partitioning overview 10
fixed partitioning for SAP ADSO targets 106 using BAPI/RFC functions 50
partner profile using IDocs 65
creating outbound and inbound parameters 20 using SAP ADSO 99
SAP ALE integration 19 SAP libraries
program ID for SAP IDoc and RFC/BAPI connections 13, 14
ALE integration 19 SAP notes
SAP, logical system 19 to write to SAP ADSO 26
sapnwrfc.ini file 16 SAP RFC/BAPI connection
creating 43
SAP RFC/BAPI Interface
R connection properties 34
SAP targets
RFC destination ADSO 100, 102
creating tRFC port 19 sapnwrfc.ini
RFC server program connection connections 16
entry in sapnwrfc.ini 16 segments and groups
RSEIDoc3 overview 67
generating metadata file for IDoc 69 specific application server connection
rules and guidelines entry in sapnwrfc.ini 16
BAPI/RFC mapplets 55 status
Informatica Intelligent Cloud Services 8
Synchronization tasks
114 Index