Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Real Time Processing

Download as doc, pdf, or txt
Download as doc, pdf, or txt
You are on page 1of 18
At a glance
Powered by AI
The key takeaways are that real-time processing allows for immediate access to dynamic data sources like financial data. Real-time sources include messaging queues, web services, and change data capture. Properties like flush latency and commit type determine how often and when data is processed and committed. Products like PowerExchange and Web Services Provider can be used to process different real-time sources.

Some of the real-time data sources discussed include messages and message queues from providers like WebSphere MQ, JMS, MSMQ, SAP, TIBCO and webMethods. Web service messages and change data from PowerExchange change data capture sources are also mentioned.

Real-time properties like terminating conditions, flush latency, and commit type determine when the Integration Service processes and commits real-time data. Terminating conditions determine when reading from the source ends, flush latency determines how often data is flushed, and commit type determines when data is committed to the target.

Real-time Processing

By PenchalaRaju.Yanamala

This chapter includes the following topics:

Overview
Understanding Real-time Data
Configuring Real-time Sessions
Terminating Conditions
Flush Latency
Commit Type
Message Recovery
Recovery File
Recovery Table
Recovery Queue and Recovery Topic
Recovery Ignore List
Stopping Real-time Sessions
Restarting and Recovering Real-time Sessions
Rules and Guidelines
Real-time Processing Example
Informatica Real-time Products

Overview

This chapter contains general information about real-time processing. Real-time


processing behavior depends on the real-time source. Exceptions are noted in
this chapter or are described in the corresponding product documentation. For
more information about real-time processing for your Informatica real-time
product, see the Informatica product documentation. For more information about
real-time processing for PowerExchange, see PowerExchange Interfaces for
PowerCenter.

You can use PowerCenter to process data in real time. Real-time processing is
on-demand processing of data from real-time sources. A real-time session reads,
processes, and writes data to targets continuously. By default, a session reads
and writes bulk data at scheduled intervals unless you configure the session for
real-time processing.

To process data in real time, the data must originate from a real-time source.
Real-time sources include JMS, WebSphere MQ, TIBCO, webMethods, MSMQ,
SAP, and web services. You might want to use real-time processing for
processes that require immediate access to dynamic data, such as financial data.

To understand real-time processing with PowerCenter, you need to be familiar


with the following concepts:

Real-time data. Real-time data includes messages and messages queues, web
services messages, and changes from a PowerExchange change data capture
source. Real-time data originates from a real-time source.
Real-time sessions. A real-time session is a session that processes real-time
source data. A session is real-time if the Integration Service generates a real-
time flush based on the flush latency configuration and all transformations
propagate the flush to the targets. Latency is the period of time from when
source data changes on a source to when a session writes the data to a target.
Real-time properties. Real-time properties
determine when the Integration Service
processes the data and commits the data to
the target.
Terminating conditions. Terminating conditions determine when
the Integration Service stops reading data from the source and ends
- the session if you do not want the session to run continuously.
-Flush latency. Flush latency determines how often the Integration Service
flushes real-time data from the source.
-Commit type. The commit type determines when the Integration Service
commits real-time data to the target.
Message recovery. If the real-time session fails, you can recover messages.
When you enable message recovery for a real-time session, the Integration
Service stores source messages or message IDs in a recovery file or table. If
the session fails, you can run the session in recovery mode to recover
messages the Integration Service could not process.

Understanding Real-time Data

You can process the following types of real-time data:

Messages and message queues. Process messages and message queues


from WebSphere MQ, JMS, MSMQ, SAP, TIBCO, and webMethods sources.
You can read from messages and message queues. You can write to
messages, messaging applications, and message queues.
Web service messages. Receive a message from a web service client through
the Web Services Hub and transform the data. You can write the data to a
target or send a message back to a web service client.
Change data from PowerExchange change data capture sources. Extract
change data that represent changes to a relational database or file source from
the change stream and write to a target.

Messages and Message Queues

The Integration Service uses the messaging and queueing architecture to


process real-time data. It can read messages from a message queue, process
the message data, and write messages to a message queue.

You can also write messages to other messaging applications. For example, the
Integration Service can read messages from a JMS source and write the data to
a TIBCO target.
PowerExchange Client for PowerCenter, PowerExchange, and the Integration
Service complete the following tasks to process change data:

1.PowerExchange Client for PowerCenter connects to PowerExchange.


2.PowerExchange extracts change data from the change stream.
3.PowerExchange passes change data to the Integration Service through
PowerExchange Client for PowerCenter.
4.The Integration Service transforms and writes data to a target.

For more information about change data, see PowerExchange Interfaces for
PowerCenter.

Configuring Real-time Sessions

When you configure a session to process data in real time, you configure session
properties that control when the session stops reading from the source. You can
configure a session to stop reading from a source after it stops receiving
messages for a set period of time, when the session reaches a message count
limit, or when the session has read messages for a set period of time. You can
also configure how the Integration Service commits data to the target and enable
message recovery for failed sessions.

You can configure the following properties for a real-time session:

Terminating conditions. Define the terminating conditions to determine when


the Integration Service stops reading from a source and ends the session. For
more information, see Terminating Conditions.
Flush latency. Define a session with flush latency to read and write real-time
data. Flush latency determines how often the session commits data to the
targets. For more information, see Flush Latency.
Commit type. Define a source- or target-based commit type for real-time
sessions. With a source-based commit, the Integration Service commits
messages based on the commit interval and the flush latency interval. With a
target-based commit, the Integration Service commits messages based on the
flush latency interval. For more information, see Commit Type.
Message recovery. Enable recovery for a real-time session to recover
messages from a failed session. For more information, see Message Recovery.

Terminating Conditions

A terminating condition determines when the Integration Service stops reading


messages from a real-time source and ends the session. When the Integration
Service reaches a terminating condition, it stops reading from the real-time
source. It processes the messages it read and commits data to the target. Then,
it ends the session.

You can configure the following terminating conditions:

Idle time
Message count
Reader time limit

If you configure multiple terminating conditions, the Integration Service stops


reading from the source when it meets the first condition. By default, the
Integration Service reads messages continuously and uses the flush latency to
determine when it flushes data from the source. After the flush, the Integration
Service resets the counters for the terminating conditions.

Idle Time
Idle time is the amount of time in seconds the Integration Service waits to receive
messages before it stops reading from the source. -1 indicates an infinite period
of time.

For example, if the idle time for a JMS session is 30 seconds, the Integration
Service waits 30 seconds after reading from JMS. If no new messages arrive in
JMS within 30 seconds, the Integration Service stops reading from JMS. It
processes the messages and ends the session.

Message Count

Message count is the number of messages the Integration Service reads from a
real-time source before it stops reading from the source. -1 indicates an infinite
number of messages.

For example, if the message count in a JMS session is 100, the Integration
Service stops reading from the source after it reads 100 messages. It processes
the messages and ends the session.

Note: The name of the message count terminating condition depends on the
Informatica product. For example, the message count for PowerExchange for
SAP NetWeaver is called Packet Count. The message count for PowerExchange
Client for PowerCenter is called UOW Count.

Reader Time Limit

Reader time limit is the amount of time in seconds that the Integration Service
reads source messages from the real-time source before it stops reading from
the source. Use reader time limit to read messages from a real-time source for a
set period of time. 0 indicates an infinite period of time.

For example, if you use a 10 second time limit, the Integration Service stops
reading from the messaging application after 10 seconds. It processes the
messages and ends the session.

Flush Latency

Use flush latency to run a session in real time. Flush latency determines how
often the Integration Service flushes data from the source. For example, if you
set the flush latency to 10 seconds, the Integration Service flushes data from the
source every 10 seconds.

For change data from a PowerExchange change data capture source, the flush
latency interval is determined by the flush latency and the unit of work (UOW)
count attributes. For more information, see PowerExchange Interfaces for
PowerCenter.

The Integration Service uses the following process when it reads data from a
real-time source and the session is configured with flush latency:

1.The Integration Service reads data from the source.


The flush latency interval begins when the Integration Service reads the first
message from the source.
At the end of the flush latency interval, the Integration Service stops reading
2.data from the source.
3.The Integration Service processes messages and writes them to the target.
The Integration Service reads from the source again until it reaches the next
4.flush latency interval.

Configure flush latency in seconds. The default value is zero, which indicates that
the flush latency is disabled and the session does not run in real time.

Configure the flush latency interval depending on how dynamic the data is and
how quickly users need to access the data. If data is outdated quickly, such as
financial trading information, then configure a lower flush latency interval so the
target tables are updated as close as possible to when the changes occurred.
For example, users need updated financial data every few minutes. However,
they need updated customer address changes only once a day. Configure a
lower flush latency interval for financial data and a higher flush latency interval for
address changes.

Use the following rules and guidelines when you configure flush latency:

The Integration Service does not buffer messages longer than the flush latency
interval.
The lower you set the flush latency interval, the more frequently the Integration
Service commits messages to the target.
If you use a low flush latency interval, the session can consume more system
resources.

If you configure a commit interval, then a combination of the flush latency and the
commit interval determines when the data is committed to the target.

Commit Type

The Integration Service commits data to the target based on the flush latency
and the commit type. You can configure a session to use the following commit
types:

Source-based commit. When you configure a source-based commit, the


Integration Service commits data to the target using a combination of the
commit interval and the flush latency interval. The first condition the Integration
Service meets triggers the end of the flush latency period. After the flush, the
counters are reset.
For example, you set the flush latency to five seconds and the source-based
commit interval to 1,000 messages. The Integration Service commits messages
to the target either after reading 1,000 messages from the source or after five
seconds.
Target-based commit. When you configure a target-based commit, the
Integration Service ignores the commit interval and commits data to the target
based on the flush latency interval.

When writing to targets in a real-time session, the Integration Service processes


commits serially and commits data to the target in real time. It does not store
data in the DTM buffer memory.
Message Recovery

When you enable message recovery for a real-time session, the Integration
Service can recover unprocessed messages from a failed session. The
Integration Service stores source messages or message IDs in a recovery file,
recovery table, recovery queue, or recovery topic. If the session fails, run the
session in recovery mode to recover the messages the Integration Service did
not process.

Depending on the real-time source and the target type, the messages or
message IDs are stored in the following storage types:

Recovery file. Messages or message IDs are stored in a designated local


recovery file. A session with a real-time source and a non-relational or non-
queue target uses the recovery file. For more information, see Recovery File.
Recovery table. Message IDs are stored in a recovery table in the target
database. A session with a JMS or WebSphere MQ source and a relational
target uses the recovery table. For more information, see Recovery Table.
Recovery queue and recovery topic. Message IDs are stored in a recovery
queue or recovery topic. A session with a JMS or WebSphere MQ source and a
JMS or WebSphere MQ target uses the recovery queue. A session with a JMS
or WebSphere MQ source and a topic target uses the recovery topic. For more
information, see Recovery Queue and Recovery Topic.

A session can use a combination of the storage types. For example, a session
with a JMS and TIBCO source uses a recovery file and recovery table.

When you recover a real-time session, the Integration Service restores the state
of operation from the point of interruption. It reads and processes the messages
in the recovery file, recovery table, recovery queue, or recovery topic. Then, it
ends the session.

During recovery, the terminating conditions do not affect the messages the
Integration Service reads from the recovery file, recovery table, recovery queue,
or recovery topic. For example, if you specified message count and idle time for
the session, the conditions apply to the messages the Integration Service reads
from the source, not the recovery file, recovery table, recovery queue, or
recovery topic.

In addition to the storage types above, the Integration Service uses a recovery
ignore list if the session fails under certain conditions. For more information, see
Recovery Ignore List.

Sessions with MSMQ sources, web service messages, or change data from a
PowerExchange change data capture source use a different recovery strategy.
For more information, see the PowerExchange for MSMQ User Guide,
PowerCenter Web Services Provider Guide, or PowerExchange Interfaces for
PowerCenter.

Prerequisites
Complete the following prerequisites before you enable message recovery for
sessions with a JMS or WebSphere MQ source and a JMS or WebSphere MQ
target:

Create the recovery queue in the JMS provider or WebSphere MQ. Or, create
the recovery topic in the JMS provider.
Create the recovery queue under the same queue manager as the message
queue so the commit scope is the same.
Configure the recovery queue to be persistent. If the recovery queue is not
persistent, data duplication can occur.

If you do not configure the prerequisites, the Integration Service stores recovery
information in a recovery file instead of a recovery queue or recovery topic.

Steps to Enable Message Recovery

Complete the following steps to enable message recovery for sessions:

In the session properties, select Resume from Last Checkpoint as the recovery
1.strategy.
Specify a recovery cache directory in the session properties at each partition
2.point.

The Integration Service stores messages in the location indicated by the


recovery cache directory. The default value recovery cache directory is
$PMCacheDir.

Recovery File

The Integration Service stores messages or message IDs in a recovery file for
real-time sessions that are enabled for recovery and include the following source
and target types:

JMS source with non-relational, non-JMS, or non-WebSphere MQ targets


WebSphere MQ source with non-relational, non-JMS, or non-WebSphere MQ
targets
SAP R/3 source and all targets
webMethods source and all targets
TIBCO source and all targets

The Integration Service temporarily stores messages or message IDs in a local


recovery file that you configure in the session properties. During recovery, the
Integration Service processes the messages in this recovery file to ensure that
data is not lost.

Message Processing
Message Recovery

When you recover a real-time session, the Integration Service reads and
processes the cached messages. After the Integration Service reads all cached
messages, it ends the session.

For sessions with JMS and WebSphere MQ sources, the Integration Service
uses the message ID in the recovery file to retrieve the message from the
source.

The Integration Service clears the recovery file after the flush latency period
expires and at the end of a successful session. If the session fails after the
Integration Service commits messages to the target but before it removes the
messages from the recovery file, targets can receive duplicate rows during
recovery.

Session Recovery Data Flush

A recovery data flush is a process that the Integration Service uses to flush
session recovery data that is in the operating system buffer to the recovery file.
You can prevent data loss if the Integration Service is not able to write the
recovery data to the recovery file. The Integration Service can fail to write
recovery data in cases of an operating system failure, hardware failure, or file
system outage. The recovery data flush applies to sessions that include a JMS or
WebSphere MQ source and non-relational, non-JMS, or non-WebSphere MQ
targets.

You can configure the Integration Service to flush recovery data from the
operating system buffer to the recovery file by setting the Integration Service
property Flush Session Recovery Data to “Auto” or “Yes” in the Administration
Console.

Recovery Table

The Integration Service stores message IDs in a recovery table for real-time
sessions that are enabled for recovery and include the following source and
target types:

JMS source with relational targets


WebSphere MQ source with relational targets

The Integration Service temporarily stores message IDs and commit numbers in
a recovery table on each target database. The commit number indicates the
number of commits that the Integration Service committed to the target. During
recovery, the Integration Service uses the commit number to determine if it wrote
the same amount of messages to all targets. The messages IDs and commit
numbers are verified against the recovery table to ensure that no data is lost or
duplicated.

Note: The source must use unique message IDs and provide access to the
messages through the message ID.

PM_REC_STATE Table

When the Integration Service runs a real-time session that uses the recovery
table and has recovery enabled, it creates a recovery table, PM_REC_STATE,
on the target database to store message IDs and commit numbers. When the
Integration Service recovers the session, it uses information in the recovery
tables to determine if it needs to write the message to the target table.

Related Topics:
Target Recovery Tables

Message Processing
Message Recovery

When you recover a real-time session, the Integration Service uses the message
ID and the commit number in the recovery table to determine whether it
committed messages to all targets.

The Integration Service commits messages to all targets if the message ID exists
in the recovery table and all targets have the same commit number. During
recovery, the Integration Service sends an acknowledgement to the source that it
processed the message.

The Integration Service does not commit messages to all targets if the targets
have different commit numbers. During recovery, the Integration Service reads
the message IDs and the transformation state from the recovery table. It
processes messages and writes them to the targets that did not have the
message. When the Integration Service reads all messages from the recovery
table, it ends the session.

If the session fails before the Integration Service commits messages to all targets
and you restart the session in cold start mode, targets can receive duplicate
rows.

Recovery Queue and Recovery Topic

The Integration Service stores message IDs in a recovery queue or recovery


topic for real-time sessions that are enabled for recovery and include the
following source and target types:

JMS source with JMS or WebSphere MQ targets


WebSphere MQ source with JMS or WebSphere MQ targets

The Integration Service temporarily stores message IDs and commit numbers in
a recovery queue or recovery topic that you created in the JMS provider or in
WebSphere MQ. The commit number indicates the number of commits that the
Integration Service committed to the target. During recovery, the Integration
Service uses the commit number to determine if it wrote the same amount of
messages to all targets. The messages IDs and commit numbers are verified
against the recovery queue or recovery topic to ensure that no data is lost or
duplicated.

The Integration Service uses the same recovery queue or recovery topic for all
queue targets in each session. Create multiple recovery queues or recovery
topics for sessions to improve performance.

If you do not specify the recovery queue or recovery topic name in the session
properties or in the JMS connection object, the Integration Service stores
recovery information in the recovery file.

Related Topics:

Recovery Table

Message Processing

Stopping Real-time Sessions


A real-time session runs continuously unless it fails or you manually stop it. You
can stop a session by issuing a stop command in pmcmd or from the Workflow
Monitor. You might want to stop a session to perform routine maintenance.

When you stop a real-time session, the Integration Service processes messages
in the pipeline based on the following real-time sources:

JMS and WebSphere MQ. The Integration Service processes messages it read
up until you issued the stop. It writes messages to the targets.
MSMQ, SAP, TIBCO, webMethods, and web service messages. The
Integration Service does not process messages if you stop a session before the
Integration Service writes all messages to the target.

When you stop a real-time session with a JMS or a WebSphere MQ source, the
Integration Service performs the following tasks:

1.The Integration Service stops reading messages from the source.


If you stop a real-time recovery session, the Integration Service stops reading
from the source after it recovers all messages.
The Integration Service processes messages in the pipeline and writes to the
2.targets.
3.The Integration Service sends an acknowledgement to the source.
The Integration Service clears the recovery table or recovery file to avoid data
4.duplication when you restart the session.

When you restart the session, the Integration Service starts reading from the
source. It restores the session and transformation state of operation to resume
the session from the point of interruption.

Restarting and Recovering Real-time Sessions

You can resume a stopped or failed real-time session. To resume a session, you
must restart or recover the session. The Integration Service can recover a
session automatically if you enabled the session for automatic task recovery.

The following sections describe recovery information that is specific to real-time


sessions.

Related Topics:
Recovering Workflows

Restarting Real-time Sessions

When you restart a session, the Integration Service resumes the session based
on the real-time source. Depending on the real-time source, it restarts the
session with or without recovery.

You can restart a task or workflow in cold start mode. When you restart a task or
workflow in cold start mode, the Integration Service discards the recovery
information and restarts the task or workflow.

Recovering Real-time Sessions


If you enabled session recovery, you can recover a failed or aborted session.
When you recover a session, the Integration Service continues to process
messages from the point of interruption. The Integration Service recovers
messages according to the real-time source.

The Integration Service uses the following session recovery types:

Automatic recovery. The Integration Service restarts the session if you


configured the workflow to automatically recover terminated tasks. The
Integration Service recovers any unprocessed data and restarts the session
regardless of the real-time source.
Manual recovery. Use a Workflow Monitor or Workflow Manager menu
command or pmcmd command to recover the session. For some real-time
sources, you must recover the session before you restart it or the Integration
Service will not process messages from the failed session.

Restart and Recover Commands

You can restart or recover a session in the Workflow Manager, Workflow Monitor,
or pmcmd. The Integration Service resumes the session based on the real-time
source.

Table 11-1 describes the behavior when you restart or recover a session with the
following commands:

Table 11-1. Restart and Recover Commands for Real-time Sessions


Command Description
- Restarts the task or workflow. For
Restart Task JMS and WebSphere MQ
- sessions, the Integration Service
Restart Workflow recovers before it restarts the task
- or workflow.
Restart Workflow from Task Note: If the session includes a
JMS, WebSphere MQ source, and
another real-time source, the
Integration Service performs
recovery for all real-time sources
before it restarts the task or
workflow.
- Recovers the task or workflow.
Recover Task
-
Recover Workflow
-
Recover Workflow from Task
- Discards the recovery information
Cold Start Task and restarts the task or workflow.
-
Cold Start Workflow
-
Cold Start Workflow from Task
Related Topics:
Steps to Recover Workflows and Tasks
Restarting a Task or Workflow Without Recovery

Rules and Guidelines

Rules and Guidelines for Real-time Sessions

Use the following rules and guidelines when you run real-time sessions:

The session fails if a mapping contains a Transaction Control transformation.


The session fails if a mapping contains any transformation with Generate
Transactions enabled.
The session fails if a mapping contains any transformation with the
transformation scope set to all input.
The session fails if a mapping contains any transformation that has row
transformation scope and receives input from multiple transaction control points.
The session fails if the load scope for the target is set to all input.
The Integration Service ignores flush latency when you run a session in debug
mode.
If the mapping contains a relational target, configure the load type for the target
to normal.
If the mapping contains an XML target definition, select Append to Document for
the On Commit option in the target definition.
The Integration Service is resilient to connection failures to WebSphere MQ. It is
not resilient to any other messaging system.

Rules and Guidelines for Message Recovery

The Integration Service fails sessions that have message recovery enabled and
contain any of the following conditions:

The source definition is the master source for a Joiner transformation.


You configure multiple source definitions to run concurrently for the same target
load order group.
The mapping contains an XML target definition.
You edit the recovery file or the mapping before you restart the session and you
run a session with a recovery strategy of Restart or Resume.
The Integration Service cannot connect to the recovery queue or recovery topic.
The Integration Service does not write the recovery message to the recovery
queue or recovery topic.

If the number of messages that the Integration Service reads or writes from the
message queue exceeds the message size limit, increase the message size limit
or decrease the flush latency.
The sample mapping includes the following components:

Source. WebSphere MQ. Each message is in XML format and contains one
purchase order.
XML Parser transformation. Receives purchase order information from the
MQ Source Qualifier transformation. It parses the purchase order ID and the
quantity from the XML file.
Lookup transformation. Looks up the supplier details for the purchase order
ID. It passes the supplier information, the purchase item ID, and item cost to the
Expression transformation.
Expression transformation. Calculates the order cost for the supplier.
Target. Oracle relational database. It contains the supplier information and the
total supplier cost.

You create and configure a session and workflow with the following properties:

Property Value
Message count 1,000
Flush latency interval 2,000 milliseconds
Commit type Source-based commit
Workflow schedule Run continuously

The following steps describe how the Integration Service processes the session
in real-time:

The Integration Service reads messages from the WebSphere MQ queue until
it reads 1,000 messages or after 2,000 milliseconds. When it meets either
1.condition, it stops reading from the WebSphere MQ queue.
The Integration Service looks up supplier information and calculates the order
2.cost.
The Integration Service writes the supplier information and order cost to the
3.Oracle relational target.
The Integration Service starts to read messages from the WebSphere MQ
4.queue again.
The Integration Service repeats steps 1 through 4 as you configured the
5.workflow to run continuously.

Informatica Real-time Products

You can use the following products to read, transform, and write real-time data:

PowerExchange for JMS. Use PowerExchange for JMS to read from JMS
sources and write to JMS targets. You can read from JMS messages, JMS
provider message queues, or JMS provider based on message topic. You can
write to JMS provider message queues or to a JMS provider based on message
topic.
JMS providers are message-oriented middleware systems that can send and
receive JMS messages. During a session, the Integration Service connects to the
Java Naming and Directory Interface (JNDI) to determine connection information.
When the Integration Service determines the connection information, it connects
to the JMS provider to read or write JMS messages.
PowerExchange for WebSphere MQ. Use PowerExchange for WebSphere
MQ to read from WebSphere MQ message queues and write to WebSphere MQ
message queues or database targets. PowerExchange for WebSphere MQ
interacts with the WebSphere MQ queue manager, message queues, and
WebSphere MQ messages during data extraction and loading.
PowerExchange for TIBCO. Use PowerExchange for TIBCO to read
messages from TIBCO and write messages to TIBCO in TIB/Rendezvous or AE
format.
The Integration Service receives TIBCO messages from a TIBCO daemon, and it
writes messages through a TIBCO daemon. The TIBCO daemon transmits the
target messages across a local or wide area network. Target listeners subscribe
to TIBCO target messages based on the message subject.
PowerExchange for webMethods. Use PowerExchange for webMethods to
read documents from webMethods sources and write documents to
webMethods targets.
The Integration Service connects to a webMethods broker that sends, receives,
and queues webMethods documents. The Integration Service reads and writes
webMethods documents based on a defined document type or the client ID. The
Integration Service also reads and writes webMethods request/reply documents.
PowerExchange for MSMQ. Use PowerExchange for MSMQ to read from
MSMQ sources and write to MSMQ targets.
The Integration Service connects to the Microsoft Messaging Queue to read data
from messages or write data to messages. The queue can be public or private
and transactional or non-transactional.
PowerExchange for SAP NetWeaver. Use PowerExchange for SAP
NetWeaver to read from SAP using outbound IDocs or write to SAP using
inbound IDocs using Application Link Enabling (ALE).
The Integration Service can read from outbound IDocs and write to a relational
target. The Integration Service can read data from a relational source and write
the data to an inbound IDoc. The Integration Service can capture changes to the
master data or transactional data in the SAP application database in real time.
PowerCenter Web Services Provider. Use PowerCenter Web Services
Provider to expose transformation logic as a service through the Web Services
Hub and write client applications to run real-time web services. You can create a
service mapping to receive a message from a web service client, transform it,
and write it to any target PowerCenter supports. You can also create a service
mapping with both a web service source and target definition to receive a
message request from a web service client, transform the data, and send the
response back to the web service client.
The Web Services Hub receives requests from web service clients and passes
them to the gateway. The Integration Service or the Repository Service process
the requests and send a response to the web service client through the Web
Services Hub.
PowerExchange. Use PowerExchange to extract and load relational and non-
relational data, extract change data, and extract change data in real time.
To extract data, the Integration Service reads change data from PowerExchange
on the machine hosting the source. You can extract and load data from multiple
sources and targets, such as DB2/390, DB2/400, and Oracle. You can also use a
data map from a PowerExchange Listener as a non-relational source. For more
information, see PowerExchange Interfaces for PowerCenter.

You might also like