Fetching Data From One Iflow Via Process Direct And Invoking A Loop Process Call With Exception Handling
WRITTEN BY
14th June 2022
INTEGRATION
Introduction
Hello
This is my first blog Post where I have been tried to describe a real time scenario of CPI where we have to call the iflow followed by another iflow and spilt the data and fetch it with looping process call. The exception messages will be stored in the data store (with global variable) which will send the receiver the mail also. The receiver also wants to maintain the log id which will be maintained in the Local integration process.
The key to the solution is The Process Direct adapter which can be used to establish communication between integration processes within the same tenant. A Looping Process Call in Cloud Integration that invokes a Local Integration Process iteratively till the conditions are met.
Pre-Requisite:
- 1. Access to SAP BTP Cockpit
– Scenario
In this blog post, we will fetch the data from odata via Request Reply then send the data via Process Direct to another iflow. In the second iflow branch the data using Parallel Multicast then send one phase data to the Loop Process call where it is splitted and the Logger script has been added for checking the log ID after deployment.
The exception sub process has been added to store the data in the data store and the content modifier send the message to the receiver.
For simple example, I just create simple integration (Fig 1 & 2) just include timer and one Mail Receiver Adapter.
Fig 1 iflow-1 with Process direct Receiver
Fig 2 iflow-2 with Process direct Receiver
⦁ Details of Adapter and Pallets Configuration
⦁ iFlow-1
Odata Configuration (Fig 3)
Use any free Odata Service here I have used Northwind and select the entities as per your requirement.
Process Direct Receiver Adapter (Fig 4)
Redirect to the address given in sender iflow sender adapter (Process direct).
- iFlow-II
Content Modifier
Set the current timestamp in header value
${date:now:dd-MM-yyyy HH:mm}
Looping Process Call
A Looping Process Call in CPI invokes a Local Integration Process iteratively till the conditions are met. The loop has to be executed till the maximum number if iteration.
Local Integration Process
In Transaction Handling select the required calling Process.
General Splitter (Fig 5)
You can choose the Root node accordingly, here I have been select the Product.
Logger Script
Logger Script is here to check the Logs which we can check from the monitoring page.
import com.sap.gateway.ip.core.customdev.util.Message;
import java.util.HashMap;
def Message processData(Message message)
{
def body = message.getBody(java.lang.String) as String;
def messageLog = messageLogFactory.getMessageLog(message);
if(messageLog != null)
{
messageLog.setStringProperty("log1","Printing Payload As Attachment")
messageLog.addAttachmentAsString("log1",body,"text/plain");
}
return message;
}
Exception Sub Process (Fig 6)
If the iflow has some error then it will go to data store through exception sub process and we can check there by fetching the data from a different iflow as we have set the Global Variable in Write variable option.
Content Modifier
Set the message in message Body to send the receiver.
Deployment (Fig 8)
After deployment the mail received as follows.
Conclusion
This blog will help you to connect two different iflow via Process direct and help to store the exception data in the data store which explain the use of Write variable and write operation. The process call also check the logs in the monitoring page by logger script.
Please like the blog post if you find the information useful. Also, please leave your thoughts and opinions in the comments section.
Good luck with your studies!…
References
1]https://blogs.sap.com/2020/01/09/sap-cloud-platform-integration-cpi-part-7-maintaining-logs-using-groovy-scripts-even-if-the-iflow-is-not-on-trace-mode./
2]https://blogs.sap.com/2021/09/16/sap-cloud-integration-cpi-hci-writing-groovy-scripts-_-with-basic-examples/
3]https://blogs.sap.com/2020/01/23/sap-cloud-platform-integration-cpi-part-11-how-to-use-a-local-integration-process/
How Can Consumer Products Brands Use SAP BTP to Build a Trusted Data Foundation for Faster Product Launches?
Consumer product companies face intense competition, evolving consumer expectations, demand for personalization, and sustainability goals. To stay ahead, brands must deliver innovative products faster with precision. This requires a unified data foundation connecting design, manufacturing, supply chain, and marketing. When data is siloed, decisions slow and inconsistencies grow. With SAP BTP, brands can centralize data, enhance accuracy, and drive agility, enabling faster, more confident decisions and accelerating innovation across the consumer products value chain.
The Enterprise Playbook for SAP BTP Success
Enterprises today walk a tightrope between legacy systems, silos, and rigid integrations on one side, and the rapid pace of innovation, AI, and evolving customer demands on the other. SAP Business Technology Platform (BTP) provides a strong foundation to bridge this gap. Success, however, depends on strategy. This enterprise playbook outlines four actionable plays showing how Incture and SAP BTP solutions help overcome key challenges and deliver scalable business outcomes.


















































