Efficiently managing data ingestion is crucial for businesses leveraging Pega Customer Decision Hub. By employing data jobs, organizations can streamline the process, ensuring a standardized and structured approach to importing data into the Customer Decision Hub. This not only enhances efficiency but also facilitates seamless monitoring of data job progress through the intuitive Customer Profile Designer.
Let's delve into the steps to create a new batch import data job:
Data Source Setup:
The customer data, typically in CSV format, resides in an Amazon S3 file repository in an external environment.
Alongside the CSV file, a manifest file provides essential details like process type, total record count, and the relative file path to the CustomerDataIngest.csv file.
Access to the repository is facilitated through defining a Repository rule in Pega Platform, with the Root path set as /DataJobs/.
It's vital to ensure that the file path in the manifest file is relative to this defined path, e.g., /DataJobs/CustomerData/ folder within the Amazon S3 repository.
Privilege Requirement:
To create a new data job, users must possess the CanUpdateJobConfiguration privilege in their Access Group.
Data Ingestion Tasks:
Data ingestion involves two primary tasks: defining a data set to read from the source and creating the data job to ingest the source data to its destination.
Dataset Creation:
Define a new file data set, such as Customerfile, from the Profile Data Sources in Customer Decision Hub.
Assign a name, select the source type, and specify the Apply to class.
Choose the repository containing the data to import and navigate to the desired file or location.
Best Practice for File Data Set Creation:
Load both the manifest and data files in the repository during the file data set creation. This practice facilitates seamless mapping of columns in the CSV files to properties in the data model during the Field mapping step.
Data Ingestion Configuration:
Configure the data job to ingest data from the source to its destination.
Begin by selecting the target data source and the source data set (e.g., Customerfile). - Choose the trigger mechanism (e.g., file detection or schedule) to initiate the run.
Define the failure policy to handle situations where the number of failed records exceeds the defined error threshold.
Confirmation and Monitoring:
Review and confirm all settings for the data job.
The ingestion process commences upon detection of a token file as defined in the data job.
Monitor the progress of the data ingestion process on the Overview tab of the data job, which displays details such as processed records, successful records, failed records, and the final status.
By following these systematic steps, organizations can effectively manage data ingestion within Pega Customer Decision Hub, ensuring a smooth and structured process from source to destination.
-Team Enigma Metaverse
Comentários