Charif

This project is maintained by C-EB

talend data integration project-commercial data management

Objective

Creation of a Data Integration System in a Data Warehouse

The project consisted of two distinct parts:

Part 1: Loading the ODS

In this phase, I set up a PostgreSQL database (COM_INGESTION_DB) and developed SQL scripts for schema creation, table creation, and data insertion. I carefully managed the ingestion process, ensuring seamless integration of flat files into the database using Talend’s powerful data integration capabilities. Key tasks included:

Data source

Six CSV files CSV files

ETL process

Extract

Extracting data from multiple sources (six CSV files)

Transform

Load

Loading data into the target database

Tools

| Tool | Purpose | |:————:|:———————————-:| | PostgreSQL | managing ODS and QWH tables | | Talend | ETL process | | Github | Hosting the project documentation |

Development

Part 1

Step 1: Database

Creation of a directory containing the daily data

Step 3: Data Integration

Part 2

Step 1: Database ==> PostgreSQL

  1. script_create_schema_dwh.sql: Creation of the schema VENTE_DWH

  2. script_create_table_dwh.sql: Creation of the various DWH tables

  3. script_create_table_ods.sql: Updating the ODS_VENTE table

  4. script_insert_context_variable.sql: Inserting data into the CONTEXT table

  5. script_select_table_dwh.sql: This script selects the data inserted into the various tables

  6. script_truncate_table_dwh.sql This script deletes the data present in the various DWH tables

Database

Step 2: Data Integration Process ==> Talend

  1. Opening the Talend project: ICOMMERCE_REPORTING
  2. Creating the context group DWH
  3. Creating metadata: Connecting to the database
  4. Creating the DWH jobs DWH Client DWH Product DWH Vente DWH generate calender

  5. Creating the main job Job charge DWH

Part 3: Building and scheduling the automatic execution of the main job (Job for populating the Operational Data Store and the Data Warehouse)

Job charge DWH