Peform the Extract, Transform, and Load (ETL) process for data pipelines
Construct ETL job pipelines using ETL tools
Manage ETL job scheduling and optimize data load
Identify gaps in the ETL pipeline and execute data normalization to enhance data processing
Gain extensive exposure to various database technologies and automation
Design, create, document, and implement comprehensive data pipelines and integration processes, encompassing both batch and real-time data processing
Perform data analysis, profiling, cleansing, lineage, mapping, and transformation
Develop ETL/ELT jobs, workflows, and deploy data solutions
Monitor and suggest improvements for data quality, reliability, efficiency, and cleanliness
Optimize and fine-tune ETL/ELT processes
Recommend and implement best practices in data management and data lifecycle processes
Prepare test data and support the creation and execution of test plans.
Collaborate with Data Architect, Data Modeler, IT team members, SMEs, vendors, and business stakeholders to comprehend data requirements and implement data solutions
Provide ongoing support for data issues and change requests, documenting all investigations, findings, recommendations, and resolutions
Requirements:
Over 2 years of work experience in the insurance or finance industry
Knowledge of IBM DataStage or equivalent ETL software
Strong communication skills to effectively communicate and present technical information in a clear and unambiguous manner
Familiarity with ETL/ELT frameworks, data warehousing concepts, data management frameworks, and data lifecycle processes
Experienced in handling and processing diverse data types (structured, semi-structured, and unstructured)
Thorough understanding of various database technologies (RDBMS, NoSQL, and columnar)
Profound comprehension of programming languages such as Python
Strong ability to work independently and collaborate with diverse teams in a multi-stakeholder environment