Fulltime

Data Engineer

Fulltime
Data Engineer
View below the job description in addition to the application form.
Responsibilities :
  • Design, develop, and maintain scalable data pipelines.
  • Build data ingestion processes using Change Data Capture and ETL / ELT jobs.
  • Build data pipeline to cleanse and transform raw data into clean data in the targeted format to be stored in Data lake.
  • Build and manage jobs to preprocess data into an optimized and highly structured data format into Data mart.
  • Write clean, efficient, and well-tested code for data processing and transformation.
  • Optimize data storage and retrieval processes to ensure high performance and scalability.
  • Collaborate with cross-functional teams to understand data requirements and deliver solutions.
  • Participate in code reviews and contribute to technical discussions.
  • Monitor and optimize data pipeline performance.
  • Document technical specifications and data pipeline documentation.
  • Troubleshoot and resolve complex technical issues related to data processing.
Preferred Qualifications:
  • Bachelor’s degree in Computer Science, Software Engineering, or equivalent practical experience. 
  • 3+ years of experience in data engineering.
  • Strong proficiency in building data pipelines using Debezium, Airflow, Spark, Hive Metastore, Trino (Presto).
  • Solid understanding of data lake and data mart architectures.
  • Experience with ETL processes and data transformation techniques.
  • Proficiency in SQL and experience with relational and NoSQL databases.
  • Strong knowledge of software design patterns and principles.
  • Experience with containerization (Docker) and version control (Git).
  • Proven track record of writing clean, maintainable code.
  • Proficiency in unit testing and test-driven development.
  • Excellent problem-solving, communication, and collaboration abilities.
  • Experience in event-driven architecture is a plus.