Data Engineer - Data Lakehouse
About this position
Responsibilities
• Design, build and maintain efficient and scalable data pipelines to support data processing and analysis
• Develop and implement ETL processes to ensure data quality and integrity
• Manage and optimize databases, including designing schema and indexing strategies, data partitioning and data archival
• Work closely with data scientists and business analysts to understand data requirements and translate them into technical solutions
• Collaborate with DevOps and IT teams to deploy, monitor and troubleshoot data systems
• Develop and maintain documentation on data infrastructure and processes
• Stay current with the latest data engineering technologies, tools and practices
Requirements
• Bachelor's degree in computer science, engineering or related field
• Minimum of 3 years of experience in data engineering roles
• Strong proficiency in SQL, ETL processes and database management systems (e.g., MySQL, PostgreSQL, MongoDB)
• Hands-on experience with AWS services for data processing, storage and analysis (e.g., S3, Redshift, EMR, Glue)
• Familiarity with programming languages such as Python or Java
• Understanding of data warehousing concepts and data modeling techniques
• Experience working with big data technologies (e.g., Hadoop, Spark) is an advantage
• Excellent problem-solving and analytical skills
• Strong communication and collaboration skills