
Job Summary
As a Senior Data Engineer at ICEA LION, you will develop, optimise, and manage our data lake, data pipelines, and data infrastructure to power analytics, reporting, advanced analytics, machine learning & AI. Your role will focus on building scalable data products that unify data across all interactions and touchpoints. Working with cross-functional teams, you will enforce data governance standards and drive a collaborative, data-driven culture.
Roles And Responsibilities
▪ Set up, manage, and maintain the company’s data analytics systems, including the operational data store, feature store, and data warehouse, to support informed decision-making.
▪ Create and manage processes (ETL/ELT) for collecting, cleaning, and transforming data from different sources.
▪ Build and improve data structures to support reporting, advanced analytics, and machine learning projects.
▪ Ensure data queries run efficiently while keeping costs low and making the best use of resources.
▪ Work closely with technology and data analytics teams to develop customized data solutions and include machine learning insights.
▪ Set up and manage machine learning workflows to make sure models run smoothly and reliably.
▪ Implement data quality checks, follow governance rules, and enforce security to protect and control access to data.
▪ Automate tasks and improve processes using tools to ensure stable and scalable data pipelines.
- Clearly document all processes and share knowledge to help the team adopt best practices.
Requirements
Academic and Professional Qualifications
▪ Bachelor’s Degree in Computer Science, Data Science, Information Technology, Engineering, or a related field.
▪ Professional Data Engineer Certification is a plus.
▪ Master’s Degree in Data Science, Computer Science, Information Systems, or a related discipline.
▪ Extensive knowledge of data warehousing concepts, including dimensional modelling and data marts.
▪ Minimum of 5-7 years in data engineering, data architecture, or a related field.
▪ Experience in leading data projects or teams is highly desirable.
▪ Advanced proficiency in SQL and NoSQL with hands-on experience creating complex queries and data transformations.
▪ Proven experience with cloud-based data engineering tools such as Cloud Storage, Data flow, Cloud Composer, Cloud Functions, and AWS Glue.
▪ Strong familiarity with ETL/ELT tools (e.g., Apache Beam, Apache Airflow, SSIS, Data flow, dbt) for building and maintaining data pipelines.
▪ Proficient in scripting languages like Bash, Python, and JavaScript to support automation and integration tasks.
▪ Skilled in managing and optimizing large datasets for performance and cost-efficiency.
▪ Excellent communication skills with a demonstrated ability to work collaboratively in cross-functional teams.
- Familiarity with Machine Learning (ML) techniques and Language Learning Models (LLMs) to support data-driven applications.