ClearPeaks is a specialist consulting firm delivering services and solutions in Everything Data - Business Intelligence, Advanced Analytics, Big Data & Cloud, and Web & Mobile Applications. Founded in 2000, we have been a trusted partner to our customers in over 15 industry verticals and functional areas, with operations spanning Europe, Middle East, the United States, and Africa.
ClearPeaks is part of a strategic alliance with synvert, a group of six successful full-service Data & Analytics (D&A) consulting firms, with a clear goal to become one of EMEA's largest D&A consulting companies.
Our services are based on the latest market-leading enterprise technology platforms, and delivered by a dynamic team of expert consultants. Our strength lies in our ability to efficiently deliver customer insight and value, gained through our decades of experience with real-world challenges.
As a Data Engineer, you will be responsible securely transfer the data between different network domains and environments. Besides that, you will be improving the performance and enhancing the structure of the data pipelines and data models.
At ClearPeaks, there are endless opportunities to get involved in different projects bringing innovative ideas while being part of leading-edge teams who are always innovating and evolving in the Data Management field.
Responsibilities
- Participate in requirements clarification and sprint planning sessions.
- Design technical solutions and implement them, including ETL Pipelines Build robust data pipelines in PySpark to extract and transform.
- Optimize ETL Processes Enhance and tune existing ETL processes for better performance, scalability, and reliability.
- Writing unit and integration tests.
- Support QA teammates in the acceptance process.
Requirements
- A degree in Statistics, Computer Science or Telecommunication Engineering, Informatics or related degree.
- Proficiency in PySpark for distributed computing and Python for ETL development.
- Strong expertise in writing and optimizing complex SQL queries, preferably with experience in databases such as PostgreSQL, MySQL, Oracle, or Snowflake.
- Experience working with data warehousing concepts and platforms, ideally Databricks.
- Familiarity with ETL tools & processes.
- Experience with dimensional modelling, normalization/denormalization, and schema design.
- Proficiency with version control tools like Git to manage codebases and collaborate on development.
- Familiarity with monitoring tools (e.g., Prometheus, Grafana, or custom monitoring scripts) to track pipeline performance.
- Experience implementing data validation, cleansing, and quality frameworks, ideally Monte Carlo.
- Comfort in working within an Agile team, including sprint planning, stand-ups, and retrospectives.
- Ability to work in a team environment with tools like Azure Dev Ops, Jira and/or Confluence.
- Strong debugging and problem-solving skills to troubleshoot complex data engineering issues.
Our offer to you!
- Work with leading edge technologies that will enable you to accelerate your career development.
- Enjoy an excellent work environment where people love what they do.
- Be part of an international and ambitious team whilst having fun.