Experience - 610 years
Location - Dubai
Key Responsibilities
Data Pipeline & ETL/ELT Development
- Design, build, and maintain scalable and high-performance data pipelines.
- Implement batch and real-time ingestion using tools like Apache Spark, Kafka, Flink, NiFi, Airflow, etc.
- Optimize data workflow performance, quality, and latency.
Data Architecture & Modeling
- Develop and maintain data models, data lakes, and data warehouses (Kimball, Data Vault, DDD models).
Cloud Data Engineering
- Build and manage data ecosystems on AWS / Azure / GCP:
- AWS: Glue, EMR, Redshift, S3, Lambda
- Azure: Data Factory, Synapse, Databricks, ADLS
- GCP: Dataflow, BigQuery, Dataproc
- Optimize cloud resources, storage formats (Parquet/ORC), and cost efficiency.
Big Data Processing
- Work with distributed data systems: Spark, Hadoop, Hive, Presto, Snowflake, Databricks, etc.
- Handle large-scale structured and unstructured datasets.
Database Management
- Manage relational and NoSQL databases (PostgreSQL, SQL Server, Cassandra, MongoDB, DynamoDB, etc.).
- Write complex SQL queries, schema design, and performance tuning.
Technical Skills
- 610 years of experience in Data Engineering / Big Data Engineering.
- Strong in Python, SQL (Scala/Java optional).
- Expertise in Spark (mandatory).
- Hands-on experience with cloud data platforms (AWS / Azure / GCP).