We are seeking a highly experienced Senior Data Engineer to design, build, and optimize scalable data pipelines and architectures within the Cloudera ecosystem. The ideal candidate will have strong expertise in ETL/ELT development, data integration, and big data performance optimization to support analytics and business objectives.
Key Responsibilities
- Design, build, and optimize scalable and reliable data pipelines and architectures.
- Develop ETL/ELT processes for structured, semi-structured, and unstructured data sources.
- Collaborate with data scientists, analysts, and stakeholders to ensure data availability and usability.
- Ensure data quality, integrity, and compliance across multiple platforms.
- Optimize database queries and big data processing for high performance.
- Monitor, maintain, and troubleshoot daily data pipelines and ETL jobs.
- Implement automation and monitoring tools to ensure reliability and efficiency.
- Document processes, workflows, incidents, and solutions to support knowledge sharing.
Required Qualifications
- Bachelor’s degree in Computer Science, Engineering, or related field.
- 10+ years of professional experience in data engineering.
- Proven hands-on expertise with Cloudera ecosystem (Spark, Hive, Kafka, HDFS).
- Advanced knowledge of SQL, Python, and data modeling techniques.
- Strong understanding of data integration and data warehousing concepts.
Preferred Qualifications
- Experience with cloud platforms (AWS, GCP, or Azure).
- Knowledge of CI/CD tools and version control systems (Git).
- Exposure to real-time data streaming (Kafka, Flink).
- Arabic Fluency
Key Competencies
- Strong problem-solving and analytical skills.
- Ability to design efficient and scalable solutions.
- Effective collaboration and communication with cross-functional teams.
- Attention to detail with a strong focus on data quality.