Responsibilities:
- Design, implement, and maintain scalable data pipelines and ETL processes in cloud environments (AWS) to support client business needs.
- Collaborate with cross-functional teams to develop solutions that meet both technical and business objectives.
- Work with clients to understand data requirements, existing systems, and opportunities for modernisation.
- Develop and maintain data models, schemas, databases, and data lakes for structured and unstructured data sources.
- Integrate data with GenAI solutions, including implementing architectures such as RAG in SageMaker.
- Troubleshoot and resolve data quality, consistency, and availability issues promptly.
- Stay updated on emerging technologies, tools, and trends in cloud computing and data engineering to foster continuous improvement and innovation.
Qualifications / Requirements:
- Bachelor's or Master's degree in Computer Science, Engineering, or a related field.
- 2-5 years of experience in data engineering, focusing on building and optimizing data pipelines in cloud environments.
- Proficiency in Python, Java, or Scala, with experience in SQL and NoSQL databases (e.g., Amazon RDS, DynamoDB).
- Hands-on experience with cloud data and analytics services (e.g., AWS Glue, Azure Data Factory, Google Cloud Dataflow).
- Strong understanding of data modeling, data warehousing, and distributed computing systems.
- Practical experience delivering ML/AI use cases.
- Excellent problem-solving skills, attention to detail, and ability to work independently and collaboratively.
- Effective communication skills to articulate technical concepts to non-technical stakeholders.
- Self-motivated learner eager to advance their career.
- Experience with agile methodologies such as Scrum or Kanban.
Skills must have:
- Python
- SQL/RDBMS
- Git/Version Control
- AWS
- Agile/Scrum
- Data skills (granularity, meaning, standardization)
- Notebooks
- Usage of GenAI in daily tasks
- Client-facing skills (demos, workshops)
Nice to have:
- IAC (e.g., CloudFormation or Terraform)
- Cloud Data Engineering (e.g., PySpark, Databricks, Snowflake)
- Linux
- Cloud Security
- Documentation
- AWS Data Engineer Associate Certification
- Data visualization tools (e.g., Quicksight, Superset, Tableau, Power BI)
- Compute services (EC2, Lambda, Batch)
- Data formats (JSON, XML, flat files)
- Data lake/mesh architectures
- Kinesis/Streaming experience
About Us:
Westcon-Comstor is a global technology distributor specializing in Cloud, Deployment, Security, UC, Networking, and Data Center solutions. With teams in over 110 offices across more than 70 countries, we are committed to transforming technology distribution and delivering exceptional partner experiences.