Overview
We are seeking a Data Engineer to design, develop, and maintain data pipelines and data products in a cloud environment, supporting data ingestion, transformation, validation, and export automation for enterprise reporting and analytics.
Responsibilities
- Design, build, and optimize data pipelines and ETL/ELT processes in cloud environments (e.g., AWS, Azure) to ingest data from multiple sources and transform it into usable data products.
- Develop scalable data models and data warehouses/lakes; implement dimensional modeling and data governance practices.
- Collaborate with data scientists, analysts, and stakeholders to understand data requirements and deliver reliable datasets for reporting and analytics.
- Monitor, troubleshoot, and optimize data pipelines for performance, reliability, and cost efficiency; implement automated QA checks and data quality controls.
- Document data sources, transformations, dependencies, and data governance processes; provide knowledge transfer and mentoring as needed.
- Participate in CI/CD and deployment of data infrastructure; containerize and deploy components using Docker and orchestration where applicable.
- Support operational and ad-hoc data requests and assist in data security and compliance practices.
Required Skills
- Strong experience building data pipelines and data warehouses in cloud environments (e.g., AWS, Azure).
- Advanced SQL skills and experience with ETL/ELT development; proficiency with data modeling (dimensional modeling, star schemas).
- Experience with data governance, lineage, and quality frameworks; familiarity with data validation techniques.
- Proficiency in Python or similar scripting languages; experience with data tools such as Apache Airflow or equivalent orchestration.
- Experience with containerization (Docker) and basic knowledge of CI/CD for data infrastructure.
- Excellent communication and collaboration skills for working with cross-functional teams.
Nice-to-Have
- Experience with Snowflake, DBT, or other modern data stack technologies.
- Experience in healthcare IT data or in regulated industries.
- Experience with analytics databases and cloud data services (S3, Snowflake, SageMaker, Data Factory, Synapse, etc.).
- Familiarity with API integration, data streaming (Kafka), and big data tools (Hadoop, Spark).
What to Expect
Contract roles with potential extensions based on project needs. Benefits details vary by engagement. Fully remote options may be available depending on client and location.
About the Company
We are a global digital solutions provider delivering technology strategy, software engineering, and product development across industries. We value inclusion, collaboration, and continuous learning.
Note: This description consolidates multiple postings and focuses on core data engineering responsibilities and qualifications. All candidate data policies and privacy notices remain in effect where applicable per client and regulatory requirements.