Enable job alerts via email!

Senior Data Engineer (Python, SQL, AWS)

DigiConnect – ogłoszenia grzecznościowe

Gdańsk

On-site

PLN 120,000 - 160,000

Full time

7 days ago
Be an early applicant

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

A fast-growing software development company is seeking a Senior Data Engineer to manage data pipelines for a leading language technology client. This role involves collaborating with stakeholders, ensuring data quality, and optimizing ETL processes using modern cloud technologies. Join a team dedicated to delivering efficient data solutions in a diverse and inclusive environment.

Benefits

Competitive salary
Flexible timesheet (remote-friendly)
Medical insurance
Private healthcare

Qualifications

  • Proficiency in Python programming, including data manipulation libraries.
  • Experience with key AWS services like S3, Glue, Lambda.
  • Understanding of ETL processes and data pipeline development.

Responsibilities

  • Develop and maintain data pipelines using Python and AWS services.
  • Implement ETL processes to extract, transform, and load data.
  • Monitor and optimize data workflows for performance and cost-efficiency.

Skills

Python
AWS
SQL
Problem Solving
Data Quality

Education

Valid Polish residence and permission to work in Poland
Fluent English (C1 or above)

Tools

Terraform
CloudFormation
CDK
Git
Apache Airflow
Docker
Kubernetes

Job description

Digiteum, a fast-growing cross-functional software development company, is looking for a Data Engineer to work on the world’s largest linguistics databases for a leading language technology company.

The Client is a global provider of quality lexical data and linguistics databases in hundreds of world languages. Engineers, startups, and blue-chip brands (Apple, Microsoft, Google, etc.) rely on the Client’s language datasets and services to build multilingual applications, lexical software, machine translation, NLP, and ML technologies.

You will join a team focused on delivering reliable and efficient data engineering solutions using modern cloud technologies. Your main responsibility will be to build and manage data pipelines that support key analytical and operational workflows.

As a Data Engineer, you will collaborate closely with stakeholders to understand data needs, design effective solutions, and ensure high standards of data quality and performance. The role also includes maintaining and optimizing ETL processes, contributing to code reviews, and supporting infrastructure deployment using Infrastructure as Code tools.

This is a hands-on engineering role, ideal for someone who enjoys problem-solving, cares about clean and efficient data workflows, and wants to contribute to a data-driven environment.

Join our team of experts as a Senior Data Engineer.

  1. Develop and maintain data pipelines using Python and AWS services (e.g., AWS Glue, Lambda, S3).
  2. Implement ETL processes to extract, transform, and load data from various sources.
  3. Collaborate with language engineers and stakeholders to gather requirements and deliver effective solutions.
  4. Ensure data quality and integrity through automated testing and validation.
  5. Monitor and optimize data workflows for performance and cost-efficiency.
  6. Document data engineering processes for knowledge sharing and support.
  7. Participate in code reviews and provide feedback.
  8. Assist in designing data models and schemas.
  9. Support infrastructure deployment and maintenance using Infrastructure as Code (IaC) tools (e.g., Terraform, CloudFormation, CDK).
  10. Proficiency in Python programming, including data manipulation libraries (e.g., Pandas, NumPy).
  11. Experience with key AWS services (e.g., S3, Glue, Lambda, RDS).
  12. Understanding of ETL processes and data pipeline development.
  13. Familiarity with SQL and database management.
  14. Experience with version control systems (e.g., Git).
  15. Knowledge of data quality and validation techniques.
  16. Strong problem-solving skills and a proactive, analytical mindset.
  17. Valid Polish residence and permission to work in Poland.
  18. Fluent English (C1 or above), with the ability to explain complex topics to diverse stakeholders.
  19. Experience with additional scripting or programming languages (e.g., SQL, Bash).
  20. Familiarity with data orchestration tools (e.g., Apache Airflow).
  21. Knowledge of data warehousing tools and concepts (e.g., Redshift).
  22. Understanding of data security and compliance best practices.
  23. Experience with Infrastructure as Code tools (e.g., Terraform, CloudFormation).
  24. Knowledge of containerization and orchestration (e.g., Docker, Kubernetes).
  25. Work with small teams of highly skilled specialists who share knowledge and welcome initiative.

We offer a competitive salary, extensive corporate life, flexible timesheet (remote-friendly), medical insurance, and private healthcare.

We support diversity and inclusivity, encouraging all qualified candidates to apply regardless of gender, age, race, ethnicity, religious belief, or physical ability. We are committed to equal opportunities and the personal and professional growth of our team members.

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.