Job Search and Career Advice Platform

Aktiviere Job-Benachrichtigungen per E-Mail!

Data Engineer

Cocus

Düsseldorf

Hybrid

EUR 60.000 - 80.000

Vollzeit

Heute
Sei unter den ersten Bewerbenden

Erstelle in nur wenigen Minuten einen maßgeschneiderten Lebenslauf

Überzeuge Recruiter und verdiene mehr Geld. Mehr erfahren

Zusammenfassung

A leading IT solutions provider in Düsseldorf is seeking a Data Engineer. Responsibilities include designing data solutions using knowledge graphs, analyzing diverse documents for transformations, creating ETL pipelines, and ensuring data quality controls. The ideal candidate should have a Master’s degree in applied mathematics or computer science, strong Python skills, and practical experience with data pipelines. The company offers a competitive salary and hybrid work model, with additional benefits like vacation days and a company pension scheme.

Leistungen

Up to 30 days of vacation
Company phone for personal use
Employee Assistance Program (EAP)
Company ticket for public transport
Bicycle leasing option
Hybrid work model

Qualifikationen

  • Strong knowledge of applied graph theory and knowledge graph solutions.
  • Several years of experience as Data Engineer & Data Scientist.
  • Excellent practical knowledge of Python, especially Pydantic.
  • Strong background in Airflow DAG development and orchestration.
  • Experience with Docker Compose and building custom images.
  • Strong practical knowledge of PostgreSQL, Neo4j & Cypher.
  • Experience with rapid experimentation using Jupyter.
  • Proven track record of secure data handling and CI/CD pipeline setup.
  • Fluent in German and English, excellent communication skills.

Aufgaben

  • Design and build a data solution using a knowledge graph.
  • Analyze documents in different formats as sources for knowledge graphs.
  • Create ETL pipelines and data quality controls.
  • Document the solution and collaborate with subject matter experts.

Kenntnisse

Applied graph theory
Data pipeline design
Python (Pydantic)
Airflow DAG development
PostgreSQL
Neo4j & Cypher
Jupyter for experimentation
Data extraction from documents
ML tools
Docker

Ausbildung

Master's degree in applied mathematics or computer science

Tools

Apache Hive
S3
Hadoop
Spark
AWS
Kafka
Scala
Jobbeschreibung

COCUS is all about People! We are proud to deliver skilled services and products developed by great talent with attitude and ambition to work in innovative IT solutions. We are partnering with worldwide industry leaders and always looking for the brightest minds to have fun working Digital & Cloud Solutions Data Services Connectivity and Security. Emotions are part of us we encourage everyone to be what they truly are in our collaborative informal transparent and open environment where everyone can contribute to the path to achieve our goals as a Team!

For our customer in German heavy industry were creating a Data solution based on data engineering tools to collect knowledge out of structured and unstructured documents in different formats and transform it into a knowledge graph for further storage and usage in graph database.

Responsibilities
  • Designing & building data solution with knowledge graph (graph modelling relationship extraction querying etc).
  • Analyzing various documents in different formats to use as a source for the knowledge graph solution.
  • Designing & implementing data models for relational and graph databases.
  • Creating ETL pipelines.
  • Designing and implementing data quality controls for structured and semi-structured data sources.
  • Prompt design vector store setup inference service orchestration.
  • Exploratory data analysis & experimentation prototyping model evaluation.
  • CI / CD & Automation: Docker build image tagging automated testing.
  • Creating documentation of the solution. Collaborate with subject matter experts on the customer side and data scientists in the team to design and build the solution.
  • An informal and friendly corporate culture that rewards innovation and teamwork through global projects in a fast‑paced environment with direct impact on solutions and applications.
Qualifications & Experience
  • Strong knowledge of applied graph theory and understanding of knowledge graph solutions.
  • Several years of experience as Data Engineer & Data Scientist with designing and building complex data pipelines and analyzing data sets for various use cases.
  • Excellent practical knowledge of Python especially Pydantic.
  • Strong background in Airflow DAG development and orchestration; other ETL tools would be a big plus.
  • Experience with Docker Compose and building custom images.
  • Strong practical knowledge of PostgreSQL Neo4j & Cypher, Redis and Celery.
  • Experience with rapid experimentation using Jupyter.
  • Proven track record of secure data handling (encryption Git‑Secret) and CI / CD pipeline setup.
  • Understanding of local LLMs including embedding engineering and vector stores.
  • Outstanding documentation skills.
  • Great communication skills.
  • Active mindset and willingness to learn.
  • Fluent German and English, excellent written and verbal.
  • Masters degree in applied mathematics and computer science.
  • Knowledge of various ETL / ELT tools like dbt, NiFi, Knime.
  • Hands‑on experience with data extraction from various document formats.
  • Practical experience with ML tools (real business use cases).
  • Experience of usage LLMs for knowledge graphs.
Key Skills
  • Apache Hive
  • S3
  • Hadoop
  • Redshift
  • Spark
  • AWS
  • Apache Pig
  • NoSQL
  • Big Data
  • Data Warehouse
  • Kafka
  • Scala
Employment & Benefits

Employment Type: Full‑Time

Experience: years

Vacancy: 1

  • Permanent employment contract and a competitive market‑aligned salary based on your experience.
  • Two pet‑friendly offices in Germany.
  • Company phone for personal use.
  • Up to 30 days of vacation.
  • Employee Assistance Program (EAP).
  • Company ticket for public transportation and available parking spaces.
  • Benefit from a company pension scheme.
  • Bicycle leasing option with a company subsidy.
  • Hybrid work model with 2 days of home office per week.
  • Referral program with a bonus to invite a friend to join the team.
Hol dir deinen kostenlosen, vertraulichen Lebenslauf-Check.
eine PDF-, DOC-, DOCX-, ODT- oder PAGES-Datei bis zu 5 MB per Drag & Drop ablegen.