
Activez les alertes d’offres d’emploi par e-mail !
Générez un CV personnalisé en quelques minutes
Décrochez un entretien et gagnez plus. En savoir plus
A European AI startup in Paris is looking for a hands-on engineer with recent experience in training neural networks on NVIDIA clusters with H100 GPUs. The role involves profiling pretraining runs, optimizing scheduling, and implementing practices to enhance performance on large GPU jobs. This freelance mission can lead to a full-time offer based on strong performance, with remote-friendly options and competitive compensation. Join a dynamic team and influence the engineering culture in a prime tech location.
Kog is a European VC-funded startup and real-time AI frontier lab building the world’s fastest AI execution layer, part of the 2030 French Tech cohort.
We are not just optimizing existing libraries; we are bypassing inefficient abstraction layers to rewrite the rules of AI inference. By coding at the Assembly level on high-end GPUs (starting with the AMD MI300X), we unlock raw performance that standard stacks leave on the table.
Our Mission: To enable true real-time AI. We are targeting 10x performance gains through a combination of low-level GPU mastery and novel model architecture. Our goal is to build the sovereign infrastructure that will power the next generation of collaborative AI agents.
Why join now? We have already achieved a 3x to 10x speedup compared to state-of-the-art alternatives (vLLM, TensorRT-LLM) by making breakthroughs in:
We have access to a large NVIDIA H100 training cluster with 200+ GPUs. Our immediate priority is to optimize LLM pretraining efficiency on this cluster. We are competent with our current setup, but not yet at the level we need. We want a hands‑on engineer who already runs pretraining at scale elsewhere and can quickly profile, correct, and document best practices so our team can execute independently.