Internships
Register
Copyright © 2000—2026 JetBrains s.r.o.

Applied Research: Knowledge Distillation for Coding LLMs.

Description

The Mellum team aims to train the best possible code models for AI Assistant features, including completion, next-edit suggestion, AI chat, and agents. Our key constraint is inference efficiency: we want to offer our models to users at no additional cost. This project explores knowledge distillation methods to elicit higher-quality responses from small models.

All development will be done in our Mellum code repository, where we use NVidia Nemotron-Bridge for training, interfacing with a k8s GPU cluster.

Responsibilities Include

  • Apply knowledge distillation in both pre- and post-training settings.

  • Survey and interpret literature on knowledge distillation.

  • Design and run experiments on real models with real training infrastructure.

  • Evaluate results and iterate on findings.

Requirements

  • Solid foundational understanding of deep learning.

  • Experience with transformer-based language models, and frameworks like pytorch.

  • Good programming skills in Python.

  • Working proficiency with GPU clusters.

Admission

Internship Projects Fall 2026

Contact details

internship@jetbrains.com

Preferred internship location

Czechia
Germany
Netherlands
Poland
Serbia
Spain
UK

Technologies

Big Data Stack
Deep learning
Docker

Area

Data Science
Machine Learning
Research

Internship timing preferences

Flexible start
Full-time preferable
Part-time acceptable

Candidate graduation status

Final-year students preferred

Additional information

Potential thesis
Applications by 08.05.2026
Interview by 30.05.2026
Feedback and final results by 12.06.2026