Why Did JetBrains Create Mellum?

Mellum is a purpose-built language model trained from scratch to do one job well: code completion.

View at original site


❓Why did JetBrains create a purpose-built LLM?

Because not every model needs to be a generalist.

Mellum is a purpose-built language model trained from scratch to do one job well: code completion. It’s fast, lightweight, and focused, prioritizing depth over breadth.

🔎 It's what we're calling a “focal model”.

And this is just the beginning. Mellum is the first in a growing family of focal models where each is designed around a specific developer task, from code completion to diff prediction and beyond.

🤗 And now, we’ve open-sourced Mellum on Hugging Face to invite transparency, collaboration, and contribution.

Researchers, engineers, and curious developers like you can explore the model, fine-tune it, and help shape what comes next 🤝

💻 Try it in your JetBrains IDE 🔬 Explore it on Hugging Face

Related Resources

Introducing Mellum: JetBrains' New LLM Built for Developers
Introducing Mellum: JetBrains' New LLM Built for Developers
JetBrains launches Mellum, a proprietary large language model (LLM) specifically designed to assist software developers.
Prompt AI Directly in the Editor
Prompt AI Directly in the Editor
With PyCharm, you now have the support of AI Assistant at your fingertips. You can interact with it right where you do most of your work – in the editor.
AI and Kotlin: A Perfect Mix
AI and Kotlin: A Perfect Mix
The latest advancements in AI coding assistance by JetBrains AI for Kotlin in your IDE.