DevConf.CZ 2026

Fine-tuning a small model for style/vibe (a Kimi distillation and beyond)
2026-06-19 , D105 (capacity 300)

Edge-runnable smaller language models have a significant appeal to privacy-conscious technical users.

Kimi K2 Instruct, an extremely large language model (1T parameter MoE), has developed a "vibe" uniquely attractive to technical users. with low sycophancy and high creativity.

Attempting to distill a heavy "vibe" into a 1.5B model (IBM Granite 4-h Nano) has run into a number of issues but produced interesting results, with different training methods and optimizers tried.


Experience level: Intermediate - attendees should be familiar with the subject

Principal Technical Writer at Red Hat in Ireland, with lots of side interests