Distillation
A technique for training a smaller, faster model to mimic the behavior of a larger, more capable one — trading some performance for dramatically lower cost and latency.
Loading...
Related terms
Read more
Last updated 2026-05-12
A technique for training a smaller, faster model to mimic the behavior of a larger, more capable one — trading some performance for dramatically lower cost and latency.
Related terms
Read more
Last updated 2026-05-12