<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Knowledge Distillation | Tobias Nauen</title><link>https://nauen-it.de/tags/knowledge-distillation/</link><atom:link href="https://nauen-it.de/tags/knowledge-distillation/index.xml" rel="self" type="application/rss+xml"/><description>Knowledge Distillation</description><generator>HugoBlox Kit (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Wed, 11 Mar 2026 00:00:00 +0000</lastBuildDate><image><url>https://nauen-it.de/media/icon.svg</url><title>Knowledge Distillation</title><link>https://nauen-it.de/tags/knowledge-distillation/</link></image><item><title>TextTeacher: What Can Language Teach About Images?</title><link>https://nauen-it.de/publications/text-teacher/</link><pubDate>Wed, 11 Mar 2026 00:00:00 +0000</pubDate><guid>https://nauen-it.de/publications/text-teacher/</guid><description>&lt;h1 id="introduction"&gt;Introduction&lt;/h1&gt;
&lt;p&gt;What if a language model could teach a vision model — and then step aside?&lt;/p&gt;
&lt;p&gt;TextTeacher adds a frozen text encoder and a lightweight alignment loss to standard image classification training.
Image captions provide per-instance semantic targets that shape the visual representation during training.
At inference, everything text-related is discarded.
The deployed model is a plain, fast vision network — no extra parameters, no added latency.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The result:&lt;/strong&gt; Up to +2.7 p.p. on ImageNet, +1.0 p.p. average transfer gain across six fine-grained benchmarks, and +8.4 p.p. under 50% label noise — all with negligible compute overhead and without any multimodal pretraining of the target model.
TextTeacher distills knowledge 33% more efficient than traditional vision-distillation.&lt;/p&gt;
&lt;h1 id="how-it-works"&gt;How it works&lt;/h1&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="TextTeacher method overview"
srcset="https://nauen-it.de/publications/text-teacher/featured_hu_e1d41f02400fecee.webp 320w, https://nauen-it.de/publications/text-teacher/featured_hu_5131e8de0a4a84b0.webp 480w, https://nauen-it.de/publications/text-teacher/featured_hu_887442f9c58fa4b6.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/text-teacher/featured_hu_e1d41f02400fecee.webp"
width="760"
height="263"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;p&gt;A lightweight text head projects the backbone&amp;rsquo;s CLS token into the embedding space of a frozen text encoder.
An auxiliary contrastive loss aligns each image&amp;rsquo;s projection with the text embedding of its caption, pulling image features toward a pre-organised semantic manifold.
The text head and encoder are dropped after training.&lt;/p&gt;
&lt;h1 id="results"&gt;Results&lt;/h1&gt;
&lt;h2 id="textteacher-on-imagenet"&gt;TextTeacher on ImageNet&lt;/h2&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="ImageNet results"
srcset="https://nauen-it.de/publications/text-teacher/images/imagenet-results_hu_1ab762c0b87d8928.webp 320w, https://nauen-it.de/publications/text-teacher/images/imagenet-results_hu_13a61e6a4fd70312.webp 480w, https://nauen-it.de/publications/text-teacher/images/imagenet-results_hu_523bc7261e57a01f.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/text-teacher/images/imagenet-results_hu_1ab762c0b87d8928.webp"
width="760"
height="752"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
Text teacher increases performance across models without additional training-time overhead.&lt;/p&gt;
&lt;h2 id="comparison-of-knowledge-distillation-methods"&gt;Comparison of Knowledge Distillation Methods&lt;/h2&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="Comparison of Baselines"
srcset="https://nauen-it.de/publications/text-teacher/images/baselines-comparison_hu_c4d97de0e820f2b.webp 320w, https://nauen-it.de/publications/text-teacher/images/baselines-comparison_hu_2ce4bdfa47931b1d.webp 480w, https://nauen-it.de/publications/text-teacher/images/baselines-comparison_hu_2492d84f2b1dacb3.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/text-teacher/images/baselines-comparison_hu_c4d97de0e820f2b.webp"
width="760"
height="730"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;p&gt;TextTeacher runs at nearly the same per-epoch cost as baseline classification (32 min/epoch vs. 31 min/epoch), while online knowledge distillation from DINOv2-L costs ~48 min/epoch. In a compute-matched setting, TextTeacher (79.1%) matches DINOv2-L online distillation trained for the full 100 epochs (79.1%) at only ~66% the wall-clock time. Adding the full preprocessing (captioning + embedding the training set) still saves ≈6 GPU-hours over 300-epoch online distillation.&lt;/p&gt;
&lt;h2 id="analysis-how-does-textteacher-work"&gt;Analysis: How does TextTeacher work?&lt;/h2&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="Training with Jump schedules"
srcset="https://nauen-it.de/publications/text-teacher/images/jump-schedules_hu_f1c8d25f03783b8d.webp 320w, https://nauen-it.de/publications/text-teacher/images/jump-schedules_hu_dcf58e1523e10bcd.webp 480w, https://nauen-it.de/publications/text-teacher/images/jump-schedules_hu_ec69f79d77bcf2ac.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/text-teacher/images/jump-schedules_hu_f1c8d25f03783b8d.webp"
width="760"
height="169"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
Left: Fréchet feature distance (FFD) to the final model over training.
With TextTeacher active from epoch 0, the representation is already close to its final semantic configuration by epoch 10 (FFD 65 vs. 450 for the baseline).
Right: Accuracy when TextTeacher is dropped at different epochs (jump schedules).
Benefits saturate around epoch 50; dropping very late (epochs 70–90) hurts stability.&lt;/p&gt;
&lt;p&gt;
&lt;figure &gt;
&lt;div class="flex justify-center "&gt;
&lt;div class="w-full" &gt;
&lt;img alt="cka similarity"
srcset="https://nauen-it.de/publications/text-teacher/images/cka-similarity_hu_4c2812521285e773.webp 320w, https://nauen-it.de/publications/text-teacher/images/cka-similarity_hu_4c64765a080ac0a8.webp 480w, https://nauen-it.de/publications/text-teacher/images/cka-similarity_hu_5b33320ad5291e02.webp 760w"
sizes="(max-width: 480px) 100vw, (max-width: 768px) 90vw, (max-width: 1024px) 80vw, 760px"
src="https://nauen-it.de/publications/text-teacher/images/cka-similarity_hu_4c2812521285e773.webp"
width="760"
height="377"
loading="lazy" data-zoomable /&gt;&lt;/div&gt;
&lt;/div&gt;&lt;/figure&gt;
TextTeacher substantially increases cross-seed similarity in deeper attention layers, indicating it organises higher-level semantic subspaces while leaving low-level feature extractors flexible.&lt;/p&gt;
&lt;p&gt;Thus, TextTeacher acts primarily as an early-phase preconditioner: most of the benefit is realised in the first 30–50 epochs, where it accelerates the formation of semantically organised features — especially in deeper layers.&lt;/p&gt;
&lt;h1 id="contributions"&gt;Contributions&lt;/h1&gt;
&lt;ul&gt;
&lt;li&gt;We pose and investigate a fundamental question: Can the semantic knowledge of a language model efficiently improve a vision model? — operationalised via a decoupled auxiliary signal from a frozen text encoder.&lt;/li&gt;
&lt;li&gt;We introduce TextTeacher, a general approach that efficiently injects textual knowledge into standard vision backbones using an auxiliary alignment loss, requiring no text at inference.&lt;/li&gt;
&lt;li&gt;We provide extensive experiments showing improvements in accuracy and transfer, and demonstrate that language-based guidance outperforms visual guidance and knowledge distillation in a compute-matched setting.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Associated Projects:&lt;/strong&gt;
,
&lt;/p&gt;</description></item></channel></rss>