@@ -65,7 +65,7 @@ <h1 class="title is-1 publication-title">The Diffusion Duality</h1>
6565 < span class ="author-block ">
6666 < a href ="https://skylion007.github.io " target ="_blank "> Aaron Gokaslan< sup > 1</ sup > </ a > ,</ span >
6767 < span class ="author-block ">
68- < a href ="https://tech.cornell.edu/people/guanghan-wang/ " target ="_blank "> Guanghan Wang< sup > 1</ sup > </ a > ,</ span >
68+ < a href ="https://www.guanghanwang.com " target ="_blank "> Guanghan Wang< sup > 1</ sup > </ a > ,</ span >
6969 < span class ="author-block ">
7070 < a href ="https://justinchiu.netlify.app " target ="_blank "> Justin Chiu< sup > 3</ sup > </ a > ,</ span >
7171 < span class ="author-block ">
@@ -201,26 +201,24 @@ <h4 class="subtitle has-text-centered">
201201<!-- Paper abstract -->
202202< section class ="section hero is-light ">
203203 < div class ="container is-max-desktop ">
204- < div class ="columns is-centered has-text-centered ">
204+ < div class ="columns is-centered ">
205205 < div class ="column is-four-fifths ">
206- < h2 class ="title is-3 "> Our Contributions</ h2 >
207- < div class ="content has-text-justified ">
208- < ol >
209- < li >
210- We show that < b > uniform-state discrete diffusion emerges from Gaussian diffusion</ b > ,
211- enabling the transfer of techniques from continuous to discrete domains.
212- </ li >
213- < li >
214- Building on this insight, we propose the DUO framework,
215- which improves training through a low-variance curriculum.
216- </ li >
217- < li >
218- We further introduce Discrete Consistency Distillation, adapting consistency
219- distillation to the discrete setting and accelerating DUO sampling
220- by < b > two orders</ b > of magnitude.
221- </ li >
222- </ ol >
223- </ div >
206+ < h2 class ="title is-3 "> Key Innovations</ h2 >
207+ < ol >
208+ < li >
209+ We show that < b > uniform-state discrete diffusion emerges from Gaussian diffusion</ b > ,
210+ enabling the transfer of techniques from continuous to discrete domains.
211+ </ li >
212+ < li >
213+ Building on this insight, we propose the DUO framework,
214+ which improves training through a low-variance curriculum.
215+ </ li >
216+ < li >
217+ We further introduce Discrete Consistency Distillation, adapting consistency
218+ distillation to the discrete setting and accelerating DUO sampling
219+ by < b > two orders</ b > of magnitude.
220+ </ li >
221+ </ ol >
224222 </ div >
225223 </ div >
226224 </ div >
0 commit comments