
Sign up to save your podcasts
Or


In this episode, Katherine Forrest and Scott Caravello break down three generative AI architectures—transformers, JEPA, and diffusion models—exploring what sets each apart and how they overlap. They also discuss Manifold-Constrained Hyper-Connections, a recent innovation aimed at improving how transformer layers communicate during training.
For the sources referenced in this episode, please see the links below:
DeepSeek AI: mHC: Manifold-Constrained Hyper Connections
##
Learn More About Paul, Weiss’s Artificial Intelligence practice:
By Paul, Weiss4.8
2323 ratings
In this episode, Katherine Forrest and Scott Caravello break down three generative AI architectures—transformers, JEPA, and diffusion models—exploring what sets each apart and how they overlap. They also discuss Manifold-Constrained Hyper-Connections, a recent innovation aimed at improving how transformer layers communicate during training.
For the sources referenced in this episode, please see the links below:
DeepSeek AI: mHC: Manifold-Constrained Hyper Connections
##
Learn More About Paul, Weiss’s Artificial Intelligence practice:

91,297 Listeners

6,881 Listeners

30,609 Listeners

2,418 Listeners

9,724 Listeners

113,121 Listeners

369,956 Listeners

10,331 Listeners

5,832 Listeners

10,254 Listeners

5,576 Listeners

16,525 Listeners

688 Listeners

1,480 Listeners

12,559 Listeners