Pretrained

DeepSeek's New LLM Architectures


Listen Later

Pierce and Richard break down DeepSeek's latest model architecture moves in Manifold-Constrained Hyper Connections and Engram memory. Are these conceptually sound? Will they hop the pond over to US frontier labs?

...more
View all episodesView all episodes
Download on the App Store

PretrainedBy Pierce Freeman & Richard Diehl Martinez