Skim Logo
Dwarkesh PatelApril 30, 2026
How GPT-5, Claude, and Gemini are actually trained and served – Reiner Pope
2:13:40
DP

How GPT-5, Claude, and Gemini are actually trained and served – Reiner Pope

RevNets: Applying Invertibility to Neural Networks — Dwarkesh Patel

From How GPT-5, Claude, and Gemini are actually trained and served – Reiner Pope. Category: Tech. Format: Commentary. This is a single keypoint from the analysis.

The RevNets paper from 2017 demonstrates how the Feistel construction can be applied to any neural network, including transformers, to make the entire network invertible. This allows for the rematerialization of activations during the backward pass, drastically cutting down the memory needed for training.

Impact: High. This architectural innovation directly addresses a major bottleneck in training large neural networks, offering a path to greater efficiency and scalability by optimizing memory usage.

In the source video, this keypoint occurs from 02:10:11 to 02:12:00.

Sources in support: Dwarkesh Patel (Host)

For the full credibility analysis, key takeaways, and other keypoints from this video, see the full analysis on skim.

This keypoint analysis was generated by skim (skim.plus), an AI-powered content analysis platform by Credible AI.