--- license: mit pipeline_tag: unconditional-image-generation --- # RAE: Diffusion Transformers with Representation Autoencoders This repository contains the official PyTorch checkpoints for Representation Autoencoders. Representation Autoencoders (RAE) are a class of autoencoders that utilize pretrained, frozen representation encoders such as DINOv2 and SigLIP2 as encoders with trained ViT decoders. RAE can be used in a two-stage training pipeline for high-fidelity image synthesis, where a Stage 2 diffusion model is trained on the latent space of a pretrained RAE to generate images. Website: https://rae-dit.github.io/ Code: https://github.com/bytetriper/RAE Paper: https://huggingface.co/papers/2510.11690