Website | NVIDIA News | NVIDIA Blog | Hugging Face | YouTube
Cosmos Tokenizer is a suite of visual tokenizers for images and videos that delivers various compression rates while maintaining high reconstruction quality. Cosmos Tokenizer can serve as an effective and efficient building block in both diffusion-based and autoregressive models for image and video generation. This model is ready for commercial use.
Our tokenizers come in two types: Continuous (C) and Discrete (D), each with Image (I) and Video (V) variants:
Continuous ( C ) | Discrete ( D ) | |
---|---|---|
Images ( I ) | Cosmos-Tokenizer-CI | Cosmos-Tokenizer-DI |
Videos ( V ) | Cosmos-Tokenizer-CV | Cosmos-Tokenizer-DV |
Given an image or a video, Cosmos Tokenizer outputs either continuous latents or discrete tokens. Cosmos Tokenizer achieves spatial compression rates of 8x8 or 16x16 and temporal compression factors of 4x or 8x, resulting in a total compression factor of up to 2048x (=8x16x16). Cosmos Tokenizer delivers 8x more total compression than state-of-the-art (SOTA) methods while simultaneously maintaining higher image quality and running up to 12x faster than the best available SOTA tokenizers.
Model Developer: NVIDIA
This release (v1.0) of Cosmos Tokenizer includes the following tokenizers:
The previous release (v0.1) of Cosmos Tokenizer included the following tokenizers:
Under the NVIDIA Open Model License, NVIDIA confirms:
We designed Cosmos Tokenizer using a lightweight and computationally efficient architecture, featuring a temporally causal design. Specifically, we employ causal temporal convolution and causal temporal attention layers to preserve the natural temporal order of video frames, ensuring seamless tokenization of images and videos using a single unified network architecture. The encoder and decoder form a symmetrical pair, which are mirrors of each other. The encoder starts with a 2-level Haar wavelet transform layer, which down-samples inputs by a factor of 4 in both spatial and temporal dimensions. Likewise, the decoder ends with an inverse wavelet transform. We employ the vanilla autoencoder (AE) formulation to model the latent space for continuous tokenizers. For discrete tokenizers, we adopt the Finite-Scalar-Quantization (FSQ) as the latent space quantizer.
Input
Output
Input
Output
Runtime Engine(s):
Supported Hardware Microarchitecture Compatibility:
Note: We have only tested Cosmos Tokenizer with BF16 precision on Ampere and Hopper GPUs. If you are using older versions of NVIDIA GPUs (e.g., NVIDIA Volta GPUs), you may need to switch to FP32 precision.
Operating System(s):
Inference Engines:
Cosmos-Tokenizer
(PyTorch)Cosmos-Tokenizer
Note: Currently, the Cosmos-Tokenizer
code is only supported on Linux.
Please clone the Cosmos-Tokenizer
from GitHub repo github.com/NVIDIA/Cosmos-Tokenizer.
git clone https://github.com/NVIDIA/Cosmos-Tokenizer.git
cd Cosmos-Tokenizer
Install dependencies
pip3 install -r requirements.txt
apt-get install -y ffmpeg
Preferably, you could build a docker image using our provided Dockerfile.
docker build -t cosmos-docker -f Dockerfile.
# You can run the container as:
docker run --gpus all -it --rm -v /home/${USER}:/home/${USER} \
--workdir ${PWD} cosmos-docker /bin/bash
Create a local directory for the pre-trained checkpoints and download the pre-trained checkpoints
Under the checkpoint directory pretrained_ckpts/<model-name>
, we provide the encoder,
decoder and the full autoencoder JIT models.
├── pretrained_ckpts/
│ ├── Cosmos-Tokenizer-1.0-CV8x8x8/
│ │ ├── encoder.jit
│ │ ├── decoder.jit
│ │ ├── autoencoder.jit
│ ...
You can use the following example commands to encode and decode images or videos. For each, the same command works for both continuous and discrete tokenization. Simply provide the proper JIT-compiled ckpt to checkpoint_enc
, checkpoint_dec
, or the full autoencoder ckpt to checkpoint
.
import torch
from cosmos_tokenizer.video_lib import CausalVideoTokenizer
model_name = "Cosmos-Tokenizer-1.0-CV8x8x8"
input_tensor = torch.randn(1, 3, 9, 512, 512).to('cuda').to(torch.bfloat16) # [B, C, T, H, W]
encoder = CausalVideoTokenizer(checkpoint_enc=f'pretrained_ckpts/{model_name}/encoder.jit')
(latent,) = encoder.encode(input_tensor)
torch.testing.assert_close(latent.shape, (1, 16, 2, 64, 64))
# The input tensor can be reconstructed by the decoder as:
decoder = CausalVideoTokenizer(checkpoint_dec=f'pretrained_ckpts/{model_name}/decoder.jit')
reconstructed_tensor = decoder.decode(latent)
torch.testing.assert_close(reconstructed_tensor.shape, input_tensor.shape)
The indices
will have the shape (1, 2, 64, 64)
and contain integral values in the range [1..64K]
, where the first of the three integral maps represents the first frame.
The codes
will contain the pre-quantization continuous latent with shape (1, 6, 2, 64, 64)
, where C=6 represents the number of FSQ levels.
Note: More inference usage commands, including both TorchScript (JIT) and PyTorch Inference APIs on real images and videos, can be found on our GitHub repository github.com/NVIDIA/Cosmos-Tokenizer.
We have evaluated the additional Cosmos Tokenizer models on DAVIS video benchmark dataset.
Tokenizer | Compression Ratio | Height | Num. of Frames | Quantization | PSNR (DAVIS) | SSIM (DAVIS) | rFVD (DAVIS) |
---|---|---|---|---|---|---|---|
CogVideoX | 4×4×4 | - | - | VAE | 31.74 | 0.860 | 19.58 |
OmniTokenizer | 4×8×8 | - | - | VAE | 29.04 | 0.710 | 117.66 |
Cosmos-Tokenizer-CV | 4×8×8 | 720 | 49 | AE | 35.28 | 0.890 | 15.93 |
Cosmos-Tokenizer-CV | 8×8×8 | 720 | 49 | AE | 34.10 | 0.850 | 30.16 |
Cosmos-Tokenizer-CV | 8×8×8 | 720 | 121 | AE | 34.32 | 0.867 | 23.49 |
Cosmos-Tokenizer-CV | 8×16×16 | 720 | 49 | AE | 32.55 | 0.770 | 93.82 |
The following table shows the number of parameters and the averaged encoding and decoding times per image or video frame, measured on a single A100 80GB GPU. For comparison, we also list the parameters and average speeds of prior state-of-the-art tokenizer(s) with the same compression ratio.
Tokenizer | Resolution | Compression Ratio | Parameters | Time (ms) |
---|---|---|---|---|
CogVideoX | 720x1280 | 4×8×8 | 216M | 414 |
OmniTokenizer | 720x1280 | 4×8×8 | 54M | 82.9 |
Cosmos-Tokenizer-CV | 720x1280 | 4×8×8 | 105M | 34.8 |
Note: We benchmarked the runtime for images under the 8x8 compression and videos under the 4×8×8 compression. Tokenizers with different compression ratios are not included in this comparison.
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
For more detailed information on ethical considerations for this model, please see the subcards of Explainability, Bias, Safety & Security, and Privacy below. Please report security vulnerabilities or NVIDIA AI Concerns here.