CogVideoX-2b / transformer /config.json
Ubuntu
init
91597cf
raw
history blame
713 Bytes
{
"_class_name": "CogVideoXTransformer3DModel",
"_diffusers_version": "0.30.0.dev0",
"activation_fn": "gelu-approximate",
"attention_bias": true,
"attention_head_dim": 64,
"dropout": 0.0,
"flip_sin_to_cos": true,
"freq_shift": 0,
"in_channels": 16,
"max_text_seq_length": 226,
"norm_elementwise_affine": true,
"norm_eps": 1e-05,
"num_attention_heads": 30,
"num_layers": 30,
"out_channels": 16,
"patch_size": 2,
"sample_frames": 49,
"sample_height": 60,
"sample_width": 90,
"spatial_interpolation_scale": 1.875,
"temporal_compression_ratio": 4,
"temporal_interpolation_scale": 1.0,
"text_embed_dim": 4096,
"time_embed_dim": 512,
"timestep_activation_fn": "silu"
}