DiT( (x_embedder): PatchEmbed( (proj): Conv2d(32, 1152, kernel_size=(1, 1), stride=(1, 1)) (norm): Identity() ) (t_embedder): TimestepEmbedder( (mlp): Sequential( (0): Linear(in_features=256, out_features=1152, bias=True) (1): SiLU() (2): Linear(in_features=1152, out_features=1152, bias=True) ) ) (y_embedder): LabelEmbedder( (embedding_table): Embedding(1001, 1152) ) (blocks): ModuleList( (0-27): 28 x DiTBlock( (norm1): LayerNorm((1152,), eps=1e-06, elementwise_affine=False) (attn): Attention( (qkv): Linear(in_features=1152, out_features=3456, bias=True) (attn_drop): Dropout(p=0.0, inplace=False) (proj): Linear(in_features=1152, out_features=1152, bias=True) (proj_drop): Dropout(p=0.0, inplace=False) ) (norm2): LayerNorm((1152,), eps=1e-06, elementwise_affine=False) (mlp): Mlp( (fc1): Linear(in_features=1152, out_features=4608, bias=True) (act): GELU(approximate='tanh') (drop1): Dropout(p=0, inplace=False) (fc2): Linear(in_features=4608, out_features=1152, bias=True) (drop2): Dropout(p=0, inplace=False) ) (adaLN_modulation): Sequential( (0): SiLU() (1): Linear(in_features=1152, out_features=6912, bias=True) ) ) ) (final_layer): FinalLayer( (norm_final): LayerNorm((1152,), eps=1e-06, elementwise_affine=False) (linear): Linear(in_features=1152, out_features=64, bias=True) (adaLN_modulation): Sequential( (0): SiLU() (1): Linear(in_features=1152, out_features=2304, bias=True) ) ) )