| name: flexTokenizer | |
| embedding_dim: 8 | |
| hidden_channels: 128 | |
| channel_multipliers: | |
| - 1 | |
| - 2 | |
| - 2 | |
| - 4 | |
| encoder_layer_configs: | |
| - 2 | |
| - 2 | |
| - 2 | |
| - 2 | |
| - 2 | |
| decoder_layer_configs: | |
| - 2 | |
| - 2 | |
| - 2 | |
| - 2 | |
| - 2 | |
| use_adaptive_norm: true | |
| use_learnable_up_down_sample: true | |
| quantizer_config: | |
| quantize_type: vq | |
| embed_dim: 8 | |
| num_embed: 65536 | |
| commitment_loss_weight: 0.25 | |
| use_l2_norm: true | |
| use_uniform_init: false | |