bezzam HF Staff commited on
Commit
cd4ea7e
·
verified ·
1 Parent(s): 0c98e36

Upload VibeVoiceForConditionalGeneration

Browse files
config.json CHANGED
@@ -1,7 +1,9 @@
1
  {
2
- "_attn_implementation_autoset": false,
3
  "acostic_vae_dim": 64,
4
  "acoustic_tokenizer_config": {
 
 
 
5
  "bias": true,
6
  "channels": 1,
7
  "depths": [
@@ -21,6 +23,7 @@
21
  5,
22
  8
23
  ],
 
24
  "ffn_expansion": 4,
25
  "hidden_act": "gelu",
26
  "hidden_size": 64,
@@ -56,6 +59,9 @@
56
  "model_type": "vibevoice",
57
  "pad_token_id": 151643,
58
  "semantic_tokenizer_config": {
 
 
 
59
  "bias": true,
60
  "channels": 1,
61
  "depths": [
@@ -75,6 +81,7 @@
75
  5,
76
  8
77
  ],
 
78
  "ffn_expansion": 4,
79
  "hidden_act": "gelu",
80
  "hidden_size": 128,
 
1
  {
 
2
  "acostic_vae_dim": 64,
3
  "acoustic_tokenizer_config": {
4
+ "architectures": [
5
+ "VibeVoiceAcousticTokenizerModel"
6
+ ],
7
  "bias": true,
8
  "channels": 1,
9
  "depths": [
 
23
  5,
24
  8
25
  ],
26
+ "dtype": "bfloat16",
27
  "ffn_expansion": 4,
28
  "hidden_act": "gelu",
29
  "hidden_size": 64,
 
59
  "model_type": "vibevoice",
60
  "pad_token_id": 151643,
61
  "semantic_tokenizer_config": {
62
+ "architectures": [
63
+ "VibeVoiceSemanticTokenizerModel"
64
+ ],
65
  "bias": true,
66
  "channels": 1,
67
  "depths": [
 
81
  5,
82
  8
83
  ],
84
+ "dtype": "bfloat16",
85
  "ffn_expansion": 4,
86
  "hidden_act": "gelu",
87
  "hidden_size": 128,
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42bfa2acc7208cff6373d24c19428577ac9aa9ec1032de42beec917d604c20aa
3
- size 4877662532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be3ff3f8377beca46483320c334beb384cd95a8149f65c2903be78d40bc22b75
3
+ size 4877662484
model.safetensors.index.json CHANGED
@@ -4,6 +4,8 @@
4
  "total_size": 18686710726
5
  },
6
  "weight_map": {
 
 
7
  "lm_head.weight": "model-00004-of-00004.safetensors",
8
  "model.acoustic_connector.fc1.bias": "model-00004-of-00004.safetensors",
9
  "model.acoustic_connector.fc1.weight": "model-00004-of-00004.safetensors",
@@ -1206,8 +1208,6 @@
1206
  "model.semantic_tokenizer.encoder.stages.6.7.gamma": "model-00004-of-00004.safetensors",
1207
  "model.semantic_tokenizer.encoder.stages.6.7.mixer.conv.bias": "model-00004-of-00004.safetensors",
1208
  "model.semantic_tokenizer.encoder.stages.6.7.mixer.conv.weight": "model-00004-of-00004.safetensors",
1209
- "model.semantic_tokenizer.encoder.stages.6.7.norm.weight": "model-00004-of-00004.safetensors",
1210
- "model.speech_bias_factor": "model-00001-of-00004.safetensors",
1211
- "model.speech_scaling_factor": "model-00001-of-00004.safetensors"
1212
  }
1213
  }
 
4
  "total_size": 18686710726
5
  },
6
  "weight_map": {
7
+ "latent_bias_factor": "model-00001-of-00004.safetensors",
8
+ "latent_scaling_factor": "model-00001-of-00004.safetensors",
9
  "lm_head.weight": "model-00004-of-00004.safetensors",
10
  "model.acoustic_connector.fc1.bias": "model-00004-of-00004.safetensors",
11
  "model.acoustic_connector.fc1.weight": "model-00004-of-00004.safetensors",
 
1208
  "model.semantic_tokenizer.encoder.stages.6.7.gamma": "model-00004-of-00004.safetensors",
1209
  "model.semantic_tokenizer.encoder.stages.6.7.mixer.conv.bias": "model-00004-of-00004.safetensors",
1210
  "model.semantic_tokenizer.encoder.stages.6.7.mixer.conv.weight": "model-00004-of-00004.safetensors",
1211
+ "model.semantic_tokenizer.encoder.stages.6.7.norm.weight": "model-00004-of-00004.safetensors"
 
 
1212
  }
1213
  }