Update audioldm/clap/encoders.py
Browse files
audioldm/clap/encoders.py
CHANGED
@@ -14,6 +14,7 @@ class CLAPAudioEmbeddingClassifierFreev2(nn.Module):
|
|
14 |
key="class",
|
15 |
sampling_rate=16000,
|
16 |
embed_mode="audio",
|
|
|
17 |
unconditional_prob=0.1,
|
18 |
random_mute=False,
|
19 |
max_random_mute_portion=0.5,
|
@@ -24,7 +25,7 @@ class CLAPAudioEmbeddingClassifierFreev2(nn.Module):
|
|
24 |
self.key = key
|
25 |
self.device = "cpu"
|
26 |
self.precision = "fp32"
|
27 |
-
self.amodel =
|
28 |
self.tmodel = "roberta" # the best text encoder in our training
|
29 |
self.enable_fusion = False # False if you do not want to use the fusion model
|
30 |
self.fusion_type = "aff_2d"
|
|
|
14 |
key="class",
|
15 |
sampling_rate=16000,
|
16 |
embed_mode="audio",
|
17 |
+
amodel = "HTSAT-tiny",
|
18 |
unconditional_prob=0.1,
|
19 |
random_mute=False,
|
20 |
max_random_mute_portion=0.5,
|
|
|
25 |
self.key = key
|
26 |
self.device = "cpu"
|
27 |
self.precision = "fp32"
|
28 |
+
self.amodel = amodel
|
29 |
self.tmodel = "roberta" # the best text encoder in our training
|
30 |
self.enable_fusion = False # False if you do not want to use the fusion model
|
31 |
self.fusion_type = "aff_2d"
|