Spaces:
Sleeping
Sleeping
modify fx norm
Browse files- inference.py +3 -1
inference.py
CHANGED
@@ -143,7 +143,9 @@ class MasteringStyleTransfer:
|
|
143 |
|
144 |
# Apply fx normalization for input audio during mastering style transfer
|
145 |
if normalize:
|
|
|
146 |
data = self.fx_normalizer.normalize_audio(data, 'mixture')
|
|
|
147 |
|
148 |
# Convert to torch tensor
|
149 |
data_tensor = torch.FloatTensor(data).unsqueeze(0)
|
@@ -154,8 +156,8 @@ class MasteringStyleTransfer:
|
|
154 |
print(f"input: {input_audio}")
|
155 |
print(f"reference: {reference_audio}")
|
156 |
input_tensor = self.preprocess_audio(input_audio, self.args.sample_rate, normalize=True)
|
157 |
-
reference_tensor = self.preprocess_audio(reference_audio, self.args.sample_rate)
|
158 |
print(f"input_tensor: {input_tensor.shape}")
|
|
|
159 |
print(f"reference_tensor: {reference_tensor.shape}")
|
160 |
|
161 |
reference_feature = self.get_reference_embedding(reference_tensor)
|
|
|
143 |
|
144 |
# Apply fx normalization for input audio during mastering style transfer
|
145 |
if normalize:
|
146 |
+
print(f"before normalization: {data.shape}")
|
147 |
data = self.fx_normalizer.normalize_audio(data, 'mixture')
|
148 |
+
print(f"after normalization: {data.shape}")
|
149 |
|
150 |
# Convert to torch tensor
|
151 |
data_tensor = torch.FloatTensor(data).unsqueeze(0)
|
|
|
156 |
print(f"input: {input_audio}")
|
157 |
print(f"reference: {reference_audio}")
|
158 |
input_tensor = self.preprocess_audio(input_audio, self.args.sample_rate, normalize=True)
|
|
|
159 |
print(f"input_tensor: {input_tensor.shape}")
|
160 |
+
reference_tensor = self.preprocess_audio(reference_audio, self.args.sample_rate)
|
161 |
print(f"reference_tensor: {reference_tensor.shape}")
|
162 |
|
163 |
reference_feature = self.get_reference_embedding(reference_tensor)
|