Spaces:
Running
Running
import os | |
import sys | |
import tqdm | |
import torch | |
import torch.nn.functional as F | |
import soundfile as sf | |
import numpy as np | |
now_dir = os.getcwd() | |
sys.path.append(now_dir) | |
from rvc.lib.utils import load_embedding | |
device = sys.argv[1] | |
n_parts = int(sys.argv[2]) | |
i_part = int(sys.argv[3]) | |
i_gpu = sys.argv[4] | |
exp_dir = sys.argv[5] | |
os.environ["CUDA_VISIBLE_DEVICES"] = str(i_gpu) | |
version = sys.argv[6] | |
is_half = bool(sys.argv[7]) | |
embedder_model = sys.argv[8] | |
try: | |
embedder_model_custom = sys.argv[9] | |
except: | |
embedder_model_custom = None | |
wav_path = f"{exp_dir}/1_16k_wavs" | |
out_path = f"{exp_dir}/3_feature256" if version == "v1" else f"{exp_dir}/3_feature768" | |
os.makedirs(out_path, exist_ok=True) | |
def read_wave(wav_path, normalize=False): | |
wav, sr = sf.read(wav_path) | |
assert sr == 16000 | |
feats = torch.from_numpy(wav) | |
feats = feats.half() if is_half else feats.float() | |
feats = feats.mean(-1) if feats.dim() == 2 else feats | |
feats = feats.view(1, -1) | |
if normalize: | |
with torch.no_grad(): | |
feats = F.layer_norm(feats, feats.shape) | |
return feats | |
print("Starting feature extraction...") | |
models, saved_cfg, task = load_embedding(embedder_model, embedder_model_custom) | |
model = models[0] | |
model = model.to(device) | |
if device not in ["mps", "cpu"]: | |
model = model.half() | |
model.eval() | |
todo = sorted(os.listdir(wav_path))[i_part::n_parts] | |
n = max(1, len(todo) // 10) | |
if len(todo) == 0: | |
print( | |
"An error occurred in the feature extraction, make sure you have provided the audios correctly." | |
) | |
else: | |
print(f"{len(todo)}") | |
with tqdm.tqdm(total=len(todo)) as pbar: | |
for idx, file in enumerate(todo): | |
try: | |
if file.endswith(".wav"): | |
wav_file_path = os.path.join(wav_path, file) | |
out_file_path = os.path.join(out_path, file.replace("wav", "npy")) | |
if os.path.exists(out_file_path): | |
continue | |
feats = read_wave(wav_file_path, normalize=saved_cfg.task.normalize) | |
padding_mask = torch.BoolTensor(feats.shape).fill_(False) | |
inputs = { | |
"source": feats.to(device), | |
"padding_mask": padding_mask.to(device), | |
"output_layer": 9 if version == "v1" else 12, | |
} | |
with torch.no_grad(): | |
logits = model.extract_features(**inputs) | |
feats = ( | |
model.final_proj(logits[0]) | |
if version == "v1" | |
else logits[0] | |
) | |
feats = feats.squeeze(0).float().cpu().numpy() | |
if np.isnan(feats).sum() == 0: | |
np.save(out_file_path, feats, allow_pickle=False) | |
else: | |
print(f"{file} - contains nan") | |
pbar.set_description(f"Processing {file} {feats.shape}") | |
except Exception as error: | |
print(error) | |
pbar.update(1) | |
print("Feature extraction completed successfully!") | |