repo
stringlengths 2
99
| file
stringlengths 13
225
| code
stringlengths 0
18.3M
| file_length
int64 0
18.3M
| avg_line_length
float64 0
1.36M
| max_line_length
int64 0
4.26M
| extension_type
stringclasses 1
value |
---|---|---|---|---|---|---|
Vecchia_GPR_var_select | Vecchia_GPR_var_select-master/code/func/KISS_GP.py | import math
import torch
import gpytorch
import numpy
class GPRegressionModel(gpytorch.models.ExactGP):
def __init__(self, train_x, train_y, likelihood):
super(GPRegressionModel, self).__init__(train_x, train_y, likelihood)
# SKI requires a grid size hyperparameter. This util can help with that
# We're setting Kronecker structure to False because we're using an additive structure decomposition
grid_size = int(gpytorch.utils.grid.choose_grid_size(train_x, kronecker_structure=False))
self.mean_module = gpytorch.means.ConstantMean()
self.covar_module = gpytorch.kernels.AdditiveStructureKernel(
gpytorch.kernels.ScaleKernel(
gpytorch.kernels.GridInterpolationKernel(
gpytorch.kernels.MaternKernel(nu=2.5),
grid_size=grid_size, num_dims=1
)
), num_dims=train_x.shape[-1]
)
def forward(self, x):
mean_x = self.mean_module(x)
covar_x = self.covar_module(x)
return gpytorch.distributions.MultivariateNormal(mean_x, covar_x)
def KISS_GP_wrap(XTrn, yTrn, XTst, yTst):
XTrnTorch = torch.from_numpy(XTrn).to(torch.float32)
yTrnTorch = torch.from_numpy(numpy.array(yTrn)).to(torch.float32)
XTstTorch = torch.from_numpy(XTst).to(torch.float32)
yTstTorch = torch.from_numpy(numpy.array(yTst)).to(torch.float32)
# Train
likelihood = gpytorch.likelihoods.GaussianLikelihood()
model = GPRegressionModel(XTrnTorch, yTrnTorch, likelihood)
model.train()
likelihood.train()
optimizer = torch.optim.Adam(model.parameters(), lr=0.1)
mll = gpytorch.mlls.ExactMarginalLogLikelihood(likelihood, model)
training_iterations = 100
for i in range(training_iterations):
optimizer.zero_grad()
output = model(XTrnTorch)
loss = -mll(output, yTrnTorch)
loss.backward()
optimizer.step()
# Predict
model.eval()
likelihood.eval()
with torch.no_grad(), gpytorch.settings.fast_pred_var():
yPred = likelihood(model(XTstTorch)).mean.view([XTstTorch.shape[0]])
rmseScr = torch.sqrt(torch.square(yTstTorch - yPred).mean() / torch.var(yTstTorch))
return dict({"yTstPred": yPred.tolist(), "rmseScr": rmseScr.item()})
## A test here
# n = 1000
# train_x = torch.rand([n, 4])
# train_y = torch.sin(train_x[:, 0] * (4 * math.pi) + torch.randn(n) * 0.2) + \
# torch.sin(train_x[:, 1] * (2 * math.pi) + torch.randn(n) * 0.2) + \
# torch.sin(train_x[:, 2] * (3 * math.pi) + torch.randn(n) * 0.2)
# nTst = 1000
# xTst = torch.rand([nTst, 4])
# yTst = torch.sin(xTst[:, 0] * (4 * math.pi) + torch.randn(n) * 0.2) + \
# torch.sin(xTst[:, 1] * (2 * math.pi) + torch.randn(n) * 0.2) + \
# torch.sin(xTst[:, 2] * (3 * math.pi) + torch.randn(n) * 0.2)
# KISS_GP_wrap(train_x.numpy(), train_y.numpy(), xTst.numpy(), yTst.numpy())
| 2,912 | 40.614286 | 108 | py |
Vecchia_GPR_var_select | Vecchia_GPR_var_select-master/code/func/faiss_NN.py | import numpy as np
import faiss
def get_NN_py(x, nNbor, y=None):
n, d = x.shape
x = np.ascontiguousarray(x.astype(np.float32))
if y is not None:
y = np.ascontiguousarray(y.astype(np.float32))
quantizer = faiss.IndexFlatL2(d)
index = faiss.IndexIVFFlat(quantizer, d, 1024)
index.train(x)
index.add(x)
index.nprobe = 256
if y is not None:
_, IAprx = index.search(y, int(nNbor))
else:
_, IAprx = index.search(x, int(nNbor))
return IAprx
| 503 | 25.526316 | 54 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/startup_config.py | #!/usr/bin/env python
"""
startup_config
Startup configuration utilities
"""
from __future__ import absolute_import
import os
import sys
import torch
import importlib
import random
import numpy as np
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def set_random_seed(random_seed, args=None):
""" set_random_seed(random_seed, args=None)
Set the random_seed for numpy, python, and cudnn
input
-----
random_seed: integer random seed
args: argue parser
"""
# initialization
torch.manual_seed(random_seed)
random.seed(random_seed)
np.random.seed(random_seed)
os.environ['PYTHONHASHSEED'] = str(random_seed)
#For torch.backends.cudnn.deterministic
#Note: this default configuration may result in RuntimeError
#see https://pytorch.org/docs/stable/notes/randomness.html
if args is None:
cudnn_deterministic = True
cudnn_benchmark = False
else:
cudnn_deterministic = args.cudnn_deterministic_toggle
cudnn_benchmark = args.cudnn_benchmark_toggle
if not cudnn_deterministic:
print("cudnn_deterministic set to False")
if cudnn_benchmark:
print("cudnn_benchmark set to True")
if torch.cuda.is_available():
torch.cuda.manual_seed_all(random_seed)
torch.backends.cudnn.deterministic = cudnn_deterministic
torch.backends.cudnn.benchmark = cudnn_benchmark
return
| 1,549 | 25.271186 | 66 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/__init__.py | 0 | 0 | 0 | py |
|
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/math_tools/sig_test.py | #!/usr/bin/env python
"""
sig_test.py
Tools for significance test.
This is used in tutorials/b2_anti_spoofing/chapter_a1_stats_test.ipynb
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
from scipy import stats
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
##########
# Functions
##########
def compute_Z_alpha2(significance_level, alternative='two-sided'):
""" z_alpha = compute_Z_alpha2(significance_level, alternative)
input
-----
significance_level: float
alternative: string, less, greatr or two-sided
output
------
Z_alpha2: float
Example
-------
print("Z_alpha/2 for 90% CI (alpha=0.10): {:.4f}".format(
compute_Z_alpha2(0.10)))
print("Z_alpha/2 for 95% CI (alpha=0.05): {:.4f}".format(
compute_Z_alpha2(0.05)))
print("Z_alpha/2 for 99% CI (alpha=0.01): {:.4f}".format(
compute_Z_alpha2(0.01)))
"""
if alternative == 'less':
Z_alpha2 = stats.norm.ppf(significance_level)
elif alternative == 'greater':
Z_alpha2 = stats.norm.isf(significance_level)
elif alternative == 'two-sided':
Z_alpha2 = stats.norm.isf(significance_level/2)
else:
raise ValueError("alternative must be "
"'less', 'greater' or 'two-sided'")
return Z_alpha2
##########
# Tools for family-wise hypothesis testing and correction
##########
def reject_null_bonferroni_naive(
z_values, num_test, significance_level, alternative='two-sided',
accept_value=True, reject_value=False):
"""result = reject_null_bonferroni_naive(z_values, significance_level)
native bonferroni correction
https://en.wikipedia.org/wiki/Bonferroni_correction
input
-----
z_values: np.array, an array of z_value
num_test: int, number of tests in the experiment
signifiance_level: float, common choise is 0.1, 0.05, or 0.01
output
------
result: np.array, same size as z_values,
if result[i] is True if z_value[i] is larger than the threshold
"""
#num_test = z_values.size
corrected_conf_level = significance_level / num_test
Z_alpha2 = compute_Z_alpha2(corrected_conf_level, alternative)
# reject null hypothesis
idx_reject = z_values > Z_alpha2
# acccept null hypothesis
idx_accept = z_values <= Z_alpha2
result = np.zeros(z_values.shape)
result[idx_accept] = accept_value
result[idx_reject] = reject_value
return result
def reject_null_sidak(
z_values, num_test, significance_level, alternative='two-sided',
accept_value=True, reject_value=False):
""" similar API to reject_null_bonferroni_naive
See Hervé Abdi, and others. Bonferroni and Šidák
Corrections for Multiple Comparisons.
Encyclopedia of Measurement and Statistics 3.
Sage Thousand Oaks, CA: 103–107. 2007.
"""
#num_test = z_values.size
#FWER = 1 - (1 - significance_level) ** (1 / num_test)
corrected_conf_level = 1 - (1 - significance_level) ** (1 / num_test)
Z_alpha2 = compute_Z_alpha2(corrected_conf_level, alternative)
idx_reject = z_values >= Z_alpha2
idx_accept = z_values < Z_alpha2
result = np.zeros(z_values.shape)
result[idx_accept] = accept_value
result[idx_reject] = reject_value
return result
def reject_null_holm_bonferroni(
z_values, num_test, significance_level, alternative='two-sided',
accept_value=True, reject_value=False):
""" similar API to reject_null_bonferroni_naive
See https://en.wikipedia.org/wiki/Holm%E2%80%93Bonferroni_method
"""
# get the index of each z_value in the sorted array
# large z corresponds to small p
# we start with the case where the z is the largest
order = np.argsort(z_values.flatten())[::-1]
# a result buffer, by default, set all to accept
result_buff = np.zeros(order.shape) + accept_value
# conduct the test
for test_idx, data_idx in enumerate(order):
# the test coming first should receive a more strict Z_alpha
corrected_conf_level = significance_level / (num_test - test_idx)
Z_alpha2 = compute_Z_alpha2(corrected_conf_level, alternative)
#print(corrected_conf_level)
if z_values.flatten()[data_idx] < Z_alpha2:
# if we cannot reject, stop testing
result_buff[data_idx] = accept_value
break
else:
result_buff[data_idx] = reject_value
return np.reshape(result_buff, z_values.shape)
def reject_null_holm_bonferroni_given_p_value(
p_values, num_test, significance_level,
accept_value=True, reject_value=False):
""" result = reject_null_holm_bonferroni_given_p_value(
p_values, num_test, significance_level,
accept_value = True, reject_value = False)
https://en.wikipedia.org/wiki/Holm%E2%80%93Bonferroni_method
input
-----
p_values: np.array, any shape, of p-values for multiple test
num_test: int, number of elements in the p_values matrix
significance_level: float
accept_value: int, float, or bool, anything. Default True
It is used to indicate
"accept the null hypothesis"
reject_value: int, float, or bool, anything. Default False
It is used to indicate
"reject the null hypothesis"
output
------
result: np.array, same shape as p_values
If result[i, ...] accepts the null hypothesis,
its value will be set to accept_value
Otherwise, the value will be equal to reject_value
See Example in
>> p_values = np.array([[0.01, 0.04], [0.03, 0.005]])
>> print(reject_null_holm_bonferroni_given_p_value(p_values, 4, 0.05))
Compare p-value 0.005 with corrected alpha 0.0125 reject NULL
Compare p-value 0.010 with corrected alpha 0.0167 reject NULL
Compare p-value 0.030 with corrected alpha 0.0250 accept NULL, stop
[[0. 1.]
[1. 0.]]
"""
# get the index of each p_value in the sorted array
# small p comes first
order = np.argsort(p_values.flatten())
# a result buffer, by default, set all to accept
result_buff = np.zeros(order.shape) + accept_value
# conduct the test
for test_idx, data_idx in enumerate(order):
# the test coming first should receive a more strict Z_alpha
corrected_conf_level = significance_level / (num_test - test_idx)
print("Compare p-value {:.3f} with corrected alpha {:.4f}".format(
p_values.flatten()[data_idx], corrected_conf_level), end=' ')
# here smaller p_value is more significant
if p_values.flatten()[data_idx] > corrected_conf_level:
# if we cannot reject, stop testing
result_buff[data_idx] = accept_value
print("accept NULL, stop")
break
else:
result_buff[data_idx] = reject_value
print("reject NULL")
return np.reshape(result_buff, p_values.shape)
if __name__ == "__main__":
print("tools for significance test")
| 7,412 | 33.47907 | 77 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/math_tools/stats.py | #!/usr/bin/env python
"""
stats.py
Tools to calcualte statistics
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import core_scripts.other_tools.display as nii_display
import core_scripts.data_io.conf as nii_dconf
import torch
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def f_var2std(var):
"""
std = f_var2std(var)
Args:
var: np.arrary, variance
Return:
std: np.array, standard-devitation
std = sqrt(variance), std[std<floor] = 1.0
"""
negative_idx = var < 0
std = np.sqrt(var)
std[negative_idx] = 1.0
floored_idx = std < nii_dconf.std_floor
std[floored_idx] = 1.0
return std
def f_online_mean_std(data, mean_old, var_old, cnt_old):
"""
mean, var, count=f_online_mean_var(data, mean, var, num_count):
online algorithm to accumulate mean and var
input
-----
data: input data as numpy.array, in shape [length, dimension]
mean: mean to be updated, np.array [dimension]
var: var to be updated, np.array [dimension]
num_count: how many data rows have been calculated before
this calling.
output
------
mean: mean, np.array [dimension]
var: var, np.array [dimension]
count: accumulated data number, = num_count + data.shape[0]
Ref. parallel algorithm
https://en.wikipedia.org/wiki/Algorithms_for_calculating_variance
"""
try:
# how many time steps (number of rows) in this data
cnt_this = data.shape[0]
# if input data is empty, don't update
if cnt_this == 0:
return mean_old, var_old, cnt_old
if data.ndim == 1:
# single dimension data, 1d array
mean_this = data.mean()
var_this = data.var()
dim = 1
else:
# multiple dimension data, 2d array
mean_this = data.mean(axis=0)
var_this = data.var(axis=0)
dim = data.shape[1]
# difference of accumulated mean and data mean
diff_mean = mean_this - mean_old
# new mean and var
new_mean = np.zeros([dim], dtype=nii_dconf.h_dtype)
new_var = np.zeros([dim], dtype=nii_dconf.h_dtype)
# update count
updated_count = cnt_old + cnt_this
# update mean
new_mean = mean_old + diff_mean * (float(cnt_this) /
(cnt_old + cnt_this))
# update var
if cnt_old == 0:
# if this is the first data
if data.ndim == 1:
# remember that var is array, not scalar
new_var[0] = var_this
else:
new_var = var_this
else:
# not first data
new_var = (var_old * (float(cnt_old) / updated_count)
+ var_this * (float(cnt_this)/ updated_count)
+ (diff_mean * diff_mean
/ (float(cnt_this)/cnt_old
+ float(cnt_old)/cnt_this
+ 2.0)))
# done
return new_mean, new_var, updated_count
except ValueError:
if data.ndim > 1:
if data.shape[1] != mean_old.shape[0] or \
data.shape[1] != var_old.shape[0]:
nii_display.f_print("Dimension incompatible", "error")
nii_display.f_die("Error in online mean var calculation")
else:
if mean_old.shape[0] != 1 or \
var_old.shape[0] != 1:
nii_display.f_print("Dimension incompatible", "error")
nii_display.f_die("Error in online mean var calculation")
def f_online_mean_cov(data, mean_old, cov_old, cnt_old):
"""
mean, cov, count=f_online_mean_cov(data, mean, cov, num_count):
online algorithm to accumulate mean and cov
input
-----
data: input data as numpy.array, in shape [length, dimension]
mean: mean to be updated, np.array [dimension]
cov: cov to be updated, np.array [dimension, dimension]
num_count: how many data rows have been calculated before
this calling.
output
------
mean: mean, np.array [dimension]
cov: cov, np.array [dimension, dimension]
count: accumulated data number, = num_count + data.shape[0]
Note that the returned cov is biased.
Ref. parallel algorithm
https://en.wikipedia.org/wiki/Algorithms_for_calculating_variance
"""
if data.ndim == 1:
nii_display.f_print("Dimension incompatible", "error")
nii_display.f_die("Error in online mean cov calculation")
try:
# how many time steps (number of rows) in this data
cnt_this = data.shape[0]
dim = data.shape[1]
# if input data is empty, don't update
if cnt_this == 0:
return mean_old, cov_old, cnt_old
# multiple dimension data, 2d array
mean_this = data.mean(axis=0)
# assumpe number of columns be the number of variables
if cnt_this == 1:
cov_this = np.zeros([dim, dim], dtype=nii_dconf.h_dtype)
else:
cov_this = np.cov(data.T)
# difference of accumulated mean and data mean
diff_mean = mean_this - mean_old
# new mean and cov
new_mean = np.zeros([dim], dtype=nii_dconf.h_dtype)
new_cov = np.zeros([dim, dim], dtype=nii_dconf.h_dtype)
# update count
updated_count = cnt_old + cnt_this
# update mean
new_mean = mean_old + diff_mean * (float(cnt_this) /
(cnt_old + cnt_this))
# update cov
if cnt_old == 0:
new_cov = cov_this
else:
# not first data
new_cov = (cov_old * (float(cnt_old) / updated_count)
+ cov_this * (float(cnt_this)/ updated_count)
+ (np.outer(diff_mean, diff_mean)
/ (float(cnt_this)/cnt_old
+ float(cnt_old)/cnt_this
+ 2.0)))
# done
return new_mean, new_cov, updated_count
except ValueError:
if data.ndim > 1:
if data.shape[1] != mean_old.shape[0] or \
data.shape[1] != cov_old.shape[0]:
nii_display.f_print("Dimension incompatible", "error")
nii_display.f_die("Error in online mean cov calculation")
else:
if mean_old.shape[0] != 1 or \
cov_old.shape[0] != 1:
nii_display.f_print("Dimension incompatible", "error")
nii_display.f_die("Error in online mean cov calculation")
def cov_tensor(X):
"""cov = cov_tensor(X)
compute covariance of tensor
input
-----
X: tensor, (batch, length, dim)
output
------
cov: tensor, (batch, dim, dim)
Note that this function returns biased cov
"""
X = data
D = X.shape[1]
mean = torch.mean(X, dim=1).unsqueeze(1)
X = X - mean
return 1/D * torch.matmul(X.transpose(-1, -2), X)
def f_online_mean_cov_tensor(data, mean_old, cov_old, cnt_old):
"""
mean, cov, count=f_online_mean_cov(data, mean, cov, num_count):
online algorithm to accumulate mean and cov
input
-----
data: tensor, in shape [batch, length, dimension]
mean: mean to be updated, tensor [batch, dimension]
cov: cov to be updated, tensor [batch, dimension, dimension]
num_count: how many data rows have been calculated before
this calling.
output
------
mean: mean, tensor [batch, dimension]
cov: cov, tensor [batch, dimension, dimension]
count: accumulated data number, = num_count + data.shape[0]
Note that the returned cov is biased.
Example:
data = torch.randn([2, 10, 5])
mean_start = torch.zeros([2, 5])
cov_start = torch.zeros([2, 5, 5])
cnt = 0
mean_old = mean_start
cov_old = cov_start
for idx in range(10):
new_mean, new_var, cnt = f_online_mean_cov_tensor(data[:, idx:idx+1, :],
mean_old, cov_old, cnt)
mean_old = new_mean
cov_old = new_var
Ref. parallel algorithm
https://en.wikipedia.org/wiki/Algorithms_for_calculating_variance
"""
# how many time steps (number of rows) in this data
batch, cnt_this, dim = data.shape
# if input data is empty, don't update
if cnt_this == 0:
return mean_old, cov_old, cnt_old
# Mean and Cov for this data
# multiple dimension data, 2d array
mean_this = data.mean(dim=1)
# assumpe number of columns be the number of variables
if cnt_this == 1:
cov_this = torch.zeros(
[dim, dim], dtype=data.dtype, device=data.device)
else:
cov_this = cov(data.T)
# Update
# difference of accumulated mean and data mean
diff_mean = mean_this - mean_old
# new mean and cov
new_mean = torch.zeros([dim], dtype=data.dtype, device=data.device)
new_cov = torch.zeros([dim, dim], dtype=data.dtype, device=data.device)
# update count
updated_count = cnt_old + cnt_this
# update mean
new_mean = mean_old + diff_mean * (float(cnt_this) /
(cnt_old + cnt_this))
# update cov
if cnt_old == 0:
new_cov = cov_this
else:
# not first data
new_cov = (cov_old * (float(cnt_old) / updated_count)
+ cov_this * (float(cnt_this)/ updated_count)
+ (torch.bmm(diff_mean.unsqueeze(-1), diff_mean.unsqueeze(1))
/ (float(cnt_this)/cnt_old
+ float(cnt_old)/cnt_this
+ 2.0)))
# done
return new_mean, new_cov, updated_count
if __name__ == "__main__":
pass
| 10,262 | 30.194529 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/math_tools/random_tools.py | #!/usr/bin/env python
"""
random.py
Tools related to randomness
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import random
import core_scripts.other_tools.display as nii_display
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
##############
# Shuffling tools
##############
def f_shuffle_slice_inplace(input_list, slice_start=None, slice_stop=None):
""" shuffle_slice(input_list, slice_start, slice_stop)
Shuffling input list (in place) in the range specified by slice_start
and slice_stop.
Based on Knuth shuffling
https://en.wikipedia.org/wiki/Fisher%E2%80%93Yates_shuffle
input
-----
input_list: list
slice_start: int, start idx of the range to be shuffled
slice_end: int, end idx of the range to be shuffled
Both slice_start and slice_end should be in the style of python index
e.g., shuffle_slice(input_list, 0, N) will shuffle the slice input[0:N]
When slice_start / slice_stop is None,
slice_start = 0 / slice_stop = len(input_list)
output
------
none: shuffling is done in place
"""
if slice_start is None or slice_start < 0:
slice_start = 0
if slice_stop is None or slice_stop > len(input_list):
slice_stop = len(input_list)
idx = slice_start
while (idx < slice_stop - 1):
idx_swap = random.randrange(idx, slice_stop)
# naive swap
tmp = input_list[idx_swap]
input_list[idx_swap] = input_list[idx]
input_list[idx] = tmp
idx += 1
return
def f_shuffle_in_block_inplace(input_list, block_size):
"""
f_shuffle_in_block_inplace(input_list, block_size)
Shuffle the input list (in place) by dividing the list input blocks and
shuffling within each block
Example:
>>> data = [1,2,3,4,5,6]
>>> random_tools.f_shuffle_in_block_inplace(data, 3)
>>> data
[3, 1, 2, 5, 4, 6]
input
-----
input_list: input list
block_size: int
output
------
None: shuffling is done in place
"""
if block_size <= 1:
# no need to shuffle if block size if 1
return
else:
list_length = len(input_list)
# range( -(- x // y) ) -> int(ceil(x / y))
for iter_idx in range( -(-list_length // block_size) ):
# shuffle within each block
f_shuffle_slice_inplace(
input_list, iter_idx * block_size, (iter_idx+1) * block_size)
return
def f_shuffle_blocks_inplace(input_list, block_size):
"""
f_shuffle_blocks_inplace(input_list, block_size)
Shuffle the input list (in place) by dividing the list input blocks and
shuffling blocks
Example:
>> data = np.arange(1, 7)
>> f_shuffle_blocks_inplace(data, 3)
>> print(data)
[4 5 6 1 2 3]
input
-----
input_list: input list
block_size: int
output
------
None: shuffling is done in place
"""
# new list
tmp_list = input_list.copy()
block_number = len(input_list) // block_size
shuffle_block_idx = [x for x in range(block_number)]
random.shuffle(shuffle_block_idx)
new_idx = None
for iter_idx in range(block_size * block_number):
block_idx = iter_idx // block_size
in_block_idx = iter_idx % block_size
new_idx = shuffle_block_idx[block_idx] * block_size + in_block_idx
input_list[iter_idx] = tmp_list[new_idx]
return
if __name__ == "__main__":
print("Definition of randomness tools")
| 3,646 | 25.23741 | 77 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/math_tools/mos_norm.py | #!/usr/bin/env python
"""
mos_norm.py
MOS score normalization tools to analysis listening test results.
This is dumped from scripts
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import random
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
####################
# rank-normalization
#
# [1] Andrew Rosenberg, and Bhuvana Ramabhadran. 2017.
# Bias and Statistical Significance in Evaluating Speech Synthesis with
# Mean Opinion Scores. In Proc. Interspeech, 3976–3980.
####################
def _rank_norm_mapping_function(data, data_range=[1, 10]):
""" data_map = _rank_norm_mapping_function(data, data_range=[1, 10])
input
-----
data: np.array, (N,), MOS scores to compute the mapping function
data_range: [int, int],
the lowest and highest possible values for the data
output
------
data_map: dict, data_map[SCORE] -> rank_normed_score
"""
#
data_map = {}
# step1. sort the scores
sorted_data = np.sort(data, kind='quicksort')
# step2. assigna rank (staring from 1) to each score
sorted_rank = np.arange(len(sorted_data)) + 1
# step3. compute the normalized scores
for score_value in np.arange(data_range[0], data_range[1]+1):
# indices are the ranks of the score
# (the score may appear multiple times, indices are their ranks)
indices = sorted_rank[sorted_data == score_value]
if indices.shape[0]:
# (mean of rank - 1) / N
data_map[score_value] = (np.mean(indices) - 1) / data.shape[0]
else:
# if the score does not appear in the data
data_map[score_value] = -1
return data_map
def rank_norm(data, data_range):
""" ranked_score = rank_norm(data)
input
-----
data: np.array, (N,), MOS scores to compute the mapping function
data_range: [int, int],
the lowest and highest possible values for the data
(even though the value may not appear in the data)
It can be [1, 5] for most of the MOS test
output
------
data_map: dict, data_map[SCORE] -> rank_normed_score
Example
-------
>>> data = np.array([2, 1, 2, 10, 4, 5, 6, 4, 5, 7])
>>> rank_norm(data, [1, 10])
[0.15, 0.0, 0.15, 0.9, 0.35, 0.55, 0.7, 0.35, 0.55, 0.8]
"""
data_map = _rank_norm_mapping_function(data, data_range)
data_new = [data_map[x] for x in data]
return data_new
if __name__ == "__main__":
print("Tools to normalize MOS scores")
| 2,648 | 28.433333 | 76 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/math_tools/__init__.py | 0 | 0 | 0 | py |
|
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/script_model_para.py | #!/usr/bin/env python
"""
A simple wrapper to show parameter of the model
Usage:
# go to the model directory, then
$: python script_model_para.py
We assume model.py and config.py are available in the project directory.
"""
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
import importlib
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def f_model_show(pt_model):
"""
f_model_show(pt_model)
Args: pt_model, a Pytorch model
Print the informaiton of the model
"""
#f_model_check(pt_model)
print(pt_model)
num = sum(p.numel() for p in pt_model.parameters() if p.requires_grad)
print("Parameter number: {:d}".format(num))
for name, p in pt_model.named_parameters():
if p.requires_grad:
print("Layer: {:s}\tPara. num: {:<10d} ({:02.1f}%)\tShape: {:s}"\
.format(name, p.numel(), p.numel()*100.0/num, str(p.shape)))
return
if __name__ == "__main__":
sys.path.insert(0, os.getcwd())
if len(sys.argv) == 3:
prj_model = importlib.import_module(sys.argv[1])
prj_conf = importlib.import_module(sys.argv[2])
else:
print("By default, load model.py and config.py")
prj_model = importlib.import_module("model")
prj_conf = importlib.import_module("config")
input_dims = sum(prj_conf.input_dims)
output_dims = sum(prj_conf.output_dims)
model = prj_model.Model(input_dims, output_dims, None)
f_model_show(model)
| 1,902 | 29.206349 | 78 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/list_tools.py | #!/usr/bin/env python
"""
list_tools.py
Tools to process list(s)
"""
from __future__ import absolute_import
import os
import sys
import collections
import core_scripts.other_tools.display as nii_warn
import core_scripts.other_tools.str_tools as nii_str_tool
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
from pathlib import Path as pathtool
def listdir_with_ext_flat(file_dir, file_ext=None):
"""
file_list = lstdir_with_ext_flat(file_dir, file_ext=None)
Return a list of file names with specified extention
Args:
file_dir: a file directory
file_ext: string, specify the extention, e.g., txt, bin
Return:
file_list: a list of file_names
"""
try:
if file_ext is None:
file_list = [os.path.splitext(x)[0] for x in os.listdir(file_dir) \
if not x.startswith('.')]
else:
file_list = [os.path.splitext(x)[0] for x in os.listdir(file_dir) \
if not x.startswith('.') and x.endswith(file_ext)]
return file_list
except OSError:
nii_warn.f_print("Cannot access %s" % (file_dir), "error")
return []
def listdir_with_ext_recur(file_dir, file_ext=None, recursive=True):
"""
file_list = lstdir_with_ext(file_dir, file_ext=None)
Return a list of file names with specified extention
Args:
file_dir: a file directory
file_ext: string, specify the extention, e.g., txt, bin
Return:
file_list: a list of file_names
"""
file_dir_tmp = file_dir if file_dir.endswith('/') else file_dir + '/'
file_list = []
for rootdir, dirs, files in os.walk(file_dir_tmp, followlinks=True):
tmp_path = rootdir.replace(file_dir_tmp, '')
# concatenate lists may be slow for large data set
# change it in the future
if file_ext:
file_list += [os.path.splitext(os.path.join(tmp_path, x))[0] \
for x in files if x.endswith(file_ext)]
else:
file_list += [os.path.splitext(os.path.join(tmp_path, x))[0] \
for x in files]
return file_list
def listdir_with_ext(file_dir, file_ext=None, recursive=False):
"""
file_list = lstdir_with_ext(file_dir, file_ext=None, recursive=False)
Return a list of file names with specified extention
Args:
file_dir: a file directory
file_ext: string, specify the extention, e.g., txt, bin
recursive: bool, whether search recursively (default False)
Return:
file_list: a list of file_names
"""
if not recursive:
return listdir_with_ext_flat(file_dir, file_ext)
else:
return listdir_with_ext_recur(file_dir, file_ext)
def common_members(list_a, list_b):
""" list_c = common_members(list_a, list_b)
Return a list (sorted) of common members in list_a, list_b
Parameters:
list_a: list
list_b: list
Returns:
list_c: a list of common members in list_a and list_b
"""
list_c = list(set(list_a).intersection(list_b))
list_c.sort()
return list_c
def list_identical(list_a, list_b):
""" flag = list_identical(list_a, list_b)
Return true/false, check whether list_a is identical to list_b
stackoverflow.com/a/19244156/403423
"""
return collections.Counter(list_a) == collections.Counter(list_b)
def list_b_in_list_a(list_a, list_b):
""" list_b_in_list_a(list_a, list_b)
Whether list_b is subset of list_a
Parameters:
list_a: list
list_b: list
Return:
flag: bool
"""
return set(list_b) <= set(list_a)
def members_in_a_not_in_b(list_a, list_b):
""" members_in_a_not_b(list_a, list_b):
Return a list of members that are in list_a but not in list_b
Args:
list_a: list
list_b: list
Return:
list
"""
return list(set(list_a) - set(list_b))
def read_list_from_text(filename, f_chop=True):
"""out_list = read_list_from_text(filename, f_chop=True)
Read a text file and return a list, where each text line is one element
Args:
filename: str, path to the file
f_chop: bool, whether trim the newline symbol at the end of each line
(default True)
Return:
output_list: list, each element is one line in the input text file
"""
data = []
with open(filename,'r') as file_ptr:
for line in file_ptr:
line = nii_str_tool.string_chop(line) if f_chop else line
data.append(line)
return data
def write_list_to_text_file(data_list, filepath, endl='\n'):
"""write_list_to_text(data_list, filepath, endl='\n')
Save a list of data to a text file
Args:
data_list: list, data list to be saved
filepath: str, path to the output text file
endl: str, ending of each new line, default \n
If each element in data_list is not str, it will be converted to
str by str().
"""
with open(filepath, 'w') as file_ptr:
for data_entry in data_list:
if type(data_entry) is str:
file_ptr.write(data_entry + endl)
else:
file_ptr.write(str(data_entry) + endl)
return
if __name__ == "__main__":
#print("Definition of tools for list operation")
input_list1 = read_list_from_text(sys.argv[1])
input_list2 = read_list_from_text(sys.argv[2])
residual_list = members_in_a_not_in_b(input_list1, input_list2)
for filename in residual_list:
print(filename)
| 6,026 | 31.578378 | 79 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/str_tools.py | #!/usr/bin/env python
"""
str_tools
tools to process string
"""
from __future__ import absolute_import
import os
import sys
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def f_realpath(f_dir, f_name, f_ext):
""" file_path = f_realpath(f_dir, f_name, f_ext)
Args:
f_dir: string, directory
f_name: string, file name
f_ext: string, file name extension
Return:
file_path: realpath
"""
file_path = os.path.join(f_dir, f_name)
if f_ext.startswith(os.extsep):
file_path = file_path + f_ext
else:
file_path = file_path + os.extsep + f_ext
return file_path
def string_chop(InStr):
""" output = string_chop(InStr)
Chop the ending '\r' and '\n' from input string
Args:
InStr: str, the input string
Return:
output: str
'\r' corresponds to '0x0d' or 13,
'\n' corresponds to '0x0a' or 10
"""
if len(InStr) >= 2 and ord(InStr[-1]) == 10 and ord(InStr[-2]) == 13:
return InStr[:-2]
elif len(InStr) >= 1 and ord(InStr[-1]) == 10:
return InStr[:-1]
else:
return InStr
if __name__ == "__main__":
print("string tools")
| 1,279 | 21.857143 | 73 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/data_warehouse.py | #!/usr/bin/env python
"""
data_warehouse
Simple tools to manage data from text file
"""
from __future__ import absolute_import
import os
import sys
import itertools
import numpy as np
from core_scripts.other_tools import list_tools
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
class DataEntry:
"""DataEntry to store data for one entry
"""
def __init__(self, data, tags, comment=""):
"""DataEntry(data, tags, comment)
args:
data: any kind of python object
tags: list of str, tags of the data entry
comment: coment
"""
self.data_value = data
self.tags = self._parse_tag(tags)
self.comment = comment
def _parse_tag(self, tags):
"""[tag_1, tag_2, tag_3] -> {1: tag1, 2: tag2, 3: tag3}
"""
temp = {x:y for x, y in enumerate(tags)}
return temp
def get_value(self):
return self.data_value
def get_tag(self, tag_idx):
return self.tags[tag_idx]
def check_tags(self, tag_indices, tag_values):
"""check_tags(tag_indices, tag_values)
check whether the specified tag is equal to the tag value
input:
tag_indices: list, self.tags[tag_index] should be accessible
tag_values: list, self.tags[tag_index] == tag_value?
output:
True: if tag_values are matched with tags if this data
"""
for tag_idx, tag_value in zip(tag_indices, tag_values):
if self.tags[tag_idx] != tag_value:
return False
return True
class DataWarehouse:
"""DataWarehouse to manage data with multi-view
"""
def __init__(self, orig_file_path, parse_value_methods, parse_tag_methods):
"""DataWarehouse(orig_file_path, parse_methods)
input:
orig_file_path: str, path to the original file
parse_methods: list of functions, to parse the data entry
"""
self.file_path = orig_file_path
self.parse_v_methods = parse_value_methods
self.parse_t_methods = parse_tag_methods
self.data_list = []
self.tag_list = {}
self.data_entries = self._parse_file()
def _parse_file(self):
# load list
data_content = list_tools.read_list_from_text(self.file_path)
for data_entry in data_content:
# iterate over parse methods
for parse_v_method, parse_t_method in \
zip(self.parse_v_methods, self.parse_t_methods):
# get value
data_value = parse_v_method(data_entry)
# get tag
tags = [x(data_entry) for x in parse_t_method]
# skip invalid line
if data_value is None or None in tags:
continue
# create data entry
tmp_data_entry = DataEntry(data_value, tags)
self.data_list.append(tmp_data_entry)
# add tag to the self.tag_list
for tag_id, tag_val in enumerate(tags):
self._add_tag(tag_id, tag_val)
return
def _add_tag(self, tag_id, tag_val):
# collect all possible tags for the tag_id-th tag
if tag_id in self.tag_list:
if not tag_val in self.tag_list[tag_id]:
self.tag_list[tag_id].append(tag_val)
else:
self.tag_list[tag_id] = [tag_val]
return
def get_view(self, tag_idxs, tag_values, score_parse = None):
""" get_view(tag_idxs, tag_values, score_parse = None)
input:
tag_idxs: list, the index of the tag slot to check
tag_values: list, the value of the tag slot to compare
score_parse: function, a function to extract score from entry
output:
data_view: list of data
"""
data_view = [x.get_value() for x in self.data_list \
if x.check_tags(tag_idxs, tag_values)]
if score_parse is not None:
return [score_parse(x) for x in data_view]
else:
return data_view
def _to_numpy(self, data_list, dims, statistics):
""" convert data_list to numpy
"""
# maximum length of one data entry
max_length = max([len(x) for x in data_list])
# create data array
if statistics is None:
data_array = np.ones([np.prod(dims), max_length]) * np.inf
for idx, data_entry in enumerate(data_list):
data_array[idx, 0:len(data_entry)] = np.array(data_entry)
return np.reshape(data_array, dims + [max_length])
else:
data_array = np.ones([np.prod(dims)])
for idx, data_entry in enumerate(data_list):
if data_entry:
data_array[idx] = statistics(data_entry)
return np.reshape(data_array, dims)
def get_views_cross(self, tag_idxs, tag_values,
score_parse=None, to_numpy=False, statistics=None):
"""get_views_cross(self, tag_idxs, tag_values,
score_parse=None, to_numpy=False, statistics=None)
input:
tag_idxs: list, list of tag indices to check
tag_values: list of list, for each tag_index,
A list of tags will be created through this cross:
tag_values[0] x tag_values[1] x ...
Then, each combination is used to retrieve the data
output data will be a tensor of
[len(tag_values[0]), len(tag_values[1]), ...]
output:
data_list:
"""
data_list = []
data_mat_size = [len(x) for x in tag_values]
tag_iter = itertools.product(*tag_values)
for tag_ent in tag_iter:
data_list.append(self.get_view(tag_idxs, tag_ent, score_parse))
if to_numpy:
return self._to_numpy(data_list, data_mat_size, statistics)
else:
return data_list
def get_tags(self, tag_idx):
if tag_idx in self.tag_list:
return self.tag_list[tag_idx]
else:
return None
if __name__ == "__main__":
print("tools for data warehouse")
| 6,499 | 32.678756 | 79 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/__init__.py | 0 | 0 | 0 | py |
|
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/random_name_mgn.py | #!/usr/bin/env python
"""
Random name manager
Used for produce protocols
"""
from __future__ import absolute_import
import os
import sys
from core_scripts.other_tools import list_tools
from core_scripts.data_io import io_tools
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
def list_loader(list_file):
""" output_list = list_loader(list_file)
Load a text file as a list of string. This function will use __cache to save
the loaded list
Args:
list_file: str, path to the input text file
Return:
output_list: list,
Note, a __cache will be created along list_file
"""
cache_dir = os.path.join(os.path.dirname(list_file), '__cache')
return io_tools.wrapper_data_load_with_cache(
list_file, list_tools.read_list_from_text, cache_dir)
class RandomNameMgn:
""" Class to manage the list of random file names
"""
def __init__(self, file_random_name_list, verbose=False):
""" RandomNameMgn(file_random_name_list, verbose=False)
Create a random name manager
Args:
file_random_name_list: str, path to the text file of random names.
verbose: bool, default False, print information during initialziation
"""
if verbose:
print("Loading random name tables")
## For unused random name list
# load entries in the list
self.unused_entries = list_loader(file_random_name_list)
# prepare dictionary
self.mapper = {x : None for x in self.unused_entries}
# reverse dictionary
self.mapper_rev = {}
# print some informaiton
if verbose:
self.print_info()
self.verbose = verbose
# done
return
def print_info(self):
mes = "Number of unused random file names: {:d}".format(
len(self.unused_entries))
print(mes)
return
def retrieve_rand_name(self, filename):
""" rand_name = retrieve_rand_name(filename)
filename: str, input, input file name
rand_name: str, output, the random file name
"""
if filename in self.mapper_rev:
return self.mapper_rev[filename]
else:
rand_name = self.unused_entries.pop()
self.mapper[rand_name] = filename
self.mapper_rev[filename]= rand_name
return rand_name
def save_unused_name(self, save_file):
""" save_unused_name(save_file)
save_file: str, input, the path to save random names that
have NOT been used
"""
with open(save_file, 'w') as file_ptr:
for entry in self.unused_entries:
file_ptr.write(entry+'\n')
if self.verbose:
self.print_info()
print("Save unused random names to {:s}".format(save_file))
return
def retrieve_filename(self, random_name):
if random_name in self.mapper:
return self.mapper[random_name]
else:
print("Random name {:s} has not been logged".format(random_name))
sys.exit(1)
if __name__ == "__main__":
print("")
| 3,241 | 29.018519 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/debug.py | #!/usr/bin/env python
"""
debug.py
Tools to help debugging
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import datetime
import numpy as np
import torch
from core_scripts.data_io import io_tools as nii_io
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def convert_data_for_debug(data):
""" data_new = convert_data_for_debug(data)
For debugging, it is convenient to has a data in numpy format
Args
----
data: tensor
Return
------
data_new: numpy array
"""
if hasattr(data, 'detach'):
return data.detach().to('cpu').numpy()
elif hasattr(data, 'cpu'):
return data.to('cpu').numpy()
elif hasattr(data, 'numpy'):
return data.numpy()
else:
return data
def qw(data, path=None):
""" write data tensor into a temporary buffer
Args
----
data: a pytorch tensor or numpy tensor
path: str, path to be write the data
if None, it will be "./debug/temp.bin"
Return
------
None
"""
if path is None:
path = 'debug/temp.bin'
try:
os.mkdir(os.path.dirname(path))
except OSError:
pass
# write to IO
nii_io.f_write_raw_mat(convert_data_for_debug(data), path)
return
def check_para(pt_model):
""" check_para(pt_model)
Quickly check the statistics on the parameters of the model
Args
----
pt_model: a Pytorch model defined based on torch.nn.Module
Return
------
None
"""
mean_buf = [p.mean() for p in pt_model.parameters() if p.requires_grad]
std_buf = [p.std() for p in pt_model.parameters() if p.requires_grad]
print(np.array([convert_data_for_debug(x) for x in mean_buf]))
print(np.array([convert_data_for_debug(x) for x in std_buf]))
return
class data_probe:
""" data_probe is defined to collect intermediate data
produced from the inference or training stage
"""
def __init__(self):
# a list to store all intermediate data
self.data_buf = []
# a list of data name
self.data_names = []
# a single array to store the data
self.data_concated = None
# default data convert method
self.data_convert_method = convert_data_for_debug
# default method to dump method
self.data_dump_method = nii_io.pickle_dump
# dump file name extension
self.dump_file_ext = '.pkl'
return
def add_data(self, input_data, name=None):
""" add_data(input_data)
Add the input data to a data list. Data will be automatically
converted by self.data_convert_method
input
-----
input_data: tensor, or numpy.array
"""
self.data_buf.append(self.data_convert_method(input_data))
if name:
self.data_names.append(name)
return
def _merge_data(self):
""" merge_data()
Merge the data in the list to a big numpy array table.
Follow the convention of this project, we assume data has shape
(batchsize, length, feat_dim)
"""
self.data_concated = np.concatenate(self.data_buf, axis=1)
return
def _dump_file_path(self, file_path, add_time_tag=True):
""" add additional infor to the ump file path
"""
if add_time_tag:
time_tag = datetime.datetime.now().strftime("%Y%m%d%H%M%S")
return file_path + '_' + time_tag + self.dump_file_ext, \
file_path + '_name_' + time_tag + self.dump_file_ext
else:
return file_path + self.dump_file_ext, \
file_path + '_name_' + self.dump_file_ext, \
def dump(self, output_path='./debug/data_dump', add_time_tag=True):
""" dump(output_path='./debug/data_dump')
input
-----
output_path: str, path to store the dumped data
"""
# add additional infor to output_path name
output_path_new, output_file_lst = self._dump_file_path(
output_path, add_time_tag)
try:
os.mkdir(os.path.dirname(output_path_new))
os.mkdir(os.path.dirname(output_file_lst))
except OSError:
pass
## merge data if it has not been done
#if self.data_concated is None:
# self.merge_data()
#nii_io.f_write_raw_mat(self.data_concated, output_path_new)
self.data_dump_method(self.data_buf, output_path_new)
print("Data dumped to {:s}".format(output_path_new))
self.data_dump_method(self.data_names, output_file_lst)
print("Data dumped to {:s}".format(output_file_lst))
self.data_concated = None
return
if __name__ == '__main__':
print("Debugging tools")
| 4,915 | 27.581395 | 75 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/display.py | #!/usr/bin/env python
"""
dispaly.py
Tools to display the commands or warnings
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import datetime
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
class DisplayColors:
HEADER = '\033[95m'
OKBLUE = '\033[94m'
OKGREEN = '\033[92m'
WARNING = '\033[91m'
FAIL = '\033[91m'
ENDC = '\033[0m'
BOLD = '\033[1m'
UNDERLINE = '\033[4m'
def f_print(message, opt='ok', end='\n', flush=False):
""" f_print(message, opt)
Print message with specific style
Args:
message: str
opt: str, "warning", "highlight", "ok", "error"
"""
if opt == 'warning':
print(DisplayColors.WARNING + str(message) + DisplayColors.ENDC,
flush = flush, end = end)
elif opt == 'highlight':
print(DisplayColors.OKGREEN + str(message) + DisplayColors.ENDC,
flush = flush, end = end)
elif opt == 'ok':
print(DisplayColors.OKBLUE + str(message) + DisplayColors.ENDC,
flush = flush, end = end)
elif opt == 'error':
print(DisplayColors.FAIL + str(message) + DisplayColors.ENDC,
flush = flush, end = end)
else:
print(message, flush=flush, end=end)
return
def f_print_w_date(message, level='h'):
""" f_print_w_date(message, level)
Print message with date shown
Args:
message: a string
level: which can be 'h' (high-level), 'm' (middle-level), 'l' (low-level)
"""
if level == 'h':
message = '--- ' + str(message) + ' ' \
+ str(datetime.datetime.now()) + ' ---'
tmp = ''.join(['-' for x in range(len(message))])
f_print(tmp)
f_print(message)
f_print(tmp)
elif level == 'm':
f_print('---' + str(message) + ' ' \
+ str(datetime.datetime.now().time()) + '---')
else:
f_print(str(message) + ' ' + str(datetime.datetime.now().time()))
sys.stdout.flush()
return
def f_die(message):
""" f_die(message)
Print message in "error" mode and exit program with sys.exit(1)
"""
f_print("Error: " + message, 'error')
sys.exit(1)
def f_eprint(*args, **kwargs):
""" f_eprint(*args, **kwargs)
Print
"""
print(*args, file=sys.stderr, **kwargs)
def f_print_message(message, flush=False, end='\n'):
f_print(message, 'normal', flush=flush, end=end)
if __name__ == "__main__":
pass
| 2,555 | 25.350515 | 79 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/other_tools/log_parser.py | #!/usr/bin/env python
"""
log_parser
tools to parse log_train and log_err
"""
from __future__ import absolute_import
from __future__ import print_function
import numpy as np
import os
import re
import sys
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def f_read_log_err(file_path):
"""
each line looks like
10753,LJ045-0082,0,9216,0, 22/12100, Time: 0.190877s, Loss: 85.994621, ...
"""
def parse_line(line_input):
line_tmps = line_input.split(',')
tmp_loss = []
for tmp in line_tmps:
if tmp.count('Time'):
tmp_time = float(tmp.lstrip(' Time:').rstrip('s'))
elif tmp.count('Loss'):
tmp_loss.append(float(tmp.lstrip(' Loss:')))
return tmp_time, tmp_loss
time_mat = []
error_mat = []
with open(file_path, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip()
if line.count("Loss"):
tmp_time, tmp_loss = parse_line(line)
time_mat.append(tmp_time)
error_mat.append(tmp_loss)
return np.array(error_mat), np.array(time_mat)
# This function is obsolete
def f_read_log_err_old(file_path, train_num, val_num):
"""
log_train, log_val = f_read_log_err(log_err, num_train_utt, num_val_utt)
input:
-----
log_err: path to the log_err file
num_train_utt: how many training utterances
num_val_utt: how many validation utterances
output:
------
log_train: np.array, average error values per epoch on training set
log_val: np.array, average error values per epoch on valiation set
"""
data_str = []
with open(file_path, 'r') as file_ptr:
for line in file_ptr:
if not line.count('skip'):
try:
tmp = int(line[0])
data_str.append(line)
except ValueError:
pass
row = len(data_str)
col = len(np.fromstring(data_str[0], dtype=np.float32, sep=','))
data = np.zeros([row,col])
for idx, line in enumerate(data_str):
data[idx, :] = np.fromstring(line, dtype=np.float32, sep=',')
print(data.shape[0])
total_num = train_num + val_num
epoch_num = int(data.shape[0] / total_num)
data_train = np.zeros([epoch_num, data.shape[1]])
data_val = np.zeros([epoch_num, data.shape[1]])
for x in range(epoch_num):
temp_data = data[x * total_num:(x+1)*total_num, :]
train_part = temp_data[0:train_num,:]
val_part = temp_data[train_num:(train_num+val_num),:]
data_train[x, :] = np.mean(train_part, axis=0)
data_val[x, :] = np.mean(val_part, axis=0)
return data_train, data_val
def pass_number(input_str):
return np.array([float(x) for x in input_str.split()]).sum()
def f_read_log_train(file_path, sep='/'):
"""
data_train, data_val, time_per_epoch = read_log_train(path_to_log_train)
input:
-----
path_to_log_train: path to the log_train file
output:
------
data_train: error values per epoch on training set
data_val: error values per epoch on valiation set
time_per_epoch: training time per epoch
"""
def parse_line(line_input, sep):
if sep == ' ':
return line_input.split()
else:
return line_input.split(sep)
read_flag = False
data_str = []
with open(file_path, 'r') as file_ptr:
for line in file_ptr:
if read_flag and line.count('|') > 2:
data_str.append(line)
if line.count('Duration'):
read_flag = True
row = len(data_str)
data_train = None
data_val = None
time_per_epoch = np.zeros(row)
for idx, line in enumerate(data_str):
try:
time_per_epoch[idx] = float(line.split('|')[1])
except ValueError:
continue
trn_data = parse_line(line.split('|')[2], sep)
val_data = parse_line(line.split('|')[3], sep)
if data_train is None or data_val is None:
data_train = np.zeros([row, len(trn_data)])
data_val = np.zeros([row, len(val_data)])
for idx2 in np.arange(len(trn_data)):
data_train[idx, idx2] = pass_number(trn_data[idx2])
data_val[idx,idx2] = pass_number(val_data[idx2])
return data_train, data_val, time_per_epoch
def read_log_err_pytorch(file_path, merge_epoch=False):
def set_size(line):
return int(line.split('/')[1].split(',')[0])
def data_line(line):
if line.count("Time:"):
return True
else:
return False
def get_data(line):
return [float(x.split(":")[1]) for x in line.split(',') if x.count("Loss:")]
trn_utt_num = None
val_utt_num = None
trn_total_num = 0
val_total_num = 0
with open(file_path, 'r') as file_ptr:
for line in file_ptr:
if not data_line(line):
continue
temp_num = set_size(line)
col_num = len(get_data(line))
if trn_utt_num is None:
trn_utt_num = temp_num
if temp_num != val_utt_num and temp_num != trn_utt_num:
val_utt_num = temp_num
if trn_utt_num == temp_num:
trn_total_num += 1
if val_utt_num == temp_num:
val_total_num += 1
if trn_utt_num is None:
print("Cannot parse file")
return
if val_utt_num is None:
print("Trn %d, no val" % (trn_utt_num))
else:
print("Trn %d, val %d" % (trn_utt_num, val_utt_num))
print("Trn data %d, val data %d" % (trn_total_num, val_total_num))
trn_data = np.zeros([trn_total_num, col_num])
val_data = np.zeros([val_total_num, col_num])
trn_utt_cnt = 0
val_utt_cnt = 0
with open(file_path, 'r') as file_ptr:
for line in file_ptr:
if not data_line(line):
continue
data = get_data(line)
temp_num = set_size(line)
if trn_utt_num == temp_num:
trn_data[trn_utt_cnt, :] = np.array(data)
trn_utt_cnt += 1
if val_utt_num == temp_num:
val_data[val_utt_cnt, :] = np.array(data)
val_utt_cnt += 1
if merge_epoch:
trn_data_new = np.zeros([trn_total_num // trn_utt_num, col_num])
val_data_new = np.zeros([val_total_num // val_utt_num, col_num])
for idx in range(min([trn_total_num // trn_utt_num, val_total_num // val_utt_num])):
trn_data_new[idx, :] = trn_data[idx*trn_utt_num:(idx+1)*trn_utt_num, :].mean(axis=0)
val_data_new[idx, :] = val_data[idx*val_utt_num:(idx+1)*val_utt_num, :].mean(axis=0)
return trn_data_new, val_data_new
else:
return trn_data, val_data
if __name__ == "__main__":
print("logParser")
| 7,064 | 31.113636 | 96 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/seq_info.py | #!/usr/bin/env python
"""
seq_info
A class to log the information for one sample.
This data sequence could be one segment within a long utterance
"""
from __future__ import absolute_import
import os
import sys
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
class SeqInfo():
""" Definition of sequence information
Save the information about one utterance (which may be a trunck from
the original data utterance)
"""
def __init__(self,
length = 0,
seq_name = '',
seg_idx = 0,
start_pos = 0,
info_id = 0):
"""
Args:
length: length this utterance segment
seq_name: name of the utterance
seg_idx: idx of this segment in the original utterance
start_pos: from which step does this segment start in the
original utterance
info_id: idx of this seq segment in training set
"""
# data length (data stored on HDD)
self.length = int(length)
# name of the data sequence
self.seq_name = seq_name
# idx of the data in data set
self.seg_idx = seg_idx
# from which time step does this data sequence starts
self.start_pos = int(start_pos)
# other information
self.info_id = info_id
# add one slot for updating sequence information,
# this is only used for sampler shuffling
self.valid_len = 0
return
def print_to_dic(self):
"""
Print to dictionary format in order to dump
"""
return {"length": self.length,
"seq_name": self.seq_name,
"seg_idx": self.seg_idx,
"start_pos": self.start_pos,
"info_id": self.info_id}
def load_from_dic(self, dic):
"""
Load seq informaiton from dictionary
"""
try:
self.length = dic["length"]
self.seq_name = dic["seq_name"]
self.seg_idx = dic["seg_idx"]
self.start_pos = dic["start_pos"]
self.info_id = dic["info_id"]
except KeyError:
nii_warn.f_die("Seq infor %s invalid" % str(dic))
def print_to_str(self):
"""
Print infor to str
"""
temp = "{:d},{},{:d},{:d},{:d}".format(
self.info_id, self.seq_name, self.seg_idx, self.length,
self.start_pos)
return temp
def parse_from_str(self, input_str):
"""
Parse a input string (which should be generated from print_to_str)
"""
temp = input_str.split(',')
self.seq_name = temp[1]
try:
self.info_id = int(temp[0])
self.seg_idx = int(temp[2])
self.length = int(temp[3])
self.start_pos = int(temp[4])
except ValueError:
nii_warn.f_die("Seq infor cannot parse {}".format(input_str))
return
def seq_length(self):
return self.length
def seq_tag(self):
return self.seq_name
def seq_start_pos(self):
return self.start_pos
def seq_len_for_sampler(self):
return self.valid_len
def update_len_for_sampler(self, valid_len):
# udpate the valid length of the data
# due to data augmentation or trimming, the actual data sequence
# may be shorter than self.length.
# this affects sampler such as shuffle_by_seq_length
self.valid_len = valid_len
return
############
### Util to parse the output from print_to_str
############
def parse_length(input_str):
return int(input_str.split(',')[3])
def parse_filename(input_str):
return input_str.split(',')[1]
def parse_idx(input_str):
return int(input_str.split(',')[0])
if __name__ == "__main__":
print("Definition of seq_info class")
| 3,979 | 27.633094 | 74 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/wav_tools.py | #!/usr/bin/env python
"""
data_io
Interface to process waveforms.
Note that functions here are based on numpy, and they are intended to be used
before data are converted into torch tensors.
data on disk -> DataSet.__getitem__() -----> Collate ----> Pytorch model
numpy.tensor torch.tensor
These functions don't work on pytorch tensors
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import wave
import scipy.io.wavfile
try:
import soundfile
except ModuleNotFoundError:
pass
import core_scripts.data_io.io_tools as nii_io_tk
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
def mulaw_encode(x, quantization_channels, scale_to_int=True):
"""x_mu = mulaw_encode(x, quantization_channels, scale_to_int=True)
mu-law companding
input
-----
x: np.array, float-valued waveforms in (-1, 1)
quantization_channels (int): Number of channels
scale_to_int: Bool
True: scale mu-law to int
False: return mu-law in (-1, 1)
output
------
x_mu: np.array, mulaw companded wave
"""
mu = quantization_channels - 1.0
x_mu = np.sign(x) * np.log1p(mu * np.abs(x)) / np.log1p(mu)
if scale_to_int:
x_mu = np.array((x_mu + 1) / 2 * mu + 0.5, dtype=np.int32)
return x_mu
def mulaw_decode(x_mu, quantization_channels, input_int=True):
"""mulaw_decode(x_mu, quantization_channels, input_int=True)
mu-law decoding
input
-----
x_mu: np.array, mu-law waveform
quantization_channels: int, Number of channels
input_int: Bool
True: convert x_mu (int) from int to float, before mu-law decode
False: directly decode x_mu (float)
output
------
x: np.array, waveform from mulaw decoding
"""
mu = quantization_channels - 1.0
if input_int:
x = x_mu / mu * 2 - 1.0
else:
x = x_mu
x = np.sign(x) * (np.exp(np.abs(x) * np.log1p(mu)) - 1.0) / mu
return x
def alaw_encode(x, quantization_channels, scale_to_int=True, A=87.6):
"""x_a = alaw_encoder(x, quantization_channels, scale_to_int=True, A=87.6)
input
-----
x: np.array, float-valued waveforms in (-1, 1)
quantization_channels (int): Number of channels
scale_to_int: Bool
True: scale mu-law to int
False: return mu-law in (-1, 1)
A: float, parameter for a-law, default 87.6
output
------
x_a: np.array, a-law companded waveform
"""
num = quantization_channels - 1.0
x_abs = np.abs(x)
flag = (x_abs * A) >= 1
x_a = A * x_abs
x_a[flag] = 1 + np.log(x_a[flag])
x_a = np.sign(x) * x_a / (1 + np.log(A))
if scale_to_int:
x_a = np.array((x_a + 1) / 2 * num + 0.5, dtype=np.int32)
return x_a
def alaw_decode(x_a, quantization_channels, input_int=True, A=87.6):
"""alaw_decode(x_a, quantization_channels, input_int=True)
input
-----
x_a: np.array, mu-law waveform
quantization_channels: int, Number of channels
input_int: Bool
True: convert x_mu (int) from int to float, before mu-law decode
False: directly decode x_mu (float)
A: float, parameter for a-law, default 87.6
output
------
x: np.array, waveform
"""
num = quantization_channels - 1.0
if input_int:
x = x_a / num * 2 - 1.0
else:
x = x_a
sign = np.sign(x)
x_a_abs = np.abs(x)
x = x_a_abs * (1 + np.log(A))
flag = x >= 1
x[flag] = np.exp(x[flag] - 1)
x = sign * x / A
return x
def waveReadAsFloat(wavFileIn):
""" sr, wavData = wavReadToFloat(wavFileIn)
Wrapper over scipy.io.wavfile
Return:
sr: sampling_rate
wavData: waveform in np.float32 (-1, 1)
"""
sr, wavdata = scipy.io.wavfile.read(wavFileIn)
if wavdata.dtype is np.dtype(np.int16):
wavdata = np.array(wavdata, dtype=np.float32) / \
np.power(2.0, 16-1)
elif wavdata.dtype is np.dtype(np.int32):
wavdata = np.array(wavdata, dtype=np.float32) / \
np.power(2.0, 32-1)
elif wavdata.dtype is np.dtype(np.float32):
pass
else:
print("Unknown waveform format %s" % (wavFileIn))
sys.exit(1)
return sr, wavdata
def waveFloatToPCMFile(waveData, wavFile, bit=16, sr=16000):
"""waveSaveFromFloat(waveData, wavFile, bit=16, sr=16000)
Save waveData (np.float32) as PCM *.wav
Args:
waveData: waveform data as np.float32
wavFile: output PCM waveform file
bit: PCM bits
sr: sampling rate
"""
# recover to 16bit range [-32768, +32767]
rawData = waveData * np.power(2.0, bit-1)
rawData[rawData >= np.power(2.0, bit-1)] = np.power(2.0, bit-1)-1
rawData[rawData < -1*np.power(2.0, bit-1)] = -1*np.power(2.0, bit-1)
# write as signed 16bit PCM
if bit == 16:
rawData = np.asarray(rawData, dtype=np.int16)
elif bit == 32:
rawData = np.asarray(rawData, dtype=np.int32)
else:
print("Only be able to save wav in int16 and int32 type")
print("Save to int16")
rawData = np.asarray(rawData, dtype=np.int16)
scipy.io.wavfile.write(wavFile, sr, rawData)
return
def flacReadAsFloat(wavFileIn):
""" sr, wavData = flacReadAsFloat(wavFileIn)
Wrapper over soundfile.read
Return:
sr: sampling_rate
wavData: waveform in np.float32 (-1, 1)
"""
if 'soundfile' in sys.modules:
x, sr = soundfile.read(wavFileIn)
else:
print("soundfile is not installed.")
print("Due to practical reason, soundfile is not included in env.yml")
print("To install soundfile with support to flac, try:")
print(" conda install libsndfile=1.0.31 -c conda-forge")
print(" conda install pysoundfile -c conda-forge")
exit(1)
return sr, x
def readWaveLength(wavFileIn):
""" length = readWaveLength(wavFileIn)
Read the length of the waveform
Input:
waveFile, str, path to the input waveform
Return:
length, int, length of waveform
"""
with wave.open(wavFileIn, 'rb') as file_ptr:
wavlength = file_ptr.getnframes()
return wavlength
def buffering(x, n, p=0, opt=None):
"""buffering(x, n, p=0, opt=None)
input
-----
x: np.array, input signal, (length, )
n: int, window length
p: int, overlap, not frame shift
outpupt
-------
output: np.array, framed buffer, (frame_num, frame_length)
Example
-------
framed = buffer(wav, 320, 80, 'nodelay')
Code from https://stackoverflow.com/questions/38453249/
"""
if opt not in ('nodelay', None):
raise ValueError('{} not implemented'.format(opt))
i = 0
if opt == 'nodelay':
# No zeros at array start
result = x[:n]
i = n
else:
# Start with `p` zeros
result = np.hstack([np.zeros(p), x[:n-p]])
i = n-p
# Make 2D array, cast to list for .append()
result = list(np.expand_dims(result, axis=0))
while i < len(x):
# Create next column, add `p` results from last col if given
col = x[i:i+(n-p)]
if p != 0:
col = np.hstack([result[-1][-p:], col])
# Append zeros if last row and not length `n`
if len(col):
col = np.hstack([col, np.zeros(n - len(col))])
# Combine result with next row
result.append(np.array(col))
i += (n - p)
return np.vstack(result).astype(x.dtype)
def windowing(framed_buffer, window_type='hanning'):
"""windowing(framed_buffer, window_type='hanning')
input
-----
framed_buffer: np.array, (frame_num, frame_length), output of buffering
window_type: str, default 'hanning'
"""
if window_type == 'hanning':
window = np.hanning(framed_buffer.shape[1])
else:
assert False, "Unknown window type in windowing"
return framed_buffer * window.astype(framed_buffer.dtype)
def silence_handler(wav, sr, fl=320, fs=80,
max_thres_below=30,
min_thres=-55,
shortest_len_in_ms=50,
flag_output=0,
flag_norm_amp=True,
flag_only_startend_sil = False,
opt_silence_handler = -1):
"""silence_handler(wav, sr, fl=320, fs=80,
max_thres_below=30,
min_thres=-55,
shortest_len_in_ms=50,
flag_output=0,
flag_norm_amp=True,
flag_only_startend_sil = False,
opt_silence_handler = 1)
Based on the Speech activity detector mentioned in Sec5.1 of
Tomi Kinnunen, and Haizhou Li.
An Overview of Text-Independent Speaker Recognition: From Features to
Supervectors. Speech Communication 52 (1).
Elsevier: 12–40. doi:10.1016/j.specom.2009.08.009. 2010.
input
-----
wav: np.array, (wav_length, ), wavform data
sr: int, sampling rate
fl: int, frame length, default 320
fs: int, frame shift, in number of waveform poings, default 80
flag_output: int, flag to select output
0: return wav_no_sil, sil_wav, time_tag
1: return wav_no_sil
2: return sil_wav
max_thres_below: int, default 30, max_enenergy - max_thres_below
is the lower threshold for speech frame
min_thres: int, default -55, the lower threshold for speech frame
shortest_len_in_ms: int, ms, default 50 ms,
segment less than this length is treated as speech
flag_norm_amp: bool, whether normalize the waveform amplitude
based on window function (default True)
flag_only_startend_sil (obsolete): bool, whether only consider silence in
the begining and end. If False, silence within the utterance
will be marked / removed (default False)
opt_silence_handler: int, option to silence trim handler
0: equivalent to flag_only_startend_sil = False
1: equivalent to flag_only_startend_sil = True
2: remove only silence between words
-1: not use this option, but follow flag_only_startend_sil
output
------
wav_no_sil: np.array, (length_1, ), waveform after removing silence
sil_wav: np.array, (length_2, ), waveform in silence regions
frame_tag: np.array, [0, 0, 0, 1, 1, ..., 1, 0, ], where 0 indicates
silence frame, and 1 indicates a non-silence frame
Note: output depends on flag_output
"""
assert fs < fl, "Frame shift should be smaller than frame length"
# frame the singal
frames = buffering(wav, fl, fl - fs, 'nodelay')
# apply window to each frame
windowed_frames = windowing(frames)
# buffer to save window prototype, this is used to normalize the amplitude
window_proto = windowing(np.ones_like(frames))
# compute the frame energy and assign a sil/nonsil flag
frame_energy = 20*np.log10(np.std(frames, axis=1)+np.finfo(np.float32).eps)
frame_energy_max = np.max(frame_energy)
frame_tag = np.bitwise_and(
(frame_energy > (frame_energy_max - max_thres_below)),
frame_energy > min_thres)
frame_tag = np.asarray(frame_tag, dtype=np.int)
# post filtering of the sil/nonsil flag sequence
seg_len_thres = shortest_len_in_ms * sr / 1000 / fs
# function to ignore short segments
def ignore_short_seg(frame_tag, seg_len_thres):
frame_tag_new = np.zeros_like(frame_tag) + frame_tag
# boundary of each segment
seg_bound = np.diff(np.concatenate(([0], frame_tag, [0])))
# start of each segment
seg_start = np.argwhere(seg_bound == 1)[:, 0]
# end of each segment
seg_end = np.argwhere(seg_bound == -1)[:, 0]
assert seg_start.shape[0] == seg_end.shape[0], \
"Fail to extract segment boundaries"
# length of segment
seg_len = seg_end - seg_start
seg_short_ids = np.argwhere(seg_len < seg_len_thres)[:, 0]
for idx in seg_short_ids:
start_frame_idx = seg_start[idx]
end_frame_idx = seg_end[idx]
frame_tag_new[start_frame_idx:end_frame_idx] = 0
return frame_tag_new
# remove short sil segments
# 1-frame_tag indicates non-speech frames
frame_process_sil = ignore_short_seg(1-frame_tag, seg_len_thres)
# reverse the sign
frame_process_sil = 1 - frame_process_sil
# remove short nonsil segments
frame_process_all = ignore_short_seg(frame_process_sil, seg_len_thres)
frame_tag = frame_process_all
if opt_silence_handler < 0:
# if only consder silence in the front and end
if flag_only_startend_sil:
tmp_nonzero = np.flatnonzero(frame_tag)
# start of the first nonsil segment
#start_nonsil = np.asarray(frame_tag == 1).nonzero()[0]
if np.any(tmp_nonzero):
start_nonsil = np.flatnonzero(frame_tag)[0]
# end of the last nonsil segment
end_nonsil = np.flatnonzero(frame_tag)[-1]
# all segments between are switched to nonsil
frame_tag[start_nonsil:end_nonsil] = 1
else:
# no non-silence data, just let it pass
pass
elif opt_silence_handler == 1:
# if only consder silence in the front and end
tmp_nonzero = np.flatnonzero(frame_tag)
# start of the first nonsil segment
#start_nonsil = np.asarray(frame_tag == 1).nonzero()[0]
if np.any(tmp_nonzero):
start_nonsil = np.flatnonzero(frame_tag)[0]
# end of the last nonsil segment
end_nonsil = np.flatnonzero(frame_tag)[-1]
# all segments between are switched to nonsil
frame_tag[start_nonsil:end_nonsil] = 1
else:
# no non-silence data, just let it pass
pass
elif opt_silence_handler == 2:
# if only consder silence in the front and end
tmp_nonzero = np.flatnonzero(frame_tag)
# start of the first nonsil segment
#start_nonsil = np.asarray(frame_tag == 1).nonzero()[0]
if np.any(tmp_nonzero):
start_nonsil = np.flatnonzero(frame_tag)[0]
# end of the last nonsil segment
end_nonsil = np.flatnonzero(frame_tag)[-1]
# all segments between are switched to nonsil
frame_tag[:start_nonsil] = 1
frame_tag[end_nonsil:] = 1
else:
# no non-silence data, just let it pass
pass
else:
pass
# separate non-speech and speech segments
# do overlap and add
# buffer for speech segments
spe_buf = np.zeros([np.sum(frame_tag) * fs + fl], dtype=wav.dtype)
spe_buf_win = np.zeros([np.sum(frame_tag) * fs + fl], dtype=wav.dtype)
# buffer for non-speech segments
sil_buf = np.zeros([np.sum(1-frame_tag) * fs + fl], dtype=wav.dtype)
sil_buf_win = np.zeros([np.sum(1-frame_tag) * fs + fl], dtype=wav.dtype)
spe_fr_pt = 0
non_fr_pt = 0
for frame_idx, flag_speech in enumerate(frame_tag):
if flag_speech:
spe_buf[spe_fr_pt*fs:spe_fr_pt*fs+fl] += windowed_frames[frame_idx]
spe_buf_win[spe_fr_pt*fs:spe_fr_pt*fs+fl] += window_proto[frame_idx]
spe_fr_pt += 1
else:
sil_buf[non_fr_pt*fs:non_fr_pt*fs+fl] += windowed_frames[frame_idx]
sil_buf_win[non_fr_pt*fs:non_fr_pt*fs+fl] += window_proto[frame_idx]
non_fr_pt += 1
# normalize the amplitude if necessary
if flag_norm_amp:
spe_buf_win[spe_buf_win < 0.0001] = 1.0
sil_buf_win[sil_buf_win < 0.0001] = 1.0
spe_buf /= spe_buf_win
sil_buf /= sil_buf_win
if flag_output == 1:
return spe_buf
elif flag_output == 2:
return sil_buf
else:
return spe_buf, sil_buf, frame_tag
###################
# wrapper functions
###################
def silence_handler_wrapper(wav, sr, fl=320, fs=80,
max_thres_below=30,
min_thres=-55,
shortest_len_in_ms=50,
flag_output=0,
flag_norm_amp=True,
flag_only_startend_sil=False):
"""Wrapper over silence_handler
Many APIs used in this project assume (length, 1) shape.
Thus, this API is a wrapper to accept (length, 1) and output (length, 1)
See more on silence_handler
"""
output = silence_handler(
wav[:, 0], sr, fl, fs, max_thres_below,
min_thres, shortest_len_in_ms,
flag_output, flag_norm_amp, flag_only_startend_sil)
if flag_output == 1:
# from (length) to (length, 1)
return np.expand_dims(output, axis=1)
elif flag_output == 2:
return np.expand_dims(output, axis=1)
else:
return np.expand_dims(output[0], axis=1), \
np.expand_dims(output[1], axis=1), \
output[2]
###################
# Other tools
###################
def wav_get_amplitude(waveform, method='max'):
"""
input
-----
wavform: np.array, (length, 1)
method: str,
'max': compute np.max(np.abs(waveform))
'mean': compute np.mean(np.abs(waveform))
output
------
amp: np.array (1)
"""
if method == 'max':
return np.max(np.abs(waveform))
else:
return np.mean(np.abs(waveform))
def wav_norm_amplitude(waveform, method='max', floor=1e-12):
"""
input
-----
wavform: np.array, (length, 1)
method: str,
'max': compute np.max(np.abs(waveform))
'mean': compute np.mean(np.abs(waveform))
output
------
amp: np.array (1)
"""
amp = wav_get_amplitude(waveform, method=method)
amp = amp + floor if amp < floor else amp
return waveform / amp
def wav_scale_amplitude_to(waveform, amp, method = 'max'):
"""
input
-----
wavform: np.array, (length, 1)
get_amp_method: str,
'max': compute np.max(np.abs(wavform))
'mean': compute np.mean(np.abs(wavform))
output
------
waveform: np.array, (length, 1)
"""
return wav_norm_amplitude(waveform, method=method) * amp
###################
# legacy functions
###################
def wavformRaw2MuLaw(wavdata, bit=16, signed=True, quanLevel = 256.0):
"""
wavConverted = wavformRaw2MuLaw(wavdata, bit=16, signed=True, \
quanLevel = 256.0)
Assume wavData is int type:
step1. convert int wav -> float wav
step2. convert linear scale wav -> mu-law wav
Args:
wavdata: np array of int-16 or int-32 waveform
bit: number of bits to encode waveform
signed: input is signed or not
quanLevel: level of quantization (default 2 ^ 8)
Returned:
wav: integer stored as float numbers
"""
if wavdata.dtype != np.int16 and wavdata.dtype != np.int32:
print("Input waveform data in not int16 or int32")
sys.exit(1)
# convert to float numbers
if signed==True:
wavdata = np.array(wavdata, dtype=np.float32) / \
np.power(2.0, bit-1)
else:
wavdata = np.array(wavdata, dtype=np.float32) / \
np.power(2.0, bit)
tmp_quan_level = quanLevel - 1
# mu-law compansion
wavtrans = np.sign(wavdata) * \
np.log(1.0 + tmp_quan_level * np.abs(wavdata)) / \
np.log(1.0 + tmp_quan_level)
wavtrans = np.round((wavtrans + 1.0) * tmp_quan_level / 2.0)
return wavtrans
def wavformMuLaw2Raw(wavdata, quanLevel = 256.0):
"""
waveformMuLaw2Raw(wavdata, quanLevel = 256.0)
Convert Mu-law waveform back to raw waveform
Args:
wavdata: np array
quanLevel: level of quantization (default: 2 ^ 8)
Return:
raw waveform: np array, float
"""
tmp_quan_level = quanLevel - 1
wavdata = wavdata * 2.0 / tmp_quan_level - 1.0
wavdata = np.sign(wavdata) * (1.0/ tmp_quan_level) * \
(np.power(quanLevel, np.abs(wavdata)) - 1.0)
return wavdata
def float2wav(rawData, wavFile, bit=16, samplingRate = 16000):
"""
float2wav(rawFile, wavFile, bit=16, samplingRate = 16000)
Convert float waveform into waveform in int
This is identitcal to waveFloatToPCMFile
To be removed
Args:
rawdata: float waveform data in np-arrary
wavFile: output file path
bit: number of bits to encode waveform in output *.wav
samplingrate:
"""
rawData = rawData * np.power(2.0, bit-1)
rawData[rawData >= np.power(2.0, bit-1)] = np.power(2.0, bit-1)-1
rawData[rawData < -1*np.power(2.0, bit-1)] = -1*np.power(2.0, bit-1)
# write as signed 16bit PCM
if bit == 16:
rawData = np.asarray(rawData, dtype=np.int16)
elif bit == 32:
rawData = np.asarray(rawData, dtype=np.int32)
else:
print("Only be able to save wav in int16 and int32 type")
print("Save to int16")
rawData = np.asarray(rawData, dtype=np.int16)
scipy.io.wavfile.write(wavFile, samplingRate, rawData)
return
#################################
# Other utilities based on Numpy
#################################
def f_overlap_cat(data_list, overlap_length):
"""Wrapper for overlap and concatenate
input:
-----
data_list: list of np.array, [(length1, dim), (length2, dim)]
output
------
data: np.array, (length1 + length2 ... - overlap_length * N, dim)
"""
data_dtype = data_list[0].dtype
if data_list[0].ndim == 1:
dim = 1
else:
dim = data_list[0].shape[1]
total_length = sum([x.shape[0] for x in data_list])
data_gen = np.zeros([total_length, dim], dtype=data_dtype)
prev_end = 0
for idx, data_trunc in enumerate(data_list):
tmp_len = data_trunc.shape[0]
if data_trunc.ndim == 1:
data_tmp = np.expand_dims(data_trunc, 1)
else:
data_tmp = data_trunc
if idx == 0:
data_gen[0:tmp_len] = data_tmp
prev_end = tmp_len
else:
win_len = min([prev_end, overlap_length, tmp_len])
win_cof = np.arange(0, win_len)/win_len
win_cof = np.expand_dims(win_cof, 1)
data_gen[prev_end - win_len:prev_end] *= 1.0 - win_cof
data_tmp[:win_len] *= win_cof
data_gen[prev_end-win_len:prev_end-win_len+tmp_len] += data_tmp
prev_end = prev_end-win_len+tmp_len
return data_gen[0:prev_end]
if __name__ == "__main__":
print("Definition of tools for wav")
| 23,187 | 31.340307 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/wav_augmentation.py | #!/usr/bin/env python
"""
Functions for waveform augmentation.
Note that
1. functions here are based on numpy, and they are intended to be used
before data are converted into torch tensors.
data on disk -> DataSet.__getitem__() -----> Collate ----> Pytorch model
numpy.tensor torch.tensor
These functions don't work on pytorch tensors
2. RawBoost functions are based on those by H.Tak and M.Todisco
See code here
https://github.com/TakHemlata/RawBoost-antispoofing
Hemlata Tak, Madhu R Kamble, Jose Patino, Massimiliano Todisco, and
Nicholas W D Evans. RawBoost: A Raw Data Boosting and Augmentation Method
Applied to Automatic Speaker Verification Anti-Spoofing. Proc. ICASSP. 2022
"""
from __future__ import absolute_import
import os
import sys
import copy
import numpy as np
from scipy import signal
from pathlib import Path
try:
from pydub import AudioSegment
except ModuleNotFoundError:
pass
import core_scripts.data_io.wav_tools as nii_wav_tools
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
################
# Tool
################
def unify_length_shape(data, target_data):
""" output = unify_length_shape(data, target_data)
input
-----
data: np.array, either (L), or (L, 1)
target_data: np.array, either (L) or (L, 1)
output
------
output: np.array that has same shape as target_data
"""
output_buf = np.zeros_like(target_data)
min_len = min([data.shape[0], target_data.shape[0]])
if data.ndim == target_data.ndim:
output_buf[:min_len] = data[:min_len]
elif data.ndim == 1:
output_buf[:min_len, 0] = data[:min_len]
elif data.ndim == 2:
output_buf[:min_len] = data[:min_len, 0]
else:
print("Implementation error in unify_length_shape")
sys.exit(1)
return output_buf
################
# Time domain
################
def wav_rand_sil_trim(wav,
sr,
random_trim_sil = False,
random_trim_nosil = False):
""" output = wav_rand_sil_trim(
wav, sr,
random_trim_sil = False,
random_trim_nosil = False)
Randomly trim the leading and ending silence.
input
-----
wav: np.array, (length, 1)
sr: int, waveform sampling rate
random_trim_sil: bool, randomly trim silence (default False)
random_trim_nosil: bool, randomly trim no-silence segments
output
------
output: np.array, (length, 1)
start_idx: int, starting time index in the input wav
end_idx: int, ending time index in the input wav
output <- wav[start_idx:end_idx]
"""
# original length
orig_len = wav.shape[0]
# frame shift for silence detection, fixed here
fs=80
# get the time flag for silence region
_, _, frame_tag = nii_wav_tools.silence_handler_wrapper(
wav, sr, fs=fs, flag_only_startend_sil = True)
# get the ending time of the leading silence
# get the starting time of the trailing silence
#
if len(np.flatnonzero(frame_tag)):
# start and end position
start_nonsil = np.flatnonzero(frame_tag)[0] * fs
end_nonsil = np.flatnonzero(frame_tag)[-1] * fs
else:
# no silence, use the original entire data
start_nonsil = 0
end_nonsil = wav.shape[0]
# if further randomly trim,
if random_trim_sil:
prob = np.random.rand()
start_nosil_new = int(start_nonsil * prob)
end_nosil_new = int((orig_len - end_nonsil)*prob) + end_nonsil
else:
start_nosil_new = start_nonsil
end_nosil_new = end_nonsil
# get the non-silence region
if start_nosil_new < end_nosil_new and start_nosil_new > 0:
input_new = wav[start_nosil_new:end_nosil_new]
else:
input_new = wav
return input_new, start_nosil_new, end_nosil_new
def wav_time_mask(input_data, wav_samp_rate):
""" output = wav_time_mask(input_data, wav_samp_rate)
Apply time mask and zero-out segments
input
-----
input_data: np.array, (length, 1)
wav_samp_rate: int, waveform sampling rate
output
------
output: np.array, (length, 1)
"""
# choose the codec
seg_width = int(np.random.rand() * 0.2 * wav_samp_rate)
start_idx = int(np.random.rand() * (input_data.shape[0] - seg_width))
if start_idx < 0:
start_idx = 0
if (start_idx + seg_width) > input_data.shape[0]:
seg_width = input_data.shape[0] - start_idx
tmp = np.ones_like(input_data)
tmp[start_idx:start_idx+seg_width] = 0
return input_data * tmp
def batch_siltrim_for_multiview(input_data_batch, wav_samp_rate,
random_trim_sil=False,
random_trim_nosil=False):
""" output = batch_trim(input_data, wav_samp_rate)
For multi-view data, trim silence
input
-----
input_data: list of np.array, (length, 1)
wav_samp_rate: int, waveform sampling rate
output
------
output: list of np.array, (length, 1)
"""
# output buffer
output_data_batch = []
# original length
orig_len = input_data_batch[0].shape[0]
# get the starting and ending of non-silence region
# (computed based on the first wave in the list)
_, start_time, end_time = wav_rand_sil_trim(
input_data_batch[0], wav_samp_rate,
random_trim_sil, random_trim_nosil)
# do trimming on all waveforms in the input list
if start_time < end_time and start_time > 0:
for data in input_data_batch:
output_data_batch.append(data[start_time:end_time])
else:
for data in input_data_batch:
output_data_batch.append(data)
return output_data_batch
def batch_pad_for_multiview(input_data_batch_, wav_samp_rate, length,
random_trim_nosil=False, repeat_pad=False):
""" output = batch_pad_for_multiview(
input_data_batch, wav_samp_rate, length, random_trim_nosil=False)
If input_data_batch is a single trial, trim it to a fixed length
For multi-view data, trim all the trials to a fixed length, using the same
random start and end
input
-----
input_data: list of np.array, (length, 1)
wav_samp_rate: int, waveform sampling rate
output
------
output: list of np.array, (length, 1)
"""
# unify the length of input data before further processing
def _ad_length(x, length, repeat_pad):
# adjust the length of the input x
if length > x.shape[0]:
if repeat_pad:
rt = int(length / x.shape[0]) + 1
tmp = np.tile(x, (rt, 1))[0:length]
else:
tmp = np.zeros([length, 1])
tmp[0:x.shape[0]] = x
else:
tmp = x[0:length]
return tmp
# use the first data in the list
firstlen = input_data_batch_[0].shape[0]
input_data_batch = [_ad_length(x, firstlen, repeat_pad) \
for x in input_data_batch_]
#
new_len = input_data_batch[0].shape[0]
if repeat_pad is False:
# if we simply trim longer sentence but not pad shorter sentence
if new_len < length:
start_len = 0
end_len = new_len
elif random_trim_nosil:
start_len = int(np.random.rand() * (new_len - length))
end_len = start_len + length
else:
start_len = 0
end_len = length
input_data_batch_ = input_data_batch
else:
if new_len < length:
start_len = 0
end_len = length
rt = int(length / new_len) + 1
# repeat multiple times
input_data_batch_ = [np.tile(x, (rt, 1)) for x in input_data_batch]
elif random_trim_nosil:
start_len = int(np.random.rand() * (new_len - length))
end_len = start_len + length
input_data_batch_ = input_data_batch
else:
start_len = 0
end_len = length
input_data_batch_ = input_data_batch
output_data_batch = [x[start_len:end_len] for x in input_data_batch_]
return output_data_batch
##################
# Frequency domain
##################
def wav_freq_mask_fixed(input_data, wav_samp_rate, start_b, end_b):
""" output = wav_freq_mask_fixed(input_data, wav_samp_rate, start_b, end_b)
Mask the frequency range, fixed
input
-----
input_data: np.array, (length, 1)
wav_samp_rate: int, waveform sampling rate
start_b: float
end_b: float
output
------
output: np.array, (length, 1)
"""
# order of the filder, fixed to be 10
# change it to a random number later
filter_order = 10
if start_b < 0.01:
sos = signal.butter(filter_order, end_b, 'highpass', output='sos')
elif end_b > 0.99:
sos = signal.butter(filter_order, start_b, 'lowpass',output='sos')
else:
sos = signal.butter(
filter_order, [start_b, end_b], 'bandstop', output='sos')
filtered = signal.sosfilt(sos, input_data[:, 0])
# change dimension
output = np.expand_dims(filtered, axis=1)
return output
def wav_freq_mask(input_data, wav_samp_rate):
""" output = wav_freq_mask(input_data, wav_samp_rate)
Randomly mask the signal in frequency domain
input
-----
input_data: np.array, (length, 1)
wav_samp_rate: int, waveform sampling rate
output
------
output: np.array, (length, 1)
"""
# order of the filder, fixed to be 10
# change it to a random number later
filter_order = 10
# maximum value of the bandwidth for frequency masking
max_band_witdh = 0.2
# actual band_w
band_w = np.random.rand() * max_band_witdh
if band_w < 0.05:
# if the bandwidth is too small, do no masking
output = input_data
else:
# start
start_b = np.random.rand() * (1 - band_w)
# end
end_b = start_b + band_w
output = wav_freq_mask_fixed(input_data, wav_samp_rate, start_b, end_b)
return output
##################
# Compression codec
##################
def wav_codec(input_dat, wav_samp_rate):
""" A wrapper to use pyDub and ffmpeg
This requires pyDub and ffmpeg installed.
"""
tmpdir = '/tmp/xwtemp'
if not os.path.isdir(tmpdir):
os.mkdir(tmpdir)
randomname = "{:s}/{:010d}".format(tmpdir, np.random.randint(100000))
while os.path.isfile(randomname + '.empty'):
randomname = "{:s}/{:010d}".format(tmpdir, np.random.randint(100000))
Path(randomname + '.empty').touch()
# write to file (16bit PCM)
nii_wav_tools.waveFloatToPCMFile(input_dat[:, 0], randomname + '.wav')
data = AudioSegment.from_wav(randomname + '.wav')
# choose the codec
rand_codec = np.random.randint(2)
if rand_codec == 0:
# mp3
rand_bitrate = np.random.randint(6)
if rand_bitrate == 0:
rate = '16k'
elif rand_bitrate == 1:
rate = '32k'
elif rand_bitrate == 2:
rate = '64k'
elif rand_bitrate == 3:
rate = '128k'
elif rand_bitrate == 4:
rate = '256k'
else:
rate = '320k'
data.export(randomname + '.tmp', format='mp3',
codec='libmp3lame', bitrate=rate)
data_codec = AudioSegment.from_mp3(randomname + '.tmp')
output_dat = np.array(data_codec.get_array_of_samples() /
np.power(2, 16-1))
# clean
#os.remove(randomname + '.wav')
try:
os.remove(randomname + '.tmp')
except FileNotFoundError:
pass
#os.remove(randomname + '.empty')
else:
# opus ogg
rand_bitrate = np.random.randint(5)
if rand_bitrate == 0:
rate = '16k'
elif rand_bitrate == 1:
rate = '32k'
elif rand_bitrate == 2:
rate = '64k'
elif rand_bitrate == 3:
rate = '128k'
else:
rate = '256k'
data.export(randomname + '.tmp', format='opus',
bitrate = rate, codec='libopus')
data_codec = AudioSegment.from_file(
randomname + '.tmp', format='ogg', codec='libopus')
data_codec = data_codec.set_frame_rate(16000)
output_dat = np.array(data_codec.get_array_of_samples() /
np.power(2, 16-1))
# clean
#os.remove(randomname + '.wav')
try:
os.remove(randomname + '.tmp')
except FileNotFoundError:
pass
#os.remove(randomname + '.empty')
try:
os.remove(randomname + '.wav')
os.remove(randomname + '.empty')
except FileNotFoundError:
pass
if output_dat.shape[0] != input_dat.shape[0]:
output_dat_ = np.zeros_like(input_dat[:, 0])
minlen = min([output_dat.shape[0], input_dat.shape[0]])
output_dat_[0:minlen] = output_dat_[0:minlen]
output_dat = output_dat_
return np.expand_dims(output_dat, axis=1)
##################
# Waveform morphing
##################
def morph_wavform(wav1, wav2, para=0.5, method=2,
fl = 320, fs = 160, nfft = 1024):
""" output = morph_waveform(wav1, wav2, method=2, para=0.5
fl = 320, fs = 160, nfft = 1024)
input
-----
wav1: np.array, (L,) or (L,1), input waveform 1
wav2: np.array, (L,) or (L,1), input waveform 2
para: float, coefficient for morphing
method: int,
method = 1, waveform level morphing
output = wav1 * para + wav2 * (1-para)
method = 2, spec amplitude morphing
amplitude = wav1_amp * para + wav2_amp * (1-para)
phase is from wav2
method = 3, phase morphing
...
amplitude is from wav1
method = 4, both spec and phase
fl: int, frame length for STFT analysis when method > 1
fs: int, frame shift for STFT analysis when method > 1
nfft: int, fft points for STFT analysis when method > 1
output
------
output: np.array, same shape as wav1 and wav2
"""
length = min([wav1.shape[0], wav2.shape[0]])
if wav1.ndim > 1:
data1 = wav1[0:length, 0]
else:
data1 = wav1[0:length]
if wav2.ndim > 1:
data2 = wav2[0:length, 0]
else:
data2 = wav2[0:length]
if method == 1 or method == 'wav':
# waveform level
data = data1 * para + data2 * (1.0 - para)
elif method == 2 or method == 'specamp':
# spectrum amplitude
_, _, Zxx1 = signal.stft(
data1, nperseg=fl, noverlap=fl - fs, nfft=nfft)
_, _, Zxx2 = signal.stft(
data2, nperseg=fl, noverlap=fl - fs, nfft=nfft)
amp1 = np.abs(Zxx1)
amp2 = np.abs(Zxx2)
pha1 = np.angle(Zxx1)
pha2 = np.angle(Zxx2)
# merge amplitude
amp = np.power(amp1, para) * np.power(amp2, (1.0 - para))
#
Zxx = amp * np.cos(pha1) + 1j * amp * np.sin(pha1)
_, data = signal.istft(
Zxx, nperseg = fl, noverlap = fl - fs, nfft = nfft)
elif method == 3 or method == 'phase':
# phase,
_, _, Zxx1 = signal.stft(
data1, nperseg=fl, noverlap=fl - fs, nfft=nfft)
_, _, Zxx2 = signal.stft(
data2, nperseg=fl, noverlap=fl - fs, nfft=nfft)
amp1 = np.abs(Zxx1)
amp2 = np.abs(Zxx2)
pha1 = np.unwrap(np.angle(Zxx1))
pha2 = np.unwrap(np.angle(Zxx2))
#amp = amp1 * para + amp2 * (1.0 - para)
pha = pha1 * para + pha2 * (1.0 - para)
Zxx = amp1 * np.cos(pha1) + 1j * amp1 * np.sin(pha)
_, data = signal.istft(
Zxx, nperseg=fl, noverlap=fl-fs, nfft=nfft)
elif method == 4 or method == 'specamp-phase':
# both
_, _, Zxx1 = signal.stft(
data1, nperseg=fl, noverlap=fl - fs, nfft=nfft)
_, _, Zxx2 = signal.stft(
data2, nperseg=fl, noverlap=fl - fs, nfft=nfft)
amp1 = np.abs(Zxx1)
amp2 = np.abs(Zxx2)
pha1 = np.unwrap(np.angle(Zxx1))
pha2 = np.unwrap(np.angle(Zxx2))
amp = np.power(amp1, para) * np.power(amp2, (1.0 - para))
pha = pha1 * para + pha2 * (1.0 - para)
Zxx = amp * np.cos(pha1) + 1j * amp * np.sin(pha)
_, data = signal.istft(
Zxx, nperseg=fl, noverlap=fl-fs, nfft=nfft)
# adjust length & shape
data = unify_length_shape(data, wav1)
return data
##################
# reverberation
##################
def wav_reverb(waveform, rir, use_fft=True, keep_alignment=False):
""" output = wav_reverb(waveform, rir, use_fft=True, keep_alignment=False)
input
-----
waveform: np.array, (length, 1), input waveform
rir: np.array, (length, 1), room impulse response
use_fft: bool,
True: use FFT to do convolution (default)
False: use scipy.lfilter to do convolution (not implemented yet)
keep_alignment: bool
True: shift RIR so that max of RIR starts from 1st time step
False: do nothing (default)
output
------
output_wav: np.array, (length, 1)
"""
if use_fft:
# handling different length
signal_length = max([waveform.shape[0], rir.shape[0]])
# buffer
waveform_buf = np.zeros([signal_length])
rir_buf = np.zeros([signal_length])
waveform_buf[:waveform.shape[0]] = waveform[:, 0]
rir_buf[:rir.shape[0]] = rir[:, 0]
# alignment
if keep_alignment:
# get the max value of RIR
max_index = np.argmax(rir, axis=0)[0]
# circular shift the buffer
rir_buf = np.roll(rir_buf, -max_index)
# fft
convolved = np.fft.rfft(waveform_buf) * np.fft.rfft(rir_buf)
# ifft
output_wav = np.fft.irfft(convolved)
# adjust volume
orig_amp = nii_wav_tools.wav_get_amplitude(
waveform, method='max')
output_wav = nii_wav_tools.wav_scale_amplitude_to(
output_wav, orig_amp, method='max')
else:
print("Not implemented")
sys.exit(1)
return np.expand_dims(output_wav, axis=1)
##################
# RawBoost
#
# https://github.com/TakHemlata/RawBoost-antispoofing/blob/main/RawBoost.py
#
# MIT license
# Copyright (c) 2021 Hemlata
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
##################
def randRange(x1, x2, integer):
y = np.random.uniform(low=x1, high=x2, size=(1,))
if integer:
y = int(y)
return y
def normWav(x, always):
if always:
x = x/np.amax(abs(x))
elif np.amax(abs(x)) > 1:
x = x/np.amax(abs(x))
return x
def genNotchCoeffs(nBands, minF, maxF, minBW, maxBW,
minCoeff, maxCoeff, minG, maxG, fs):
b = 1
for i in range(0, nBands):
fc = randRange(minF,maxF,0);
bw = randRange(minBW,maxBW,0);
c = randRange(minCoeff,maxCoeff,1);
if c/2 == int(c/2):
c = c + 1
f1 = fc - bw/2
f2 = fc + bw/2
if f1 <= 0:
f1 = 1/1000
if f2 >= fs/2:
f2 = fs/2-1/1000
b = np.convolve(
signal.firwin(c, [float(f1), float(f2)], window='hamming', fs=fs),
b)
G = randRange(minG,maxG,0);
_, h = signal.freqz(b, 1, fs=fs)
b = pow(10, G/20)*b/np.amax(abs(h))
return b
def filterFIR(x,b):
N = b.shape[0] + 1
xpad = np.pad(x, (0, N), 'constant')
y = signal.lfilter(b, 1, xpad)
y = y[int(N/2):int(y.shape[0]-N/2)]
return y
def LnL_convolutive_noise(x,
fs=16000, N_f=5, nBands=5,
minF=20, maxF=8000, minBW=100, maxBW=1000,
minCoeff=10, maxCoeff=100, minG=0, maxG=0,
minBiasLinNonLin=5, maxBiasLinNonLin=20):
# Linear and non-linear convolutive noise
y = [0] * x.shape[0]
for i in range(0, N_f):
if i == 1:
minG = minG-minBiasLinNonLin;
maxG = maxG-maxBiasLinNonLin;
b = genNotchCoeffs(nBands, minF, maxF, minBW,
maxBW, minCoeff, maxCoeff,minG,maxG,fs)
y = y + filterFIR(np.power(x, (i+1)), b)
y = y - np.mean(y)
y = normWav(y,0)
return y
def ISD_additive_noise(x, P=10, g_sd=2):
# Impulsive signal dependent noise
beta = randRange(0, P, 0)
y = copy.deepcopy(x)
x_len = x.shape[0]
n = int(x_len*(beta/100))
p = np.random.permutation(x_len)[:n]
f_r= np.multiply(((2*np.random.rand(p.shape[0]))-1),
((2*np.random.rand(p.shape[0]))-1))
r = g_sd * x[p] * f_r
y[p] = x[p] + r
y = normWav(y,0)
return y
def SSI_additive_noise(x, SNRmin, SNRmax, nBands, minF, maxF, minBW, maxBW,
minCoeff, maxCoeff, minG, maxG, fs):
# Stationary signal independent noise
noise = np.random.normal(0, 1, x.shape[0])
b = genNotchCoeffs(nBands, minF, maxF, minBW, maxBW,
minCoeff, maxCoeff, minG, maxG, fs)
noise = filterFIR(noise, b)
noise = normWav(noise,1)
SNR = randRange(SNRmin, SNRmax, 0)
noise = noise / np.linalg.norm(noise,2) * np.linalg.norm(x,2) / 10.0**(0.05 * SNR)
x = x + noise
return x
def RawBoostWrapper12(x, fs=16000):
""" RawBoost strategy 1+2
"""
if x.ndim > 1:
x_ = x[:, 0]
else:
x_ = x
y = LnL_convolutive_noise(x_, fs)
y = ISD_additive_noise(y)
# make the length equal
length = min(x.shape[0], y.shape[0])
y_ = np.zeros_like(x_)
y_[0:length] = y[0:length]
if x.ndim > 1:
y_ = np.expand_dims(y_, axis=1)
return y_
if __name__ == "__main__":
print("Waveform augmentation tools")
| 23,456 | 29.463636 | 86 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/customize_dataset.py | #!/usr/bin/env python
"""
customized dataset
NII_MergeDataSetLoader (to one minibatch):
We want to load dataset 1, 2, and 3,
We also want to draw sample from each dataset for one minibatch.
One epoch over the merged datasets will be decided by the smallest dataset
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import torch
import torch.utils.data
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.default_data_io as nii_default_dset
import core_scripts.data_io.customize_collate_fn as nii_collate_fn
import core_scripts.data_io.customize_sampler as nii_sampler_fn
import core_scripts.data_io.conf as nii_dconf
import core_scripts.data_io.seq_info as nii_seqinfo
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
###############################################
# Dataset definition to merge multiple datasets
###############################################
class merge_loader():
""" merge_loader
Data loader for customized data with m_concate_set = None
By default, draw equal number of samples from each subset
__iter__():
__next__(): load data and merge into minibatch
Note: pay attention to adjust the data index in the dataset.
See comment of adjust_utt_idx
"""
def __init__(self, datasets):
# list of datasets
self.m_datasets = datasets
# initialized iterators
self.m_loaders = [x.get_loader() for x in self.m_datasets]
# utterance index shift
# if we merge dataset 1, 2, 3, then
# index for dataset 2: index += dataset_1.get_seq_num()
# index for dataset 3: index += dataset_1 + dataset_2.get_seq_num()
self.m_idx_shift = np.cumsum([0] +
[x.get_seq_num() for x in self.m_datasets])
return
def adjust_utt_idx(self, data_tuple, dataset_idx):
""" adjust_utt_idx(data_tutple, dataset_idx)
when merging dataset 1, 2, 3 ...
index for dataset 2: index += dataset_1.get_seq_num()
index for dataset 3: index += dataset_1 + dataset_2.get_seq_num()
We have to call dataset.f_adjust_idx because it is the dataset itself
that knows how to parse the data_tuple
"""
return self.m_datasets[dataset_idx].get_dataset().f_adjust_idx(
data_tuple, self.m_idx_shift[dataset_idx])
def __iter__(self):
"""
create the list of iterators
"""
self.m_loader_iter = [iter(x) for x in self.m_loaders]
return self
def __next__(self):
""" try to load data from m_datasets, and merge them into a
single minibatch
"""
try:
data_list = []
for dataset_idx, dataloader in enumerate(self.m_loader_iter):
data_list.append(
self.adjust_utt_idx(next(dataloader), dataset_idx))
#
return nii_collate_fn.customize_collate_from_batch(data_list)
except StopIteration:
raise StopIteration
class ConcatDataset(torch.utils.data.Dataset):
""" ConcatDataset
Torch.Dataset that concatenate multiple sub-datasets.
Adopted from
https://discuss.pytorch.org/t/train-simultaneously-on-two-datasets/649/2
But here we concatenate data corpora directly. Minibatch may contain data
from each sub corpus
"""
def __init__(self, datasets):
""" ConcatDataset(datasets)
Args
----
datasets: list of torch.utils.data.Dataset or derived classes
it must have __len__, __getitem__, and adjust_utt_idx
"""
# all the sub sets
self.datasets = datasets
self.num_subset = len(datasets)
# initial the len and len_top, len_bot
self.__init_sub()
return
def __init_sub(self):
# len of each sub set
self.len_buffer = [x.__len__() for x in self.datasets]
# for later use, to decide from which subset we draw the sample
self.len_top = np.cumsum(self.len_buffer)
self.len_bot = np.cumsum([0] + self.len_buffer[:-1])
# name <--> idx mapper (idx in the whole dataset)
self.name_idx_map = {}
for idx_u, idx_d, subset in \
zip(self.len_top, self.len_bot, self.datasets):
data_name_list = subset.f_get_seq_name_list()
for idx_data, data_name in enumerate(data_name_list):
# similar to the login in getitem, we need to add the
# shift idx_d for each subset
self.name_idx_map[data_name] = idx_data + idx_d
# done
return
def __getitem__(self, i):
""" getitem from the corresponding subcorpus
"""
# for example, data1 = [a], data2 = [b, c]
# self.len_buffer = [1, 2]
# self.len_top = [1, 3]
# self.len_bot = [0, 1]
# __getitem__(0) -> data1[0-0] = a
# __getitem__(1) -> data2[1-1] = b
# __getitem__(2) -> data2[2-1] = c
#
# Notice that the index of b is 0 in data2 but 1 in concatenated set
# The index must be adjusted using f_adjust_idx
for idx_u, idx_d, subset in \
zip(self.len_top, self.len_bot, self.datasets):
if i < idx_u:
# get the data from subdataset
orig_data_tuple = subset.__getitem__(i - idx_d)
# adjust the data idx
return subset.f_adjust_idx(orig_data_tuple, idx_d)
else:
# keep going to the next subset
pass
nii_warn.f_die("Merge dataset: fatal error in __getitem__")
return None
def __len__(self):
return sum(self.len_buffer)
def f_get_seq_len_list(self):
""" length = f_get_seq_len_list()
Total length of utterance in the merged dataset
"""
tmp = []
for sub_dataset in self.datasets:
tmp += sub_dataset.f_get_seq_len_list()
return tmp
def f_get_updated_seq_len_for_sampler_list(self):
""" Similar to f_get_seq_len_list
but it returns the updated data sequence length only for
length-based shuffling in sampler
"""
tmp = []
for sub_dataset in self.datasets:
tmp += sub_dataset.f_get_updated_seq_len_for_sampler_list()
return tmp
def f_update_seq_len_for_sampler_list(self, data_info):
"""
"""
for one_info in data_info:
data_idx = nii_seqinfo.parse_idx(one_info)
data_len = nii_seqinfo.parse_length(one_info)
for idx_u, idx_d, subset in \
zip(self.len_top, self.len_bot, self.datasets):
if data_idx < idx_u:
subset.f_update_seq_len_for_sampler_list(data_idx, data_len)
break
else:
pass
return
def f_manage_data(self, lst_data_idx, opt):
""" f_manage_data(self, lst_data_idx, opt)
"""
# manage the data in each subset
for idx_u, idx_d, subset in \
zip(self.len_top, self.len_bot, self.datasets):
# adjust the index for each sub dataset
tmp_data_idx = [x - idx_d for x in lst_data_idx \
if x >= idx_d and x < idx_u]
subset.f_manage_data(tmp_data_idx, opt)
# re-initialize len, len_top, and len_bot
self.__init_sub()
return None
def f_get_seq_idx_from_name(self, data_names):
""" list_of_idx = f_get_seq_idx_from_name(data_names)
"""
return [self.name_idx_map[x] for x in data_names]
class NII_MergeDataSetLoader():
""" DatasetLoader for loading multiple data corpora as a single one
Similar to NIIDataSetLoader, this merges the DataSet and DataLoader
into a single class.
"""
def __init__(self,
dataset_name, \
list_file_list, \
list_input_dirs, input_exts, input_dims, input_reso, \
input_norm, \
list_output_dirs, output_exts, output_dims, output_reso, \
output_norm, \
stats_path, \
data_format = nii_dconf.h_dtype_str, \
params = None, \
truncate_seq = None, \
min_seq_len = None,
save_mean_std = True, \
wav_samp_rate = None, \
flag_lang = 'EN', \
way_to_merge = 'concatenate',
global_arg = None,
dset_config = None,
input_augment_funcs = None,
output_augment_funcs = None,
inoutput_augment_func = None):
""" Signature is similar to default_io.NIIDataSetLoader.
file_list, input_dirs, and output_dirs are different.
One additional optional argument is way_to_merge.
Args
----
data_set_name: a string to name this dataset
this will be used to name the statistics files
such as the mean/std for this dataset
list_file_list: a list of file_name path
list_input_dirs: a list of lists of dirs for input features
input_exts: a list of input feature name extentions
input_dims: a list of input feature dimensions
input_reso: a list of input feature temporal resolution,
or None
input_norm: a list of bool, whether normalize input feature or not
list_output_dirs: a list of lists of dirs for output features
output_exts: a list of output feature name extentions
output_dims: a list of output feature dimensions
output_reso: a list of output feature temporal resolution,
or None
output_norm: a list of bool, whether normalize target feature or not
stats_path: path to the directory of statistics(mean/std)
data_format: method to load the data
'<f4' (default): load data as float32m little-endian
'htk': load data as htk format
params: parameter for torch.utils.data.DataLoader
truncate_seq: None or int,
truncate data sequence into smaller truncks
truncate_seq > 0 specifies the trunck length
min_seq_len: None (default) or int, minimum length of an utterance
utterance shorter than min_seq_len will be ignored
save_mean_std: bool, True (default): save mean and std
wav_samp_rate: None (default) or int, if input data has waveform,
please set sampling rate. It is used by _data_writer
flag_lang: str, 'EN' (default), if input data has text, text will
be converted into code indices. flag_lang indicates the
language for the text processer. It is used by _data_reader
wav_to_merge: string, 'concatenate' (default) or 'merge'
'concatenate': simply concatenate multiple corpora
'merge': create minibatch by merging data from each copora
global_arg: argument parser returned by arg_parse.f_args_parsed()
default None
dset_config: object, data set configuration, default None
input_augment_funcs: list of functions for input data transformation
default None
output_augment_funcs: list of output data transformation functions
default None
inoutput_augment_func: a single data augmentation function
default None
Methods
-------
get_loader(): return a torch.util.data.DataLoader
get_dataset(): return a torch.util.data.DataSet
"""
########
# check whether input_dirs and output_dirs are lists
if type(list_input_dirs[0]) is list and \
type(list_output_dirs[0]) is list and \
type(list_file_list) is list and \
len(list_input_dirs) == len(list_output_dirs) and \
len(list_input_dirs) == len(list_file_list):
pass
else:
mes = "NII_MergeDataSetLoader: input_dirs, output_dirs, "
mes += "and file_list should be list of lists. "
mes += "They should have equal length. But we have:"
mes += "{:s}\n{:s}\n{:s}".format(
str(list_input_dirs), str(list_output_dirs),
str(list_file_list))
nii_warn.f_die(mes)
if type(dataset_name) is list:
if len(dataset_name) != len(list_input_dirs):
mes = "dataset_name should have {:d} elements. ".format(
len(list_file_list))
mes += "But we have: {:s}".format(str(dataset_name))
nii_warn.f_die(mes)
elif len(list(set(dataset_name))) != len(list_input_dirs):
mes = "dataset_name has duplicated elements: {:s}".format(
str(dataset_name))
nii_warn.f_die(mes)
else:
tmp_dnames = dataset_name
else:
tmp_dnames = [dataset_name + '_sub_{:d}'.format(idx) \
for idx in np.arange(len(list_input_dirs))]
#######
# create individual datasets
lst_dset = []
cnt = 0
for sub_input_dirs, sub_output_dirs, sub_file_list, tmp_name in \
zip(list_input_dirs, list_output_dirs, list_file_list, tmp_dnames):
inaug = input_augment_funcs[cnt] if input_augment_funcs else None
ouaug = output_augment_funcs[cnt] if output_augment_funcs else None
lst_dset.append(
nii_default_dset.NIIDataSetLoader(
tmp_name, sub_file_list, \
sub_input_dirs, input_exts, input_dims, input_reso, \
input_norm, \
sub_output_dirs, output_exts, output_dims, output_reso, \
output_norm, \
stats_path, data_format, params, truncate_seq, min_seq_len,\
save_mean_std, wav_samp_rate, flag_lang, \
global_arg, dset_config, inaug, ouaug,\
inoutput_augment_func))
cnt += 1
# list of the datasets
self.m_datasets = lst_dset
#######
# merge multiple datasets (i.e., build the DataLoader)
self.way_to_merge = way_to_merge
if way_to_merge == 'concatenate':
# to create DataLoader, we need the pytorch.dataset
self.m_concate_set = ConcatDataset(
[x.get_dataset() for x in self.m_datasets])
# concatenate multiple datasets
if params is None:
tmp_params = nii_dconf.default_loader_conf
else:
tmp_params = params.copy()
# save parameters
self.m_params = tmp_params
# create data loader
self.m_loader = self.build_loader_concate_merge()
else:
self.m_concate_set = None
# sample mini-batches of equal size from each sub dataset
# use specific dataloader
self.m_loader = merge_loader(lst_dset)
self.m_params = lst_dset[0].get_loader_params()
return
def build_loader_concate_merge(self):
"""
build dataloader for a merged dataset
"""
# legacy implementation, no need to use
####
# Although members in l_dset have Dataloader, we need to
# create a dataloder for the concatenate dataset
###
tmp_params = self.m_params.copy()
# creatr sampler
if 'sampler' in tmp_params:
tmp_sampler = None
if tmp_params['sampler'] == nii_sampler_fn.g_str_sampler_bsbl:
if 'batch_size' in tmp_params \
and tmp_params['batch_size'] > 1:
# initialize the sampler
tmp_sampler = nii_sampler_fn.SamplerBlockShuffleByLen(
self.m_concate_set.f_get_seq_len_list(),
tmp_params['batch_size'])
# turn off automatic shuffle
tmp_params['shuffle'] = False
else:
nii_warn.f_print("{:s} off as batch-size is 1".format(
nii_sampler_fn.g_str_sampler_bsbl))
tmp_params['sampler'] = tmp_sampler
# collate function
if 'batch_size' in tmp_params and tmp_params['batch_size'] > 1:
# use customize_collate to handle data with unequal length
# we cannot use default collate_fn
collate_fn = nii_collate_fn.customize_collate
else:
collate_fn = None
# use default DataLoader
return torch.utils.data.DataLoader(
self.m_concate_set, collate_fn=collate_fn, **tmp_params)
def get_loader_params(self):
return self.m_params
def get_loader(self):
""" get_loader():
Return the dataLoader (torch.util.data.DataLoader)
"""
return self.m_loader
def get_dataset(self):
""" get_dataset():
Return the dataset (torch.util.data.Dataset)
"""
return self.m_datasets
def get_data_mean_std(self):
"""
"""
# temporary solution: just use the first one
return self.m_datasets[0].get_data_mean_std()
def print_info(self):
"""
"""
nii_warn.f_print_message("Merge datasets by: " + self.way_to_merge)
for dset in self.m_datasets:
dset.print_info()
return
def putitem(self, output_data, save_dir, filename_prefix, data_infor_str):
""" Decompose the output_data from network into
separate files
"""
# Since all datasets have similar configuration on feat dim,
# use anyone is OK
self.m_datasets[0].putitem(output_data, save_dir, filename_prefix,
data_infor_str)
def get_in_dim(self):
""" Return the dimension of input features
"""
# Since all datasets have similar configuration on feat dim,
# use anyone is OK
return self.m_datasets[0].get_in_dim()
def get_out_dim(self):
""" Return the dimension of output features
"""
# Since all datasets have similar configuration on feat dim,
# use anyone is OK
return self.m_datasets[0].get_out_dim()
def get_seq_num(self):
""" Return the number of sequences (after truncation)
"""
return sum([x.get_seq_num() for x in self.m_datasets])
def get_seq_info(self):
""" Return the full information of each data,
including name, index, segmentation information
"""
tmp = []
for dataset in self.m_datasets:
tmp += dataset.get_seq_info()
return tmp
def get_seq_name_list(self):
""" list = get_seq_name_list()
Return a list of data sequence name
"""
tmp = []
for dataset in self.m_datasets:
tmp += dataset.get_seq_name_list()
return tmp
def get_seq_idx_from_name(self, data_names):
""" idx = get_seq_idx_from_name(data_names)
Return a list of data idx corresponding to the data file names
"""
# re-build data loader
if self.way_to_merge == 'concatenate':
return self.m_concate_set.f_get_seq_idx_from_name(data_names)
else:
nii_warn.f_print("Not implemented get_seq_idx_from_name")
nii_warn.f_die("--way-to-merge-datasets concatenate")
return None
def update_seq_len_in_sampler_sub(self, data_info):
"""
"""
# assume data_info logs the new data length that can be used for
# sampler shuffle_by_length
if self.way_to_merge == 'concatenate':
self.m_concate_set.f_update_seq_len_for_sampler_list(data_info)
else:
print("Not implemented")
sys.exit(1)
return
def update_seq_len_in_sampler(self):
""" update_seq_len()
Update sequence length if sequence length has been changed
(for example, during silence trim process)
This is necessary when using shuffle_by_seq_length sampler
and the sequences were trimmed in data augmentation function.
"""
# call each subdataset and update the sequence length
for idx, _ in enumerate(self.m_datasets):
self.m_datasets[idx].update_seq_len_in_sampler()
# update loader of this database
if self.way_to_merge == 'concatenate':
if self.m_params['sampler'] == nii_sampler_fn.g_str_sampler_bsbl \
and hasattr(self.m_loader.sampler, 'update_seq_length'):
self.m_loader.sampler.update_seq_length(
self.m_concate_set.f_get_updated_seq_len_for_sampler_list())
return
def manage_data(self, lst_data_idx, opt):
""" manage_data(data_index_list, opt)
Manage data in the dataset. Current, we can either keep or delete
the specified list of data
Args:
lst_data_idx: list of data indices, lst_data_idx[n] is
the index of the sample in the merged dataset
opt: 'keep', keep only data in idx
'delete', delete data in idx
"""
# re-build data loader
if self.way_to_merge == 'concatenate':
self.m_concate_set.f_manage_data(lst_data_idx, opt)
if self.get_seq_num() < 1:
# DataLoader will raise Error when no data is DataSet
self.m_loader = None
else:
self.m_loader = self.build_loader_concate_merge()
else:
nii_warn.f_print("Active learning requires ")
nii_warn.f_die("--way-to-merge-datasets concatenate")
return
def add_dataset(self, new_data_wrapper):
"""add_dataset(new_data_wrapper)
Add an existing merged dataset to this dataset
Args:
new_data_wrapper: must be a NII_MergeDataSetLoader
"""
self.m_datasets = self.m_datasets + new_data_wrapper.m_datasets
if self.way_to_merge == 'concatenate':
self.m_concate_set = ConcatDataset(
[x.get_dataset() for x in self.m_datasets])
self.m_loader = self.build_loader_concate_merge()
else:
nii_warn.f_print("Active learning requires ")
nii_warn.f_die("--way-to-merge-datasets concatenate")
return
if __name__ == "__main__":
print("Definition of customized Pytorch dataset")
| 23,353 | 37.665563 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/io_tools.py | #!/usr/bin/env python
"""
io_tools
Functions to load data
"""
from __future__ import absolute_import
import os
import sys
import json
import pickle
import numpy as np
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
def f_read_raw_mat(filename, col, data_format='f4', end='l'):
"""data = f_read_raw_mat(filename, col, data_format='float', end='l')
Read the binary data from filename
Return data, which is a (N, col) array
input
-----
filename: str, path to the binary data on the file system
col: int, number of column assumed by the data matrix
format: str, please use the Python protocal to write format
default: 'f4', float32
end: str, little endian 'l' or big endian 'b'?
default: 'l'
output
------
data: np.array, shape (N, col), where N is the number of rows
decided by total_number_elements // col
"""
f = open(filename,'rb')
if end=='l':
data_format = '<'+data_format
elif end=='b':
data_format = '>'+data_format
else:
data_format = '='+data_format
datatype = np.dtype((data_format,(col,)))
data = np.fromfile(f,dtype=datatype)
f.close()
if data.ndim == 2 and data.shape[1] == 1:
return data[:,0]
else:
return data
def f_read_raw_mat_length(filename, data_format='f4'):
"""len = f_read_raw_mat_length(filename, data_format='f4')
Read length of data, i.e., number of elements in the data file.
If data is in shape (N, M), then len = N * M
input
-----
filename: str, path to the binary data on the file system
format: str, please use the Python protocal to write format
default: 'f4', float32
output
------
len: int, number of data elements in the data file
"""
f = open(filename,'rb')
tmp = f.seek(0, 2)
bytes_num = f.tell()
f.close()
if data_format == 'f4':
return int(bytes_num / 4)
else:
return bytes_num
def f_read_htk(filename, data_format='f4', end='l'):
"""data = read_htk(filename, data_format='f4', end='l')
Read HTK File and return the data as numpy.array
input
-----
filename: str, path to the binary HTK data on file system
data_format: str, format of the returned data
default: 'f4' float32
end: little endian 'l' or big endian 'b'?
default: 'l'
output
------
data: numpy.array
"""
if end=='l':
data_format = '<'+data_format
data_formatInt4 = '<i4'
data_formatInt2 = '<i2'
elif end=='b':
data_format = '>'+data_format
data_formatInt4 = '>i4'
data_formatInt2 = '>i2'
else:
data_format = '='+data_format
data_formatInt4 = '=i4'
data_formatInt2 = '=i2'
head_type = np.dtype([('nSample',data_formatInt4),
('Period',data_formatInt4),
('SampleSize',data_formatInt2),
('kind',data_formatInt2)])
f = open(filename,'rb')
head_info = np.fromfile(f,dtype=head_type,count=1)
"""if end=='l':
data_format = '<'+data_format
elif end=='b':
data_format = '>'+data_format
else:
data_format = '='+data_format
"""
if 'f' in data_format:
sample_size = int(head_info['SampleSize'][0]/4)
else:
print("Error in read_htk: input should be float32")
return False
datatype = np.dtype((data_format,(sample_size,)))
data = np.fromfile(f,dtype=datatype)
f.close()
return data
def f_read_htk_length(filename, data_format='f4', end='l'):
"""length = read_htk(filename, data_format='f4', end='l')
Read HTK File and return the number of data elements in the file
Read HTK File and return the data as numpy.array
input
-----
filename: str, path to the binary HTK data on file system
data_format: str, format of the returned data
default: 'f4' float32
end: little endian 'l' or big endian 'b'?
default: 'l'
output
------
length: int, number of data elements in the file
"""
if end=='l':
data_format = '<'+data_format
data_formatInt4 = '<i4'
data_formatInt2 = '<i2'
elif end=='b':
data_format = '>'+data_format
data_formatInt4 = '>i4'
data_formatInt2 = '>i2'
else:
data_format = '='+data_format
data_formatInt4 = '=i4'
data_formatInt2 = '=i2'
head_type = np.dtype([('nSample',data_formatInt4),
('Period',data_formatInt4),
('SampleSize',data_formatInt2),
('kind',data_formatInt2)])
f = open(filename,'rb')
head_info = np.fromfile(f,dtype=head_type,count=1)
f.close()
sample_size = int(head_info['SampleSize'][0]/4)
return sample_size
def f_write_raw_mat(data, filename, data_format='f4', end='l'):
"""flag = write_raw_mat(data, filename, data_format='f4', end='l')
Write data to file on the file system as binary data
input
-----
data: np.array, data to be saved
filename: str, path of the file to save the data
data_format: str, data_format for numpy
default: 'f4', float32
end: str little endian 'l' or big endian 'b'?
default: 'l'
output
------
flag: bool, whether the writing is done or not
"""
if not isinstance(data, np.ndarray):
print("Error write_raw_mat: input should be np.array")
return False
f = open(filename,'wb')
if len(data_format)>0:
if end=='l':
data_format = '<'+data_format
elif end=='b':
data_format = '>'+data_format
else:
data_format = '='+data_format
datatype = np.dtype(data_format)
temp_data = data.astype(datatype)
else:
temp_data = data
temp_data.tofile(f,'')
f.close()
return True
def f_append_raw_mat(data, filename, data_format='f4', end='l'):
"""flag = write_raw_mat(data, filename, data_format='f4', end='l')
Append data to an existing file on the file system as binary data
input
-----
data: np.array, data to be saved
filename: str, path of the file to save the data
data_format: str, data_format for numpy
default: 'f4', float32
end: str little endian 'l' or big endian 'b'?
default: 'l'
output
------
flag: bool, whether the writing is done or not
"""
if not isinstance(data, np.ndarray):
print("Error write_raw_mat: input shoul be np.array")
return False
f = open(filename,'ab')
if len(data_format)>0:
if end=='l':
data_format = '<'+data_format
elif end=='b':
data_format = '>'+data_format
else:
data_format = '='+data_format
datatype = np.dtype(data_format)
temp_data = data.astype(datatype)
else:
temp_data = data
temp_data.tofile(f,'')
f.close()
return True
def f_write_htk(data, targetfile,
sampPeriod=50000, sampKind=9, data_format='f4', end='l'):
"""
write_htk(data,targetfile,
sampPeriod=50000,sampKind=9,data_format='f4',end='l')
Write data as HTK-compatible format
input
-----
data: np.array, data to be saved
targetfile: str, path of the file to save the data
...
output
------
"""
if data.ndim==1:
nSamples, vDim = data.shape[0], 1
else:
nSamples, vDim = data.shape
if data_format=='f4':
sampSize = vDim * 4;
else:
sampSize = vDim * 8;
f = open(targetfile,'wb')
if len(data_format)>0:
if end=='l':
data_format1 = '<i4'
data_format2 = '<i2'
elif end=='b':
data_format1 = '>i4'
data_format2 = '>i2'
else:
data_format1 = '=i4'
data_format2 = '=i2'
temp_data = np.array([nSamples, sampPeriod],
dtype=np.dtype(data_format))
temp_data.tofile(f, '')
temp_data = np.array([sampSize, sampKind], dtype=np.dtype(data_format2))
temp_data.tofile(f, '')
if len(data_format)>0:
if end=='l':
data_format = '<'+data_format
elif end=='b':
data_format = '>'+data_format
else:
data_format = '='+data_format
datatype = np.dtype(data_format)
temp_data = data.astype(datatype)
else:
temp_data = data
temp_data.tofile(f, '')
f.close()
return True
def read_dic(file_path):
""" dic = read_dic(file_path)
Read a json file from file_path and return a dictionary
input
-----
file_path: string, path to the file
output
------
dic: a dictionary
"""
try:
data = json.load( open(file_path) )
except IOError:
print("Cannot find %s" % (file_path))
sys.exit(1)
except json.decoder.JSONDecodeError:
print("Cannot parse %s" % (file_path))
sys.exit(1)
return data
def write_dic(dic, file_path):
""" write_dic(dic, file_path)
Write a dictionary to file
input
-----
dic: dictionary to be dumped
file_path: file to store the dictionary
"""
try:
json.dump(dic, open(file_path, 'w'))
except IOError:
print("Cannot write to %s " % (file_path))
sys.exit(1)
def file_exist(file_path):
""" file_exit(file_path)
Whether file exists
"""
return os.path.isfile(file_path) or os.path.islink(file_path)
def pickle_dump(data, file_path):
""" pickle_dump(data, file_path)
Dump data into a pickle file
inputs:
data: python object, data to be dumped
file_path: str, path to save the pickle file
"""
try:
os.mkdir(os.path.dirname(file_path))
except OSError:
pass
with open(file_path, 'wb') as file_ptr:
pickle.dump(data, file_ptr)
return
def pickle_load(file_path):
""" data = pickle_load(file_path)
Load data from a pickle dump file
inputs:
file_path: str, path of the pickle file
output:
data: python object
"""
with open(file_path, 'rb') as file_ptr:
data = pickle.load(file_ptr)
return data
def wrapper_data_load_with_cache(file_path, method_data_load,
cache_dir='__cache',
use_cached_data=True, verbose=False):
"""wrapper_data_load_with_cache(file_path, method_data_load,
cache_dir='__cache',
use_cached_data=True, verbose=False):
Load data from file and save data as pickle file in cache.
input
-----
file_path: str, path of input file
method_data_load: python function, funtion to load the data
cache_dir: str, default __cache, the directory to save cached pickle file
use_cached_data: bool, default True, use cached data when available
verbose: bool, default False, print information on reading/writing
output
------
data: python object decided by method_data_load
This method is useful to load large text file. No need to parse text
everytime because the data will be saved as pickle file in cache after
the first time of execution
Example:
from core_scripts.data_io import io_tools
from core_scripts.other_tools import list_tools
data = io_tools.wrapper_data_load_with_cache('test_file',
list_tools.read_list_from_text)
"""
try:
os.mkdir(cache_dir)
except OSError:
pass
cache_file_path = '_'.join(file_path.split(os.path.sep))
cache_file_path = os.path.join(cache_dir, cache_file_path)
cache_file_path += '.pkl'
if use_cached_data and os.path.isfile(cache_file_path):
if verbose:
print("Load cached data {:s}".format(cache_file_path))
return pickle_load(cache_file_path)
else:
data = method_data_load(file_path)
pickle_dump(data, cache_file_path)
if verbose:
print("Load data {:s}".format(file_path))
print("Save cahced data {:s}".format(cache_file_path))
return data
if __name__ == "__main__":
print("Definition of tools for I/O operation")
| 12,660 | 27.775 | 79 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/default_data_io.py | #!/usr/bin/env python
"""
data_io
Interface to load data
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
from inspect import signature
import torch
import torch.utils.data
import core_scripts.other_tools.list_tools as nii_list_tools
import core_scripts.other_tools.display as nii_warn
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.data_io.io_tools as nii_io_tk
import core_scripts.data_io.wav_tools as nii_wav_tk
import core_scripts.data_io.text_process.text_io as nii_text_tk
import core_scripts.data_io.conf as nii_dconf
import core_scripts.data_io.seq_info as nii_seqinfo
import core_scripts.math_tools.stats as nii_stats
import core_scripts.data_io.customize_collate_fn as nii_collate_fn
import core_scripts.data_io.customize_sampler as nii_sampler_fn
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
###
## functions wrappers to read/write data for this data_io
###
def _data_reader(file_path, dim, flag_lang, g2p_tool):
""" A wrapper to read raw binary data, waveform, or text
"""
file_name, file_ext = os.path.splitext(file_path)
if file_ext == '.wav':
sr, data = nii_wav_tk.waveReadAsFloat(file_path)
if data.ndim > 1 and data.shape[-1] != dim:
nii_warn.f_print("Expect {:d} channel(s)".format(dim), 'error')
nii_warn.f_die("But {:s} has {:d} channel(s)".format(
file_path, data.shape[-1]))
elif file_ext == '.flac':
sr, data = nii_wav_tk.flacReadAsFloat(file_path)
if data.ndim > 1 and data.shape[-1] != dim:
nii_warn.f_print("Expect {:d} channel(s)".format(dim), 'error')
nii_warn.f_die("But {:s} has {:d} channel(s)".format(
file_path, data.shape[-1]))
elif file_ext == '.txt':
data = nii_text_tk.textloader(file_path, flag_lang, g2p_tool)
else:
data = nii_io_tk.f_read_raw_mat(file_path, dim)
return data
def _data_writer(data, file_path, sr = 16000):
""" A wrapper to write raw binary data or waveform
"""
file_name, file_ext = os.path.splitext(file_path)
if file_ext == '.wav':
nii_wav_tk.waveFloatToPCMFile(data, file_path, sr = sr)
elif file_ext == '.txt':
nii_warn.f_die("Cannot write to {:s}".format(file_path))
else:
nii_io_tk.f_write_raw_mat(data, file_path)
return
def _data_len_reader(file_path):
""" A wrapper to read length of data
"""
file_name, file_ext = os.path.splitext(file_path)
if file_ext == '.wav':
#sr, data = nii_wav_tk.waveReadAsFloat(file_path)
#length = data.shape[0]
length = nii_wav_tk.readWaveLength(file_path)
elif file_ext == '.flac':
sr, data = nii_wav_tk.flacReadAsFloat(file_path)
length = data.shape[0]
elif file_ext == '.txt':
# txt, no need to account length
# note that this is for tts task
length = 0
else:
length = nii_io_tk.f_read_raw_mat_length(file_path)
return length
###
# Definition of DataSet
###
class NIIDataSet(torch.utils.data.Dataset):
""" General class for NII speech dataset
For definition of customized Dataset, please refer to
https://pytorch.org/tutorials/beginner/data_loading_tutorial.html
"""
def __init__(self,
dataset_name, \
file_list, \
input_dirs, input_exts, input_dims, input_reso, \
input_norm, \
output_dirs, output_exts, output_dims, output_reso, \
output_norm, \
stats_path, \
data_format = nii_dconf.h_dtype_str, \
truncate_seq = None, \
min_seq_len = None, \
save_mean_std = True, \
wav_samp_rate = None, \
flag_lang = 'EN', \
global_arg = None, \
dset_config = None, \
input_augment_funcs = None, \
output_augment_funcs = None,
inoutput_augment_func = None):
"""
args
----
dataset_name: name of this data set
file_list: a list of file name strings (without extension)
or, path to the file that contains the file names
input_dirs: a list of dirs from which input feature is loaded
input_exts: a list of input feature name extentions
input_dims: a list of input feature dimensions
input_reso: a list of input feature temporal resolutions
input_norm: a list of bool, whether normalize input feature or not
output_dirs: a list of dirs from which output feature is loaded
output_exts: a list of output feature name extentions
output_dims: a list of output feature dimensions
output_reso: a list of output feature temporal resolutions
output_norm: a list of bool, whether normalize target feature or not
stat_path: path to the directory that saves mean/std,
utterance length
data_format: method to load the data
'<f4' (default): load data as float32m little-endian
'htk': load data as htk format
truncate_seq: None (default) or int, truncate sequence into truncks.
truncate_seq > 0 specifies the trunck length
min_seq_len: None (default) or int, minimum length of an utterance
utterance shorter than min_seq_len will be ignored
save_mean_std: bool, True (default): save mean and std
wav_samp_rate: None (default) or int, if input data has waveform,
please set sampling rate. It is used by _data_writer
flag_lang: str, 'EN' (default), if input data has text, the text will
be converted into code indices. flag_lang indicates the
language for the text processer. It is used by _data_reader
global_arg: argument parser returned by arg_parse.f_args_parsed()
default None
dset_config: object, dataset configuration, default None
input_augment_funcs: list of functions for input data transformation
default None
output_augment_funcs: list of output data transformation functions
default None
inoutput_augment_func: a single data augmentation function,
default None
"""
# initialization
self.m_set_name = dataset_name
self.m_file_list = file_list
self.m_input_dirs = input_dirs
self.m_input_exts = input_exts
self.m_input_dims = input_dims
self.m_output_dirs = output_dirs
self.m_output_exts = output_exts
self.m_output_dims = output_dims
if len(self.m_input_dirs) != len(self.m_input_exts) or \
len(self.m_input_dirs) != len(self.m_input_dims):
nii_warn.f_print("Input dirs, exts, dims, unequal length",
'error')
nii_warn.f_print(str(self.m_input_dirs), 'error')
nii_warn.f_print(str(self.m_input_exts), 'error')
nii_warn.f_print(str(self.m_input_dims), 'error')
nii_warn.f_die("Please check input dirs, exts, dims")
if len(self.m_output_dims) != len(self.m_output_exts) or \
(self.m_output_dirs and \
len(self.m_output_dirs) != len(self.m_output_exts)):
nii_warn.f_print("Output dirs, exts, dims, unequal length", \
'error')
nii_warn.f_die("Please check output dirs, exts, dims")
# fill in m_*_reso and m_*_norm
def _tmp_f(list2, default_value, length):
if list2 is None:
return [default_value for x in range(length)]
else:
return list2
self.m_input_reso = _tmp_f(input_reso, 1, len(input_dims))
self.m_input_norm = _tmp_f(input_norm, True, len(input_dims))
self.m_output_reso = _tmp_f(output_reso, 1, len(output_dims))
self.m_output_norm = _tmp_f(output_norm, True, len(output_dims))
if len(self.m_input_reso) != len(self.m_input_dims):
nii_warn.f_die("len(input_reso) != len(input_dims) in config")
if len(self.m_output_reso) != len(self.m_output_dims):
nii_warn.f_die("len(output_reso) != len(input_dims) in config")
if len(self.m_input_norm) != len(self.m_input_dims):
nii_warn.f_die("len(input_norm) != len(input_dims) in config")
if len(self.m_output_norm) != len(self.m_output_dims):
nii_warn.f_die("len(output_norm) != len(output_dims) in config")
if global_arg is not None:
self.m_ignore_length_invalid = global_arg.ignore_length_invalid_data
self.m_ignore_cached_finfo = global_arg.ignore_cached_file_infor
self.m_force_skip_scanning = global_arg.force_skip_datadir_scanning
else:
self.m_ignore_length_invalid = False
self.m_ignore_cached_finfo = False
self.m_force_skip_scanning = False
# check augmentation funcctions
if input_augment_funcs:
if len(input_augment_funcs) != len(self.m_input_dims):
nii_warn.f_die("len(input_augment_funcs) != len(input_dims)")
self.m_inaug_funcs = input_augment_funcs
else:
self.m_inaug_funcs = []
if output_augment_funcs:
if len(output_augment_funcs) != len(self.m_output_dims):
nii_warn.f_die("len(output_augment_funcs) != len(output_dims)")
self.m_ouaug_funcs = output_augment_funcs
else:
self.m_ouaug_funcs = []
if inoutput_augment_func:
self.m_inouaug_func = inoutput_augment_func
else:
self.m_inouaug_func = None
# dimensions
self.m_input_all_dim = sum(self.m_input_dims)
self.m_output_all_dim = sum(self.m_output_dims)
self.m_io_dim = self.m_input_all_dim + self.m_output_all_dim
self.m_truncate_seq = truncate_seq
self.m_min_seq_len = min_seq_len
self.m_save_ms = save_mean_std
# in case there is waveform data in input or output features
self.m_wav_sr = wav_samp_rate
# option to process waveform with simple VAD
if global_arg is not None:
self.m_opt_wav_handler = global_arg.opt_wav_silence_handler
else:
self.m_opt_wav_handler = 0
# in case there is text data in input or output features
self.m_flag_lang = flag_lang
self.m_g2p_tool = None
if hasattr(dset_config, 'text_process_options') and \
type(dset_config.text_process_options) is dict:
self.m_flag_lang = dset_config.text_process_options['flag_lang']
if 'g2p_tool' in dset_config.text_process_options:
self.m_g2p_tool = dset_config.text_process_options['g2p_tool']
# sanity check on resolution configuration
# currently, only input features can have different reso,
# and the m_input_reso must be the same for all input features
if any([x != self.m_input_reso[0] for x in self.m_input_reso]):
nii_warn.f_print("input_reso {:s}".format(str(self.m_input_reso)),
'error')
nii_warn.f_print("NIIDataSet not support", 'error', end='')
nii_warn.f_die(" different input_reso")
if any([x != self.m_output_reso[0] for x in self.m_output_reso]):
nii_warn.f_print("output_reso {:s}".format(str(self.m_output_reso)),
'error')
nii_warn.f_print("NIIDataSet not support", 'error', end='')
nii_warn.f_die(" different output_reso")
if np.any(np.array(self.m_output_reso) < 0):
nii_warn.f_print("NIIDataSet not support negative reso",
'error', end='')
nii_warn.f_die(" Output reso: {:s}".format(str(self.m_output_reso)))
if np.any(np.array(self.m_input_reso) < 0):
nii_warn.f_print("input_reso: {:s}".format(str(self.m_input_reso)))
nii_warn.f_print("Data IO for unaligned input and output pairs")
if truncate_seq is not None:
nii_warn.f_print("truncate is set to None", 'warning')
self.m_truncate_seq = None
self.m_min_seq_len = None
# no need to contrain output_reso = 1
#if any([x != 1 for x in self.m_output_reso]):
# nii_warn.f_print("NIIDataSet only supports", 'error', end='')
# nii_warn.f_die(" output_reso = [1, 1, ... 1]")
#self.m_single_reso = self.m_input_reso[0]
self.m_single_reso = np.max(self.m_input_reso + self.m_output_reso)
# To make sure that target waveform length is exactly equal
# to the up-sampled sequence length
# self.m_truncate_seq must be changed to be N * up_sample
if self.m_truncate_seq is not None:
# assume input resolution is the same
self.m_truncate_seq = self.f_adjust_len(self.m_truncate_seq)
# similarly on self.m_min_seq_len
if self.m_min_seq_len is not None:
# assume input resolution is the same
self.m_min_seq_len = self.f_adjust_len(self.m_min_seq_len)
# method to load/write raw data
if data_format == nii_dconf.h_dtype_str:
self.f_load_data = lambda x, y: _data_reader(
x, y, self.m_flag_lang, self.m_g2p_tool)
self.f_length_data = _data_len_reader
self.f_write_data = lambda x, y: _data_writer(x, y, self.m_wav_sr)
else:
nii_warn.f_print("Unsupported dtype {:s}".format(data_format))
nii_warn.f_die("Only supports {:s} ".format(nii_dconf.h_dtype_str))
# whether input file name in list contains part of the path
# this will be confirmed after reading the file list in the next step
self.flag_filename_with_path = False
# log down statiscs
# 1. length of each data utterance
# 2. mean / std of feature feature file
def get_name(stats_path, set_name, file_name):
tmp = set_name + '_' + file_name
return os.path.join(stats_path, tmp)
if global_arg is not None and global_arg.path_cache_file:
nii_warn.f_print("Cached files are re-directed to {:s}".format(
global_arg.path_cache_file))
tmp_stats_path = global_arg.path_cache_file
else:
tmp_stats_path = stats_path
self.m_ms_input_path = get_name(tmp_stats_path, self.m_set_name, \
nii_dconf.mean_std_i_file)
self.m_ms_output_path = get_name(tmp_stats_path, self.m_set_name, \
nii_dconf.mean_std_o_file)
self.m_data_len_path = get_name(tmp_stats_path, self.m_set_name, \
nii_dconf.data_len_file)
# load and check the validity of data list
self.f_check_file_list(self.m_data_len_path)
# initialize data length and mean /std, read prepared data stats
flag_cal_len = self.f_init_data_len_stats(self.m_data_len_path)
flag_cal_mean_std = self.f_init_mean_std(self.m_ms_input_path,
self.m_ms_output_path)
# if data information is not available, read it again from data
if flag_cal_len or flag_cal_mean_std:
self.f_calculate_stats(flag_cal_len, flag_cal_mean_std)
# if some additional flags are turned on
if hasattr(global_arg, "flag_reverse_data_loading_order") and \
global_arg.flag_reverse_data_loading_order:
self.m_flag_reverse_load_order = True
else:
self.m_flag_reverse_load_order = False
#
if hasattr(global_arg, "force_update_seq_length") and \
global_arg.force_update_seq_length:
self.m_force_update_seq_length = True
else:
self.m_force_update_seq_length = False
# check
if self.__len__() < 1:
nii_warn.f_print("Fail to load any data", "error")
nii_warn.f_print("Possible reasons: ", "error")
mes = "1. Old cache {:s}. Do rm it.".format(self.m_data_len_path)
mes += "\n2. input_dirs, input_exts, "
mes += "output_dirs, or output_exts incorrect."
mes += "\n3. all data are less than minimum_len in length. "
mes += "\nThe last case may happen if truncate_seq == mininum_len "
mes += "and truncate_seq % input_reso != 0. Then, the actual "
mes += "truncate_seq becomes truncate_seq//input_reso*input_reso "
mes += "and it will be shorter than minimum_len. Please change "
mes += "truncate_seq and minimum_len so that "
mes += "truncate_seq % input_reso == 0."
nii_warn.f_print(mes, "error")
nii_warn.f_die("Please check configuration file")
# done
return
def __len__(self):
""" __len__():
Return the number of samples in the list
"""
return len(self.m_seq_info)
def __getitem__(self, idx_input):
""" __getitem__(self, idx):
Return input, output
For test set data, output can be None
"""
# option to select the (N - i + 1)-th sample
if self.m_flag_reverse_load_order:
idx = len(self.m_seq_info) - idx_input - 1
else:
idx = idx_input
# get the sample information
try:
tmp_seq_info = self.m_seq_info[idx]
except IndexError:
nii_warn.f_die("Sample {:d} is not in seq_info".format(idx))
# file_name
file_name = tmp_seq_info.seq_tag()
# For input data
input_reso = self.m_input_reso[0]
seq_len = int(tmp_seq_info.seq_length() // input_reso)
s_idx = int(tmp_seq_info.seq_start_pos() // input_reso)
e_idx = s_idx + seq_len
# in case the input length not account using tmp_seq_info.seq_length
if seq_len < 0:
seq_len = 0
s_idx = 0
e_idx = 0
input_dim = self.m_input_all_dim
in_data = np.zeros([seq_len, input_dim], dtype=nii_dconf.h_dtype)
s_dim = 0
e_dim = 0
# loop over each feature type
for t_dir, t_ext, t_dim, t_res in \
zip(self.m_input_dirs, self.m_input_exts, \
self.m_input_dims, self.m_input_reso):
e_dim = s_dim + t_dim
# get file path and load data
file_path = nii_str_tk.f_realpath(t_dir, file_name, t_ext)
try:
tmp_d = self.f_load_data(file_path, t_dim)
except IOError:
nii_warn.f_die("Cannot find {:s}".format(file_path))
# write data
if t_res < 0:
# if this is for input data not aligned with output
# make sure that the input is in shape (seq_len, dim)
# f_load_data should return data in shape (seq_len, dim)
if tmp_d.ndim == 1:
in_data = np.expand_dims(tmp_d, axis=1)
elif tmp_d.ndim == 2:
in_data = tmp_d
else:
nii_warn.f_die("IO not support {:s}".format(file_path))
elif tmp_d.shape[0] == 1:
# input data has only one frame, duplicate
if tmp_d.ndim > 1:
in_data[:,s_dim:e_dim] = tmp_d[0,:]
elif t_dim == 1:
in_data[:,s_dim] = tmp_d
else:
nii_warn.f_die("Dimension wrong {:s}".format(file_path))
else:
# check
try:
# normal case
if tmp_d.ndim > 1:
# write multi-dimension data
in_data[:,s_dim:e_dim] = tmp_d[s_idx:e_idx,:]
elif t_dim == 1:
# write one-dimension data
in_data[:,s_dim] = tmp_d[s_idx:e_idx]
else:
nii_warn.f_die("Dimension wrong {:s}".format(file_path))
except ValueError:
if in_data.shape[0] != tmp_d[s_idx:e_idx].shape[0]:
mes = 'Expected length is {:d}.\n'.format(e_idx-s_idx)
mes += "Loaded length "+str(tmp_d[s_idx:e_idx].shape[0])
mes += '\nThis may be due to an incompatible cache *.dic.'
mes += '\nPlease check the length in *.dic'
mes += '\nPlease delete it if the cached length is wrong.'
nii_warn.f_print(mes)
nii_warn.f_die("fail to load {:s}".format(file_name))
else:
nii_warn.f_print("unknown data io error")
nii_warn.f_die("fail to load {:s}".format(file_name))
s_dim = e_dim
# load output data
if self.m_output_dirs:
output_reso = self.m_output_reso[0]
seq_len = int(tmp_seq_info.seq_length() // output_reso)
s_idx = int(tmp_seq_info.seq_start_pos() // output_reso)
e_idx = s_idx + seq_len
out_dim = self.m_output_all_dim
out_data = np.zeros([seq_len, out_dim], \
dtype = nii_dconf.h_dtype)
s_dim = 0
e_dim = 0
for t_dir, t_ext, t_dim in zip(self.m_output_dirs, \
self.m_output_exts, \
self.m_output_dims):
e_dim = s_dim + t_dim
# get file path and load data
file_path = nii_str_tk.f_realpath(t_dir, file_name, t_ext)
try:
tmp_d = self.f_load_data(file_path, t_dim)
except IOError:
nii_warn.f_die("Cannot find {:s}".format(file_path))
if tmp_d.shape[0] == 1:
if tmp_d.ndim > 1:
out_data[:,s_dim:e_dim] = tmp_d[0,:]
elif t_dim == 1:
out_data[:,s_dim]=tmp_d
else:
nii_warn.f_die("Dimension wrong {:s}".format(file_path))
else:
try:
if tmp_d.ndim > 1:
out_data[:,s_dim:e_dim] = tmp_d[s_idx:e_idx,:]
elif t_dim == 1:
out_data[:,s_dim]=tmp_d[s_idx:e_idx]
else:
nii_warn.f_die("Dim wrong {:s}".format(file_path))
except ValueError:
if out_data.shape[0] != tmp_d[s_idx:e_idx].shape[0]:
mes = 'Expected length is ' + str(e_idx-s_idx)
mes += ". Loaded "+str(tmp_d[s_idx:e_idx].shape[0])
mes += 'This may be due to an old cache *.dic.'
mes += '\nPlease check the length in *.dic\n'
mes += 'Please delete it if cached length is wrong.'
nii_warn.f_print(mes)
nii_warn.f_die("fail to load " +file_name)
else:
nii_warn.f_print("unknown data io error")
nii_warn.f_die("fail to load " +file_name)
s_dim = s_dim + t_dim
else:
out_data = []
# post processing if necessary
in_data, out_data, tmp_seq_info, idx = self.f_post_data_process(
in_data, out_data, tmp_seq_info, idx)
# return data
return in_data, out_data, tmp_seq_info.print_to_str(), idx
def f_post_data_process(self, in_data, out_data, seq_info, idx):
"""A wrapper to process the data after loading from files
"""
if self.m_opt_wav_handler == 0 \
and not self.m_inaug_funcs and not self.m_ouaug_funcs \
and not self.m_inouaug_func:
# no any post-processing process
return in_data, out_data, seq_info, idx
else:
# Do post processing one by one
# The order is:
# waveform silence handler -> input augementation functions ->
# output augmentation functions -> input&output augmentation
#
# Everthing can be handled in input&output augmentation.
# But to be compatible with old codes, we keep them all.
# It is recommended to use the unified input&output augmentation
###
# buffer infor
###
# create a new sequence information buffer for the input and output
tmp_seq_info = nii_seqinfo.SeqInfo(
seq_info.length, seq_info.seq_name, seq_info.seg_idx,
seq_info.start_pos, seq_info.info_id)
###
# waveform silence handler
###
# waveform handler, this is kept for compatibility
if self.m_opt_wav_handler > 0:
if len(self.m_input_exts) == 1 \
and self.m_input_exts[0][-3:] == 'wav':
if self.m_opt_wav_handler == 1:
tmp_flag_output = self.m_opt_wav_handler
tmp_only_twoends = False
elif self.m_opt_wav_handler == 2:
tmp_flag_output = self.m_opt_wav_handler
tmp_only_twoends = False
elif self.m_opt_wav_handler == 3:
tmp_flag_output = 1
tmp_only_twoends = True
else:
print("Unknown option for wav handler {:d}".format(
self.m_opt_wav_handler))
sys.exit(1)
in_data_n = nii_wav_tk.silence_handler_wrapper(
in_data, self.m_wav_sr,
flag_output = tmp_flag_output,
flag_only_startend_sil = tmp_only_twoends)
# this is temporary setting, if in_data.shape[0]
# corresponds to waveform length, update it
if tmp_seq_info.length == in_data.shape[0]:
tmp_seq_info.length = in_data_n.shape[0]
if self.m_force_update_seq_length:
seq_info.update_len_for_sampler(in_data_n.shape[0])
else:
in_data_n = in_data
if len(self.m_output_exts) == 1 \
and self.m_output_exts[0][-3:] == 'wav':
out_data_n = nii_wav_tk.silence_handler_wrapper(
out_data, self.m_wav_sr,
flag_output = self.m_opt_wav_handler,
flag_only_startend_sil = (self.m_opt_wav_handler==3))
# this is temporary setting, use length if it is compatible
if tmp_seq_info.length == out_data.shape[0]:
tmp_seq_info.length = out_data_n.shape[0]
if self.m_force_update_seq_length:
seq_info.update_len_for_sampler(out_data_n.shape[0])
else:
out_data_n = out_data
else:
in_data_n = in_data
out_data_n = out_data
###
# augmentation functions for input data
###
if self.m_inaug_funcs:
if len(self.m_input_exts) == 1:
# only a single input feature,
sig = signature(self.m_inaug_funcs[0])
if len(sig.parameters) == 1:
in_data_n = self.m_inaug_funcs[0](in_data_n)
elif len(sig.parameters) == 2:
in_data_n = self.m_inaug_funcs[0](in_data_n, seq_info)
else:
in_data_n = self.m_inaug_funcs[0](in_data_n)
# more rules should be applied to handle the data length
# here, simply set length
if type(in_data_n) == np.ndarray:
if tmp_seq_info.length > in_data_n.shape[0]:
tmp_seq_info.length = in_data_n.shape[0]
elif type(in_data_n) == dict:
if 'length' in in_data_n:
tmp_seq_info.length = in_data_n['length']
if 'data' in in_data_n:
in_data_n = in_data_n['data']
else:
print("Input data aug method does not return data")
sys.exit(1)
if self.m_force_update_seq_length:
# Update the data length so that correct data length
# can be used for --sampler block_shuffle_by_length
#
#tmp_len = seq_info.length
seq_info.update_len_for_sampler(tmp_seq_info.length)
#print("{:s} {:s} {:d} -> {:d}".format(
# seq_info.seq_name, seq_info.print_to_str(),
# tmp_len, seq_info.valid_len),
# flush=True)
else:
# multiple input features,
# must check whether func changes the feature length
# only fun that keeps the length will be applied
s_dim = 0
for func, dim in zip(self.m_inaug_funcs, self.m_input_dims):
e_dim = s_dim + dim
tmp_data = func(in_data_n[:, s_dim:e_dim])
if tmp_data.shape[0] == in_data_n.shape[0]:
in_data_n[:, s_dim:e_dim] = tmp_data
s_dim = s_dim + dim
###
# augmentation functions for output data
###
if self.m_ouaug_funcs:
if len(self.m_output_exts) == 1:
# only a single output feature type
sig = signature(self.m_ouaug_funcs[0])
if len(sig.parameters) == 1:
out_data_n = self.m_ouaug_funcs[0](out_data_n)
elif len(sig.parameters) == 2:
out_data_n = self.m_ouaug_funcs[0](out_data_n, seq_info)
else:
out_data_n = self.m_ouaug_funcs[0](out_data_n)
# more rules should be applied to handle the data length
# here, simply set length
#if tmp_seq_info.length > out_data_n.shape[0]:
# tmp_seq_info.length = out_data_n.shape[0]
else:
# multiple output features,
# must check whether func changes the feature length
# only fun that keeps the length will be applied
s_dim = 0
for func, dim in zip(self.m_ouaug_funcs,self.m_output_dims):
e_dim = s_dim + dim
tmp_data = func(out_data_n[:,s_dim:e_dim])
if tmp_data.shape[0] == out_data_n.shape[0]:
out_data_n[:, s_dim:e_dim] = tmp_data
s_dim = s_dim + dim
###
# a unified augmentation function for input and output
###
if self.m_inouaug_func:
# update input output features
in_data_n, out_data_n, tmp_len = self.m_inouaug_func(
in_data_n, out_data_n)
# update sequence length
tmp_seq_info.length = tmp_len
if self.m_force_update_seq_length:
seq_info.update_len_for_sampler(tmp_seq_info.length)
return in_data_n, out_data_n, tmp_seq_info, idx
def f_get_num_seq(self):
""" __len__():
Return the number of samples in the list
"""
return len(self.m_seq_info)
def f_get_seq_len_list(self):
""" Return length of each sequence as list
"""
return [x.seq_length() for x in self.m_seq_info]
def f_get_updated_seq_len_for_sampler_list(self):
""" Similar to f_get_seq_len_list
but it returns the updated data sequence length only for
length-based shuffling in sampler
"""
return [x.seq_len_for_sampler() for x in self.m_seq_info]
def f_update_seq_len_for_sampler_list(self, data_idx, data_len):
try:
self.m_seq_info[data_idx].update_len_for_sampler(data_len)
except IndexError:
nii_warn.f_die("Fail to index data {:d}".format(data_idx))
return
def f_get_mean_std_tuple(self):
return (self.m_input_mean, self.m_input_std,
self.m_output_mean, self.m_output_std)
def f_filename_has_folderpath(self):
""" Return True if file name in self.m_file_list contains '/',
Which indicates that the file name is path/filename
"""
return any([x.count(os.path.sep)>0 for x in self.m_file_list])
def f_check_file_list(self, data_len_buf_path):
""" f_check_file_list(data_len_buf_path):
Check the file list after initialization
Make sure that the file in file_list appears in every
input/output feature directory.
If not, get a file_list in which every file is avaiable
in every input/output directory
input
-----
data_len_buf_path: str, path to the data length buffer
"""
if self.m_file_list is None:
# get a initial file list if self.m_file_list is None
#
# if file list is not provided, we only search the directory
# without recursing sub directories
self.m_file_list = nii_list_tools.listdir_with_ext(
self.m_input_dirs[0], self.m_input_exts[0])
elif not isinstance(self.m_file_list, list):
# if m_file_list is a string
# load file list
if isinstance(self.m_file_list, str) and \
os.path.isfile(self.m_file_list):
# read the list if m_file_list is a str
self.m_file_list = nii_list_tools.read_list_from_text(
self.m_file_list)
else:
nii_warn.f_print("Cannot read {:s}".format(self.m_file_list))
nii_warn.f_print("Read file list from directories")
self.m_file_list = nii_list_tools.listdir_with_ext(
self.m_input_dirs[0], self.m_input_exts[0])
else:
# self.m_file_list is a list
pass
if type(self.m_file_list) is list and len(self.m_file_list) < 1:
mes = "either input data list is wrong"
mes += ", or {:s} is empty".format(self.m_input_dirs[0])
mes += "\nPlease check the folder and data list"
nii_warn.f_die(mes)
# decide whether the file name in self.m_file_list contains
# sub folders
flag_recur = self.f_filename_has_folderpath()
self.flag_filename_with_path = flag_recur
# if the stats cache will be loaded, let's skip the checking process
if os.path.isfile(data_len_buf_path) and not self.m_ignore_cached_finfo:
nii_warn.f_print("Skip scanning directories")
return
# check the list of files exist in all input/output directories
if not self.m_force_skip_scanning:
for tmp_d, tmp_e in zip(self.m_input_dirs, self.m_input_exts):
# read a file list from the input directory
tmp_list = nii_list_tools.listdir_with_ext(
tmp_d, tmp_e, flag_recur)
# get the common set of the existing files and those in list
tmp_new_list = nii_list_tools.common_members(
tmp_list, self.m_file_list)
if len(tmp_new_list) < 1:
nii_warn.f_print("Possible error when scanning:", 'error')
nii_warn.f_print(" {:s} for {:s}".format(tmp_d, tmp_e), 'error')
nii_warn.f_print('Some file names to be scanned:', 'error')
nii_warn.f_print(' ' + ' '.join(self.m_file_list[0:10]),'error')
if self.m_file_list[0].endswith(tmp_e):
nii_warn.f_print('Names should not have {:s}'.format(tmp_e))
if os.path.isfile(self.m_file_list[0]):
mes = "The above name seems not to be the data name. "
mes += "It seems to be a file path. "
mes += "\nPlease check test_list, trn_list, val_list."
nii_warn.f_print(mes, 'error')
self.m_file_list = tmp_new_list
break
else:
self.m_file_list = tmp_new_list
if len(self.m_file_list) < 1:
nii_warn.f_print("\nNo input features found after scanning",'error')
nii_warn.f_print("Please check %s" \
% (str(self.m_input_dirs)), 'error')
nii_warn.f_print("They should contain all files in file list",
'error')
nii_warn.f_print("Please also check filename extentions %s" \
% (str(self.m_input_exts)), 'error')
nii_warn.f_print("They should be correctly specified", 'error')
nii_warn.f_die("Failed to read input features")
# check output files if necessary
if self.m_output_dirs and not self.m_force_skip_scanning:
for tmp_d, tmp_e in zip(self.m_output_dirs, \
self.m_output_exts):
tmp_list = nii_list_tools.listdir_with_ext(tmp_d, tmp_e,
flag_recur)
self.m_file_list = nii_list_tools.common_members(
tmp_list, self.m_file_list)
if len(self.m_file_list) < 1:
nii_warn.f_print("\nNo output data found", 'error')
nii_warn.f_print("Please check %s" \
% (str(self.m_output_dirs)), 'error')
nii_warn.f_print("They should contain all files in file list",
'error')
nii_warn.f_print("Please also check filename extentions %s" \
% (str(self.m_output_exts)), 'error')
nii_warn.f_print("They should be correctly specified", 'error')
nii_warn.f_die("Failed to read output features")
else:
#nii_warn.f_print("Not loading output features")
pass
# done
return
def f_valid_len(self, t_1, t_2, min_length):
""" f_valid_time_steps(time_step1, time_step2, min_length)
When either t_1 > min_length or t_2 > min_length, check whether
time_step1 and time_step2 are too different
"""
if max(t_1, t_2) > min_length:
if (np.abs(t_1 - t_2) * 1.0 / t_1) > 0.1:
return False
return True
def f_check_specific_data(self, file_name):
""" check the data length of a specific file
"""
tmp_dirs = self.m_input_dirs.copy()
tmp_exts = self.m_input_exts.copy()
tmp_dims = self.m_input_dims.copy()
tmp_reso = self.m_input_reso.copy()
tmp_dirs.extend(self.m_output_dirs)
tmp_exts.extend(self.m_output_exts)
tmp_dims.extend(self.m_output_dims)
tmp_reso.extend(self.m_output_reso)
# loop over each input/output feature type
for t_dir, t_ext, t_dim, t_res in \
zip(tmp_dirs, tmp_exts, tmp_dims, tmp_reso):
file_path = nii_str_tk.f_realpath(t_dir, file_name, t_ext)
if not nii_io_tk.file_exist(file_path):
nii_warn.f_die("%s not found" % (file_path))
else:
t_len = self.f_length_data(file_path) // t_dim
print("%s, length %d, dim %d, reso: %d" % \
(file_path, t_len, t_dim, t_res))
return
def f_log_data_len(self, file_name, t_len, t_reso):
""" f_log_data_len(file_name, t_len, t_reso):
Log down the length of the data file.
When comparing the different input/output features for the same
file_name, only keep the shortest length
"""
# We need to exclude features that should not be considered when
# calculating the sequence length
# 1. sentence-level vector (t_len = 1)
# 2. unaligned feature (text in text-to-speech) (t_reso < 0)
valid_flag = t_len > 1 and t_reso > 0
if valid_flag:
# the length for the sequence with the fast tempoeral rate
# For example, acoustic-feature -> waveform 16kHz,
# if acoustic-feature is one frame per 5ms,
# tmp_len = acoustic feature frame length * (5 * 16)
# where t_reso = 5*16 is the up-sampling rate of acoustic feature
tmp_len = t_len * t_reso
# save length when have not read the file
if file_name not in self.m_data_length:
self.m_data_length[file_name] = tmp_len
# check length
if t_len == 1:
# cannot come here, keep this line as history
# if this is an utterance-level feature, it has only 1 frame
pass
elif self.f_valid_len(self.m_data_length[file_name], tmp_len, \
nii_dconf.data_seq_min_length):
# if the difference in length is small
if self.m_data_length[file_name] > tmp_len:
self.m_data_length[file_name] = tmp_len
else:
nii_warn.f_print("Sequence length mismatch:", 'error')
self.f_check_specific_data(file_name)
nii_warn.f_print("Please check the above features", 'error')
if self.m_ignore_length_invalid:
nii_warn.f_print("ignore-length-invalid-data is on")
nii_warn.f_print("ignore {:s}".format(file_name))
return False
else:
nii_warn.f_print("Or remove them from data list", 'error')
nii_warn.f_print("Or --ignore-length-invalid-data",'error')
nii_warn.f_die("Possible invalid data %s" % (file_name))
# adjust the length so that, when reso is used,
# the sequence length will be N * reso
tmp = self.m_data_length[file_name]
self.m_data_length[file_name] = self.f_adjust_len(tmp)
else:
# do nothing for unaligned input or sentence-level input
pass
return True
def f_adjust_len(self, length):
""" When input data will be up-sampled by self.m_single_reso,
Make sure that the sequence length at the up-sampled level is
= N * self.m_single_reso
For data without up-sampling m_single_reso = 1
"""
return length // self.m_single_reso * self.m_single_reso
def f_precheck_data_length(self):
""" For unaligned input and output, there is no way to know the
target sequence length before hand during inference stage
self.m_data_length will be empty
"""
if not self.m_data_length and not self.m_output_dirs and \
all([x < 0 for x in self.m_input_reso]):
# inference stage, when only input is given
# manually create a fake data length for each utterance
for file_name in self.m_file_list:
self.m_data_length[file_name] = 0
return
def f_log_seq_info(self):
""" After m_data_length has been created, create seq_info
"""
for file_name in self.m_file_list:
# if file_name is not logged, ignore this file
if file_name not in self.m_data_length:
nii_warn.f_eprint("Exclude %s from dataset" % (file_name))
continue
# if not truncate, save the seq_info directly
# otherwise, save truncate_seq info
length_remain = self.m_data_length[file_name]
start_pos = 0
seg_idx = 0
if self.m_truncate_seq is not None:
while(length_remain > 0):
info_idx = len(self.m_seq_info)
seg_length = min(self.m_truncate_seq, length_remain)
seq_info = nii_seqinfo.SeqInfo(seg_length,
file_name, seg_idx,
start_pos, info_idx)
if self.m_min_seq_len is None or \
seg_length >= self.m_min_seq_len:
self.m_seq_info.append(seq_info)
seg_idx += 1
start_pos += seg_length
length_remain -= seg_length
else:
info_idx = len(self.m_seq_info)
seq_info = nii_seqinfo.SeqInfo(length_remain,
file_name, seg_idx,
start_pos, info_idx)
if self.m_min_seq_len is None or \
length_remain >= self.m_min_seq_len:
self.m_seq_info.append(seq_info)
# get the total length
self.m_data_total_length = self.f_sum_data_length()
return
def f_init_mean_std(self, ms_input_path, ms_output_path):
""" f_init_mean_std
Initialzie mean and std vectors for input and output
"""
self.m_input_mean = np.zeros([self.m_input_all_dim])
self.m_input_std = np.ones([self.m_input_all_dim])
self.m_output_mean = np.zeros([self.m_output_all_dim])
self.m_output_std = np.ones([self.m_output_all_dim])
flag = True
if not self.m_save_ms:
# assume mean/std will be loaded from the network
# for example, for validation and test sets
flag = False
if not any(self.m_input_norm + self.m_output_norm):
# none of the input / output features needs norm
flag = False
if os.path.isfile(ms_input_path) and \
os.path.isfile(ms_output_path):
# load mean and std if exists
ms_input = self.f_load_data(ms_input_path, 1)
ms_output = self.f_load_data(ms_output_path, 1)
if ms_input.shape[0] != (self.m_input_all_dim * 2) or \
ms_output.shape[0] != (self.m_output_all_dim * 2):
if ms_input.shape[0] != (self.m_input_all_dim * 2):
nii_warn.f_print("%s incompatible" % (ms_input_path),
'warning')
if ms_output.shape[0] != (self.m_output_all_dim * 2):
nii_warn.f_print("%s incompatible" % (ms_output_path),
'warning')
nii_warn.f_print("mean/std will be recomputed", 'warning')
else:
self.m_input_mean = ms_input[0:self.m_input_all_dim]
self.m_input_std = ms_input[self.m_input_all_dim:]
self.m_output_mean = ms_output[0:self.m_output_all_dim]
self.m_output_std = ms_output[self.m_output_all_dim:]
nii_warn.f_print("Load mean/std from %s and %s" % \
(ms_input_path, ms_output_path))
flag = False
return flag
def f_sum_data_length(self):
"""
"""
return sum([x.seq_length() for x in self.m_seq_info])
def f_init_data_len_stats(self, data_path):
"""
flag = f_init_data_len_stats(self, data_path)
Check whether data length has been stored in data_pat.
If yes, load data_path and return False
Else, return True
"""
self.m_seq_info = []
self.m_data_length = {}
self.m_data_total_length = 0
flag = True
if os.path.isfile(data_path) and not self.m_ignore_cached_finfo:
# load data length from pre-stored *.dic
dic_seq_infos = nii_io_tk.read_dic(self.m_data_len_path)
for dic_seq_info in dic_seq_infos:
seq_info = nii_seqinfo.SeqInfo()
seq_info.load_from_dic(dic_seq_info)
self.m_seq_info.append(seq_info)
seq_tag = seq_info.seq_tag()
if seq_tag not in self.m_data_length:
self.m_data_length[seq_tag] = seq_info.seq_length()
else:
self.m_data_length[seq_tag] += seq_info.seq_length()
self.m_data_total_length = self.f_sum_data_length()
# check whether *.dic contains files in filelist
# note: one file is not found in self.m_data_length if it
# is shorter than the truncate_seq
if nii_list_tools.list_identical(self.m_file_list,\
self.m_data_length.keys()):
nii_warn.f_print("Read sequence info: %s" % (data_path))
flag = False
elif nii_list_tools.list_b_in_list_a(self.m_file_list,
self.m_data_length.keys()):
nii_warn.f_print("Read sequence info: %s" % (data_path))
nii_warn.f_print(
"However %d samples are ignoed" % \
(len(self.m_file_list)-len(self.m_data_length)))
tmp = nii_list_tools.members_in_a_not_in_b(
self.m_file_list, self.m_data_length.keys())
for tmp_name in tmp:
nii_warn.f_eprint("Exclude %s from dataset" % (tmp_name))
flag = False
else:
nii_warn.f_print("Incompatible cache: %s" % (data_path))
tmp = nii_list_tools.members_in_a_not_in_b(
self.m_data_length.keys(), self.m_file_list)
nii_warn.f_print("Possibly invalid data (a few examples):")
for tmp_name in tmp[:10]:
nii_warn.f_print(tmp_name)
nii_warn.f_print("...\nYou may carefully check these data.")
nii_warn.f_print("\nThey may not be in the provided data list.")
nii_warn.f_print("Re-read data statistics")
self.m_seq_info = []
self.m_data_length = {}
self.m_data_total_length = 0
# check wheteher truncating length has been changed
if self.m_truncate_seq is not None and flag is False:
tmp_max_len = max([x.seq_length() for x in self.m_seq_info])
if tmp_max_len != self.m_truncate_seq:
mes = "WARNING: truncate_seq conflicts with cached infor. "
mes += "Please delete cache files *.dic if you want to"
mes += " use the new truncate_seq"
nii_warn.f_print(mes, "warning")
return flag
def f_save_data_len(self, data_len_path):
"""
"""
if not self.m_ignore_cached_finfo:
nii_io_tk.write_dic([x.print_to_dic() for x in self.m_seq_info], \
data_len_path)
return
def f_save_mean_std(self, ms_input_path, ms_output_path):
"""
"""
# save mean and std
ms_input = np.zeros([self.m_input_all_dim * 2])
ms_input[0:self.m_input_all_dim] = self.m_input_mean
ms_input[self.m_input_all_dim :] = self.m_input_std
self.f_write_data(ms_input, ms_input_path)
ms_output = np.zeros([self.m_output_all_dim * 2])
ms_output[0:self.m_output_all_dim] = self.m_output_mean
ms_output[self.m_output_all_dim :] = self.m_output_std
self.f_write_data(ms_output, ms_output_path)
return
def f_print_info(self):
"""
"""
mes = "Dataset {}:".format(self.m_set_name)
mes += "\n Time steps: {:d} ".format(self.m_data_total_length)
if self.m_truncate_seq is not None:
mes += "\n Truncate length: {:d}".format(self.m_truncate_seq)
mes += "\n Data sequence num: {:d}".format(len(self.m_seq_info))
tmp_min_len = min([x.seq_length() for x in self.m_seq_info])
tmp_max_len = max([x.seq_length() for x in self.m_seq_info])
mes += "\n Maximum sequence length: {:d}".format(tmp_max_len)
mes += "\n Minimum sequence length: {:d}".format(tmp_min_len)
if self.m_min_seq_len is not None:
mes += "\n Shorter sequences are ignored"
mes += "\n Inputs\n Dirs:"
for subdir in self.m_input_dirs:
mes += "\n {:s}".format(subdir)
mes += "\n Exts:{:s}".format(str(self.m_input_exts))
mes += "\n Dims:{:s}".format(str(self.m_input_dims))
mes += "\n Reso:{:s}".format(str(self.m_input_reso))
mes += "\n Norm:{:s}".format(str(self.m_input_norm))
mes += "\n Outputs\n Dirs:"
for subdir in self.m_output_dirs:
mes += "\n {:s}".format(subdir)
mes += "\n Exts:{:s}".format(str(self.m_output_exts))
mes += "\n Dims:{:s}".format(str(self.m_output_dims))
mes += "\n Reso:{:s}".format(str(self.m_output_reso))
mes += "\n Norm:{:s}".format(str(self.m_output_norm))
if self.m_opt_wav_handler > 0:
# wav handler
if len(self.m_input_exts) == 1 \
and self.m_input_exts[0][-3:] == 'wav':
mes += "\n Waveform silence handler will be used on input"
else:
mes += "\n Waveform silence handler NOT used on input"
if len(self.m_input_exts) > 1:
mes += "\t because multiple input features are used"
if len(self.m_output_exts) == 1 \
and self.m_output_exts[0][-3:] == 'wav':
mes += "\n Waveform silence handler will be used on output"
else:
mes += "\n Waveform silence handler NOT used on output"
if len(self.m_output_exts) > 1:
mes += "\t because multiple output features are used"
if self.m_inaug_funcs:
mes += "\n Use input feature transformation functions"
if len(self.m_input_exts) > 1:
mes += "\n Functions that change data length are ignored"
mes += "\n If it is intend to change data length, "
mes += "\n please use inoutput_augment_func"
if self.m_ouaug_funcs:
mes += "\n Use output feature transformation functions"
if len(self.m_output_exts) > 1:
mes += "\n Functions that change data length are ignored"
mes += "\n If it is intend to change data length, "
mes += "\n please use inoutput_augment_func"
if self.m_inouaug_func:
mes += "\n Use a unified function to alter input and output data"
if self.m_flag_reverse_load_order:
mes += "\n Reverse the data loading order from dataset "
nii_warn.f_print_message(mes)
return
def f_calculate_stats(self, flag_cal_data_len, flag_cal_mean_std):
""" f_calculate_stats
Log down the number of time steps for each file
Calculate the mean/std
"""
# check
#if not self.m_output_dirs:
# nii_warn.f_print("Calculating mean/std", 'error')
# nii_warn.f_die("But output_dirs is not provided")
# prepare the directory, extension, and dimensions
tmp_dirs = self.m_input_dirs.copy()
tmp_exts = self.m_input_exts.copy()
tmp_dims = self.m_input_dims.copy()
tmp_reso = self.m_input_reso.copy()
tmp_norm = self.m_input_norm.copy()
tmp_dirs.extend(self.m_output_dirs)
tmp_exts.extend(self.m_output_exts)
tmp_dims.extend(self.m_output_dims)
tmp_reso.extend(self.m_output_reso)
tmp_norm.extend(self.m_output_norm)
# starting dimension of one type of feature
s_dim = 0
# ending dimension of one type of feature
e_dim = 0
# print information
load_cnt = 0
total_cnt = len(tmp_dirs) * len(self.m_file_list)
# print progress
nii_warn.f_print("Get data statistis (may be slow due to data I/O)")
bar_len = 50
loading_marker = total_cnt // bar_len + 1
nii_warn.f_print("".join(['-' for x in range(bar_len-2)])+">|", 'plain')
# list of invalid data
invalid_data_lst = []
# loop over each input/output feature type
for t_dir, t_ext, t_dim, t_reso, t_norm in \
zip(tmp_dirs, tmp_exts, tmp_dims, tmp_reso, tmp_norm):
s_dim = e_dim
e_dim = s_dim + t_dim
t_cnt = 0
mean_i, var_i = np.zeros([t_dim]), np.zeros([t_dim])
# loop over all the data
for file_name in self.m_file_list:
load_cnt += 1
if load_cnt % loading_marker == 0:
nii_warn.f_print('>', end='', flush=True, opt='')
# get file path
file_path = nii_str_tk.f_realpath(t_dir, file_name, t_ext)
if not nii_io_tk.file_exist(file_path):
nii_warn.f_die("%s not found" % (file_path))
# read the length of the data
if flag_cal_data_len:
t_len = self.f_length_data(file_path) // t_dim
if not self.f_log_data_len(file_name, t_len, t_reso):
# this data is not valid, ignore it
# but it is OK to use it to compute mean/std
invalid_data_lst.append(file_name)
# accumulate the mean/std recursively
if flag_cal_mean_std:
t_data = self.f_load_data(file_path, t_dim)
# if the is F0 data, only consider voiced data
if t_ext in nii_dconf.f0_unvoiced_dic:
unvoiced_value = nii_dconf.f0_unvoiced_dic[t_ext]
t_data = t_data[t_data > unvoiced_value]
# mean_i, var_i, t_cnt will be updated using online
# accumulation method
mean_i, var_i, t_cnt = nii_stats.f_online_mean_std(
t_data, mean_i, var_i, t_cnt)
# save mean and std for one feature type
if flag_cal_mean_std:
# if not normalize this dimension, set mean=0, std=1
if not t_norm:
mean_i[:] = 0
var_i[:] = 1
if s_dim < self.m_input_all_dim:
self.m_input_mean[s_dim:e_dim] = mean_i
std_i = nii_stats.f_var2std(var_i)
self.m_input_std[s_dim:e_dim] = std_i
else:
tmp_s = s_dim - self.m_input_all_dim
tmp_e = e_dim - self.m_input_all_dim
self.m_output_mean[tmp_s:tmp_e] = mean_i
std_i = nii_stats.f_var2std(var_i)
self.m_output_std[tmp_s:tmp_e] = std_i
if flag_cal_data_len:
# remove invalid data (remove duplicated entries first)
invalid_data_lst = list(set(invalid_data_lst))
for tmp_file_name in invalid_data_lst:
self.m_data_length.pop(tmp_file_name)
#
self.f_precheck_data_length()
# create seq_info
self.f_log_seq_info()
# save len information
self.f_save_data_len(self.m_data_len_path)
if flag_cal_mean_std:
self.f_save_mean_std(self.m_ms_input_path,
self.m_ms_output_path)
nii_warn.f_print('')
# done
return
def f_putitem(self, output_data, save_dir, filename_prefix, data_infor_str):
"""
"""
# Change the dimension to (length, dim)
if output_data.ndim == 3 and output_data.shape[0] == 1:
# When input data is (batchsize=1, length, dim)
output_data = output_data[0]
elif output_data.ndim == 2 and output_data.shape[0] == 1:
# When input data is (batchsize=1, length)
output_data = np.expand_dims(output_data[0], -1)
else:
nii_warn.f_print("Output data format not supported.", "error")
nii_warn.f_print("Format is not (batch, len, dim)", "error")
nii_warn.f_die("Please use batch_size = 1 in generation")
# Save output
if output_data.shape[1] != self.m_output_all_dim:
nii_warn.f_print("Output data dim != expected dim", "error")
nii_warn.f_print("Output:%d" % (output_data.shape[1]), \
"error")
nii_warn.f_print("Expected:%d" % (self.m_output_all_dim), \
"error")
nii_warn.f_die("Please check configuration")
if not os.path.isdir(save_dir):
try:
os.makedirs(save_dir, exist_ok=True)
except OSError:
nii_warn.f_die("Cannot carete {}".format(save_dir))
# read the sentence information
tmp_seq_info = nii_seqinfo.SeqInfo()
tmp_seq_info.parse_from_str(data_infor_str)
# write the data
file_name = tmp_seq_info.seq_tag()
if len(filename_prefix):
file_name = filename_prefix + file_name
seq_length = tmp_seq_info.seq_length()
s_dim = 0
e_dim = 0
for t_ext, t_dim, t_reso in \
zip(self.m_output_exts, self.m_output_dims, self.m_output_reso):
e_dim = s_dim + t_dim
file_path = nii_str_tk.f_realpath(save_dir, file_name, t_ext)
# if this file_name contains part of the path, make sure that the
# parent folder has been created
if self.flag_filename_with_path:
tmp_save_dir = os.path.dirname(file_path)
if not os.path.isdir(tmp_save_dir):
try:
os.makedirs(tmp_save_dir, exist_ok=True)
except OSError:
nii_warn.f_die("Cannot carete {}".format(tmp_save_dir))
# check the length and write the data
if seq_length > 0:
expect_len = seq_length // t_reso
# confirm that the generated file length is as expected
if output_data.shape[0] < expect_len:
nii_warn.f_print("Warning {:s}".format(file_path), "error")
nii_warn.f_print("Generated data is shorter than expected")
nii_warn.f_print("Please check the generated file")
if s_dim == 0 and e_dim == output_data.shape[1]:
# if there is only one output feature, directly output it
self.f_write_data(output_data[:expect_len], file_path)
else:
# else, output the corresponding dimentions
self.f_write_data(output_data[:expect_len, s_dim:e_dim],
file_path)
elif seq_length == 0:
# if seq_length == 0, this is for unaligned input
if s_dim == 0 and e_dim == output_data.shape[1]:
self.f_write_data(output_data, file_path)
else:
self.f_write_data(output_data[s_dim:e_dim], file_path)
else:
nii_warn.f_die("Error: seq_length < 0 in generation")
return
def f_input_dim(self):
"""
f_input_dim()
return the total dimension of input features
"""
return self.m_input_all_dim
def f_output_dim(self):
"""
f_output_dim
return the total dimension of output features
"""
return self.m_output_all_dim
def f_adjust_idx(self, data_tuple, idx_shift):
"""
f_adjust_idx
This is to be used by customize_dataset for idx adjustment.
When multiple data sets are merged, the idx from __getitem__
should be adjusted.
Only data_io itselts knows how to identify idx from the output of
__getitem__, we need to define the function here
"""
if isinstance(data_tuple[-1], list) \
or isinstance(data_tuple[-1], torch.Tensor):
# if data_tuple has been collated
for idx in np.arange(len(data_tuple[-1])):
data_tuple[-1][idx] += idx_shift
else:
# if data_tuple is from __getitem()__
data_tuple = (data_tuple[0], data_tuple[1],
data_tuple[2], data_tuple[-1] + idx_shift)
return data_tuple
def f_manage_data(self, idx, opt):
"""
f_mange_seq(self, idx)
Args:
idx: list of int, list of data indices
opt: 'keep', keep only data in idx
'delete', delete data in idx
"""
if type(idx) is not list:
nii_warn.f_die("f_delete_seq(idx) expects idx to be list")
# get a new list of data for this database
if opt == 'delete':
# convert to set of int
idx_set = set([int(x) for x in idx])
tmp_idx = [x for x in range(self.__len__()) if x not in idx_set]
else:
tmp_idx = [int(x) for x in idx]
# keep the specified data indices
self.m_seq_info = [self.m_seq_info[x] for x in tmp_idx \
if x < self.__len__() and x >= 0]
# re-compute the total length of data
self.m_data_total_length = self.f_sum_data_length()
return
def f_get_seq_name_list(self):
""" return list of data of names in the dataset
"""
return [x.seq_tag() for x in self.m_seq_info]
def f_get_seq_info(self):
return [x.print_to_str() for x in self.m_seq_info]
def f_get_seq_idx_from_name(self, data_names):
""" return the data index given the data names
This function is not used so often.
"""
data_list = self.f_get_seq_name_list()
try:
return [data_list.index(x) for x in data_names]
except ValueError:
nii_warn.f_print("Not all data names are in this dataset")
nii_warn.f_print("Return []")
return []
class NIIDataSetLoader:
""" NIIDataSetLoader:
A wrapper over torch.utils.data.DataLoader and DataSet
self.m_dataset will be the dataset
self.m_loader will be the dataloader
"""
def __init__(self,
dataset_name, \
file_list, \
input_dirs, input_exts, input_dims, input_reso, \
input_norm, \
output_dirs, output_exts, output_dims, output_reso, \
output_norm, \
stats_path, \
data_format = nii_dconf.h_dtype_str, \
params = None, \
truncate_seq = None, \
min_seq_len = None,
save_mean_std = True, \
wav_samp_rate = None, \
flag_lang = 'EN',
global_arg = None,
dset_config = None,
input_augment_funcs = None,
output_augment_funcs = None,
inoutput_augment_func = None):
"""
NIIDataSetLoader(
data_set_name,
file_list,
input_dirs, input_exts, input_dims, input_reso, input_norm,
output_dirs, output_exts, output_dims, output_reso, output_norm,
stats_path,
data_format = '<f4',
params = None,
truncate_seq = None,
min_seq_len = None,
save_mean_std = True, \
wav_samp_rate = None, \
flag_lang = 'EN',
global_arg = None,
dset_config = None,
input_augment_funcs = None,
output_augment_funcs = None,
inoutput_augment_func = None):
Args
----
data_set_name: a string to name this dataset
this will be used to name the statistics files
such as the mean/std for this dataset
file_list: a list of file name strings (without extension)
or, path to the file that contains the file names
input_dirs: a list of dirs from which input feature is loaded
input_exts: a list of input feature name extentions
input_dims: a list of input feature dimensions
input_reso: a list of input feature temporal resolution,
or None
input_norm: a list of bool, whether normalize input feature or not
output_dirs: a list of dirs from which output feature is loaded
output_exts: a list of output feature name extentions
output_dims: a list of output feature dimensions
output_reso: a list of output feature temporal resolution,
or None
output_norm: a list of bool, whether normalize target feature or not
stats_path: path to the directory of statistics(mean/std)
data_format: method to load the data
'<f4' (default): load data as float32m little-endian
'htk': load data as htk format
params: parameter for torch.utils.data.DataLoader
truncate_seq: None or int,
truncate data sequence into smaller truncks
truncate_seq > 0 specifies the trunck length
min_seq_len: None (default) or int, minimum length of an utterance
utterance shorter than min_seq_len will be ignored
save_mean_std: bool, True (default): save mean and std
wav_samp_rate: None (default) or int, if input data has waveform,
please set sampling rate. It is used by _data_writer
flag_lang: str, 'EN' (default), if input data has text, text will
be converted into code indices. flag_lang indicates the
language for the text processer, used by _data_reader
global_arg: argument parser returned by arg_parse.f_args_parsed()
default None
input_augment_funcs: list of functions for input data augmentation,
default None
output_augment_funcs: list of functions for output data augmentation
default None
inoutput_augment_func: a single data augmentation function
default None
Methods
-------
get_loader(): return a torch.util.data.DataLoader
get_dataset(): return a torch.util.data.DataSet
"""
nii_warn.f_print_w_date("Loading dataset %s" % (dataset_name),
level="h")
# create torch.util.data.DataSet
self.m_dataset = NIIDataSet(dataset_name, \
file_list, \
input_dirs, input_exts, \
input_dims, input_reso, \
input_norm, \
output_dirs, output_exts, \
output_dims, output_reso, \
output_norm, \
stats_path, data_format, \
truncate_seq, min_seq_len,\
save_mean_std, \
wav_samp_rate, \
flag_lang, \
global_arg,\
dset_config, \
input_augment_funcs,
output_augment_funcs,
inoutput_augment_func)
# create torch.util.data.DataLoader
if params is None:
tmp_params = nii_dconf.default_loader_conf
else:
tmp_params = params.copy()
# save parameters
self.m_params = tmp_params
# create data loader
self.m_loader = self.build_loader()
# done
return
def build_loader(self):
"""
"""
# initialize sampler if necessary
tmp_params = self.m_params.copy()
if 'sampler' in tmp_params:
tmp_sampler = None
if tmp_params['sampler'] == nii_sampler_fn.g_str_sampler_bsbl:
if 'batch_size' in tmp_params and tmp_params['batch_size']>1:
# initialize the sampler
tmp_sampler = nii_sampler_fn.SamplerBlockShuffleByLen(
self.m_dataset.f_get_seq_len_list(),
tmp_params['batch_size'])
# turn off automatic shuffle
tmp_params['shuffle'] = False
else:
nii_warn.f_print("{:s} off as batch-size is 1".format(
nii_sampler_fn.g_str_sampler_bsbl))
#nii_warn.f_die("Sampler requires batch size > 1")
tmp_params['sampler'] = tmp_sampler
# collate function
if 'batch_size' in tmp_params and tmp_params['batch_size'] > 1:
# for batch-size > 1, use customize_collate to handle
# data with different length
collate_fn = nii_collate_fn.customize_collate
else:
collate_fn = None
# return the loader
return torch.utils.data.DataLoader(
self.m_dataset, collate_fn=collate_fn, **tmp_params)
def get_loader_params(self):
return self.m_params
def get_loader(self):
""" get_loader():
Return the dataLoader (torch.util.data.DataLoader)
"""
return self.m_loader
def get_dataset(self):
""" get_dataset():
Return the dataset (torch.util.data.Dataset)
"""
return self.m_dataset
def get_data_mean_std(self):
"""
"""
return self.m_dataset.f_get_mean_std_tuple()
def print_info(self):
"""
"""
self.m_dataset.f_print_info()
print(str(self.m_params))
return
def get_seq_name_list(self):
return self.m_dataset.f_get_seq_name_list()
def get_seq_info(self):
return self.m_dataset.f_get_seq_info()
def get_seq_idx_from_name(self, data_names):
return self.m_dataset.f_get_seq_idx_from_name(data_names)
def putitem(self, output_data, save_dir, filename_prefix, data_infor_str):
""" Decompose the output_data from network into
separate files
"""
self.m_dataset.f_putitem(output_data, save_dir, filename_prefix,
data_infor_str)
def get_in_dim(self):
""" Return the dimension of input features
"""
return self.m_dataset.f_input_dim()
def get_out_dim(self):
""" Return the dimension of output features
"""
return self.m_dataset.f_output_dim()
def get_seq_num(self):
""" Return the number of sequences (after truncation)
"""
return self.m_dataset.f_get_num_seq()
def adjust_utt_idx(self, data_tuple, utt_idx_shift):
""" Return data tuple with adjusted utterance index in merged dataset
This is used by customize_dataset.
"""
return self.m_dataset.f_adjust_idx(data_tuple, utt_idx_shift)
def manage_data(self, data_idx, opt):
"""
manage_data(self, data_idx)
Args:
data_idx: list of indices, samples with these indices will be deleted
opt: 'keep', keep only data in idx
'delete', delete data in idx
"""
# delete the data from dataset
self.m_dataset.f_delete_seq(data_idx, opt)
# rebuild dataloader
self.m_loader = self.build_loader()
return
def update_seq_len_in_sampler_sub(self, data_info):
"""
"""
data_idx = seq_info.parse_idx(one_info)
data_len = seq_info.parse_length(one_info)
self.m_dataset.f_update_seq_len_for_sampler_list(data_idx, data_len)
return
def update_seq_len_in_sampler(self):
"""update_seq_len()
Update sequence length if sequence length has been changed
(for example, during silence trim process)
This is necessary when using shuffle_by_seq_length sampler
and the sequences were trimmed in data augmentation function.
"""
# only useful for shuffle_by_seq_length sampler
if self.m_params['sampler'] == nii_sampler_fn.g_str_sampler_bsbl:
if hasattr(self.m_loader.sampler, 'update_seq_length'):
self.m_loader.sampler.update_seq_length(
self.m_dataset.f_get_updated_seq_len_for_sampler_list())
else:
print("Unknown error in update_seq_len_in_sampler")
sys.exit(1)
return
if __name__ == "__main__":
pass
| 79,969 | 42.39121 | 84 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/customize_sampler.py | #!/usr/bin/env python
"""
customized sampler
1. Block shuffler based on sequence length
Like BinnedLengthSampler in https://github.com/fatchord/WaveRNN
e.g., data length [1, 2, 3, 4, 5, 6] -> [3,1,2, 6,5,4] if block size =3
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import torch
import torch.utils.data
import torch.utils.data.sampler as torch_sampler
import core_scripts.math_tools.random_tools as nii_rand_tk
import core_scripts.other_tools.display as nii_warn
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
# name of the sampler
g_str_sampler_bsbl = 'block_shuffle_by_length'
###############################################
# Sampler definition
###############################################
class SamplerBlockShuffleByLen(torch_sampler.Sampler):
""" Sampler with block shuffle based on sequence length
e.g., data length [1, 2, 3, 4, 5, 6] -> [3,1,2, 6,5,4] if block size =3
"""
def __init__(self, buf_dataseq_length, batch_size):
""" SamplerBlockShuffleByLength(buf_dataseq_length, batch_size)
args
----
buf_dataseq_length: list or np.array of int,
length of each data in a dataset
batch_size: int, batch_size
"""
if batch_size == 1:
mes = "Sampler block shuffle by length requires batch-size>1"
nii_warn.f_die(mes)
# hyper-parameter, just let block_size = batch_size * 3
self.m_block_size = batch_size * 4
# idx sorted based on sequence length
self.m_idx = np.argsort(buf_dataseq_length)
return
def __iter__(self):
""" Return a iterator to be iterated.
"""
tmp_list = list(self.m_idx.copy())
# shuffle within each block
# e.g., [1,2,3,4,5,6], block_size=3 -> [3,1,2,5,4,6]
nii_rand_tk.f_shuffle_in_block_inplace(tmp_list, self.m_block_size)
# shuffle blocks
# e.g., [3,1,2,5,4,6], block_size=3 -> [5,4,6,3,1,2]
nii_rand_tk.f_shuffle_blocks_inplace(tmp_list, self.m_block_size)
# return a iterator, list is iterable but not a iterator
# https://www.programiz.com/python-programming/iterator
return iter(tmp_list)
def __len__(self):
""" Sampler requires __len__
https://pytorch.org/docs/stable/data.html#torch.utils.data.Sampler
"""
return len(self.m_idx)
def update_seq_length(self, buf_dataseq_length):
"""Update sequence length if necessary
This will resort the sequences based on updated sequence length
"""
if len(buf_dataseq_length) == len(self.m_idx):
self.m_idx = np.argsort(buf_dataseq_length)
else:
print("Incompatible sequence length input: updata_seq_length")
sys.exit(1)
return
if __name__ == "__main__":
print("Definition of customized_sampler")
| 2,994 | 30.861702 | 75 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/dsp_tools.py | #!/usr/bin/env python
"""
dsp_tools
Interface to process waveforms with DSP tools
Note that functions here are based on numpy, and they are intended to be used
before data are converted into torch tensors.
data on disk -> DataSet.__getitem__() -----> Collate ----> Pytorch model
numpy.tensor torch.tensor
These functions don't work on pytorch tensors
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import scipy
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
class Melspec(object):
"""Melspec
A simple class to produce and invert mel-spectrogram
Note that this not compatible with librosa.melspec
Most of the API is written by Dr. Shinji Takaki
"""
def __init__(self, sf=16000, fl=400, fs=80, fftl=1024, mfbsize=80,
melmin=0, melmax=None, ver=1):
"""Melspec(sf, fl, fs, fftl, mfbsize, melmin, melmax)
Args
----
sf: int, sampling rate
fl: int, frame length (number of waveform points)
fs: int, frame shift
fftl: int, FFT points
mfbsize: int, mel-filter bank size
melmin: float, lowest freq. covered by mel-filter bank, default 0
melmax: float, highest freq. covered by mel-filter bank, default sf/2
Note
----
configuration for Voiceprivacy challenge:
dsp_tools.Melspec(fftl=1024, fl=400, fs=160, ver=2)
"""
#
self.ver = ver
# sampling rate
self.sf = sf
# frame length
self.fl = fl
# frame shift
self.fs = fs
# fft length
self.fftl = fftl
# mfbsize
self.mfbsize = mfbsize
# mel.min frequency (in Hz)
self.melmin = melmin
# mel.max frequency (in Hz)
if melmax is None:
self.melmax = sf/2
else:
self.melmax = melmax
# windows
self.window = np.square(np.blackman(self.fl).astype(np.float32))
winpower = np.sqrt(np.sum(self.window))
if self.ver == 2:
self.window = np.blackman(self.fl).astype(np.float32) / winpower
else:
self.window = self.window / winpower
# create mel-filter bank
self.melfb = self._melfbank(self.melmin, self.melmax)
# eps = 1.0E-12
self.eps = 1.0E-12
return
def _freq2mel(self, freq):
return 1127.01048 * np.log(freq / 700.0 + 1.0)
def _mel2freq(self, mel):
return (np.exp(mel / 1127.01048) - 1.0) * 700.0
def _melfbank(self, melmin, melmax):
linear_freq = 1000.0
mfbsize = self.mfbsize - 1
bFreq = np.linspace(0, self.sf / 2.0, self.fftl//2 + 1,
dtype=np.float32)
minMel = self._freq2mel(melmin)
maxMel = self._freq2mel(melmax)
iFreq = self._mel2freq(np.linspace(minMel, maxMel, mfbsize + 2,
dtype=np.float32))
linear_dim = np.where(iFreq<linear_freq)[0].size
iFreq[:linear_dim+1] = np.linspace(iFreq[0], iFreq[linear_dim],
linear_dim+1)
diff = np.diff(iFreq)
so = np.subtract.outer(iFreq, bFreq)
lower = -so[:mfbsize] / np.expand_dims(diff[:mfbsize], 1)
upper = so[2:] / np.expand_dims(diff[1:], 1)
fb = np.maximum(0, np.minimum(lower, upper))
enorm = 2.0 / (iFreq[2:mfbsize+2] - iFreq[:mfbsize])
fb *= enorm[:, np.newaxis]
fb0 = np.hstack([np.array(2.0*(self.fftl//2)/self.sf, np.float32),
np.zeros(self.fftl//2, np.float32)])
fb = np.vstack([fb0, fb])
return fb
def _melfbank_pinv(self, melfb):
"""get the pseudo inverse of melfb
"""
return
def _frame(self, X):
"""framing
"""
X = np.concatenate([np.zeros(self.fl//2, np.float32), X,
np.zeros(self.fl//2, np.float32)])
frame_num = (X.shape[0] - self.fl) // self.fs + 1
F = np.zeros([frame_num, self.fl])
for frame_idx in np.arange(frame_num):
F[frame_idx, :] = X[frame_idx*self.fs : frame_idx*self.fs+self.fl]
return F
def _anawindow(self, F):
W = F * self.window
return W
def _rfft(self, W):
Y = np.fft.rfft(W, n=self.fftl).astype(np.complex64)
return Y
def _amplitude(self, Y):
A = np.fmax(np.absolute(Y), self.eps)
return A
def _logmelfbspec(self, A):
M = np.log(np.dot(A, self.melfb.T))
return M
def _preprocess(self, X):
if self.ver == 2:
# in ver2, assume wave in 16 bits
return X * np.power(2, 15)
else:
return X
def analyze(self, X):
"""Mel = analysze(X)
input: X, np.array, waveform data, (length, )
output: Mel, np.array, melspec., (frame_length, melfb_size)
"""
X = self._preprocess(X)
M = self._amplitude(self._rfft(self._anawindow(self._frame(X))))
M = self._logmelfbspec(M)
return M
class LPClite(object):
""" A lite LPC analyzer & synthesizr
Note that this is based on numpy, not Pytorch
It can be used for pre-processing when loading data, or use
it as data transformation function
(see message at top)
Example:
# load waveform
sr, wav = wav_tools.waveReadAsFloat(wav_path)
m_lpc = LPClite(320, 80)
# LPC analysis
lpc_coef, _, rc, gain, err, err_overlapped = m_lpc.analysis(wav)
# LPC synthesis
wav_re = m_lpc.synthesis(lpc_coef, err, gain)
# rc to LPC
lpc_coef_tmp = m_lpc._rc2lpc(lpc_coef)
np.std(lpc_coef_tmp - lpc_coef)
"""
def __init__(self, fl=320, fs=80, order=29, window='blackman',
flag_emph=True, emph_coef=0.97):
"""LPClite(fl=320, fs=80, order=30, window='blackman')
Args
----
fl: int, frame length
fs: int, frame shift
order: int, order of LPC, [1, a_1, a_2, ..., a_order]
window: str, 'blackman' or 'hanning'
flag_emph: bool, whether use pre-emphasis (default True)
emph_coef: float, coefficit for pre-emphasis filter (default 0.97)
Note that LPC model is defined as:
1 Gain
-------- ---------------------------------------------
1- bz^-1 a_0 + a_1 z^-1 + ... + a_order z^-(order)
b = emph_coef if flag_emph is True
b = 0 otherwise
"""
self.fl = fl
self.fs = fs
#
self.order = order
self.flag_emph = flag_emph
self.emph_coef = emph_coef
if np.abs(emph_coef) >= 1.0:
print("Warning: emphasis coef {:f} set to 0.97".format(emph_coef))
self.emph_coef = 0.97
if window == 'hanning':
self.win = np.hanning(self.fl)
else:
self.win = np.blackman(self.fl)
return
def analysis(self, wav):
"""lpc_coef, ld_err, gamma, gain, framed_err, err_signal = analysis(wav)
LPC analysis on each frame
input
-----
wav: np.array, (length, 1)
output
------
lpc_coef: np.array, LPC coeff, (frame_num, lpc_order + 1)
ld_err: np.array, LD analysis error, (frame_num, lpc_order + 1)
gamma: np.array, reflection coefficients, (frame_num,lpc_order)
gain: np.array, gain, (frame_num, 1)
framed_err: np.array, LPC error per frame, (frame_num, frame_length)
eer_signal: np.array, overlap-added excitation (length, 1)
Note that framed_err is the excitation signal from LPC analysis on each
frame. eer_signal is the overlap-added excitation signal.
"""
if self.flag_emph:
wav_tmp = self._preemphasis(wav)
else:
wav_tmp = wav
# framing & windowing
frame_wined = self._windowing(self._framing(wav_tmp[:, 0]))
# auto-correlation
auto = self._auto_correlation(frame_wined)
# LD analysis
lpc_coef, lpc_err, gamma_array, gain = self._levison_durbin(auto)
# get LPC excitation signals in each frame
framed_err = self._lpc_analysis_core(lpc_coef, frame_wined, gain)
# overlap-add for excitation signal
err_signal = self._overlapadd(framed_err)
return lpc_coef, lpc_err, gamma_array, gain, framed_err, err_signal
def synthesis(self, lpc_coef, framed_err, gain):
"""wav = synthesis(lpc_coef, framed_err, gain):
LPC synthesis (and overlap-add)
input
-----
lpc_coef: np.array, LPC coeff, (frame_num, lpc_order + 1)
framed_err: np.array, LPC excitations, (frame_num, frame_length)
gain: np.array, LPC gain, (frame_num, 1)
output
------
wav: np.array, (length, 1)
This function does LPC synthesis in each frame and create
the output waveform by overlap-adding
"""
framed_x = self._lpc_synthesis_core(lpc_coef, framed_err, gain)
wav_tmp = self._overlapadd(framed_x)
if self.flag_emph:
wav_tmp = self._deemphasis(wav_tmp)
return wav_tmp
def _preemphasis(self, wav):
""" wav_out = _preemphasis(wav)
input
-----
wav: np.array, (length)
output
------
wav: np.array, (length)
"""
wav_out = np.zeros_like(wav) + wav
wav_out[1:] = wav_out[1:] - wav_out[0:-1] * self.emph_coef
return wav_out
def _deemphasis(self, wav):
""" wav_out = _deemphasis(wav)
input
-----
wav: np.array, (length)
output
------
wav: np.array, (length)
"""
wav_out = np.zeros_like(wav) + wav
for idx in range(1, wav.shape[0]):
wav_out[idx] = wav_out[idx] + wav_out[idx-1] * self.emph_coef
return wav_out
def _framing(self, wav):
"""F = _framed(wav)
Framing the signal
input
-----
wav: np.array, (length)
output
------
F: np.array, (frame_num, frame_length)
"""
frame_num = (wav.shape[0] - self.fl) // self.fs + 1
F = np.zeros([frame_num, self.fl], dtype=wav.dtype)
for frame_idx in np.arange(frame_num):
F[frame_idx, :] = wav[frame_idx*self.fs : frame_idx*self.fs+self.fl]
return F
def _windowing(self, framed_x):
"""windowing
"""
return framed_x * self.win
def _overlapadd(self, framed_x):
"""wav = _overlapadd(framed_x)
Do overlap-add on framed (and windowed) signal
input
-----
framed_x: np.array, (frame_num, frame_length)
output
------
wav: np.array, (length, 1)
length = (frame_num - 1) * frame_shift + frame_length
"""
# waveform length
wavlen = (framed_x.shape[0] - 1) * self.fs + self.fl
wavbuf = np.zeros([wavlen])
# buf to save overlapped windows (to normalize the signal amplitude)
protobuf = np.zeros([wavlen])
win_prototype = self._windowing(self._framing(np.ones_like(protobuf)))
# overlap and add
for idx in range(framed_x.shape[0]):
frame_s = idx * self.fs
wavbuf[frame_s : frame_s + self.fl] += framed_x[idx]
protobuf[frame_s : frame_s + self.fl] += win_prototype[idx]
# remove the impact of overlapped windows
#protobuf[protobuf<1e-05] = 1.0
wavbuf = wavbuf / protobuf.mean()
return np.expand_dims(wavbuf, axis=1)
def _lpc_analysis_core(self, lpc_coef, framed_x, gain):
"""framed_err = _lpc_analysis_core(lpc_coef, framed_x, gain)
LPC analysis on frame
MA filtering: e[n] = \sum_k=0 a_k x[n-k] / gain
input
-----
lpc_coef: np.array, (frame_num, order + 1)
framed_x: np.array, (frame_num, frame_length)
gain: np.array, (frame_num, 1)
output
------
framed_err: np.array, (frame_num, frame_length)
Note that lpc_coef[n, :] = (1, a_1, a_2, ..., a_order) for n-th frame
framed_x[n, :] = (x[0], x[1], ..., x[frame_len]) for n-th frame
"""
#
frame_num = framed_x.shape[0]
frame_len = framed_x.shape[1]
# lpc order (without the a_0 term)
order = lpc_coef.shape[1] - 1
# pad zero, every frame has [0, ..., 0, x[0], x[1], ..., x[frame_len]]
tmp_framed = np.concatenate(
[np.zeros([frame_num, order + 1]), framed_x], axis=1)
# flip to (x[frame_len], ... x[1], x[0], 0, ..., 0)
tmp_framed = tmp_framed[:, ::-1]
# LPC excitation buffer
framed_err = np.zeros_like(framed_x)
# e[n] = \sum_k=0 a[k] x[n-k]
# do this for all frames and n simultaneously
for k in range(self.order + 1):
# a[k]
tmp_coef = lpc_coef[:, k:k+1]
# For each frame
# RHS = [x[n-k], x[n-k-1], ..., ] * a[k]
#
# By doing this for k in [0, order]
# LHS = [e[n], e[n-1], ...]
# [x[n-0], x[n-0-1], ..., ] * a[0]
# + [x[n-1], x[n-1-1], ..., ] * a[1]
# + [x[n-2], x[n-2-1], ..., ] * a[2]
# + ...
# We get the excitation for one frame
# This process is conducted for all frames at the same time
framed_err += tmp_framed[:, 0:frame_len] * tmp_coef
# roll to [x[n-k-1], x[n-k-2], ..., ]
tmp_framed = np.roll(tmp_framed, -1, axis=1)
# revese to (e[0], e[1], ..., e[frame_len])
return framed_err[:, ::-1] / gain
def _lpc_synthesis_core(self, lpc_coef, framed_err, gain):
"""framed_x = _lpc_synthesis_core(lpc_coef, framed_err, gain)
AR filtering: x[n] = gain * e[n] - \sum_k=0 a_k x[n-k]
LPC synthesis on frame
input
-----
lpc_coef: np.array, (frame_num, order + 1)
framed_err: np.array, (frame_num, frame_length)
gain: np.array, (frame_num, 1)
output
------
framed_x: np.array, (frame_num, frame_length)
Note that
lpc_coef[n, :] = (1, a_1, a_2, ..., a_order), for n-th frame
framed_x[n, :] = (x[0], x[1], ..., x[frame_len]), for n-th frame
"""
frame_num = framed_err.shape[0]
frame_len = framed_err.shape[1]
order = lpc_coef.shape[1] - 1
# pad zero
# the buffer looks like
# [[0, 0, 0, 0, 0, ... x[0], x[1], x[frame_length -1]], -> 1st frame
# [0, 0, 0, 0, 0, ... x[0], x[1], x[frame_length -1]], -> 2nd frame
# ...]
framed_x = np.concatenate(
[np.zeros([frame_num, order]), np.zeros_like(framed_err)], axis=1)
# flip the cofficients of each frame as [a_order, ..., a_1, 1]
lpc_coef_tmp = lpc_coef[:, ::-1]
# synthesis (all frames are down at the same time)
for idx in range(frame_len):
# idx+order so that it points to the shifted time idx
# idx+order
# [0, 0, 0, 0, 0, ... x[0], x[1], ... x[idx], ]
# gain * e[n]
framed_x[:, idx+order] = framed_err[:, idx] * gain[:, 0]
# [x[idx-1-order], ..., x[idx-1]] * [a_order, a_1]
pred = np.sum(framed_x[:, idx:idx+order] * lpc_coef_tmp[:, :-1],
axis=1)
# gain * e[n] - [x[idx-1-order], ..., x[idx-1]] * [a_order, a_1]
framed_x[:, idx+order] = framed_x[:, idx+order] - pred
# [0, 0, 0, 0, 0, ... x[0], x[1], ... ] -> [x[0], x[1], ...]
return framed_x[:, order:]
def _auto_correlation(self, framed_x):
""" autocorr = _auto_correlation(framed_x)
input
-----
framed_x: np.array, (frame_num, frame_length), frame windowed signal
output
------
autocorr: np.array, auto-correlation coeff (frame_num, lpc_order+1)
"""
# (frame_num, order)
autocor = np.zeros([framed_x.shape[0], self.order+1])
# loop and compute auto-corr (for all frames simultaneously)
for i in np.arange(self.order+1):
autocor[:, i] = np.sum(
framed_x[:, 0:self.fl-i] * framed_x[:, i:],
axis=1)
#print(autocor[0, i])
#autocor[:, i] = 0
#for idx in np.arange(self.fl):
# if (idx + i) < self.fl:
# autocor[:, i] += framed_x[:, idx] * framed_x[:, idx + i]
# else:
# break
#print(autocor[0, i])
# (frame_num, order)
return autocor
def _levison_durbin(self, autocor):
"""lpc_coef_ou, lpc_err, gamma_array, gain = _levison_durbin(autocor)
Levison durbin
input
-----
autocor: np.array, auto-correlation, (frame_num, lpc_order+1)
output
------
lpc_coef: np.array, LPC coefficients, (frame_num, lpc_order+1)
lpc_err: np.array, LPC error, (frame_num, lpc_order+1)
gamma: np.array, reflection coeff, (frame_num, lpc_order)
gain: np.array, gain, (frame_num, 1)
Note that lpc_coef[n] = (1, a_2, ... a_order) for n-th frame
"""
# (frame_num, order)
frame_num, order = autocor.shape
order = order - 1
polyOrder = order + 1
# to log down the invalid frames
tmp_order = np.zeros([frame_num], dtype=np.int32) + polyOrder
lpc_coef = np.zeros([frame_num, 2, polyOrder])
lpc_err = np.zeros([frame_num, polyOrder])
gamma_array = np.zeros([frame_num, order])
gain = np.zeros([frame_num])
lpc_err[:, 0] = autocor[:, 0]
lpc_coef[:, 0, 0] = 1.0
for index in np.arange(1, polyOrder):
lpc_coef[:, 1, index] = 1.0
# compute gamma
# step1.
gamma = np.sum(lpc_coef[:, 0, 0:(index)] * autocor[:, 1:(index+1)],
axis=1)
# step2. check validity of lpc_err
ill_idx = lpc_err[:,index-1] < 1e-07
# also frames that should have been stopped in previous iter
ill_idx = np.bitwise_or(ill_idx, tmp_order < polyOrder)
# step3. make invalid frame gamma=0
gamma[ill_idx] = 0
gamma[~ill_idx] = gamma[~ill_idx] / lpc_err[~ill_idx,index-1]
gamma_array[:, index-1] = gamma
# step4. log down the ill frames
tmp_order[ill_idx] = index
lpc_coef[:, 1, 0] = -1.0 * gamma
if index > 1:
lpc_coef[:, 1, 1:index] = lpc_coef[:, 0, 0:index-1] \
+ lpc_coef[:, 1, 0:1] * lpc_coef[:, 0, 0:index-1][:, ::-1]
lpc_err[:, index] = lpc_err[:, index-1] * (1 - gamma * gamma)
lpc_coef[:, 0, :] = lpc_coef[:, 1, :]
# flip to (1, a_1, ..., a_order)
lpc_coef = lpc_coef[:, 0, ::-1]
# output LPC coefficients
lpc_coef_ou = np.zeros([frame_num, polyOrder])
# if high-order LPC analysis is not working
# each frame may require a different truncation length
for idx in range(frame_num):
lpc_coef_ou[idx, 0:tmp_order[idx]] = lpc_coef[idx, 0:tmp_order[idx]]
# get the gain, when tmp_order = polyOrder, tmp_order-2 -> order-1,
# last element of the lpc_err buffer
gain = np.sqrt(lpc_err[np.arange(len(tmp_order)), tmp_order-2])
# if the gain is zero, it means analysis error is zero,
gain[gain < 1e-07] = 1.0
# (frame_num, order)
return lpc_coef_ou, lpc_err, gamma_array, np.expand_dims(gain, axis=1)
def _rc2lpc(self, rc):
"""lpc_coef = _rc2lpc(rc)
from reflection coefficients to LPC coefficients
forward Levinson recursion
input
-----
rc: np.array, (frame_num, lpc_order)
output
------
lpc_coef, np.array, (frame_num, lpc_order+1)
Note that LPC model is defined as:
Gain
---------------------------------------------
a_0 + a_1 z^-1 + ... + a_order z^-(order)
Thus, the reflection coefficitns [gamma_1, ... gamma_order]
"""
# (frame_num, order)
frame_num, order = rc.shape
polyOrder = order + 1
lpc_coef = np.zeros([frame_num, 2, polyOrder])
lpc_coef[:, 0, 0] = 1.0
for index in np.arange(1, polyOrder):
lpc_coef[:, 1, index] = 1.0
gamma = rc[:, index-1]
lpc_coef[:, 1, 0] = -1.0 * gamma
if index > 1:
lpc_coef[:, 1, 1:index] = lpc_coef[:, 0, 0:index-1] \
+ lpc_coef[:, 1, 0:1] * lpc_coef[:, 0, 0:index-1][:, ::-1]
lpc_coef[:, 0, :] = lpc_coef[:, 1,:]
lpc_coef = lpc_coef[:, 0, ::-1]
return lpc_coef
def f0resize(input_f0, input_reso, output_reso):
"""output_f0 = f0size(input_f0, input_reso, output_reso)
input
-----
input_f0: array, (length, )
input_reso: int, frame_shift, ms
output_reso: int, frame_shift, ms
output
------
output_f0: array, (length2, )
where length2 ~ np.ceil(length * input_reso / output_reso)
"""
# function to merge two f0 value
# average them unless there is u/v mismatch
def merge_f0(val1, val2):
if val1 < 1 and val2 < 1:
return (val1 + val2)/2
elif val1 < 1:
return val2
elif val2 < 1:
return val1
else:
return (val1 + val2)/2
def retrieve_f0(buf, idx):
if idx > 0 and idx < buf.shape[0]:
return buf[idx]
else:
return 0
# input length
input_len = input_f0.shape[0]
# output length
output_len = int(np.ceil(input_len * input_reso / output_reso))
# output buffer
output_f0 = np.zeros([output_len])
for idx in np.arange(output_len):
input_idx = idx * output_reso / input_reso
input_idx_left = int(np.floor(input_idx))
input_idx_right = int(np.ceil(input_idx))
# get the nearest value from input f0
val1 = retrieve_f0(input_f0, input_idx_left)
val2 = retrieve_f0(input_f0, input_idx_right)
output_f0[idx] = merge_f0(val1, val2)
return output_f0
def spectra_substraction(input1, input2, ratio=0.1,
frame_length = 512, frame_shift = 256, fft_n = 512):
"""
output = spectra_substraction(input1, input2, ratio=0.1,
frame_length = 512, frame_shift = 256, fft_n = 512)
input
-----
input1: array, (length1 ), waveform to be denoised
input2: array, (length2 ), waveform background noise
ratio: float, weight to average spectra of noise
frame_length, frame_shift, fft_n
output
------
output: array, (length 1)
"""
_, _, input_spec1 = scipy.signal.stft(
input1, nperseg = frame_length,
noverlap = frame_length - frame_shift, nfft=fft_n)
_, _, input_spec2 = scipy.signal.stft(
input1, nperseg = frame_length,
noverlap = frame_length - frame_shift, nfft=fft_n)
# ampltiude and phase
amp1 = np.abs(input_spec1)
pha1 = np.angle(input_spec1)
# nosie average spectrum
amp2 = np.abs(input_spec2)
amp2 = amp2.mean(axis=1, keepdims=1)
#idx = np.bitwise_and(amp1 > 0.0000001, amp2 > 0.0000001)
#amp_new = amp1
#amp_new[idx] = np.exp(np.log(amp1[idx]) - np.log((amp2[idx] * ratio)))
# spectra substraction
amp_new = amp1 - amp2 * ratio
# keep amplitude none-negative
amp_new[amp_new<0] = 0.0
# reconstruct
spec_new = amp_new * np.cos(pha1) + 1j * amp_new * np.sin(pha1)
_, output = scipy.signal.istft(
spec_new, nperseg=frame_length,
noverlap=frame_length - frame_shift, nfft = fft_n)
return output
def GriffinLim(sp_amp, n_iter, fl, fs, fft_n,
window='hann', momentum=0.99, init='rand'):
"""
wav = GriffinLim(sp_amp, n_iter, fl, fs, fft_n,
window='hann', momentum=0.99, init='rand')
Code based on librosa API.
input
-----
sp_amp: array, (frame, fft_n//2+1), spectrum amplitude (linear domain)
n_iter: int, number of GL iterations
fl: int, frame length
fs: int, frame shift
fft_n: int, number of FFT points,
window: str, default hann window
momentum: float, momentum for fast GL iteration default 0.99
init: str, initialization method of initial phase, default rand
output
------
wav: array, (length, ), reconstructed waveform
Example
-------
nfft = 512
fl = 512
fs = 256
_, _, data_stft = scipy.signal.stft(data1, window='hann', nperseg=fl,
noverlap=fl - fs, nfft = nfft)
data_stft = np.abs(data_stft)
wav = GriffinLim(data_stft, 32, fl, fs, nfft)
"""
def angle_to_complex(x):
return np.cos(x) + 1j * np.sin(x)
# check data shape
if sp_amp.shape[0] != fft_n // 2 + 1:
spec_amp = sp_amp.T
if spec_amp.shape[0] != fft_n // 2 + 1:
print("Input sp_amp has shape {:s}".format(str(sp_amp)))
print("FFT bin number is {:d}, incompatible".format(fft_n))
else:
spec_amp = sp_amp
# small value
eps = 0.0000001
# buffer for angles
angles = np.zeros(spec_amp.shape, dtype=np.complex64)
# initialize phase
if init == "rand":
angles[:] = angle_to_complex(2*np.pi * np.random.rand(*spec_amp.shape))
else:
angles[:] = 1.0
# Place-holders for temporary data and reconstructed buffer
rebuilt = None
tprev = None
inverse = None
# Absorb magnitudes into angles
angles *= spec_amp
for _ in range(n_iter):
# Invert
_, inverse = scipy.signal.istft(angles, window = window,
nperseg=fl, noverlap=fl - fs, nfft = fft_n)
# rebuild
_, _, rebuilt = scipy.signal.stft(inverse, window = window,
nperseg=fl, noverlap=fl - fs, nfft = fft_n)
# update
angles[:] = rebuilt
if tprev is not None:
angles -= (momentum / (1 + momentum)) * tprev
angles /= np.abs(angles) + eps
angles *= spec_amp
#
rebuilt, tprev = tprev, rebuilt
# reconstruct
_, wav = scipy.signal.istft(angles, window = window,
nperseg=fl, noverlap=fl - fs, nfft = fft_n)
return wav
def warp_function_bilinear(normed_freq, alpha=0):
""" warped_freq = warp_function_quadratic(normed_freq)
Frequency warping using bi-linear function
input
-----
normed_freq: np.array, (N, ), normalized frequency values
between 0 and pi
alpha: float, warping coefficient. alpha=0 means no warping
output
------
warped_freq: np.array, (N, ), warpped normalized frequency
Example
-------
orig_rad = np.arange(0, 512)/512 * np.pi
warp_rad = warp_function_bilinear(orig_rad, alpha=0.3)
"""
if np.any(normed_freq < 0) or np.any(normed_freq > np.pi):
print("Input freq. out of range")
sys.exit(1)
nom = (1 - alpha * alpha) * np.sin(normed_freq)
den = (1 + alpha * alpha) * np.cos(normed_freq) - 2 * alpha
output = np.arctan(nom / den)
output[output < 0] = output[output < 0] + np.pi
return output
def warp_interpolation(spec, alpha, warp_func=None):
"""output = wrap_interpolation(spec, spec)
Do frequency Warping and linear interpolation of spectrum.
This is used for Vocal-tract pertubation
input
-----
spec: spectra evelope, (L, N), where L is the frame number
alpha: float, coefficients for warping
warp_func: a warp function,
if None, we will use warp_function_bilinear in dsp_tools.py
output
------
output: spectra evelope, (L, N), where L is the frame number
Example
-------
# let us vocal-tract length perturbation
# let's do warping on spectrum envelope
# we use pyworld to extract spectrum envelope
import pyworld as pw
x, sf = some_waveread_function(audio_file)
# World analysis
_f0, t = pw.dio(x, sf) # raw pitch extractor
f0 = pw.stonemask(x, _f0, t, sf) # pitch refinement
sp = pw.cheaptrick(x, f0, t, sf) # extract smoothed spectrogram
ap = pw.d4c(x, f0, t, sf) # extract aperiodicity
# Synthesis without warpping
y = pw.synthesize(f0, sp, ap, sf)
# Synthesis after warpping
alpha = 0.1
sp_wrapped = warp_interpolation(sp, warp_function_bilinear, alpha)
ap_wrapped = warp_interpolation(ap, warp_function_bilinear, alpha)
y_wrapped = pw.synthesize(f0, sp_wrapped, ap_wrapped, sf)
# please listen and compare y and y_wrapped
"""
nbins = spec.shape[1]
orig_rad = np.arange(0, nbins) / nbins * np.pi
warp_rad = warp_func(orig_rad, alpha=alpha)
if np.mean(np.abs(warp_rad - orig_rad)) < 0.0001:
return spec
else:
output = np.zeros_like(spec)
for rad_idx in np.arange(nbins):
warp = warp_rad[rad_idx]
warp_idx = warp / np.pi * nbins
idx_left = int(np.floor(warp_idx))
idx_right = int(np.ceil(warp_idx))
if idx_left < 0:
idx_left = 0
if idx_right >= nbins:
idx_right = nbins - 1
if idx_left == idx_right:
w_l, w_r = 0.0, 1.0
else:
w_l = warp_idx - idx_left
w_r = idx_right - warp_idx
# weighted sum for interpolation
output[:,rad_idx] = spec[:,idx_left] * w_l + spec[:,idx_right] * w_r
return output
if __name__ == "__main__":
print("DSP tools using numpy")
# Example for downing LPC analysis
sr, data1 = wav_tools.waveReadAsFloat('media/arctic_a0001.wav')
m_lpc = LPClite(320, 80)
# LPC analysis
lpc_coef, _, rc, gain, err, err_overlapped = m_lpc.analysis(
np.expand_dims(data1, axis=1))
# LPC synthesis
wav_re = m_lpc.synthesis(lpc_coef, err, gain)
# excitation with Gain
excitation_new = m_lpc._overlapadd(err * gain)
# need to import
# from tutorials.plot_tools import plot_API
# from tutorials.plot_tools import plot_lib
plot_API.plot_API([wav_re[:, 0], data1,
err_overlapped[:, 0],
excitation_new[:, 0]],
plot_lib.plot_spec, 'v')
plot_API.plot_API([wav_re[:, 0] - err_overlapped[:, 0]],
plot_lib.plot_spec, 'single')
# RC to LPC
lpc_coef_tmp = m_lpc._rc2lpc(rc)
print(np.std(lpc_coef_tmp - lpc_coef))
| 32,385 | 31.680121 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/__init__.py | 0 | 0 | 0 | py |
|
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/conf.py | #!/usr/bin/env python
"""
config.py
Configurations for data_io
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import torch
import torch.utils.data
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
# ---------------------------
# Numerical configuration
# ---------------------------
# data type for host
h_dtype = np.float32
# data type string format for numpy
h_dtype_str = '<f4'
# data type for device (GPU)
d_dtype = torch.float32
# std_floor
std_floor = 0.00000001
# ---------------------------
# File name configuration
# ---------------------------
# name of the mean/std file for input features
mean_std_i_file = 'mean_std_input.bin'
# name of the mean/std file for output features
mean_std_o_file = 'mean_std_output.bin'
# name of the the uttrerance length file
data_len_file = 'utt_length.dic'
# ---------------------------
# F0 extention and unvoiced value
# ---------------------------
# dictionary: key is F0 file extention, value is unvoiced value
f0_unvoiced_dic = {'.f0' : 0}
# ---------------------------
# Data configuration
# ---------------------------
# minimum length of data. Sequence shorter than this will be ignored
data_seq_min_length = 40
# default configuration for torch.DataLoader
default_loader_conf = {'batch_size':1, 'shuffle':False, 'num_workers':0}
| 1,383 | 21.688525 | 72 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/customize_collate_fn.py | #!/usr/bin/env python
"""
customize_collate_fn
Customized collate functions for DataLoader, based on
github.com/pytorch/pytorch/blob/master/torch/utils/data/_utils/collate.py
PyTorch is BSD-style licensed, as found in the LICENSE file.
"""
from __future__ import absolute_import
import os
import sys
import torch
import re
import collections
#from torch._six import container_abcs, string_classes, int_classes
from torch._six import string_classes
"""
The primary motivation is to handle batch of data with varied length.
Default default_collate cannot handle that because of stack:
github.com/pytorch/pytorch/blob/master/torch/utils/data/_utils/collate.py
Here we modify the default_collate to take into consideration of the
varied length of input sequences in a single batch.
Notice that the customize_collate_fn only pad the sequences.
For batch input to the RNN layers, additional pack_padded_sequence function is
necessary. For example, this collate_fn does something similar to line 56-66,
but not line 117 in this repo:
https://gist.github.com/HarshTrivedi/f4e7293e941b17d19058f6fb90ab0fec
"""
__author__ = "Xin Wang"
__email__ = "[email protected]"
np_str_obj_array_pattern = re.compile(r'[SaUO]')
customize_collate_err_msg = (
"customize_collate: batch must contain tensors, numpy arrays, numbers, "
"dicts or lists; found {}")
def pad_sequence(batch, padding_value=0.0):
""" output_batch = pad_sequence(batch)
input
-----
batch: list of tensor, [data_1, data2, ...], and data_1 is (len, dim, ...)
output
------
output_batch: list of tensor, [data_1_padded, data_2_padded, ...]
Pad a batch of data sequences to be same length (maximum length in batch).
This function is based on
pytorch.org/docs/stable/_modules/torch/nn/utils/rnn.html#pad_sequence.
Output list of tensor can be stacked into (batchsize, len, dim,...).
See customize_collate(batch) below
"""
# get the rest of the dimensions (dim, ...)
dim_size = batch[0].size()
trailing_dims = dim_size[1:]
# get the maximum length
max_len = max([s.size(0) for s in batch])
if all(x.shape[0] == max_len for x in batch):
# if all data sequences in batch have the same length, no need to pad
return batch
else:
# else, we need to pad
out_dims = (max_len, ) + trailing_dims
output_batch = []
for i, tensor in enumerate(batch):
# check the rest of dimensions
if tensor.size()[1:] != trailing_dims:
print("Data in batch has different dimensions:")
for data in batch:
print(str(data.size()))
raise RuntimeError('Fail to create batch data')
# save padded results
out_tensor = tensor.new_full(out_dims, padding_value)
out_tensor[:tensor.size(0), ...] = tensor
output_batch.append(out_tensor)
return output_batch
def customize_collate(batch):
""" customize_collate(batch)
Collate a list of data into batch. Modified from default_collate.
"""
elem = batch[0]
elem_type = type(elem)
if isinstance(elem, torch.Tensor):
# this is the main part to handle varied length data in a batch
# batch = [data_tensor_1, data_tensor_2, data_tensor_3 ... ]
#
batch_new = pad_sequence(batch)
out = None
if torch.utils.data.get_worker_info() is not None:
# If we're in a background process, concatenate directly into a
# shared memory tensor to avoid an extra copy
# allocate the memory based on maximum numel
numel = max([x.numel() for x in batch_new]) * len(batch_new)
storage = elem.storage()._new_shared(numel)
# updated according to latest collate function
# otherwise, it raises warning
# pytorch/blob/master/torch/utils/data/_utils/collate.py
out = elem.new(storage).resize_(
len(batch_new), *list(batch_new[0].size()))
#print(batch_new.shape[0], batch_new.shape[1])
return torch.stack(batch_new, 0, out=out)
elif elem_type.__module__ == 'numpy' and elem_type.__name__ != 'str_' \
and elem_type.__name__ != 'string_':
if elem_type.__name__ == 'ndarray' or elem_type.__name__ == 'memmap':
# array of string classes and object
if np_str_obj_array_pattern.search(elem.dtype.str) is not None:
raise TypeError(customize_collate_err_msg.format(elem.dtype))
# this will go to loop in the last case
return customize_collate([torch.as_tensor(b) for b in batch])
elif elem.shape == (): # scalars
return torch.as_tensor(batch)
elif isinstance(elem, float):
return torch.tensor(batch, dtype=torch.float64)
#elif isinstance(elem, int_classes):
elif isinstance(elem, int):
return torch.tensor(batch)
elif isinstance(elem, string_classes):
return batch
#elif isinstance(elem, container_abcs.Mapping):
elif isinstance(elem, collections.abc.Mapping):
return {key: customize_collate([d[key] for d in batch]) for key in elem}
elif isinstance(elem, tuple) and hasattr(elem, '_fields'): # namedtuple
return elem_type(*(customize_collate(samples) \
for samples in zip(*batch)))
#elif isinstance(elem, container_abcs.Sequence):
elif isinstance(elem, collections.abc.Sequence):
# check to make sure that the elements in batch have consistent size
it = iter(batch)
elem_size = len(next(it))
if not all(len(elem) == elem_size for elem in it):
raise RuntimeError('each element in batch should be of equal size')
# zip([[A, B, C], [a, b, c]]) -> [[A, a], [B, b], [C, c]]
transposed = zip(*batch)
return [customize_collate(samples) for samples in transposed]
raise TypeError(customize_collate_err_msg.format(elem_type))
def pad_sequence_batch(list_batch, padding_value=0.0):
""" output_batch = pad_sequence(list_batch)
input
-----
batch: list of batch, [batch_1, batch_2, ...], and batch_1 is
(batch_size, len, dim1, dim2, ...)
output
------
output_batch: list of tensor, [batch_1_padded, batch_2_padded, ...]
Different from pad_sequence, list_batch is a list of batched tensors
"""
# each batched_tensor has shape (batch, length, dim1, ...)
# get dimensions for (dim1, ...)
dim_size = list_batch[0].size()
if len(dim_size) <= 2:
return list_batch
trailing_dims = dim_size[2:]
# get the maximum length for each batched tensor
max_len = max([s.size(1) for s in list_batch])
if all(x.shape[1] == max_len for x in list_batch):
# if all data sequences in batch have the same length, no need to pad
return list_batch
else:
output_batch = []
for i, tensor in enumerate(list_batch):
# shape (batch, max_len, dim1, dim2, ...)
out_dims = (tensor.shape[0], max_len, ) + trailing_dims
# check the rest of dimensions
if tensor.size()[2:] != trailing_dims:
print("Data in batch has different dimensions:")
raise RuntimeError('Fail to pad batched data')
# save padded results
out_tensor = tensor.new_full(out_dims, padding_value)
out_tensor[:, :tensor.size(1), ...] = tensor
output_batch.append(out_tensor)
return output_batch
def customize_collate_from_batch(batch):
""" output = customize_collate_from_batch
input
-----
batch: list of tensor, [tensor1, tensor2, ...], where
each tensor has shape (batch, length, dim1, dim2, ...)
output
------
output: tensor (batch_sum, length, dim1, dim2, ...)
Similar to customize_collate, but input is a list of batch data that have
been collated through customize_collate.
The difference is use torch.cat rather than torch.stack to merge tensors.
Also, list of data is directly concatenated
This is used in customize_dataset when merging data from multiple datasets.
It is better to separate this function from customize_collate
"""
elem = batch[0]
elem_type = type(elem)
if isinstance(elem, torch.Tensor):
batch_new = pad_sequence_batch(batch)
out = None
if torch.utils.data.get_worker_info() is not None:
numel = max([x.numel() for x in batch_new]) * len(batch_new)
storage = elem.storage()._new_shared(numel)
# we need to resize_ to suppress a warning
# this is based on
# pytorch/blob/master/torch/utils/data/_utils/collate.py
# [batch_1, length, dim], [batch_2, length, dim] ...
# batch_new[0][0].size() -> length, dim, ...
# [x.shape[0] for x in batch_new] -> [batch_1, batch_2, ...]
out = elem.new(storage).resize_(
sum([x.shape[0] for x in batch_new]),
*list(batch_new[0][0].size()))
# here is the difference
# concateante (batch1, length, dim1, dim2, ...) (batch2, length, ...)
# into (batch1+batch2+..., length, dim1, dim2, ...)
return torch.cat(batch_new, 0, out=out)
elif elem_type.__module__ == 'numpy' and elem_type.__name__ != 'str_' \
and elem_type.__name__ != 'string_':
if elem_type.__name__ == 'ndarray' or elem_type.__name__ == 'memmap':
if np_str_obj_array_pattern.search(elem.dtype.str) is not None:
raise TypeError(customize_collate_err_msg.format(elem.dtype))
return customize_collate_from_batch(
[torch.as_tensor(b) for b in batch])
elif elem.shape == (): # scalars
return torch.as_tensor(batch)
elif isinstance(elem, float):
return torch.tensor(batch, dtype=torch.float64)
#elif isinstance(elem, int_classes):
elif isinstance(elem, int):
return torch.tensor(batch)
elif isinstance(elem, string_classes):
return batch
elif isinstance(elem, tuple):
# concatenate two tuples
tmp = elem
for tmp_elem in batch[1:]:
tmp += tmp_elem
return tmp
#elif isinstance(elem, container_abcs.Sequence):
elif isinstance(elem, collections.abc.Sequence):
it = iter(batch)
elem_size = len(next(it))
if not all(len(elem) == elem_size for elem in it):
raise RuntimeError('each element in batch should be of equal size')
transposed = zip(*batch)
return [customize_collate_from_batch(samples) for samples in transposed]
raise TypeError(customize_collate_err_msg.format(elem_type))
if __name__ == "__main__":
print("Definition of customized collate function")
| 11,034 | 36.662116 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/text_process/toolkit_en.py | #!/usr/bin/env python
"""
Simple text processer for English
Based on https://github.com/fatchord/WaveRNN
"""
import os
import sys
import re
from core_scripts.data_io.text_process import toolkit_all
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
######
## Pool of symbols
######
# symbols
_pad = '_'
# 'space' is treated as a symbol
_punctuation = '!\'(),-.:;? '
# space
_space = ' '
_eos = '~'
_letters = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz'
_skip_symbols = ['_', '~']
# ARPAbet symbols
# http://www.speech.cs.cmu.edu/cgi-bin/cmudict?in=C+M+U+Dictionary
# vowels are combined with the three possible lexical stree symbols 0 1 2
# see http://svn.code.sf.net/p/cmusphinx/code/trunk/cmudict/cmudict-0.7b.symbols
_arpabet_symbols_raw = [
'AA', 'AA0', 'AA1', 'AA2', 'AE', 'AE0', 'AE1', 'AE2',
'AH', 'AH0', 'AH1', 'AH2', 'AO', 'AO0', 'AO1', 'AO2',
'AW', 'AW0', 'AW1', 'AW2', 'AY', 'AY0', 'AY1', 'AY2',
'B', 'CH', 'D', 'DH', 'EH', 'EH0', 'EH1', 'EH2',
'ER', 'ER0', 'ER1', 'ER2', 'EY', 'EY0', 'EY1', 'EY2',
'F', 'G', 'HH', 'IH', 'IH0', 'IH1', 'IH2', 'IY', 'IY0',
'IY1', 'IY2', 'JH', 'K', 'L', 'M', 'N', 'NG', 'OW', 'OW0',
'OW1', 'OW2', 'OY', 'OY0', 'OY1', 'OY2', 'P', 'R', 'S',
'SH', 'T', 'TH', 'UH', 'UH0', 'UH1', 'UH2', 'UW', 'UW0',
'UW1', 'UW2', 'V', 'W', 'Y', 'Z', 'ZH']
_arpabet_symbol_marker = '@'
_arpabet_symbols = [_arpabet_symbol_marker + x for x in _arpabet_symbols_raw]
# create pool of symbols
_symbols = [_pad] + list(_eos) + list(_letters) + list(_punctuation) \
+ _arpabet_symbols
# create the mapping table
# x+1 so that 0 can be reserved for other purposes
_symbol_to_index = {y: x for x, y in enumerate(_symbols)}
def symbol_num():
return len(_symbols)
def symbol2index(x):
return _symbol_to_index[x]
def index2symbol(x):
return _symbols[x]
def eos_index():
return _symbol_to_index[_eos]
#####
## Functions for text normalization
## I cannot write a full fludged text normalizer here.
## Just for place holder
#####
_whitespace_re = re.compile(r'\s+')
_number_map = {'1': 'one', '2': 'two', '3': 'three',
'4': 'four', '5': 'five', '6': 'six',
'7': 'seven', '8': 'eight', '9': 'nine', '0': 'zero'}
def text_numbers(text):
""" Place holder, just convert individual number to alphabet
"""
def _tmp(tmp_text):
if all([x in _number_map for x in tmp_text]):
return ' '.join([_number_map[x] for x in tmp_text])
else:
return tmp_text
tmp = ' '.join([_tmp(x) for x in text.split()])
if text.startswith(' '):
tmp = ' ' + tmp
return tmp
def text_case_convert(text):
""" By default, use lower case
"""
return text.lower()
def text_whitespace_convert(text):
""" Collapse all redundant white spaces
e.g., 'qweq 1231 123151' -> 'qweq 1231 123151'
"""
return re.sub(_whitespace_re, ' ', text)
def text_normalizer(text):
""" Text normalizer
In this code, only lower case conversion and white space is handled
"""
return text_whitespace_convert(text_numbers(text_case_convert(text)))
def g2poutput_process(sym_seq):
""" remove unnecessary space (inplace)
"""
punc_list = [x for x in _punctuation]
new_sym_seq = []
for idx, sym in enumerate(sym_seq):
if sym == _space:
# skip initial space
if idx == 0:
continue
# skip space before punctunation
if idx < len(sym_seq) - 1 and sym_seq[idx+1] in punc_list:
continue
# skip space after puncutation
if idx > 0 and sym_seq[idx-1] in punc_list:
continue
new_sym_seq.append(sym)
return new_sym_seq
#####
## Functions to convert symbol to index
#####
def flag_convert_symbol(symbol):
""" check whether input symbol should be converted or not
input
-----
symbol: str
output
------
bool
"""
return symbol in _symbol_to_index and symbol not in _skip_symbols
def rawtext2indices(text):
""" Look up the table and return the index for input symbol in input text
input
-----
text: str
output
------
list of indices
for example, 'text' -> [23, 16, 28, 23]
"""
output = [symbol2index(x) for x in text if flag_convert_symbol(x)]
return output
def arpabet2indices(arpa_text):
""" Look up the table and return the index for input symbol in input text
input
-----
arpa_text: str
output
------
list of indices
for example, 'AH_HH' -> [12 19]
"""
_fun_at = lambda x: _arpabet_symbol_marker + x if x != _space else x
tmp = [_fun_at(x) for x in arpa_text.split(_pad)]
output = [symbol2index(x) for x in tmp if flag_convert_symbol(x)]
return output
#####
## Main function
#####
def text2code(text, flag_append_eos=True):
""" Convert English text and ARPAbet into code symbols (int)
"""
if text.startswith(toolkit_all._curly_symbol):
# phonemic annotation, no normalization
output = arpabet2indices(text.lstrip(toolkit_all._curly_symbol))
else:
# normal text, do normalization before conversion
# text normalization
text_normalized = text_normalizer(text)
output = rawtext2indices(text_normalized)
# if necessary, add the eos symbol
if flag_append_eos:
output.append(symbol2index(_eos))
return output
def code2text(codes):
""" Convert index back to text
Unfinished. ARPAbet cannot be reverted in this function
"""
# x-1 because _symbol_to_index
txt_tmp = [index2symbol(x) for x in codes]
txt_tmp = ''.join(txt_tmp)
return text_whitespace_convert(txt_tmp.replace(_arpabet_symbol_marker, ' '))
if __name__ == "__main__":
print("Definition of text processing toolkit for English")
| 6,026 | 25.550661 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/text_process/text_io.py | #!/usr/bin/env python
"""
Simple converted to convert text string into indices
Used for text-to-speech synthesis
Based on https://github.com/fatchord/WaveRNN
"""
import os
import sys
import re
import numpy as np
from core_scripts.other_tools import display as nii_warn
from core_scripts.data_io.text_process import toolkit_all
from core_scripts.data_io.text_process import toolkit_en
from core_scripts.other_tools import str_tools as nii_str_tk
from core_scripts.data_io import conf as nii_dconf
try:
from g2p_en import G2p
g_g2p_API = G2p()
except ModuleNotFoundError:
#nii_warn.f_print("g2p_en is not found, we cannot use it", opt='warning')
pass
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
def text2code(text, flag_lang='EN'):
""" Convert text string into code indices
input
-----
text: string
flag_lang: string, 'EN': English
output
------
code_seq: list of integers
"""
code_seq = []
# parse the curly bracket
text_trunks = toolkit_all.parse_curly_bracket(text)
# parse
if flag_lang == 'EN':
# English text
for idx, text_trunk in enumerate(text_trunks):
# whether this is the last trunk
flag_eos = idx >= (len(text_trunks)-1)
code_seq += toolkit_en.text2code(text_trunk, flag_eos)
else:
# unsupporte languages
nii_warn.f_die("Error: text2code cannot handle {:s}".format(flag_lang))
# convert to numpy format
code_seq = np.array(code_seq, dtype=nii_dconf.h_dtype)
return code_seq
def code2text(codes, flag_lang='EN'):
""" Convert text string into code indices
input
-----
code_seq: numpy arrays of integers
flag_lang: string, 'EN': English
output
------
text: string
"""
# convert numpy array backto indices
codes_tmp = [int(x) for x in codes]
output_text = ''
if flag_lang == 'EN':
output_text = toolkit_en.code2text(codes_tmp)
else:
nii_warn.f_die("Error: code2text cannot handle {:s}".format(flag_lang))
return output_text
def symbol_num(flag_lang='EN'):
""" Return the number of symbols defined for one language
input
-----
flag_lange: string, 'EN': English
output
------
integer
"""
if flag_lang == 'EN':
return toolkit_en.symbol_num()
else:
nii_warn.f_die("Error: symbol_num cannot handle {:s}".format(flag_lang))
return 0
def eos_index(flag_lang='EN'):
""" Return the index for end of sentence symbol
input
-----
flag_lange: string, 'EN': English
output
------
integer
"""
if flag_lang == 'EN':
return toolkit_en.eos_index()
else:
nii_warn.f_die("Error: eos_index cannot handle {:s}".format(flag_lang))
return 0
def textloader(file_path, flag_lang='EN', g2p_tool=None):
""" Load text and return the sybmol sequences
input
-----
file_path: string, absolute path to the text file
flag_lang: string, 'EN' by default, the language option to process text
output
------
output: np.array of shape (L), where L is the number of chars
"""
# load lines and chop '\n', join into one line
text_buffer = [nii_str_tk.string_chop(x) for x in open(file_path, 'r')]
text_buffer = ' '.join(text_buffer)
# convert to indices
if g2p_tool is None:
return text2code(text_buffer, flag_lang)
else:
return g2p2code(text_buffer, flag_lang)
def g2p2code(text, flag_lang='EN'):
""" Load text, do g2p, produce the indices
The g2p tool is based on https://github.com/Kyubyong/g2p
input
-----
text: string
flag_lang: string, 'EN': English
output
------
code_seq: list of integers
"""
# precheck whether input has ARPABet symbol
text_trunks = toolkit_all.parse_curly_bracket(text)
if len(text_trunks) > 1 or text_trunks[0][0] == toolkit_all._curly_symbol:
# unsupporte languages
nii_warn.f_die("Error: g2p2code don't process: {:s}".format(text))
if flag_lang == 'EN':
# g2p
try:
output = g_g2p_API(text)
except NameError:
nii_warn.f_die("G2p tool not found. Please install g2p_en")
# remove the trailing space when there is
output = toolkit_en.g2poutput_process(output)
# pack it inside '{ }' to be compatible with the API in toolkit_en
output = '{' + '_'.join(output) + '}'
#
code_seq = text2code(output, flag_lang)
else:
# unsupporte languages
nii_warn.f_die("Error: text2code cannot handle {:s}".format(flag_lang))
return code_seq
if __name__ == "__main__":
print("Definition of text2code tools")
text = 'hello we are {AY2_AY2_ _AY2_AY2} the same 123'
indices = text2code(text)
text2 = code2text(indices)
print(text)
print(indices)
print(text2)
print(code2text(textloader('./tmp.txt')))
| 5,102 | 25.169231 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/data_io/text_process/toolkit_all.py | #!/usr/bin/env python
"""
Simple text processer for all languages
Based on https://github.com/fatchord/WaveRNN
"""
import os
import sys
import re
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#####
## Parse the curly bracket
#####
# from https://github.com/fatchord/WaveRNN
_curly_re = re.compile(r'(.*?)\{(.+?)\}(.*)')
# symbol to indicate phonemic annotation
_curly_symbol = '{'
def parse_curly_bracket(text):
""" Prase the text based on curly brackets
Inspired by https://github.com/fatchord/WaveRNN: when input text
is mixed with raw text and phonemic annotation, the {} pair indicates
the phonemic part
input
-----
text: str
output
------
text_list: list of str
For example, 'text {AH II} test' -> ['text ', '{AH II', ' test']
"""
text_list = []
text_tmp = text
while len(text_tmp):
re_matched = _curly_re.match(text_tmp)
if re_matched:
# e.g., 'text {AH II} test'
# group(1), group(2) -> ['text ', '{AH II']
text_list.append(re_matched.group(1))
# the '{' in front of AH II should be trimmed off when parsing the
# text string later
text_list.append(_curly_symbol + re_matched.group(2))
# group(3) -> ' test'
text_tmp = re_matched.group(3)
else:
text_list.append(text_tmp)
break
# exclude the empty truck, for example in ['', '{AH II']
text_list_filtered = [x for x in text_list if len(x) > 0]
return text_list_filtered
if __name__ == "__main__":
print("Definition of text processing tools for all languages")
| 1,725 | 24.382353 | 78 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/op_manager/lr_scheduler.py | #!/usr/bin/env python
"""
op_manager
A simple wrapper to create lr scheduler
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import torch
import torch.optim as torch_optim
import torch.optim.lr_scheduler as torch_optim_steplr
import core_scripts.other_tools.display as nii_warn
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
class LRScheduler():
""" Wrapper over different types of learning rate Scheduler
"""
def __init__(self, optimizer, args):
# learning rate decay
self.lr_decay = args.lr_decay_factor
# lr scheduler type
# please check arg_parse.py for the number ID
self.lr_scheduler_type = args.lr_scheduler_type
# patentience for ReduceLROnPlateau
self.lr_patience = args.lr_patience
# step size for stepLR
self.lr_stepLR_size = args.lr_steplr_size
if self.lr_decay > 0:
if self.lr_scheduler_type == 1:
# StepLR
self.lr_scheduler = torch.optim.lr_scheduler.StepLR(
optimizer = optimizer, step_size = self.lr_stepLR_size,
gamma = self.lr_decay)
elif self.lr_scheduler_type == 2:
# StepLR
self.lr_scheduler = torch.optim.lr_scheduler.ExponentialLR(
optimizer = optimizer, gamma = self.lr_decay)
elif self.lr_scheduler_type == 3:
# Cosine
self.lr_shceduler = torch.optim.lr_scheduler.CosineAnnealingWarmRestarts(
optimizer = optimizer, T_0 = self.lr_patience)
else:
# by default, ReduceLROnPlateau
self.lr_scheduler = torch_optim_steplr.ReduceLROnPlateau(
optimizer=optimizer, factor=self.lr_decay,
patience=self.lr_patience)
self.flag = True
else:
self.lr_scheduler = None
self.flag =False
return
def f_valid(self):
""" Whether this LR scheduler is valid
"""
return self.flag
def f_print_info(self):
""" Print information about the LR scheduler
"""
if not self.flag:
mes = ""
else:
if self.lr_scheduler_type == 1:
mes = "\n LR scheduler, StepLR [gamma %f, step %d]" % (
self.lr_decay, self.lr_stepLR_size)
elif self.lr_scheduler_type == 2:
mes = "\n LR scheduler, ExponentialLR [gamma %f]" % (
self.lr_decay)
else:
mes = "\n LR scheduler, ReduceLROnPlateau "
mes += "[decay %f, patience %d]" % (
self.lr_decay, self.lr_patience)
return mes
def f_last_lr(self):
""" Return the last lr
"""
if self.f_valid():
if hasattr(self.lr_scheduler, "get_last_lr"):
return self.lr_scheduler.get_last_lr()
else:
return self.lr_scheduler._last_lr
else:
return []
def f_load_state_dict(self, state):
if self.f_valid():
self.lr_scheduler.load_state_dict(state)
return
def f_state_dict(self):
if self.f_valid():
return self.lr_scheduler.state_dict()
else:
return None
def f_step(self, loss_val):
if self.f_valid():
if self.lr_scheduler_type == 1:
self.lr_scheduler.step()
elif self.lr_scheduler_type == 2:
self.lr_scheduler.step()
else:
self.lr_scheduler.step(loss_val)
return
def f_allow_early_stopping(self):
if self.f_valid():
if self.lr_scheduler_type == 1:
return True
elif self.lr_scheduler_type == 2:
return True
else:
# ReduceLROnPlateau no need to use early stopping
return False
else:
return True
if __name__ == "__main__":
print("Definition of lr_scheduler")
| 4,244 | 29.106383 | 89 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/op_manager/op_manager.py | #!/usr/bin/env python
"""
op_manager
A simple wrapper to create optimizer
"""
from __future__ import absolute_import
import os
import sys
import numpy as np
import torch
import torch.optim as torch_optim
import torch.optim.lr_scheduler as torch_optim_steplr
import core_scripts.other_tools.list_tools as nii_list_tools
import core_scripts.other_tools.display as nii_warn
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.op_manager.conf as nii_op_config
import core_scripts.op_manager.op_process_monitor as nii_op_monitor
import core_scripts.op_manager.lr_scheduler as nii_lr_scheduler
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
class OptimizerWrapper():
""" Wrapper over optimizer
"""
def __init__(self, model, args):
""" Initialize an optimizer over model.parameters()
"""
# check valildity of model
if not hasattr(model, "parameters"):
nii_warn.f_print("model is not torch.nn", "error")
nii_warn.f_die("Error in creating OptimizerWrapper")
# set optimizer type
self.op_flag = args.optimizer
self.lr = args.lr
self.l2_penalty = args.l2_penalty
# grad clip norm is directly added in nn_manager
self.grad_clip_norm = args.grad_clip_norm
# create optimizer
if self.op_flag == "Adam":
if self.l2_penalty > 0:
self.optimizer = torch_optim.Adam(model.parameters(),
lr=self.lr,
weight_decay=self.l2_penalty)
else:
self.optimizer = torch_optim.Adam(model.parameters(),
lr=self.lr)
elif self.op_flag == 'AdamW':
if self.l2_penalty > 0:
self.optimizer = torch_optim.AdamW(model.parameters(),
lr=self.lr,
weight_decay=self.l2_penalty)
else:
self.optimizer = torch_optim.AdamW(model.parameters(),
lr=self.lr)
else:
nii_warn.f_print("%s not availabel" % (self.op_flag), "error")
nii_warn.f_die("Please add optimizer to op_manager")
# number of epochs
if args.active_learning_cycle_num:
# for active learning, epochs * number_of_cycle
self.epochs = args.epochs * np.abs(args.active_learning_cycle_num)
else:
self.epochs = args.epochs
self.no_best_epochs = args.no_best_epochs
# lr scheduler
self.lr_scheduler = nii_lr_scheduler.LRScheduler(self.optimizer, args)
return
def print_info(self):
""" print message of optimizer
"""
mes = "Optimizer:\n Type: {} ".format(self.op_flag)
mes += "\n Learing rate: {:2.6f}".format(self.lr)
mes += "\n Epochs: {:d}".format(self.epochs)
mes += "\n No-best-epochs: {:d}".format(self.no_best_epochs)
if self.lr_scheduler.f_valid():
mes += self.lr_scheduler.f_print_info()
if self.l2_penalty > 0:
mes += "\n With weight penalty {:f}".format(self.l2_penalty)
if self.grad_clip_norm > 0:
mes += "\n With grad clip norm {:f}".format(self.grad_clip_norm)
nii_warn.f_print_message(mes)
def get_epoch_num(self):
return self.epochs
def get_no_best_epoch_num(self):
return self.no_best_epochs
def get_lr_info(self):
if self.lr_scheduler.f_valid():
# no way to look into the updated lr rather than using _last_lr
tmp = ''
for updated_lr in self.lr_scheduler.f_last_lr():
if np.abs(self.lr - updated_lr) > 0.0000001:
tmp += "{:.2e} ".format(updated_lr)
if tmp:
tmp = " LR -> " + tmp
return tmp
else:
return None
if __name__ == "__main__":
print("Optimizer Wrapper")
| 4,193 | 33.661157 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/op_manager/op_display_tools.py | #!/usr/bin/env python
"""
op_display_tools
Functions to print/display the training/inference information
"""
from __future__ import print_function
import os
import sys
import numpy as np
import core_scripts.other_tools.display as nii_display
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def print_gen_info(seq_name, time, sample_idx=None):
""" Print the information during inference
"""
if sample_idx is None:
mes = "Generating {}, time cost: {:.3f}s".format(seq_name, time)
else:
mes = "Generating {:d}, {}, time cost: {:.3f}s".format(
sample_idx, seq_name, time)
nii_display.f_print_message(mes)
return mes + '\n'
def _print_loss(loss_array):
mes = ""
if loss_array.shape[0] == 1:
mes += "%12.4f " % (loss_array[0])
else:
mes = []
for data in loss_array:
if data > 1:
mes.append('%6.2f' % (data))
else:
mes.append('%1.4f' % (data))
mes = ' '.join(mes)
mes += "| "
return mes
def print_train_info(epoch, time_tr, loss_tr, time_val,
loss_val, isbest, lr_info):
""" Print the information during training
"""
mes = "{:>7d} | ".format(epoch)
mes = mes + "{:>12.1f} | ".format(time_tr + time_val)
mes += _print_loss(loss_tr)
mes += _print_loss(loss_val)
#mes = mes + "{:>12.4f} | ".format(loss_tr)
#mes = mes + "{:>12.4f} | ".format(loss_val)
if isbest:
mes = mes + "{:>5s}".format("yes")
else:
mes = mes + "{:>5s}".format("no")
if lr_info:
mes = mes + ' ' + lr_info
nii_display.f_print_message(mes, flush=True)
return mes + '\n'
def print_log_head():
""" Print the head information
"""
nii_display.f_print_message("{:->62s}".format(""))
mes = "{:>7s} | ".format("Epoch")
mes = mes + "{:>12s} | ".format("Duration(s)")
mes = mes + "{:>12s} | ".format("Train loss")
mes = mes + "{:>12s} | ".format("Dev loss")
mes = mes + "{:>5s}".format("Best")
nii_display.f_print_message(mes)
nii_display.f_print_message("{:->62s}".format(""), flush=True)
return mes + '\n'
def print_log_tail():
""" Print the tail line
"""
nii_display.f_print_message("{:->62s}".format(""))
return
if __name__ == "__main__":
print("Op_display_tools")
| 2,409 | 26.701149 | 72 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/op_manager/op_process_monitor.py | #!/usr/bin/env python
"""
op_process_monitor
A simple monitor on training / inference process
"""
from __future__ import print_function
import os
import sys
import numpy as np
import core_scripts.other_tools.display as nii_display
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
class Monitor():
""" A monitor to log down all the training /
inference informations
"""
def __init__(self, epoch_num, seq_num):
self.loss_mat = np.zeros([epoch_num, seq_num, 1])
self.time_mat = np.zeros([epoch_num, seq_num])
self.seq_names = {}
self.epoch_num = epoch_num
self.seq_num = seq_num
self.cur_epoch = 0
self.best_error = None
self.best_epoch = None
self.loss_flag = None
def clear(self):
self.loss_mat.fill(0)
self.time_mat.fill(0)
self.cur_epoch = 0
self.seq_names = {}
self.best_error = None
self.best_epoch = None
self.loss_flag = None
def get_state_dic(self):
""" create a dictionary to save process
"""
state_dic = {}
state_dic['loss_mat'] = self.loss_mat
state_dic['time_mat'] = self.time_mat
state_dic['epoch_num'] = self.epoch_num
state_dic['seq_num'] = self.seq_num
state_dic['cur_epoch'] = self.cur_epoch
state_dic['best_error'] = self.best_error
state_dic['best_epoch'] = self.best_epoch
state_dic['loss_flag'] = self.loss_flag
# no need to save self.seq_names
return state_dic
def load_state_dic(self, state_dic):
""" resume training, load the information
"""
try:
if self.seq_num != state_dic['seq_num']:
nii_display.f_print("Number of training data changed", 'error')
mes = "Please make sure that you are "
mes += "using the same train/dev sets as before."
mes += "\nOr\nUse --ignore-training-history-in-trained-model"
mes += " to ignore the training log from previous settings"
nii_display.f_print(mes)
nii_display.f_die("Fail to load pretrained model")
if self.epoch_num == state_dic['epoch_num']:
self.loss_mat = state_dic['loss_mat']
self.time_mat = state_dic['time_mat']
else:
# if training epoch is increased, resize the shape
tmp_loss_mat = state_dic['loss_mat']
self.loss_mat = np.resize(
self.loss_mat,
[self.epoch_num, self.seq_num, tmp_loss_mat.shape[2]])
self.loss_mat[0:tmp_loss_mat.shape[0]] = tmp_loss_mat
self.time_mat[0:tmp_loss_mat.shape[0]] = state_dic['time_mat']
self.seq_num = state_dic['seq_num']
# since the saved cur_epoch has been finished
self.cur_epoch = state_dic['cur_epoch'] + 1
self.best_error = state_dic['best_error']
self.best_epoch = state_dic['best_epoch']
self.loss_flag = state_dic['loss_flag']
self.seq_names = {}
except KeyError:
nii_display.f_die("Invalid op_process_monitor state_dic")
def print_error_for_batch(self, cnt_idx, seq_idx, epoch_idx):
"""print error for each mini-batch
"""
try:
t_1 = self.loss_mat[epoch_idx, seq_idx]
t_2 = self.time_mat[epoch_idx, seq_idx]
mes = "{}, ".format(self.seq_names[seq_idx])
mes += "{:d}/{:d}, ".format(cnt_idx+1, self.seq_num)
mes += "Time: {:.6f}s".format(t_2)
for loss_indi in t_1:
mes += ", Loss: {:.6f}".format(loss_indi)
nii_display.f_eprint(mes, flush=True)
except IndexError:
nii_display.f_die("Unknown sample index in Monitor")
except KeyError:
nii_display.f_die("Unknown sample index in Monitor")
return
def get_time(self, epoch):
return np.sum(self.time_mat[epoch, :])
def get_loss(self, epoch):
# return a array
return np.mean(self.loss_mat[epoch, :], axis=0)
def get_loss_for_learning_stopping(self, epoch_idx):
# get_loss return the loss for displaying, not all the loss
# are used as criterion for early stopping or learning rate adjusting
return self._get_loss_for_learning_stopping(epoch_idx)
def get_epoch(self):
return self.cur_epoch
def get_max_epoch(self):
return self.epoch_num
def _get_loss_for_learning_stopping(self, epoch_idx):
# compute the average loss values
if epoch_idx > self.cur_epoch:
nii_display.f_print("To find loss for future epochs", 'error')
nii_display.f_die("Op_process_monitor: error")
if epoch_idx < 0:
nii_display.f_print("To find loss for NULL epoch", 'error')
nii_display.f_die("Op_process_monitor: error")
loss_this = np.sum(self.loss_mat[epoch_idx, :, :], axis=0)
# compute only part of the loss for early stopping when necessary
if self.loss_flag is not None:
loss_this = np.sum(loss_this * self.loss_flag)
else:
loss_this = np.sum(loss_this)
return loss_this
def print_error_for_epoch(self, epoch):
loss = np.mean(self.loss_mat[epoch, :])
time_sum = np.sum(self.time_mat[epoch, :])
mes = "Epoch {:d}: ".format(epoch)
mes += 'Time: {:.6f}, Loss: {:.6f}'.format(time_sum, loss)
nii_display.f_print_message(mes)
return "{}\n".format(mes)
def log_loss(self, loss, loss_flag, time_cost, seq_info, seq_idx, \
epoch_idx):
""" Log down the loss
"""
self.seq_names[seq_idx] = seq_info
if self.loss_mat.shape[-1] != len(loss):
self.loss_mat = np.resize(self.loss_mat,
[self.loss_mat.shape[0],
self.loss_mat.shape[1],
len(loss)])
self.loss_flag = loss_flag
self.loss_mat[epoch_idx, seq_idx, :] = loss
self.time_mat[epoch_idx, seq_idx] = time_cost
self.cur_epoch = epoch_idx
return
def log_epoch(self, epoch_idx):
self.cur_epoch = epoch_idx
return
def is_new_best(self):
"""
check whether epoch is the new_best
"""
loss_this = self._get_loss_for_learning_stopping(self.cur_epoch)
if self.best_error is None or loss_this < self.best_error:
self.best_error = loss_this
self.best_epoch = self.cur_epoch
return True
else:
return False
def should_early_stop(self, no_best_epoch_num):
"""
check whether to stop training early
"""
if (self.cur_epoch - self.best_epoch) >= no_best_epoch_num:
#
#tmp = []
#for idx in np.arange(no_best_epoch_num+1):
# tmp.append(self._get_loss_for_learning_stopping(
# self.cur_epoch - idx))
#if np.sum(np.diff(tmp) < 0) >= no_best_epoch_num:
# return True
#else:
# return False
return True
else:
return False
if __name__ == "__main__":
print("Definition of nn_process_monitor")
| 7,550 | 35.302885 | 79 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/op_manager/conf.py | #!/usr/bin/env python
"""
config.py
Default configurations for optimizer
Note: these default configs only apply when input argument
doesn't specify the parameters
"""
from __future__ import absolute_import
import os
import sys
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
| 332 | 15.65 | 58 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/nn_manager/nn_manager_GAN.py | #!/usr/bin/env python
"""
nn_manager_gan
A simple wrapper to run the training / testing process for GAN
"""
from __future__ import print_function
import time
import datetime
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.display as nii_display
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.op_manager.op_process_monitor as nii_monitor
import core_scripts.op_manager.op_display_tools as nii_op_display_tk
import core_scripts.nn_manager.nn_manager_tools as nii_nn_tools
import core_scripts.nn_manager.nn_manager_conf as nii_nn_manage_conf
import core_scripts.other_tools.debug as nii_debug
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#############################################################
def f_run_one_epoch_GAN(
args, pt_model_G, pt_model_D,
loss_wrapper, \
device, monitor, \
data_loader, epoch_idx,
optimizer_G = None, optimizer_D = None, \
target_norm_method = None):
"""
f_run_one_epoch_GAN:
run one poech over the dataset (for training or validation sets)
Args:
args: from argpase
pt_model_G: pytorch model (torch.nn.Module) generator
pt_model_D: pytorch model (torch.nn.Module) discriminator
loss_wrapper: a wrapper over loss function
loss_wrapper.compute(generated, target)
device: torch.device("cuda") or torch.device("cpu")
monitor: defined in op_procfess_monitor.py
data_loader: pytorch DataLoader.
epoch_idx: int, index of the current epoch
optimizer_G: torch optimizer or None, for generator
optimizer_D: torch optimizer or None, for discriminator
if None, the back propgation will be skipped
(for developlement set)
target_norm_method: method to normalize target data
(by default, use pt_model.normalize_target)
"""
# timer
start_time = time.time()
# loop over samples
for data_idx, (data_in, data_tar, data_info, idx_orig) in \
enumerate(data_loader):
#############
# prepare
#############
# send data to device
if optimizer_G is not None:
optimizer_G.zero_grad()
if optimizer_D is not None:
optimizer_D.zero_grad()
# Put data to devices
# to device (we assume noise will be generated by the model itself)
# here we only provide external condition
data_in = data_in.to(device, dtype=nii_dconf.d_dtype)
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
else:
nii_display.f_die("target data is required")
############################
# 1. Generate the sample
############################
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder & autoregressive model
if isinstance(data_tar, torch.Tensor):
data_tar_tm = data_tar.to(device, dtype=nii_dconf.d_dtype)
if args.model_forward_with_file_name:
data_gen = pt_model_G(data_in, data_tar_tm, data_info)
else:
data_gen = pt_model_G(data_in, data_tar_tm)
else:
nii_display.f_print("--model-forward-with-target is set")
nii_display.f_die("but data_tar is not loaded")
else:
if args.model_forward_with_file_name:
# specifcal case when model.forward requires data_info
data_gen = pt_model_G(data_in, data_info)
else:
# normal case for model.forward(input)
data_gen = pt_model_G(data_in)
############################
# 2. update discrminator
############################
pt_model_D.zero_grad()
if optimizer_D is not None:
optimizer_D.zero_grad()
# compute discriminator loss
if hasattr(pt_model_D, 'loss_for_D'):
errD = pt_model_D.loss_for_D(data_tar, data_gen.detach(), data_in)
else:
# for compatiblity
# get the discrminator's outputs for real and fake data
# data_gen.detach() is required
# https://github.com/pytorch/examples/issues/116
# https://stackoverflow.com/questions/46774641/
d_out_fake = pt_model_D(data_gen.detach(), data_in)
d_out_real = pt_model_D(data_tar, data_in)
errD_real = loss_wrapper.compute_gan_D_real(d_out_real)
errD_fake = loss_wrapper.compute_gan_D_fake(d_out_fake)
errD = errD_real + errD_fake
# update discriminator weight
if optimizer_D is not None:
errD.backward()
optimizer_D.step()
############################
# 3. update generator
############################
pt_model_G.zero_grad()
if optimizer_G is not None:
optimizer_G.zero_grad()
# compute the loss for generator
if hasattr(pt_model_D, 'loss_for_G'):
errG = pt_model_D.loss_for_G(data_tar, data_gen, data_in)
if hasattr(pt_model_G, 'loss_aux'):
errG += pt_model_G.loss_aux(data_tar, data_gen, data_in)
else:
# get the discrminator's outputs again
d_out_fake_for_G = pt_model_D(data_gen, data_in)
d_out_real_for_G = pt_model_D(data_tar, data_in)
# for compatibility
errG_gan = loss_wrapper.compute_gan_G(d_out_fake_for_G)
# if defined, calculate auxilliart loss
if hasattr(loss_wrapper, "compute_aux"):
errG_aux = loss_wrapper.compute_aux(data_gen, data_tar)
else:
errG_aux = torch.zeros_like(errG_gan)
# if defined, calculate feat-matching loss
if hasattr(loss_wrapper, "compute_feat_match"):
errG_feat = loss_wrapper.compute_feat_match(
d_out_real, d_out_fake_for_G)
else:
errG_feat = torch.zeros_like(errG_gan)
# sum loss for generator
errG = errG_gan + errG_aux + errG_feat
if optimizer_G is not None:
errG.backward()
optimizer_G.step()
# construct the loss for logging and early stopping
# only use errG_aux for early-stopping
#loss_computed = [
# [errG_aux, errD_real, errD_fake, errG_gan, errG_feat],
# [True, False, False, False, False]]
loss_computed = [[errG, errD], [True, True]]
# to handle cases where there are multiple loss functions
_, loss_vals, loss_flags = nii_nn_tools.f_process_loss(loss_computed)
# save the training process information to the monitor
end_time = time.time()
batchsize = len(data_info)
for idx, data_seq_info in enumerate(data_info):
# loss_value is supposed to be the average loss value
# over samples in the the batch, thus, just loss_value
# rather loss_value / batchsize
monitor.log_loss(loss_vals, loss_flags, \
(end_time-start_time) / batchsize, \
data_seq_info, idx_orig.numpy()[idx], \
epoch_idx)
# print infor for one sentence
if args.verbose == 1:
# here we use args.batch_size because len(data_info)
# may be < args.batch_size.
monitor.print_error_for_batch(
data_idx * args.batch_size + idx,
idx_orig.numpy()[idx],
epoch_idx)
#
# start the timer for a new batch
start_time = time.time()
# Save intermediate model for every n mini-batches (optional).
# Note that if we re-start trainining with this intermediate model,
# the data will start from the 1st sample, not the one where we stopped
if args.save_model_every_n_minibatches > 0 \
and (data_idx+1) % args.save_model_every_n_minibatches == 0 \
and optimizer is not None and data_idx > 0:
cp_names = nii_nn_manage_conf.CheckPointKey()
for pt_model, optimizer, model_tag in \
zip([pt_model_G, pt_model_D], [optimizer_G, optimizer_D],
model_tags):
tmp_model_name = nii_nn_tools.f_save_epoch_name(
args, epoch_idx,
'_{:05d}_{:s}'.format(data_idx+1, model_tag))
# save
tmp_dic = {
cp_names.state_dict : pt_model.state_dict(),
cp_names.optimizer : optimizer.state_dict(),
}
torch.save(tmp_dic, tmp_model_name)
# If debug mode is used, only run a specified number of mini-batches
if args.debug_batch_num > 0 and data_idx >= (args.debug_batch_num - 1):
nii_display.f_print("Debug mode is on. This epoch is finished")
break
# lopp done
return
def f_run_one_epoch_WGAN(
args, pt_model_G, pt_model_D,
loss_wrapper, \
device, monitor, \
data_loader, epoch_idx,
optimizer_G = None, optimizer_D = None, \
target_norm_method = None):
"""
f_run_one_epoch_WGAN:
similar to f_run_one_epoch_GAN, but for WGAN
"""
# timer
start_time = time.time()
# This should be moved to model definition
# number of critic (default 5)
num_critic = 5
# clip value
wgan_clamp = 0.01
# loop over samples
for data_idx, (data_in, data_tar, data_info, idx_orig) in \
enumerate(data_loader):
# send data to device
if optimizer_G is not None:
optimizer_G.zero_grad()
if optimizer_D is not None:
optimizer_D.zero_grad()
# prepare data
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
# there is no way to normalize the data inside loss
# thus, do normalization here
if target_norm_method is None:
normed_target = pt_model_G.normalize_target(data_tar)
else:
normed_target = target_norm_method(data_tar)
else:
nii_display.f_die("target data is required")
# to device (we assume noise will be generated by the model itself)
# here we only provide external condition
data_in = data_in.to(device, dtype=nii_dconf.d_dtype)
############################
# Update Discriminator
############################
# train with real
pt_model_D.zero_grad()
d_out_real = pt_model_D(data_tar)
errD_real = loss_wrapper.compute_gan_D_real(d_out_real)
if optimizer_D is not None:
errD_real.backward()
d_out_real_mean = d_out_real.mean()
# train with fake
# generate sample
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder & autoregressive model
if isinstance(data_tar, torch.Tensor):
data_tar_tm = data_tar.to(device, dtype=nii_dconf.d_dtype)
if args.model_forward_with_file_name:
data_gen = pt_model_G(data_in, data_tar_tm, data_info)
else:
data_gen = pt_model_G(data_in, data_tar_tm)
else:
nii_display.f_print("--model-forward-with-target is set")
nii_display.f_die("but data_tar is not loaded")
else:
if args.model_forward_with_file_name:
# specifcal case when model.forward requires data_info
data_gen = pt_model_G(data_in, data_info)
else:
# normal case for model.forward(input)
data_gen = pt_model_G(data_in)
# data_gen.detach() is required
# https://github.com/pytorch/examples/issues/116
d_out_fake = pt_model_D(data_gen.detach())
errD_fake = loss_wrapper.compute_gan_D_fake(d_out_fake)
if optimizer_D is not None:
errD_fake.backward()
d_out_fake_mean = d_out_fake.mean()
errD = errD_real + errD_fake
if optimizer_D is not None:
optimizer_D.step()
# clip weights of discriminator
for p in pt_model_D.parameters():
p.data.clamp_(-wgan_clamp, wgan_clamp)
############################
# Update Generator
############################
pt_model_G.zero_grad()
d_out_fake_for_G = pt_model_D(data_gen)
errG_gan = loss_wrapper.compute_gan_G(d_out_fake_for_G)
errG_aux = loss_wrapper.compute_aux(data_gen, data_tar)
errG = errG_gan + errG_aux
# only update after num_crictic iterations on discriminator
if data_idx % num_critic == 0 and optimizer_G is not None:
errG.backward()
optimizer_G.step()
d_out_fake_for_G_mean = d_out_fake_for_G.mean()
# construct the loss for logging and early stopping
# only use errG_aux for early-stopping
loss_computed = [[errG_aux, errG_gan, errD_real, errD_fake,
d_out_real_mean, d_out_fake_mean,
d_out_fake_for_G_mean],
[True, False, False, False, False, False, False]]
# to handle cases where there are multiple loss functions
loss, loss_vals, loss_flags = nii_nn_tools.f_process_loss(loss_computed)
# save the training process information to the monitor
end_time = time.time()
batchsize = len(data_info)
for idx, data_seq_info in enumerate(data_info):
# loss_value is supposed to be the average loss value
# over samples in the the batch, thus, just loss_value
# rather loss_value / batchsize
monitor.log_loss(loss_vals, loss_flags, \
(end_time-start_time) / batchsize, \
data_seq_info, idx_orig.numpy()[idx], \
epoch_idx)
# print infor for one sentence
if args.verbose == 1:
monitor.print_error_for_batch(data_idx*batchsize + idx,\
idx_orig.numpy()[idx], \
epoch_idx)
#
# start the timer for a new batch
start_time = time.time()
# lopp done
return
def f_train_wrapper_GAN(
args, pt_model_G, pt_model_D, loss_wrapper, device, \
optimizer_G_wrapper, optimizer_D_wrapper, \
train_dataset_wrapper, \
val_dataset_wrapper = None, \
checkpoint_G = None, checkpoint_D = None):
"""
f_train_wrapper_GAN(
args, pt_model_G, pt_model_D, loss_wrapper, device,
optimizer_G_wrapper, optimizer_D_wrapper,
train_dataset_wrapper, val_dataset_wrapper = None,
check_point = None):
A wrapper to run the training process
Args:
args: argument information given by argpase
pt_model_G: generator, pytorch model (torch.nn.Module)
pt_model_D: discriminator, pytorch model (torch.nn.Module)
loss_wrapper: a wrapper over loss functions
loss_wrapper.compute_D_real(discriminator_output)
loss_wrapper.compute_D_fake(discriminator_output)
loss_wrapper.compute_G(discriminator_output)
loss_wrapper.compute_G(fake, real)
device: torch.device("cuda") or torch.device("cpu")
optimizer_G_wrapper:
a optimizer wrapper for generator (defined in op_manager.py)
optimizer_D_wrapper:
a optimizer wrapper for discriminator (defined in op_manager.py)
train_dataset_wrapper:
a wrapper over training data set (data_io/default_data_io.py)
train_dataset_wrapper.get_loader() returns torch.DataSetLoader
val_dataset_wrapper:
a wrapper over validation data set (data_io/default_data_io.py)
it can None.
checkpoint_G:
a check_point that stores every thing to resume training
checkpoint_D:
a check_point that stores every thing to resume training
"""
nii_display.f_print_w_date("Start model training")
##############
## Preparation
##############
# get the optimizers for Generator and Discrminators
optimizer_G_wrapper.print_info()
optimizer_D_wrapper.print_info()
optimizer_G = optimizer_G_wrapper.optimizer
optimizer_D = optimizer_D_wrapper.optimizer
epoch_num = optimizer_G_wrapper.get_epoch_num()
no_best_epoch_num = optimizer_G_wrapper.get_no_best_epoch_num()
# get data loader for training set
train_dataset_wrapper.print_info()
train_data_loader = train_dataset_wrapper.get_loader()
train_seq_num = train_dataset_wrapper.get_seq_num()
# get the training process monitor
monitor_trn = nii_monitor.Monitor(epoch_num, train_seq_num)
# if validation data is provided, get data loader for val set
if val_dataset_wrapper is not None:
val_dataset_wrapper.print_info()
val_data_loader = val_dataset_wrapper.get_loader()
val_seq_num = val_dataset_wrapper.get_seq_num()
monitor_val = nii_monitor.Monitor(epoch_num, val_seq_num)
else:
monitor_val = None
# training log information
train_log = ''
model_tags = ["_G", "_D"]
# prepare for DataParallism if available
# pytorch.org/tutorials/beginner/blitz/data_parallel_tutorial.html
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
nii_display.f_die("data_parallel not implemented for GAN")
else:
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
flag_multi_device = False
normtarget_f = None
# put models to device
pt_model_G.to(device, dtype=nii_dconf.d_dtype)
pt_model_D.to(device, dtype=nii_dconf.d_dtype)
# print the network
# check the definition of GAN generator. It must a valid model definition
nii_display.f_print("Setup generator")
nii_nn_tools.f_model_show(pt_model_G, model_type='GAN')
# check the definition of GAN discrminator. It is free to define
nii_display.f_print("Setup discriminator")
nii_nn_tools.f_model_show(pt_model_D,
do_model_def_check=False, model_type='GAN')
# check the loss function
if loss_wrapper is not None:
mes = "Separate Loss() in model.py is not recommended.\n"
mes += "It is better to define loss in generator and discrminator.\n"
mes += " ModelGenerator.loss_aux: for auxialliary loss of generator\n"
mes += " ModelDiscriminator.loss_for_G: loss for generator\n"
mes += " ModelDiscriminator.loss_for_D: loss for discrminator\n"
mes += "Each loss function should have a signature like: \n"
mes += " loss = loss_func(natural_data, generated_data, condition)"
nii_display.f_print(mes)
nii_nn_tools.f_loss_show(loss_wrapper, model_type='GAN')
###############################
## Resume training if necessary
###############################
# resume training or initialize the model if necessary
cp_names = nii_nn_manage_conf.CheckPointKey()
if checkpoint_G is not None or checkpoint_D is not None:
for checkpoint, optimizer, pt_model, model_name in \
zip([checkpoint_G, checkpoint_D], [optimizer_G, optimizer_D],
[pt_model_G, pt_model_D], ["Generator", "Discriminator"]):
nii_display.f_print("For %s" % (model_name))
if type(checkpoint) is dict:
# checkpoint
# load model parameter and optimizer state
if cp_names.state_dict in checkpoint:
# wrap the state_dic in f_state_dict_wrapper
# in case the model is saved when DataParallel is on
pt_model.load_state_dict(
nii_nn_tools.f_state_dict_wrapper(
checkpoint[cp_names.state_dict],
flag_multi_device))
# load optimizer state
if cp_names.optimizer in checkpoint:
optimizer.load_state_dict(checkpoint[cp_names.optimizer])
# optionally, load training history
if not args.ignore_training_history_in_trained_model:
#nii_display.f_print("Load ")
if cp_names.trnlog in checkpoint:
monitor_trn.load_state_dic(
checkpoint[cp_names.trnlog])
if cp_names.vallog in checkpoint and monitor_val:
monitor_val.load_state_dic(
checkpoint[cp_names.vallog])
if cp_names.info in checkpoint:
train_log = checkpoint[cp_names.info]
nii_display.f_print("Load check point, resume training")
else:
nii_display.f_print("Load pretrained model and optimizer")
elif checkpoint is not None:
# only model status
#pt_model.load_state_dict(checkpoint)
pt_model.load_state_dict(
nii_nn_tools.f_state_dict_wrapper(
checkpoint, flag_multi_device))
nii_display.f_print("Load pretrained model")
else:
nii_display.f_print("No pretrained model")
# done for resume training
######################
### User defined setup
######################
# Not implemented yet
######################
### Start training
######################
# other variables
flag_early_stopped = False
start_epoch = monitor_trn.get_epoch()
epoch_num = monitor_trn.get_max_epoch()
# select one wrapper, based on the flag in loss definition
if hasattr(loss_wrapper, "flag_wgan") and loss_wrapper.flag_wgan:
f_wrapper_gan_one_epoch = f_run_one_epoch_WGAN
else:
f_wrapper_gan_one_epoch = f_run_one_epoch_GAN
# print
_ = nii_op_display_tk.print_log_head()
nii_display.f_print_message(train_log, flush=True, end='')
# loop over multiple epochs
for epoch_idx in range(start_epoch, epoch_num):
# training one epoch
pt_model_D.train()
pt_model_G.train()
f_wrapper_gan_one_epoch(
args, pt_model_G, pt_model_D,
loss_wrapper, device, \
monitor_trn, train_data_loader, \
epoch_idx, optimizer_G, optimizer_D,
normtarget_f)
time_trn = monitor_trn.get_time(epoch_idx)
loss_trn = monitor_trn.get_loss(epoch_idx)
# if necessary, do validataion
if val_dataset_wrapper is not None:
# set eval() if necessary
if args.eval_mode_for_validation:
pt_model_G.eval()
pt_model_D.eval()
with torch.no_grad():
f_wrapper_gan_one_epoch(
args, pt_model_G, pt_model_D,
loss_wrapper, \
device, \
monitor_val, val_data_loader, \
epoch_idx, None, None, normtarget_f)
time_val = monitor_val.get_time(epoch_idx)
loss_dev = monitor_val.get_loss(epoch_idx)
else:
time_val, loss_dev = 0, np.zeros_like(loss_trn)
if val_dataset_wrapper is not None:
flag_new_best = monitor_val.is_new_best()
else:
flag_new_best = True
# print information
train_log += nii_op_display_tk.print_train_info(
epoch_idx, time_trn, loss_trn, time_val, loss_dev,
flag_new_best, optimizer_G_wrapper.get_lr_info())
# save the best model
if flag_new_best or args.force_save_lite_trained_network_per_epoch:
for pt_model, tmp_tag in zip([pt_model_G, pt_model_D], model_tags):
tmp_best_name = nii_nn_tools.f_save_trained_name(args, tmp_tag)
torch.save(pt_model.state_dict(), tmp_best_name)
# save intermediate model if necessary
if not args.not_save_each_epoch:
# save model discrminator and generator
for pt_model, optimizer, model_tag in \
zip([pt_model_G, pt_model_D], [optimizer_G, optimizer_D],
model_tags):
tmp_model_name = nii_nn_tools.f_save_epoch_name(
args, epoch_idx, model_tag)
if monitor_val is not None:
tmp_val_log = monitor_val.get_state_dic()
else:
tmp_val_log = None
# save
tmp_dic = {
cp_names.state_dict : pt_model.state_dict(),
cp_names.info : train_log,
cp_names.optimizer : optimizer.state_dict(),
cp_names.trnlog : monitor_trn.get_state_dic(),
cp_names.vallog : tmp_val_log
}
torch.save(tmp_dic, tmp_model_name)
if args.verbose == 1:
nii_display.f_eprint(str(datetime.datetime.now()))
nii_display.f_eprint("Save {:s}".format(tmp_model_name),
flush=True)
# early stopping
if monitor_val is not None and \
monitor_val.should_early_stop(no_best_epoch_num):
flag_early_stopped = True
break
# loop done
nii_op_display_tk.print_log_tail()
if flag_early_stopped:
nii_display.f_print("Training finished by early stopping")
else:
nii_display.f_print("Training finished")
nii_display.f_print("Model is saved to", end = '')
for model_tag in model_tags:
nii_display.f_print("{}".format(
nii_nn_tools.f_save_trained_name(args, model_tag)))
return
if __name__ == "__main__":
print("nn_manager for GAN")
| 27,182 | 39.211538 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/nn_manager/nn_manager_tools.py | #!/usr/bin/env python
"""
nn_manager
utilities used by nn_manager
"""
from __future__ import print_function
from collections import OrderedDict
import numpy as np
import torch
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.other_tools.display as nii_display
import core_scripts.nn_manager.nn_manager_conf as nii_nn_manage_conf
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#############################################################
def f_state_dict_wrapper(state_dict, data_parallel=False):
""" a wrapper to take care of state_dict when using DataParallism
f_model_load_wrapper(state_dict, data_parallel):
state_dict: pytorch state_dict
data_parallel: whether DataParallel is used
https://discuss.pytorch.org/t/solved-keyerror-unexpected-
key-module-encoder-embedding-weight-in-state-dict/1686/3
"""
if data_parallel is True:
# if data_parallel is used
new_state_dict = OrderedDict()
for k, v in state_dict.items():
if not k.startswith('module'):
# if key is not starting with module, add it
name = 'module.' + k
else:
name = k
new_state_dict[name] = v
return new_state_dict
else:
new_state_dict = OrderedDict()
for k, v in state_dict.items():
if not k.startswith('module'):
name = k
else:
# remove module.
name = k[7:]
new_state_dict[name] = v
return new_state_dict
def f_process_loss(loss):
""" loss, loss_value = f_process_loss(loss):
Input:
loss: returned by loss_wrapper.compute
It can be a torch.tensor or a list of torch.tensor
When it is a list, it should look like:
[[loss_1, loss_2, loss_3],
[true/false, true/false, true.false]]
where true / false tells whether the loss should be taken into
consideration for early-stopping
Output:
loss: a torch.tensor
loss_value: a torch number of a list of torch number
"""
if type(loss) is list:
loss_sum = loss[0][0]
loss_list = [loss[0][0].item()]
if len(loss[0]) > 1:
for loss_tmp in loss[0][1:]:
loss_sum += loss_tmp
loss_list.append(loss_tmp.item())
return loss_sum, loss_list, loss[1]
else:
return loss, [loss.item()], [True]
def f_load_checkpoint(checkpoint, args, flag_multi_device, pt_model, optimizer,
monitor_trn, monitor_val, lr_scheduler):
""" f_load_checkpoint(checkpoint, args, pt_model, optimizer,
monitor_trn, monitor_val, lr_scheduler)
Load checkpoint.
Input:
checkpoint: check point saved by the script. Either a dict or a pt model
args: command line arguments when running the script
flag_multi_device: bool, does this code uses multiple GPUs?
Input (which will be modified by the function)
pt_model: Pytorch model, this will load the model saved in checkpoint
optimizer: optimizer, this will load the optimizer saved in checkpoint
monitor_trn: log of loss on training set
monitor_val: log of loss on validation set
lr_scheduler: scheudler of learning rate
Output:
train_log: str, text log of training loss
"""
#
train_log = ''
if checkpoint is None:
# no checkpoint
return train_log
# checkpoint exist
cp_names = nii_nn_manage_conf.CheckPointKey()
if args.allow_mismatched_pretrained_model:
if type(checkpoint) is dict:
# if it is a epoch*.pt, ignore training histories
# only load the model parameter
nii_display.f_print("allow-mismatched-pretrained-model is on")
nii_display.f_print("ignore training history in pre-trained model")
pretrained_dict = f_state_dict_wrapper(
checkpoint[cp_names.state_dict], flag_multi_device)
else:
# it is only a dictionary of trained parameters
pretrained_dict = f_state_dict_wrapper(
checkpoint, flag_multi_device)
# target model dict
model_dict = pt_model.state_dict()
# methods similar to f_load_pretrained_model_partially
# 1. filter out mismatched keys
pre_dict_tmp = {
k: v for k, v in pretrained_dict.items() \
if k in model_dict \
and model_dict[k].numel() == pretrained_dict[k].numel()}
mismatch_keys = [k for k in model_dict.keys() if k not in pre_dict_tmp]
if mismatch_keys:
print("Partially load model, ignoring buffers: {:s}".format(
' '.join(mismatch_keys)))
# 2. overwrite entries in the existing state dict
model_dict.update(pre_dict_tmp)
# 3. load the new state dict
pt_model.load_state_dict(model_dict)
else:
# the usual case
# only model status
pt_model.load_state_dict(pretrained_dict)
nii_display.f_print("Load pretrained model")
else:
if type(checkpoint) is dict:
# checkpoint is a dict (trained model + optimizer + other logs)
# load model parameter and optimizer state
if cp_names.state_dict in checkpoint:
# wrap the state_dic in f_state_dict_wrapper
# in case the model is saved when DataParallel is on
pt_model.load_state_dict(
f_state_dict_wrapper(checkpoint[cp_names.state_dict],
flag_multi_device))
# load optimizer state
if cp_names.optimizer in checkpoint and \
not args.ignore_optimizer_statistics_in_trained_model:
optimizer.load_state_dict(checkpoint[cp_names.optimizer])
# optionally, load training history
if not args.ignore_training_history_in_trained_model:
#nii_display.f_print("Load ")
if cp_names.trnlog in checkpoint:
monitor_trn.load_state_dic(checkpoint[cp_names.trnlog])
if cp_names.vallog in checkpoint and monitor_val:
monitor_val.load_state_dic(checkpoint[cp_names.vallog])
if cp_names.info in checkpoint:
train_log = checkpoint[cp_names.info]
if cp_names.lr_scheduler in checkpoint and \
checkpoint[cp_names.lr_scheduler] and lr_scheduler.f_valid():
lr_scheduler.f_load_state_dict(
checkpoint[cp_names.lr_scheduler])
nii_display.f_print("Load check point, resume training")
else:
nii_display.f_print("Load pretrained model and optimizer")
else:
# the usual case
# only model status
pt_model.load_state_dict(
f_state_dict_wrapper(checkpoint, flag_multi_device))
nii_display.f_print("Load pretrained model")
return train_log
def f_load_checkpoint_for_inference(checkpoint, pt_model):
""" f_load_checkpoint_for_inference(checkpoint, pt_model)
Load checkpoint for model inference
No matter what is inside the checkpoint, only load the model parameters
"""
cp_names = nii_nn_manage_conf.CheckPointKey()
if type(checkpoint) is dict and cp_names.state_dict in checkpoint:
pt_model.load_state_dict(checkpoint[cp_names.state_dict])
else:
pt_model.load_state_dict(checkpoint)
return
def f_load_pretrained_model_partially(model, model_paths, model_name_prefix):
""" f_load_pretrained_model_partially(model, model_paths, model_name_prefix)
Initialize part of the model with pre-trained models.
This function can be used directly. It is also called by nn_manager.py
if model.g_pretrained_model_path and model.g_pretrained_model_prefix are
defined.
For reference:
https://discuss.pytorch.org/t/how-to-load-part-of-pre-trained-model/1113/3
Input:
-----
model: torch model
model_paths: list of str, list of path to pre-trained models (.pt files)
model_prefix: list of str, list of model name prefix used by model
Output:
------
None
For example,
case1: A module in a pretrained model may be called model.***
This module will be model.m_part1.*** in the new model.
Then the prefix is "m_part1."
new_model.m_part1.*** <- pre_trained_model.***
case2: A module in a pretrained model may be called model.***
This module will still be model..*** in the new model.
Then the prefix is ""
new_model.*** <- pre_trained_model.***
Call f(model, ['./asr.pt', './tts.pt'], ['asr.', 'tts.']), then
model.asr <- load_dict(asr.pt)
model.tts <- load_dict(tts.pt)
"""
cp_names = nii_nn_manage_conf.CheckPointKey()
# change string to list
if type(model_paths) is str:
model_path_tmp = [model_paths]
else:
model_path_tmp = model_paths
if type(model_name_prefix) is str:
model_prefix_tmp = [model_name_prefix]
else:
model_prefix_tmp = model_name_prefix
# get the dictionary format of new model
model_dict = model.state_dict()
# for each pre-trained model
for model_path, prefix in zip(model_path_tmp, model_prefix_tmp):
if prefix == '':
pass
elif prefix[-1] != '.':
# m_part1. not m_part
prefix += '.'
pretrained_dict = torch.load(model_path)
# if this is a epoch***.pt, load only the network weight
if cp_names.state_dict in pretrained_dict:
pretrained_dict = pretrained_dict[cp_names.state_dict]
# 1. filter out unnecessary keys
pretrained_dict = {prefix + k: v \
for k, v in pretrained_dict.items() \
if prefix + k in model_dict}
print("Load model {:s} as {:s} ({:d} parameter buffers, ".format(
model_path, prefix, len(pretrained_dict.keys())), end=' ')
print("{:d} parameters)".format(
sum([pretrained_dict[x].numel() for x in pretrained_dict.keys()])))
# 2. overwrite entries in the existing state dict
model_dict.update(pretrained_dict)
# 3. load the new state dict
model.load_state_dict(model_dict)
return
def f_save_epoch_name(args, epoch_idx, suffix='', prefix=''):
""" str = f_save_epoch_name(args, epoch_idx)
Return the name of the model file saved during training
Args:
args: argument object by arg_parse, we will use
args.save_epoch_name, args.save_model_dir, args.save_model_ext
epoch_idx:, int, epoch index
suffix: a suffix to the name (default '')
Return:
str: name of epoch state file, str, e.g. epoch_001.pt
"""
tmp_name = args.save_epoch_name + prefix
tmp_name = "{}_{:03d}".format(tmp_name, epoch_idx)
tmp_name = tmp_name + suffix
return nii_str_tk.f_realpath(args.save_model_dir, tmp_name, \
args.save_model_ext)
def f_save_trained_name(args, suffix=''):
""" str = f_save_trained_name(args)
Return the name of the best trained model file
Args:
args: argument object by arg_parse
args.save_trained_name, args.save_model_dir, args.save_model_ext
suffix: a suffix added to the name (default '')
Return:
str: name of trained network file, e.g., trained_network.pt
"""
return nii_str_tk.f_realpath(
args.save_model_dir, args.save_trained_name + suffix,
args.save_model_ext)
def f_model_check(pt_model, model_type=None):
""" f_model_check(pt_model)
Check whether the model contains all the necessary keywords
Args:
----
pt_model: a Pytorch model
model_type_flag: str or None, a flag indicating the type of network
Return:
-------
"""
nii_display.f_print("Model check:")
if model_type in nii_nn_manage_conf.nn_model_keywords_bags:
keywords_bag = nii_nn_manage_conf.nn_model_keywords_bags[model_type]
else:
keywords_bag = nii_nn_manage_conf.nn_model_keywords_default
for tmpkey in keywords_bag.keys():
flag_mandatory, mes = keywords_bag[tmpkey]
# mandatory keywords
if flag_mandatory:
if not hasattr(pt_model, tmpkey):
nii_display.f_print("Please implement %s (%s)" % (tmpkey, mes))
nii_display.f_die("[Error]: found no %s in Model" % (tmpkey))
else:
print("[OK]: %s found" % (tmpkey))
else:
if not hasattr(pt_model, tmpkey):
print("[OK]: %s is ignored, %s" % (tmpkey, mes))
else:
print("[OK]: use %s, %s" % (tmpkey, mes))
# done
nii_display.f_print("Model check done\n")
return
def f_model_show(pt_model, do_model_def_check=True, model_type=None):
""" f_model_show(pt_model, do_model_check=True)
Print the informaiton of the model
Args:
pt_model, a Pytorch model
do_model_def_check, bool, whether check model definition (default True)
model_type: str or None (default None), what type of network
Return:
None
"""
if do_model_def_check:
f_model_check(pt_model, model_type)
nii_display.f_print("Model infor:")
print(pt_model)
num = sum(p.numel() for p in pt_model.parameters() if p.requires_grad)
nii_display.f_print("Parameter number: {:d}\n".format(num), "normal")
return
def f_set_grad_to_none(pt_model):
""" f_set_grad_to_one(pt_model)
Set the grad of trainable weights to None.
Even if grad value is 0, the weight may change due to l1 norm, moment,
or so on. It is better to explicitly set the grad of the parameter to None
https://discuss.pytorch.org/t/the-grad-is-zero-the-value-change/22765/2
"""
for p in pt_model.parameters():
if p.requires_grad:
p.requires_grad = False
p.grad = None
return
def f_loss_check(loss_module, model_type=None):
""" f_loss_check(pt_model)
Check whether the loss module contains all the necessary keywords
Args:
----
loss_module, a class
model_type, a str or None
Return:
-------
"""
nii_display.f_print("Loss check")
if model_type in nii_nn_manage_conf.loss_method_keywords_bags:
keywords_bag = nii_nn_manage_conf.loss_method_keywords_bags[model_type]
else:
keywords_bag = nii_nn_manage_conf.loss_method_keywords_default
for tmpkey in keywords_bag.keys():
flag_mandatory, mes = keywords_bag[tmpkey]
# mandatory keywords
if flag_mandatory:
if not hasattr(loss_module, tmpkey):
nii_display.f_print("Please implement %s (%s)" % (tmpkey, mes))
nii_display.f_die("[Error]: found no %s in Loss" % (tmpkey))
else:
# no need to print other information here
pass #print("[OK]: %s found" % (tmpkey))
else:
if not hasattr(loss_module, tmpkey):
# no need to print other information here
pass #print("[OK]: %s is ignored, %s" % (tmpkey, mes))
else:
print("[OK]: use %s, %s" % (tmpkey, mes))
# done
nii_display.f_print("Loss check done")
return
def f_loss_show(loss_module, do_loss_def_check=True, model_type=None):
""" f_model_show(pt_model, do_model_check=True)
Print the informaiton of the model
Args:
pt_model, a Pytorch model
do_model_def_check, bool, whether check model definition (default True)
model_type: str or None (default None), what type of network
Return:
None
"""
# no need to print other information here
# because loss is usually not a torch.Module
#nii_display.f_print("Loss infor:")
if do_loss_def_check:
f_loss_check(loss_module, model_type)
#print(loss_module)
return
########################
# data buffer operations
########################
def f_split_data(data_in, data_tar, max_length, overlap):
""" in_list, tar_list = f_split_data(data_in, data_tar, length, overlap)
Args:
data_in: tensor, (batch, length, dim)
data_tar: tensor, (batch, length, dim)
length: int, max lengnth of each trunk
overlap: int, trunc will have this number of overlap
Return:
data_in_list: list of tensors
data_tar_list: list of tensors
"""
if not isinstance(data_in, torch.Tensor):
print("Not implemented for a list of data")
sys.exit(1)
if max_length <= 0:
print("Not implemented for a negative trunc length")
sys.exit(1)
if overlap > (max_length - 1):
overlap = max_length - 1
tmp_trunc_len = max_length - overlap
trunc_num = data_in.shape[1] // tmp_trunc_len
if trunc_num > 0:
# ignore the short segment at the end
if data_in.shape[1] % tmp_trunc_len > overlap:
trunc_num += 1
else:
# however, if input is too short, just don not segment
if data_in.shape[1] % tmp_trunc_len > 0:
trunc_num += 1
data_in_list = []
data_tar_list = []
for trunc_idx in range(trunc_num):
start_idx = trunc_idx * tmp_trunc_len
end_idx = start_idx + max_length
data_in_list.append(data_in[:, start_idx:end_idx])
if isinstance(data_tar, torch.Tensor):
data_tar_list.append(data_tar[:, start_idx:end_idx])
else:
data_tar_list.append([])
return data_in_list, data_tar_list, overlap
def f_overlap_data(data_list, overlap_length):
""" data_gen = f_overlap_data(data_list, overlap_length)
Input:
data_list: list of tensors, in (batch, length, dim) or (batch, length)
overlap_length: int, overlap_length
Output:
data_gen: tensor, (batch, length, dim)
"""
batch = data_list[0].shape[0]
data_device = data_list[0].device
data_dtype = data_list[0].dtype
if len(data_list[0].shape) == 2:
dim = 1
else:
dim = data_list[0].shape[2]
total_length = sum([x.shape[1] for x in data_list])
data_gen = torch.zeros([batch, total_length, dim], dtype=data_dtype,
device = data_device)
prev_end = 0
for idx, data_trunc in enumerate(data_list):
tmp_len = data_trunc.shape[1]
if len(data_trunc.shape) == 2:
data_tmp = torch.unsqueeze(data_trunc, -1)
else:
data_tmp = data_trunc
if idx == 0:
data_gen[:, 0:tmp_len] = data_tmp
prev_end = tmp_len
else:
win_len = min([prev_end, overlap_length, tmp_len])
win_cof = torch.arange(0, win_len,
dtype=data_dtype, device=data_device)/win_len
win_cof = win_cof.unsqueeze(0).unsqueeze(-1)
data_gen[:, prev_end-win_len:prev_end] *= 1.0 - win_cof
data_tmp[:, :win_len] *= win_cof
data_gen[:, prev_end-win_len:prev_end-win_len+tmp_len] += data_tmp
prev_end = prev_end-win_len+tmp_len
return data_gen[:, 0:prev_end]
##############
#
##############
def data2device(data_in, device, data_type):
if isinstance(data_in, torch.Tensor):
data_ = data_in.to(device, dtype = data_type)
elif isinstance(data_in, list) and data_in:
data_ = [data2device(x, device, data_type) for x in data_in]
else:
data_ = None
if data_ is None:
nii_display.f_die("[Error]: fail to cast data to device")
return data_
if __name__ == "__main__":
print("nn_manager_tools")
| 20,331 | 33.815068 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/nn_manager/nn_manager.py | #!/usr/bin/env python
"""
nn_manager
A simple wrapper to run the training / testing process
"""
from __future__ import print_function
import time
import datetime
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
import core_scripts.data_io.conf as nii_dconf
import core_scripts.data_io.seq_info as nii_seqinfo
import core_scripts.other_tools.list_tools as nii_list_tk
import core_scripts.other_tools.display as nii_display
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.op_manager.op_process_monitor as nii_monitor
import core_scripts.op_manager.op_display_tools as nii_op_display_tk
import core_scripts.nn_manager.nn_manager_tools as nii_nn_tools
import core_scripts.nn_manager.nn_manager_conf as nii_nn_manage_conf
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#############################################################
def f_run_one_epoch(args,
pt_model, loss_wrapper, \
device, monitor, \
data_loader, epoch_idx, optimizer = None, \
target_norm_method = None,
data_set_wrapper = None):
"""
f_run_one_epoch:
run one poech over the dataset (for training or validation sets)
Args:
args: from argpase
pt_model: pytorch model (torch.nn.Module)
loss_wrapper: a wrapper over loss function
loss_wrapper.compute(generated, target)
device: torch.device("cuda") or torch.device("cpu")
monitor: defined in op_procfess_monitor.py
data_loader: pytorch DataLoader.
epoch_idx: int, index of the current epoch
optimizer: torch optimizer or None
if None, the back propgation will be skipped
(for developlement set)
target_norm_method: method to normalize target data
(by default, use pt_model.normalize_target)
data_set_wrapper: pytorch Dataset. It is only used to update
information
"""
# timer
start_time = time.time()
# loop over samples
for data_idx, (data_in, data_tar, data_info, idx_orig) in \
enumerate(data_loader):
#############
# prepare
#############
# idx_orig is the original idx in the dataset
# which can be different from data_idx when shuffle = True
#idx_orig = idx_orig.numpy()[0]
#data_seq_info = data_info[0]
# send data to device
if optimizer is not None:
if args.size_accumulate_grad > 1:
# if accumulate-gradient is on, only zero out grad here
if data_idx % args.size_accumulate_grad == 0:
optimizer.zero_grad()
else:
# zero grad is gradient accumulation is not used
optimizer.zero_grad()
############
# compute output
############
data_in = nii_nn_tools.data2device(data_in, device, nii_dconf.d_dtype)
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder & autoregressive model
if isinstance(data_tar, torch.Tensor):
data_tar_tm = data_tar.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_tar, list) and data_tar:
# if the data_tar is a list of tensors
data_tar_tm = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
else:
nii_display.f_print("--model-forward-with-target is set")
nii_display.f_die("but data_tar is not loaded, or a tensor")
if args.model_forward_with_file_name:
data_gen = pt_model(data_in, data_tar_tm, data_info)
else:
data_gen = pt_model(data_in, data_tar_tm)
else:
if args.model_forward_with_file_name:
# specifcal case when model.forward requires data_info
data_gen = pt_model(data_in, data_info)
else:
# normal case for model.forward(input)
data_gen = pt_model(data_in)
#####################
# compute loss and do back propagate
#####################
# Two cases
# 1. if loss is defined as pt_model.loss, then let the users do
# normalization inside the pt_mode.loss
# 2. if loss_wrapper is defined as a class independent from model
# there is no way to normalize the data inside the loss_wrapper
# because the normalization weight is saved in pt_model
if hasattr(pt_model, 'loss'):
# case 1, pt_model.loss is available
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_tar, list) and data_tar:
data_tar = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
else:
data_tar = []
loss_computed = pt_model.loss(data_gen, data_tar)
else:
# case 2, loss is defined independent of pt_model
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
# there is no way to normalize the data inside loss
# thus, do normalization here
if target_norm_method is None:
normed_target = pt_model.normalize_target(data_tar)
else:
normed_target = target_norm_method(data_tar)
elif isinstance(data_tar, list) and data_tar:
data_tar = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
if target_norm_method is None:
normed_target = pt_model.normalize_target(data_tar)
else:
normed_target = target_norm_method(data_tar)
else:
normed_target = []
# return the loss from loss_wrapper
# loss_computed may be [[loss_1, loss_2, ...],[flag_1, flag_2,.]]
# which contain multiple loss and flags indicating whether
# the corresponding loss should be taken into consideration
# for early stopping
# or
# loss_computed may be simply a tensor loss
loss_computed = loss_wrapper.compute(data_gen, normed_target)
loss_values = [0]
# To handle cases where there are multiple loss functions
# when loss_comptued is [[loss_1, loss_2, ...],[flag_1, flag_2,.]]
# loss: sum of [loss_1, loss_2, ...], for backward()
# loss_values: [loss_1.item(), loss_2.item() ..], for logging
# loss_flags: [True/False, ...], for logging,
# whether loss_n is used for early stopping
# when loss_computed is loss
# loss: loss
# los_vals: [loss.item()]
# loss_flags: [True]
loss, loss_values, loss_flags = nii_nn_tools.f_process_loss(
loss_computed)
# Back-propgation using the summed loss
if optimizer is not None and loss.requires_grad:
if args.size_accumulate_grad > 1:
# if gradient accumulation is on
# adjust loss based on the number of batches accumulated
loss = loss / args.size_accumulate_grad
loss.backward()
# do updating only after specified number of mini-batches
if (data_idx + 1) % args.size_accumulate_grad == 0:
# apply gradient clip
if args.grad_clip_norm > 0:
grad_norm = torch.nn.utils.clip_grad_norm_(
pt_model.parameters(), args.grad_clip_norm)
# update parameters
optimizer.step()
else:
# else
# backward propagation
loss.backward()
# apply gradient clip
if args.grad_clip_norm > 0:
grad_norm = torch.nn.utils.clip_grad_norm_(
pt_model.parameters(), args.grad_clip_norm)
# update parameters
optimizer.step()
# save the training process information to the monitor
end_time = time.time()
batchsize = len(data_info)
for idx, data_seq_info in enumerate(data_info):
# loss_value is supposed to be the average loss value
# over samples in the the batch, thus, just loss_value
# rather loss_value / batchsize
monitor.log_loss(loss_values, loss_flags, \
(end_time-start_time) / batchsize, \
data_seq_info, idx_orig.numpy()[idx], \
epoch_idx)
# print infor for one sentence
if args.verbose == 1:
# here we use args.batch_size because len(data_info)
# may be < args.batch_size.
monitor.print_error_for_batch(
data_idx * args.batch_size + idx,\
idx_orig.numpy()[idx], \
epoch_idx)
#
# start the timer for a new batch
start_time = time.time()
# Save intermediate model for every n mini-batches (optional).
# Note that if we re-start trainining with this intermediate model,
# the data will start from the 1st sample, not the one where we stopped
if args.save_model_every_n_minibatches > 0 \
and (data_idx+1) % args.save_model_every_n_minibatches == 0 \
and optimizer is not None and data_idx > 0:
cp_names = nii_nn_manage_conf.CheckPointKey()
tmp_model_name = nii_nn_tools.f_save_epoch_name(
args, epoch_idx, '_{:05d}'.format(data_idx+1))
# save
tmp_dic = {
cp_names.state_dict : pt_model.state_dict(),
cp_names.optimizer : optimizer.state_dict()
}
torch.save(tmp_dic, tmp_model_name)
# If debug mode is used, only run a specified number of mini-batches
if args.debug_batch_num > 0 and data_idx >= (args.debug_batch_num - 1):
nii_display.f_print("Debug mode is on. This epoch is finished")
break
# other procedures
if data_set_wrapper and args.force_update_seq_length:
# update data length for sampler
# when using multi-thread (workers > 0), the information updated
# in each subthread will not be reflected here.
# we need to do this update manually
data_set_wrapper.update_seq_len_in_sampler_sub(data_info)
# loop done
return
def f_train_wrapper(args, pt_model, loss_wrapper, device, \
optimizer_wrapper, \
train_dataset_wrapper, \
val_dataset_wrapper = None, \
checkpoint = None):
"""
f_train_wrapper(args, pt_model, loss_wrapper, device,
optimizer_wrapper
train_dataset_wrapper, val_dataset_wrapper = None,
check_point = None):
A wrapper to run the training process
Args:
args: argument information given by argpase
pt_model: pytorch model (torch.nn.Module)
loss_wrapper: a wrapper over loss function
loss_wrapper.compute(generated, target)
device: torch.device("cuda") or torch.device("cpu")
optimizer_wrapper:
a wrapper over optimizer (defined in op_manager.py)
optimizer_wrapper.optimizer is torch.optimizer
train_dataset_wrapper:
a wrapper over training data set (data_io/default_data_io.py)
train_dataset_wrapper.get_loader() returns torch.DataSetLoader
val_dataset_wrapper:
a wrapper over validation data set (data_io/default_data_io.py)
it can None.
check_point:
a check_point that stores every thing to resume training
"""
nii_display.f_print_w_date("Start model training")
##############
## Preparation
##############
# get the optimizer
optimizer_wrapper.print_info()
optimizer = optimizer_wrapper.optimizer
lr_scheduler = optimizer_wrapper.lr_scheduler
epoch_num = optimizer_wrapper.get_epoch_num()
no_best_epoch_num = optimizer_wrapper.get_no_best_epoch_num()
# get data loader for training set
train_dataset_wrapper.print_info()
train_data_loader = train_dataset_wrapper.get_loader()
train_seq_num = train_dataset_wrapper.get_seq_num()
# get the training process monitor
monitor_trn = nii_monitor.Monitor(epoch_num, train_seq_num)
# if validation data is provided, get data loader for val set
if val_dataset_wrapper is not None:
val_dataset_wrapper.print_info()
val_data_loader = val_dataset_wrapper.get_loader()
val_seq_num = val_dataset_wrapper.get_seq_num()
monitor_val = nii_monitor.Monitor(epoch_num, val_seq_num)
else:
monitor_val = None
# training log information
train_log = ''
# prepare for DataParallism if available
# pytorch.org/tutorials/beginner/blitz/data_parallel_tutorial.html
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
flag_multi_device = True
nii_display.f_print("\nUse %d GPUs\n" % (torch.cuda.device_count()))
# no way to call normtarget_f after pt_model is in DataParallel
normtarget_f = pt_model.normalize_target
pt_model = nn.DataParallel(pt_model)
else:
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
flag_multi_device = False
normtarget_f = None
pt_model.to(device, dtype=nii_dconf.d_dtype)
# print the network
nii_nn_tools.f_model_show(pt_model)
nii_nn_tools.f_loss_show(loss_wrapper)
cp_names = nii_nn_manage_conf.CheckPointKey()
###############################
## Resume training if necessary
###############################
# resume training or initialize the model if necessary
train_log = nii_nn_tools.f_load_checkpoint(
checkpoint, args, flag_multi_device, pt_model,
optimizer, monitor_trn, monitor_val, lr_scheduler)
######################
### User defined setup
######################
if hasattr(pt_model, "other_setups"):
nii_display.f_print("Conduct User-defined setup")
pt_model.other_setups()
# This should be merged with other_setups
if hasattr(pt_model, "g_pretrained_model_path") and \
hasattr(pt_model, "g_pretrained_model_prefix"):
nii_display.f_print("Load pret-rained models as part of this mode")
nii_nn_tools.f_load_pretrained_model_partially(
pt_model, pt_model.g_pretrained_model_path,
pt_model.g_pretrained_model_prefix)
######################
### Start training
######################
# other variables
flag_early_stopped = False
start_epoch = monitor_trn.get_epoch()
epoch_num = monitor_trn.get_max_epoch()
# print
_ = nii_op_display_tk.print_log_head()
nii_display.f_print_message(train_log, flush=True, end='')
# loop over multiple epochs
for epoch_idx in range(start_epoch, epoch_num):
# training one epoch
pt_model.train()
# set validation and other flags if necessary
# this tells the model whether the current epoch is for validation
if hasattr(pt_model, 'validation'):
pt_model.validation = False
mes = "Warning: model.validation is deprecated, "
mes += "please use model.g_flag_validation"
nii_display.f_print(mes, 'warning')
if hasattr(pt_model, 'flag_validation'):
pt_model.flag_validation = False
if hasattr(pt_model, 'g_flag_validation'):
pt_model.g_flag_validation = False
# set epoch number
if hasattr(pt_model, 'g_epoch_idx'):
pt_model.g_epoch_idx = epoch_idx
# run one epoch
f_run_one_epoch(args, pt_model, loss_wrapper, device, \
monitor_trn, train_data_loader, \
epoch_idx, optimizer, normtarget_f, \
train_dataset_wrapper)
time_trn = monitor_trn.get_time(epoch_idx)
loss_trn = monitor_trn.get_loss(epoch_idx)
# if necessary, do validataion
if val_dataset_wrapper is not None:
# set eval() if necessary
if args.eval_mode_for_validation:
pt_model.eval()
# set validation flag if necessary
if hasattr(pt_model, 'validation'):
pt_model.validation = True
mes = "Warning: model.validation is deprecated, "
mes += "please use model.flag_validation"
nii_display.f_print(mes, 'warning')
if hasattr(pt_model, 'flag_validation'):
pt_model.flag_validation = True
with torch.no_grad():
f_run_one_epoch(args, pt_model, loss_wrapper, \
device, \
monitor_val, val_data_loader, \
epoch_idx, None, normtarget_f, \
val_dataset_wrapper)
time_val = monitor_val.get_time(epoch_idx)
loss_val = monitor_val.get_loss(epoch_idx)
# update lr rate scheduler if necessary
if lr_scheduler.f_valid():
lr_scheduler.f_step(
monitor_val.get_loss_for_learning_stopping(epoch_idx))
else:
#time_val = monitor_val.get_time(epoch_idx)
#loss_val = monitor_val.get_loss(epoch_idx)
time_val, loss_val = 0, np.zeros_like(loss_trn)
if val_dataset_wrapper is not None:
flag_new_best = monitor_val.is_new_best()
else:
flag_new_best = True
# print information
train_log += nii_op_display_tk.print_train_info(
epoch_idx, time_trn, loss_trn, time_val, loss_val,
flag_new_best, optimizer_wrapper.get_lr_info())
# save the best model
if flag_new_best or args.force_save_lite_trained_network_per_epoch:
tmp_best_name = nii_nn_tools.f_save_trained_name(args)
torch.save(pt_model.state_dict(), tmp_best_name)
# save intermediate model if necessary
if not args.not_save_each_epoch:
tmp_model_name = nii_nn_tools.f_save_epoch_name(args, epoch_idx)
if monitor_val is not None:
tmp_val_log = monitor_val.get_state_dic()
else:
tmp_val_log = None
if lr_scheduler.f_valid():
lr_scheduler_state = lr_scheduler.f_state_dict()
else:
lr_scheduler_state = None
# save
tmp_dic = {
cp_names.state_dict : pt_model.state_dict(),
cp_names.info : train_log,
cp_names.optimizer : optimizer.state_dict(),
cp_names.trnlog : monitor_trn.get_state_dic(),
cp_names.vallog : tmp_val_log,
cp_names.lr_scheduler : lr_scheduler_state
}
torch.save(tmp_dic, tmp_model_name)
if args.verbose == 1:
nii_display.f_eprint(str(datetime.datetime.now()))
nii_display.f_eprint("Save {:s}".format(tmp_model_name),
flush=True)
# Early stopping
# note: if LR scheduler is used, early stopping will be
# disabled
if lr_scheduler.f_allow_early_stopping() and \
monitor_val is not None and \
monitor_val.should_early_stop(no_best_epoch_num):
flag_early_stopped = True
break
# Update datasets informaiton if necessary
if args.force_update_seq_length:
# update the sequence length logged in sampler if sequence
# length are to be changed during data loading process
train_dataset_wrapper.update_seq_len_in_sampler()
val_dataset_wrapper.update_seq_len_in_sampler()
# loop done
nii_op_display_tk.print_log_tail()
if flag_early_stopped:
nii_display.f_print("Training finished by early stopping")
else:
nii_display.f_print("Training finished")
nii_display.f_print("Model is saved to", end = '')
nii_display.f_print("{}".format(nii_nn_tools.f_save_trained_name(args)))
return
def f_inference_wrapper(args, pt_model, device, \
test_dataset_wrapper, checkpoint):
""" Wrapper for inference
"""
# prepare dataloader
test_data_loader = test_dataset_wrapper.get_loader()
test_seq_num = test_dataset_wrapper.get_seq_num()
test_dataset_wrapper.print_info()
# cuda device
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
nii_display.f_print(
"DataParallel for inference is not implemented", 'warning')
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
# print the network
pt_model.to(device, dtype=nii_dconf.d_dtype)
nii_nn_tools.f_model_show(pt_model)
# load trained model parameters from checkpoint
nii_nn_tools.f_load_checkpoint_for_inference(checkpoint, pt_model)
# decide the range of the data index to generate
range_genidx_start = args.inference_sample_start_index
if args.inference_sample_end_index < 0:
range_genidx_end = len(test_data_loader)
else:
range_genidx_end = args.inference_sample_end_index
if range_genidx_start >= range_genidx_end:
mes = "--inference-sample-start-index should be smaller than"
mes += " --inference-sample-end-index"
nii_display.f_die(mes)
# print information
nii_display.f_print("Start inference (generation):", 'highlight')
nii_display.f_print("Generate minibatch indexed within [{:d},{:d})".format(
range_genidx_start, range_genidx_end))
# if a list of file to be processed is provided
inf_datalist_path = args.inference_data_list
if len(inf_datalist_path):
inf_datalist = nii_list_tk.read_list_from_text(inf_datalist_path)
mes = "And only data in {:s} is processed".format(inf_datalist_path)
nii_display.f_print(mes)
else:
inf_datalist = None
# other information
if hasattr(args, 'trunc_input_length_for_inference') and \
args.trunc_input_length_for_inference > 0:
mes = "Generation in segment-by-segment mode (truncation length "
mes += "{:d})".format(args.trunc_input_length_for_inference)
nii_display.f_print(mes)
# output buffer, filename buffer
#output_buf = []
#filename_buf = []
# start generation
pt_model.eval()
total_start_time = time.time()
total_accumulate = 0
sample_account = 0
with torch.no_grad():
start_time_load = time.time()
# run generation
for data_idx, (data_in, data_tar, data_info, idx_orig) in \
enumerate(test_data_loader):
# decide whether to process this data sample or not
if data_idx < range_genidx_start:
# not in range
nii_display.f_print("skip {:s}".format(str(data_info)))
continue
elif data_idx >= range_genidx_end:
# not in range
nii_display.f_print("stopped by --inference-sample-end-index")
break
else:
# and if the data is in the list
if inf_datalist is not None:
# to be completed. this only works for batchsize=1
seqname = nii_seqinfo.SeqInfo()
seqname.parse_from_str(data_info[0])
if seqname.seq_tag() in inf_datalist:
pass
else:
nii_display.f_print("skip {:s}".format(str(data_info)))
continue
else:
pass
# send data to device and convert data type
if isinstance(data_in, torch.Tensor):
data_in = data_in.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_in, list) and data_in:
data_in = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_in]
else:
nii_display.f_die("data_in is not a tensor or list of tensors")
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_tar, list) and data_tar:
data_tar = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
else:
pass
# load down time for debugging
start_time_inf = time.time()
time_cost_load = (start_time_inf - start_time_load)/len(data_info)
# in case the model defines inference function explicitly
if hasattr(pt_model, "inference"):
infer_func = pt_model.inference
else:
infer_func = pt_model.forward
if hasattr(args, 'trunc_input_length_for_inference') and \
args.trunc_input_length_for_inference > 0:
# generate the data segment by segment, then do overlap and add
in_list, tar_list, overlap = nii_nn_tools.f_split_data(
data_in, data_tar, args.trunc_input_length_for_inference,
args.trunc_input_overlap)
gen_list = []
for in_tmp, tar_tmp in zip(in_list, tar_list):
# compute output
if args.model_forward_with_target:
if args.model_forward_with_file_name:
data_gen = infer_func(in_tmp, tar_tmp, data_info)
else:
data_gen = infer_func(in_tmp, tar_tmp)
else:
if args.model_forward_with_file_name:
data_gen = infer_func(in_tmp, data_info)
else:
data_gen = infer_func(in_tmp)
gen_list.append(data_gen)
# generation model may "up-sample" the input, we need to know
# output_segment_length // input_segment_length
if len(gen_list) > 0 and len(in_list) > 0:
upsamp_fac = gen_list[0].shape[1] // in_list[0].shape[1]
data_gen = nii_nn_tools.f_overlap_data(
gen_list, upsamp_fac * overlap)
else:
print("Gneration failed on {:s}".format(data_info))
sys.exit(1)
else:
# normal case: generate the output sequence as a whole
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder
if args.model_forward_with_file_name:
data_gen = infer_func(data_in, data_tar, data_info)
else:
data_gen = infer_func(data_in, data_tar)
else:
if args.model_forward_with_file_name:
data_gen = infer_func(data_in, data_info)
else:
data_gen = infer_func(data_in)
# log time for debugging
start_time_save = time.time()
time_cost_inf = start_time_save - start_time_inf
# average time for each sequence when batchsize > 1
time_cost_inf = time_cost_inf / len(data_info)
# write the generated data to file
if data_gen is None:
nii_display.f_print("No output saved: %s" % (str(data_info)),\
'warning')
else:
#output_buf.append(data_gen)
#filename_buf.append(data_info)
try:
data_gen = pt_model.denormalize_output(data_gen)
data_gen_np = data_gen.to("cpu").numpy()
except AttributeError:
mes = "Output data is not torch.tensor. Please check "
mes += "model.forward or model.inference"
nii_display.f_die(mes)
# save output (in case batchsize > 1, )
for idx, seq_info in enumerate(data_info):
#nii_display.f_print(seq_info)
test_dataset_wrapper.putitem(data_gen_np[idx:idx+1],\
args.output_dir, \
args.output_filename_prefix, \
seq_info)
# time to generate
start_time_load = time.time()
time_cost_save = (start_time_load - start_time_save)/len(data_info)
# print information
time_cost = time_cost_load + time_cost_inf + time_cost_save
for idx, seq_info in enumerate(data_info):
if args.verbose == 2:
print("{:s} {:f} {:f} {:f}".format(
seq_info,time_cost_load, time_cost_inf, time_cost_save))
sample_account += 1
_ = nii_op_display_tk.print_gen_info(
seq_info, time_cost, sample_account)
# log time for debugging
total_accumulate += time_cost * len(data_info)
#
total_time = time.time() - total_start_time
nii_display.f_print("Inference time cost: {:f}s".format(total_time))
#nii_display.f_print("{:f}".format(total_accumulate))
# done for
# done with
nii_display.f_print("Output data has been saved to %s" % (args.output_dir))
# finish up if necessary
if hasattr(pt_model, "finish_up_inference"):
pt_model.finish_up_inference()
# done
return
def f_convert_epoch_to_trained(args, pt_model, device, checkpoint):
""" Convert a checkpoint to trained_network.pt
(remove gradients and other statistics for training)
"""
# cuda device
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
nii_display.f_print(
"DataParallel is not implemented here", 'warning')
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
# print the network
pt_model.to(device, dtype=nii_dconf.d_dtype)
nii_nn_tools.f_model_show(pt_model)
# load trained model parameters from checkpoint
nii_nn_tools.f_load_checkpoint_for_inference(checkpoint, pt_model)
# start generation
nii_display.f_print("Start conversion:", 'highlight')
tmp_best_name = nii_nn_tools.f_save_trained_name(args)
torch.save(pt_model.state_dict(), tmp_best_name)
nii_display.f_print("Model is saved to", end = '')
nii_display.f_print("{}".format(tmp_best_name))
# done
return
if __name__ == "__main__":
print("nn_manager")
| 32,606 | 39.606476 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/nn_manager/nn_manager_profile.py | #!/usr/bin/env python
"""
A trimmed version of nn_manager.py for profiling
This requires Pytorch-1.8
https://pytorch.org/tutorials/intermediate/tensorboard_profiler_tutorial.html
It requires a specific ../../sandbox/different_main/main_profile.py to run.
"""
from __future__ import print_function
import time
import datetime
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.profiler
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.display as nii_display
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.op_manager.op_process_monitor as nii_monitor
import core_scripts.op_manager.op_display_tools as nii_op_display_tk
import core_scripts.nn_manager.nn_manager_tools as nii_nn_tools
import core_scripts.nn_manager.nn_manager_conf as nii_nn_manage_conf
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#############################################################
def f_run_one_epoch_profile(args,
pt_model, loss_wrapper, \
device, monitor, \
data_loader, epoch_idx, optimizer = None, \
target_norm_method = None):
"""
f_run_one_epoch:
run one poech over the dataset (for training or validation sets)
Args:
args: from argpase
pt_model: pytorch model (torch.nn.Module)
loss_wrapper: a wrapper over loss function
loss_wrapper.compute(generated, target)
device: torch.device("cuda") or torch.device("cpu")
monitor: defined in op_procfess_monitor.py
data_loader: pytorch DataLoader.
epoch_idx: int, index of the current epoch
optimizer: torch optimizer or None
if None, the back propgation will be skipped
(for developlement set)
target_norm_method: method to normalize target data
(by default, use pt_model.normalize_target)
"""
# timer
start_time = time.time()
#######################
# options for profile
#######################
try:
prof_opt = [int(x) for x in args.wait_warmup_active_repeat.split('-')]
except ValueError:
nii_display.f_die("Fail to parse --wait-warmup-active-repeat")
if len(prof_opt) != 4:
nii_display.f_die("Fail to parse --wait-warmup-active-repeat")
# number of steps for profiling
num_steps = (prof_opt[0] + prof_opt[1] + prof_opt[2]) * prof_opt[3]
# output dir
prof_outdir = args.profile_output_dir
with torch.profiler.profile(
schedule=torch.profiler.schedule(wait=prof_opt[0],
warmup=prof_opt[1],
active=prof_opt[2],
repeat=prof_opt[3]),
on_trace_ready=torch.profiler.tensorboard_trace_handler(prof_outdir),
record_shapes=True,
profile_memory=False,
with_stack=True
) as prof:
# loop over samples
for data_idx, (data_in, data_tar, data_info, idx_orig) in \
enumerate(data_loader):
# If debug mode is used, only run a specified number of mini-batches
if data_idx >= num_steps:
nii_display.f_print("Profiling mode is on. Epoch is finished")
break
#############
# prepare
#############
# idx_orig is the original idx in the dataset
# which can be different from data_idx when shuffle = True
#idx_orig = idx_orig.numpy()[0]
#data_seq_info = data_info[0]
# send data to device
if optimizer is not None:
optimizer.zero_grad()
############
# compute output
############
if isinstance(data_in, torch.Tensor):
data_in = data_in.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_in, list) and data_in:
data_in=[x.to(device, dtype=nii_dconf.d_dtype) for x in data_in]
else:
nii_display.f_die("data_in is not a tensor or list of tensors")
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder & autoregressive model
if isinstance(data_tar, torch.Tensor):
data_tar_tm = data_tar.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_tar, list) and data_tar:
# if the data_tar is a list of tensors
data_tar_tm = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
else:
nii_display.f_print("--model-forward-with-target is set")
nii_display.f_die("but data_tar is not loaded, or a tensor")
if args.model_forward_with_file_name:
data_gen = pt_model(data_in, data_tar_tm, data_info)
else:
data_gen = pt_model(data_in, data_tar_tm)
else:
if args.model_forward_with_file_name:
# specifcal case when model.forward requires data_info
data_gen = pt_model(data_in, data_info)
else:
# normal case for model.forward(input)
data_gen = pt_model(data_in)
#####################
# compute loss and do back propagate
#####################
# Two cases
# 1. if loss is defined as pt_model.loss, then let the users do
# normalization inside the pt_mode.loss
# 2. if loss_wrapper is defined as a class independent from model
# there is no way to normalize the data inside the loss_wrapper
# because the normalization weight is saved in pt_model
if hasattr(pt_model, 'loss'):
# case 1, pt_model.loss is available
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_tar, list) and data_tar:
data_tar = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
else:
data_tar = []
loss_computed = pt_model.loss(data_gen, data_tar)
else:
# case 2, loss is defined independent of pt_model
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
# there is no way to normalize the data inside loss
# thus, do normalization here
if target_norm_method is None:
normed_target = pt_model.normalize_target(data_tar)
else:
normed_target = target_norm_method(data_tar)
elif isinstance(data_tar, list) and data_tar:
data_tar = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
if target_norm_method is None:
normed_target = pt_model.normalize_target(data_tar)
else:
normed_target = target_norm_method(data_tar)
else:
normed_target = []
# return the loss from loss_wrapper
# loss_computed may be [[loss_1, loss_2, ...],[flag_1,flag_2,.]]
# which contain multiple loss and flags indicating whether
# the corresponding loss should be taken into consideration
# for early stopping
# or
# loss_computed may be simply a tensor loss
loss_computed = loss_wrapper.compute(data_gen, normed_target)
loss_values = [0]
# To handle cases where there are multiple loss functions
# when loss_comptued is [[loss_1, loss_2, ...],[flag_1, flag_2,.]]
# loss: sum of [loss_1, loss_2, ...], for backward()
# loss_values: [loss_1.item(), loss_2.item() ..], for logging
# loss_flags: [True/False, ...], for logging,
# whether loss_n is used for early stopping
# when loss_computed is loss
# loss: loss
# los_vals: [loss.item()]
# loss_flags: [True]
loss, loss_values, loss_flags = nii_nn_tools.f_process_loss(
loss_computed)
# Back-propgation using the summed loss
if optimizer is not None and loss.requires_grad:
# backward propagation
loss.backward()
# apply gradient clip
if args.grad_clip_norm > 0:
grad_norm = torch.nn.utils.clip_grad_norm_(
pt_model.parameters(), args.grad_clip_norm)
# update parameters
optimizer.step()
# save the training process information to the monitor
end_time = time.time()
batchsize = len(data_info)
for idx, data_seq_info in enumerate(data_info):
# loss_value is supposed to be the average loss value
# over samples in the the batch, thus, just loss_value
# rather loss_value / batchsize
monitor.log_loss(loss_values, loss_flags, \
(end_time-start_time) / batchsize, \
data_seq_info, idx_orig.numpy()[idx], \
epoch_idx)
# print infor for one sentence
if args.verbose == 1:
# here we use args.batch_size because len(data_info)
# may be < args.batch_size.
monitor.print_error_for_batch(
data_idx * args.batch_size + idx,\
idx_orig.numpy()[idx], \
epoch_idx)
#
# start the timer for a new batch
start_time = time.time()
#
prof.step()
# loop done
# done with
return
def f_train_wrapper(args, pt_model, loss_wrapper, device, \
optimizer_wrapper, \
train_dataset_wrapper, \
val_dataset_wrapper = None, \
checkpoint = None):
"""
f_train_wrapper(args, pt_model, loss_wrapper, device,
optimizer_wrapper
train_dataset_wrapper, val_dataset_wrapper = None,
check_point = None):
A wrapper to run the training process
Args:
args: argument information given by argpase
pt_model: pytorch model (torch.nn.Module)
loss_wrapper: a wrapper over loss function
loss_wrapper.compute(generated, target)
device: torch.device("cuda") or torch.device("cpu")
optimizer_wrapper:
a wrapper over optimizer (defined in op_manager.py)
optimizer_wrapper.optimizer is torch.optimizer
train_dataset_wrapper:
a wrapper over training data set (data_io/default_data_io.py)
train_dataset_wrapper.get_loader() returns torch.DataSetLoader
val_dataset_wrapper:
a wrapper over validation data set (data_io/default_data_io.py)
it can None.
check_point:
a check_point that stores every thing to resume training
"""
nii_display.f_print_w_date("Start model training")
##############
## Preparation
##############
# get the optimizer
optimizer_wrapper.print_info()
optimizer = optimizer_wrapper.optimizer
lr_scheduler = optimizer_wrapper.lr_scheduler
epoch_num = optimizer_wrapper.get_epoch_num()
no_best_epoch_num = optimizer_wrapper.get_no_best_epoch_num()
# get data loader for training set
train_dataset_wrapper.print_info()
train_data_loader = train_dataset_wrapper.get_loader()
train_seq_num = train_dataset_wrapper.get_seq_num()
# get the training process monitor
monitor_trn = nii_monitor.Monitor(epoch_num, train_seq_num)
# if validation data is provided, get data loader for val set
if val_dataset_wrapper is not None:
val_dataset_wrapper.print_info()
val_data_loader = val_dataset_wrapper.get_loader()
val_seq_num = val_dataset_wrapper.get_seq_num()
monitor_val = nii_monitor.Monitor(epoch_num, val_seq_num)
else:
monitor_val = None
# training log information
train_log = ''
# prepare for DataParallism if available
# pytorch.org/tutorials/beginner/blitz/data_parallel_tutorial.html
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
flag_multi_device = True
nii_display.f_print("\nUse %d GPUs\n" % (torch.cuda.device_count()))
# no way to call normtarget_f after pt_model is in DataParallel
normtarget_f = pt_model.normalize_target
pt_model = nn.DataParallel(pt_model)
else:
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
flag_multi_device = False
normtarget_f = None
pt_model.to(device, dtype=nii_dconf.d_dtype)
# print the network
nii_nn_tools.f_model_show(pt_model)
nii_nn_tools.f_loss_show(loss_wrapper)
cp_names = nii_nn_manage_conf.CheckPointKey()
###############################
## Resume training if necessary
###############################
# resume training or initialize the model if necessary
train_log = nii_nn_tools.f_load_checkpoint(
checkpoint, args, flag_multi_device, pt_model,
optimizer, monitor_trn, monitor_val, lr_scheduler)
######################
### User defined setup
######################
if hasattr(pt_model, "other_setups"):
nii_display.f_print("Conduct User-defined setup")
pt_model.other_setups()
# This should be merged with other_setups
if hasattr(pt_model, "g_pretrained_model_path") and \
hasattr(pt_model, "g_pretrained_model_prefix"):
nii_display.f_print("Load pret-rained models as part of this mode")
nii_nn_tools.f_load_pretrained_model_partially(
pt_model, pt_model.g_pretrained_model_path,
pt_model.g_pretrained_model_prefix)
######################
### Start training
######################
# other variables
flag_early_stopped = False
start_epoch = monitor_trn.get_epoch()
#epoch_num = monitor_trn.get_max_epoch()
epoch_num = 1
# print
_ = nii_op_display_tk.print_log_head()
nii_display.f_print_message(train_log, flush=True, end='')
# loop over multiple epochs
for epoch_idx in range(start_epoch, epoch_num):
# training one epoch
pt_model.train()
# set validation flag if necessary
if hasattr(pt_model, 'validation'):
pt_model.validation = False
mes = "Warning: model.validation is deprecated, "
mes += "please use model.flag_validation"
nii_display.f_print(mes, 'warning')
if hasattr(pt_model, 'flag_validation'):
pt_model.flag_validation = False
f_run_one_epoch_profile(
args, pt_model, loss_wrapper, device, \
monitor_trn, train_data_loader, \
epoch_idx, optimizer, normtarget_f)
time_trn = monitor_trn.get_time(epoch_idx)
loss_trn = monitor_trn.get_loss(epoch_idx)
# No validation for profiling
if False and val_dataset_wrapper is not None:
# set eval() if necessary
if args.eval_mode_for_validation:
pt_model.eval()
# set validation flag if necessary
if hasattr(pt_model, 'validation'):
pt_model.validation = True
mes = "Warning: model.validation is deprecated, "
mes += "please use model.flag_validation"
nii_display.f_print(mes, 'warning')
if hasattr(pt_model, 'flag_validation'):
pt_model.flag_validation = True
with torch.no_grad():
f_run_one_epoch(args, pt_model, loss_wrapper, \
device, \
monitor_val, val_data_loader, \
epoch_idx, None, normtarget_f)
time_val = monitor_val.get_time(epoch_idx)
loss_val = monitor_val.get_loss(epoch_idx)
# update lr rate scheduler if necessary
if lr_scheduler.f_valid():
lr_scheduler.f_step(loss_val)
else:
time_val = monitor_val.get_time(epoch_idx)
loss_val = monitor_val.get_loss(epoch_idx)
if val_dataset_wrapper is not None:
flag_new_best = monitor_val.is_new_best()
else:
flag_new_best = True
# print information
train_log += nii_op_display_tk.print_train_info(
epoch_idx, time_trn, loss_trn, time_val, loss_val,
flag_new_best, optimizer_wrapper.get_lr_info())
# not save the best model for profiling
if False and flag_new_best:
tmp_best_name = nii_nn_tools.f_save_trained_name(args)
torch.save(pt_model.state_dict(), tmp_best_name)
# not save intermediate model if necessary
if False and not args.not_save_each_epoch:
tmp_model_name = nii_nn_tools.f_save_epoch_name(args, epoch_idx)
if monitor_val is not None:
tmp_val_log = monitor_val.get_state_dic()
else:
tmp_val_log = None
if lr_scheduler.f_valid():
lr_scheduler_state = lr_scheduler.f_state_dict()
else:
lr_scheduler_state = None
# save
tmp_dic = {
cp_names.state_dict : pt_model.state_dict(),
cp_names.info : train_log,
cp_names.optimizer : optimizer.state_dict(),
cp_names.trnlog : monitor_trn.get_state_dic(),
cp_names.vallog : tmp_val_log,
cp_names.lr_scheduler : lr_scheduler_state
}
torch.save(tmp_dic, tmp_model_name)
if args.verbose == 1:
nii_display.f_eprint(str(datetime.datetime.now()))
nii_display.f_eprint("Save {:s}".format(tmp_model_name),
flush=True)
# Early stopping
# note: if LR scheduler is used, early stopping will be
# disabled
if lr_scheduler.f_allow_early_stopping() and \
monitor_val is not None and \
monitor_val.should_early_stop(no_best_epoch_num):
flag_early_stopped = True
break
# loop done
nii_op_display_tk.print_log_tail()
if flag_early_stopped:
nii_display.f_print("Profiling finished")
else:
nii_display.f_print("Profiling finished")
#nii_display.f_print("Model is saved to", end = '')
#nii_display.f_print("{}".format(nii_nn_tools.f_save_trained_name(args)))
nii_display.f_print("Profiling log is saved to {:s}".format(
args.profile_output_dir))
return
def f_inference_wrapper(args, pt_model, device, \
test_dataset_wrapper, checkpoint):
""" Wrapper for inference
"""
# prepare dataloader
test_data_loader = test_dataset_wrapper.get_loader()
test_seq_num = test_dataset_wrapper.get_seq_num()
test_dataset_wrapper.print_info()
# cuda device
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
nii_display.f_print(
"DataParallel for inference is not implemented", 'warning')
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
# print the network
pt_model.to(device, dtype=nii_dconf.d_dtype)
nii_nn_tools.f_model_show(pt_model)
# load trained model parameters from checkpoint
nii_nn_tools.f_load_checkpoint_for_inference(checkpoint, pt_model)
# start generation
nii_display.f_print("Start inference (generation):", 'highlight')
# output buffer, filename buffer
output_buf = []
filename_buf = []
pt_model.eval()
with torch.no_grad():
# run generation
for _, (data_in, data_tar, data_info, idx_orig) in \
enumerate(test_data_loader):
# send data to device and convert data type
if isinstance(data_in, torch.Tensor):
data_in = data_in.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_in, list) and data_in:
data_in = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_in]
else:
nii_display.f_die("data_in is not a tensor or list of tensors")
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
elif isinstance(data_tar, list) and data_tar:
data_tar = [x.to(device, dtype=nii_dconf.d_dtype) \
for x in data_tar]
else:
pass
start_time = time.time()
# in case the model defines inference function explicitly
if hasattr(pt_model, "inference"):
infer_func = pt_model.inference
else:
infer_func = pt_model.forward
# compute output
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder
if args.model_forward_with_file_name:
data_gen = infer_func(data_in, data_tar, data_info)
else:
data_gen = infer_func(data_in, data_tar)
else:
if args.model_forward_with_file_name:
data_gen = infer_func(data_in, data_info)
else:
data_gen = infer_func(data_in)
time_cost = time.time() - start_time
# average time for each sequence when batchsize > 1
time_cost = time_cost / len(data_info)
if data_gen is None:
nii_display.f_print("No output saved: %s" % (str(data_info)),\
'warning')
else:
output_buf.append(data_gen)
filename_buf.append(data_info)
# print information
for idx, seq_info in enumerate(data_info):
_ = nii_op_display_tk.print_gen_info(seq_info, time_cost)
# Writing generatd data to disk
nii_display.f_print("Writing output to %s" % (args.output_dir))
for data_gen, data_info in zip(output_buf, filename_buf):
if data_gen is not None:
try:
data_gen = pt_model.denormalize_output(data_gen)
data_gen_np = data_gen.to("cpu").numpy()
except AttributeError:
mes = "Output data is not torch.tensor. Please check "
mes += "model.forward or model.inference"
nii_display.f_die(mes)
# save output (in case batchsize > 1, )
for idx, seq_info in enumerate(data_info):
nii_display.f_print(seq_info)
test_dataset_wrapper.putitem(data_gen_np[idx:idx+1],\
args.output_dir, \
args.output_filename_prefix, \
seq_info)
# done for
# done with
nii_display.f_print("Output data has been saved to %s" % (args.output_dir))
# finish up if necessary
if hasattr(pt_model, "finish_up_inference"):
pt_model.finish_up_inference()
# done
return
if __name__ == "__main__":
print("nn_manager used for profiling")
| 25,076 | 38.931529 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/nn_manager/nn_manager_conf.py | #!/usr/bin/env python
"""
nn_manager_conf
A few definitions of nn_manager
"""
from __future__ import print_function
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
####
# Name of keys in checkpoint
# Epochs saved as checkpoint will have these fields
# state_dict: network weights
# info: printed information from the nn_manager
# optimizer: optimizer state
# trnlog: log of training error on training set
# vallog: log of validation error on validation set
# lr_scheduler: status for learning rate scheduler
####
class CheckPointKey:
state_dict = 'state_dict'
info = 'info'
optimizer = 'optimizer'
trnlog = 'train_log'
vallog = 'val_log'
lr_scheduler = 'lr_scheduler'
####
# Methods that a Model should have
# name: (flag_mandatory, comment)
# name: str, name of the method
# flag_mandatory: bool, whether this method is mandatory to have
# comment: str, comment string on the method
#
####
# keywords for typical models
nn_model_keywords_default = {
#
# mandatory methods that a model must define
'prepare_mean_std': (True, "method to initialize mean/std"),
'normalize_input': (True, "method to normalize input features"),
'normalize_target': (True, "method to normalize target features"),
'denormalize_output': (True, "method to de-normalize output features"),
'forward': (True, "main method for forward"),
#
# optional methods that a model can define
# inference:
# mainly for AR model. In training, forward() is used while in generation
# inference() is used. For other models, inference may be unnecessary
'inference': (False, "alternative method for inference"),
#
# loss:
# A model can define loss inside its module. This is convenient when
# loss requires some additional information stored in the model.
# If model.loss() is not defined, Loss defined in model.py will be used.
'loss': (False, 'loss defined within model module'),
#
# other_setups:
# other setups functions that should be applied before training
'other_setups': (False, "other setup functions before training"),
#
# flag_validation:
# model.training tells whether this is under trainining or not
# model.flag_validation tells whether the model is on train set or val set
'flag_validation': (False, 'flag to indicate train or validation set'),
'validation': (False, 'deprecated. Please use model.flag_validation'),
#
# finish_up:
# model.finish_up_inference will conduct finish_up work after looping over
# all the data for inference (see nn_manager.py f_inference_wrapper)
#
'finish_up_inference': (False, 'method to finish up work after inference')
}
# create the bag for each type of models
nn_model_keywords_bags = {'default': nn_model_keywords_default}
####
# Methods that a Loss should have
# name: (flag_mandatory, comment)
# name: str, name of the method
# flag_mandatory: bool, whether this method is mandatory to have
# comment: str, comment string on the method
#
####
# Loss function for a typical model
loss_method_keywords_default = {
'compute': (True, "method to comput loss")
}
# Loss function for a GAN model
loss_method_keywords_GAN = {
'compute_gan_D_real': (True, "method to comput loss for GAN dis. on real"),
'compute_gan_D_fake': (True, "method to comput loss for GAN dis. on fake"),
'compute_gan_G': (True, "method to comput loss for GAN gen."),
'compute_aux': (False, "(onlt for GAN-based model), auxialliary loss"),
'compute_feat_match': (False, '(only for GAN-based model), feat-matching'),
'flag_wgan': (False, '(only for GAN-based model), w-gan')
}
# create the bag for each type of Loss functions
loss_method_keywords_bags = {'default': loss_method_keywords_default,
'GAN': loss_method_keywords_GAN}
if __name__ == "__main__":
print("Configurations for nn_manager")
| 3,996 | 34.061404 | 79 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/nn_manager/nn_manager_GAN_ob.py | #!/usr/bin/env python
"""
nn_manager_gan
A simple wrapper to run the training / testing process for GAN
"""
from __future__ import print_function
import time
import datetime
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.display as nii_display
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.op_manager.op_process_monitor as nii_monitor
import core_scripts.op_manager.op_display_tools as nii_op_display_tk
import core_scripts.nn_manager.nn_manager_tools as nii_nn_tools
import core_scripts.nn_manager.nn_manager_conf as nii_nn_manage_conf
import core_scripts.other_tools.debug as nii_debug
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#############################################################
def f_run_one_epoch_GAN(
args, pt_model_G, pt_model_D,
loss_wrapper, \
device, monitor, \
data_loader, epoch_idx,
optimizer_G = None, optimizer_D = None, \
target_norm_method = None):
"""
f_run_one_epoch_GAN:
run one poech over the dataset (for training or validation sets)
Args:
args: from argpase
pt_model_G: pytorch model (torch.nn.Module) generator
pt_model_D: pytorch model (torch.nn.Module) discriminator
loss_wrapper: a wrapper over loss function
loss_wrapper.compute(generated, target)
device: torch.device("cuda") or torch.device("cpu")
monitor: defined in op_procfess_monitor.py
data_loader: pytorch DataLoader.
epoch_idx: int, index of the current epoch
optimizer_G: torch optimizer or None, for generator
optimizer_D: torch optimizer or None, for discriminator
if None, the back propgation will be skipped
(for developlement set)
target_norm_method: method to normalize target data
(by default, use pt_model.normalize_target)
"""
# timer
start_time = time.time()
# loop over samples
for data_idx, (data_in, data_tar, data_info, idx_orig) in \
enumerate(data_loader):
#############
# prepare
#############
# send data to device
if optimizer_G is not None:
optimizer_G.zero_grad()
if optimizer_D is not None:
optimizer_D.zero_grad()
# normalize the target data (for input for discriminator)
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
# there is no way to normalize the data inside loss
# thus, do normalization here
if target_norm_method is None:
normed_target = pt_model_G.normalize_target(data_tar)
else:
normed_target = target_norm_method(data_tar)
else:
nii_display.f_die("target data is required")
# to device (we assume noise will be generated by the model itself)
# here we only provide external condition
data_in = data_in.to(device, dtype=nii_dconf.d_dtype)
############################
# Update Discriminator
############################
####
# train with real
####
pt_model_D.zero_grad()
d_out_real = pt_model_D(data_tar, data_in)
errD_real = loss_wrapper.compute_gan_D_real(d_out_real)
if optimizer_D is not None:
errD_real.backward()
# this should be given by pt_model_D or loss wrapper
#d_out_real_mean = d_out_real.mean()
###
# train with fake
###
# generate sample
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder & autoregressive model
if isinstance(data_tar, torch.Tensor):
data_tar_tm = data_tar.to(device, dtype=nii_dconf.d_dtype)
if args.model_forward_with_file_name:
data_gen = pt_model_G(data_in, data_tar_tm, data_info)
else:
data_gen = pt_model_G(data_in, data_tar_tm)
else:
nii_display.f_print("--model-forward-with-target is set")
nii_display.f_die("but data_tar is not loaded")
else:
if args.model_forward_with_file_name:
# specifcal case when model.forward requires data_info
data_gen = pt_model_G(data_in, data_info)
else:
# normal case for model.forward(input)
data_gen = pt_model_G(data_in)
# data_gen.detach() is required
# https://github.com/pytorch/examples/issues/116
# https://stackoverflow.com/questions/46774641/
d_out_fake = pt_model_D(data_gen.detach(), data_in)
errD_fake = loss_wrapper.compute_gan_D_fake(d_out_fake)
if optimizer_D is not None:
errD_fake.backward()
# get the summed error for discrminator (only for displaying)
errD = errD_real + errD_fake
# update discriminator weight
if optimizer_D is not None:
optimizer_D.step()
############################
# Update Generator
############################
pt_model_G.zero_grad()
d_out_fake_for_G = pt_model_D(data_gen, data_in)
errG_gan = loss_wrapper.compute_gan_G(d_out_fake_for_G)
# if defined, calculate auxilliart loss
if hasattr(loss_wrapper, "compute_aux"):
errG_aux = loss_wrapper.compute_aux(data_gen, data_tar)
else:
errG_aux = torch.zeros_like(errG_gan)
# if defined, calculate feat-matching loss
if hasattr(loss_wrapper, "compute_feat_match"):
errG_feat = loss_wrapper.compute_feat_match(
d_out_real, d_out_fake_for_G)
else:
errG_feat = torch.zeros_like(errG_gan)
# sum loss for generator
errG = errG_gan + errG_aux + errG_feat
if optimizer_G is not None:
errG.backward()
optimizer_G.step()
# construct the loss for logging and early stopping
# only use errG_aux for early-stopping
loss_computed = [
[errG_aux, errD_real, errD_fake, errG_gan, errG_feat],
[True, False, False, False, False]]
# to handle cases where there are multiple loss functions
_, loss_vals, loss_flags = nii_nn_tools.f_process_loss(loss_computed)
# save the training process information to the monitor
end_time = time.time()
batchsize = len(data_info)
for idx, data_seq_info in enumerate(data_info):
# loss_value is supposed to be the average loss value
# over samples in the the batch, thus, just loss_value
# rather loss_value / batchsize
monitor.log_loss(loss_vals, loss_flags, \
(end_time-start_time) / batchsize, \
data_seq_info, idx_orig.numpy()[idx], \
epoch_idx)
# print infor for one sentence
if args.verbose == 1:
monitor.print_error_for_batch(data_idx*batchsize + idx,\
idx_orig.numpy()[idx], \
epoch_idx)
#
# start the timer for a new batch
start_time = time.time()
# lopp done
return
def f_run_one_epoch_WGAN(
args, pt_model_G, pt_model_D,
loss_wrapper, \
device, monitor, \
data_loader, epoch_idx,
optimizer_G = None, optimizer_D = None, \
target_norm_method = None):
"""
f_run_one_epoch_WGAN:
similar to f_run_one_epoch_GAN, but for WGAN
"""
# timer
start_time = time.time()
# This should be moved to model definition
# number of critic (default 5)
num_critic = 5
# clip value
wgan_clamp = 0.01
# loop over samples
for data_idx, (data_in, data_tar, data_info, idx_orig) in \
enumerate(data_loader):
# send data to device
if optimizer_G is not None:
optimizer_G.zero_grad()
if optimizer_D is not None:
optimizer_D.zero_grad()
# prepare data
if isinstance(data_tar, torch.Tensor):
data_tar = data_tar.to(device, dtype=nii_dconf.d_dtype)
# there is no way to normalize the data inside loss
# thus, do normalization here
if target_norm_method is None:
normed_target = pt_model_G.normalize_target(data_tar)
else:
normed_target = target_norm_method(data_tar)
else:
nii_display.f_die("target data is required")
# to device (we assume noise will be generated by the model itself)
# here we only provide external condition
data_in = data_in.to(device, dtype=nii_dconf.d_dtype)
############################
# Update Discriminator
############################
# train with real
pt_model_D.zero_grad()
d_out_real = pt_model_D(data_tar)
errD_real = loss_wrapper.compute_gan_D_real(d_out_real)
if optimizer_D is not None:
errD_real.backward()
d_out_real_mean = d_out_real.mean()
# train with fake
# generate sample
if args.model_forward_with_target:
# if model.forward requires (input, target) as arguments
# for example, for auto-encoder & autoregressive model
if isinstance(data_tar, torch.Tensor):
data_tar_tm = data_tar.to(device, dtype=nii_dconf.d_dtype)
if args.model_forward_with_file_name:
data_gen = pt_model_G(data_in, data_tar_tm, data_info)
else:
data_gen = pt_model_G(data_in, data_tar_tm)
else:
nii_display.f_print("--model-forward-with-target is set")
nii_display.f_die("but data_tar is not loaded")
else:
if args.model_forward_with_file_name:
# specifcal case when model.forward requires data_info
data_gen = pt_model_G(data_in, data_info)
else:
# normal case for model.forward(input)
data_gen = pt_model_G(data_in)
# data_gen.detach() is required
# https://github.com/pytorch/examples/issues/116
d_out_fake = pt_model_D(data_gen.detach())
errD_fake = loss_wrapper.compute_gan_D_fake(d_out_fake)
if optimizer_D is not None:
errD_fake.backward()
d_out_fake_mean = d_out_fake.mean()
errD = errD_real + errD_fake
if optimizer_D is not None:
optimizer_D.step()
# clip weights of discriminator
for p in pt_model_D.parameters():
p.data.clamp_(-wgan_clamp, wgan_clamp)
############################
# Update Generator
############################
pt_model_G.zero_grad()
d_out_fake_for_G = pt_model_D(data_gen)
errG_gan = loss_wrapper.compute_gan_G(d_out_fake_for_G)
errG_aux = loss_wrapper.compute_aux(data_gen, data_tar)
errG = errG_gan + errG_aux
# only update after num_crictic iterations on discriminator
if data_idx % num_critic == 0 and optimizer_G is not None:
errG.backward()
optimizer_G.step()
d_out_fake_for_G_mean = d_out_fake_for_G.mean()
# construct the loss for logging and early stopping
# only use errG_aux for early-stopping
loss_computed = [[errG_aux, errG_gan, errD_real, errD_fake,
d_out_real_mean, d_out_fake_mean,
d_out_fake_for_G_mean],
[True, False, False, False, False, False, False]]
# to handle cases where there are multiple loss functions
loss, loss_vals, loss_flags = nii_nn_tools.f_process_loss(loss_computed)
# save the training process information to the monitor
end_time = time.time()
batchsize = len(data_info)
for idx, data_seq_info in enumerate(data_info):
# loss_value is supposed to be the average loss value
# over samples in the the batch, thus, just loss_value
# rather loss_value / batchsize
monitor.log_loss(loss_vals, loss_flags, \
(end_time-start_time) / batchsize, \
data_seq_info, idx_orig.numpy()[idx], \
epoch_idx)
# print infor for one sentence
if args.verbose == 1:
monitor.print_error_for_batch(data_idx*batchsize + idx,\
idx_orig.numpy()[idx], \
epoch_idx)
#
# start the timer for a new batch
start_time = time.time()
# lopp done
return
def f_train_wrapper_GAN(
args, pt_model_G, pt_model_D, loss_wrapper, device, \
optimizer_G_wrapper, optimizer_D_wrapper, \
train_dataset_wrapper, \
val_dataset_wrapper = None, \
checkpoint_G = None, checkpoint_D = None):
"""
f_train_wrapper_GAN(
args, pt_model_G, pt_model_D, loss_wrapper, device,
optimizer_G_wrapper, optimizer_D_wrapper,
train_dataset_wrapper, val_dataset_wrapper = None,
check_point = None):
A wrapper to run the training process
Args:
args: argument information given by argpase
pt_model_G: generator, pytorch model (torch.nn.Module)
pt_model_D: discriminator, pytorch model (torch.nn.Module)
loss_wrapper: a wrapper over loss functions
loss_wrapper.compute_D_real(discriminator_output)
loss_wrapper.compute_D_fake(discriminator_output)
loss_wrapper.compute_G(discriminator_output)
loss_wrapper.compute_G(fake, real)
device: torch.device("cuda") or torch.device("cpu")
optimizer_G_wrapper:
a optimizer wrapper for generator (defined in op_manager.py)
optimizer_D_wrapper:
a optimizer wrapper for discriminator (defined in op_manager.py)
train_dataset_wrapper:
a wrapper over training data set (data_io/default_data_io.py)
train_dataset_wrapper.get_loader() returns torch.DataSetLoader
val_dataset_wrapper:
a wrapper over validation data set (data_io/default_data_io.py)
it can None.
checkpoint_G:
a check_point that stores every thing to resume training
checkpoint_D:
a check_point that stores every thing to resume training
"""
nii_display.f_print_w_date("Start model training")
##############
## Preparation
##############
# get the optimizer
optimizer_G_wrapper.print_info()
optimizer_D_wrapper.print_info()
optimizer_G = optimizer_G_wrapper.optimizer
optimizer_D = optimizer_D_wrapper.optimizer
epoch_num = optimizer_G_wrapper.get_epoch_num()
no_best_epoch_num = optimizer_G_wrapper.get_no_best_epoch_num()
# get data loader for training set
train_dataset_wrapper.print_info()
train_data_loader = train_dataset_wrapper.get_loader()
train_seq_num = train_dataset_wrapper.get_seq_num()
# get the training process monitor
monitor_trn = nii_monitor.Monitor(epoch_num, train_seq_num)
# if validation data is provided, get data loader for val set
if val_dataset_wrapper is not None:
val_dataset_wrapper.print_info()
val_data_loader = val_dataset_wrapper.get_loader()
val_seq_num = val_dataset_wrapper.get_seq_num()
monitor_val = nii_monitor.Monitor(epoch_num, val_seq_num)
else:
monitor_val = None
# training log information
train_log = ''
model_tags = ["_G", "_D"]
# prepare for DataParallism if available
# pytorch.org/tutorials/beginner/blitz/data_parallel_tutorial.html
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
nii_display.f_die("data_parallel not implemented for GAN")
else:
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
flag_multi_device = False
normtarget_f = None
pt_model_G.to(device, dtype=nii_dconf.d_dtype)
pt_model_D.to(device, dtype=nii_dconf.d_dtype)
# print the network
nii_display.f_print("Setup generator")
nii_nn_tools.f_model_show(pt_model_G, model_type='GAN')
nii_display.f_print("Setup discriminator")
nii_nn_tools.f_model_show(pt_model_D, do_model_def_check=False,
model_type='GAN')
nii_nn_tools.f_loss_show(loss_wrapper, model_type='GAN')
###############################
## Resume training if necessary
###############################
# resume training or initialize the model if necessary
cp_names = nii_nn_manage_conf.CheckPointKey()
if checkpoint_G is not None or checkpoint_D is not None:
for checkpoint, optimizer, pt_model, model_name in \
zip([checkpoint_G, checkpoint_D], [optimizer_G, optimizer_D],
[pt_model_G, pt_model_D], ["Generator", "Discriminator"]):
nii_display.f_print("For %s" % (model_name))
if type(checkpoint) is dict:
# checkpoint
# load model parameter and optimizer state
if cp_names.state_dict in checkpoint:
# wrap the state_dic in f_state_dict_wrapper
# in case the model is saved when DataParallel is on
pt_model.load_state_dict(
nii_nn_tools.f_state_dict_wrapper(
checkpoint[cp_names.state_dict],
flag_multi_device))
# load optimizer state
if cp_names.optimizer in checkpoint:
optimizer.load_state_dict(checkpoint[cp_names.optimizer])
# optionally, load training history
if not args.ignore_training_history_in_trained_model:
#nii_display.f_print("Load ")
if cp_names.trnlog in checkpoint:
monitor_trn.load_state_dic(
checkpoint[cp_names.trnlog])
if cp_names.vallog in checkpoint and monitor_val:
monitor_val.load_state_dic(
checkpoint[cp_names.vallog])
if cp_names.info in checkpoint:
train_log = checkpoint[cp_names.info]
nii_display.f_print("Load check point, resume training")
else:
nii_display.f_print("Load pretrained model and optimizer")
elif checkpoint is not None:
# only model status
#pt_model.load_state_dict(checkpoint)
pt_model.load_state_dict(
nii_nn_tools.f_state_dict_wrapper(
checkpoint, flag_multi_device))
nii_display.f_print("Load pretrained model")
else:
nii_display.f_print("No pretrained model")
# done for resume training
######################
### User defined setup
######################
# Not implemented yet
######################
### Start training
######################
# other variables
flag_early_stopped = False
start_epoch = monitor_trn.get_epoch()
epoch_num = monitor_trn.get_max_epoch()
# select one wrapper, based on the flag in loss definition
if hasattr(loss_wrapper, "flag_wgan") and loss_wrapper.flag_wgan:
f_wrapper_gan_one_epoch = f_run_one_epoch_WGAN
else:
f_wrapper_gan_one_epoch = f_run_one_epoch_GAN
# print
_ = nii_op_display_tk.print_log_head()
nii_display.f_print_message(train_log, flush=True, end='')
# loop over multiple epochs
for epoch_idx in range(start_epoch, epoch_num):
# training one epoch
pt_model_D.train()
pt_model_G.train()
f_wrapper_gan_one_epoch(
args, pt_model_G, pt_model_D,
loss_wrapper, device, \
monitor_trn, train_data_loader, \
epoch_idx, optimizer_G, optimizer_D,
normtarget_f)
time_trn = monitor_trn.get_time(epoch_idx)
loss_trn = monitor_trn.get_loss(epoch_idx)
# if necessary, do validataion
if val_dataset_wrapper is not None:
# set eval() if necessary
if args.eval_mode_for_validation:
pt_model_G.eval()
pt_model_D.eval()
with torch.no_grad():
f_wrapper_gan_one_epoch(
args, pt_model_G, pt_model_D,
loss_wrapper, \
device, \
monitor_val, val_data_loader, \
epoch_idx, None, None, normtarget_f)
time_val = monitor_val.get_time(epoch_idx)
loss_val = monitor_val.get_loss(epoch_idx)
else:
time_val, loss_val = 0, 0
if val_dataset_wrapper is not None:
flag_new_best = monitor_val.is_new_best()
else:
flag_new_best = True
# print information
train_log += nii_op_display_tk.print_train_info(
epoch_idx, time_trn, loss_trn, time_val, loss_val,
flag_new_best, optimizer_G_wrapper.get_lr_info())
# save the best model
if flag_new_best:
for pt_model, tmp_tag in zip([pt_model_G, pt_model_D], model_tags):
tmp_best_name = nii_nn_tools.f_save_trained_name(args, tmp_tag)
torch.save(pt_model.state_dict(), tmp_best_name)
# save intermediate model if necessary
if not args.not_save_each_epoch:
# save model discrminator and generator
for pt_model, optimizer, model_tag in \
zip([pt_model_G, pt_model_D], [optimizer_G, optimizer_D],
model_tags):
tmp_model_name = nii_nn_tools.f_save_epoch_name(
args, epoch_idx, model_tag)
if monitor_val is not None:
tmp_val_log = monitor_val.get_state_dic()
else:
tmp_val_log = None
# save
tmp_dic = {
cp_names.state_dict : pt_model.state_dict(),
cp_names.info : train_log,
cp_names.optimizer : optimizer.state_dict(),
cp_names.trnlog : monitor_trn.get_state_dic(),
cp_names.vallog : tmp_val_log
}
torch.save(tmp_dic, tmp_model_name)
if args.verbose == 1:
nii_display.f_eprint(str(datetime.datetime.now()))
nii_display.f_eprint("Save {:s}".format(tmp_model_name),
flush=True)
# early stopping
if monitor_val is not None and \
monitor_val.should_early_stop(no_best_epoch_num):
flag_early_stopped = True
break
# loop done
nii_op_display_tk.print_log_tail()
if flag_early_stopped:
nii_display.f_print("Training finished by early stopping")
else:
nii_display.f_print("Training finished")
nii_display.f_print("Model is saved to", end = '')
for model_tag in model_tags:
nii_display.f_print("{}".format(
nii_nn_tools.f_save_trained_name(args, model_tag)))
return
if __name__ == "__main__":
print("nn_manager for GAN")
| 24,536 | 38.009539 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/nn_manager/nn_manager_AL.py | #!/usr/bin/env python
"""
nn_manager_AL
A simple wrapper to run the training for active learning
Note:
1. The mode to continue to training does not guanrantee exactly
the same result because selection is based on random sampling. The
random seed for data selection differs.
"""
from __future__ import print_function
import os
import time
import datetime
import numpy as np
import copy
import re
import torch
import torch.nn as nn
import torch.nn.functional as F
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.display as nii_display
import core_scripts.other_tools.str_tools as nii_str_tk
import core_scripts.op_manager.op_process_monitor as nii_monitor
import core_scripts.op_manager.op_display_tools as nii_op_display_tk
import core_scripts.nn_manager.nn_manager_tools as nii_nn_tools
import core_scripts.nn_manager.nn_manager_conf as nii_nn_manage_conf
import core_scripts.nn_manager.nn_manager as nii_nn_manager_base
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
#############################################################
__g_info_separator = ':'
__g_name_separator = ';'
__g_type_tags = ['add', 'remove']
def __print_new_sample_list(cycle_idx, dataset_wrapper, data_idx):
""" print information on the newly added data
"""
mes = 'Active learning cycle {:d}, add samples'.format(cycle_idx)
mes += __g_info_separator + ' '
mes += __g_name_separator.join(dataset_wrapper.get_seq_info())
mes += __g_info_separator + ' '
mes += __g_name_separator.join([str(x) for x in data_idx])
#mes += '\nNumber of samples: {:d}'.format(len(data_idx))
nii_display.f_eprint(mes)
return mes
def __print_excl_sample_list(cycle_idx, dataset_wrapper, data_idx):
""" print information on the newly removed data
"""
mes = 'Before learning cycle {:d}, remove'.format(cycle_idx)
mes += __g_info_separator + ' '
mes += __g_name_separator.join(dataset_wrapper.get_seq_info())
mes += __g_info_separator + ' '
mes += __g_name_separator.join([str(x) for x in data_idx])
#mes += '\nNumber of removed samples: {:d}'.format(len(data_idx))
nii_display.f_eprint(mes)
return mes
def __save_sample_list_buf(list_buff, cache_path):
with open(cache_path, 'w') as file_ptr:
for data_str in list_buff:
file_ptr.write(data_str + '\n')
return
def __cache_name(path, cycle_idx):
return '{:s}_{:03d}.txt'.format(path, cycle_idx)
def __parse_sample_list(mes):
"""
Active learning cycle K, add samples: file1, file2, file3
->
K, add, [file1, file2, file3]
"""
# cycle index
cycle_id = re.findall("[0-9]+", mes.split(__g_info_separator)[0])
cycle_id = int(cycle_id[0])
# type of method
if re.findall(__g_type_tags[1], mes.split(__g_info_separator)[0]):
tag = __g_type_tags[1]
else:
tag = __g_type_tags[0]
# assume that : is not included in the file name
filepart = mes.split(__g_info_separator)[2]
# return the sample list
return cycle_id, tag, \
[int(x.rstrip().lstrip()) for x in filepart.split(__g_name_separator)]
def __load_cached_data_list_file(cache_path):
with open(cache_path, 'r') as file_ptr:
output = [__parse_sample_list(x) for x in file_ptr]
return output
def __print_cycle(cycle_idx, train_s, pool_s):
""" print information added to the error log
"""
return "AL cycle {:d}, {:d}, {:d}".format(cycle_idx, train_s, pool_s)
def __print_AL_info(num_al_cycle, epoch_per_cycle, num_sample_al_cycle, args):
""" print head information to summarize the AL settings
"""
mes = "\nActive learning (pool-based) settings:"
nii_display.f_print(mes)
mes = 'Number of active learning cycle: {:d}'.format(num_al_cycle)
mes += '\nNumber of epochs per cycle: {:d}'.format(epoch_per_cycle)
mes += '\nNumber of new samples per cycle: {:d}'.format(num_sample_al_cycle)
if args.active_learning_use_new_data_only:
mes += '\nUse retrieved data for fine-tuning model'
else:
mes += '\nUse seed + retrieved data for model training'
if args.active_learning_with_replacement:
mes += '\nRetrieve data w/ replacement'
else:
mes += '\nRetrieve data w/o replacement'
mes += '\n'
nii_display.f_print(mes, 'normal')
return
def _f_copy_subset(dataset_wrapper, data_idx):
""" return a deepcopy of dataset that contains data specified by data_idx
"""
# create data that contains selected data only
# the database only contains data index, so it is fast to do deepcopy
tmp_data_wrapper = copy.deepcopy(dataset_wrapper)
tmp_data_wrapper.manage_data(data_idx, 'keep')
return tmp_data_wrapper
def _f_add_data(pool_dataset_wrapper, train_dataset_wrapper, data_idx, args):
"""
"""
# create a copy of the data to be selected from the pool
# the database only contains data index, so it is fast to do deepcopy
tmp_data_wrapper = _f_copy_subset(pool_dataset_wrapper, data_idx)
# Delete data from original pool if we sample without replacement
if not args.active_learning_with_replacement:
pool_dataset_wrapper.manage_data(data_idx, 'delete')
#pool_data_loader = pool_dataset_wrapper.get_loader()
#
if args.active_learning_use_new_data_only:
# only augmented data
nii_display.f_die("Not implemented yet")
else:
# base dataset + augmented data
train_dataset_wrapper.add_dataset(tmp_data_wrapper)
return
def _f_remove_data(pool_dataset_wrapper, data_idx, args):
"""
"""
pool_dataset_wrapper.manage_data(data_idx, 'delete')
return
def f_train_wrapper(args, pt_model, loss_wrapper, device, \
optimizer_wrapper, \
train_dataset_wrapper, \
pool_dataset_wrapper, \
val_dataset_wrapper = None, \
checkpoint = None):
"""
f_train_wrapper(args, pt_model, loss_wrapper, device,
optimizer_wrapper
train_dataset_wrapper, val_dataset_wrapper = None,
check_point = None):
A wrapper to run the training process
Args:
args: argument information given by argpase
pt_model: pytorch model (torch.nn.Module)
loss_wrapper: a wrapper over loss function
loss_wrapper.compute(generated, target)
device: torch.device("cuda") or torch.device("cpu")
optimizer_wrapper:
a wrapper over optimizer (defined in op_manager.py)
optimizer_wrapper.optimizer is torch.optimizer
train_dataset_wrapper:
a wrapper over training data set
train_dataset_wrapper.get_loader() returns torch.DataSetLoader
pool_dataset_wrapper:
a wrapper over pool data set for AL
train_dataset_wrapper.get_loader() returns torch.DataSetLoader
val_dataset_wrapper:
a wrapper over validation data set (data_io/default_data_io.py)
it can None.
check_point:
a check_point that stores every thing to resume training
"""
nii_display.f_print_w_date("Start model training")
##############
## Preparation
##############
##
# Configurations
##
# total number of epoch = epoch per cycle * number of AL cycles
total_epoch_num = optimizer_wrapper.get_epoch_num()
num_al_cycle = np.abs(args.active_learning_cycle_num)
if num_al_cycle == 0:
nii_display.f_die("Number of active learning cycles must be > 0")
epoch_per_cycle = total_epoch_num // num_al_cycle
# set the number of samples take per cycle
num_sample_al_cycle = args.active_learning_new_sample_per_cycle
# if not set, take batch-size of samples per cycle
if num_sample_al_cycle < 1:
num_sample_al_cycle = args.batch_size
#nii_display.f_print("Add {:d} new samples per cycle".format(
# num_sample_al_cycle))
# patience for early stopping on development set
no_best_epoch_num = optimizer_wrapper.get_no_best_epoch_num()
##
# data loader, optimizer, model, ...
##
# get the optimizer
if not args.active_learning_pre_train_epoch_num:
# this information should have been printed during pre-training
optimizer_wrapper.print_info()
optimizer = optimizer_wrapper.optimizer
lr_scheduler = optimizer_wrapper.lr_scheduler
# get data loader for seed training set
if not args.active_learning_pre_train_epoch_num:
train_dataset_wrapper.print_info()
train_data_loader = train_dataset_wrapper.get_loader()
train_seq_num = train_dataset_wrapper.get_seq_num()
# get pool data set for active learning
pool_dataset_wrapper.print_info()
pool_data_loader = pool_dataset_wrapper.get_loader()
pool_seq_num = pool_dataset_wrapper.get_seq_num()
# get the training process monitor
monitor_trn = nii_monitor.Monitor(total_epoch_num, train_seq_num)
# if validation data is provided, get data loader for val set
if val_dataset_wrapper is not None:
if not args.active_learning_pre_train_epoch_num:
val_dataset_wrapper.print_info()
val_data_loader = val_dataset_wrapper.get_loader()
val_seq_num = val_dataset_wrapper.get_seq_num()
monitor_val = nii_monitor.Monitor(total_epoch_num, val_seq_num)
else:
monitor_val = None
# prepare for DataParallism if available
# pytorch.org/tutorials/beginner/blitz/data_parallel_tutorial.html
if torch.cuda.device_count() > 1 and args.multi_gpu_data_parallel:
nii_display.f_die("Not implemented for multiple GPU")
else:
nii_display.f_print("\nUse single GPU: %s\n" % \
(torch.cuda.get_device_name(device)))
flag_multi_device = False
normtarget_f = None
pt_model.to(device, dtype=nii_dconf.d_dtype)
##
# misc
##
# print the network
if not args.active_learning_pre_train_epoch_num:
nii_nn_tools.f_model_show(pt_model)
nii_nn_tools.f_loss_show(loss_wrapper)
# key names, used when saving *.epoch.pt
cp_names = nii_nn_manage_conf.CheckPointKey()
# training log information
train_log = ''
# buffer for selected data list
al_mes_buff = []
###############################
## Resume training if necessary
###############################
##
# load epoch*.pt (which contains no infor on previously selected data)
##
train_log = nii_nn_tools.f_load_checkpoint(
checkpoint, args, flag_multi_device, pt_model,
optimizer, monitor_trn, monitor_val, lr_scheduler)
##
# load selected or removed utterances in previous cycles,
##
# index of the starting cycle (default 0)
start_cycle = 0
if len(args.active_learning_cache_dataname_path) \
and os.path.isfile(args.active_learning_cache_dataname_path):
nii_display.f_print("Load cache of selected (removed data)")
# Load from file
cached_data_status = __load_cached_data_list_file(
args.active_learning_cache_dataname_path)
# For each cycle, update the pool and training set
for entry in cached_data_status:
# retrieve the log
cycle_id, method_type, data_idx = entry[0], entry[1], entry[2]
if method_type == __g_type_tags[1]:
# for removing
# print the information
mes = __print_excl_sample_list(
cycle_id, _f_copy_subset(pool_dataset_wrapper, data_idx),
data_idx)
# remove previously removed data from pool
_f_remove_data(pool_dataset_wrapper, data_idx, args)
else:
# for selected data (either active or passive)
# print the information
mes = __print_new_sample_list(
cycle_id, _f_copy_subset(pool_dataset_wrapper, data_idx),
data_idx)
# add previously selected from the pool to the training set
_f_add_data(pool_dataset_wrapper, train_dataset_wrapper,
data_idx, args)
#
pool_data_loader = pool_dataset_wrapper.get_loader()
train_data_loader = train_dataset_wrapper.get_loader()
#
al_mes_buff.append(mes)
if len(cached_data_status):
# data selectio and removing should have been done in cycle_id
# thus, we start from the next cycle
start_cycle = cycle_id + 1
else:
pass
######################
### User defined setup
######################
if hasattr(pt_model, "other_setups"):
nii_display.f_print("Conduct User-defined setup")
pt_model.other_setups()
# This should be merged with other_setups
if hasattr(pt_model, "g_pretrained_model_path") and \
hasattr(pt_model, "g_pretrained_model_prefix"):
nii_display.f_print("Load pret-rained models as part of this mode")
nii_nn_tools.f_load_pretrained_model_partially(
pt_model, pt_model.g_pretrained_model_path,
pt_model.g_pretrained_model_prefix)
##############################
### Start active learning loop
##############################
# other variables
# initialize flag to save state of early stopping
flag_early_stopped = False
# initialize the starting epoch number
# this counts the epoch number across different cycles
start_epoch = monitor_trn.get_epoch()
epoch_counter = start_epoch
# get the total number of epochs to run
total_epoch_num = monitor_trn.get_max_epoch()
# a buf to store the path of trained models per cycle
saved_model_path_buf = []
# print active learining general information
__print_AL_info(num_al_cycle, epoch_per_cycle, num_sample_al_cycle, args)
# print training log (if available from resumed checkpoint)
_ = nii_op_display_tk.print_log_head()
nii_display.f_print_message(train_log, flush=True, end='')
# sanity check
if start_epoch // epoch_per_cycle != start_cycle:
nii_display.f_print("Training cycle in {:s} != that in {:s}".format(
args.trained_model, args.active_learning_cache_dataname_path))
nii_display.f_print(" {:d} // {:d} != {:d}".format(
start_epoch, epoch_per_cycle, start_cycle))
nii_display.f_die("Fail to resume training")
#
# currently, we can only restat from the 1st epoch in each active learning
# cycle. Note that, monitor_trn.get_epoch() returns the current epoch idx
if start_epoch > 0 and start_epoch % epoch_per_cycle != 0:
mes = "The checkpoint is not the last epoch in one cycle"
nii_display.f_print(mes)
nii_display.f_die("Fail to resume training")
# loop over cycles
for cycle_idx in range(start_cycle, num_al_cycle):
# Pool data has been used up. Training ends.
if pool_dataset_wrapper.get_seq_num() < 1:
break
########
# select the samples
########
# There are many methods to select samples
# we require pt_model to define one of the method
# I. Pool-based, no knowedge on seed data:
# al_retrieve_data(pool_data_loader,
# num_sample_al_cycle)
# Only use model to score each data in pool_data_loader
#
# II. Pool-based, w/ knowedge on seed data
# al_retrieve_data_knowing_train(train_data_loader,
# pool_data_loader,
# num_sample_al_cycle)
# Select sample from pool given knowlege of train seed data
#
# III. Pool-based, but to exclude data first
# al_exclude_data(train_data_loader,pool_data_loader,
# num_sample_al_cycle)
# Exclude samples from the pool set
# If provided, this function will be called first before executing
# Pool-based II and I
#
# save current model flag
tmp_train_flag = True if pt_model.training else False
if args.active_learning_train_model_for_retrieval:
pt_model.train()
else:
pt_model.eval()
# exclude data if necesary
if hasattr(pt_model, 'al_exclude_data'):
# select least useful data
data_idx = pt_model.al_exclude_data(
pool_data_loader, num_sample_al_cycle)
# convert data index to int
data_idx = [int(x) for x in data_idx]
# print datat to be excluded
mes = __print_excl_sample_list(
cycle_idx, _f_copy_subset(pool_dataset_wrapper, data_idx),
data_idx)
al_mes_buff.append(mes)
# remove the pool
_f_remove_data(pool_dataset_wrapper, data_idx, args)
pool_data_loader = pool_dataset_wrapper.get_loader()
# retrieve data from the pool to training set
if hasattr(pt_model, 'al_retrieve_data'):
data_idx = pt_model.al_retrieve_data(
pool_data_loader, num_sample_al_cycle)
elif hasattr(pt_model, 'al_retrieve_data_knowing_train'):
data_idx = pt_model.al_retrieve_data_knowing_train(
train_data_loader, pool_data_loader, num_sample_al_cycle)
else:
nii_display.f_die("model must define al_retrieve_data")
# convert data index to int
data_idx = [int(x) for x in data_idx]
# set flag back
if tmp_train_flag:
pt_model.train()
else:
pt_model.eval()
########
# Create Dataset wrapper for the new data add new data to train set
########
# _f_add_data alters pool_dataset_wrapper
# thus, we print the list based on pool_dataset before _f_add_data
mes = __print_new_sample_list(
cycle_idx, _f_copy_subset(pool_dataset_wrapper, data_idx), data_idx)
al_mes_buff.append(mes)
_f_add_data(pool_dataset_wrapper, train_dataset_wrapper, data_idx, args)
# prepare for training
# get the data loader from the new training and pool sets
pool_data_loader = pool_dataset_wrapper.get_loader()
train_data_loader = train_dataset_wrapper.get_loader()
# number of samples in current training and pool sets
train_seq_num = train_dataset_wrapper.get_seq_num()
pool_seq_num = pool_dataset_wrapper.get_seq_num()
# because the number of training data changes in each cycle, we need to
# create a temporary monitor for training
tmp_monitor_trn = nii_monitor.Monitor(epoch_per_cycle, train_seq_num)
########
# training using the new training set
# epoch_counter: a global counter of training epoch, across all cycles
# tmp_start_epoch: index of starting epoch within one cycle
# tmp_epoch_idx: index of epoch within one cycle
tmp_start_epoch = epoch_counter % epoch_per_cycle
for tmp_epoch_idx in range(tmp_start_epoch, epoch_per_cycle):
# If the model has a member for g_epoch_idx
# save the index
if hasattr(pt_model, 'g_epoch_idx'):
pt_model.g_epoch_idx = epoch_counter
# If the model has a member for g_epoch_idx
# save the index
# cycle index should be updated after selecting the data
if hasattr(pt_model, 'g_cycle_idx'):
pt_model.g_cycle_idx = cycle_idx
# training one epoch
pt_model.train()
if hasattr(pt_model, 'flag_validation'):
pt_model.flag_validation = False
nii_nn_manager_base.f_run_one_epoch(
args, pt_model, loss_wrapper, device, \
tmp_monitor_trn, train_data_loader, \
tmp_epoch_idx, optimizer, normtarget_f)
# get the time and loss for this epoch
time_trn = tmp_monitor_trn.get_time(tmp_epoch_idx)
loss_trn = tmp_monitor_trn.get_loss(tmp_epoch_idx)
# if necessary, forward pass on development set
if val_dataset_wrapper is not None:
# set eval() if necessary
if args.eval_mode_for_validation:
pt_model.eval()
if hasattr(pt_model, 'flag_validation'):
pt_model.flag_validation = True
with torch.no_grad():
nii_nn_manager_base.f_run_one_epoch(
args, pt_model, loss_wrapper, \
device, monitor_val, val_data_loader, \
epoch_counter, None, normtarget_f)
time_val = monitor_val.get_time(epoch_counter)
loss_val = monitor_val.get_loss(epoch_counter)
# update lr rate scheduler if necessary
if lr_scheduler.f_valid():
lr_scheduler.f_step(loss_val)
else:
time_val = monitor_val.get_time(epoch_counter)
loss_val = monitor_val.get_loss(epoch_counter)
#time_val, loss_val = 0, 0
# wether this is the new best trained epoch?
if val_dataset_wrapper is not None:
flag_new_best = monitor_val.is_new_best()
else:
flag_new_best = True
# print information
info_mes = [optimizer_wrapper.get_lr_info(),
__print_cycle(cycle_idx, train_seq_num, pool_seq_num)]
train_log += nii_op_display_tk.print_train_info(
tmp_epoch_idx, time_trn, loss_trn, time_val, loss_val,
flag_new_best, ', '.join([x for x in info_mes if x]))
# save the best model if necessary
if flag_new_best or args.force_save_lite_trained_network_per_epoch:
tmp_best_name = nii_nn_tools.f_save_trained_name(args)
torch.save(pt_model.state_dict(), tmp_best_name)
# save intermediate model if necessary
# we only say the last epoch in each cycle
if not args.not_save_each_epoch \
and tmp_epoch_idx == (epoch_per_cycle - 1):
# we save the global epoch counter into the checkpoint
monitor_trn.log_epoch(epoch_counter)
# name of the checkpoint
tmp_model_name = nii_nn_tools.f_save_epoch_name(
args, cycle_idx, '_epoch_{:03d}'.format(tmp_epoch_idx),
'_al_cycle')
if monitor_val is not None:
tmp_val_log = monitor_val.get_state_dic()
else:
tmp_val_log = None
if lr_scheduler.f_valid():
lr_scheduler_state = lr_scheduler.f_state_dict()
else:
lr_scheduler_state = None
# save
tmp_dic = {
cp_names.state_dict : pt_model.state_dict(),
cp_names.info : train_log,
cp_names.optimizer : optimizer.state_dict(),
cp_names.trnlog : monitor_trn.get_state_dic(),
cp_names.vallog : tmp_val_log,
cp_names.lr_scheduler : lr_scheduler_state
}
torch.save(tmp_dic, tmp_model_name)
if args.verbose == 1:
nii_display.f_eprint(str(datetime.datetime.now()))
nii_display.f_eprint("Save {:s}".format(tmp_model_name),
flush=True)
#
epoch_counter += 1
# loop done for epoch per cycle
# always save the trained model for each cycle
suffix = '_al_cycle_{:03d}'.format(cycle_idx)
tmp_best_name = nii_nn_tools.f_save_trained_name(args, suffix)
torch.save(pt_model.state_dict(), tmp_best_name)
saved_model_path_buf.append(tmp_best_name)
# save selected data for each cycle
__save_sample_list_buf(
al_mes_buff,
__cache_name(args.active_learning_cache_dataname_save, cycle_idx))
# loop for AL cycle
nii_op_display_tk.print_log_tail()
nii_display.f_print("Training finished")
nii_display.f_print("Models from each cycle are saved to:")
for path in saved_model_path_buf:
nii_display.f_print("{}".format(path), 'normal')
return
if __name__ == "__main__":
print("nn_manager_AL")
| 25,279 | 38.254658 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/config_parse/arg_parse.py | #!/usr/bin/env python
"""
config_parse
Argument parse
"""
from __future__ import absolute_import
import os
import sys
import argparse
import core_scripts.other_tools.list_tools as nii_list_tools
import core_scripts.other_tools.display as nii_display
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#############################################################
# argparser
#
def f_args_parsed(argument_input = None):
""" Arg_parse
"""
parser = argparse.ArgumentParser(
description='General argument parse')
######
# lib
mes = 'module of model definition (default model, model.py will be loaded)'
parser.add_argument('--module-model', type=str, default="model", help=mes)
mes = 'module of configuration (default config, config.py will be loaded)'
parser.add_argument('--module-config', type=str, default="config",
help=mes)
mes = 'module of auxiliary model definition (in case this is needed)'
parser.add_argument('--module-model-aux', type=str, default="", help=mes)
######
# Training settings
mes = 'batch size for training/inference (default: 1)'
parser.add_argument('--batch-size', type=int, default=1, help=mes)
mes = 'number of mini-batches to accumulate (default: 1)'
parser.add_argument('--size-accumulate-grad', type=int, default=1, help=mes)
mes = 'number of epochs to train (default: 50)'
parser.add_argument('--epochs', type=int, default=50, help=mes)
mes = 'number of no-best epochs for early stopping (default: 5)'
parser.add_argument('--no-best-epochs', type=int, default=5, help=mes)
mes = 'force to save trained-network.pt per epoch, '
mes += 'no matter whether the epoch is currently the best.'
parser.add_argument('--force-save-lite-trained-network-per-epoch',
action='store_true', default=False, help=mes)
mes = 'sampler (default: None). Default sampler is random shuffler. '
mes += 'Option 1: block_shuffle_by_length, shuffle data by length'
parser.add_argument('--sampler', type=str, default='None', help=mes)
parser.add_argument('--lr', type=float, default=0.0001,
help='learning rate (default: 0.0001)')
mes = 'learning rate decaying factor, using '
mes += 'torch.optim.lr_scheduler.ReduceLROnPlateau(patience=no-best-epochs,'
mes += ' factor=lr-decay-factor). By default, no decaying is used.'
mes += ' Training stopped after --no-best-epochs.'
parser.add_argument('--lr-decay-factor', type=float, default=-1.0, help=mes)
mes = 'lr scheduler: 0: ReduceLROnPlateau (default); 1: StepLR; '
mes += 'this option is set on only when --lr-decay-factor > 0. '
mes += 'Please check core_scripts/op_manager/lr_scheduler.py '
mes += 'for detailed hyper config for each type of lr scheduler'
parser.add_argument('--lr-scheduler-type', type=int, default=0, help=mes)
mes = 'lr patience: patience for torch_optim_steplr.ReduceLROnPlateau '
mes += 'this option is used only when --lr-scheduler-type == 0. '
parser.add_argument('--lr-patience', type=int, default=5, help=mes)
mes = 'lr step size: step size for torch.optim.lr_scheduler.StepLR'
mes += 'this option is used only when --lr-scheduler-type == 1. '
parser.add_argument('--lr-steplr-size', type=int, default=5, help=mes)
mes = 'L2 penalty on weight (default: not use). '
mes += 'It corresponds to the weight_decay option in Adam'
parser.add_argument('--l2-penalty', type=float, default=-1.0, help=mes)
mes = 'gradient norm (torch.nn.utils.clip_grad_norm_ of Pytorch)'
mes += 'default (-1, not use)'
parser.add_argument('--grad-clip-norm', type=float, default=-1.0,
help=mes)
parser.add_argument('--no-cuda', action='store_true', default=False,
help='disables CUDA training')
parser.add_argument('--seed', type=int, default=1,
help='random seed (default: 1)')
mes = 'turn model.eval() on validation set (default: false)'
parser.add_argument('--eval-mode-for-validation', \
action='store_true', default=False, help=mes)
mes = 'if model.forward(input, target), please set this option on. '
mes += 'This is used for autoregressive model, auto-encoder, and so on. '
mes += 'When --model-forward-with-file-name is also on, '
mes += 'model.forward(input, target, file_name) should be defined'
parser.add_argument('--model-forward-with-target', \
action='store_true', default=False, help=mes)
mes = 'if model.forward(input, file_name), please set option on. '
mes += 'This is used with forward requires file name of the data. '
mes += 'When --model-forward-with-target is also on, '
mes += 'model.forward(input, target, file_name) should be defined'
parser.add_argument('--model-forward-with-file-name', \
action='store_true', default=False, help=mes)
mes = 'shuffle data? (default true). Set --shuffle will turn off shuffling'
parser.add_argument('--shuffle', action='store_false', \
default=True, help=mes)
mes = 'number of parallel workers to load data (default: 0)'
parser.add_argument('--num-workers', type=int, default=0, help=mes)
mes = 'use DataParallel to levarage multiple GPU (default: False)'
parser.add_argument('--multi-gpu-data-parallel', \
action='store_true', default=False, help=mes)
mes = 'way to concatenate multiple datasets: '
mes += 'concatenate: simply merge two datasets as one large dataset. '
mes += 'batch_merge: make a minibatch by drawing one sample from each set. '
mes += '(default: concatenate)'
parser.add_argument('--way-to-merge-datasets', type=str, \
default='concatenate', help=mes)
mes = "Ignore invalid data? the length of features does not match"
parser.add_argument('--ignore-length-invalid-data',
action='store_true', default=False, help=mes)
mes = "Ignore existing cache file dic"
parser.add_argument('--ignore-cached-file-infor',
action='store_true', default=False, help=mes)
mes = "External directory to store cache file dic"
parser.add_argument('--path-cache-file', type=str, default="", help=mes)
mes = "Skip scanning data directories (by default False)"
parser.add_argument('--force-skip-datadir-scanning',
action='store_true', default=False, help=mes)
######
# options to save model / checkpoint
parser.add_argument('--save-model-dir', type=str, \
default="./", \
help='save model to this direcotry (default ./)')
mes = 'do not save model after every epoch (default: False)'
parser.add_argument('--not-save-each-epoch', action='store_true', \
default=False, help=mes)
mes = 'name prefix of saved model (default: epoch)'
parser.add_argument('--save-epoch-name', type=str, default="epoch", \
help=mes)
mes = 'name of trained model (default: trained_network)'
parser.add_argument('--save-trained-name', type=str, \
default="trained_network", help=mes)
parser.add_argument('--save-model-ext', type=str, default=".pt",
help='extension name of model (default: .pt)')
mes = 'save model after every N mini-batches (default: 0, not use)'
parser.add_argument('--save-model-every-n-minibatches', type=int,
default=0, help=mes)
#######
# options for active learning
mes = 'Number of active leaning cycles'
parser.add_argument('--active-learning-cycle-num', type=int, default=0,
help = mes)
mes = 'Whetehr use base traing set with new samples? (default True)'
parser.add_argument('--active-learning-use-new-data-only',
action='store_true', default=False, help = mes)
mes = 'Number of samples selected per cycle? (default =batch size)'
parser.add_argument('--active-learning-new-sample-per-cycle', type=int,
default=0, help = mes)
mes = 'Use model.train() during data retrieval (defaul False)'
parser.add_argument('--active-learning-train-model-for-retrieval',
action='store_true', default=False, help = mes)
mes = 'Retrieve data with replacement (defaul True)'
parser.add_argument('--active-learning-with-replacement',
action='store_true', default=False, help = mes)
mes = 'Number of pre-trainining epochs before active learniing (defaul 0)'
parser.add_argument('--active-learning-pre-train-epoch-num', type=int,
default=0, help=mes)
mes = 'Name of the cache file to store names of selected or removed data'
parser.add_argument('--active-learning-cache-dataname-save', type=str,
default="cache_al_data_log", help=mes)
mes = 'Path to the cache file that stores names of selected or removed data'
parser.add_argument('--active-learning-cache-dataname-path', type=str,
default="", help=mes)
#######
# options to load model
mes = 'a trained model for inference or resume training '
parser.add_argument('--trained-model', type=str, \
default="", help=mes + "(default: '')")
mes = 'do not load previous training error information.'
mes += " Load only model para. and optimizer state (default: false)"
parser.add_argument('--ignore-training-history-in-trained-model',
action='store_true', \
default=False, help=mes)
mes = 'do not load previous training statistics in optimizer.'
mes += " (default: false)"
parser.add_argument('--ignore-optimizer-statistics-in-trained-model',
action='store_true', \
default=False, help=mes)
mes = 'load pre-trained model even if there is mismatch on the number of'
mes += " parameters. Mismatched part will not be loaded (default: false)"
parser.add_argument('--allow-mismatched-pretrained-model',
action='store_true', \
default=False, help=mes)
mes = 'run inference mode (default: False, run training script)'
parser.add_argument('--inference', action='store_true', \
default=False, help=mes)
mes = 'run model conversion script (default: False)'
parser.add_argument('--epoch2pt', action='store_true', \
default=False, help=mes)
mes = 'inference only on data whose minibatch index is within the range of '
mes = mes + '[--inference-sample-start-index, --inference-sample-end-index)'
mes = mes + 'default: 0, starting from the 1st data'
parser.add_argument('--inference-sample-start-index', type=int, default=0,
help=mes)
mes = 'inference only on data whose minibatch index is within the range of '
mes = mes + '[--inference-sample-start-index, --inference-sample-end-index)'
mes = mes + 'default: -1, until the end of all data'
parser.add_argument('--inference-sample-end-index', type=int, default=-1,
help=mes)
mes = 'inference data list. A list of file names that should '
mes = mes + 'be processed during the inference stage. '
mes = mes + 'If such a data list is provided, only data listed will '
mes = mes + 'be processed.'
parser.add_argument('--inference-data-list', type=str, default="",
help=mes)
#######
# options to output
mes = 'path to save generated data (default: ./output)'
parser.add_argument('--output-dir', type=str, default="./output", \
help=mes)
# options to output
mes = 'prefix added to file name (default: no string)'
parser.add_argument('--output-filename-prefix', type=str, default="", \
help=mes)
mes = 'truncate input data sequences so that the max length < N.'
mes += ' (default: -1, not do truncating at all)'
parser.add_argument('--trunc-input-length-for-inference', type=int,
default=-1, help=mes)
mes = 'truncate input data overlap length (default: 5)'
parser.add_argument('--trunc-input-overlap', type=int, default=5, help=mes)
mes = 'which optimizer to use (Adam | SGD, default: Adam)'
parser.add_argument('--optimizer', type=str, default='Adam', help=mes)
mes = 'verbose level 0: nothing; 1: print error per utterance'
mes = mes + ' (default: 1)'
parser.add_argument('--verbose', type=int, default=1,
help=mes)
#######
# options for debug mode
mes = 'debug mode, each epoch only uses a specified number of mini-batches'
mes += ' (default: 0, not used)'
parser.add_argument('--debug-batch-num', type=int, default=0, help=mes)
#######
# options for user defined
mes = 'a temporary flag without specific purpose.'
mes += 'User should define args.temp_flag only for temporary usage.'
parser.add_argument('--temp-flag', type=str, default='', help=mes)
mes = 'reverse the order when loading data from the dataset.'
mes += 'This should not not used if --sampler block_shuffle_by_length '
parser.add_argument('--flag-reverse-data-loading-order',
action='store_true', default=False, help=mes)
#######
# backend options
parser.add_argument('--cudnn-deterministic-toggle', action='store_false', \
default=True,
help='use cudnn-deterministic? (default true)')
parser.add_argument('--cudnn-benchmark-toggle', action='store_true', \
default=False,
help='use cudnn-benchmark? (default false)')
#######
# profile options
mes = "options to setup Pytorch profile. It must be a string like A-B-C-D"
mes += ' where A, B, C, D are integers. Meanining of these options are in'
mes += ' torch.profiler.schedule. Default 1-1-3-2.'
parser.add_argument('--wait-warmup-active-repeat', type=str,
default='1-1-3-2',
help=mes)
mes = "directory to save profiling output. Default ./log_profile"
parser.add_argument('--profile-output-dir', type=str,
default='./log_profile')
#######
# data options
mes = 'option to set silence_handler on waveform data.\n'
mes += ' 0: do nothing, use the data as it is (default) \n'
mes += ' 1: remove segments with small energy, use other segments\n'
mes += ' 2: keep only segments with small energy, remove other segments\n'
mes += ' 3: remove segments with small energy only at begining and end\n'
mes += 'Code in core_scripts.data_io.wav_tools.silence_handler. '
mes += 'This option is used when input or output contains only waveform. '
mes += 'It only processes waveform. Other features will not be trimmed.'
parser.add_argument('--opt-wav-silence-handler', type=int,
default=0, help=mes)
mes = 'update data length in internal buffer if data length is changed '
mes += 'by augmentation method. This is useful, for example, when using '
mes += '--sampler block_shuffle_by_length --opt-wav-silence-handler 3 '
mes += 'or using other data augmentation method changes data length.'
parser.add_argument('--force-update-seq-length', action='store_true', \
default=False, help=mes)
#
# done
if argument_input is not None:
return parser.parse_args(argument_input)
else:
return parser.parse_args()
if __name__ == "__main__":
pass
| 16,184 | 41.704485 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/config_parse/config_parse.py | #!/usr/bin/env python
"""
config_parse
Configuration parser
"""
from __future__ import absolute_import
import os
import sys
import configparser
import core_scripts.other_tools.list_tools as nii_list_tools
import core_scripts.other_tools.display as nii_display
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
class ConfigParse:
""" ConfigParse
class to parse input configuration file
"""
def __init__(self, config_path):
""" initialization
"""
# get configuration path
self.m_config_path = None
if os.path.isfile(config_path):
self.m_config_path = config_path
else:
nii_display.f_die("Cannot find %s" % (config_path), 'error')
# path configuration file
self.m_config = self.f_parse()
if self.m_config is None:
nii_display.f_die("Fail to parse %s" % (config_path), 'error')
# done
return
def f_parse(self):
""" f_parse
parse the configuration file
"""
if self.m_config_path is not None:
tmp_config = configparser.ConfigParser()
tmp_config.read(self.m_config_path)
return tmp_config
else:
nii_display.f_print("No config file provided", 'error')
return None
def f_retrieve(self, keyword, section_name=None, config_type=None):
""" f_retrieve(self, keyword, section_name=None, config_type=None)
retrieve the keyword from config file
Return:
value: string, int, float
Parameters:
keyword: 'keyword' to be retrieved
section: which section is this keyword in the config.
None will search all the config sections and
return the first
config_type: which can be 'int', 'float', or None.
None will return the value as a string
"""
tmp_value = None
if section_name is None:
# if section is not given, search all the sections
for section_name in self.m_config.sections():
tmp_value = self.f_retrieve(keyword, section_name, \
config_type)
if tmp_value is not None:
break
elif section_name in self.m_config.sections() or \
section_name == 'DEFAULT':
tmp_sec = self.m_config[section_name]
# search a specific section
if config_type == 'int':
tmp_value = tmp_sec.getint(keyword, fallback=None)
elif config_type == 'float':
tmp_value = tmp_sec.getfloat(keyword, fallback=None)
elif config_type == 'bool':
tmp_value = tmp_sec.getboolean(keyword, fallback=None)
else:
tmp_value = tmp_sec.get(keyword, fallback=None)
else:
nii_display.f_die("Unknown section %s" % (section_name))
return tmp_value
| 3,108 | 31.051546 | 74 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/core_scripts/config_parse/__init__.py | 0 | 0 | 0 | py |
|
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/tutorials/plot_tools/plot_rainbow.py | #!/usr/bin/python
"""
Modified based on https://gist.github.com/jesseengel/e223622e255bd5b8c9130407397a0494
Note:
1. Librosa is required
2. Rainbowgram is a way for plotting magnitiude and phase CQT spectrogram,
not a new type of feature
"""
import os
import sys
import numpy as np
import matplotlib
import matplotlib.pyplot as plt
from matplotlib.colors import LinearSegmentedColormap
import scipy.io.wavfile
import librosa
cdict = {'red': ((0.0, 0.0, 0.0),
(1.0, 0.0, 0.0)),
'green': ((0.0, 0.0, 0.0),
(1.0, 0.0, 0.0)),
'blue': ((0.0, 0.0, 0.0),
(1.0, 0.0, 0.0)),
'alpha': ((0.0, 1.0, 1.0),
(1.0, 0.0, 0.0))
}
my_mask = matplotlib.colors.LinearSegmentedColormap('MyMask', cdict)
plt.register_cmap(cmap=my_mask)
def transform(data, sampling_rate, frame_shift,
notes_per_octave, over_sample,
octaves, res_factor, base_note):
ccqt = librosa.cqt(data, sr=sampling_rate, hop_length=frame_shift,
bins_per_octave=int(notes_per_octave*over_sample),
n_bins=int(octaves * notes_per_octave * over_sample),
filter_scale=res_factor,
fmin=librosa.note_to_hz(base_note))
mag, pha = librosa.magphase(ccqt)
return mag, pha
def plot_rainbow(waveform_data, sampling_rate,
fig, axis,
frame_shift = 256, frame_length = 512, fft_points = 4096,
base_note = 'C1', over_sample = 4, octaves = 8,
notes_per_octave = 12, warping_fac = 2.0,
res_factor = 0.8, magRange = None, phaRange = None):
#https://gist.github.com/jesseengel/e223622e255bd5b8c9130407397a0494
# the range to plot
base = int(base_note[1])
yticks_labels = ['C' + str(x+base) for x in np.arange(octaves)]
yticks = [notes_per_octave * 0 * over_sample + x * over_sample * notes_per_octave for x in np.arange(len(yticks_labels)) ]
if magRange is None or phaRange is None:
magRange = [None, None]
phaRange = [None, None]
data = waveform_data
sr = sampling_rate
mag, pha = transform(data, sr, frame_shift, notes_per_octave, over_sample, octaves, res_factor, base_note)
phaData = np.diff(np.unwrap(np.angle(pha))) / np.pi
magData = (librosa.power_to_db(mag ** 2, amin=1e-13, top_db=70, ref=np.max) / 70) + 1.
magData = magData[:, 0:(magData.shape[1]-1)]
axis.matshow(phaData, cmap=plt.cm.rainbow, aspect='auto', origin='lower')
axis.matshow(magData, cmap=my_mask, aspect='auto', origin='lower')
axis.set_yticks(yticks)
axis.set_yticklabels(yticks_labels)
axis.set_ylabel('Octave')
return fig, axis
if __name__ == "__main__":
print("plot_rainbogram")
| 2,897 | 35.225 | 126 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/tutorials/plot_tools/plot_lib.py | #!/usr/bin/env python
"""
Library of plotting functions
"""
from __future__ import absolute_import
from __future__ import print_function
from mpl_toolkits.mplot3d import Axes3D
import matplotlib
import matplotlib.pyplot as plt
import numpy as np
import sys
import copy
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
##################
## Basic functions
##################
def plot_signal(data, fig, axis, config_dic):
""" plot signal as 1D sequence, using matplotlib.plot
Args
----
data: np.array, (L, 2) or (L, 1) or (L, )
fig: matplotlib figure handle
axis: matplotlib axis handle
config: dictionary for axis.bar
L: number of vertial bar groups
When data is (L, 2), use data[:, 0] as x, data[:, 1] as y
Optional field in config_dic['plot_bar']:
6. other options for matplotlib.pyplot.bar
"""
if type(data) is list:
print("plot_signal no longer support list input")
print("data will be converted to np.array")
x = data[0]
y = data[1]
if "plot_signal" in config_dic:
axis.plot(x, y, **config_dic["plot_signal"])
else:
axis.plot(x, y)
elif type(data) is np.ndarray:
if data.ndim == 2 and data.shape[1] == 2:
x = data[:, 0]
y = data[:, 1]
elif data.ndim == 2 and data.shape[1] == 1:
y = data[:, 0]
x = np.arange(y.shape[0])
elif data.ndim == 1:
y = data
x = np.arange(y.shape[0])
else:
print("dimension of data is not supported")
sys.exit(1)
if "plot_signal" in config_dic:
axis.plot(x, y, **config_dic["plot_signal"])
else:
# default configuration
axis.plot(x, y)
else:
print("Type of data is not np.array")
sys.exit(1)
return fig, axis
def plot_hist(data, fig, axis, config_dic):
if "plot_hist" in config_dic:
axis.hist(data, **config_dic["plot_hist"])
else:
# default configuration
axis.hist(data, histtype='step', density=True, bins=100)
return fig, axis
def plot_imshow(data, fig, axis, config_dic):
"""
"""
plot_colorbar = False
if "plot_imshow" in config_dic:
temp_dic = copy.deepcopy(config_dic["plot_imshow"])
if "colorbar" in temp_dic:
plot_colorbar = temp_dic['colorbar']
temp_dic.pop("colorbar")
pos = axis.imshow(data, **temp_dic)
else:
# default configuration
pos = axis.imshow(data, cmap='jet', origin='lower', aspect='auto')
if type(plot_colorbar) is dict:
fig.colorbar(pos, **plot_colorbar)
elif plot_colorbar:
fig.colorbar(pos)
else:
pass
return fig, axis
def plot_scatter(data, fig, axis, config_dic):
"""
"""
if type(data) is list:
x = data[0]
y = data[1]
# size
s = data[2] if len(data) > 2 else None
# color
c = data[3] if len(data) > 3 else None
else:
x = data[:, 0]
y = data[:, 1]
s = data[:, 2] if data.shape[1] > 2 else None
c = data[:, 3] if data.shape[1] > 3 else None
if "plot_scatter" in config_dic:
if "s" in config_dic["plot_scatter"]:
if "c" in config_dic["plot_scatter"]:
axis.scatter(x, y, **config_dic["plot_scatter"])
else:
axis.scatter(x, y, c = c, **config_dic["plot_scatter"])
else:
if "c" in config_dic["plot_scatter"]:
axis.scatter(x, y, s, **config_dic["plot_scatter"])
else:
axis.scatter(x, y, s, c, **config_dic["plot_scatter"])
else:
# default configuration
axis.scatter(x, y, s, c)
return fig, axis
def plot_bar(data, fig, axis, config):
"""plot_bar(data, fig, axis, config_dic)
Args
----
data: np.array, (L, M)
fig: matplotlib figure handle
axis: matplotlib axis handle
config: dictionary for axis.bar
L: number of vertial bar groups
M: number of sub-bar in each group
If yerr is to be used, please save it in config_dic['plot_bar']['yerr']
config_dic['plot_bar']['yerr'] should have the same shape as data.
Optional field in config_dic['plot_bar']:
1. ['x']: np.array or list, location of each bar group on x-axis
2. ['show_number']: str, format of showing numbers on top of each var
"{:{form}}".format(form=config_dic['plot_bar']['show_number'])
4. ['color_list']: list of str, list of colors for each column of data
5. ['edgecolor_list']: list of str, list of edge colors for each column of data
6. other options for matplotlib.pyplot.bar
"""
if type(data) is list:
data = np.asarray(data)
if data.ndim == 1:
data = np.expand_dims(data, 1)
width = 0.3
x_pos = np.arange(data.shape[0])
yerr = None
show_number = None
group_size = data.shape[1]
config_dic = copy.deepcopy(config)
color_list = [None for x in range(group_size)]
edgecolor_list = [None for x in range(group_size)]
if "plot_bar" in config_dic:
if "x" in config_dic["plot_bar"]:
x_pos = config_dic["plot_bar"]['x']
config_dic['plot_bar'].pop('x')
if "width" in config_dic['plot_bar']:
width = config_dic['plot_bar']['width']
config_dic['plot_bar'].pop('width')
if "color" in config_dic["plot_bar"]:
color_list = [config_dic["plot_bar"]['color'] for x in range(group_size)]
config_dic["plot_bar"].pop('color')
if "edgecolor" in config_dic['plot_bar']:
edgecolor_list = [config_dic["plot_bar"]['edgecolor'] for x in range(group_size)]
config_dic["plot_bar"].pop('edgecolor')
if "yerr" in config_dic['plot_bar']:
yerr = config_dic['plot_bar']['yerr']
config_dic['plot_bar'].pop('yerr')
if "show_number" in config_dic['plot_bar']:
show_number = config_dic['plot_bar']['show_number']
config_dic['plot_bar'].pop('show_number')
if "color_list" in config_dic['plot_bar']:
color_list = config_dic['plot_bar']['color_list']
config_dic['plot_bar'].pop('color_list')
if "edgecolor_list" in config_dic['plot_bar']:
edgecolor_list = config_dic['plot_bar']['edgecolor_list']
config_dic['plot_bar'].pop('edgecolor_list')
if "color_list" in config_dic:
color_list = config_dic['color_list']
config_dic.pop('color_list')
print("color_list should be in dic['plot_bar']")
for group_idx in range(group_size):
x_shift = group_idx * width - width * (group_size - 1) / 2
if yerr is None:
sub_yerr = None
else:
sub_yerr = yerr[:, group_idx]
if "plot_bar" in config_dic:
axis.bar(x_pos + x_shift, data[:, group_idx], width=width,
color = color_list[group_idx],
edgecolor = edgecolor_list[group_idx],
yerr = sub_yerr, **config_dic["plot_bar"])
else:
axis.bar(x_pos + x_shift, data[:, group_idx], width=width,
color = color_list[group_idx],
edgecolor = edgecolor_list[group_idx],
yerr = sub_yerr)
if show_number is not None:
for x, y in zip(x_pos, data[:, group_idx]):
axis.text(x + x_shift, y*1.01,
"{num:{form}}".format(num=y, form=show_number),
horizontalalignment="center")
return fig, axis
def plot_boxplot(data, fig, axis, config):
"""
"""
# get config for obxplot
bp_config = copy.deepcopy(config["plot_boxplot"]) \
if "plot_boxplot" in config else {}
if "plot_marker" in bp_config:
marker_config = bp_config["plot_marker"]
bp_config.pop("plot_marker")
else:
marker_config = {}
# filter data
data_for_plot = []
data_mean = []
if type(data) is list:
for col in range(len(data)):
idx = ~np.bitwise_or(np.isinf(data[col]), np.isnan(data[col]))
data_for_plot.append(data[col][idx])
data_mean.append(data[col][idx].mean())
else:
for col in range(data.shape[1]):
idx = ~np.bitwise_or(np.isinf(data[:, col]), np.isnan(data[:, col]))
data_for_plot.append(data[idx, col])
data_mean.append(data[idx, col].mean())
#
axis.boxplot(data_for_plot, **bp_config)
xpos = bp_config["positions"] if "positions" in bp_config \
else np.arange(len(data_for_plot))+1
if marker_config:
axis.plot(xpos, data_mean, **marker_config)
return fig, axis
def plot_err_bar(data, fig, axis, config_dic):
""" plot_err_bar
"""
if type(data) is list:
if len(data) == 3:
x = data[0]
y = data[1]
err = data[2]
elif len(data) == 2:
y = data[0]
err = data[1]
x = np.arange(y.shape[0])
else:
print("data must have 3 or 2 elements")
sys.exit(1)
else:
if data.shape[1] == 3:
x = data[:, 0]
y = data[:, 1]
err = data[:, 2]
elif data.shape[1] == 2:
x = np.arange(data.shape[0])
y = data[:, 0]
err = data[:, 1]
else:
print("data must have 3 or 2 columns")
sys.exit(1)
if "plot_err_bar" in config_dic:
axis.errorbar(x, y, yerr=err, **config_dic["plot_err_bar"])
else:
axis.errorbar(x, y, yerr=err)
return fig, axis
def plot_stacked_bar(data, fig, axis, config):
"""plot_bar(data, fig, axis, config_dic)
Args
----
data: np.array, (L, M)
fig: matplotlib figure handle
axis: matplotlib axis handle
config: dictionary for axis.bar
L: number of bars
M: number of values to be stacked in each group
If yerr is to be used, please save it in config_dic['plot_bar']['yerr']
config_dic['plot_bar']['yerr'] should have the same shape as data.
Optional field in config_dic['plot_bar']:
1. ['x']: np.array or list, location of each bar group on x-axis
2. ['show_number']: str, format of showing numbers on top of each var
"{:{form}}".format(form=config_dic['plot_bar']['show_number'])
4. ['color_list']: list of str, list of colors for each column of data
5. ['edgecolor_list']: list of str, list of edge colors for each column of data
6. other options for matplotlib.pyplot.bar
"""
if type(data) is list:
data = np.asarray(data)
if data.ndim == 1:
data = np.expand_dims(data, 1)
width = 0.3
x_pos = np.arange(data.shape[0])
yerr = None
show_number = None
group_size = data.shape[1]
config_dic = copy.deepcopy(config)
color_list = [None for x in range(group_size)]
edgecolor_list = [None for x in range(group_size)]
if "plot_bar" in config_dic:
if "x" in config_dic["plot_bar"]:
x_pos = config_dic["plot_bar"]['x']
config_dic['plot_bar'].pop('x')
if "width" in config_dic['plot_bar']:
width = config_dic['plot_bar']['width']
config_dic['plot_bar'].pop('width')
if "color" in config_dic["plot_bar"]:
color_list = [config_dic["plot_bar"]['color'] for x in range(group_size)]
config_dic["plot_bar"].pop('color')
if "edgecolor" in config_dic['plot_bar']:
edgecolor_list = [config_dic["plot_bar"]['edgecolor'] for x in range(group_size)]
config_dic["plot_bar"].pop('edgecolor')
if "yerr" in config_dic['plot_bar']:
yerr = config_dic['plot_bar']['yerr']
config_dic['plot_bar'].pop('yerr')
if "show_number" in config_dic['plot_bar']:
show_number = config_dic['plot_bar']['show_number']
config_dic['plot_bar'].pop('show_number')
if "color_list" in config_dic['plot_bar']:
color_list = config_dic['plot_bar']['color_list']
config_dic['plot_bar'].pop('color_list')
if "edgecolor_list" in config_dic['plot_bar']:
edgecolor_list = config_dic['plot_bar']['edgecolor_list']
config_dic['plot_bar'].pop('edgecolor_list')
if "color_list" in config_dic:
color_list = config_dic['color_list']
config_dic.pop('color_list')
print("color_list should be in dic['plot_bar']")
stacked_accum = np.zeros([data.shape[0]])
for group_idx in range(group_size):
if yerr is None:
sub_yerr = None
else:
sub_yerr = yerr[:, group_idx]
if "plot_bar" in config_dic:
axis.bar(x_pos, data[:, group_idx], width=width,
color = color_list[group_idx], bottom=stacked_accum,
edgecolor = edgecolor_list[group_idx],
yerr = sub_yerr, **config_dic["plot_bar"])
else:
axis.bar(x_pos, data[:, group_idx], width=width,
color = color_list[group_idx], bottom=stacked_accum,
edgecolor = edgecolor_list[group_idx],
yerr = sub_yerr)
stacked_accum += data[:, group_idx]
if show_number is not None:
for x, y in zip(x_pos, data[:, group_idx]):
axis.text(x, y*1.01,
"{num:{form}}".format(num=y, form=show_number),
horizontalalignment="center")
return fig, axis
def plot_table(data, fig, axis, config_dic):
"""plot_table(data, fig, axis, config_dic)
Use plot_imshow to show table and print numbers
"""
print_format = "1.2f"
font_color = "r"
if "plot_table" in config_dic:
tmp_config = copy.deepcopy(config_dic["plot_table"])
if "print_format" in tmp_config:
print_format = tmp_config['print_format']
tmp_config.pop('print_format')
if "font_color" in tmp_config:
font_color = tmp_config['font_color']
tmp_config.pop('font_color')
else:
tmp_config = {'cmap':'jet', 'origin':'lower', 'aspect':'auto'}
axis.imshow(data, **tmp_config)
config_dic['xlim'] = (-0.5, data.shape[1]-0.5)
config_dic['ylim'] = (-0.5, data.shape[0]-0.5)
for row_idx in range(data.shape[0]):
for col_idx in range(data.shape[1]):
axis.text(col_idx, row_idx,
"{num:{form}}".format(
num=data[row_idx,col_idx],form=print_format),
ha='center', va='center', c=font_color)
return fig, axis
def plot_barh(dataset, fig, axis, configs):
"""
"""
max_bin_width = np.max([np.max(x) for x in dataset])
if 'plot_barh' in configs:
if 'show_percentage' in configs['plot_barh']:
flag_show_per = configs['plot_barh']['show_percentage']
if 'ad_max_bin_width' in configs['plot_barh']:
max_bin_width *= configs['plot_barh']['ad_max_bin_width']
if 'color' in configs['plot_barh']:
color = configs['plot_barh']['color']
else:
color = 'Grey'
if flag_show_per:
dataset_tmp = dataset / np.sum(dataset, axis=1, keepdims=True) * 100
max_bin_width = 100 * 1.05
else:
dataset_tmp = dataset
# location of x
x_locs = [x * max_bin_width for x in np.arange(len(dataset))]
# temporary buffer
max_length = np.max([x.shape[0] for x in dataset])
# left most pos
left_most = x_locs[-1]
right_most = 0
for idx, (x_loc, data) in enumerate(zip(x_locs, dataset_tmp)):
# decide the left alignment position
lefts = x_loc - 0 * data
if 'plot_barh' in configs and 'align' in configs['plot_barh']:
if configs['plot_barh']['align'] == 'middle':
lefts = x_loc - 0.5 * data
elif configs['plot_barh']['align'] == 'right':
lefts = x_loc - data
# plot
axis.barh(np.arange(len(data)), data, height=1.0, left=lefts,
color=color)
# show text
if flag_show_per:
for idx, (data_value, left_pos) in enumerate(zip(data, lefts)):
axis.text(left_pos, idx,
'{:2.1f}'.format(data_value),
ha='right', va='center')
left_most = np.min([left_most, left_pos - max_bin_width * 0.4])
#right_most = np.max([right_most, left_pos + max(data)])
right_most = left_pos + max_bin_width * 1.05
if flag_show_per:
axis.set_xlim([left_most, right_most])
if 'xticklabels' in configs:
axis.set_xticks(x_locs)
axis.set_xticklabels(configs['xticklabels'])
if 'yticklabels' in configs:
axis.set_yticks(np.arange(max_length))
axis.set_yticklabels(configs['yticklabels'])
return fig, axis
############################
## Specific functions
## classification
############################
from scipy import special as scipy_special
def probit(x):
""" probit function to scale the axis
based on __probit__(p)
https://
projets-lium.univ-lemans.fr/sidekit/_modules/sidekit/bosaris/detplot.html
"""
return np.sqrt(2) * scipy_special.erfinv(2.0 * x - 1)
def plot_det(data, fig, axis, config_dic):
""" plot DET curves
fig, axis = plot_det(data, fig, axis, config_dic)
This function will do probit conversion
input
-----
data: [frr, far] computed by compute_det_curve
fig: fig handler
axis: axis handler
config_dic: configuration dictionary
output
------
fig: fig handler
axis: axis handler
"""
if type(data) is list:
# warping through probit
x = probit(data[1]) # far
y = probit(data[0]) # frr
# we will use plot_signal as back-end function for plotting DET curves
tmp_config_dic = config_dic.copy()
if "plot_det" in config_dic:
tmp_config_dic["plot_signal"] = config_dic["plot_det"]
# grid option
if "grid" in config_dic and config_dic["grid"]["b"] is False:
pass
else:
#axis.plot([probit(0.0001), probit(0.99)],
# [probit(0.0001), probit(0.99)],
# c='lightgrey', linestyle='--')
#axis.plot([probit(0.5), probit(0.5)],
# [probit(0.0001), probit(0.99)],
# c='lightgrey', linestyle='--')
#axis.plot([probit(0.0001), probit(0.99)],
# [probit(0.5), probit(0.5)],
# c='lightgrey', linestyle='--')
pass
# plot using the plot_signal function
plot_signal(np.stack([x, y], axis=1), fig, axis, tmp_config_dic)
# options on label
if "xlabel" not in config_dic:
axis.set_xlabel("False alarm probablity ({:s})".format("\%"))
if "ylabel" not in config_dic:
axis.set_ylabel("Miss probability ({:s})".format("\%"))
# ticks
if "xticks" not in config_dic:
xticks_to_use = [0.005, 0.05, 0.1, 0.2, 0.3, 0.5, 0.9]
else:
xticks_to_use = config_dic["xticks"]
config_dic.pop("xticks", None)
if "yticks" not in config_dic:
yticks_to_use = [0.005, 0.05, 0.1, 0.2, 0.3, 0.5, 0.9]
else:
yticks_to_use = config_dic["yticks"]
config_dic.pop("yticks", None)
xticks_to_use_probit = [probit(x) for x in xticks_to_use]
yticks_to_use_probit = [probit(x) for x in yticks_to_use]
axis.set_xticks(xticks_to_use_probit)
axis.set_xticklabels(["%d" % (x * 100) for x in xticks_to_use])
axis.set_yticks(yticks_to_use_probit)
axis.set_yticklabels(["%d" % (x * 100) for x in yticks_to_use])
if "xlim" in config_dic:
config_dic["xlim"] = [probit(x) for x in config_dic["xlim"]]
if "ylim" in config_dic:
config_dic["ylim"] = [probit(x) for x in config_dic["ylim"]]
if "grid" not in config_dic:
axis.grid(True)
# whether show EER on the figure
if "eer" in config_dic and config_dic['eer']:
abs_diffs = np.abs(data[1] - data[0])
min_index = np.argmin(abs_diffs)
eer = np.mean((data[1][min_index], data[0][min_index]))
axis.text(probit(eer), probit(eer),
"EER {:2.3}\%".format(eer * 100))
else:
print("plot_det requires input data = [far, frr]")
return fig, axis
############################
## Specific functions
## signal processing
############################
import scipy
import scipy.signal
import numpy as np
def _spec(data, fft_bins=4096, frame_shift=40, frame_length=240):
f, t, cfft = scipy.signal.stft(
data, nfft=fft_bins,
noverlap=frame_length-frame_shift, nperseg=frame_length)
return f,t,cfft
def _amplitude(cfft):
mag = np.power(np.power(np.real(cfft),2) + np.power(np.imag(cfft),2), 0.5)
return mag
def _amplitude_to_db(mag):
return 20*np.log10(mag+ np.finfo(np.float32).eps)
def _spec_amplitude(data, fft_bins=4096, frame_shift=40, frame_length=240):
_, _, cfft = _spec(data, fft_bins, frame_shift, frame_length)
mag = _amplitude(cfft)
return _amplitude_to_db(mag)
def plot_spec(data, fig, axis, config_dic):
"""
fig, axis = plot_spec(data, fig, axis, config_dic)
This function will plot spectrogram, given configuration in config_dic
input
-----
data: data
fig: fig handler
axis: axis handler
config_dic: configuration dictionary
output
------
fig: fig handler
axis: axis handler
"""
if type(data) is list:
print("plot_spectrogram only supports data array input, ")
print("but it receives list of data")
sys.exit(1)
# default configuration
tmp_dic = copy.deepcopy(config_dic["plot_spec"]) \
if "plot_spec" in config_dic else {}
if "sampling_rate" in tmp_dic:
sr = tmp_dic["sampling_rate"]
tmp_dic.pop("sampling_rate")
else:
sr = None
if "frame_shift" in tmp_dic:
fs = tmp_dic["frame_shift"]
tmp_dic.pop("frame_shift")
else:
fs = 80
if "frame_length" in tmp_dic:
fl = tmp_dic["frame_length"]
tmp_dic.pop("frame_length")
else:
fl = 320
if "fft_bins" in tmp_dic:
fn = tmp_dic["fft_bins"]
fn.pop("fft_bins")
else:
fn = 1024
# stft
spec = _spec_amplitude(data, fn, fs, fl)
tmp_config_dic = config_dic.copy()
if tmp_dic:
tmp_config_dic["plot_imshow"] = tmp_dic
plot_imshow(spec, fig, axis, tmp_config_dic)
# options on label
if "xlabel" not in config_dic:
axis.set_xlabel("Frame index")
if "ylabel" not in config_dic:
if sr is None:
axis.set_ylabel("Frequency bins")
else:
axis.set_ylabel("Frequency (Hz)")
# ticks
if "yticks" not in config_dic:
yticks_to_use = [(fn//2+1)//2, fn//2+1]
else:
yticks_to_use = config_dic["yticks"]
config_dic.pop("yticks", None)
axis.set_yticks(yticks_to_use)
if sr is not None:
freq_str = ["{:4.1f}".format(x * sr // 2 // 1000 / (fn//2+1)) \
for x in yticks_to_use]
axis.set_yticklabels(freq_str)
return fig, axis
if __name__ == "__main__":
print("Definition of plot_lib")
| 24,448 | 32.445964 | 93 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/tutorials/plot_tools/plot_API.py | #!/usr/bin/env python
"""
API for plotting
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import matplotlib
import matplotlib.pyplot as plt
from matplotlib.gridspec import GridSpec
from matplotlib.pyplot import cm
import shutil
# Set latex and font type
from matplotlib import rc
if shutil.which('tex') and shutil.which('latex'):
rc('font',**{'family':'sans-serif','sans-serif':['Helvetica']})
rc('text', usetex=True)
else:
print("Cannot find tex or latex. matplotlib.rc usetex is disabled")
# default fontsize
rc('font', size=9)
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############
## Utilities
############
_marker_bag = ["*","o","v","^","<",">","x","s","p","P",
"h","H","+",".","D","d","|","_"]
_line_style_bag = ['-','--', '-.', ':']
_color_bag = ['b', 'g', 'r', 'c', 'm', 'y', 'k']
def get_marker(idx):
return _marker_bag[idx % len(_marker_bag)]
def get_colors(total, colormap='jet'):
"""Function to get colors from a specified colormap
data = get_colors(total, colormap='jet')
input:
total: int, total number of colors to be sampled
colormap: str, name of the color map
ttps://matplotlib.org/3.1.1/gallery/color/colormap_reference.html
output:
data: np.array, [color_idx, color_RGB_dims]
Example usage:
# plot histogram using different colors
data1 = np.random.randn(1000)
color_num = 5
mapname = 'jet'
datas = [data1 + x*5 for x in range(color_num)]
config = [{'plot_hist': {'alpha':1.0, 'bins': 100, 'density': True,
'histtype': 'stepfilled',
'color': plot_API.get_color(x, color_num, mapname)}} \
for x in range(color_num)]
plot_API.plot_API(datas, plot_lib.plot_hist, 'single',
{'sub_plot': config, 'title': mapname.replace('_', '-')})
"""
color_bag = []
color_func = None
# pre-defined color bags
if colormap == 'self_1':
# black -> blue -> red -> ... old matplotlib style
color_bag = np.array([[78, 78, 78], [132,145,252],
[253,134,141], [110,143,82],
[229,173,69], [139,139,139]])/255.0
elif colormap == 'self_2':
# blue -> red
color_bag = np.array([[178, 180, 253], [253, 178, 179]])/255.0
elif colormap == 'self_3':
# grey scales
color_bag = np.array([[1.0, 1.0, 1.0], [0.95, 0.95, 0.95],
[0.90, 0.90, 0.90], [0.85, 0.85, 0.85],
[0.80, 0.80, 0.80]])
elif colormap == 'self_4':
# grey -> black -> blue -> red
color_bag = np.array([[174,174,174], [13,13,13], [11,36,251],
[252,13,27], [55,183,164], [189,27,189],
[26,120,148], [110,143,82]])/255.0;
elif colormap == 'self_5':
# self_1 withput black
color_bag = np.array([[132,145,252], [253,134,141],
[110,143,82], [229,173,69],
[139,139,139], [200,200,200]])/255.0;
elif colormap == 'self_6':
# grey
color_bag = np.array([[243,243,243],[202,202,202],
[160,160,160]])/255.0
elif colormap == 'self_7':
# office colors, black -> red -> blue -> green
color_bag = np.array([[0, 0, 0],[182,0,0],[60,103,188],
[101, 163, 62]])/255.0
elif colormap == 'self_8':
# black -> red -> blue -> green
color_bag = np.array([[0, 0, 0],[255,0,0],[0, 0, 255]])/255.0
else:
color_func = cm.get_cmap(colormap)
if color_func is not None:
indices = np.linspace(0,1,total)
return [color_func(idx) for idx in indices]
else:
return [color_bag[x % len(color_bag)] for x in range(total)]
def get_color(idx, total, colormap='jet'):
return get_colors(total, colormap)[idx]
def correct_label(label_str, replace_str='-'):
return label_str.replace('_', replace_str)
#################
## initialization
#################
# figure size
default_fig_size = (10, 5)
matplotlib.rcParams['figure.figsize'] = default_fig_size
###############
## function API
###############
def plot_API(data, plot_func, split_mode='single',
config_dic={}, verbose=False):
"""
fig, axis = plot_API(data, plot_func, split_mode='single', config_dic={})
Plot a figure on data, using plot_func with config_dic
input
-----
data: either a np.array or a list of np.array
plot_func: the function to plot a single figure. see plot_lib
split_mode:
'single': plot data in a single figure
'v': plot data in separate rows. data must be a list
'h': plot data in separate columns. data must be a list
'grid': plot in a grid. config_dic['ncols'] and config_dic['nrows']
are used to decide the number of rows and columns for gridspec
config_dic: configuration dictionary
verbose: whether print out the config_dic information for each figure
(default: False)
output
------
fig: fig handler
axis: axis handler of the last figure
"""
# figure size
if "figsize" not in config_dic or config_dic["figsize"] is None:
figsize = default_fig_size
else:
figsize = config_dic["figsize"]
fig = plt.figure(figsize=figsize)
#
if split_mode == 'single':
# plot in a single figure
axis = fig.add_subplot(111)
if type(data) is list:
for idx, data_entry in enumerate(data):
tmp_config = process_config_dic(config_dic, idx)
fig, axis = plot_func(data_entry, fig, axis, tmp_config)
util_options(fig, axis, tmp_config)
else:
fig, axis = plot_func(data, fig, axis, config_dic)
util_options(fig, axis, config_dic)
if verbose:
print(str(config_dic))
elif split_mode == 'grid' or split_mode == 'h' or split_mode == 'v':
# plot in a grid
# data must be a list
if type(data) is not list:
print("split_mode == h requires list of input data")
sys.exit(1)
# decide number of row and col
if split_mode == 'h':
# horizontol mode
nrows = 1
ncols = len(data)
elif split_mode == 'v':
# vertical mode
ncols = 1
nrows = len(data)
else:
# grid mode
nrows, ncols = get_nrow_ncol_config(len(data), config_dic)
# grid
wspace = config_dic["wspace"] if "wspace" in config_dic else None
hspace = config_dic["hspace"] if "hspace" in config_dic else None
gs = GridSpec(nrows, ncols, figure=fig, wspace=wspace, hspace=hspace)
# buffers
xlim_bag = [np.inf, -np.inf]
ylim_bag = [np.inf, -np.inf]
axis_bags = []
# plot
for idx, data_entry in enumerate(data):
if "col_first" in config_dic and config_dic["col_first"]:
col_idx = idx % ncols
row_idx = idx // ncols
else:
row_idx = idx % nrows
col_idx = idx // nrows
axis = fig.add_subplot(gs[row_idx, col_idx])
tmp_config = process_config_dic(config_dic, idx)
fig, axis = plot_func(data_entry, fig, axis, tmp_config)
util_options(fig, axis, tmp_config)
if verbose:
print(str(tmp_config))
axis_bags.append(axis)
xlim_bag = [min(xlim_bag[0], axis.get_xlim()[0]),
max(xlim_bag[1], axis.get_xlim()[1])]
ylim_bag = [min(ylim_bag[0], axis.get_ylim()[0]),
max(ylim_bag[1], axis.get_ylim()[1])]
if "sharey" in config_dic and config_dic["sharey"]:
if col_idx > 0:
axis.set_yticks([])
axis.set_ylabel("")
if "sharex" in config_dic and config_dic["sharex"]:
if row_idx < nrows -1:
axis.set_xticks([])
axis.set_xlabel("")
for axis in axis_bags:
if "sharex" in config_dic and config_dic["sharex"]:
axis.set_xlim(xlim_bag)
if "sharey" in config_dic and config_dic["sharey"]:
axis.set_ylim(ylim_bag)
else:
print("Unknown split_mode {:s}".format(split_mode))
axis = None
return fig, axis
def plot_API2(data_list, plot_funcs, grid_configs,
config_dic, verbose=False):
"""
fig, axis = plot_API2(data_list, plot_func_list, grid_config_list,
config_dic={})
On the same figure, plot data using different plot_funcs
This func will loop over
enumerante(zip(data_list, plot_func_list, grid_list))
input
-----
data_list: list of np.array or list of list of np.arrays
plot_func_list: list of functions to plot a single figure
grid_config_list: it can be
list of grid configuration:
[[[row_start, row_end], [col_start, col_end]],# for 1st subfig
[[row_start, row_end], [col_start, col_end]],# for 2nd subfig
...
]
or 'v': align figure vertically
or 'h': align figure horizontally
or 'grid': align figure in a grid
config_dic: configuration dictionary
verbose: whether print out the config_dic information for each figure
(default: False)
output
------
fig: fig handler
axis: axis handler of the last figure
"""
# figure size
if "figsize" not in config_dic or config_dic["figsize"] is None:
figsize = default_fig_size
else:
figsize = config_dic["figsize"]
fig = plt.figure(figsize=figsize)
# get the grid
if type(grid_configs) is str:
grid_configs = get_grid_config(len(data_list),
grid_configs, config_dic)
# analyze grid configuration
nrows = max([x[0][1] for x in grid_configs])
ncols = max([x[1][1] for x in grid_configs])
wspace = config_dic["wspace"] if "wspace" in config_dic else None
hspace = config_dic["hspace"] if "hspace" in config_dic else None
gs = GridSpec(nrows, ncols, figure=fig, wspace=wspace, hspace=hspace)
# buffers
xlim_bag = [np.inf, -np.inf]
ylim_bag = [np.inf, -np.inf]
axis_bags = []
axis_dic = {}
def sublabel(pos_a, pos_b, pos_c, pos_d):
return "{:d}{:d}{:d}{:d}".format(pos_a, pos_b, pos_c, pos_d)
# plot
for idx, (data_entry, plot_func, grid_config) in \
enumerate(zip(data_list, plot_funcs, grid_configs)):
axis_label = sublabel(grid_config[0][0], grid_config[0][1],
grid_config[1][0], grid_config[1][1])
if axis_label in axis_dic:
axis = axis_dic[axis_label]
else:
axis = fig.add_subplot(gs[grid_config[0][0]:grid_config[0][1],
grid_config[1][0]:grid_config[1][1]])
axis_dic[axis_label] = axis
tmp_config = process_config_dic(config_dic, idx)
if type(data_entry) is list:
for idx2, sub_data_entry in enumerate(data_entry):
sub_tmp_config = process_config_dic(tmp_config, idx2)
fig, axis = plot_func(sub_data_entry, fig, axis,
sub_tmp_config)
util_options(fig, axis, sub_tmp_config)
if verbose:
print(str(tmp_config))
else:
fig, axis = plot_func(data_entry, fig, axis, tmp_config)
util_options(fig, axis, tmp_config)
if verbose:
print(str(tmp_config))
axis_bags.append(axis)
xlim_bag = [min(xlim_bag[0], axis.get_xlim()[0]),
max(xlim_bag[1], axis.get_xlim()[1])]
ylim_bag = [min(ylim_bag[0], axis.get_ylim()[0]),
max(ylim_bag[1], axis.get_ylim()[1])]
if "sharey" in config_dic and config_dic["sharey"]:
if grid_config[1][0] > 0:
axis.set_yticks([])
axis.set_ylabel("")
if "sharex" in config_dic and config_dic["sharex"]:
if grid_config[0][1] < nrows:
axis.set_xticks([])
axis.set_xlabel("")
for axis in axis_bags:
if "sharex" in config_dic and config_dic["sharex"]:
axis.set_xlim(xlim_bag)
if "sharey" in config_dic and config_dic["sharey"]:
axis.set_ylim(ylim_bag)
return fig, axis_bags
##############
## Utilities
##############
def get_nrow_ncol_config(data_num, config_dic):
if "ncols" in config_dic and "nrows" not in config_dic:
ncols = config_dic["ncols"]
nrows = int(np.ceil(data_num * 1.0 / ncols))
elif "ncols" not in config_dic and "nrows" in config_dic:
nrows = config_dic["nrows"]
ncols = int(np.ceil(data_num * 1.0 / nrows))
elif "ncols" not in config_dic and "nrows" not in config_dic:
nrows = int(np.sqrt(data_num))
ncols = int(np.ceil(data_num * 1.0 / nrows))
else:
nrows = config_dic["nrows"]
ncols = config_dic["ncols"]
if nrows * ncols < data_num:
print("nrows * ncols < number of data in list")
sys.exit(1)
return nrows, ncols
def get_grid_config(data_num, option, config):
"""grid_spec = get_grid_config(data_num, option)
used by plot_API2 for creating the grid spec
"""
if option == 'h':
output = [[[0, 1], [x, x+1]] for x in range(data_num)]
elif option == 'v':
output = [[[x, x+1], [0, 1]] for x in range(data_num)]
else:
if option != 'grid':
print("Cannot create grid for option:" + str(option))
print("Use grid by default")
nrows, ncols = get_nrow_ncol_config(data_num, config)
output = []
for row_idx in range(nrows):
for col_idx in range(ncols):
output.append([[row_idx, row_idx+1],
[col_idx, col_idx+1]])
return output
def util_options(fig, axis, config_dic):
"""util_options(fig, axis, config_dic)
Wrappers to set fig and axis using options in config_dic
"""
if "xlabel" in config_dic:
axis.set_xlabel(config_dic["xlabel"])
if "ylabel" in config_dic:
axis.set_ylabel(config_dic["ylabel"])
if "title" in config_dic:
axis.set_title(config_dic["title"])
if "xlim" in config_dic:
axis.set_xlim(config_dic["xlim"])
if "ylim" in config_dic:
axis.set_ylim(config_dic["ylim"])
if "legend" in config_dic:
axis.legend(**config_dic["legend"])
if "xticks" in config_dic:
if type(config_dic["xticks"]) is dict:
axis.set_xticks(**config_dic["xticks"])
else:
axis.set_xticks(config_dic["xticks"])
if "xticklabels" in config_dic:
if type(config_dic["xticklabels"]) is dict:
axis.set_xticklabels(**config_dic["xticklabels"])
else:
axis.set_xticklabels(config_dic["xticklabels"])
if "yticks" in config_dic:
axis.set_yticks(config_dic["yticks"])
if "yticklabels" in config_dic:
if type(config_dic["yticklabels"]) is dict:
axis.set_yticklabels(**config_dic["yticklabels"])
else:
axis.set_yticklabels(config_dic["yticklabels"])
if "grid" in config_dic:
axis.grid(**config_dic["grid"])
return fig, axis
def process_config_dic(input_dic, idx):
""" input_dic may contain global configuration and sub configuration
for each sub-figures.
>> config = {"xlabel": "time",
"ylabel": "amplitude",
"sharey": True,
"sharex": True,
"plot": {"alpha": 0.3},
"sub1": [{"legend": {"labels": ["s1", "s2"], "loc":2}},
{"legend": {"labels": ["s3", "s4"], "loc":2}}]}
>> plot_API.process_config_dic(config, 0)
{'xlabel': 'time',
'ylabel': 'amplitude',
'sharey': True,
'sharex': True,
'plot': {'alpha': 0.3},
'legend': {'labels': ['s1', 's2'], 'loc': 2}}
"""
global_dic = {x:y for x, y in input_dic.items() if not x.startswith("sub")}
for x, y in input_dic.items():
if x.startswith("sub"):
if type(y) is not list:
print("{:s} is not list".format(str(y)))
sys.exit(1)
if idx < len(y):
sub_dic = y[idx]
global_dic.update(sub_dic)
return global_dic
if __name__ == "__main__":
print("Definition of plot_API")
| 17,325 | 34.00202 | 79 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/tutorials/plot_tools/table_API.py | #!/usr/bin/env python
"""
Library of utilities for printing latex table
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import matplotlib
import matplotlib.pyplot as plt
from matplotlib.gridspec import GridSpec
from matplotlib.pyplot import cm
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#####################
## Latex table
#####################
def return_one_row_latex(content_buffer):
return " & ".join(content_buffer) + r"\\ " + "\n"
def return_one_row_text(content_buffer):
return " ".join(content_buffer) + "\n"
def fill_cell(text, length, sep=''):
return "{str:^{wid}}".format(str=text, wid=length) + sep
def wrap_value(data, wrap_factor=0):
if wrap_factor == 0:
return data
else:
ratio = (1+wrap_factor) / (1-wrap_factor)
return np.power((1 - np.power(1 - data, ratio)), 1/ratio)
def return_latex_color_cell(value, val_min, val_max, scale, wrap, color_func):
# clip the value for color rendering
value = np.clip(value, val_min, val_max)
# normalized value
if scale < 0:
value = wrap_value((value - val_min) / (val_max - val_min), wrap)*-scale
value = -scale - value
else:
value = wrap_value((value - val_min) / (val_max - val_min), wrap)*scale
# only use RGB, not RGBA
color_code = color_func(value)[:-1]
color_code = ', '.join(["{:0.2f}".format(x) for x in color_code])
return r"\cellcolor[rgb]{" + color_code + "}"
def is_valid_float(val):
try:
float(val)
except ValueError:
return False
else:
if val != np.inf and val == val:
return True
else:
return False
def return_valid_number_idx(data_array):
"""return the index of data ceil that has valid nummerical value
"""
is_numeric_3 = np.vectorize(is_valid_float, otypes = [bool])
return is_numeric_3(data_array)
def print_table(data_array, column_tag, row_tag,
print_format = "1.2f",
with_color_cell = True,
colormap='Greys',
colorscale = 0.5,
colorwrap = 0,
col_sep = '',
print_latex_table=True,
print_text_table=True,
print_format_along_row=True,
color_minmax_in = 'global',
pad_data_column = 0,
pad_dummy_col = 0,
func_after_row = None,
data_display_array = None):
"""
print a latex table given the data (np.array) and tags
step1. table will be normalized so that values will be (0, 1.0)
step2. each normalzied_table[i,j] will be assigned a RGB color tuple
based on color_func( normalzied_table[i,j] * color_scale)
input
-----
data_array: np.array [M, N]
column_tag: list of str, length N, tag in the first row
row_tag: list of str, length M, tags in first col of each row
print_format: str or list of str, specify the format to print number
default "1.2f"
print_format_along_row: bool, when print_format is a list, is this
list specified for rows? Default True
If True, row[n] will use print_format[n]
If False, col[n] will use print_format[n]
with_color_cell: bool, default True,
whether to use color in each latex cell
colormap: str, color map name (matplotlib)
colorscale: float, default 0.5,
normalized table value will be scaled
color = color_func(nomrlized_table[i,j] * colorscale)
list of float
depends on configuration of color_minmax_in
if color_minmax_in = 'row', colorscale[i] for the i-th row
if color_minmax_in = 'col', colorscale[j] for the j-th row
np.array
color_minmax_in cannot be 'row' or 'col'.
colorscale[i, j] is used for normalized_table[i, j]
colorwrap: float, default 0, wrap the color-value mapping curve
colorwrap > 0 works like mels-scale curve
col_sep: str, additional string to separate columns.
You may use '\t' or ',' for CSV
print_latex_table: bool, print the table as latex command (default True)
print_text_table: bool, print the table as text format (default True)
color_minmax_in: how to decide the max and min to compute cell color?
'global': get the max and min values from the input matrix
'row': get the max and min values from the current row
'col': get the max and min values from the current column
(min, max): given the min and max values
default is global
pad_data_column: int, pad columns on the left or right of data matrix
(the tag column will still be on the left)
0: no padding (default)
-N: pad N dummy data columns to the left
N: pad N dummy data columns to the right
pad_dummy_col: int, pad columns to the left or right of the table
(the column will be padded to the left of head column)
0: no padding (default)
N: pad N columns to the left
output
------
latext_table, text_table
Tables will be printed to the screen.
The latex table will be surrounded by begin{tabular}...end{tabular}
It can be directly pasted to latex file.
However, it requires usepackage{colortbl} to show color in table cell.
"""
# default column and row are empty string
if column_tag is None:
column_tag = ["" for data in data_array[0, :]]
if row_tag is None:
row_tag = ["" for data in data_array]
#
if data_display_array is None:
data_display_array = data_array + np.nan
flag_data_display = False
else:
flag_data_display = True
# if padding of the data array is necessary
if pad_data_column < 0:
column_tag = ["" for x in range(-pad_data_column)] + column_tag
dummy_col = np.zeros([data_array.shape[0], -pad_data_column]) + np.nan
data_array = np.concatenate([dummy_col, data_array], axis=1)
data_display_array = np.concatenate([dummy_col, data_display_array], axis=1)
elif pad_data_column > 0:
column_tag = ["" for x in range(pad_data_column)] + column_tag
dummy_col = np.zeros([data_array.shape[0], pad_data_column]) + np.nan
data_array = np.concatenate([data_array, dummy_col], axis=1)
data_display_array = np.concatenate([data_display_array, dummy_col], axis=1)
else:
pass
# check print_format
if type(print_format) is not list:
if print_format_along_row:
# repeat the tag
print_format = [print_format for x in row_tag]
else:
print_format = [print_format for x in column_tag]
else:
if print_format_along_row:
assert len(print_format) == len(row_tag)
else:
assert len(print_format) == len(column_tag)
# color configuration
color_func = cm.get_cmap(colormap)
#data_idx = return_valid_number_idx(data_array)
#value_min = np.min(data_array[data_idx])
#value_max = np.max(data_array[data_idx])
def get_latex_color(data_array, row_idx, col_idx, color_minmax_in):
x = data_array[row_idx, col_idx]
if color_minmax_in == 'row':
data_idx = return_valid_number_idx(data_array[row_idx])
value_min = np.min(data_array[row_idx][data_idx])
value_max = np.max(data_array[row_idx][data_idx])
if type(colorscale) is list:
colorscale_tmp = colorscale[row_idx]
elif color_minmax_in == 'col':
data_idx = return_valid_number_idx(data_array[:, col_idx])
value_min = np.min(data_array[:, col_idx][data_idx])
value_max = np.max(data_array[:, col_idx][data_idx])
if type(colorscale) is list:
colorscale_tmp = colorscale[col_idx]
elif type(color_minmax_in) is tuple or type(color_minmax_in) is list:
value_min = color_minmax_in[0]
value_max = color_minmax_in[1]
if type(colorscale) is np.ndarray:
colorscale_tmp = colorscale[row_idx, col_idx]
else:
data_idx = return_valid_number_idx(data_array)
value_min = np.min(data_array[data_idx])
value_max = np.max(data_array[data_idx])
if type(colorscale) is np.ndarray:
colorscale_tmp = colorscale[row_idx, col_idx]
if type(colorscale) is not list:
colorscale_tmp = colorscale
# return a color command for latex cell
return return_latex_color_cell(x, value_min, value_max,
colorscale_tmp, colorwrap, color_func)
# maximum width for tags in 1st column
row_tag_max_len = max([len(x) for x in row_tag])
# maximum width for data and tags for other columns
if print_format_along_row:
tmp_len = []
for idx, data_row in enumerate(data_array):
if len(print_format[0]):
if flag_data_display:
max_len = max([len(x) for x in data_display_array[idx]])
else:
max_len = max([len("{num:{form}}".format(num=x,
form=print_format[idx])) \
for x in data_row])
tmp_len.append(max_len)
else:
tmp_len.append(0)
else:
tmp_len = []
for idx, data_col in enumerate(data_array.T):
if len(print_format[0]):
if flag_data_display:
max_len = max([len(x) for x in data_display_array[:, idx]])
else:
max_len = max([len("{num:{form}}".format(num=x,
form=print_format[idx])) \
for x in data_col])
tmp_len.append(max_len)
else:
tmp_len.append(0)
col_tag_max_len = max([len(x) for x in column_tag] + tmp_len)
# prepare buffer
text_buffer = ""
latex_buffer = ""
text_cell_buffer = []
latex_cell_buffer = []
# latex head
if pad_dummy_col > 0:
latex_buffer += r"\begin{tabular}{" \
+ ''.join(['c' for x in column_tag + ['']])
latex_buffer += ''.join(['c' for x in range(pad_dummy_col)]) + r"}"+"\n"
else:
latex_buffer += r"\begin{tabular}{" \
+ ''.join(['c' for x in column_tag + ['']]) + r"}"+"\n"
latex_buffer += r"\toprule" + "\n"
# head row
# for latex
hrow = [fill_cell("", row_tag_max_len)] \
+ [fill_cell(x, col_tag_max_len) for x in column_tag]
if pad_dummy_col > 0:
hrow = [fill_cell("", 1) for x in range(pad_dummy_col)] + hrow
latex_buffer += return_one_row_latex(hrow)
latex_buffer += r"\midrule" + "\n"
latex_cell_buffer.append(hrow)
# for plain text (add additional separator for each column)
hrow = [fill_cell("", row_tag_max_len, col_sep)] \
+ [fill_cell(x, col_tag_max_len, col_sep) for x in column_tag]
text_buffer += return_one_row_text(hrow)
text_cell_buffer.append(hrow)
# contents
row = data_array.shape[0]
col = data_array.shape[1]
for row_idx in np.arange(row):
# row head
row_content_latex = [fill_cell(row_tag[row_idx], row_tag_max_len)]
row_content_text = [fill_cell(row_tag[row_idx],row_tag_max_len,col_sep)]
if pad_dummy_col > 0:
row_content_latex = [fill_cell("", 1) for x in range(pad_dummy_col)] \
+ row_content_latex
# each column in the raw
for col_idx in np.arange(col):
if print_format_along_row:
tmp_print_format = print_format[row_idx]
else:
tmp_print_format = print_format[col_idx]
if is_valid_float(data_array[row_idx,col_idx]):
if len(tmp_print_format):
num_str = "{num:{form}}".format(
num=data_array[row_idx,col_idx],
form=tmp_print_format)
else:
num_str = ""
latex_color_cell = get_latex_color(
data_array, row_idx, col_idx,
color_minmax_in)
elif type(data_array[row_idx,col_idx]) is str:
if len(tmp_print_format):
num_str = "{num:{form}}".format(
num=data_array[row_idx,col_idx],
form=tmp_print_format)
else:
num_str = ""
latex_color_cell = ''
else:
num_str = ''
latex_color_cell = ''
if not with_color_cell:
latex_color_cell = ''
if flag_data_display:
num_str = data_display_array[row_idx, col_idx]
row_content_text.append(
fill_cell(num_str, col_tag_max_len, col_sep))
row_content_latex.append(
fill_cell(latex_color_cell + ' ' + num_str, col_tag_max_len))
# latex table content
latex_buffer += return_one_row_latex(row_content_latex)
latex_cell_buffer.append(row_content_latex)
# text content
text_buffer += return_one_row_text(row_content_text)
text_cell_buffer.append(row_content_text)
if func_after_row is not None:
latex_buffer += func_after_row(row_idx)
latex_buffer += r"\bottomrule" + "\n"
latex_buffer += r"\end{tabular}" + "\n"
if print_latex_table:
print(latex_buffer)
if print_text_table:
print(text_buffer)
return latex_buffer, text_buffer, latex_cell_buffer, text_cell_buffer
def concatenate_table(table_list, ignore_initial=True,
add_separator=1, latex=True):
"""
"""
rows = [len(x) for x in table_list]
if len(list(set(rows))) > 1:
print("Input tables have different row numbers")
return None
output_text = ""
output_table = []
for row in range(len(table_list[0])):
temp = []
for idx, subtable in enumerate(table_list):
if ignore_initial:
temp += subtable[row][1:]
else:
temp += subtable[row]
if add_separator and idx < len(table_list)-1:
temp += ['' for x in range(add_separator)]
output_table.append(temp)
output_text += return_one_row_latex(temp)
# latex head
latex_buffer = r"\begin{tabular}{" \
+ ''.join(['c' for x in temp + ['']]) + r"}" + "\n"
latex_buffer += output_text
latex_buffer += r"\end{tabular}" + "\n"
return latex_buffer, output_table
if __name__ == "__main__":
print("Tools for printing table for latex")
# example
data = np.random.randn(5, 3)
col_tags = ['A', 'B', 'C']
row_tags = ['1', '2', '3', '4', '5']
_ = print_table(data, col_tags, row_tags)
# Latex code of the colored table will be printed
| 15,821 | 35.96729 | 87 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/tutorials/plot_tools/plot_misc.py | #!/usr/bin/env python
"""
Misc tools used before the start of plot_API and plot_lib
"""
from __future__ import absolute_import
from __future__ import print_function
from mpl_toolkits.mplot3d import Axes3D
import matplotlib
import matplotlib.pyplot as plt
matplotlib.rcParams['figure.figsize'] = (10, 5)
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def plot_signal(data, fig, axis, xlabel='', ylabel='', title=''):
"""
"""
axis.plot(data)
axis.set_xlabel(xlabel)
axis.set_ylabel(ylabel)
axis.set_title(title)
axis.set_xlim(0, data.shape[0]-1)
return fig, axis
def plot_signal_stem(data, fig, axis, xlabel, ylabel):
"""
"""
axis.stem(data, use_line_collection=True)
axis.set_xlabel(xlabel)
axis.set_ylabel(ylabel)
axis.set_xlim(0, data.shape[0]-1)
return fig, axis
def plot_imshow(data, fig, axis, xlabel, ylabel):
axis.imshow(data, aspect='auto', origin='lower', cmap='RdBu')
axis.set_xlabel(xlabel)
axis.set_ylabel(ylabel)
return fig, axis
def plot_spectrogram(data, fig, axis, xlabel, ylabel, sampling_rate=None, title=''):
axis.imshow(data, aspect='auto', origin='lower', cmap='jet')
axis.set_xlabel(xlabel)
axis.set_ylabel(ylabel)
if sampling_rate:
yticks = [data.shape[0]//2, data.shape[0]//4*3, data.shape[0]]
ytickslabels = ["%1.1f" % (x / data.shape[0] * sampling_rate / 2.0 / 1000.0) for x in yticks]
axis.set_yticks(yticks)
axis.set_yticklabels(ytickslabels)
if title:
axis.set_title(title)
return fig, axis
def plot_surface(data, fig, ax, xlabel='', ylabel='', zlabel='', angleX=30, angleY=30):
X = np.arange(data.shape[0])
Y = np.arange(data.shape[1])
X, Y = np.meshgrid(X, Y)
Z = data
surf = ax.plot_surface(X, Y, Z.T, cmap='RdBu')
ax.set_xlabel(xlabel)
ax.set_ylabel(ylabel)
ax.set_zlabel(zlabel)
ax.view_init(angleX, angleY)
return fig, ax
import numpy as np
import matplotlib
def plot_matrix(data, fig, axis_left=0.0,
axis_bottom=0.0, axis_length=1.0, colormap='Greys',
color_norm = None,
colorgrad_y=True, colorgrad_x=True, alpha=1.0,
int_width=1, deci_width=1):
axis = fig.add_axes([axis_left, axis_bottom, axis_length, axis_length])
# get colormap for each data point
cmap = matplotlib.cm.get_cmap(colormap)
color_num = colorgrad_x * data.shape[1] + colorgrad_y * (data.shape[0]) * (data.shape[1] ** colorgrad_x)
color_num = int(color_num)
n_more_color = int(np.sqrt(data.shape[0] * data.shape[1]))
color_val = np.linspace(0, 1.0, color_num+n_more_color)
#
bias = 0
x_range = np.arange(data.shape[1]+1)
y_range = np.arange(data.shape[0]+1)
x_max = x_range[-1]
y_max = y_range[-1]
for x_idx in x_range:
axis.plot([x_idx + bias, x_idx + bias], [bias, y_max+bias], 'k')
for y_idx in y_range:
axis.plot([bias, x_max + bias], [y_idx + bias, y_idx+bias], 'k')
for x_idx in x_range[:-1]:
for y_idx in y_range[:-1]:
# plot from top to down
y_idx_reverse = y_max - 1 - y_idx
axis.text(x_idx+0.5+bias, y_idx_reverse+0.5+bias,
"%*.*f" % (int_width, deci_width, data[y_idx, x_idx]),
color='k', horizontalalignment='center',
verticalalignment='center')
x_tmp = np.array([x_idx, x_idx+1, x_idx+1, x_idx]) + bias
y_tmp = np.array([y_idx_reverse, y_idx_reverse,
y_idx_reverse+1, y_idx_reverse+1])+bias
# get the color-idx
color_idx = x_idx * colorgrad_x \
+ y_idx * colorgrad_y * (data.shape[1] ** colorgrad_x)
if color_norm is not None:
cell_color = cmap(color_norm(data[y_idx, x_idx]))
else:
cell_color = cmap(color_val[color_idx])
axis.fill(x_tmp, y_tmp, color=cell_color, alpha=alpha)
axis.axis('off')
return axis
def plot_tensor(data_tensor, shift=0.1, colormap='Greys',
color_on_value=False,
colorgrad_y=True, colorgrad_x=True, alpha=1.0,
title='', int_width=1, deci_width=1):
"""plot_tensor(data, shift=0.1, colormap='Greys',
color_on_value=False,
colorgrad_y=True, colorgrad_x=True, alpha=1.0,
title='', int_width=1, deci_width=1):
Plot 3D tensor data.
data: tensor of shape (batchsize, length, dim)
shift=0.1: space between different data in the batch
colormap='Greys': color map for coloring
color_on_value=False: select color of the cell based on data value
colorgrad_y=True: when color_on_value=False, select color based on row index
colorgrad_x=True: when color_on_value=False, select color based on column index
alpha=1.0: alpha for matplotlib.plot
title='': title of this data
int_width=1: x in %x.yf, when displaying numbers in figure
deci_width=1: y in %x.yf, when displaying numbers in figure
"""
try:
data = data_tensor.numpy()
except AttributeError:
data = data_tensor
if data.ndim != 3:
print("input data is not a 3d tensor ")
return None,None
fig_width = data.shape[2]/2 + (data.shape[0]-1)*shift
fig_height = data.shape[1]/2 + (data.shape[0]-1)*shift
fig = plt.figure(figsize=(fig_width, fig_height))
axis_start = 0.0
axis_end = 1.0
axis_length = axis_end - shift * (data.shape[0] - 1) - axis_start
if color_on_value:
color_norm = lambda x: (x - data.min())/(data.max() - data.min()+0.0001)*0.6
else:
color_norm = None
axis = []
for idx in np.arange(data.shape[0]):
axis.append(
plot_matrix(
data[idx], fig,
axis_start + shift * idx,
axis_start + shift * (data.shape[0]-1-idx),
axis_length,
colormap, color_norm,
colorgrad_y, colorgrad_x, alpha, int_width, deci_width))
if len(title):
fig.text(0.5, 0.99, title, ha='center')
return fig, axis
if __name__ == "__main__":
print("Misc tools from ../plot_lib.py")
| 6,363 | 34.553073 | 108 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/util_loss_metric.py | #!/usr/bin/env python
"""
util_loss_metric
Loss functions or metrics
References
[1] Weitang Liu, Xiaoyun Wang, John Owens, and Yixuan Li.
Energy-Based Out-of-Distribution Detection.
In Proc. NIPS, 33:21464–21475. 2020.
[2] Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna,
Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan
Supervised Contrastive Learning. Proc.NIPS. 2020.
[3] Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz.
Mixup: Beyond Empirical Risk Minimization. In Proc. ICLR. 2018.
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
#####################
# negative energy [1]
#####################
def neg_energy(logits, temperature=1):
""" neg_eng = neg_energy(logits, temperature=1)
neg_eng[x] = -T \log \sum_y \exp (logits[x, y] / T)
See [1]
input
-----
logits: tensor, shape (batch, dim)
temperature: float, temperature hyperparameter
output
------
neg_eng: tensor, shape (batch,)
"""
eng = - temperature * torch.logsumexp(logits / temperature, dim=1)
return eng
def neg_energy_reg_loss(energy, margin_in, margin_out, flag_in):
""" loss = neg_energy_reg_loss(energy, margin_in, margin_out, flag_in)
See [1] eqs.(8-9)
input
-----
energy: tensor, any shape is OK
margin_in: float, margin for the in-dist. data
margin_out: float, margin for the out-dist. data
flag_in: bool, if the input data is in-dist. data
output
------
loss: scalar
"""
if flag_in:
loss = torch.pow(torch_nn_func.relu(energy - margin_in), 2).mean()
else:
loss = torch.pow(torch_nn_func.relu(margin_out - energy), 2).mean()
return loss
#####################
# supervised contrastive loss [2]
#####################
def supcon_loss(input_feat,
labels = None, mask = None, sim_metric = None,
t=0.07, contra_mode='all', length_norm=False):
"""
loss = SupConLoss(feat,
labels = None, mask = None, sim_metric = None,
t=0.07, contra_mode='all')
input
-----
feat: tensor, feature vectors z [bsz, n_views, ...].
labels: ground truth of shape [bsz].
mask: contrastive mask of shape [bsz, bsz], mask_{i,j}=1 if sample j
has the same class as sample i. Can be asymmetric.
sim_metric: func, function to measure the similarity between two
feature vectors
t: float, temperature
contra_mode: str, default 'all'
'all': use all data in class i as anchors
'one': use 1st data in class i as anchors
length_norm: bool, default False
if True, l2 normalize feat along the last dimension
output
------
A loss scalar.
Based on https://github.com/HobbitLong/SupContrast/blob/master/losses.py
Supervised Contrastive Learning: https://arxiv.org/pdf/2004.11362.pdf.
Example:
feature = torch.rand([16, 2, 1000], dtype=torch.float32)
feature = torch_nn_func.normalize(feature, dim=-1)
label = torch.tensor([1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 5, 1, 1, 1, 1, 1],
dtype=torch.long)
loss = supcon_loss(feature, labels=label)
"""
if length_norm:
feat = torch_nn_func.normalize(input_feat, dim=-1)
else:
feat = input_feat
# batch size
bs = feat.shape[0]
# device
dc = feat.device
# dtype
dt = feat.dtype
# number of view
nv = feat.shape[1]
# get the mask
# mask[i][:] indicates the data that has the same class label as data i
if labels is not None and mask is not None:
raise ValueError('Cannot define both `labels` and `mask`')
elif labels is None and mask is None:
mask = torch.eye(bs, dtype=dt, device=dc)
elif labels is not None:
labels = labels.view(-1, 1)
if labels.shape[0] != bs:
raise ValueError('Num of labels does not match num of features')
mask = torch.eq(labels, labels.T).type(dt).to(dc)
else:
mask = mask.type(dt).to(dc)
# prepare feature matrix
# -> (num_view * batch, feature_dim, ...)
contrast_feature = torch.cat(torch.unbind(feat, dim=1), dim=0)
#
if contra_mode == 'one':
# (batch, feat_dim, ...)
anchor_feature = feat[:, 0]
anchor_count = 1
elif contra_mode == 'all':
anchor_feature = contrast_feature
anchor_count = nv
else:
raise ValueError('Unknown mode: {}'.format(self.contrast_mode))
# compute logits
# logits_mat is a matrix of size [num_view * batch, num_view * batch]
# or [batch, num_view * batch]
if sim_metric is not None:
logits_mat = torch.div(
sim_metric(anchor_feature, contrast_feature), t)
else:
logits_mat = torch.div(
torch.matmul(anchor_feature, contrast_feature.T), t)
# mask based on the label
# -> same shape as logits_mat
mask_ = mask.repeat(anchor_count, nv)
# mask on each data itself (
self_mask = torch.scatter(
torch.ones_like(mask_), 1,
torch.arange(bs * anchor_count).view(-1, 1).to(dc),
0)
#
mask_ = mask_ * self_mask
# for numerical stability, remove the max from logits
# see https://en.wikipedia.org/wiki/LogSumExp trick
# for numerical stability
logits_max, _ = torch.max(logits_mat * self_mask, dim=1, keepdim=True)
logits_mat_ = logits_mat - logits_max.detach()
# compute log_prob
exp_logits = torch.exp(logits_mat_ * self_mask) * self_mask
log_prob = logits_mat_ - torch.log(exp_logits.sum(1, keepdim=True))
# compute mean of log-likelihood over positive
mean_log_prob_pos = (mask_ * log_prob).sum(1) / mask_.sum(1)
# loss
loss = - mean_log_prob_pos
loss = loss.view(anchor_count, bs).mean()
return loss
################
# Mixup
################
class MixUpCE(torch_nn.Module):
def __init__(self, weight = None):
super(MixUpCE, self).__init__()
self.m_loss1 = torch_nn.CrossEntropyLoss(weight=weight,reduction='none')
self.m_loss2 = torch_nn.CrossEntropyLoss(weight=weight,reduction='none')
return
def forward(self, logits, y1, y2=None, gammas=None):
""" loss = MixUpCE.forward(logits, y1, y2, gammas)
This API computes the mixup cross-entropy.
Logits is assumed to be f( gammas * x1 + (1-gammas) * x2).
Thus, this API only compute the CE:
gammas * Loss(logits, y1) + (1 - gammas) * Loss(logits, y2)
Note that if y2 and gammas are None, it uses common CE
input
-----
logits: tensor, (batch, dim)
y1: tensor, (batch, )
y2: tensor, (batch, )
gammas: tensor, (batch, )
output
------
loss: scalar
"""
if y2 is None and gammas is None:
loss_val = self.m_loss1(logits, y1)
else:
loss_val = gammas * self.m_loss1(logits, y1)
loss_val += (1-gammas) * self.m_loss2(logits, y2)
return loss_val.mean()
#####################
# Distillation related
#####################
def kld_distill(logits_s, logits_t, temp=20):
""" KLD-based distillation loss
input
-----
logits_s: tensor, (batch, ..., dim), student output logits
where dim is #. output categories
logits_t: tensor, (batch, ..., dim), teacher output logits
temp: float, temperature, default=20
output
------
loss: scalar
"""
KD_loss = torch_nn_func.kl_div(
torch_nn_func.log_softmax(logits_s / temp, dim = -1),
torch_nn_func.log_softmax(logits_t / temp, dim = -1),
reduction = 'batchmean',
log_target = True) * temp * temp
return KD_loss
#####################
# Rank consistency
#####################
def rank_consistency(x, metric = None, anchor = None, diff_mat = None):
"""loss = rank_consistency(x, metric)
input
-----
x: tensor, (batch, D1, D2 ...)
metric: a function or None
This function must be f(x1, x2) -> scalar
where x1 and x2 are in shape (D1, D2 ...)
if None, negative cosine similarity for
x1 and x2 of shape (D1, )
anchor: tensor, (batch, D1, D2, ...), as anchor
or None
If None, one of difference vector in the
matrix will be selected as anchor
diff_mat: tensor, (batch, batch, D1, D2 ...)
of None
if diff_mat is provided, x will be ignored
output
------
loss: scalar, loss value
Example
-------
>> x = torch.randn(4, 10)
>> x[1] = x[0] + 1.0
>> x[2] = x[0] + 2.0
>> x[3] = x[0] + 3.0
>> rank_consistency(x)
tensor(-1.)
"""
if diff_mat is None:
# (batch, batch, dim)
# diff_mat[i, j] = x[j] - x[i]
diff_mat = x - x.unsqueeze(1)
# batch size
bs = diff_mat.shape[0]
# loss to be accumulated
loss = 0.0
# metric
if metric is None:
# default use negative cosine_similarity
metric = lambda x1, x2: -torch_nn_func.cosine_similarity(x1, x2, dim=0)
#
if bs < 3:
return loss
# get anchor
if anchor is None:
# choose the diff_mat[1, 0] as the anchor
anchor_row_idx = 1
anchor_col_idx = 0
anchor = diff_mat[anchor_row_idx, anchor_col_idx]
else:
# anchor is provided externally
anchor_row_idx = -1
anchor_col_idx = -1
# loop over the matrix, compare the off-diagnoal elements
# with the anchor
count = 0.0
for col_idx in np.arange(bs-1):
for row_idx in np.arange(col_idx+1, bs):
if col_idx == anchor_col_idx and anchor_row_idx == row_idx:
continue
loss += metric(anchor, diff_mat[row_idx, col_idx])
count += 1
loss = loss / count
return loss
def rank_consistency_v2(x, metric = None, diff_mat = None):
"""loss = rank_consistency_v2(x, metric)
input
-----
x: tensor, (batch, D1, D2 ...)
metric: a function or None
This function must be f(x1, x2) -> scalar
where x1 and x2 are in shape (D1, D2 ...)
if None, negative cosine similarity for
x1 and x2 of shape (D1, )
diff_mat: tensor, (batch, batch, D1, D2 ...)
of None
if diff_mat is provided, x will be ignored
output
------
loss: scalar, loss value
Example
-------
>> x = torch.randn(4, 10)
>> x[1] = x[0] + 1.0
>> x[2] = x[0] + 2.0
>> x[3] = x[0] + 3.0
>> metric = lambda x1, x2: \
torch_nn_func.margin_ranking_loss(x1, x2, torch.ones_like(x1), 0.1)
>> rank_consistencyV2(x, metric)
tensor(.0)
"""
if diff_mat is None:
# (batch, batch, dim)
# diff_mat[i, j] = x[j] - x[i]
diff_mat = x - x.unsqueeze(1)
# batch size
bs = diff_mat.shape[0]
# loss to be accumulated
loss = 0.0
# metric
if metric is None:
# default use margin_ranking_loss
metric = lambda x1, x2: torch_nn_func.margin_ranking_loss(
x1, x2, torch.ones_like(x1), 0.1)
#
if bs < 3:
return loss
count = 0.0
# loop over the matrix, column first
for col_idx in np.arange(bs-2):
for row_idx in np.arange(col_idx+2, bs):
# assume diff[i, j] should be diff[i-1, j]
loss += metric(diff_mat[row_idx-1, col_idx],
diff_mat[row_idx, col_idx])
count += 1
# loop over the matrix, column first
for row_idx in np.arange(2, bs):
for col_idx in np.arange(1, row_idx):
# assume diff[i, j] should be diff[i, j-1]
loss += metric(diff_mat[row_idx, col_idx],
diff_mat[row_idx, col_idx-1])
count += 1
loss = loss / count
return loss
def rank_consistency_v3(x, metric = None):
"""loss = rank_consistency_v3(x, metric)
input
-----
x: tensor, (batch, D1, D2 ...)
metric: a function or None
This function must be f(x1, x2) -> scalar
where x1 and x2 are in shape (D1, D2 ...)
if None, negative cosine similarity for
x1 and x2 of shape (D1, )
output
------
loss: scalar, loss value
Example
-------
>> x = torch.randn(4, 10)
>> x[1] = x[0] + 1.0
>> x[2] = x[0] + 2.0
>> x[3] = x[0] + 3.0
>> metric = lambda x1, x2: \
torch_nn_func.margin_ranking_loss(x1, x2, torch.ones_like(x1), 0.1)
>> rank_consistency_v3(x, metric)
tensor(.0)
"""
# batch size
bs = x.shape[0]
# loss to be accumulated
loss = 0.0
# metric
if metric is None:
# default use margin_ranking_loss
# x1 should be ranked higher
metric = lambda x1, x2: torch_nn_func.margin_ranking_loss(
x1, x2, torch.ones_like(x1), 0.1)
#
if bs < 2:
return loss
count = 0.0
# loop over the rows
for row_idx1 in np.arange(1, bs):
for row_idx2 in np.arange(0, row_idx1):
loss += metric(x[row_idx1],
x[row_idx2])
count += 1
loss = loss / count
return loss
if __name__ == "__main__":
print("loss and metric")
| 14,085 | 27.456566 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_nn.py | ##!/usr/bin/env python
"""
Common blocks for neural networks
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
from scipy import signal as scipy_signal
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import torch.nn.init as torch_init
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
# For blstm
class BLSTMLayer(torch_nn.Module):
""" Wrapper over dilated conv1D
Input tensor: (batchsize=1, length, dim_in)
Output tensor: (batchsize=1, length, dim_out)
We want to keep the length the same
"""
def __init__(self, input_dim, output_dim):
super(BLSTMLayer, self).__init__()
if output_dim % 2 != 0:
print("Output_dim of BLSTMLayer is {:d}".format(output_dim))
print("BLSTMLayer expects a layer size of even number")
sys.exit(1)
# bi-directional LSTM
self.l_blstm = torch_nn.LSTM(input_dim, output_dim // 2, \
bidirectional=True)
def forward(self, x):
# permute to (length, batchsize=1, dim)
blstm_data, _ = self.l_blstm(x.permute(1, 0, 2))
# permute it backt to (batchsize=1, length, dim)
return blstm_data.permute(1, 0, 2)
#
# 1D dilated convolution that keep the input/output length
class Conv1dKeepLength(torch_nn.Conv1d):
""" Wrapper for causal convolution
Input tensor: (batchsize=1, length, dim_in)
Output tensor: (batchsize=1, length, dim_out)
https://github.com/pytorch/pytorch/issues/1333
Note: Tanh is applied
"""
def __init__(self, input_dim, output_dim, dilation_s, kernel_s,
causal = False, stride = 1, groups=1, bias=True, \
tanh = True, pad_mode='constant'):
super(Conv1dKeepLength, self).__init__(
input_dim, output_dim, kernel_s, stride=stride,
padding = 0, dilation = dilation_s, groups=groups, bias=bias)
self.causal = causal
# input & output length will be the same
if self.causal:
# left pad to make the convolution causal
self.pad_le = dilation_s * (kernel_s - 1)
self.pad_ri = 0
else:
# pad on both sizes
self.pad_le = dilation_s * (kernel_s - 1) // 2
self.pad_ri = dilation_s * (kernel_s - 1) - self.pad_le
# we may wrap other functions too
if tanh:
self.l_ac = torch_nn.Tanh()
else:
self.l_ac = torch_nn.Identity()
self.pad_mode = pad_mode
#
return
def forward(self, data):
# permute to (batchsize=1, dim, length)
# add one dimension (batchsize=1, dim, ADDED_DIM, length)
# pad to ADDED_DIM
# squeeze and return to (batchsize=1, dim, length)
# https://github.com/pytorch/pytorch/issues/1333
x = torch_nn_func.pad(data.permute(0, 2, 1).unsqueeze(2), \
(self.pad_le, self.pad_ri, 0, 0),
mode = self.pad_mode).squeeze(2)
# tanh(conv1())
# permmute back to (batchsize=1, length, dim)
output = self.l_ac(super(Conv1dKeepLength, self).forward(x))
return output.permute(0, 2, 1)
#
# Moving average
class MovingAverage(Conv1dKeepLength):
""" Wrapper to define a moving average smoothing layer
Note: MovingAverage can be implemented using TimeInvFIRFilter too.
Here we define another Module dicrectly on Conv1DKeepLength
"""
def __init__(self, feature_dim, window_len, causal=False):
super(MovingAverage, self).__init__(
feature_dim, feature_dim, 1, window_len, causal,
groups=feature_dim, bias=False, tanh=False)
# set the weighting coefficients
torch_nn.init.constant_(self.weight, 1/window_len)
# turn off grad for this layer
for p in self.parameters():
p.requires_grad = False
def forward(self, data):
return super(MovingAverage, self).forward(data)
#
# FIR filter layer
class TimeInvFIRFilter(Conv1dKeepLength):
""" Wrapper to define a FIR filter over Conv1d
Note: FIR Filtering is conducted on each dimension (channel)
independently: groups=channel_num in conv1d
"""
def __init__(self, feature_dim, filter_coef,
causal=True, flag_train=False):
""" __init__(self, feature_dim, filter_coef,
causal=True, flag_train=False)
feature_dim: dimension of input data
filter_coef: 1-D tensor of filter coefficients
causal: FIR is causal or not (default: true)
flag_train: whether train the filter coefficients (default: false)
Input data: (batchsize=1, length, feature_dim)
Output data: (batchsize=1, length, feature_dim)
"""
super(TimeInvFIRFilter, self).__init__(
feature_dim, feature_dim, 1, filter_coef.shape[0], causal,
groups=feature_dim, bias=False, tanh=False)
if filter_coef.ndim == 1:
# initialize weight using provided filter_coef
with torch.no_grad():
tmp_coef = torch.zeros([feature_dim, 1,
filter_coef.shape[0]])
tmp_coef[:, 0, :] = filter_coef
tmp_coef = torch.flip(tmp_coef, dims=[2])
self.weight = torch.nn.Parameter(tmp_coef,
requires_grad=flag_train)
else:
print("TimeInvFIRFilter expects filter_coef to be 1-D tensor")
print("Please implement the code in __init__ if necessary")
sys.exit(1)
def forward(self, data):
return super(TimeInvFIRFilter, self).forward(data)
#
# Up sampling
class UpSampleLayer(torch_nn.Module):
""" Wrapper over up-sampling
Input tensor: (batchsize=1, length, dim)
Ouput tensor: (batchsize=1, length * up-sampling_factor, dim)
"""
def __init__(self, feature_dim, up_sampling_factor, smoothing=False):
super(UpSampleLayer, self).__init__()
# wrap a up_sampling layer
self.scale_factor = up_sampling_factor
self.l_upsamp = torch_nn.Upsample(scale_factor=self.scale_factor)
if smoothing:
self.l_ave1 = MovingAverage(feature_dim, self.scale_factor)
self.l_ave2 = MovingAverage(feature_dim, self.scale_factor)
else:
self.l_ave1 = torch_nn.Identity()
self.l_ave2 = torch_nn.Identity()
return
def forward(self, x):
# permute to (batchsize=1, dim, length)
up_sampled_data = self.l_upsamp(x.permute(0, 2, 1))
# permute it backt to (batchsize=1, length, dim)
# and do two moving average
return self.l_ave1(self.l_ave2(up_sampled_data.permute(0, 2, 1)))
class upsampleByTransConv(torch_nn.Module):
"""upsampleByTransConv
Upsampling layer using transposed convolution
"""
def __init__(self, feat_dim, output_dim, upsample_rate, window_ratio=5):
"""upsampleByTransConv(feat_dim, upsample_rate, window_ratio=5)
Args
----
feat_dim: int, input feature should be (batch, length, feat_dim)
upsample_rate, int, output feature will be
(batch, length*upsample_rate, feat_dim)
window_ratio: int, default 5, window length of transconv will be
upsample_rate * window_ratio
"""
super(upsampleByTransConv, self).__init__()
window_l = upsample_rate * window_ratio
self.m_layer = torch_nn.ConvTranspose1d(
feat_dim, output_dim, window_l, stride=upsample_rate)
self.m_uprate = upsample_rate
return
def forward(self, x):
""" y = upsampleByTransConv(x)
input
-----
x: tensor, (batch, length, feat_dim)
output
------
y: tensor, (batch, length*upsample_rate, output_dim)
"""
l = x.shape[1] * self.m_uprate
y = self.m_layer(x.permute(0, 2, 1))[:, :, 0:l]
return y.permute(0, 2, 1).contiguous()
class TimeVarFIRFilter(torch_nn.Module):
""" TimeVarFIRFilter
Given sequences of filter coefficients and a signal, do filtering
Filter coefs: (batchsize, signal_length, filter_order = K)
Signal: (batchsize, signal_length, 1)
For batch 0:
For n in [1, sequence_length):
output(0, n, 1) = \sum_{k=1}^{K} signal(0, n-k, 1)*coef(0, n, k)
Note: filter coef (0, n, :) is only used to compute the output
at (0, n, 1)
"""
def __init__(self):
super(TimeVarFIRFilter, self).__init__()
def forward(self, signal, f_coef):
"""
Filter coefs: (batchsize=1, signal_length, filter_order = K)
Signal: (batchsize=1, signal_length, 1)
Output: (batchsize=1, signal_length, 1)
For n in [1, sequence_length):
output(0, n, 1) = \sum_{k=1}^{K} signal(0, n-k, 1)*coef(0, n, k)
This method may be not efficient:
Suppose signal [x_1, ..., x_N], filter [a_1, ..., a_K]
output [y_1, y_2, y_3, ..., y_N, *, * ... *]
= a_1 * [x_1, x_2, x_3, ..., x_N, 0, ..., 0]
+ a_2 * [ 0, x_1, x_2, x_3, ..., x_N, 0, ..., 0]
+ a_3 * [ 0, 0, x_1, x_2, x_3, ..., x_N, 0, ..., 0]
"""
signal_l = signal.shape[1]
order_k = f_coef.shape[-1]
# pad to (batchsize=1, signal_length + filter_order-1, dim)
padded_signal = torch_nn_func.pad(signal, (0, 0, 0, order_k - 1))
y = torch.zeros_like(signal)
# roll and weighted sum, only take [0:signal_length]
for k in range(order_k):
y += torch.roll(padded_signal, k, dims=1)[:, 0:signal_l, :] \
* f_coef[:, :, k:k+1]
# done
return y
class SignalsConv1d(torch_nn.Module):
""" Filtering input signal with time invariant filter
Note: FIRFilter conducted filtering given fixed FIR weight
SignalsConv1d convolves two signals
Note: this is based on torch.nn.functional.conv1d
"""
def __init__(self):
super(SignalsConv1d, self).__init__()
def forward(self, signal, system_ir):
""" output = forward(signal, system_ir)
signal: (batchsize, length1, dim)
system_ir: (length2, dim)
output: (batchsize, length1, dim)
"""
if signal.shape[-1] != system_ir.shape[-1]:
print("Error: SignalsConv1d expects shape:")
print("signal (batchsize, length1, dim)")
print("system_id (batchsize, length2, dim)")
print("But received signal: {:s}".format(str(signal.shape)))
print(" system_ir: {:s}".format(str(system_ir.shape)))
sys.exit(1)
padding_length = system_ir.shape[0] - 1
groups = signal.shape[-1]
# pad signal on the left
signal_pad = torch_nn_func.pad(signal.permute(0, 2, 1),\
(padding_length, 0))
# prepare system impulse response as (dim, 1, length2)
# also flip the impulse response
ir = torch.flip(system_ir.unsqueeze(1).permute(2, 1, 0), \
dims=[2])
# convolute
output = torch_nn_func.conv1d(signal_pad, ir, groups=groups)
return output.permute(0, 2, 1)
# Sinc filter generator
class SincFilter(torch_nn.Module):
""" SincFilter
Given the cut-off-frequency, produce the low-pass and high-pass
windowed-sinc-filters.
If input cut-off-frequency is (batchsize=1, signal_length, 1),
output filter coef is (batchsize=1, signal_length, filter_order).
For each time step in [1, signal_length), we calculate one
filter for low-pass sinc filter and another for high-pass filter.
Example:
import scipy
import scipy.signal
import numpy as np
filter_order = 31
cut_f = 0.2
sinc_layer = SincFilter(filter_order)
lp_coef, hp_coef = sinc_layer(torch.ones(1, 10, 1) * cut_f)
w, h1 = scipy.signal.freqz(lp_coef[0, 0, :].numpy(), [1])
w, h2 = scipy.signal.freqz(hp_coef[0, 0, :].numpy(), [1])
plt.plot(w, 20*np.log10(np.abs(h1)))
plt.plot(w, 20*np.log10(np.abs(h2)))
plt.plot([cut_f * np.pi, cut_f * np.pi], [-100, 0])
"""
def __init__(self, filter_order):
super(SincFilter, self).__init__()
# Make the filter oder an odd number
# [-(M-1)/2, ... 0, (M-1)/2]
#
self.half_k = (filter_order - 1) // 2
self.order = self.half_k * 2 +1
def hamming_w(self, n_index):
""" prepare hamming window for each time step
n_index (batchsize=1, signal_length, filter_order)
For each step, n_index.shape is [-(M-1)/2, ... 0, (M-1)/2]
where,
n_index[0, 0, :] = [-(M-1)/2, ... 0, (M-1)/2]
n_index[0, 1, :] = [-(M-1)/2, ... 0, (M-1)/2]
...
output (batchsize=1, signal_length, filter_order)
output[0, 0, :] = hamming_window
output[0, 1, :] = hamming_window
...
"""
# Hamming window
return 0.54 + 0.46 * torch.cos(2 * np.pi * n_index / self.order)
def sinc(self, x):
""" Normalized sinc-filter sin( pi * x) / pi * x
https://en.wikipedia.org/wiki/Sinc_function
Assume x (batchsize, signal_length, filter_order) and
x[0, 0, :] = [-half_order, - half_order+1, ... 0 ..., half_order]
x[:, :, self.half_order] -> time index = 0, sinc(0)=1
"""
y = torch.zeros_like(x)
y[:,:,0:self.half_k]=torch.sin(np.pi * x[:, :, 0:self.half_k]) \
/ (np.pi * x[:, :, 0:self.half_k])
y[:,:,self.half_k+1:]=torch.sin(np.pi * x[:, :, self.half_k+1:])\
/ (np.pi * x[:, :, self.half_k+1:])
y[:,:,self.half_k] = 1
return y
def forward(self, cut_f):
""" lp_coef, hp_coef = forward(self, cut_f)
cut-off frequency cut_f (batchsize=1, length, dim = 1)
lp_coef: low-pass filter coefs (batchsize, length, filter_order)
hp_coef: high-pass filter coefs (batchsize, length, filter_order)
"""
# create the filter order index
with torch.no_grad():
# [- (M-1) / 2, ..., 0, ..., (M-1)/2]
lp_coef = torch.arange(-self.half_k, self.half_k + 1,
device=cut_f.device)
# [[[- (M-1) / 2, ..., 0, ..., (M-1)/2],
# [- (M-1) / 2, ..., 0, ..., (M-1)/2],
# ...
# ],
# [[- (M-1) / 2, ..., 0, ..., (M-1)/2],
# [- (M-1) / 2, ..., 0, ..., (M-1)/2],
# ...
# ]]
lp_coef = lp_coef.repeat(cut_f.shape[0], cut_f.shape[1], 1)
hp_coef = torch.arange(-self.half_k, self.half_k + 1,
device=cut_f.device)
hp_coef = hp_coef.repeat(cut_f.shape[0], cut_f.shape[1], 1)
# temporary buffer of [-1^n] for gain norm in hp_coef
tmp_one = torch.pow(-1, hp_coef)
# unnormalized filter coefs with hamming window
lp_coef = cut_f * self.sinc(cut_f * lp_coef) \
* self.hamming_w(lp_coef)
hp_coef = (self.sinc(hp_coef) \
- cut_f * self.sinc(cut_f * hp_coef)) \
* self.hamming_w(hp_coef)
# normalize the coef to make gain at 0/pi is 0 dB
# sum_n lp_coef[n]
lp_coef_norm = torch.sum(lp_coef, axis=2).unsqueeze(-1)
# sum_n hp_coef[n] * -1^n
hp_coef_norm = torch.sum(hp_coef * tmp_one, axis=2).unsqueeze(-1)
lp_coef = lp_coef / lp_coef_norm
hp_coef = hp_coef / hp_coef_norm
# return normed coef
return lp_coef, hp_coef
class BatchNorm1DWrapper(torch_nn.BatchNorm1d):
"""
"""
def __init__(self, num_features, eps=1e-05, momentum=0.1,
affine=True, track_running_stats=True):
super(BatchNorm1DWrapper, self).__init__(
num_features, eps, momentum, affine, track_running_stats)
def forward(self, data):
output = super(BatchNorm1DWrapper, self).forward(data.permute(0, 2, 1))
return output.permute(0, 2, 1)
class SignalFraming(torch_nn.Conv1d):
""" SignalFraming(w_len, h_len, w_type='Hamming')
Do framing on the signal. The implementation is based on conv1d
Args:
-----
w_len: window length (frame length)
h_len: hop length (frame shift)
w_type: type of window, (default='Hamming')
Hamming: Hamming window
else: square window
Note:
-----
input signal (batch, length, 1)
output signal (batch, frame_num, frame_length)
where frame_num = length + (frame_length - frame_num)
Compatibility with Librosa framing need to be checked
"""
def __init__(self, w_len, h_len, w_type='Hamming'):
super(SignalFraming, self).__init__(1, w_len, w_len, stride=h_len,
padding = 0, dilation = 1, groups=1, bias=False)
self.m_wlen = w_len
self.m_wtype = w_type
self.m_hlen = h_len
if w_type == 'Hamming':
self.m_win = scipy_signal.windows.hamming(self.m_wlen)
else:
self.m_win = np.ones([self.m_wlen])
# for padding
if h_len > w_len:
print("Error: SignalFraming(w_len, h_len)")
print("w_len cannot be < h_len")
sys.exit(1)
self.m_mat = np.diag(self.m_win)
self.m_pad_len_l = (w_len - h_len)//2
self.m_pad_len_r = (w_len - h_len) - self.m_pad_len_l
# filter [output_dim = frame_len, 1, input_dim=frame_len]
# No need to flip the filter coefficients
with torch.no_grad():
tmp_coef = torch.zeros([w_len, 1, w_len])
tmp_coef[:, 0, :] = torch.tensor(self.m_mat)
self.weight = torch.nn.Parameter(tmp_coef, requires_grad = False)
return
def forward(self, signal):
"""
signal: (batchsize, length1, 1)
output: (batchsize, num_frame, frame_length)
Note:
"""
if signal.shape[-1] > 1:
print("Error: SignalFraming expects shape:")
print("signal (batchsize, length, 1)")
sys.exit(1)
# 1. switch dimension from (batch, length, dim) to (batch, dim, length)
# 2. pad signal on the left to (batch, dim, length + pad_length)
signal_pad = torch_nn_func.pad(signal.permute(0, 2, 1),\
(self.m_pad_len_l, self.m_pad_len_r))
# switch dimension from (batch, dim, length) to (batch, length, dim)
return super(SignalFraming, self).forward(signal_pad).permute(0, 2, 1)
class Conv1dStride(torch_nn.Conv1d):
""" Wrapper for normal 1D convolution with stride (optionally)
Input tensor: (batchsize, length, dim_in)
Output tensor: (batchsize, length2, dim_out)
However, we wish that length2 = floor(length / stride)
Therefore,
padding_total_length - dilation_s * (kernel_s - 1) -1 + stride = 0
or,
padding_total_length = dilation_s * (kernel_s - 1) + 1 - stride
Conv1dBundle(input_dim, output_dim, dilation_s, kernel_s,
causal = False, stride = 1, groups=1, bias=True, \
tanh = True, pad_mode='constant')
input_dim: int, input dimension (input channel)
output_dim: int, output dimension (output channel)
kernel_s: int, kernel size of filter
dilation_s: int, dilation for convolution
causal: bool, whether causal convolution, default False
stride: int, stride size, default 1
groups: int, group for conv1d, default 1
bias: bool, whether add bias, default True
tanh: bool, whether use tanh activation, default True
pad_mode: str, padding method, default "constant"
"""
def __init__(self, input_dim, output_dim, kernel_s, dilation_s=1,
causal = False, stride = 1, groups=1, bias=True, \
tanh = True, pad_mode='constant'):
super(Conv1dStride, self).__init__(
input_dim, output_dim, kernel_s, stride=stride,
padding = 0, dilation = dilation_s, groups=groups, bias=bias)
self.pad_mode = pad_mode
self.causal = causal
# padding size
# input & output length will be the same
if self.causal:
# left pad to make the convolution causal
self.pad_le = dilation_s * (kernel_s - 1) + 1 - stride
self.pad_ri = 0
else:
# pad on both sizes
self.pad_le = (dilation_s*(kernel_s-1)+1-stride) // 2
self.pad_ri = (dilation_s*(kernel_s-1)+1-stride) - self.pad_le
# activation functions
if tanh:
self.l_ac = torch_nn.Tanh()
else:
self.l_ac = torch_nn.Identity()
def forward(self, data):
# https://github.com/pytorch/pytorch/issues/1333
# permute to (batchsize=1, dim, length)
# add one dimension as (batchsize=1, dim, ADDED_DIM, length)
# pad to ADDED_DIM
# squeeze and return to (batchsize=1, dim, length+pad_length)
x = torch_nn_func.pad(
data.permute(0, 2, 1).unsqueeze(2), \
(self.pad_le, self.pad_ri,0,0), \
mode = self.pad_mode).squeeze(2)
# tanh(conv1())
# permmute back to (batchsize=1, length, dim)
output = self.l_ac(super(Conv1dStride, self).forward(x))
return output.permute(0, 2, 1)
class MaxPool1dStride(torch_nn.MaxPool1d):
""" Wrapper for maxpooling
Input tensor: (batchsize, length, dim_in)
Output tensor: (batchsize, length2, dim_in)
However, we wish that length2 = floor(length / stride)
Therefore,
padding_total_length - dilation_s * (kernel_s - 1) -1 + stride = 0
or,
padding_total_length = dilation_s * (kernel_s - 1) + 1 - stride
MaxPool1dStride(kernel_s, stride, dilation_s=1)
"""
def __init__(self, kernel_s, stride, dilation_s=1):
super(MaxPool1dStride, self).__init__(
kernel_s, stride, 0, dilation_s)
# pad on both sizes
self.pad_le = (dilation_s*(kernel_s-1)+1-stride) // 2
self.pad_ri = (dilation_s*(kernel_s-1)+1-stride) - self.pad_le
def forward(self, data):
# https://github.com/pytorch/pytorch/issues/1333
# permute to (batchsize=1, dim, length)
# add one dimension as (batchsize=1, dim, ADDED_DIM, length)
# pad to ADDED_DIM
# squeeze and return to (batchsize=1, dim, length+pad_length)
x = torch_nn_func.pad(
data.permute(0, 2, 1).unsqueeze(2), \
(self.pad_le, self.pad_ri,0,0)).squeeze(2).contiguous()
# tanh(conv1())
# permmute back to (batchsize=1, length, dim)
output = super(MaxPool1dStride, self).forward(x)
return output.permute(0, 2, 1)
class AvePool1dStride(torch_nn.AvgPool1d):
""" Wrapper for average pooling
Input tensor: (batchsize, length, dim_in)
Output tensor: (batchsize, length2, dim_in)
However, we wish that length2 = floor(length / stride)
Therefore,
padding_total_length - dilation_s * (kernel_s - 1) -1 + stride = 0
or,
padding_total_length = dilation_s * (kernel_s - 1) + 1 - stride
MaxPool1dStride(kernel_s, stride, dilation_s=1)
"""
def __init__(self, kernel_s, stride):
super(AvePool1dStride, self).__init__(
kernel_s, stride, 0)
# pad on both sizes
self.pad_le = ((kernel_s-1)+1-stride) // 2
self.pad_ri = ((kernel_s-1)+1-stride) - self.pad_le
def forward(self, data):
# https://github.com/pytorch/pytorch/issues/1333
# permute to (batchsize=1, dim, length)
# add one dimension as (batchsize=1, dim, ADDED_DIM, length)
# pad to ADDED_DIM
# squeeze and return to (batchsize=1, dim, length+pad_length)
x = torch_nn_func.pad(
data.permute(0, 2, 1).unsqueeze(2), \
(self.pad_le, self.pad_ri,0,0)).squeeze(2).contiguous()
# tanh(conv1())
# permmute back to (batchsize=1, length, dim)
output = super(AvePool1dStride, self).forward(x)
return output.permute(0, 2, 1)
class Maxout1D(torch_nn.Module):
""" Maxout activation (along 1D)
Maxout(d_in, d_out, pool_size)
From https://github.com/pytorch/pytorch/issues/805
Arguments
---------
d_in: feature input dimension
d_out: feature output dimension
pool_size: window size of max-pooling
Usage
-----
l_maxout1d = Maxout1D(d_in, d_out, pool_size)
data_in = torch.rand([1, T, d_in])
data_out = l_maxout1d(data_in)
"""
def __init__(self, d_in, d_out, pool_size):
super().__init__()
self.d_in, self.d_out, self.pool_size = d_in, d_out, pool_size
self.lin = torch_nn.Linear(d_in, d_out * pool_size)
def forward(self, inputs):
# suppose inputs (batchsize, length, dim)
# shape (batchsize, length, out-dim, pool_size)
shape = list(inputs.size())
shape[-1] = self.d_out
shape.append(self.pool_size)
max_dim = len(shape) - 1
# shape (batchsize, length, out-dim * pool_size)
out = self.lin(inputs)
# view to (batchsize, length, out-dim, pool_size)
# maximize on the last dimension
m, i = out.view(*shape).max(max_dim)
return m
class MaxFeatureMap2D(torch_nn.Module):
""" Max feature map (along 2D)
MaxFeatureMap2D(max_dim=1)
l_conv2d = MaxFeatureMap2D(1)
data_in = torch.rand([1, 4, 5, 5])
data_out = l_conv2d(data_in)
Input:
------
data_in: tensor of shape (batch, channel, ...)
Output:
-------
data_out: tensor of shape (batch, channel//2, ...)
Note
----
By default, Max-feature-map is on channel dimension,
and maxout is used on (channel ...)
"""
def __init__(self, max_dim = 1):
super().__init__()
self.max_dim = max_dim
def forward(self, inputs):
# suppose inputs (batchsize, channel, length, dim)
shape = list(inputs.size())
if self.max_dim >= len(shape):
print("MaxFeatureMap: maximize on %d dim" % (self.max_dim))
print("But input has %d dimensions" % (len(shape)))
sys.exit(1)
if shape[self.max_dim] // 2 * 2 != shape[self.max_dim]:
print("MaxFeatureMap: maximize on %d dim" % (self.max_dim))
print("But this dimension has an odd number of data")
sys.exit(1)
shape[self.max_dim] = shape[self.max_dim]//2
shape.insert(self.max_dim, 2)
# view to (batchsize, 2, channel//2, ...)
# maximize on the 2nd dim
m, i = inputs.view(*shape).max(self.max_dim)
return m
class SelfWeightedPooling(torch_nn.Module):
""" SelfWeightedPooling module
Inspired by
https://github.com/joaomonteirof/e2e_antispoofing/blob/master/model.py
To avoid confusion, I will call it self weighted pooling
Using self-attention format, this is similar to softmax(Query, Key)Value
where Query is a shared learnarble mm_weight, Key and Value are the input
Sequence.
l_selfpool = SelfWeightedPooling(5, 1, False)
with torch.no_grad():
input_data = torch.rand([3, 10, 5])
output_data = l_selfpool(input_data)
"""
def __init__(self, feature_dim, num_head=1, mean_only=False):
""" SelfWeightedPooling(feature_dim, num_head=1, mean_only=False)
Attention-based pooling
input (batchsize, length, feature_dim) ->
output
(batchsize, feature_dim * num_head), when mean_only=True
(batchsize, feature_dim * num_head * 2), when mean_only=False
args
----
feature_dim: dimension of input tensor
num_head: number of heads of attention
mean_only: whether compute mean or mean with std
False: output will be (batchsize, feature_dim*2)
True: output will be (batchsize, feature_dim)
"""
super(SelfWeightedPooling, self).__init__()
self.feature_dim = feature_dim
self.mean_only = mean_only
self.noise_std = 1e-5
self.num_head = num_head
# transformation matrix (num_head, feature_dim)
self.mm_weights = torch_nn.Parameter(
torch.Tensor(num_head, feature_dim), requires_grad=True)
torch_init.kaiming_uniform_(self.mm_weights)
return
def _forward(self, inputs):
""" output, attention = forward(inputs)
inputs
------
inputs: tensor, shape (batchsize, length, feature_dim)
output
------
output: tensor
(batchsize, feature_dim * num_head), when mean_only=True
(batchsize, feature_dim * num_head * 2), when mean_only=False
attention: tensor, shape (batchsize, length, num_head)
"""
# batch size
batch_size = inputs.size(0)
# feature dimension
feat_dim = inputs.size(2)
# input is (batch, legth, feature_dim)
# change mm_weights to (batchsize, feature_dim, num_head)
# weights will be in shape (batchsize, length, num_head)
weights = torch.bmm(inputs,
self.mm_weights.permute(1, 0).contiguous()\
.unsqueeze(0).repeat(batch_size, 1, 1))
# attention (batchsize, length, num_head)
attentions = torch_nn_func.softmax(torch.tanh(weights),dim=1)
# apply attention weight to input vectors
if self.num_head == 1:
# We can use the mode below to compute self.num_head too
# But there is numerical difference.
# original implementation in github
# elmentwise multiplication
# weighted input vector: (batchsize, length, feature_dim)
weighted = torch.mul(inputs, attentions.expand_as(inputs))
else:
# weights_mat = (batch * length, feat_dim, num_head)
weighted = torch.bmm(
inputs.view(-1, feat_dim, 1),
attentions.view(-1, 1, self.num_head))
# weights_mat = (batch, length, feat_dim * num_head)
weighted = weighted.view(batch_size, -1, feat_dim * self.num_head)
# pooling
if self.mean_only:
# only output the mean vector
representations = weighted.sum(1)
else:
# output the mean and std vector
noise = self.noise_std * torch.randn(
weighted.size(), dtype=weighted.dtype, device=weighted.device)
avg_repr, std_repr = weighted.sum(1), (weighted+noise).std(1)
# concatenate mean and std
representations = torch.cat((avg_repr,std_repr),1)
# done
return representations, attentions
def forward(self, inputs):
""" output = forward(inputs)
inputs
------
inputs: tensor, shape (batchsize, length, feature_dim)
output
------
output: tensor
(batchsize, feature_dim * num_head), when mean_only=True
(batchsize, feature_dim * num_head * 2), when mean_only=False
"""
output, _ = self._forward(inputs)
return output
def debug(self, inputs):
return self._forward(inputs)
class Conv1dForARModel(Conv1dKeepLength):
""" Definition of dilated Convolution for autoregressive model
This module is based on block_nn.py/Conv1DKeepLength.
However, Conv1DKeepLength doesn't assume step-by-step generation
for autogressive model.
This Module further adds the method to generate output in AR model
Example:
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import tutorials.plot_lib as nii_plot
# Compare the results of two layers
batchsize = 1
input_dim = 1
output_dim = 1
length = 5
dilation = 2
kernel_s = 3
# Layers
conv1 = nii_nn.Conv1dKeepLength(
input_dim, output_dim, dilation, kernel_s,
causal=True, tanh=False, bias=True)
conv2 = Conv1dForARModel(input_dim, output_dim, dilation, kernel_s,
tanh=False, bias=True)
conv2.weight = conv1.weight
conv2.bias = conv1.bias
# Test
input = torch.rand([batchsize, length, input_dim])
with torch.no_grad():
output = conv1(input)
output2 = conv2(input)
out = torch.zeros([batchsize, length, output_dim])
for step in range(length):
out[:, step:step+1, :] = conv2(input[:, step:step+1, :], step)
print(output - output2)
print(output - out)
#nii_plot.plot_tensor(input, deci_width=2)
#nii_plot.plot_tensor(output, deci_width=2)
#nii_plot.plot_tensor(output2, deci_width=2)
#nii_plot.plot_tensor(out, deci_width=2)
"""
def __init__(self, input_dim, output_dim, dilation_s, kernel_s,
bias=True, tanh = True, causal=True):
""" Conv1dForARModel(input_dim, output_dim, dilation_s, kernel_s,
bias=True, tanh=True)
args
----
input_dim: int, input tensor should be (batchsize, length, input_dim)
output_dim: int, output tensor will be (batchsize, length, output_dim)
dilation_s: int, dilation size
kernel_s: int, kernel size
bias: bool, whether use bias term, default True
tanh: bool, whether apply tanh on the output, default True
causal: bool, whether the convoltuion is causal, default True
Note that causal==False, step-by-step AR generation will raise Error
"""
super(Conv1dForARModel, self).__init__(
input_dim, output_dim, dilation_s, kernel_s, \
causal = causal, stride = 1, groups=1, bias=bias, tanh = tanh)
# configuration options
self.use_bias = bias
self.use_tanh = tanh
self.kernel_s = kernel_s
self.dilation_s = dilation_s
self.out_dim = output_dim
self.causal = causal
# See slide http://tonywangx.github.io/slide.html#misc CURRENNT WaveNet,
# page 50-56 for example on kernel_s = 2
#
# buffer length, depends on kernel size and dilation size
# kernel_size = 3, dilation_size = 1 -> * * * -> buffer_len = 3
# kernel_size = 3, dilation_size = 2 -> * . * . * -> buffer_len = 5
self.buffer_len = (kernel_s - 1) * dilation_s + 1
self.buffer_data = None
# self.buffer_conv1d = None
return
def forward(self, x, step_idx = None):
""" output = forward(x, step_idx)
input
-----
x: tensor, in shape (batchsize, length, input_dim)
step_idx: int, the index of the current time step
or None
output
------
output: tensor, in shape (batchsize, length, output_dim)
If step_idx is True
------------------------
this is same as common conv1d forward method
If self.training is False
------------------------
This method assumes input and output tensors
are for one time step, i.e., length = 1 for both x and output.
This method should be used in a loop, for example:
model.eval()
for idx in range(total_time_steps):
...
output[:, idx:idx+1, :] = forward(x[:, idx:idx+1, :])
...
This Module will use a buffer to store the intermediate results.
See slide http://tonywangx.github.io/slide.html#misc CURRENNT WaveNet,
page 50-56 for example on kernel_s = 2
"""
if step_idx is None:
# normal training mode, use the common conv forward method
return super(Conv1dForARModel, self).forward(x)
else:
if self.causal is False:
print("Step-by-step generation cannot work on non-causal conv")
print("Please use causal=True for Conv1dForARModel")
sys.exit(1)
# step-by-step for generation in AR model
# initialize buffer if necessary
if step_idx == 0:
self.buffer_data = torch.zeros(
[x.shape[0], self.buffer_len, x.shape[-1]],
dtype=x.dtype, device=x.device)
#self.buffer_conv1d = torch.zeros(
# [x.shape[0], self.kernel_s, x.shape[-1]],
# dtype=x.dtype, device=x.device)
# Put new input data into buffer
# the position index to put the input data
tmp_ptr_save = step_idx % self.buffer_len
# assume x is (batchsize, length=1, input_dim), thus
# only take x[:, 0, :]
self.buffer_data[:, tmp_ptr_save, :] = x[:, 0, :]
## Method 1: do multiplication and summing
##
## initialize
#output_tensor = torch.zeros(
# [x.shape[0], self.out_dim], dtype=x.dtype, device=x.device)
## loop over the kernel
#for ker_idx in range(self.kernel_s):
# # which buffer should be retrieved for this kernel idx
# tmp_data_idx = (step_idx - ker_idx * self.dilation_s) \
# % self.buffer_len
# # apply the kernel and sum the product
# # note that self.weight[:, :, -1] is the 1st kernel
# output_tensor += torch.matmul(
# self.buffer_data[:, tmp_data_idx, :],
# self.weight[:, :, self.kernel_s - ker_idx - 1].T)
## Method 2: take advantage of conv1d API
# Method 2 is slower than Method1 when kernel size is small
## create a input buffer to conv1d
#idxs = [(step_idx - x * self.dilation_s) % self.buffer_len \
# for x in range(self.kernel_s)][::-1]
#self.buffer_conv1d = self.buffer_data[:, idxs, :].permute(0, 2, 1)
#output_tensor = torch_nn_func.conv1d(self.buffer_conv1d,
# self.weight).permute(0, 2, 1)
# Method 3:
batchsize = x.shape[0]
# which data buffer should be retrieved for each kernel
# [::-1] is necessary because self.weight[:, :, -1] corresponds to
# the first buffer, [:, :, -2] to the second ...
index_buf = [(step_idx - y * self.dilation_s) % self.buffer_len \
for y in range(self.kernel_s)][::-1]
# concanate buffers as a tensor [batchsize, input_dim * kernel_s]
# concanate weights as a tensor [input_dim * kernel_s, output_dim]
# (out_dim, in_dim, kernel_s)-permute->(out_dim, kernel_s, in_dim)
# (out_dim, kernel_s, in_dim)-reshape->(out_dim, in_dim * kernel_s)
output_tensor = torch.mm(
self.buffer_data[:, index_buf, :].view(batchsize, -1),
self.weight.permute(0, 2, 1).reshape(self.out_dim, -1).T)
# apply bias and tanh if necessary
if self.use_bias:
output_tensor += self.bias
if self.use_tanh:
output_tensor = torch.tanh(output_tensor)
# make it to (batch, length=1, output_dim)
return output_tensor.unsqueeze(1)
class AdjustTemporalResoIO(torch_nn.Module):
def __init__(self, list_reso, target_reso, list_dims):
"""AdjustTemporalResoIO(list_reso, target_reso, list_dims)
Module to change temporal resolution of input tensors.
Args
----
list_reso: list, list of temporal resolutions.
list_reso[i] should be the temporal resolution of the
(i+1)-th tensor
target_reso: int, target temporal resolution to be changed
list_dims: list, list of feat_dim for tensors
assume tensor to have shape (batchsize, time length, feat_dim)
Note
----
target_reso must be <= max(list_reso)
all([target_reso % x == 0 for x in list_reso if x < target_reso])
all([x % target_reso == 0 for x in list_reso if x < target_reso])
Suppose a tensor A (batchsize, time_length1, feat_dim_1) has
temporal resolution of 1. Tensor B has temporal resolution
k and is aligned with A. Then B[:, n, :] corresponds to
A[:, k*n:k*(n+1), :].
For example:
let k = 3, batchsize = 1, feat_dim = 1
---------------> time axis
0 1 2 3 4 5 6 7 8
A[0, 0:9, 0] = [ a b c d e f g h i ]
B[0, 0:3, 0] = [ * & ^ ]
[*] is aligned with [a b c]
[&] is aligned with [d e f]
[^] is aligned with [g h i]
Assume the input tensor list is [A, B]:
list_reso = [1, 3]
list_dims = [A.shape[-1], B.shape[-1]]
If target_reso = 3, then
B will not be changed
A (batchsize=1, time_length1=9, feat_dim=1) will be A_new (1, 3, 3)
B [0, 0:3, 0] = [ * & ^ ]
A_new[0, 0:3, :] = [ [a, [d, [g, ]
b, e, h,
c] f] i]
More concrete examples:
input_dims = [5, 3]
rates = [1, 6]
target_rate = 2
l_adjust = AdjustTemporalRateIO(rates, target_rate, input_dims)
data1 = torch.rand([2, 2*6, 5])
data2 = torch.rand([2, 2, 3])
data1_new, data2_new = l_adjust([data1, data2])
# Visualization requires matplotlib and tutorial.plot_lib as nii_plot
nii_plot.plot_tensor(data1)
nii_plot.plot_tensor(data1_new)
nii_plot.plot_tensor(data2)
nii_plot.plot_tensor(data2_new)
"""
super(AdjustTemporalResoIO, self).__init__()
list_reso = np.array(list_reso)
list_dims = np.array(list_dims)
# save
self.list_reso = list_reso
self.fatest_reso = min(list_reso)
self.slowest_reso = max(list_reso)
self.target_reso = target_reso
# check
if any(list_reso < 0):
print("Expects positive resolution in AdjustTemporalResoIO")
sys.exit(1)
if self.target_reso < 0:
print("Expects positive target_reso in AdjustTemporalResoIO")
sys.exit(1)
if any([x % self.target_reso != 0 for x in self.list_reso \
if x > self.target_reso]):
print("Resolution " + str(list_reso) + " incompatible")
print(" with target resolution {:d}".format(self.target_reso))
sys.exit(1)
if any([self.target_reso % x != 0 for x in self.list_reso \
if x < self.target_reso]):
print("Resolution " + str(list_reso) + " incompatible")
print(" with target resolution {:d}".format(self.target_reso))
sys.exit(1)
self.dim_change = []
self.reso_change = []
self.l_upsampler = []
for x, dim in zip(self.list_reso, list_dims):
if x > self.target_reso:
# up sample
# up-sample don't change feat dim, just duplicate frames
self.dim_change.append(1)
self.reso_change.append(x // self.target_reso)
self.l_upsampler.append(
nii_nn.UpSampleLayer(dim, x // self.target_reso))
elif x < self.target_reso:
# down sample
# for down-sample, we fold the multiple feature frames into one
self.dim_change.append(self.target_reso // x)
# use a negative number to indicate down-sample
self.reso_change.append(-self.target_reso // x)
self.l_upsampler.append(None)
else:
self.dim_change.append(1)
self.reso_change.append(1)
self.l_upsampler.append(None)
self.l_upsampler = torch_nn.ModuleList(self.l_upsampler)
# log down the dimensions after resolution change
self.dim = []
if list_dims is not None and len(list_dims) == len(self.dim_change):
self.dim = [x * y for x, y in zip(self.dim_change, list_dims)]
return
def get_dims(self):
return self.dim
def forward(self, tensor_list):
""" tensor_list = AdjustTemporalResoIO(tensor_list):
Adjust the temporal resolution of the input tensors.
For up-sampling, the tensor is duplicated
For down-samplin, multiple time steps are concated into a single vector
input
-----
tensor_list: list, list of tensors,
(batchsize, time steps, feat dim)
output
------
tensor_list: list, list of tensors,
(batchsize, time_steps * N, feat_dim * Y)
where N is the resolution change option in self.reso_change,
Y is the factor to change dimension in self.dim_change
"""
output_tensor_list = []
for in_tensor, dim_fac, reso_fac, l_up in \
zip(tensor_list, self.dim_change, self.reso_change,
self.l_upsampler):
batchsize = in_tensor.shape[0]
timelength = in_tensor.shape[1]
if reso_fac == 1:
# no change
output_tensor_list.append(in_tensor)
elif reso_fac < 0:
# down sample by concatenating
reso_fac *= -1
expected_len = timelength // reso_fac
trim_length = expected_len * reso_fac
if expected_len == 0:
# if input tensor length < down_sample factor
output_tensor_list.append(
torch.reshape(in_tensor[:, 0:1, :],
(batchsize, 1, -1)))
else:
# make sure that
output_tensor_list.append(
torch.reshape(in_tensor[:, 0:trim_length, :],
(batchsize, expected_len, -1)))
else:
# up-sampling by duplicating
output_tensor_list.append(l_up(in_tensor))
return output_tensor_list
class LSTMZoneOut(torch_nn.Module):
"""LSTM layer with zoneout
This module replies on LSTMCell
"""
def __init__(self, in_feat_dim, out_feat_dim,
bidirectional=False, residual_link=False, bias=True):
"""LSTMZoneOut(in_feat_dim, out_feat_dim,
bidirectional=False, residual_link=False, bias=True)
Args
----
in_feat_dim: int, input tensor should be (batch, length, in_feat_dim)
out_feat_dim: int, output tensor will be (batch, length, out_feat_dim)
bidirectional: bool, whether bidirectional, default False
residual_link: bool, whether residual link over LSTM, default False
bias: bool, bias option in torch.nn.LSTMCell, default True
When bidirectional is True, out_feat_dim must be an even number
When residual_link is True, out_feat_dim must be equal to in_feat_dim
"""
super(LSTMZoneOut, self).__init__()
# config parameters
self.in_dim = in_feat_dim
self.out_dim = out_feat_dim
self.flag_bi = bidirectional
self.flag_res = residual_link
self.bias = bias
# check
if self.flag_res and self.out_dim != self.in_dim:
print("Error in LSTMZoneOut w/ residual: in_feat_dim!=out_feat_dim")
sys.exit(1)
if self.flag_bi and self.out_dim % 2 > 0:
print("Error in Bidirecional LSTMZoneOut: out_feat_dim is not even")
sys.exit(1)
# layer
if self.flag_bi:
self.l_lstm1 = torch_nn.LSTMCell(
self.in_dim, self.out_dim//2, self.bias)
self.l_lstm2 = torch_nn.LSTMCell(
self.in_dim, self.out_dim//2, self.bias)
else:
self.l_lstm1 = torch_nn.LSTMCell(
self.in_dim, self.out_dim, self.bias)
self.l_lstm2 = None
return
def _zoneout(self, pre, cur, p=0.1):
"""zoneout wrapper
"""
if self.training:
with torch.no_grad():
mask = torch.zeros_like(pre).bernoulli_(p)
return pre * mask + cur * (1-mask)
else:
return cur
def forward(self, x):
"""y = LSTMZoneOut(x)
input
-----
x: tensor, (batchsize, length, in_feat_dim)
output
------
y: tensor, (batchsize, length, out_feat_dim)
"""
batchsize = x.shape[0]
length = x.shape[1]
# output tensor
y = torch.zeros([batchsize, length, self.out_dim],
device=x.device, dtype=x.dtype)
# recurrent
if self.flag_bi:
# for bi-directional
hid1 = torch.zeros([batchsize, self.out_dim//2],
device=x.device, dtype=x.dtype)
hid2 = torch.zeros_like(hid1)
cell1 = torch.zeros_like(hid1)
cell2 = torch.zeros_like(hid1)
for time in range(length):
# reverse time idx
rtime = length-time-1
# compute in both forward and reverse directions
hid1_new, cell1_new = self.l_lstm1(x[:,time, :], (hid1, cell1))
hid2_new, cell2_new = self.l_lstm2(x[:,rtime, :], (hid2, cell2))
hid1 = self._zoneout(hid1, hid1_new)
hid2 = self._zoneout(hid2, hid2_new)
y[:, time, 0:self.out_dim//2] = hid1
y[:, length-time-1, self.out_dim//2:] = hid2
else:
# for uni-directional
hid1 = torch.zeros([batchsize, self.out_dim],
device=x.device, dtype=x.dtype)
cell1 = torch.zeros_like(hid1)
for time in range(length):
hid1_new, cell1_new = self.l_lstm1(x[:, time, :], (hid1, cell1))
hid1 = self._zoneout(hid1, hid1_new)
y[:, time, :] = hid1
# residual part
if self.flag_res:
y = y+x
return y
class LinearInitialized(torch_nn.Module):
"""Linear layer with specific initialization
"""
def __init__(self, weight_mat, flag_train=True):
"""LinearInitialized(weight_mat, flag_trainable=True)
Args
----
weight_mat: tensor, (input_dim, output_dim),
the weight matrix for initializing the layer
flag_train: bool, where trainable or fixed, default True
This can be used for trainable filter bank. For example:
import sandbox.util_frontend as nii_front_end
l_fb = LinearInitialized(nii_front_end.linear_fb(fn, sr, filter_num))
y = l_fb(x)
"""
super(LinearInitialized, self).__init__()
self.weight = torch_nn.Parameter(weight_mat, requires_grad=flag_train)
return
def forward(self, x):
"""y = LinearInitialized(x)
input
-----
x: tensor, (batchsize, ..., input_feat_dim)
output
------
y: tensor, (batchsize, ..., output_feat_dim)
Note that weight is in shape (input_feat_dim, output_feat_dim)
"""
return torch.matmul(x, self.weight)
class GRULayer(torch_nn.Module):
"""GRULayer
There are two modes for forward
1. forward(x) -> process sequence x
2. forward(x[n], n) -> process n-th step of x
Example:
data = torch.randn([2, 10, 3])
m_layer = GRULayer(3, 3)
out = m_layer(data)
out_2 = torch.zeros_like(out)
for idx in range(data.shape[1]):
out_2[:, idx:idx+1, :] = m_layer._forwardstep(
data[:, idx:idx+1, :], idx)
"""
def __init__(self, in_size, out_size, flag_bidirec=False):
"""GRULayer(in_size, out_size, flag_bidirec=False)
Args
----
in_size: int, dimension of input feature per step
out_size: int, dimension of output feature per step
flag_bidirec: bool, whether this is bi-directional GRU
"""
super(GRULayer, self).__init__()
self.m_in_size = in_size
self.m_out_size = out_size
self.m_flag_bidirec = flag_bidirec
self.m_gru = torch_nn.GRU(
in_size, out_size, batch_first=True,
bidirectional=flag_bidirec)
# for step-by-step generation
self.m_grucell = None
self.m_buffer = None
return
def _get_gru_cell(self):
# dump GRU layer to GRU cell for step-by-step generation
self.m_grucell = torch_nn.GRUCell(self.m_in_size, self.m_out_size)
self.m_grucell.weight_hh.data = self.m_gru.weight_hh_l0.data
self.m_grucell.weight_ih.data = self.m_gru.weight_ih_l0.data
self.m_grucell.bias_hh.data = self.m_gru.bias_hh_l0.data
self.m_grucell.bias_ih.data = self.m_gru.bias_ih_l0.data
return
def _forward(self, x):
"""y = _forward(x)
input
-----
x: tensor, (batch, length, inputdim)
output
------
y: tensor, (batch, length, out-dim)
"""
out, hn = self.m_gru(x)
return out
def _forwardstep(self, x, step_idx):
"""y = _forwardstep(x)
input
-----
x: tensor, (batch, 1, inputdim)
output
------
y: tensor, (batch, 1, out-dim)
"""
if self.m_flag_bidirec:
print("Bi-directional GRU not supported for step-by-step mode")
sys.exit(1)
else:
if step_idx == 0:
# load weight as grucell
if self.m_grucell is None:
self._get_gru_cell()
# buffer
self.m_buffer = torch.zeros(
[x.shape[0], self.m_out_size],
device=x.device, dtype=x.dtype)
self.m_buffer = self.m_grucell(x[:, 0, :], self.m_buffer)
# (batch, dim) -> (batch, 1, dim)
return self.m_buffer.unsqueeze(1)
def forward(self, x, step_idx=None):
"""y = forward(x, step_idx=None)
input
-----
x: tensor, (batch, length, inputdim)
output
------
y: tensor, (batch, length, out-dim)
When step_idx >= 0, length must be 1, forward(x[:, n:n+1, :], n)
will process the x at the n-th step. The hidden state will be saved
in the buffer and used for n+1 step
"""
if step_idx is None:
# given full context
return self._forward(x)
else:
# step-by-step processing
return self._forwardstep(x, step_idx)
class LSTMLayer(torch_nn.Module):
"""LSTMLayer
There are two modes for forward
1. forward(x) -> process sequence x
2. forward(x[n], n) -> process n-th step of x
Example:
data = torch.randn([2, 10, 3])
m_layer = LSTMLayer(3, 3)
out = m_layer(data)
out_2 = torch.zeros_like(out)
for idx in range(data.shape[1]):
out_2[:, idx:idx+1, :] = m_layer._forwardstep(
data[:, idx:idx+1, :], idx)
"""
def __init__(self, in_size, out_size, flag_bidirec=False):
"""LSTMLayer(in_size, out_size, flag_bidirec=False)
Args
----
in_size: int, dimension of input feature per step
out_size: int, dimension of output feature per step
flag_bidirec: bool, whether this is bi-directional GRU
"""
super(LSTMLayer, self).__init__()
self.m_in_size = in_size
self.m_out_size = out_size
self.m_flag_bidirec = flag_bidirec
self.m_lstm = torch_nn.LSTM(
input_size=in_size, hidden_size=out_size,
batch_first=True,
bidirectional=flag_bidirec)
# for step-by-step generation
self.m_lstmcell = None
self.m_c_buf = None
self.m_h_buf = None
return
def _get_lstm_cell(self):
# dump LSTM layer to LSTM cell for step-by-step generation
self.m_lstmcell = torch_nn.LSTMCell(self.m_in_size, self.m_out_size)
self.m_lstmcell.weight_hh.data = self.m_lstm.weight_hh_l0.data
self.m_lstmcell.weight_ih.data = self.m_lstm.weight_ih_l0.data
self.m_lstmcell.bias_hh.data = self.m_lstm.bias_hh_l0.data
self.m_lstmcell.bias_ih.data = self.m_lstm.bias_ih_l0.data
return
def _forward(self, x):
"""y = _forward(x)
input
-----
x: tensor, (batch, length, inputdim)
output
------
y: tensor, (batch, length, out-dim)
"""
out, hn = self.m_lstm(x)
return out
def _forwardstep(self, x, step_idx):
"""y = _forwardstep(x)
input
-----
x: tensor, (batch, 1, inputdim)
output
------
y: tensor, (batch, 1, out-dim)
"""
if self.m_flag_bidirec:
print("Bi-directional GRU not supported for step-by-step mode")
sys.exit(1)
else:
if step_idx == 0:
# For the 1st time step, prepare the LSTM Cell and buffer
# load weight as LSTMCell
if self.m_lstmcell is None:
self._get_lstm_cell()
# buffer
self.m_c_buf = torch.zeros([x.shape[0], self.m_out_size],
device=x.device, dtype=x.dtype)
self.m_h_buf = torch.zeros_like(self.m_c_buf)
# do generation
self.m_h_buf, self.m_c_buf = self.m_lstmcell(
x[:, 0, :], (self.m_h_buf, self.m_c_buf))
# (batch, dim) -> (batch, 1, dim)
return self.m_h_buf.unsqueeze(1)
def forward(self, x, step_idx=None):
"""y = forward(x, step_idx=None)
input
-----
x: tensor, (batch, length, inputdim)
output
------
y: tensor, (batch, length, out-dim)
When step_idx >= 0, length must be 1, forward(x[:, n:n+1, :], n)
will process the x at the n-th step. The hidden state will be saved
in the buffer and used for n+1 step
"""
if step_idx is None:
# given full context
return self._forward(x)
else:
# step-by-step processing
return self._forwardstep(x, step_idx)
class DropoutForMC(torch_nn.Module):
"""Dropout layer for Bayesian model
THe difference is that we do dropout even in eval stage
"""
def __init__(self, p, dropout_flag=True):
super(DropoutForMC, self).__init__()
self.p = p
self.flag = dropout_flag
return
def forward(self, x):
return torch_nn_func.dropout(x, self.p, training=self.flag)
if __name__ == "__main__":
print("Definition of block NN")
| 61,631 | 36.172497 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_resnet_new.py | ##!/usr/bin/env python
"""
ResNet model
Modified based on https://github.com/joaomonteirof/e2e_antispoofing
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import torch.nn.init as torch_init
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
class ResNetBlock1D(torch_nn.Module):
""" ResNetBlock1D(inplane, outplane, dilation=1, stride=1,
kernel=[1, 3, 1], expansion = 4)
Args
----
inplane: int, input feature dimension.
outplane: int, output feature dimension.
dilation: int, convolution dilation
stride: int, stride size
kernel: [int, int, int], the kernel size of the 3 conv layers
expansion: int, ratio for the bottleneck
"""
def __init__(self, inplane, outplane, dilation=1, stride=1,
kernel=[1, 3, 1], expansion = 4, act_type='ReLU'):
super(ResNetBlock1D, self).__init__()
#
self.ins = inplane
self.outs = outplane
self.expansion = expansion
self.hid = self.ins // expansion
dl = dilation
#
# block1 (batch, input_dim, length) -> (batch, hid_dim, length)
pad = self._get_pad(1, dilation, kernel[0])
self.conv1 = torch_nn.Sequential(
torch_nn.Conv1d(self.ins, self.hid, kernel[0], 1, pad, dilation),
torch_nn.BatchNorm1d(self.hid),
self._get_act(act_type))
# block2 (batch, hid_dim, length) -> (batch, hid_dim, length // stride)
pad = self._get_pad(stride, dilation, kernel[1])
self.conv2 = torch_nn.Sequential(
torch_nn.Conv1d(self.hid, self.hid, kernel[1], stride, pad, dl),
torch_nn.BatchNorm1d(self.hid),
self._get_act(act_type))
# block3
pad = self._get_pad(1, dilation, kernel[2])
self.conv3 = torch_nn.Sequential(
torch_nn.Conv1d(self.hid, self.outs, kernel[2], 1, pad, dl),
torch_nn.BatchNorm1d(self.outs))
self.output_act = self._get_act(act_type)
# change input dimension if necessary
if self.ins != self.outs or stride != 1:
pad = self._get_pad(stride, dilation, kernel[1])
self.changeinput = torch_nn.Sequential(
torch_nn.Conv1d(self.ins,self.outs, kernel[1], stride, pad, dl),
torch_nn.BatchNorm1d(self.outs))
else:
self.changeinput = torch_nn.Identity()
return
def _get_act(self, act_type):
if act_type == 'LeakyReLU':
return torch_nn.LeakyReLU()
elif act_type == 'ELU':
return torch_nn.ELU()
elif act_type == 'GELU':
return torch_nn.GELU()
else:
return torch_nn.ReLU()
def _get_pad(self, stride, dilation, kernel):
pad = (dilation * (kernel - 1) + 1 - stride) // 2
return pad
def forward(self, input_data):
"""output = ResNetBlock(input_data)
input: tensor, (batchsize, dimension, length)
output: tensor, (batchsize, dimension, length)
"""
output = self.conv1(input_data)
output = self.conv2(output)
output = self.conv3(output)
output = output + self.changeinput(input_data)
output = self.output_act(output)
return output
class ResNet1D(torch_nn.Module):
"""
"""
def __init__(self, inplane, outplanes, kernels, dilations, strides, ratios,
block_module = ResNetBlock1D, act_type = 'ReLU'):
super(ResNet1D, self).__init__()
#
tmp_ins = [inplane] + outplanes[:-1]
tmp_outs = outplanes
layer_list = []
for indim, outdim, kernel, dilation, stride, expand in zip(
tmp_ins, tmp_outs, kernels, dilations, strides, ratios):
layer_list.append(
block_module(indim, outdim, dilation, stride, kernel,
expand, act_type))
self.m_layers = torch_nn.Sequential(*layer_list)
return
def forward(self, input_data, length_first=True):
""" output = ResNet(input_data, swap_dim=True)
input
-----
input_data: tensor, (batch, input_dim, length),
or (batch, length, input_dim)
length_first: bool, True, this is used when input_data is
(batch, length, input_dim). Otherwise, False
output
------
output_data: tensor, (batch, length, input_dim) if length_first True
else (batch, input_dim, length)
"""
if length_first:
return self.m_layers(input_data.permute(0, 2, 1).contiguous()).permute(0, 2, 1).contiguous()
else:
return self.m_layers(input_data)
if __name__ == "__main__":
print("Implementation of ResNet for 1D signals")
| 5,137 | 32.363636 | 104 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_attention.py | ##!/usr/bin/env python
"""
Blocks for attention mechanism
Implementation is based on https://github.com/soobinseo/Transformer-TTS.git,
but code is re-facotrized:
DotScaledAttention and MultiheadAttention are separated.
The former is the core attention softmax(QK^T/sqrt(d))V,
with optional mask to mask dummy query and dummy key-value
that zero-padded due to the varied sequence length in batch
The former further includes the mask due to causal dependency
between output and input
The latter does split-> transform -> DotScaledAtt -> concat -> transform
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2023, Xin Wang"
# ================================================
# DotScaledAttention & multi-head implementation
# ================================================
class DotScaledAttention(torch_nn.Module):
"""DotScaledAttention in Transformer
O = q_mask * softmax( (Q K^\top / \sqrt(d)) k_mask ) V, where
Q: (batch, length1, dimension)
K: (batch, length2, dimension)
V: (batch, length2, dimension2)
k_mask: (batch, length1)
q_mask: (batch, length1)
Example:
l_m1 = DotScaledAttention()
q_data = torch.rand([5, 100, 64])
k_data2 = torch.rand([5, 40, 64])
v_data3 = torch.rand([5, 40, 32])
q_mask = torch.ones([5, 100])
q_mask[0, -4:] = 0
q_mask[1, -5:] = 0
q_mask_bin = q_mask.eq(0)
k_mask = torch.ones([5, 40])
k_mask[0, -4:] = 0
k_mask[1, -5:] = 0
k_mask_bin = k_mask.eq(0)
o1, a1 = l_m1(q_data, k_data2, v_data3, q_mask_bin, k_mask_bin)
# causal
l_m1 = DotScaledAttention(True)
data = torch.rand([5, 100, 64])
q_mask = torch.ones([5, 100])
q_mask_bin = q_mask.eq(0)
o1, a1 = l_m1(data, data, data, q_mask_bin, q_mask_bin)
o1[0, 1] - a1[0, 1, 0] * data[0, 0] - a1[0, 1, 1] * data[0, 1]
"""
def __init__(self, flag_causal=False, dropout=None):
super(DotScaledAttention, self).__init__()
self.flag_causal = flag_causal
if dropout is not None:
self.m_drop = torch_nn.Dropout(p=dropout)
else:
self.m_drop = None
return
def forward(self, Q, K, V, q_mask=None, k_mask=None):
"""O = DotScaledAttention(Q, K, V, q_mask=None, k_mask=None)
O = q_mask * softmax( (Q K^\top / \sqrt(d)) k_mask ) V
input:
------
Q: tensor, (batch, length1, dimension)
K: tensor, (batch, length2, dimension)
V: tensor, (batch, length2, dimension2)
k_mask: None or tensor, (batch, length2)
q_mask: None or tensor, (batch, length1)
output
------
O: tensor, (batch, length1, dimension2)
attn: tensor, (batch, length1, length2), attention matrix
k_mask[i] is a mask for the i-th key/value, k_mask[i, j]==True
indicates that K[i][j] and V[i][j] should be masked.
attention[i][:, j] should be zero
q_mask[i] is a mask for the i-the query, q_mask[i, j]==True
indicates that output O[i][j] should be masked
"""
bsize = Q.shape[0]
feat_dim = Q.shape[-1]
q_len = Q.shape[1]
k_len = K.shape[1]
# Q K^\top
# attn has shape (length1, length2)
attn = torch.bmm(Q, K.transpose(1, 2)) / np.sqrt(feat_dim)
# apply k_mask to mask dummy key/value (by setting attn to 0)
if k_mask is not None:
# (batch, length2) -> (batch, length1, length2) by duplicating
mask_tmp = k_mask.unsqueeze(1).repeat(1, q_len, 1)
# if causal dependency, add diagonal mask
# mask_tmp[:, i, >i] should be True
if self.flag_causal and q_len == k_len:
# length2 must be == length1
# create upper triagle (length1, length1)
tria_tmp = torch.triu(torch.ones_like(mask_tmp[0]), diagonal=1)
# repeat to batch
tria_tmp = tria_tmp.unsqueeze(0).repeat(bsize, 1, 1).gt(0)
# overlap the upper-triangle matrix with the k_mask
mask_tmp = torch.bitwise_or(mask_tmp, tria_tmp)
elif self.flag_causal and q_len == k_len:
# even if no need to mask dummy input, it is necessary to
# mask for causal self-attention
mask_tmp = torch.triu(torch.ones([k_len, k_len]), diagonal=1)
# repeat to batch
mask_tmp = mask_tmp.unsqueeze(0).repeat(bsize, 1, 1).gt(0)
mask_tmp = mask_tmp.to(device=Q.device)
else:
# no k_mask provided, neither is k_mask provided
mask_tmp = None
# mask the attn matrix if necessary
if mask_tmp != None:
attn = attn.masked_fill(mask_tmp, -2 ** 32 +1)
# softmax, over length2 of the (batch, length1, length2)
attn = torch_nn_func.softmax(attn, dim=-1)
# apply q_mask
if q_mask is not None:
# (batch, length1, 1) -> (batch, length1, length2)
mask_tmp = q_mask.unsqueeze(-1).repeat(1, 1, k_len)
# mask query (row) that should be dummy
attn = attn.masked_fill(mask_tmp, 0)
# apply dropout is necessary
if self.m_drop is not None:
attn = self.m_drop(attn)
# o = attn * V
O = torch.bmm(attn, V)
return O, attn
class MultiheadAttention(torch_nn.Module):
"""Multihead Attention in Transformer
V, K, Q -> linear -> split -> DotScaledAttention -> concate -> linear
Q: (batch, lengthQ, feat_dimK)
K: (batch, lengthK, feat_dimK)
V: (batch, lengthK, feat_dimV)
k_mask: (batch, lengthK)
q_mask: (batch, lengthQ)
Example:
q_data = torch.rand([5, 100, 64])
k_data2 = torch.rand([5, 40, 64])
v_data3 = torch.rand([5, 40, 32])
q_mask = torch.ones([5, 100])
q_mask[0, -4:] = 0
q_mask[1, -5:] = 0
q_mask_bin = q_mask.eq(0)
k_mask = torch.ones([5, 40])
k_mask[0, -4:] = 0
k_mask[1, -5:] = 0
k_mask_bin = k_mask.eq(0)
l_m = MultiheadAttention(64, 32, 4)
data_out = l_m.forward(v_data3, k_data2, q_data, k_mask_bin, q_mask_bin)
"""
def __init__(self, feat_dim_k, feat_dim_v, num_head=4,
flag_cat_q=True, flag_causal=False, dropout=None,
with_bias=False, flag_norm_before=False):
"""MultiheadAttention(num_head=4, flag_cat_q=True)
Args
----
feat_dim_k: int, feat_dimension of Query and Key
feat_dim_v: int, feat_dimension of Value
num_head: int, number of heads
flag_cat_q: bool, if true, concate(query, attention's output)
flag_causal: bool, causal dependency in self-attention
with_bias: bool, bias in feedforward layer for multi-head splitting?
(default False)
dropout: float or None, dropout rate on attention matrix
(default None)
flag_norm_before: bool, whether do layer normalize before attention
(default False). If true, the input q, k, and v should
be layerer normed before given to forward()
When flag_causal is True, Q, K, V must have same temporal length
"""
super(MultiheadAttention, self).__init__()
# log information
self.flag_causal = flag_causal
self.num_head = num_head
# feedforward layers
if feat_dim_k % self.num_head > 0 or feat_dim_v % self.num_head > 0:
print("feat_dim_k cannot be divided by num_head")
sys.exit(1)
self.m_q_fc = torch_nn.Linear(feat_dim_k, feat_dim_k, bias=with_bias)
self.m_k_fc = torch_nn.Linear(feat_dim_k, feat_dim_k, bias=with_bias)
self.m_v_fc = torch_nn.Linear(feat_dim_v, feat_dim_v, bias=with_bias)
torch_nn.init.xavier_uniform_(
self.m_q_fc.weight, gain=torch_nn.init.calculate_gain('linear'))
torch_nn.init.xavier_uniform_(
self.m_k_fc.weight, gain=torch_nn.init.calculate_gain('linear'))
torch_nn.init.xavier_uniform_(
self.m_v_fc.weight, gain=torch_nn.init.calculate_gain('linear'))
# core attention
self.m_attn = DotScaledAttention(self.flag_causal, dropout)
# dropout
if dropout is not None:
self.m_drop = torch_nn.Dropout(p=dropout)
else:
self.m_drop = None
# output linear layer
self.flag_cat_q = flag_cat_q
if self.flag_cat_q:
self.m_output = torch_nn.Linear(feat_dim_k+feat_dim_v, feat_dim_v)
else:
self.m_output = torch_nn.Linear(feat_dim_v, feat_dim_v)
torch_nn.init.xavier_uniform_(
self.m_output.weight, gain=torch_nn.init.calculate_gain('linear'))
#
self.m_layernorm = torch_nn.LayerNorm(feat_dim_v)
self.flag_norm_before = flag_norm_before
if feat_dim_k != feat_dim_v:
print("Warning: query/key and value differ in feature dimensions.")
print("Residual connection will not be used")
return
def forward(self, value, key, query, k_mask=None, q_mask=None):
"""O, attn = MultiheadAttention(value, key, query, k_mask, q_mask)
input:
------
Q: (batch, lengthQ, feat_dimK)
K: (batch, lengthK, feat_dimK)
V: (batch, lengthK, feat_dimV)
k_mask: None or tensor, (batch, length2)
q_mask: None or tensor, (batch, length1)
output
------
O: tensor, (batch, length1, dimension2)
attn: tensor, (batch, length1, length2), attention matrix
k_mask[i] is a mask for the i-th key/value, k_mask[i, j]==True
indicates that K[i][j] and V[i][j] should be masked.
attention[i][:, j] should be zero
q_mask[i] is a mask for the i-the query, q_mask[i, j]==True
indicates that output O[i][j] should be masked
"""
bsize = value.size(0)
k_len = key.size(1)
q_len = query.size(1)
if self.flag_causal and k_len != q_len:
print("Causal Attention, Q,V,K must have same length in time")
sys.exit(1)
# transform and split the input Q, K, V
def _trans_split(data_mat, trans_func, head):
bsize, length, dim = data_mat.shape
# (batch, length, feat_dim) -> (batch, length, feat_dimV)
# -> (batch, lengthK, num_head, feat_dimV / num_head)
tmp_mat = trans_func(data_mat).view(bsize, length, head, -1)
# -> ( num_head, batch, lengthK, feat_dimV / num_head)
tmp_mat = tmp_mat.permute(2, 0, 1, 3).contiguous()
# -> ( num_head * batch, lengthK, feat_dimV / num_head)
tmp_mat = tmp_mat.view(-1, length, tmp_mat.shape[-1])
return tmp_mat
value_mul = _trans_split(value, self.m_v_fc, self.num_head)
key_mul = _trans_split(key, self.m_k_fc, self.num_head)
query_mul = _trans_split(query, self.m_q_fc, self.num_head)
# duplicate masks to multi heads
if q_mask is not None:
q_mask_tmp = q_mask.repeat(self.num_head, 1)
else:
q_mask_tmp = None
if k_mask is not None:
k_mask_tmp = k_mask.repeat(self.num_head, 1)
else:
k_mask_tmp = None
# attention and sum
o_mul, attn = self.m_attn(query_mul, key_mul, value_mul,
q_mask_tmp, k_mask_tmp)
# recover it back
# ( num_head * batch, lengthQ, feat_dimV / num_head) ->
# ( num_head, batch, lengthQ, feat_dimV / num_head) ->
o_mul = o_mul.view(self.num_head, bsize, q_len, -1)
# -> ( batch, lengthQ, feat_dimV)
o_mat = o_mul.permute(1, 2, 0, 3).contiguous().view(bsize, q_len, -1)
# concatenate the input query and output of attention if necessary
if self.flag_cat_q:
# (batch, lengthQ, feat_dimQ + feat_dimV)
o_mat = torch.cat([o_mat, query], dim=-1)
# linear
o_mat = self.m_output(o_mat)
# dropout
if self.m_drop:
o_mat = self.m_drop(o_mat)
# residual & layer norm
if o_mat.shape[-1] == query.shape[-1]:
o_mat = o_mat + query
# layer normalize after
if not self.flag_norm_before:
o_mat = self.m_layernorm(o_mat)
return o_mat, attn
# ====================
# misc
# ====================
def position_encoding(n_pos, n_dim, padding_idx=None):
"""Position encoding in Transformer
input:
------
n_pos: int, pos, number of possible positions
n_dim: int, n_dim//2 = i, number of hidden dimensions
output:
------
sin_tab: np.array, (n_pos, n_dim)
sin_tab[n, 2i] = sin(n / 10000 ^ (2i / n_dim))
sin_tab[n, 2i+1] = cos(n / 10000 ^ (2i / n_dim))
Example:
data = position_encoding(1024, 512, 0)
"""
# make sure that n_dim is an even number
if n_dim % 2 > 0:
print("position_encoding: n_dim should be an even number")
sys.exit(1)
# create the table
sin_tab = np.zeros([n_pos, n_dim])
for idx in np.arange(n_dim // 2):
# period: 10000 ^ (2i / n_dim)
pd = np.power(10000, 2 * idx / n_dim)
# sin(n / 10000 ^ (2i / n_dim))
sin_tab[:, 2 * idx] = np.sin( np.arange(n_pos) / pd)
# cos(n / 10000 ^ ((2i+1) / n_dim))
sin_tab[:, 2 * idx+1] = np.cos( np.arange(n_pos) / pd)
# remove the dummy positioning encoding
if padding_idx is not None:
sin_tab[padding_idx] = 0
return sin_tab
class FeedforwardBlock(torch_nn.Module):
"""Feedforward block in Transformer
"""
def __init__(self, feat_dim):
super(FeedforwardBlock, self).__init__()
self.m_block = torch_nn.Sequential(
torch_nn.Linear(feat_dim, feat_dim * 4),
torch_nn.ReLU(),
torch_nn.Linear(feat_dim * 4, feat_dim)
#torch_nn.Dropout(p=0.1)
)
self.m_layernorm = torch_nn.LayerNorm(feat_dim)
# initialization
torch_nn.init.xavier_uniform_(
self.m_block[0].weight, gain=torch_nn.init.calculate_gain('relu'))
torch_nn.init.xavier_uniform_(
self.m_block[2].weight, gain=torch_nn.init.calculate_gain('linear'))
return
def forward(self, feat):
""" out = FeedforwardBlock(feat)
input
-----
feat: tensor, (batch, length, feat_dim)
output
------
output: tensor, (batch, length, feat_dim)
"""
return self.m_layernorm(self.m_block(feat) + feat)
class FeedforwardBlockv2(torch_nn.Module):
"""Feedforward block in Transformer
"""
def __init__(self, feat_dim, dropout=0.0, flag_norm_before=False):
super(FeedforwardBlockv2, self).__init__()
self.m_block = torch_nn.Sequential(
torch_nn.Linear(feat_dim, feat_dim * 4),
torch_nn.ReLU(),
torch_nn.Dropout(p=dropout),
torch_nn.Linear(feat_dim * 4, feat_dim)
)
self.flag_norm_before = flag_norm_before
self.m_layernorm = torch_nn.LayerNorm(feat_dim)
# initialization
torch_nn.init.xavier_uniform_(
self.m_block[0].weight, gain=torch_nn.init.calculate_gain('relu'))
torch_nn.init.xavier_uniform_(
self.m_block[-1].weight, gain=torch_nn.init.calculate_gain('linear'))
return
def forward(self, feat):
""" out = FeedforwardBlock(feat)
input
-----
feat: tensor, (batch, length, feat_dim)
output
------
output: tensor, (batch, length, feat_dim)
"""
if not self.flag_norm_before:
return self.m_layernorm(self.m_block(feat) + feat)
else:
return self.m_block(feat) + feat
if __name__ == "__main__":
print("block_attention.py")
| 16,852 | 33.464213 | 81 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_wavenet.py | #!/usr/bin/env python
"""
model.py for WaveNet
version: 1
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import time
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import core_scripts.other_tools.debug as nii_debug
import core_scripts.other_tools.display as nii_warn
import sandbox.block_nn as nii_nn
import sandbox.block_dist as nii_dist
import sandbox.util_dsp as nii_dsp
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
##############
#
class CondModule(torch_nn.Module):
""" Conditiona module: upsample and transform input features
"""
def __init__(self, input_dim, output_dim, up_sample, \
blstm_s = 64, cnn_kernel_s = 3):
""" CondModule(input_dim, output_dim, up_sample,
blstm_s=64, cnn_kernel_s=3)
Args
----
input_dim: int, input tensor should be (batchsize, len1, input_dim)
output_dim: int, output tensor will be (batchsize, len2, output_dim)
up_sample: int, up-sampling rate, len2 = len1 * up_sample
blstm_s: int, layer size of the Bi-LSTM layer
cnn_kernel_s: int, kernel size of the conv1d
"""
super(CondModule, self).__init__()
# configurations
self.input_dim = input_dim
self.output_dim = output_dim
self.up_sample = up_sample
self.blstm_s = blstm_s
self.cnn_kernel_s = cnn_kernel_s
# layers
self.l_blstm = nii_nn.BLSTMLayer(input_dim, self.blstm_s)
self.l_conv1d = nii_nn.Conv1dKeepLength(
self.blstm_s, output_dim, 1, self.cnn_kernel_s)
self.l_upsamp = nii_nn.UpSampleLayer(
self.output_dim, self.up_sample, True)
def forward(self, feature):
""" transformed_feat = forward(input_feature)
input
-----
feature: (batchsize, length, input_dim)
output
------
transformed_feat: tensor (batchsize, length*up_sample, out_dim)
"""
return self.l_upsamp(self.l_conv1d(self.l_blstm(feature)))
class WaveNetBlock(torch_nn.Module):
""" WaveNet block based on dilated-1D, gated-activation, and skip-connect.
Based on http://tonywangx.github.io/slide.html#misc CURRENNT WaveNet,
page 19-31.
"""
def __init__(self, input_dim, skip_ch_dim, gated_act_dim, cond_dim,
dilation_size, cnn_kernel_size=2, causal=True):
""" WaveNetBlock(input_dim, skip_ch_dim, gated_act_dim, cond_dim,
dilation_size, cnn_kernel_size = 2)
Args
----
input_dim: int, input tensor should be (batch-size, length, input_dim)
this is the dimension of residual channel
skip_ch_dim: int, tensors to be send to output blocks is in shape
(batch-size, length, skip_ch_dim)
gated_act_dim: int, tensors given by tanh(.) * sig(.) is in shape
(batch-size, length, gated_act_dim)
cond_dim: int, conditional feature (batchsize, length, cond_dim)
dilation_size: int, dilation size of the conv
cnn_kernel_size: int, kernel size of dilated conv1d (default, 2)
causal: bool, whether this block is used in AR model (default, True)
Note that causal==False will raise error if step-by-step generation
is conducted by inference(input_feat, cond_feat, step_idx) with
step_idx != None.
For causal==False, always use inference(input_feat, cond_feat, None)
"""
super(WaveNetBlock, self).__init__()
#####
# configurations
#####
# input tensor: (batchsize, length, self.input_dim)
self.input_dim = input_dim
# tensor sent to next WaveNetBlock, same shape as input
self.res_ch_dim = input_dim
#
self.skip_ch_dim = skip_ch_dim
self.gated_act_dim = gated_act_dim
self.cond_dim = cond_dim
self.dilation_size = dilation_size
self.conv_kernel_s = cnn_kernel_size
######
# layers
######
# dilated convolution
self.l_conv1d = nii_nn.Conv1dForARModel(
self.input_dim, self.gated_act_dim * 2, self.dilation_size,
self.conv_kernel_s, tanh=False)
# condition feature transform
self.l_cond_trans = torch_nn.Sequential(
torch_nn.Linear(self.cond_dim, self.gated_act_dim*2),
torch_nn.LeakyReLU())
# transformation after gated act
self.l_res_trans = torch_nn.Linear(self.gated_act_dim, self.res_ch_dim)
# transformation for skip channels
self.l_skip_trans = torch_nn.Linear(self.res_ch_dim, self.skip_ch_dim)
return
def _forward(self, input_feat, cond_feat, step_idx=None):
""" res_feat, skip_feat = forward(input_feat, cond_feat)
input
-----
input_feat: input feature tensor, (batchsize, length, input_dim)
cond_feat: condition feature tensor, (batchsize, length, cond_dim)
step_idx: None: tranining phase
int: idx of the time step during step-by-step generation
output
------
res_feat: residual channel feat tensor, (batchsize, length, input_dim)
skip_feat: skip channel feat tensor, , (batchsize, length, skip_dim)
"""
# dilated 1d convolution
hid = self.l_conv1d(input_feat, step_idx)
# transform and add condition feature
hid = hid + self.l_cond_trans(cond_feat)
# gated activation
hid = torch.tanh(hid[:, :, 0:self.gated_act_dim]) \
* torch.sigmoid(hid[:, :, self.gated_act_dim:])
# res-channel transform
res_feat = self.l_res_trans(hid) + input_feat
# skip-channel transform
skip_feat = self.l_skip_trans(res_feat)
# done
return res_feat, skip_feat
def forward(self, input_feat, cond_feat):
""" res_feat, skip_feat = forward(input_feat, cond_feat)
input
-----
input_feat: input feature tensor, (batchsize, length, input_dim)
cond_feat: condition feature tensor, (batchsize, length, cond_dim)
output
------
res_feat: residual channel feat tensor, (batchsize, length, input_dim)
skip_feat: skip channel feat tensor, , (batchsize, length, skip_dim)
Note that input_dim refers to the residual channel dimension.
Thus, input_feat should be embedding(audio), not audio.
"""
return self._forward(input_feat, cond_feat)
def inference(self, input_feat, cond_feat, step_idx):
""" res_feat, skip_feat = inference(input_feat, cond_feat, step_idx)
input
-----
input_feat: input feature tensor, (batchsize, length, input_dim)
cond_feat: condition feature tensor, (batchsize, length, cond_dim)
step_idx: int, idx of the time step during step-by-step generation
output
------
res_feat: residual channel feat tensor, (batchsize, length, input_dim)
skip_feat: skip channel feat tensor, , (batchsize, length, skip_dim)
"""
return self._forward(input_feat, cond_feat, step_idx)
class WaveNetBlock_v2(torch_nn.Module):
""" WaveNet block based on dilated-1D, gated-activation, and skip-connect.
Based on http://tonywangx.github.io/slide.html#misc CURRENNT WaveNet
(page 19-31) and WN in Pytorch WaveGlow.
The difference from WaveNetBlock
1. weight_norm
2. skip_channel is computed from gated-activation's output, not res_channel
"""
def __init__(self, input_dim, skip_ch_dim, gated_act_dim, cond_dim,
dilation_size, cnn_kernel_size=2, causal=True):
""" WaveNetBlock(input_dim, skip_ch_dim, gated_act_dim, cond_dim,
dilation_size, cnn_kernel_size = 2)
Args
----
input_dim: int, input tensor should be (batch-size, length, input_dim)
skip_ch_dim: int, tensors to be send to output blocks is in shape
(batch-size, length, skip_ch_dim)
gated_act_dim: int, tensors given by tanh(.) * sig(.) is in shape
(batch-size, length, gated_act_dim)
cond_dim: int, conditional feature (batchsize, length, cond_dim)
dilation_size: int, dilation size of the conv
cnn_kernel_size: int, kernel size of dilated conv1d (default, 2)
causal: bool, whether this block is used for AR model (default, True)
Note that when causal == False, step-by-step generation using step_index
will raise error.
"""
super(WaveNetBlock_v2, self).__init__()
#####
# configurations
#####
# input tensor: (batchsize, length, self.input_dim)
self.input_dim = input_dim
# tensor sent to next WaveNetBlock, same shape as input
self.res_ch_dim = input_dim
#
self.skip_ch_dim = skip_ch_dim
self.gated_act_dim = gated_act_dim
self.cond_dim = cond_dim
self.dilation_size = dilation_size
self.conv_kernel_s = cnn_kernel_size
######
# layers
######
# dilated convolution
tmp_layer = nii_nn.Conv1dForARModel(
self.input_dim, self.gated_act_dim * 2, self.dilation_size,
self.conv_kernel_s, tanh=False, causal = causal)
self.l_conv1d = torch.nn.utils.weight_norm(tmp_layer, name='weight')
# condition feature transform
tmp_layer = torch_nn.Linear(self.cond_dim, self.gated_act_dim*2)
self.l_cond_trans = torch.nn.utils.weight_norm(tmp_layer, name='weight')
# transformation after gated act
tmp_layer = torch_nn.Linear(self.gated_act_dim, self.res_ch_dim)
self.l_res_trans = torch.nn.utils.weight_norm(tmp_layer, name='weight')
# transformation for skip channels
#tmp_layer = torch_nn.Linear(self.res_ch_dim, self.skip_ch_dim)
tmp_layer = torch_nn.Linear(self.gated_act_dim, self.skip_ch_dim)
self.l_skip_trans = torch.nn.utils.weight_norm(tmp_layer, name='weight')
return
def _forward(self, input_feat, cond_feat, step_idx=None):
""" res_feat, skip_feat = forward(input_feat, cond_feat)
input
-----
input_feat: input feature tensor, (batchsize, length, input_dim)
cond_feat: condition feature tensor, (batchsize, length, cond_dim)
step_idx: None: tranining phase
int: idx of the time step during step-by-step generation
output
------
res_feat: residual channel feat tensor, (batchsize, length, input_dim)
skip_feat: skip channel feat tensor, , (batchsize, length, skip_dim)
"""
# dilated 1d convolution, add condition feature
hid = self.l_conv1d(input_feat, step_idx) + self.l_cond_trans(cond_feat)
# gated activation
hid = torch.tanh(hid[:, :, 0:self.gated_act_dim]) \
* torch.sigmoid(hid[:, :, self.gated_act_dim:])
# res-channel transform
res_feat = self.l_res_trans(hid) + input_feat
# skip-channel transform
# if we use skip_feat = self.l_skip_trans(res_feat), this cause
# exploding output when using skip_feat to produce scale and bias
# of affine transformation (e.g., in WaveGlow)
skip_feat = self.l_skip_trans(hid)
# done
return res_feat, skip_feat
def forward(self, input_feat, cond_feat):
""" res_feat, skip_feat = forward(input_feat, cond_feat)
input
-----
input_feat: input feature tensor, (batchsize, length, input_dim)
cond_feat: condition feature tensor, (batchsize, length, cond_dim)
output
------
res_feat: residual channel feat tensor, (batchsize, length, input_dim)
skip_feat: skip channel feat tensor, , (batchsize, length, skip_dim)
"""
return self._forward(input_feat, cond_feat)
def inference(self, input_feat, cond_feat, step_idx):
""" res_feat, skip_feat = inference(input_feat, cond_feat, step_idx)
input
-----
input_feat: input feature tensor, (batchsize, length, input_dim)
cond_feat: condition feature tensor, (batchsize, length, cond_dim)
step_idx: int, idx of the time step during step-by-step generation
output
------
res_feat: residual channel feat tensor, (batchsize, length, input_dim)
skip_feat: skip channel feat tensor, , (batchsize, length, skip_dim)
"""
return self._forward(input_feat, cond_feat, step_idx)
class OutputBlock(torch_nn.Module):
"""Output block to produce waveform distribution given skip-channel features
"""
def __init__(self, input_dim, output_dim, hid_dim=512):
""" OutputBlock(input_dim, output_dim)
Args
----
input_dim: int, input tensor should be (batchsize, length, input_dim)
it should be the sum of skip-channel features
output_dim: int, output tensor will be (batchsize, length, output_dim)
hid_dim: int, dimension of intermediate linear layers
"""
super(OutputBlock, self).__init__()
# config
self.input_dim = input_dim
self.output_dim = output_dim
self.hid_dim = hid_dim
# transformation layers before softmax
self.l_trans = torch_nn.Sequential(
torch_nn.Linear(self.input_dim, self.hid_dim // 2),
torch_nn.LeakyReLU(),
torch_nn.Linear(self.hid_dim // 2, self.hid_dim),
torch_nn.LeakyReLU(),
torch_nn.Linear(self.hid_dim, self.output_dim))
# output distribution
self.l_dist = nii_dist.DistCategorical(self.output_dim)
return
def forward(self, input_feat, target):
"""loss = forward(input_feat, target)
This method is supposed to be used to compute the loss
input
-----
input_feat: tensor in shape (batchsize, length, input_dim)
target: waveform tensor in shape (batchsize, length, dim=1)
output
------
loss: tensor or scalar
"""
# transform hidden feature vector to logit
tmp_logit = self.l_trans(input_feat)
# calculate the likelihood
return self.l_dist(tmp_logit, target)
def inference(self, input_feat):
"""output = inference(input_feat)
input
-----
input_feat: tensor in shape (batchsize, length, input_dim)
output
------
target: waveform tensor in shape (batchsize, length, dim=1)
"""
# transform hidden feature vector to logit
tmp_logit = self.l_trans(input_feat)
return self.l_dist.inference(tmp_logit)
################################
## Example of WaveNet definition
################################
class WaveNet_v1(torch_nn.Module):
""" Model definition of WaveNet
Example definition of WaveNet, version 1
"""
def __init__(self, in_dim, up_sample_rate, num_bits = 10, wnblock_ver=1,
pre_emphasis=True):
""" WaveNet(in_dim, up_sample_rate, num_bits=10, wnblock_ver=1,
pre_emphasis=False)
Args
----
in_dim: int, dimension of condition feature (batch, length, in_dim)
up_sample_rate, int, condition feature will be up-sampled by
using this rate
num_bits: int, number of bits for mu-law companding, default 10
wnblock_ver: int, version of the WaveNet Block, default 1
wnblock_ver = 1 uses WaveNetBlock
wnblock_ver = 2 uses WaveNetBlock_v2
pre_emphasis: bool, whether use pre-emphasis on the target waveform
up_sample_rate can be calculated using frame_shift of condition feature
and waveform sampling rate. For example, 16kHz waveform, condition
feature (e.g., Mel-spectrogram) extracted using 5ms frame shift, then
up_sample_rate = 16000 * 0.005 = 80. In other words, every frame will
be replicated 80 times.
"""
super(WaveNet_v1, self).__init__()
#################
## model config
#################
# number of bits for mu-law
self.num_bits = num_bits
self.num_classes = 2 ** self.num_bits
# up-sample rate
self.up_sample = up_sample_rate
# wavenet blocks
# residual channel dim
self.res_ch_dim = 64
# gated activate dim
self.gate_act_dim = 64
# condition feature dim
self.cond_dim = 64
# skip channel dim
self.skip_ch_dim = 256
# dilation size
self.dilations = [2 ** (x % 10) for x in range(30)]
# input dimension of (conditional feature)
self.input_dim = in_dim
# version of wavenet block
self.wnblock_ver = wnblock_ver
# whether pre-emphasis
self.pre_emphasis = pre_emphasis
###############
## network definition
###############
# condition module
self.l_cond = CondModule(self.input_dim, self.cond_dim, self.up_sample)
# waveform embedding layer
self.l_wav_emb = torch_nn.Embedding(self.num_classes, self.res_ch_dim)
# dilated convolution layers
tmp_wav_blocks = []
for dilation in self.dilations:
if self.wnblock_ver == 2:
tmp_wav_blocks.append(
WaveNetBlock_v2(
self.res_ch_dim, self.skip_ch_dim, self.gate_act_dim,
self.cond_dim, dilation))
else:
tmp_wav_blocks.append(
WaveNetBlock(
self.res_ch_dim, self.skip_ch_dim, self.gate_act_dim,
self.cond_dim, dilation))
self.l_wavenet_blocks = torch_nn.ModuleList(tmp_wav_blocks)
# output block
self.l_output = OutputBlock(self.skip_ch_dim, self.num_classes)
# done
return
def _waveform_encode_target(self, target_wav):
return nii_dsp.mulaw_encode(target_wav, self.num_classes)
def _waveform_decode_target(self, gen_wav):
return nii_dsp.mulaw_decode(gen_wav, self.num_classes)
def forward(self, input_feat, wav):
"""loss = forward(self, input_feat, wav)
input
-----
input_feat: tensor, input features (batchsize, length1, input_dim)
wav: tensor, target waveform (batchsize, length2, 1)
it should be raw waveform, flot valued, between (-1, 1)
it will be companded using mu-law automatically
output
------
loss: tensor / scalar
Note: returned loss can be directly used as the loss value
no need to write Loss()
"""
# step1. prepare the target waveform and feedback waveform
# do mu-law companding
# shifting by 1 time step for feedback waveform
with torch.no_grad():
if self.pre_emphasis:
wav[:, 1:, :] = wav[:, 1:, :] - 0.97 * wav[:, 0:-1, :]
wav = wav.clamp(-1, 1)
# mu-law companding (int values)
# note that _waveform_encoder_target will produce int values
target_wav = self._waveform_encode_target(wav)
# feedback wav
fb_wav = torch.zeros(
target_wav.shape, device=wav.device, dtype=target_wav.dtype)
fb_wav[:, 1:] = target_wav[:, :-1]
# step2. condition feature
hid_cond = self.l_cond(input_feat)
# step3. feedback waveform embedding
hid_wav_emb = self.l_wav_emb(fb_wav.squeeze(-1))
# step4. stacks of wavenet
# buffer to save skip-channel features
skip_ch_feat = torch.zeros(
[target_wav.shape[0],target_wav.shape[1], self.skip_ch_dim],
device=input_feat.device, dtype=input_feat.dtype)
res_ch_feat = hid_wav_emb
for l_wavblock in self.l_wavenet_blocks:
res_ch_feat, tmp_skip_ch_feat = l_wavblock(res_ch_feat, hid_cond)
skip_ch_feat += tmp_skip_ch_feat
# step5. get output
likelihood = self.l_output(skip_ch_feat, target_wav)
return likelihood
def inference(self, input_feat):
"""wav = inference(mels)
input
-----
input_feat: tensor, input features (batchsize, length1, input_dim)
output
------
wav: tensor, target waveform (batchsize, length2, 1)
Note: length2 will be = length1 * self.up_sample
"""
# prepare
batchsize = input_feat.shape[0]
wavlength = input_feat.shape[1] * self.up_sample
time_idx_marker = wavlength // 10
#
# step1. condition features
hid_cond = self.l_cond(input_feat)
# step2. do computation step-by-step
# initialzie the buffer
gen_wav_buf = torch.zeros(
[batchsize, wavlength, 1],
dtype=input_feat.dtype, device=input_feat.device)
fb_wav_buf = torch.zeros(
[batchsize, 1, 1],
dtype=input_feat.dtype, device=input_feat.device)
skip_ch_feat = torch.zeros(
[batchsize, 1, self.skip_ch_dim],
dtype=input_feat.dtype, device=input_feat.device)
# loop over all time steps
print("Total time steps: {:d}. Progress: ".format(wavlength),
end=' ', flush=True)
for time_idx in range(wavlength):
# show messages
if time_idx % 500 == 1:
print(time_idx, end=' ', flush=True)
# feedback
if time_idx > 0:
fb_wav_buf = gen_wav_buf[:, time_idx-1:time_idx, :]
# initialize skip
skip_ch_feat *= 0
# embedding
hid_wav_emb = self.l_wav_emb(fb_wav_buf.squeeze(-1).to(torch.int64))
# condition feature for current time step
# for other time steps, intermediate feat is saved by wave blocks
hid_cond_tmp = hid_cond[:, time_idx:time_idx+1, :]
# loop over wavblocks
res_ch_feat = hid_wav_emb
for l_wavblock in self.l_wavenet_blocks:
res_ch_feat, tmp_skip_ch_feat = l_wavblock.inference(
res_ch_feat, hid_cond_tmp, time_idx)
skip_ch_feat += tmp_skip_ch_feat
# draw sample
drawn_sample = self.l_output.inference(skip_ch_feat)
gen_wav_buf[:, time_idx:time_idx+1, :] = drawn_sample
# decode mu-law
wave = self._waveform_decode_target(gen_wav_buf)
# de-emphasis if necessary
if self.pre_emphasis:
for idx in range(wave.shape[1] - 1):
wave[:, idx+1, :] = wave[:, idx+1, :] + 0.97 * wave[:, idx, :]
return wave
if __name__ == "__main__":
print("Definition of model")
| 23,685 | 34.887879 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/eval_asvspoof.py | #!/usr/bin/env python
"""
Functions for evaluation - asvspoof and related binary classification tasks
Python Function from min tDCF on asvspoof.org
All functions before tDCF_wrapper are licensed by Creative Commons.
----- License ----
This work is licensed under the Creative Commons
Attribution-NonCommercial-ShareAlike 4.0 International
License. To view a copy of this license, visit
http://creativecommons.org/licenses/by-nc-sa/4.0/
or send a letter to
Creative Commons, 444 Castro Street, Suite 900,
Mountain View, California, 94041, USA.
------------------
"""
from __future__ import print_function
import os
import sys
import numpy as np
import core_scripts.data_io.io_tools as nii_io
class CustomDict:
def __init__(self, missing_value=-1):
self.databuf = {}
self.misval = missing_value
return
def __setitem__(self, key, value):
self.databuf[key] = value
return
def __getitem__(self, key):
if key in self.databuf:
return self.databuf[key]
else:
return self.misval
def keys(self):
return self.databuf.keys()
def protocol_parse_asvspoof2019(protocol_filepath):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = CustomDict()
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
temp_buffer = np.loadtxt(protocol_filepath, dtype='str')
for row in temp_buffer:
if row[-1] == 'bonafide':
data_buffer[row[1]] = 1
else:
data_buffer[row[1]] = 0
else:
print("Cannot load {:s}".format(protocol_filepath))
print("Use an empty dictionary")
return data_buffer
def protocol_parse_general(protocol_filepaths, sep=' '):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = CustomDict()
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if cols[-1] == 'bonafide':
data_buffer[cols[1]] = 1
else:
data_buffer[cols[1]] = 0
return data_buffer
def protocol_parse_attack_label_asvspoof2019(protocol_filepath):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
output:
-------
data_buffer: dic, data_bufer[filename] -> attack type
"""
data_buffer = {}
temp_buffer = np.loadtxt(protocol_filepath, dtype='str')
for row in temp_buffer:
if row[-2] == '-':
data_buffer[row[1]] = 'bonafide'
else:
data_buffer[row[1]] = row[-2]
return data_buffer
def obtain_asv_error_rates(tar_asv, non_asv, spoof_asv, asv_threshold):
# False alarm and miss rates for ASV
Pfa_asv = sum(non_asv >= asv_threshold) / non_asv.size
Pmiss_asv = sum(tar_asv < asv_threshold) / tar_asv.size
# Rate of rejecting spoofs in ASV
if spoof_asv.size == 0:
Pmiss_spoof_asv = None
Pfa_spoof_asv = None
else:
Pmiss_spoof_asv = np.sum(spoof_asv < asv_threshold) / spoof_asv.size
Pfa_spoof_asv = np.sum(spoof_asv >= asv_threshold) / spoof_asv.size
return Pfa_asv, Pmiss_asv, Pmiss_spoof_asv, Pfa_spoof_asv
def compute_det_curve(target_scores, nontarget_scores):
n_scores = target_scores.size + nontarget_scores.size
all_scores = np.concatenate((target_scores, nontarget_scores))
labels = np.concatenate((np.ones(target_scores.size),
np.zeros(nontarget_scores.size)))
# Sort labels based on scores
indices = np.argsort(all_scores, kind='mergesort')
labels = labels[indices]
# Compute false rejection and false acceptance rates
tar_trial_sums = np.cumsum(labels)
nontarget_trial_sums = (nontarget_scores.size -
(np.arange(1, n_scores + 1) - tar_trial_sums))
frr = np.concatenate((np.atleast_1d(0), tar_trial_sums/target_scores.size))
# false rejection rates
far = np.concatenate((np.atleast_1d(1),
nontarget_trial_sums / nontarget_scores.size))
# false acceptance rates
thresholds = np.concatenate((np.atleast_1d(all_scores[indices[0]] - 0.001),
all_scores[indices]))
# Thresholds are the sorted scores
return frr, far, thresholds
def compute_eer(target_scores, nontarget_scores):
""" Returns equal error rate (EER) and the corresponding threshold. """
frr, far, thresholds = compute_det_curve(target_scores, nontarget_scores)
abs_diffs = np.abs(frr - far)
min_index = np.argmin(abs_diffs)
eer = np.mean((frr[min_index], far[min_index]))
return eer, thresholds[min_index]
def compute_tDCF_legacy(
bonafide_score_cm, spoof_score_cm,
Pfa_asv, Pmiss_asv, Pmiss_spoof_asv, cost_model, print_cost=False):
"""
Compute Tandem Detection Cost Function (t-DCF) [1] for a fixed ASV system.
In brief, t-DCF returns a detection cost of a cascaded system of this form,
Speech waveform -> [CM] -> [ASV] -> decision
where CM stands for countermeasure and ASV for automatic speaker
verification. The CM is therefore used as a 'gate' to decided whether or
not the input speech sample should be passed onwards to the ASV system.
Generally, both CM and ASV can do detection errors. Not all those errors
are necessarily equally cost, and not all types of users are necessarily
equally likely. The tandem t-DCF gives a principled with to compare
different spoofing countermeasures under a detection cost function
framework that takes that information into account.
INPUTS:
bonafide_score_cm A vector of POSITIVE CLASS (bona fide or human)
detection scores obtained by executing a spoofing
countermeasure (CM) on some positive evaluation trials
trial represents a bona fide case.
spoof_score_cm A vector of NEGATIVE CLASS (spoofing attack)
detection scores obtained by executing a spoofing
CM on some negative evaluation trials.
Pfa_asv False alarm (false acceptance) rate of the ASV
system that is evaluated in tandem with the CM.
Assumed to be in fractions, not percentages.
Pmiss_asv Miss (false rejection) rate of the ASV system that
is evaluated in tandem with the spoofing CM.
Assumed to be in fractions, not percentages.
Pmiss_spoof_asv Miss rate of spoof samples of the ASV system that
is evaluated in tandem with the spoofing CM. That
is, the fraction of spoof samples that were
rejected by the ASV system.
cost_model A struct that contains the parameters of t-DCF,
with the following fields.
Ptar Prior probability of target speaker.
Pnon Prior probability of nontarget speaker
(zero-effort impostor)
Psoof Prior probability of spoofing attack.
Cmiss_asv Cost of ASV falsely rejecting target.
Cfa_asv Cost of ASV falsely accepting nontarget.
Cmiss_cm Cost of CM falsely rejecting target.
Cfa_cm Cost of CM falsely accepting spoof.
print_cost Print a summary of the cost parameters and the
implied t-DCF cost function?
OUTPUTS:
tDCF_norm Normalized t-DCF curve across the different CM
system operating points; see [2] for more details.
Normalized t-DCF > 1 indicates a useless
countermeasure (as the tandem system would do
better without it). min(tDCF_norm) will be the
minimum t-DCF used in ASVspoof 2019 [2].
CM_thresholds Vector of same size as tDCF_norm corresponding to
the CM threshold (operating point).
NOTE:
o In relative terms, higher detection scores values are assumed to
indicate stronger support for the bona fide hypothesis.
o You should provide real-valued soft scores, NOT hard decisions. The
recommendation is that the scores are log-likelihood ratios (LLRs)
from a bonafide-vs-spoof hypothesis based on some statistical model.
This, however, is NOT required. The scores can have arbitrary range
and scaling.
o Pfa_asv, Pmiss_asv, Pmiss_spoof_asv are in fractions, not percentages.
References:
[1] T. Kinnunen, K.-A. Lee, H. Delgado, N. Evans, M. Todisco,
M. Sahidullah, J. Yamagishi, D.A. Reynolds: "t-DCF: a Detection
Cost Function for the Tandem Assessment of Spoofing Countermeasures
and Automatic Speaker Verification", Proc. Odyssey 2018: the
Speaker and Language Recognition Workshop, pp. 312--319,
Les Sables d'Olonne,
France, June 2018
https://www.isca-speech.org/archive/Odyssey_2018/pdfs/68.pdf)
[2] ASVspoof 2019 challenge evaluation plan
TODO: <add link>
"""
# Sanity check of cost parameters
if cost_model['Cfa_asv'] < 0 or cost_model['Cmiss_asv'] < 0 or \
cost_model['Cfa_cm'] < 0 or cost_model['Cmiss_cm'] < 0:
print('WARNING: Usually the cost values should be positive!')
if cost_model['Ptar'] < 0 or cost_model['Pnon'] < 0 or \
cost_model['Pspoof'] < 0 or \
np.abs(cost_model['Ptar'] + cost_model['Pnon'] + cost_model['Pspoof'] - 1) > 1e-10:
sys.exit('ERROR: Your prior probabilities should be positive and sum up to one.')
# Unless we evaluate worst-case model, we need to have some spoof tests against asv
if Pmiss_spoof_asv is None:
sys.exit('ERROR: you should provide miss rate of spoof tests against your ASV system.')
# Sanity check of scores
combined_scores = np.concatenate((bonafide_score_cm, spoof_score_cm))
if np.isnan(combined_scores).any() or np.isinf(combined_scores).any():
sys.exit('ERROR: Your scores contain nan or inf.')
# Sanity check that inputs are scores and not decisions
n_uniq = np.unique(combined_scores).size
if n_uniq < 3:
sys.exit('ERROR: You should provide soft CM scores - not binary decisions')
# Obtain miss and false alarm rates of CM
Pmiss_cm, Pfa_cm, CM_thresholds = compute_det_curve(bonafide_score_cm, spoof_score_cm)
# Constants - see ASVspoof 2019 evaluation plan
C1 = cost_model['Ptar'] * (cost_model['Cmiss_cm'] - cost_model['Cmiss_asv'] * Pmiss_asv) - \
cost_model['Pnon'] * cost_model['Cfa_asv'] * Pfa_asv
C2 = cost_model['Cfa_cm'] * cost_model['Pspoof'] * (1 - Pmiss_spoof_asv)
# Sanity check of the weights
if C1 < 0 or C2 < 0:
sys.exit('You should never see this error but I cannot evalute tDCF with negative weights - please check whether your ASV error rates are correctly computed?')
# Obtain t-DCF curve for all thresholds
tDCF = C1 * Pmiss_cm + C2 * Pfa_cm
# Normalized t-DCF
tDCF_norm = tDCF / np.minimum(C1, C2)
# Everything should be fine if reaching here.
if print_cost:
print('t-DCF evaluation from [Nbona={}, Nspoof={}] trials\n'.format(bonafide_score_cm.size, spoof_score_cm.size))
print('t-DCF MODEL')
print(' Ptar = {:8.5f} (Prior probability of target user)'.format(cost_model['Ptar']))
print(' Pnon = {:8.5f} (Prior probability of nontarget user)'.format(cost_model['Pnon']))
print(' Pspoof = {:8.5f} (Prior probability of spoofing attack)'.format(cost_model['Pspoof']))
print(' Cfa_asv = {:8.5f} (Cost of ASV falsely accepting a nontarget)'.format(cost_model['Cfa_asv']))
print(' Cmiss_asv = {:8.5f} (Cost of ASV falsely rejecting target speaker)'.format(cost_model['Cmiss_asv']))
print(' Cfa_cm = {:8.5f} (Cost of CM falsely passing a spoof to ASV system)'.format(cost_model['Cfa_cm']))
print(' Cmiss_cm = {:8.5f} (Cost of CM falsely blocking target utterance which never reaches ASV)'.format(cost_model['Cmiss_cm']))
print('\n Implied normalized t-DCF function (depends on t-DCF parameters and ASV errors), s=CM threshold)')
if C2 == np.minimum(C1, C2):
print(' tDCF_norm(s) = {:8.5f} x Pmiss_cm(s) + Pfa_cm(s)\n'.format(C1 / C2))
else:
print(' tDCF_norm(s) = Pmiss_cm(s) + {:8.5f} x Pfa_cm(s)\n'.format(C2 / C1))
return tDCF_norm, CM_thresholds
def compute_tDCF(
bonafide_score_cm, spoof_score_cm,
Pfa_asv, Pmiss_asv, Pfa_spoof_asv, cost_model, print_cost):
"""
Compute Tandem Detection Cost Function (t-DCF) [1] for a fixed ASV system.
In brief, t-DCF returns a detection cost of a cascaded system of this form,
Speech waveform -> [CM] -> [ASV] -> decision
where CM stands for countermeasure and ASV for automatic speaker
verification. The CM is therefore used as a 'gate' to decided whether or
not the input speech sample should be passed onwards to the ASV system.
Generally, both CM and ASV can do detection errors. Not all those errors
are necessarily equally cost, and not all types of users are necessarily
equally likely. The tandem t-DCF gives a principled with to compare
different spoofing countermeasures under a detection cost function
framework that takes that information into account.
INPUTS:
bonafide_score_cm A vector of POSITIVE CLASS (bona fide or human)
detection scores obtained by executing a spoofing
countermeasure (CM) on some positive evaluation trials.
trial represents a bona fide case.
spoof_score_cm A vector of NEGATIVE CLASS (spoofing attack)
detection scores obtained by executing a spoofing
CM on some negative evaluation trials.
Pfa_asv False alarm (false acceptance) rate of the ASV
system that is evaluated in tandem with the CM.
Assumed to be in fractions, not percentages.
Pmiss_asv Miss (false rejection) rate of the ASV system that
is evaluated in tandem with the spoofing CM.
Assumed to be in fractions, not percentages.
Pmiss_spoof_asv Miss rate of spoof samples of the ASV system that
is evaluated in tandem with the spoofing CM. That
is, the fraction of spoof samples that were
rejected by the ASV system.
cost_model A struct that contains the parameters of t-DCF,
with the following fields.
Ptar Prior probability of target speaker.
Pnon Prior probability of nontarget speaker (zero-effort impostor)
Psoof Prior probability of spoofing attack.
Cmiss Cost of tandem system falsely rejecting target speaker.
Cfa Cost of tandem system falsely accepting nontarget speaker.
Cfa_spoof Cost of tandem system falsely accepting spoof.
print_cost Print a summary of the cost parameters and the
implied t-DCF cost function?
OUTPUTS:
tDCF_norm Normalized t-DCF curve across the different CM
system operating points; see [2] for more details.
Normalized t-DCF > 1 indicates a useless
countermeasure (as the tandem system would do
better without it). min(tDCF_norm) will be the
minimum t-DCF used in ASVspoof 2019 [2].
CM_thresholds Vector of same size as tDCF_norm corresponding to
the CM threshold (operating point).
NOTE:
o In relative terms, higher detection scores values are assumed to
indicate stronger support for the bona fide hypothesis.
o You should provide real-valued soft scores, NOT hard decisions. The
recommendation is that the scores are log-likelihood ratios (LLRs)
from a bonafide-vs-spoof hypothesis based on some statistical model.
This, however, is NOT required. The scores can have arbitrary range
and scaling.
o Pfa_asv, Pmiss_asv, Pmiss_spoof_asv are in fractions, not percentages.
References:
[1] T. Kinnunen, H. Delgado, N. Evans,K.-A. Lee, V. Vestman,
A. Nautsch, M. Todisco, X. Wang, M. Sahidullah, J. Yamagishi,
and D.-A. Reynolds, "Tandem Assessment of Spoofing Countermeasures
and Automatic Speaker Verification: Fundamentals," IEEE/ACM Transaction on
Audio, Speech and Language Processing (TASLP).
[2] ASVspoof 2019 challenge evaluation plan
https://www.asvspoof.org/asvspoof2019/asvspoof2019_evaluation_plan.pdf
"""
# Sanity check of cost parameters
if cost_model['Cfa'] < 0 or cost_model['Cmiss'] < 0 or \
cost_model['Cfa'] < 0 or cost_model['Cmiss'] < 0:
print('WARNING: Usually the cost values should be positive!')
if cost_model['Ptar'] < 0 or cost_model['Pnon'] < 0 or cost_model['Pspoof'] < 0 or \
np.abs(cost_model['Ptar'] + cost_model['Pnon'] + cost_model['Pspoof'] - 1) > 1e-10:
sys.exit('ERROR: Your prior probabilities should be positive and sum up to one.')
# Unless we evaluate worst-case model, we need to have some spoof tests against asv
if Pfa_spoof_asv is None:
sys.exit('ERROR: you should provide false alarm rate of spoof tests against your ASV system.')
# Sanity check of scores
combined_scores = np.concatenate((bonafide_score_cm, spoof_score_cm))
if np.isnan(combined_scores).any() or np.isinf(combined_scores).any():
sys.exit('ERROR: Your scores contain nan or inf.')
# Sanity check that inputs are scores and not decisions
n_uniq = np.unique(combined_scores).size
if n_uniq < 3:
sys.exit('ERROR: You should provide soft CM scores - not binary decisions')
# Obtain miss and false alarm rates of CM
Pmiss_cm, Pfa_cm, CM_thresholds = compute_det_curve(bonafide_score_cm, spoof_score_cm)
# Constants - see ASVspoof 2019 evaluation plan
C0 = cost_model['Ptar'] * cost_model['Cmiss'] * Pmiss_asv + cost_model['Pnon']*cost_model['Cfa']*Pfa_asv
C1 = cost_model['Ptar'] * cost_model['Cmiss'] - (cost_model['Ptar'] * cost_model['Cmiss'] * Pmiss_asv + cost_model['Pnon'] * cost_model['Cfa'] * Pfa_asv)
C2 = cost_model['Pspoof'] * cost_model['Cfa_spoof'] * Pfa_spoof_asv;
# Sanity check of the weights
if C0 < 0 or C1 < 0 or C2 < 0:
sys.exit('You should never see this error but I cannot evalute tDCF with negative weights - please check whether your ASV error rates are correctly computed?')
# Obtain t-DCF curve for all thresholds
tDCF = C0 + C1 * Pmiss_cm + C2 * Pfa_cm
# Obtain default t-DCF
tDCF_default = C0 + np.minimum(C1, C2)
# Normalized t-DCF
tDCF_norm = tDCF / tDCF_default
# Everything should be fine if reaching here.
if print_cost:
print('t-DCF evaluation from [Nbona={}, Nspoof={}] trials\n'.format(bonafide_score_cm.size, spoof_score_cm.size))
print('t-DCF MODEL')
print(' Ptar = {:8.5f} (Prior probability of target user)'.format(cost_model['Ptar']))
print(' Pnon = {:8.5f} (Prior probability of nontarget user)'.format(cost_model['Pnon']))
print(' Pspoof = {:8.5f} (Prior probability of spoofing attack)'.format(cost_model['Pspoof']))
print(' Cfa = {:8.5f} (Cost of tandem system falsely accepting a nontarget)'.format(cost_model['Cfa']))
print(' Cmiss = {:8.5f} (Cost of tandem system falsely rejecting target speaker)'.format(cost_model['Cmiss']))
print(' Cfa_spoof = {:8.5f} (Cost of tandem sysmte falsely accepting spoof)'.format(cost_model['Cfa_spoof']))
print('\n Implied normalized t-DCF function (depends on t-DCF parameters and ASV errors), t_CM=CM threshold)')
print(' tDCF_norm(t_CM) = {:8.5f} + {:8.5f} x Pmiss_cm(t_CM) + {:8.5f} x Pfa_cm(t_CM)\n'.format(C0/tDCF_default, C1/tDCF_default, C2/tDCF_default))
print(' * The optimum value is given by the first term (0.06273). This is the normalized t-DCF obtained with an error-free CM system.')
print(' * The minimum normalized cost (minimum over all possible thresholds) is always <= 1.00.')
print('')
return tDCF_norm, CM_thresholds
def tDCF_wrapper(bonafide_cm_scores, spoof_cm_scores,
tar_asv_scores=None, non_asv_scores=None,
spoof_asv_scores=None,
flag_verbose=False, flag_legacy=True):
"""
mintDCF, eer, eer_thre = tDCF_wrapper(bonafide_cm_scores, spoof_cm_scores,
tar_asv_scores=None, non_asv_scores=None,
spoof_asv_scores=None, flag_verbose=False, flag_legacy=True)
input
-----
bonafide_cm_scores: np.array of bona fide scores
spoof_cm_scores: np.array of spoof scores
tar_asv_scores: np.array of ASV target scores, or None
non_asv_scores: np.array of ASV non-target scores, or None
spoof_asv_scores: np.array of ASV spoof trial scores, or None,
flag_verbose: print detailed messages
flag_legacy: True: use legacy min-tDCF in ASVspoof2019
False: use min-tDCF revised
output
------
mintDCF: scalar, value of min-tDCF
eer: scalar, value of EER
eer_thre: scalar, value of threshold corresponding to EER
"""
if flag_legacy:
Pspoof = 0.05
cost_model = {
'Pspoof': Pspoof, # Prior probability of a spoofing attack
'Ptar': (1 - Pspoof) * 0.99, # Prior probability of target speaker
'Pnon': (1 - Pspoof) * 0.01, # Prior probability of nontarget speaker
'Cmiss_asv': 1, # Cost of ASV system falsely rejecting target speaker
'Cfa_asv': 10, # Cost of ASV system falsely accepting nontarget speaker
'Cmiss_cm': 1, # Cost of CM system falsely rejecting target speaker
'Cfa_cm': 10, # Cost of CM system falsely accepting spoof
}
else:
Pspoof = 0.05
cost_model = {
'Pspoof': Pspoof, # Prior probability of a spoofing attack
'Ptar': (1 - Pspoof) * 0.99, # Prior probability of target speaker
'Pnon': (1 - Pspoof) * 0.01, # Prior probability of nontarget speaker
'Cmiss': 1, # Cost of tandem system falsely rejecting target speaker
'Cfa': 10, # Cost of tandem system falsely accepting nontarget speaker
'Cfa_spoof': 10, # Cost of tandem system falsely accepting spoof
}
# read provided ASV scores
if tar_asv_scores is None or non_asv_scores is None or \
spoof_asv_scores is None:
file_name = os.path.dirname(__file__)+ \
'/data/asvspoof2019/ASVspoof2019.LA.asv.eval.gi.trl.scores.bin'
data = nii_io.f_read_raw_mat(file_name, 2)
tar_asv_scores = data[data[:, 1] == 2, 0]
non_asv_scores = data[data[:, 1] == 1, 0]
spoof_asv_scores = data[data[:, 1] == 0, 0]
eer_asv, asv_threshold = compute_eer(tar_asv_scores, non_asv_scores)
eer_cm, eer_threshold = compute_eer(bonafide_cm_scores, spoof_cm_scores)
[Pfa_asv,Pmiss_asv,Pmiss_spoof_asv,Pfa_spoof_asv] = obtain_asv_error_rates(
tar_asv_scores, non_asv_scores, spoof_asv_scores, asv_threshold)
if flag_legacy:
tDCF_curve, CM_thresholds = compute_tDCF_legacy(
bonafide_cm_scores, spoof_cm_scores,
Pfa_asv, Pmiss_asv, Pmiss_spoof_asv, cost_model, flag_verbose)
else:
tDCF_curve, CM_thresholds = compute_tDCF(
bonafide_cm_scores, spoof_cm_scores,
Pfa_asv, Pmiss_asv, Pfa_spoof_asv, cost_model, flag_verbose)
min_tDCF_index = np.argmin(tDCF_curve)
min_tDCF = tDCF_curve[min_tDCF_index]
return min_tDCF, eer_cm, eer_threshold
def tDCF_wrapper2(bonafide_score_cm, spoof_score_cm, C0, C1, C2):
""" mintDCF, eer = tDCF_wrapper2(bonafide_score_cm,
spoof_score_cm, C0, C1, C2)
compute_tDCF can be factorized into two parts:
C012 computation and min t-DCF computation.
This is for min t-DCF computation, given the values of C012
input
-----
bonafide_score_cm np.array, score of bonafide data
spoof_score_cm np.array, score of spoofed data
C0 scalar, coefficient for min tDCF computation
C1 scalar, coefficient for min tDCF computation
C2 scalar, coefficient for min tDCF computation
output
------
eer scalar, value of EER
mintDCF scalar, value of min tDCF
For C0, C1, C2, see Appendix Eqs.(1-2) in evaluation plan [1],
or Eqs.(10-11) in [2]
References:
[1] T. Kinnunen, H. Delgado, N. Evans,K.-A. Lee, V. Vestman,
A. Nautsch, M. Todisco, X. Wang, M. Sahidullah, J. Yamagishi,
and D.-A. Reynolds, "Tandem Assessment of Spoofing Countermeasures
and Automatic Speaker Verification: Fundamentals," IEEE/ACM Transaction on
Audio, Speech and Language Processing (TASLP).
[2] ASVspoof 2019 challenge evaluation plan
https://www.asvspoof.org/asvspoof2019/asvspoof2019_evaluation_plan.pdf
"""
# Sanity check of scores
combined_scores = np.concatenate((bonafide_score_cm, spoof_score_cm))
if np.isnan(combined_scores).any() or np.isinf(combined_scores).any():
sys.exit('ERROR: Your scores contain nan or inf.')
# Sanity check that inputs are scores and not decisions
n_uniq = np.unique(combined_scores).size
if n_uniq < 3:
sys.exit('ERROR: You should provide soft CM scores - not binary decisions')
# Obtain miss and false alarm rates of CM
Pmiss_cm, Pfa_cm, CM_thresholds = compute_det_curve(
bonafide_score_cm, spoof_score_cm)
# =====
# tDCF
# =====
if np.isnan(C0) or np.isnan(C1) or np.isnan(C2):
# this is a case where
mintDCF = np.nan
else:
# tDCF values
tDCF = C0 + C1 * Pmiss_cm + C2 * Pfa_cm
# Obtain default t-DCF
tDCF_default = C0 + np.minimum(C1, C2)
# Normalized t-DCF
tDCF_norm = tDCF / tDCF_default
# min t-DCF
mintDCF = tDCF_norm[tDCF_norm.argmin()]
# ====
# EER
# ====
abs_diffs = np.abs(Pmiss_cm - Pfa_cm)
min_index = np.argmin(abs_diffs)
eer = np.mean((Pmiss_cm[min_index], Pfa_cm[min_index]))
return mintDCF, eer
def ASVspoof2019_evaluate(bonafide_cm_scores, bonafide_cm_file_names,
spoof_cm_scores, spoof_cm_file_names, verbose=False,
protocol_alternative=None):
""" Decompose scores for each attack. For ASVspoof2019
ASVspoof2019_decompose(bonafide_cm_scores, bonafide_cm_file_names,
spoof_cm_scores, spoof_cm_file_names, verbose=False)
input
-----
bonafide_cm_scores: np.array of bonafide scores
bonafide_cm_file_names: file name list corresponding to bonafide_cm_scores
spoof_cm_scores: np.array of spoofed scores (all attack types)
spoof_cm_file_names: file name list corresponding to spoof_cm_scores
verbose: print information from tDCF computation (default: False)
protocol_alternative: alternative protocol to ASVspoof2019 (default: None)
output
------
min_tDCF: np.array of min tDCF for each attack
eer_cm: np.array of EER for each attack
eer_threshold: np.array of threshold for EER (not min tDCF threshod)
spoof_attack_types: list of attack types
"""
if protocol_alternative is not None:
# if provided alternative procotol, use it.
# this is for protocol tweaking
file_name = protocol_alternative
else:
# official protocol
file_name = os.path.dirname(__file__)+ '/data/asvspoof2019/protocol.txt'
protocol_data = np.genfromtxt(file_name,
dtype=[('spk', 'U10'), ('file', 'U20'),
('misc', 'U5'), ('spoof', 'U5'),
('type','U10')], delimiter=" ")
spoof_type_dic = {protocol_data[x][1]:protocol_data[x][3] for x in \
range(protocol_data.shape[0])}
spoof_attack_types = list(set([x[3] for x in protocol_data]))
spoof_attack_types.sort()
# default set to -1
min_tDCF = np.zeros([len(spoof_attack_types) + 1]) - 1
eer_cm = np.zeros([len(spoof_attack_types) + 1]) - 1
eer_threshold = np.zeros([len(spoof_attack_types) + 1])
# decompose results
decomposed_spoof_scores = []
for idx, spoof_attack_type in enumerate(spoof_attack_types):
tmp_spoof_scores = [spoof_cm_scores[x] for x, y in \
enumerate(spoof_cm_file_names) \
if spoof_type_dic[y] == spoof_attack_type]
tmp_spoof_scores = np.array(tmp_spoof_scores)
decomposed_spoof_scores.append(tmp_spoof_scores.copy())
if len(tmp_spoof_scores):
x1, x2, x3 = tDCF_wrapper(bonafide_cm_scores, tmp_spoof_scores)
min_tDCF[idx] = x1
eer_cm[idx] = x2
eer_threshold[idx] = x3
# pooled results
x1, x2, x3 = tDCF_wrapper(bonafide_cm_scores, spoof_cm_scores)
min_tDCF[-1] = x1
eer_cm[-1] = x2
eer_threshold[-1] = x3
spoof_attack_types.append("pooled")
decomposed_spoof_scores.append(spoof_cm_scores)
for idx in range(len(spoof_attack_types)):
if verbose and eer_cm[idx] > -1:
print("{:s}\tmin-tDCF: {:2.5f}\tEER: {:2.3f}%\t Thre:{:f}".format(
spoof_attack_types[idx], min_tDCF[idx], eer_cm[idx] * 100,
eer_threshold[idx]))
decomposed_spoof_scores = [decomposed_spoof_scores[x] \
for x, y in enumerate(min_tDCF) if y > -1]
spoof_attack_types = [spoof_attack_types[x] \
for x, y in enumerate(min_tDCF) if y > -1]
eer_threshold = [eer_threshold[x] \
for x, y in enumerate(min_tDCF) if y > -1]
eer_cm = [eer_cm[x] for x, y in enumerate(min_tDCF) if y > -1]
min_tDCF = [y for x, y in enumerate(min_tDCF) if y > -1]
return min_tDCF, eer_cm, eer_threshold, spoof_attack_types, \
decomposed_spoof_scores
##############
# for Pytorch models in this project
##############
def parse_pytorch_output_txt(score_file_path):
""" parse_pytorch_output_txt(file_path)
parse the score files generated by the pytorch models
input
-----
file_path: path to the log file
output
------
bonafide: np.array, bonafide scores
bonafide_names: list of file names corresponding to bonafide scores
spoofed: np.array, spoofed scores
spoofed_names: list of file names corresponding to spoofed scores
"""
bonafide = []
spoofed = []
bonafide_names = []
spoofed_names = []
with open(score_file_path, 'r') as file_ptr:
for line in file_ptr:
if line.startswith('Output,'):
temp = line.split(',')
flag = int(temp[2])
if np.isnan(float(temp[3])):
print(line)
continue
if flag:
bonafide.append(float(temp[3]))
bonafide_names.append(temp[1].strip())
else:
spoofed.append(float(temp[3]))
spoofed_names.append(temp[1].strip())
bonafide = np.array(bonafide)
spoofed = np.array(spoofed)
return bonafide, bonafide_names, spoofed, spoofed_names
def ASVspoof2019_decomposed_results(score_file_path, flag_return_results=False,
flag_verbose=True):
""" Get the results from input score log file
ASVspoof2019_decomposed_results(score_file_path, flag_return_results=False,
flag_verbose=True)
input
-----
score_file_path: path to the score file produced by the Pytorch code
flag_return_results: whether return the results (default False)
flag_verbose: print EERs and mintDCFs for each attack (default True)
output
------
if flag_return_results is True:
mintDCFs: list of min tDCF, for each attack
eers: list of EER, for each attack
cm_thres: list of threshold for EER, for each attack
spoof_types: list of spoof attack types
spoof_scores: list of spoof file scores (np.array)
bona: bonafide score
"""
bona, b_names, spoofed, s_names = parse_pytorch_output_txt(score_file_path)
mintDCFs, eers, cm_thres, spoof_types, spoof_scores = ASVspoof2019_evaluate(
bona, b_names, spoofed, s_names, flag_verbose)
if flag_return_results:
return mintDCFs, eers, cm_thres, spoof_types, spoof_scores, bona
else:
return
def ASVspoofNNN_decomposed_results(score_file_path,
flag_return_results=False,
flag_verbose=True,
protocol_alternative=None):
""" Similar to ASVspoof2019_decomposed_results, but use alternative protocol
"""
bona, b_names, spoofed, s_names = parse_pytorch_output_txt(score_file_path)
mintDCFs, eers, cm_thres, spoof_types, spoof_scores = ASVspoof2019_evaluate(
bona, b_names, spoofed, s_names, flag_verbose, protocol_alternative)
if flag_return_results:
return mintDCFs, eers, cm_thres, spoof_types, spoof_scores, bona
else:
return
##############
# for testing using ./data/cm_dev.txt and asv_dev.txt
##############
def read_asv_txt_file(file_path):
data = np.genfromtxt(
file_path, dtype=[('class', 'U10'),('type', 'U10'),
('score','f4')], delimiter=" ")
data_new = np.zeros([data.shape[0], 2])
for idx, data_entry in enumerate(data):
data_new[idx, 0] = data_entry[-1]
if data_entry[1] == 'target':
data_new[idx, 1] = 2
elif data_entry[1] == 'nontarget':
data_new[idx, 1] = 1
else:
data_new[idx, 1] = 0
return data_new
def read_cm_txt_file(file_path):
data = np.genfromtxt(
file_path, dtype=[('class', 'U10'),('type', 'U10'),
('flag', 'U10'),
('score','f4')], delimiter=" ")
data_new = np.zeros([data.shape[0], 2])
for idx, data_entry in enumerate(data):
data_new[idx, 0] = data_entry[-1]
if data_entry[-2] == 'bonafide':
data_new[idx, 1] = 1
else:
data_new[idx, 1] = 0
return data_new
if __name__ == "__main__":
# these two files are deleted from the repo
# for usage on the evaluation tool,
# please check ../tutorials/README.md
asv_scores = read_asv_txt_file('./data/asvspoof2019/asv_dev.txt')
cm_scores = read_cm_txt_file('./data/asvspoof2019/cm_dev.txt')
tar_asv = asv_scores[asv_scores[:, 1]==2, 0]
non_asv = asv_scores[asv_scores[:, 1]==1, 0]
spoof_asv = asv_scores[asv_scores[:, 1]==0, 0]
bona_cm = cm_scores[cm_scores[:, 1]==1, 0]
spoof_cm = cm_scores[cm_scores[:, 1]==0, 0]
mintdcf, eer, eer_threshold = tDCF_wrapper(
bona_cm, spoof_cm, tar_asv, non_asv, spoof_asv)
print("min tDCF: {:f}".format(mintdcf))
print("EER: {:f}%".format(eer*100))
| 37,875 | 41.461883 | 167 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_dist.py | ##!/usr/bin/env python
"""
Module definition for distributions
Definition of distributions for generative models.
Each module should define two methods: forward and inference.
1. forward(input_feat, target): computes distribution given input_feat and
likelihood given target_data
2. inference(input_feat): computes distribution given input_feat and draw sample
Note that Modules defined in core_modules/*.py are for discrminative models.
There is no method for inference. But they may be combined with this code
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
class DistCategorical(torch_nn.Module):
"""Output layer that implements categorical distribution
This Module implements two methods: forward and inference.
forward(input_feat, target): computes the categorical
distribution given input_feat and likelihood given target_data
inference(input_feat): computes the categorical
distribution given input_feat and generate output
Input_feat is the logits before softmax. It will be converted
into a probablity vector inside this Module.
In other words, input_feat does not need to be a probablity vector.
Example:
dim = 4
logits = torch.rand([2, 3, dim])
logits[0, 1, 0] += 9.9
logits[0, 2, 1] += 9.9
logits[0, 0, 2] += 9.9
logits[1, 1, 1] += 9.9
logits[1, 2, 2] += 9.9
logits[1, 0, 0] += 9.9
target = torch.tensor([[[2], [0], [1]], [[0], [1], [2]]])
l_cat = DistCategorical(dim)
samples = l_cat.inference(logits)
print(logits)
print(samples)
loss = l_cat.forward(logits, target)
print(loss)
"""
def __init__(self, category_size):
""" DistCategorical(category_size)
Args
----
category_size: int, number of category
"""
super(DistCategorical, self).__init__()
self.category_size = category_size
self.loss = torch_nn.CrossEntropyLoss()
def _check_input(self, input_feat):
""" check whether input feature vector has the correct dimension
torch.dist does not check, it will gives output no matter what
the shape of input_feat
"""
if input_feat.shape[-1] != self.category_size:
mes = "block_dist.DistCategorical expects input_feat with "
mes += "last dimension of size {:d}. ".format(self.category_size)
mes += "But receives {:d}".format(input_feat.shape[-1])
raise Exception(mes)
return True
def forward(self, input_feat, target):
""" likelihood = forward(input_feat, target)
input
-----
input_feat: tensor (batchsize, length, categorize_size)
tensor to be converted into categorical distribution
target: (batchsize, length, dim=1)
tensor to be used to evaluate the likelihood
output
------
likelihood: tensor scaler
Note that input_feat is the logits, it will be converted to prob. vec
through LogSoftmax, Then NLLoss is computed.
In practice, we directly use this API that will do the two steps
https://pytorch.org/docs/stable/generated/torch.nn.CrossEntropyLoss.html
"""
self._check_input(input_feat)
# transpose input_feat to (batchsize, cateogrical_size, length)
# squeeze target to (batchsize, length)
return self.loss(input_feat.transpose(1, 2), target.squeeze(-1))
def inference(self, input_feat):
""" sample = inference(input_feat)
input
-----
input_feat: tensor (batchsize, length, categorize_size)
tensor to be converted into categorical distribution
output
------
sample: (batchsize, length, dim=1)
Note that input_feat is the logits, it will be converted to prob. vec
through Softmax in this method
"""
# check
self._check_input(input_feat)
# compute probability
prob_vec = torch_nn_func.softmax(input_feat, dim=2)
# distribution
distrib = torch.distributions.Categorical(prob_vec)
# draw samples and save
sample = torch.zeros(
[input_feat.shape[0], input_feat.shape[1], 1],
dtype=input_feat.dtype, device=input_feat.device)
sample[:, :, 0] = distrib.sample()
return sample
if __name__ == "__main__":
print("Definition of distributions modules")
| 4,823 | 31.16 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_resnet.py | ##!/usr/bin/env python
"""
ResNet model
Modified based on https://github.com/joaomonteirof/e2e_antispoofing
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
from scipy import signal as scipy_signal
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import torch.nn.init as torch_init
import sandbox.block_nn as nii_nn
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
##
class PreActBlock(torch_nn.Module):
""" Pre-activation version of the BasicBlock
"""
expansion = 1
def __init__(self, in_planes, planes, stride, *args, **kwargs):
super(PreActBlock, self).__init__()
# input batchnorm
self.bn1 = torch_nn.BatchNorm2d(in_planes)
# conv1
self.conv1 = torch_nn.Conv2d(
in_planes, planes, kernel_size=3, stride=stride,
padding=1, bias=False)
self.bn2 = torch_nn.BatchNorm2d(planes)
self.conv2 = torch_nn.Conv2d(
planes, planes, kernel_size=3, stride=1, padding=1, bias=False)
if stride != 1 or in_planes != self.expansion * planes:
self.shortcut = torch_nn.Sequential(
torch_nn.Conv2d(in_planes, self.expansion * planes,
kernel_size=1, stride=stride, bias=False))
def forward(self, x):
out = torch_nn_func.relu(self.bn1(x))
shortcut = self.shortcut(out) if hasattr(self, 'shortcut') else x
out = self.conv1(out)
out = self.conv2(torch_nn_func.relu(self.bn2(out)))
out += shortcut
return out
class PreActBottleneck(torch_nn.Module):
""" Pre-activation version of the original Bottleneck module.
"""
expansion = 4
def __init__(self, in_planes, planes, stride, *args, **kwargs):
super(PreActBottleneck, self).__init__()
#
self.bn1 = torch_nn.BatchNorm2d(in_planes)
self.conv1 = torch_nn.Conv2d(
in_planes, planes, kernel_size=1, bias=False)
self.bn2 = torch_nn.BatchNorm2d(planes)
self.conv2 = torch_nn.Conv2d(
planes, planes, kernel_size=3, stride=stride, padding=1, bias=False)
self.bn3 = torch_nn.BatchNorm2d(planes)
self.conv3 = torch_nn.Conv2d(
planes, self.expansion * planes, kernel_size=1, bias=False)
if stride != 1 or in_planes != self.expansion * planes:
self.shortcut = torch_nn.Sequential(
torch_nn.Conv2d(in_planes, self.expansion*planes,
kernel_size=1, stride=stride, bias=False))
def forward(self, x):
out = torch_nn_func.relu(self.bn1(x))
shortcut = self.shortcut(out) if hasattr(self, 'shortcut') else x
out = self.conv1(out)
out = self.conv2(torch_nn_func.relu(self.bn2(out)))
out = self.conv3(torch_nn_func.relu(self.bn3(out)))
out += shortcut
return out
def conv3x3(in_planes, out_planes, stride=1):
return torch_nn.Conv2d(in_planes, out_planes,
kernel_size=3, stride=stride, padding=1, bias=False)
def conv1x1(in_planes, out_planes, stride=1):
return torch_nn.Conv2d(in_planes, out_planes,
kernel_size=1, stride=stride, bias=False)
RESNET_CONFIGS = {'18': [[2, 2, 2, 2], PreActBlock],
'28': [[3, 4, 6, 3], PreActBlock],
'34': [[3, 4, 6, 3], PreActBlock],
'50': [[3, 4, 6, 3], PreActBottleneck],
'101': [[3, 4, 23, 3], PreActBottleneck]
}
class ResNet(torch_nn.Module):
def __init__(self, enc_dim, resnet_type='18', nclasses=2):
self.in_planes = 16
super(ResNet, self).__init__()
layers, block = RESNET_CONFIGS[resnet_type]
self._norm_layer = torch_nn.BatchNorm2d
# laye 1
self.conv1 = torch_nn.Conv2d(1, 16, kernel_size=(9, 3),
stride=(3, 1), padding=(1, 1), bias=False)
self.bn1 = torch_nn.BatchNorm2d(16)
self.activation = torch_nn.ReLU()
self.layer1 = self._make_layer(block, 64, layers[0], stride=1)
self.layer2 = self._make_layer(block, 128, layers[1], stride=2)
self.layer3 = self._make_layer(block, 256, layers[2], stride=2)
self.layer4 = self._make_layer(block, 512, layers[3], stride=2)
self.conv5 = torch_nn.Conv2d(
512 * block.expansion, 256, kernel_size=(3, 3),
stride=(1, 1), padding=(0, 1), bias=False)
self.bn5 = torch_nn.BatchNorm2d(256)
self.fc = torch_nn.Linear(256 * 2, enc_dim)
if nclasses >= 2:
self.fc_mu = torch_nn.Linear(enc_dim, nclasses)
else:
self.fc_mu = torch_nn.Linear(enc_dim, 1)
self.initialize_params()
self.attention = nii_nn.SelfWeightedPooling(256)
def initialize_params(self):
for layer in self.modules():
if isinstance(layer, torch.nn.Conv2d):
torch_init.kaiming_normal_(layer.weight, a=0, mode='fan_out')
elif isinstance(layer, torch.nn.Linear):
torch_init.kaiming_uniform_(layer.weight)
elif isinstance(layer, torch.nn.BatchNorm2d) or \
isinstance(layer, torch.nn.BatchNorm1d):
layer.weight.data.fill_(1)
layer.bias.data.zero_()
def _make_layer(self, block, planes, num_blocks, stride=1):
norm_layer = self._norm_layer
downsample = None
if stride != 1 or self.in_planes != planes * block.expansion:
downsample = torch_nn.Sequential(
conv1x1(self.in_planes, planes * block.expansion, stride),
norm_layer(planes * block.expansion))
layers = []
layers.append(
block(self.in_planes, planes, stride, downsample,
1, 64, 1, norm_layer))
self.in_planes = planes * block.expansion
for _ in range(1, num_blocks):
layers.append(
block(self.in_planes, planes, 1, groups=1,
base_width=64, dilation=False, norm_layer=norm_layer))
return torch_nn.Sequential(*layers)
def forward(self, x, without_pooling=False):
x = self.conv1(x)
x = self.activation(self.bn1(x))
x = self.layer1(x)
x = self.layer2(x)
x = self.layer3(x)
x = self.layer4(x)
x = self.conv5(x)
x = self.activation(self.bn5(x)).squeeze(2)
if without_pooling:
return x
else:
stats = self.attention(x.permute(0, 2, 1).contiguous())
feat = self.fc(stats)
mu = self.fc_mu(feat)
return feat, mu
if __name__ == "__main__":
print("Definition of Resnet for anti-spoofing")
| 6,982 | 33.399015 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_glow.py | #!/usr/bin/env python
"""
Building blocks for glow
"""
from __future__ import absolute_import
import os
import sys
import time
import numpy as np
import scipy.linalg
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import torch.nn.init as torch_init
import sandbox.block_nn as nii_nn
import core_scripts.data_io.conf as nii_io_conf
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def sum_over_keep_batch(data):
# (batch, dim1, dim2, ..., ) -> (batch)
# sum over dim1, dim2, ...
sum_dims = [x for x in range(data.ndim)][1:]
return torch.sum(data, dim=sum_dims)
def sum_over_keep_batch2(data, factor):
# (batch, dim1, dim2, ..., ) -> (batch)
# device each value by factor and
# sum over dim1, dim2, ...
sum_dims = [x for x in range(data.ndim)][1:]
return torch.sum(data / factor, dim=sum_dims)
class ActNorm(torch_nn.Module):
"""Activation Normalization
Activation normalization layer used in
Kingma, D. P. & Dhariwal, P. Glow
Generative Flow with Invertible 1x1 Convolutions.
arXiv Prepr. arXiv1807.03039 (2018)
For debug:
m_actnorm = ActNorm(5, flag_detjac=True)
data = torch.rand([2, 5, 5])
out, detjac = m_actnorm(data)
data_new = m_actnorm.reverse(out)
print(detjac)
#print(data.mean(dim=[0, 1]))
#print(data.std(dim=[0, 1]))
#print(m_actnorm.m_bias)
#print(m_actnorm.m_scale)
print(torch.sum(torch.log(torch.abs(m_actnorm.m_scale))) * 5 * 2)
print(data - data_new)
"""
def __init__(self, feat_dim, flag_detjac=False):
"""ActNorm(feat_dim, flag_detjac)
Args
----
feat_dim: int, feature dimension (channel for image),
input tensor (batch, ..., feature dimension)
flag_detjac: bool, whether output determinant of jacobian
Note that, it assumes y -> H(.) -> x, where H(.) is ActNorm.forward,
it then returns |det(dH(y)/dy)|
"""
super(ActNorm, self).__init__()
# flag
# whether return det of jacobian matrix
self.flag_detjac = flag_detjac
#
self.feat_dim = feat_dim
# parameter
self.m_scale = torch_nn.Parameter(torch.ones(feat_dim),
requires_grad=True)
self.m_bias = torch_nn.Parameter(torch.zeros(feat_dim),
requires_grad=True)
# flag to prevent re-initialization of the scale and bias
self.m_init_flag = torch_nn.Parameter(torch.zeros(1),
requires_grad=False)
return
def _log(self, x):
# add a floor
#return torch.log(x + torch.finfo(x.dtype).eps)
return torch.log(x)
def _detjac(self, factor=1):
"""
"""
# \sum log |s|, this same value is used for all data
# in this mini-batch, no need to duplicate to (batch,)
return torch.sum(self._log(torch.abs(self.m_scale)) / factor)
def _detjac_size_factor(self, y):
""" h * w * detjac
we need to compute h * w
"""
with torch.no_grad():
# tensor in shape (batch, d1, d2, ... feat_dim)
# then the factor will be d1 x d2 ...
data_size = torch.tensor(y.shape[1:-1])
data_factor = torch.prod(data_size)
return data_factor
def _init_scale_m(self, y):
""" initialize scale and bias for transformation
"""
with torch.no_grad():
# (batch, ... ,feat_dim) -> (-1, feat_dim)
tmp_y = y.view(-1, self.feat_dim)
# get mean and std per feat_dim
m = torch.mean(tmp_y, dim=0)
std = torch.std(tmp_y, dim=0) + 1e-6
# because the transform is (y + bias) * scale
# save scale = 1/std and bias = -m
self.m_scale.data = 1 / std
self.m_bias.data = -1 * m
# prevent further initialization
self.m_init_flag += 1
return
def forward(self, y, factor=1):
"""x = ActNorm.forward(y)
input
-----
y: tensor, (batch, dim1, ..., feat_dim)
output
------
x: tensor, (batch, dim1, ..., feat_dim)
if self.flag_detjac, also returns log_detjac (scalar)
"""
# do initialization for the 1st time
if self.m_init_flag.item() < 1:
self._init_scale_m(y)
# in initial stage, this is equivalent to (y - m)/std
x = (y + self.m_bias) * self.m_scale
if self.flag_detjac:
log_detjac = self._detjac(factor) * self._detjac_size_factor(y)
return x, log_detjac
else:
return x
def reverse(self, x):
"""y = ActNorm.reverse(x)
input
-----
x: tensor, (batch, dim1, ..., feat_dim)
output
------
y: tensor, (batch, dim1, ..., feat_dim)
"""
return x / self.m_scale - self.m_bias
class InvertibleTrans(torch_nn.Module):
"""InvertibleTransformation
Invertible transformation layer used in
Kingma, D. P. & Dhariwal, P. Glow
Generative Flow with Invertible 1x1 Convolutions.
arXiv Prepr. arXiv1807.03039 (2018)
1x1 convolution is implemented using torch.matmul
Example:
feat_dim = 5
m_trans = InvertibleTrans(feat_dim, flag_detjac=True)
data = torch.rand([2, feat_dim, feat_dim])
out, detjac = m_trans(data)
data_new = m_trans.reverse(out)
print(data_new - data)
print(detjac)
"""
def __init__(self, feat_dim, flag_detjac=False):
"""InvertibleTrans(feat_dim, flag_detjac)
Args
----
feat_dim: int, feature dimension (channel for image),
input tensor (batch, ..., feature dimension)
flag_detjac: bool, whether output determinant of jacobian
It assumes y -> H(.) -> x, where H(.) is InvertibleTrans.forward,
it then returns |det(dH(y)/dy)|
"""
super(InvertibleTrans, self).__init__()
#
self.feat_dim = feat_dim
# create initial permutation, lower, and upper triangle matrices
seed_mat = np.random.randn(feat_dim, feat_dim)
# qr decomposition, rotation_mat is a unitary matrix
rotation_mat, _ = scipy.linalg.qr(seed_mat)
# LU decomposition
permute_mat, lower_mat, upper_mat = scipy.linalg.lu(rotation_mat)
# mask matrix (with zero on the diagonal line)
u_mask = np.triu(np.ones_like(seed_mat), k=1)
d_mask = u_mask.T
# permuate matrix, fixed
self.m_permute_mat = torch_nn.Parameter(
torch.tensor(permute_mat.copy(), dtype=nii_io_conf.d_dtype),
requires_grad=False)
# Lower triangle matrix, trainable
self.m_lower_tria = torch_nn.Parameter(
torch.tensor(lower_mat.copy(), dtype=nii_io_conf.d_dtype),
requires_grad=True)
# Uppper triangle matrix, trainable
self.m_upper_tria = torch_nn.Parameter(
torch.tensor(upper_mat.copy(), dtype=nii_io_conf.d_dtype),
requires_grad=True)
# diagonal line
tmp_diag_line = torch.tensor(
upper_mat.diagonal().copy(),dtype=nii_io_conf.d_dtype)
# use log(|s|)
self.m_log_abs_diag = torch_nn.Parameter(
torch.log(torch.abs(tmp_diag_line)), requires_grad=True)
# save the sign of s as fixed parameter
self.m_diag_sign = torch_nn.Parameter(
torch.sign(tmp_diag_line), requires_grad=False)
# mask and all-1 diangonal line
self.m_l_mask = torch_nn.Parameter(
torch.tensor(d_mask.copy(), dtype=nii_io_conf.d_dtype),
requires_grad=False)
self.m_u_mask = torch_nn.Parameter(
torch.tensor(u_mask.copy(), dtype=nii_io_conf.d_dtype),
requires_grad=False)
self.m_eye = torch_nn.Parameter(
torch.eye(self.feat_dim, dtype=nii_io_conf.d_dtype),
requires_grad=False)
# buffer for inverse matrix
self.flag_invered = False
self.m_inver = torch_nn.Parameter(
torch.tensor(permute_mat.copy(), dtype=nii_io_conf.d_dtype),
requires_grad=False)
#
self.flag_detjac = flag_detjac
return
def _inverse(self):
""" inverse of the transformation matrix
"""
return torch.inverse(self._compose_mat())
def _compose_mat(self):
""" compose the transformation matrix
W = P L (U + sign * exp( log|s|))
"""
# U + sign * exp(log|s|)
tmp_u = torch.diag(self.m_diag_sign * torch.exp(self.m_log_abs_diag))
tmp_u = tmp_u + self.m_upper_tria * self.m_u_mask
# L
tmp_l = self.m_lower_tria * self.m_l_mask + self.m_eye
return torch.matmul(self.m_permute_mat, torch.matmul(tmp_l, tmp_u))
def _log(self, x):
# add a floor
#return torch.log(x + torch.finfo(x.dtype).eps)
return torch.log(x)
def _detjac(self, factor=1):
"""
"""
# \sum log|s|
# no need to duplicate to each data in the batch
# they all use the same detjac
return torch.sum(self.m_log_abs_diag / factor)
def _detjac_size_factor(self, y):
with torch.no_grad():
# tensor in shape (batch, d1, d2, ... feat_dim)
# then the factor will be d1 x d2 ...
data_size = torch.tensor(y.shape[1:-1])
data_factor = torch.prod(data_size)
return data_factor
def forward(self, y, factor=1):
# y W
# for other implementation, this is done with conv2d 1x1 convolution
# to be consistent, we can use .T to transpose the matrix first
if self.flag_detjac:
detjac = self._detjac(factor) * self._detjac_size_factor(y)
return torch.matmul(y, self._compose_mat()), detjac
else:
return torch.matmul(y, self._compose_mat()),
def reverse(self, x):
if self.training:
# if it is for training, compute inverse everytime
self.m_inver.data = self._inverse().clone()
else:
# during inference, only do this once
if self.flag_invered is False:
self.m_inver.data = self._inverse().clone()
# only compute inverse matrix once
self.flag_invered = True
return torch.matmul(x, self.m_inver)
class ZeroInitConv2dForGlow(torch_nn.Module):
"""ZeroIniConv2dForGlow
Last Conv2d layer of Glow uses zero-initialized conv2d
This is only used for images
"""
def __init__(self, in_feat_dim, out_feat_dim, kernel_size=3, padding=1):
super().__init__()
# conv
self.m_conv = torch_nn.Conv2d(in_feat_dim, out_feat_dim,
kernel_size, padding=0)
self.m_conv.weight.data.zero_()
self.m_conv.bias.data.zero_()
# scale parameter, following https://github.com/rosinality/glow-pytorch/
self.m_scale = torch_nn.Parameter(
torch.zeros(out_feat_dim, dtype=nii_io_conf.d_dtype))
#
self.m_pad_size = padding
return
def _zerobias(self):
self.m_conv.bias.data.zero_()
return
def _normal_weight(self):
self.m_conv.weight.data.normal_(0, 0.05)
return
def forward(self, x):
p = self.m_pad_size
# pad
y = torch_nn_func.pad(x.permute(0, 3, 1, 2), [p,p,p,p], value=1)
# conv
y = self.m_conv(y).permute(0, 2, 3, 1).contiguous()
# scale parameter, following https://github.com/rosinality/glow-pytorch/
return y * torch.exp(self.m_scale * 3)
class Conv2dForGlow(torch_nn.Module):
"""Conv2dForGlow
Other Conv2d layer of Glow uses zero-initialized conv2d
This is only used for images
"""
def __init__(self, in_feat_dim, out_feat_dim, kernel_size=3, padding=1):
super().__init__()
self.m_conv = torch_nn.Conv2d(in_feat_dim, out_feat_dim,
kernel_size, padding=padding)
return
def _zerobias(self):
self.m_conv.bias.data.zero_()
return
def _normal_weight(self):
self.m_conv.weight.data.normal_(0, 0.05)
return
def forward(self, x):
return self.m_conv(x.permute(0, 3, 1, 2)).permute(0,2,3,1).contiguous()
class AffineCouplingGlow(torch_nn.Module):
"""AffineCouplingGlow
AffineCoupling block in Glow
Example:
m_affine = AffineCouplingGlow(10, 32, flag_affine=False,flag_detjac=True)
data = torch.randn([2, 4, 4, 10])
data_out, detjac = m_affine(data)
data_inv = m_affine.reverse(data_out)
print(data_inv - data)
print(detjac)
"""
def __init__(self, feat_dim, conv_out_dim=512,
flag_affine=True, flag_detjac=False):
"""AffineCouplingGlow(feat_dim, conv_out_dim=512,
flag_affine=True, flag_detjac=False)
Args:
-----
feat_dim: int, dimension of input feature (channel number of image)
feat_dim must be an even number
conv_out_dim: int, dimension of output feature of the intermediate
conv layer, default 512
flag_affine: bool, whether use affine or additive transformation?
default True
flag_detjac: bool, whether return the determinant of Jacobian,
default False
It assumes that y -> H(.) -> x, where H(.) is AffineCouplingGlow.forward
When flag_affine == True, H(y) = concante([y1, exp(s) \odot y_2 + b])
When flag_affine == False, H(y) = concante([y1, y_2 + b])
where, [s, b] = NN(y1)
"""
super(AffineCouplingGlow, self).__init__()
self.flag_affine = flag_affine
self.flag_detjac = flag_detjac
if feat_dim % 2 > 0:
print("AffineCoulingGlow(feat_dim), feat_dim is an odd number?!")
sys.exit(1)
if self.flag_affine:
self.m_nn_outdim = feat_dim
else:
self.m_nn_outdim = feat_dim//2
# create network
self.m_conv = torch_nn.Sequential(
Conv2dForGlow(feat_dim//2, conv_out_dim, kernel_size=3, padding=1),
torch_nn.ReLU(),
Conv2dForGlow(conv_out_dim, conv_out_dim, kernel_size=1, padding=0),
torch_nn.ReLU(),
ZeroInitConv2dForGlow(conv_out_dim, self.m_nn_outdim,
kernel_size=3, padding=1)
)
# no bias, normal initial weight
self.m_conv[0]._zerobias()
self.m_conv[0]._normal_weight()
self.m_conv[2]._zerobias()
self.m_conv[2]._normal_weight()
return
def _detjac(self, log_scale, factor=1):
# (batch, dim1, dim2, ..., feat_dim) -> (batch)
# sum over dim1, ... feat_dim
return sum_over_keep_batch(log_scale/factor)
def _nn_trans(self, y1):
if self.flag_affine:
log_scale, bias = self.m_conv(y1).chunk(2, -1)
# follow openai implementation
scale = torch.sigmoid(log_scale + 2)
log_scale = torch.log(scale)
else:
bias = self.m_conv(y1)
scale = torch.ones_like(y1)
log_scale = torch.zeros_like(y1)
return scale, bias, log_scale
def forward(self, y, factor=1):
"""AffineCoulingGlow(y)
input
-----
y: tensor, (batch, dim1, dim2, ..., feat_dim)
output
------
out: tensor, (batch, dim1, dim2, ..., feat_dim)
"""
# split
y1, y2 = y.chunk(2, -1)
scale, bias, log_scale = self._nn_trans(y1)
# transform
x1 = y1
x2 = (y2 + bias) * scale
# concatenate
x = torch.cat([x1, x2], dim=-1)
if self.flag_detjac:
return x, self._detjac(log_scale, factor)
else:
return x
def reverse(self, x):
# split
x1, x2 = x.chunk(2, -1)
# reverse transform
y1 = x1
scale, bias, log_scale = self._nn_trans(y1)
y2 = x2 / scale - bias
#
return torch.cat([y1, y2], dim=-1)
class SqueezeForGlow(torch_nn.Module):
"""SqueezeForGlow
Squeeze layer for Glow
See doc of __init__ for different operation modes
Example:
data = torch.randn([2, 4, 4, 3])
m_squeeze = SqueezeForGlow()
data_squeezed = m_squeeze(data)
data_unsqu = m_squeeze.reverse(data_squeezed)
print(data)
print(data_squeezed)
print(torch.std(data_unsqu - data))
print(data[0, :, :, 0])
print(data_squeezed[0, :, :, 0])
print(data_squeezed[0, :, :, 1])
print(data_squeezed[0, :, :, 2])
print(data_squeezed[0, :, :, 3])
"""
def __init__(self, mode = 1):
"""SqueezeForGlow(mode=1)
Args
----
mode: int, 1: for image
2: for audio
mode == 1:
(batch, height, width, channel)->(batch, height/2, width/2, channel*4)
"""
super(SqueezeForGlow, self).__init__()
self.m_mode = mode
return
def get_squeeze_factor(self):
if self.m_mode == 1:
# for image, the channel number will be compressed by 4
return 4
def forward(self, x):
"""
"""
if self.m_mode == 1:
# assume (batch, height, weight, channel)
if len(x.shape) != 4:
print("SqueezeForGlow(mode=1)")
print(", input should be (batch, height, weight, channel)")
sys.exit(1)
batch, height, width, channel = x.shape
# (batch, height, 2, width, 2, channel)
x_squeezed = x.view(batch, height // 2, 2, width // 2, 2, channel)
# (batch, height, width, channel * 2 * 2)
x_squeezed = x_squeezed.permute(0, 1, 3, 5, 2, 4).contiguous()
x_squeezed = x_squeezed.view(batch, height//2, width//2, channel*4)
else:
print("SqueezeForGlow not implemented")
return x_squeezed
def reverse(self, x_squeezed):
if self.m_mode == 1:
# assume (batch, height, weight, channel)
if len(x_squeezed.shape) != 4:
print("SqueezeForGlow(mode=1)")
print(", input should be (batch, height, weight, channel)")
sys.exit(1)
batch, height, width, channel = x_squeezed.shape
x = x_squeezed.view(batch, height, width, channel // 4, 2, 2)
# (batch, height * 2, width * 2, channel)
x = x.permute(0, 1, 4, 2, 5, 3).contiguous()
x = x.view(batch, height*2, width*2, channel//4)
else:
print("SqueezeForGlow not implemented")
return x
class PriorTransform(torch_nn.Module):
"""Prior transformation at the end of each Glow block
This is not written in paper but implemented in official code.
https://github.com/rosinality/glow-pytorch/issues/11
This is wrapper around the split operation. However, additional
affine transformation is included.
Given y,
If flag_split == True:
x, z_1 <- y.split()
z_0 <- (z_1 - f_bias(x)) / f_scale(x)
In native implementation, we can directly evaluate N(z_1; 0, I).
However, this block further converts z_1 -> z_0
If flag_split == False:
if flag_final_block == True:
z_1 <- y
z_0 <- (z_1 - f_bias(0)) / f_scale(0), final latent
x <- None , no input for the next Glowblock
else
x <- y , which is used for the next Glowblock
x <- (x - f_bias(0)) / f_scale(0), input to the next GlowBlock
z_0 <- None , no split output
"""
def __init__(self, feat_dim, flag_split, flag_final_block):
"""PriorTransform(feat_dim)
Args
----
feat_dim: int, feature dimension or channel number
input tensor should be (batch, dim1, dim2, ..., feat_dim)
image should be (batch, height, weight, feat_dim)
flag_split: bool, split or not split
flag_final_block: bool, whether this is the for the final block
"""
super(PriorTransform, self).__init__()
self.flag_split = flag_split
if flag_split:
self.m_nn = ZeroInitConv2dForGlow(feat_dim // 2, feat_dim)
else:
self.m_nn = ZeroInitConv2dForGlow(feat_dim, feat_dim * 2)
self.flag_final_block = flag_final_block
if flag_final_block and flag_split:
print("PriorTransform flag_split and flag_final_block are True")
print("This is unexpected. please check model definition")
sys.exit(1)
return
def _detjac(self, log_scale, factor=1):
# log|\prod 1/exp(log_scale)| = -\sum log_scale
# note that we should return a tensor (batch,)
return sum_over_keep_batch(-1 * log_scale / factor)
def forward(self, y, factor=1):
"""PriorTransform(y)
y -> H() -> [x, z_0]
input
-----
y: (batch, dim1, ..., feat_dim)
output
------
x: tensor or None, input to the next GlowBlock
z_0: tensor or None, latent variable for evaluating N(z_0; 0, I)
log_detjac: scalar
Note that
If self.flag_split==True, x, z_0 will (batch, dim1, ..., feat_dim//2)
If self.flag_split==False and self.flag_final_block==True:
x = None, which indicates no input for the next GlowBlock
z_0, (batch, dim1, ..., feat_dim)
If self.flag_split==False and self.flag_final_block==False:
z_0 = None, which indicates no latent output from this GlowBlock
x, (batch, dim1, ..., feat_dim), input to the next GlowBlock
"""
if not self.flag_split:
zeros = torch.zeros_like(y)
z_mean, z_log_std = self.m_nn(zeros).chunk(2, -1)
if self.flag_final_block:
# For z_1 <- y
# z_0 <- (z_1 - f_bias(zero)) / f_scale(zero)
# x <- None
z_0 = (y - z_mean) / torch.exp(z_log_std)
x = None
else:
# z_0 <- None
# x <- (z_1 - f_bias(zero)) / f_scale(zero)
z_0 = None
x = (y - z_mean) / torch.exp(z_log_std)
else:
# For x, z_1 <- y.split()
# z_0 <- (z_1 - f_bias(x)) / f_scale(x)
x, z_1 = y.chunk(2, -1)
z_mean, z_log_std = self.m_nn(x).chunk(2, -1)
z_0 = (z_1 - z_mean) / torch.exp(z_log_std)
return x, z_0, self._detjac(z_log_std, factor)
def reverse(self, x, z_out):
"""PriorTransform(y)
y <- H() <- x, z_0
input
-----
x: tensor or None
z_0: tensor or None
output
------
y: (batch, dim1, ..., feat_dim)
Note that
If self.flag_split==True
x, z_out should be (batch, dim1, ..., feat_dim//2)
If self.flag_split==False and self.flag_final_block==True:
x = None, which indicates no input for from the following GlowBlock
z_0, (batch, dim1, ..., feat_dim)
If self.flag_split==False and self.flag_final_block==False:
z_0 = None, which indicates no latent additional this GlowBlock
x, (batch, dim1, ..., feat_dim), input from the following GlowBlock
"""
if self.flag_split:
if x is not None:
z_mean, z_log_std = self.m_nn(x).chunk(2, -1)
z_tmp = z_out * torch.exp(z_log_std) + z_mean
y_tmp = torch.cat([x, z_tmp], -1)
else:
print("PriorTransform.reverse receives None")
sys.exit(1)
else:
if self.flag_final_block:
zeros = torch.zeros_like(z_out)
z_mean, z_log_std = self.m_nn(zeros).chunk(2, -1)
y_tmp = z_out * torch.exp(z_log_std) + z_mean
else:
zeros = torch.zeros_like(x)
z_mean, z_log_std = self.m_nn(zeros).chunk(2, -1)
y_tmp = x * torch.exp(z_log_std) + z_mean
return y_tmp
class FlowstepGlow(torch_nn.Module):
"""FlowstepGlow
One flow step in Glow
"""
def __init__(self, feat_dim, flag_affine=True, conv_coup_dim=512):
"""FlowstepGlow(feat_dim, flag_affine=True)
Args:
-----
feat_dim: int, dimension of input feature (channel number of image)
feat_dim must be an even number
flag_affine: bool, whether use affine or additive transformation in
AffineCouplingGlow layer (see AffineCouplingGlow)
default True.
conv_coup_dim: int, dimension of intermediate cnn layer in coupling
default 512, (see AffineCouplingGlow)
It assumes that y -> H(.) -> x, where H(.) is FlowstepGlow.forward
"""
super(FlowstepGlow, self).__init__()
self.flag_affine = flag_affine
# layers
self.m_actnorm = ActNorm(feat_dim, flag_detjac=True)
self.m_invtrans = InvertibleTrans(feat_dim, flag_detjac=True)
self.m_coupling = AffineCouplingGlow(feat_dim, conv_coup_dim,
flag_affine, flag_detjac=True)
return
def forward(self, y):
x_tmp, log_tmp1 = self.m_actnorm(y)
x_tmp, log_tmp2 = self.m_invtrans(x_tmp)
x_tmp, log_tmp3 = self.m_coupling(x_tmp)
return x_tmp, log_tmp1 + log_tmp2 + log_tmp3
def reverse(self, x):
# prevent accidental reverse during training
y_tmp = self.m_coupling.reverse(x)
y_tmp = self.m_invtrans.reverse(y_tmp)
y_tmp = self.m_actnorm.reverse(y_tmp)
return y_tmp
class GlowBlock(torch_nn.Module):
"""GlowBlock
One Glow block, squeeze + step_of_flow + (split), Fig2.(b) in original paper
Example:
m_glow = GlowBlock(3, num_flow_step=32)
data = torch.randn([2, 64, 64, 3])
x, z, detjac = m_glow(data)
m_glow.eval()
data_new = m_glow.reverse(x, z)
#print(m_glow.training)
#print(x, z)
print(torch.std(data_new - data))
"""
def __init__(self, feat_dim, num_flow_step=12, conv_coup_dim = 512,
flag_split=True, flag_final_block=False,
flag_affine=True, squeeze_mode=1):
"""GlowBlock(feat_dim, num_flow_step=12, conv_coup_dim = 512,
flag_split=True, flag_affine=True, squeeze_mode=1)
Args
----
feat_dim: int, dimension of input feature (channel number of image)
feat_dim must be an even number
num_flow_step: int, number of flow steps, default 12
conv_coup_dim: int, dimension of intermediate cnn layer in coupling
default 512, (see AffineCouplingGlow)
flag_split: bool, whether split out.
Last GlowBlock uses flag_split=False
default True
flag_final_block: bool, whether this is the final GlowBlock
default False
flag_affine: bool, whether use affine or additive transformation in
AffineCouplingGlow layer (see AffineCouplingGlow)
default True.
squeeze_mode: int, mode for squeeze, default 1 (see SqueezeForGlow)
"""
super(GlowBlock, self).__init__()
# squeeze
self.m_squeeze = SqueezeForGlow(squeeze_mode)
# number of feat-dim after sequeeze (other channels)
squeezed_feat_dim = feat_dim * self.m_squeeze.get_squeeze_factor()
# steps of flow
self.m_flow_steps = []
for i in range(num_flow_step):
self.m_flow_steps.append(
FlowstepGlow(squeezed_feat_dim, flag_affine, conv_coup_dim))
self.m_flow_steps = torch_nn.ModuleList(self.m_flow_steps)
# prior transform
self.flag_split = flag_split
self.flag_final_block = flag_final_block
if self.flag_final_block and self.flag_split:
print("GlowBlock flag_split and flag_final_block are True")
print("This is unexpected. Please check model definition")
sys.exit(1)
self.m_prior = PriorTransform(
squeezed_feat_dim, self.flag_split, self.flag_final_block)
return
def forward(self, y):
"""x, z, log_detjac = GlowBlock(y)
input
-----
y: tensor, (batch, height, width, channel)
output
------
x: tensor, (batch, height, width, channel//2),
z: tensor, (batch, height, width, channel//2),
log_detjac: tensor or scalar
For multi-scale glow, z is the whitenned noise
"""
log_detjac = 0
# squeeze
y_suqeezed = self.m_squeeze(y)
# flows
x_tmp = y_suqeezed
for m_flow in self.m_flow_steps:
x_tmp, log_detjac_tmp = m_flow(x_tmp)
log_detjac += log_detjac_tmp
# prior transform
x, z, log_detjac_tmp = self.m_prior(x_tmp)
log_detjac += log_detjac_tmp
# [x, z] should have the same size as input y_suqeezed
return x, z, log_detjac
def reverse(self, x, z):
"""
"""
# prior
x_tmp = self.m_prior.reverse(x, z)
# flow
for m_flow in self.m_flow_steps[::-1]:
x_tmp = m_flow.reverse(x_tmp)
# squeeze
y = self.m_squeeze.reverse(x_tmp)
return y
class Glow(torch_nn.Module):
"""Glow
"""
def __init__(self, feat_dim, flow_step_num=32, flow_block_num=4,
flag_affine=False, conv_coup_dim=512, squeeze_mode=1):
"""Glow(feat_dim, flow_step_num=32, flow_block_num=4,
flag_affine=True, conv_coup_dim=512, squeeze_mode=1)
Args
----
feat_dim: int, dimension of feature, or channel of input image
flow_step_num: int, number of flow steps per block, default 32
flow_block_num: int, number of flow blocks, default 4
flag_affine: bool, whether use affine transformation or not
default True, see AffineCouplingLayer
conv_coup_dim: int, channel size of intermediate conv layer in
coupling layer NN(). see AffineCouplingLayer
squeeze_mode: int, mode for suqeezing.
1 for image. See squeezeLayer
"""
super(Glow, self).__init__()
self.m_blocks = []
self.m_flag_splits = []
for i in range(flow_block_num):
# Whether the block uses split or not is completely determined by
# whether this block is the last block or not
# last block does not split output
flag_split = True if i < (flow_block_num - 1) else False
# save this information for generating random noise
self.m_flag_splits.append(flag_split)
# whether this is the final block
flag_final_block = True if i == (flow_block_num - 1) else False
self.m_blocks.append(
GlowBlock(
feat_dim * (2**i), flow_step_num, conv_coup_dim,
flag_split=flag_split, flag_final_block=flag_final_block,
flag_affine=flag_affine,
squeeze_mode=1))
self.m_blocks = torch_nn.ModuleList(self.m_blocks)
return
def _normal_lh(self, noise):
# likelihood of normal distribution on the given noise
return -0.5 * np.log(2 * np.pi) - 0.5 * noise ** 2
def forward(self, y):
"""Glow.forward(y)
Conducts y -> H(.) -> z, where z is supposed to be Gaussian noise
input
-----
y: tensor, (batch, dim1, dim2, ..., feat_dim)
for image, (batch, height, width, channel)
output
------
z: list of tensor, random noise from each block
neg_logp_y: scalar, - log p(y)
logp_z: scalar, -log N(z), averaged over batch and pixels
logdet: scalar, -|det dH(.)/dy|, averaged over batch and pixels
Because Glow uses multi-scale structure, z will be a list of noise
"""
batch_size = y.shape[0]
# for image, np.log(2) computes bit
# np.prod([dim for dim in y.shape[1:]]) is the image size in pixels
factor = np.log(2) * np.prod([dim for dim in y.shape[1:]])
z_bags = []
log_detjac = 0
log_pz = 0
h_tmp = y
for m_block in self.m_blocks:
h_tmp, z_tmp, log_detjac_tmp = m_block(h_tmp)
z_bags.append(z_tmp)
log_detjac += log_detjac_tmp / factor
# keep log_pz for each data in batch (batchsize,)
log_pz += sum_over_keep_batch(self._normal_lh(z_tmp)) / factor
# average over batch and pixels
neg_logp_y = -(log_pz + log_detjac).mean()
return z_bags, neg_logp_y, \
log_pz.mean(), log_detjac.mean()
def reverse(self, z_bags):
""" y = Glow.reverse(z_bags)
input
-----
z_bags: list of tensors
output
------
y: tensor, (batch, dim1, dim2, ..., feat_dim)
The random noise in z_bags should be compatible with the
model. You may use Glow.get_z_noises to retrieve a z_bags
"""
for i, (z, m_block) in enumerate(zip(z_bags[::-1],
self.m_blocks[::-1])):
if i == 0:
# the last block without split
y_tmp = m_block.reverse(None, z)
else:
y_tmp = m_block.reverse(y_tmp, z)
return y_tmp
def get_z_noises(self, image_size, noise_std=0.7, batchsize=16):
"""z_bags = Glow.get_z_noises(image_size, noise_std=0.7, batchsize=16)
Return a list of random noises for random sampling
input
-----
image_size: int, size of the image, assume image is square,
this number just specifies the height / width
noise_std: float, std of Gaussian noise, default 0.7
batchsize: int, batch size of this random data, default 16
output
------
z_bags: list of tensors
Shape of the random noise in z_bags is decided by Glow configuration.
Glow.reverse(z_bags) can be used to produce image from this z_bags
"""
device = next(self.parameters()).device
z_bags = []
tmp_im_size = image_size
tmp_chan = 3
for flag_split in self.m_flag_splits:
if flag_split:
tmp_im_size = tmp_im_size // 2
tmp_chan = tmp_chan * 2
else:
tmp_im_size = tmp_im_size // 2
tmp_chan = tmp_chan * 4
z_tmp = torch.randn([batchsize, tmp_im_size, tmp_im_size, tmp_chan],
dtype=nii_io_conf.d_dtype, device=device)
z_bags.append(z_tmp * noise_std)
return z_bags
if __name__ == "__main__":
print("Definition of Glow and its components")
| 36,862 | 33.809254 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/eval_sig_test.py | #!/usr/bin/env python
"""
eval_sig_test
Utilities for statistical test on EERs
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import core_scripts.math_tools.sig_test as sig_test
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
##################################
# Functions for computing z-values
#
##################################
def compute_z_independent(far_a, frr_a, far_b, frr_b, NI, NC):
"""z = compute_HTER_independent(hter_a, hter_b, NI, NC)
Bengio, S. & Mariéthoz, J. A statistical significance test for
person authentication. in Proc. Odyssey (2004).
Fig2. independent case
input
-----
far_a: float, far of system a, which be >=0 and <=1
frr_a: float, frr of system a, which be >=0 and <=1
far_b: float, far of system b, which be >=0 and <=1
frr_b: float, frr of system b, which be >=0 and <=1
NI: int, the number of impostor accesses.
NC: int, the number of client accesses.
output
------
z: float, statitics of the hypothesis test
"""
#
hter_a = (far_a + frr_a)/2
hter_b = (far_b + frr_b)/2
denominator = (far_a * (1 - far_a) + far_b * (1 - far_b)) / 4 / NI
denominator += (frr_a * (1 - frr_a) + frr_b * (1 - frr_b)) / 4 / NC
return np.abs(hter_a - hter_b) / np.sqrt(denominator)
def compute_z_dependent(far_ab, frr_ab, far_ba, frr_ba, NI, NC):
"""z = compute_HTER_independent(hter_a, hter_b, NI, NC)
Bengio, S. & Mariéthoz, J. A statistical significance test for
person authentication. in Proc. Odyssey (2004).
Fig2. dependent case
input
-----
far_ab: float, see paper
frr_ab: float, see paper
far_ba: float, see paper
frr_ba: float, see paper
NI: int, the number of impostor accesses.
NC: int, the number of client accesses.
output
------
z: float, statitics of the hypothesis test
"""
#
if far_ab == far_ba and frr_ab == frr_ba:
return 0
else:
denominator = np.sqrt((far_ab + far_ba) / (4 * NI)
+ (frr_ab + frr_ba) / (4 * NC))
return np.abs(far_ab + frr_ab - far_ba - frr_ba) / denominator
def get_eer(scores_positive, scores_negative):
"""eer, threshold = get_eer(scores_positive, scores_negative)
compute Equal Error Rate given input scores
input
-----
scores_positive: np.array, scores of positive class
scores_negative: np.array, scores of negative class
output
------
eer: float, equal error rate
threshold: float, the threshold for the err
"""
return compute_eer(scores_positive, scores_negative)
def get_far_frr_dependent(bona_score_a, spoof_score_a, threshold_a,
bona_score_b, spoof_score_b, threshold_b,
NI, NC):
"""
"""
far_ab_idx = np.bitwise_and(spoof_score_a < threshold_a,
spoof_score_b >= threshold_b)
far_ba_idx = np.bitwise_and(spoof_score_a >= threshold_a,
spoof_score_b < threshold_b)
frr_ab_idx = np.bitwise_and(bona_score_a >= threshold_a,
bona_score_b < threshold_b)
frr_ba_idx = np.bitwise_and(bona_score_a < threshold_a,
bona_score_b >= threshold_b)
far_ab = np.sum(far_ab_idx) / NI
far_ba = np.sum(far_ba_idx) / NI
frr_ab = np.sum(frr_ab_idx) / NC
frr_ba = np.sum(frr_ba_idx) / NC
return far_ab, far_ba, frr_ab, frr_ba
#######
# API for EER testing
#
#######
def sig_test_holm_bonf_method(eer_bags, NC, NI, significance_level=0.05,
flag_reverse_indicator=False):
"""test_results = sig_test_holm_bonf_method(eer_bags, sig_level, NI, NC):
input
-----
eer_bags: np.array, shape (N, M), where N is the number of systems,
and M is the number of random runs
M can be 1, which means no multi-runs
NC: int, number of bona fide trials in test set
NI: int, number of spoofed trials in test set
sig_level: float, significance level, default 0.05
flag_reverse_indicator: bool, by default, no significance difference
is indicated by value 1.0 in output test_results
flag_reverse_indicator = True will set no-sig-diff
output
------
test_results: np.array, shape (N*M, N*M),
test_results[i*M+j, l*M+n] shows the significance
test between the j-th run of i-th system
and the n-th run of the l-th system
Note:
test_results[i*M+j, l*M+n] == True: accept NULL hypothesis,
no significant different
"""
# get the reject/accept
#significance_level = 0.05
num_system = eer_bags.shape[0]
runs = eer_bags.shape[1]
z_values = np.zeros([num_system * runs, num_system * runs])
test_results = np.zeros(z_values.shape)
for sys_1_idx in range(num_system):
for sys_2_idx in range(num_system):
# compute z_values
# significance test must be conducted within this pair of system
z_value_tmp = np.zeros([runs, runs])
for run_idx1 in range(runs):
for run_idx2 in range(runs):
idx1 = sys_1_idx * runs + run_idx1
idx2 = sys_2_idx * runs + run_idx2
z_values[idx1, idx2] = compute_z_independent(
eer_bags[sys_1_idx, run_idx1],
eer_bags[sys_1_idx, run_idx1],
eer_bags[sys_2_idx, run_idx2],
eer_bags[sys_2_idx, run_idx2],
NI, NC)
z_value_tmp[run_idx1, run_idx2] = z_values[idx1, idx2]
# save results
if not flag_reverse_indicator:
test_results = sig_test.reject_null_holm_bonferroni(
z_values, z_values.size, significance_level)
else:
test_results = sig_test.reject_null_holm_bonferroni(
z_values, z_values.size, significance_level,
accept_value = False, reject_value = True)
return test_results
if __name__ == "__main__":
print("Scripts eval_sig_test")
| 6,449 | 32.247423 | 77 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_nsf.py | ##!/usr/bin/env python
"""
Major blocks defined for NSF
These blocks are originall defined in ../project/01_nsf/*/.model.py
Definition are gathered here for convience.
CondModule, SourceModule, and FilterModule are not copied here since
they may change according to the model for certain application
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
from scipy import signal as scipy_signal
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import torch.nn.init as torch_init
import sandbox.block_nn as nii_nn
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#######
# Neural filter block
#######
class NeuralFilterBlock(torch_nn.Module):
""" Wrapper over a single filter block
NeuralFilterBlock(signal_size, hidden_size, kernel_size, conv_num=10)
args
----
signal_size: int, input signal is in shape (batch, length, signal_size)
hidden_size: int, output of conv layers is (batch, length, hidden_size)
kernel_size: int, kernel size of the conv layers
conv_num: number of conv layers in this neural filter block (default 10)
legacy_scale: Bool, whether load scale as parameter or magic number
To be compatible with old models that defines self.scale
No impact on the result, just different ways to load a
fixed self.scale
"""
def __init__(self, signal_size, hidden_size, kernel_size=3, conv_num=10,
legacy_scale = False):
super(NeuralFilterBlock, self).__init__()
self.signal_size = signal_size
self.hidden_size = hidden_size
self.kernel_size = kernel_size
self.conv_num = conv_num
self.dilation_size = [np.power(2, x) for x in np.arange(conv_num)]
# ff layer to expand dimension
self.l_ff_1 = torch_nn.Linear(signal_size, hidden_size, \
bias=False)
self.l_ff_1_tanh = torch_nn.Tanh()
# dilated conv layers
tmp = [nii_nn.Conv1dKeepLength(hidden_size, hidden_size, x, \
kernel_size, causal=True, bias=False) \
for x in self.dilation_size]
self.l_convs = torch_nn.ModuleList(tmp)
# ff layer to de-expand dimension
self.l_ff_2 = torch_nn.Linear(hidden_size, hidden_size//4,
bias=False)
self.l_ff_2_tanh = torch_nn.Tanh()
self.l_ff_3 = torch_nn.Linear(hidden_size//4, signal_size,
bias=False)
self.l_ff_3_tanh = torch_nn.Tanh()
# a simple scale: to be consistent with CURRENNT implementation
if legacy_scale:
# in case this scale is defined as model parameter in
# some old models
self.scale = torch_nn.Parameter(
torch.tensor([0.1]), requires_grad=False)
else:
# simple hyper-parameter should be OK
self.scale = 0.1
return
def forward(self, signal, context):
"""
input
-----
signal (batchsize, length, signal_size)
context (batchsize, length, hidden_size)
context is produced from the condition module
output
------
output: (batchsize, length, signal_size)
"""
# expand dimension
tmp_hidden = self.l_ff_1_tanh(self.l_ff_1(signal))
# loop over dilated convs
# output of a d-conv is input + context + d-conv(input)
for l_conv in self.l_convs:
tmp_hidden = tmp_hidden + l_conv(tmp_hidden) + context
# to be consistent with legacy configuration in CURRENNT
tmp_hidden = tmp_hidden * self.scale
# compress the dimesion and skip-add
tmp_hidden = self.l_ff_2_tanh(self.l_ff_2(tmp_hidden))
tmp_hidden = self.l_ff_3_tanh(self.l_ff_3(tmp_hidden))
output_signal = tmp_hidden + signal
return output_signal
############################
# Source signal generator
############################
class SineGen(torch_nn.Module):
""" Definition of sine generator
SineGen(samp_rate, harmonic_num = 0,
sine_amp = 0.1, noise_std = 0.003,
voiced_threshold = 0,
flag_for_pulse=False)
args
----
samp_rate: flaot, sampling rate in Hz
harmonic_num: int, number of harmonic overtones (default 0, i.e., only F0)
sine_amp: float, amplitude of sine-wavefrom (default 0.1)
noise_std: float, std of Gaussian noise (default 0.003)
voiced_threshold: int, F0 threshold for U/V classification (default 0)
F0 < voiced_threshold will be set as unvoiced regions
flag_for_pulse: Bool, whether this SinGen is used inside PulseGen
(default False)
Note: when flag_for_pulse is True, the first time step of a voiced
segment is always sin(np.pi) or cos(0)
"""
def __init__(self, samp_rate, harmonic_num = 0, sine_amp = 0.1,
noise_std = 0.003, voiced_threshold = 0, flag_for_pulse=False):
super(SineGen, self).__init__()
self.sine_amp = sine_amp
self.noise_std = noise_std
self.harmonic_num = harmonic_num
self.dim = self.harmonic_num + 1
self.sampling_rate = samp_rate
self.voiced_threshold = voiced_threshold
self.flag_for_pulse = flag_for_pulse
return
def _f02uv(self, f0):
# generate uv signal
uv = torch.ones_like(f0)
uv = uv * (f0 > self.voiced_threshold)
return uv
def _f02sine(self, f0_values):
"""
input
-----
f0_values: (batchsize, length_in_time, dim)
where dim is the number of fundamental tone plus harmonic overtones
f0_values are supposed to be up-sampled. In other words, length should
be equal to the number of waveform sampling points.
output
------
sine_values: (batchsize, length_in_times, dim)
sine_values[i, :, k] is decided by the F0s in f0_values[i, :, k]
"""
# convert to F0 in rad. The interger part n can be ignored
# because 2 * np.pi * n doesn't affect phase
rad_values = (f0_values / self.sampling_rate) % 1
# initial phase noise (no noise for fundamental component)
rand_ini = torch.rand(f0_values.shape[0], f0_values.shape[2],\
device = f0_values.device)
rand_ini[:, 0] = 0
rad_values[:, 0, :] = rad_values[:, 0, :] + rand_ini
# instantanouse phase sine[t] = sin(2*pi \sum_i=1 ^{t} rad)
if not self.flag_for_pulse:
# for normal case
# To prevent torch.cumsum numerical overflow,
# it is necessary to add -1 whenever \sum_k=1^n rad_value_k > 1.
# Buffer tmp_over_one_idx indicates the time step to add -1.
# This will not change F0 of sine because (x-1) * 2*pi = x *2*pi
tmp_over_one = torch.cumsum(rad_values, 1) % 1
tmp_over_one_idx = (tmp_over_one[:, 1:, :] -
tmp_over_one[:, :-1, :]) < 0
cumsum_shift = torch.zeros_like(rad_values)
cumsum_shift[:, 1:, :] = tmp_over_one_idx * -1.0
sines = torch.sin(torch.cumsum(rad_values + cumsum_shift, dim=1)
* 2 * np.pi)
else:
# If necessary, make sure that the first time step of every
# voiced segments is sin(pi) or cos(0)
# This is used for pulse-train generation
# identify the last time step in unvoiced segments
uv = self._f02uv(f0_values)
uv_1 = torch.roll(uv, shifts=-1, dims=1)
uv_1[:, -1, :] = 1
u_loc = (uv < 1) * (uv_1 > 0)
# get the instantanouse phase
tmp_cumsum = torch.cumsum(rad_values, dim=1)
# different batch needs to be processed differently
for idx in range(f0_values.shape[0]):
temp_sum = tmp_cumsum[idx, u_loc[idx, :, 0], :]
temp_sum[1:, :] = temp_sum[1:, :] - temp_sum[0:-1, :]
# stores the accumulation of i.phase within
# each voiced segments
tmp_cumsum[idx, :, :] = 0
tmp_cumsum[idx, u_loc[idx, :, 0], :] = temp_sum
# rad_values - tmp_cumsum: remove the accumulation of i.phase
# within the previous voiced segment.
i_phase = torch.cumsum(rad_values - tmp_cumsum, dim=1)
# get the sines
sines = torch.cos(i_phase * 2 * np.pi)
return sines
def forward(self, f0):
""" sine_tensor, uv = forward(f0)
input
-----
F0: tensor, in shape (batchsize, length, dim=1)
up-sampled F0, length should be equal to the waveform length
Input F0 should be discontinuous.
F0 for unvoiced steps should be 0
output
------
sine_tensor: tensor, (batchsize, length, output_dim)
output uv: tensor, (batchsize, length, 1)
noise: tensor, (batchsize, length, 1)
note that output_dim = 1 + harmonic_num
"""
with torch.no_grad():
f0_buf = torch.zeros(f0.shape[0], f0.shape[1], self.dim, \
device=f0.device)
# fundamental component
f0_buf[:, :, 0] = f0[:, :, 0]
for idx in np.arange(self.harmonic_num):
# idx + 2: the (idx+1)-th overtone, (idx+2)-th harmonic
f0_buf[:, :, idx+1] = f0_buf[:, :, 0] * (idx+2)
# generate sine waveforms
sine_waves = self._f02sine(f0_buf) * self.sine_amp
# generate uv signal
#uv = torch.ones(f0.shape)
#uv = uv * (f0 > self.voiced_threshold)
uv = self._f02uv(f0)
# noise: for unvoiced should be similar to sine_amp
# std = self.sine_amp/3 -> max value ~ self.sine_amp
#. for voiced regions is self.noise_std
noise_amp = uv * self.noise_std + (1-uv) * self.sine_amp / 3
noise = noise_amp * torch.randn_like(sine_waves)
# first: set the unvoiced part to 0 by uv
# then: additive noise
sine_waves = sine_waves * uv + noise
return sine_waves, uv, noise
class PulseGen(torch_nn.Module):
""" Definition of Pulse train generator
There are many ways to implement pulse generator.
Here, PulseGen is based on SinGen.
This is used in cyclic-noise NSF
"""
def __init__(self, samp_rate, pulse_amp = 0.1,
noise_std = 0.003, voiced_threshold = 0):
super(PulseGen, self).__init__()
self.pulse_amp = pulse_amp
self.sampling_rate = samp_rate
self.voiced_threshold = voiced_threshold
self.noise_std = noise_std
self.l_sinegen = SineGen(self.sampling_rate, harmonic_num=0,\
sine_amp=self.pulse_amp, noise_std=0,\
voiced_threshold=self.voiced_threshold,\
flag_for_pulse=True)
def forward(self, f0):
""" Pulse train generator
pulse_train, uv = forward(f0)
input
-----
F0: tensor, (batchsize, length, dim=1)
up-sampled F0
f0 for unvoiced steps should be 0
length should be equal to the expected waveform length
output
------
pulse_train: tensor, (batchsize, length, dim)
sine_wave: tensor, (batchsize, length, dim), sine waveform that
is used to derive the pulse train
uv: tensor, (batchsize, length, 1), u/v flag
pulse_noise: tensor, (batchsize, length, dim), additive noise in
pulse_train
"""
with torch.no_grad():
sine_wav, uv, noise = self.l_sinegen(f0)
# sine without additive noise
pure_sine = sine_wav - noise
# step t corresponds to a pulse if
# sine[t] > sine[t+1] & sine[t] > sine[t-1]
# & sine[t-1], sine[t+1], and sine[t] are voiced
# or
# sine[t] is voiced, sine[t-1] is unvoiced
# we use torch.roll to simulate sine[t+1] and sine[t-1]
sine_1 = torch.roll(pure_sine, shifts=1, dims=1)
uv_1 = torch.roll(uv, shifts=1, dims=1)
uv_1[:, 0, :] = 0
sine_2 = torch.roll(pure_sine, shifts=-1, dims=1)
uv_2 = torch.roll(uv, shifts=-1, dims=1)
uv_2[:, -1, :] = 0
loc = (pure_sine > sine_1) * (pure_sine > sine_2) \
* (uv_1 > 0) * (uv_2 > 0) * (uv > 0) \
+ (uv_1 < 1) * (uv > 0)
# pulse train without noise
pulse_train = pure_sine * loc
# additive noise to pulse train
# note that noise from sinegen is zero in voiced regions
pulse_noise = torch.randn_like(pure_sine) * self.noise_std
# with additive noise on pulse, and unvoiced regions
pulse_train += pulse_noise * loc + pulse_noise * (1 - uv)
return pulse_train, sine_wav, uv, pulse_noise
if __name__ == "__main__":
print("Definition of major components in NSF")
| 13,906 | 37.311295 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/dist.py | #!/usr/bin/env python
"""
util_dist.py
Utilities for probabilistic distribution
Code adapted from
https://github.com/fatchord/WaveRNN/blob/master/utils/distribution.py
I haven't used these codes before.
To be moved to core_scripts/math_tools
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def log_sum_exp(x):
""" Adapted from
https://github.com/fatchord/WaveRNN/blob/master/utils/distribution.py
numerically stable log_sum_exp implementation that prevents overflow """
# TF ordering
axis = len(x.size()) - 1
m, _ = torch.max(x, dim=axis)
m2, _ = torch.max(x, dim=axis, keepdim=True)
return m + torch.log(torch.sum(torch.exp(x - m2), dim=axis))
def discretized_mix_logistic_loss(y_hat, y, num_classes=65536,
log_scale_min=None, reduce=True):
""" Adapted from
https://github.com/fatchord/WaveRNN/blob/master/utils/distribution.py
https://github.com/r9y9/wavenet_vocoder/blob/master/wavenet_vocoder/mixture.py
"""
if log_scale_min is None:
log_scale_min = float(np.log(1e-14))
y_hat = y_hat.permute(0,2,1)
assert y_hat.dim() == 3
assert y_hat.size(1) % 3 == 0
nr_mix = y_hat.size(1) // 3
# (B x T x C)
y_hat = y_hat.transpose(1, 2)
# unpack parameters. (B, T, num_mixtures) x 3
logit_probs = y_hat[:, :, :nr_mix]
means = y_hat[:, :, nr_mix:2 * nr_mix]
log_scales = torch.clamp(y_hat[:, :, 2 * nr_mix:3 * nr_mix],
min=log_scale_min)
# B x T x 1 -> B x T x num_mixtures
y = y.expand_as(means)
centered_y = y - means
inv_stdv = torch.exp(-log_scales)
plus_in = inv_stdv * (centered_y + 1. / (num_classes - 1))
cdf_plus = torch.sigmoid(plus_in)
min_in = inv_stdv * (centered_y - 1. / (num_classes - 1))
cdf_min = torch.sigmoid(min_in)
# log probability for edge case of 0 (before scaling)
# equivalent: torch.log(F.sigmoid(plus_in))
log_cdf_plus = plus_in - F.softplus(plus_in)
# log probability for edge case of 255 (before scaling)
# equivalent: (1 - F.sigmoid(min_in)).log()
log_one_minus_cdf_min = -F.softplus(min_in)
# probability for all other cases
cdf_delta = cdf_plus - cdf_min
mid_in = inv_stdv * centered_y
# log probability in the center of the bin, to be used in extreme cases
# (not actually used in our code)
log_pdf_mid = mid_in - log_scales - 2. * F.softplus(mid_in)
# tf equivalent
"""
log_probs = tf.where(x < -0.999, log_cdf_plus,
tf.where(x > 0.999, log_one_minus_cdf_min,
tf.where(cdf_delta > 1e-5,
tf.log(tf.maximum(cdf_delta, 1e-12)),
log_pdf_mid - np.log(127.5))))
"""
# TODO: cdf_delta <= 1e-5 actually can happen. How can we choose the value
# for num_classes=65536 case? 1e-7? not sure..
inner_inner_cond = (cdf_delta > 1e-5).float()
inner_inner_out = inner_inner_cond * \
torch.log(torch.clamp(cdf_delta, min=1e-12)) + \
(1. - inner_inner_cond) * (log_pdf_mid - np.log((num_classes - 1) / 2))
inner_cond = (y > 0.999).float()
inner_out = inner_cond * log_one_minus_cdf_min + \
(1. - inner_cond) * inner_inner_out
cond = (y < -0.999).float()
log_probs = cond * log_cdf_plus + (1. - cond) * inner_out
log_probs = log_probs + F.log_softmax(logit_probs, -1)
if reduce:
return -torch.mean(log_sum_exp(log_probs))
else:
return -log_sum_exp(log_probs).unsqueeze(-1)
def sample_from_discretized_mix_logistic(y, log_scale_min=None):
"""
Sample from discretized mixture of logistic distributions
Args:
y (Tensor): B x C x T
log_scale_min (float): Log scale minimum value
Returns:
Tensor: sample in range of [-1, 1].
"""
if log_scale_min is None:
log_scale_min = float(np.log(1e-14))
assert y.size(1) % 3 == 0
nr_mix = y.size(1) // 3
# B x T x C
y = y.transpose(1, 2)
logit_probs = y[:, :, :nr_mix]
# sample mixture indicator from softmax
temp = logit_probs.data.new(logit_probs.size()).uniform_(1e-5, 1.0 - 1e-5)
temp = logit_probs.data - torch.log(- torch.log(temp))
_, argmax = temp.max(dim=-1)
# (B, T) -> (B, T, nr_mix)
one_hot = F.one_hot(argmax, nr_mix).float()
# select logistic parameters
means = torch.sum(y[:, :, nr_mix:2 * nr_mix] * one_hot, dim=-1)
log_scales = torch.clamp(torch.sum(
y[:, :, 2 * nr_mix:3 * nr_mix] * one_hot, dim=-1), min=log_scale_min)
# sample from logistic & clip to interval
# we don't actually round to the nearest 8bit value when sampling
u = means.data.new(means.size()).uniform_(1e-5, 1.0 - 1e-5)
x = means + torch.exp(log_scales) * (torch.log(u) - torch.log(1. - u))
x = torch.clamp(torch.clamp(x, min=-1.), max=1.)
return x
if __name__ == "__main__":
print("dist")
| 5,244 | 32.196203 | 82 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_rawnet.py | #!/usr/bin/env python
"""
This file contains code for RawNet2
Hemlata Tak, Jose Patino, Massimiliano Todisco, Andreas Nautsch,
Nicholas Evans, and Anthony Larcher. End-to-End Anti-Spoofing with RawNet2.
In Proc. ICASSP, 6369--6373. 2020.
Implementation based on RawNet in
https://github.com/asvspoof-challenge/2021/
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torchaudio
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import core_scripts.other_tools.debug as nii_debug
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
class SincConv2(torch_nn.Module):
"""
"""
@staticmethod
def to_mel(hz):
return 2595 * np.log10(1 + hz / 700)
@staticmethod
def to_hz(mel):
return 700 * (10 ** (mel / 2595) - 1)
def __init__(self, num_filters, kernel_size, in_channels=1,
sample_rate = 16000, num_freq_bin = 257,
stride = 1, dilation = 1,
flag_pad = True, flag_trainable=False):
"""
SincConv2(num_filters, kernel_size, in_channels=1,
sample_rate = 16000, num_freq_bins = 257,
stride = 1, dilation = 1,
flag_pad = True, flag_trainable=False)
Args
----
num_filters: int, number of sinc-filters
kernel_size: int, length of each sinc-filter
in_channels: int, dimension of input signal,
(batchsize, length, in_channels)
sample_rate: int, sampling rate
num_freq_bin: number of frequency bins, not really important
here. Default 257
stride: int, stride of convoluiton, default 1
dilation: int, dilaion of conv, default 1
flag_pad: bool, whether pad the sequence to make input and
output have equal length, default True
flag_trainable: bool, whether the filter is trainable
default False
Num_filters and in_channels decide the output tensor dimension
If input is (batchsize, length, in_channels), output will be
(batchsize, length, in_channels * num_filters)
This is done through depwise convolution,
https://pytorch.org/docs/stable/generated/torch.nn.Conv1d.html
i.e., each input dimension will go through all the num_filters.
"""
super(SincConv2,self).__init__()
self.m_out_channels = num_filters
self.m_in_channels = in_channels
self.m_sample_rate=sample_rate
# Forcing the filters to be odd (i.e, perfectly symmetrics)
self.m_kernel_size = kernel_size
if kernel_size % 2 == 0:
self.m_kernel_size = self.m_kernel_size + 1
self.m_stride = stride
self.m_dilation = dilation
# Pad to original length
if flag_pad:
self.m_padding = dilation * (self.m_kernel_size - 1) + 1 - stride
if stride % 2 == 0:
print("Warning: padding in SincCov is not perfect because of")
print("stride {:d}".format(stride))
self.m_padding = self.m_padding // 2
else:
self.m_padding = 0
# initialize filterbanks using Mel scale
f = int(self.m_sample_rate / 2) * np.linspace(0, 1, num_freq_bin)
# Hz to mel conversion
fmel = self.to_mel(f)
fmelmax = np.max(fmel)
fmelmin = np.min(fmel)
filbandwidthsmel = np.linspace(fmelmin, fmelmax, self.m_out_channels+1)
# Mel to Hz conversion
filbandwidthsf = self.to_hz(filbandwidthsmel)
# mel band
self.m_mel = filbandwidthsf
# time index
self.m_hsupp = torch.arange(-(self.m_kernel_size-1)/2,
(self.m_kernel_size-1)/2+1)
# filter coeffs
self.m_filters = torch.zeros(self.m_out_channels, self.m_kernel_size)
# create filter coefficient
for i in range(self.m_out_channels):
fmin = self.m_mel[i]
fmax = self.m_mel[i+1]
hHigh = np.sinc(2 * fmax * self.m_hsupp / self.m_sample_rate)
hHigh = (2 * fmax / self.m_sample_rate) * hHigh
hLow = np.sinc(2 * fmin * self.m_hsupp / self.m_sample_rate)
hLow = (2 * fmin / self.m_sample_rate) * hLow
# band pass filters
hideal = hHigh - hLow
# applying windowing
self.m_filters[i,:] = torch.tensor(
np.hamming(self.m_kernel_size) * hideal)
# repeat to (output_channels * in_channels)
self.m_filters = self.m_filters.repeat(self.m_in_channels, 1)
# save as model parameter
self.m_filters = self.m_filters.view(
self.m_out_channels * self.m_in_channels, 1, self.m_kernel_size)
self.m_filters = torch_nn.Parameter(
self.m_filters, requires_grad=flag_trainable)
return
def forward(self,x):
"""SincConv(x)
input
-----
x: tensor, shape (batchsize, length, feat_dim)
output
------
y: tensor, shape (batchsize, length, output-channel)
"""
return torch_nn_func.conv1d(
x.permute(0, 2, 1), self.m_filters, stride=self.m_stride,
padding=self.m_padding, dilation=self.m_dilation,
bias=None, groups=x.shape[-1]).permute(0, 2, 1)
class FMS(torch_nn.Module):
"""filter-wise feature map scaling
Hemlata Tak, Jose Patino, Massimiliano Todisco, Andreas Nautsch,
Nicholas Evans, and Anthony Larcher.
End-to-End Anti-Spoofing with RawNet2.
In Proc. ICASSP, 6369--6373. 2020.
Example:
l_fms = FMS(5)
with torch.no_grad():
data = torch.randn(2, 1000, 5)
out = l_fms(data)
"""
def __init__(self, feat_dim):
"""FMS(feat_dim)
Args
----
feat_dim: int, dimension of input, in shape (batch, length, dim)
"""
super(FMS, self).__init__()
self.m_dim = feat_dim
self.m_pooling = torch_nn.AdaptiveAvgPool1d(1)
self.m_dim_change = torch_nn.Linear(feat_dim, feat_dim)
self.m_act = torch_nn.Sigmoid()
return
def forward(self, x):
"""FMS(x)
input
-----
x: tensor, (batch, length, dim)
output
-----
y: tensor, (batch, length, dim)
"""
if x.shape[-1] != self.m_dim:
print("FMS expects data of dim {:d}".format(self.m_dim))
sys.exit(1)
# pooling expects (batch, dim, length)
# y will be (batch, dim, 1)
y = self.m_pooling(x.permute(0, 2, 1))
# squeeze to (batch, dim), unsqueeze to (batch, 1, dim, )
y = self.m_act(self.m_dim_change(y.squeeze(-1))).unsqueeze(1)
# scaling and shifting
return (x * y + y)
class Residual_block(torch_nn.Module):
"""Residual block used in RawNet2 for Anti-spoofing
"""
def __init__(self, nb_filts, flag_bn_input = False):
"""Residual_block(bn_filts, flga_bn_input)
Args
----
bn_filts: list of int, [input_channel, output_channel]
flag_bn_input: bool, whether do BatchNorm and LReLU
default False
"""
super(Residual_block, self).__init__()
# whether batch normalize input
if flag_bn_input:
self.bn1 = torch_nn.Sequential(
torch_nn.BatchNorm1d(num_features = nb_filts[0]),
torch_nn.LeakyReLU(negative_slope=0.3))
else:
self.bn1 = None
self.conv = torch_nn.Sequential(
torch_nn.Conv1d(in_channels = nb_filts[0],
out_channels = nb_filts[1],
kernel_size = 3,
padding = 1,
stride = 1),
torch_nn.BatchNorm1d(num_features = nb_filts[1]),
torch_nn.Conv1d(in_channels = nb_filts[1],
out_channels = nb_filts[1],
padding = 1,
kernel_size = 3,
stride = 1)
)
# for dimension change
if nb_filts[0] != nb_filts[1]:
self.dim_change = torch_nn.Conv1d(
in_channels = nb_filts[0],
out_channels = nb_filts[1],
padding = 0,
kernel_size = 1,
stride = 1)
else:
self.dim_change = None
# maxpooling
self.mp = torch_nn.MaxPool1d(3)
return
def forward(self, x):
""" y= Residual_block(x)
input
-----
x: tensor, (batchsize, length, dim)
output
------
y: tensor, (batchsize, length, dim2)
"""
identity = x.permute(0, 2, 1)
if self.bn1 is None:
out = x.permute(0, 2, 1)
else:
out = self.bn1(x.permute(0, 2, 1))
out = self.conv(out)
if self.dim_change is not None:
identity = self.dim_change(identity)
out += identity
out = self.mp(out)
return out.permute(0, 2, 1)
class RawNet(torch_nn.Module):
"""RawNet based on
https://github.com/asvspoof-challenge/2021/
"""
def __init__(self, num_sinc_filter, sinc_filter_len, in_dim, sampling_rate,
res_ch_1, res_ch_2, gru_node, gru_layer, emb_dim, num_class):
super(RawNet, self).__init__()
# sinc filter layer
self.m_sinc_conv = SincConv2(
num_sinc_filter,
kernel_size = sinc_filter_len,
in_channels = in_dim,
sample_rate = sampling_rate,
flag_pad = False,
flag_trainable=False)
# res block group
self.m_resgroup = torch_nn.Sequential(
nii_nn.BatchNorm1DWrapper(num_sinc_filter),
torch_nn.SELU(),
Residual_block([num_sinc_filter, res_ch_1], flag_bn_input=False),
FMS(res_ch_1),
Residual_block([res_ch_1, res_ch_1], flag_bn_input=True),
FMS(res_ch_1),
Residual_block([res_ch_1, res_ch_2], flag_bn_input=True),
FMS(res_ch_2),
Residual_block([res_ch_2, res_ch_2], flag_bn_input=True),
FMS(res_ch_2),
Residual_block([res_ch_2, res_ch_2], flag_bn_input=True),
FMS(res_ch_2),
Residual_block([res_ch_2, res_ch_2], flag_bn_input=True),
FMS(res_ch_2),
)
# GRU part
self.m_before_gru = torch_nn.Sequential(
nii_nn.BatchNorm1DWrapper(res_ch_2),
torch_nn.SELU()
)
self.m_gru = torch_nn.GRU(input_size = res_ch_2,
hidden_size = gru_node,
num_layers = gru_layer,
batch_first = True)
self.m_emb = torch_nn.Linear(in_features = gru_node,
out_features = emb_dim)
# output score
self.m_output = torch_nn.Linear(in_features = emb_dim,
out_features = num_class,
bias=True)
#
self.logsoftmax = torch_nn.LogSoftmax(dim=1)
return
def _compute_embedding(self, x):
"""
input
-----
x: tensor, (batch, length, dim)
output
------
y: tensor, (batch, emb_dim)
"""
batch, length, dim = x.shape
#
x = self.m_sinc_conv(x)
x = self.m_resgroup(x)
x, _ = self.m_gru(self.m_before_gru(x))
return self.m_emb(x[:, -1, :])
def _compute_score(self, emb, inference=True):
"""
input
-----
emb: tensor, (batch, emb_dim)
output
------
score: tensor, (batch, num_class)
Score here refers to
"""
# we should not use logsoftmax if we will use CrossEntropyLoss
flag_logsoftmax = False
if inference:
# no softmax
return self.m_output(emb)
elif flag_logsoftmax:
# Logsoftmax for training loss
# this is used when the training criterion is NLLoss
return self.logsoftmax(self.m_output(emb))
else:
return self.m_output(emb)
def forward(self, x):
"""
input
-----
x: tensor, (batch, length, dim)
output
------
y: tensor, (batch, num_class)
y is the log-probablity after softmax
"""
emb = self._compute_embedding(x)
return self._compute_score(emb, inference=False)
def inference(self, x):
"""
input
-----
x: tensor, (batch, length, dim)
output
------
y: tensor, (batch, num_class)
y is the input activation to softmax
"""
emb = self._compute_embedding(x)
return self._compute_score(emb, inference=True)
if __name__ == "__main__":
print("Definition of RawNet2")
| 13,819 | 31.28972 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/eval_music.py | #!/usr/bin/env python
"""
Functions for evaluation - music data
This is just wrappers around MIR_eval http://craffel.github.io/mir_eval/
"""
from __future__ import print_function
import os
import sys
import numpy as np
try:
import mir_eval
except ModuleNotFoundError:
print("Please install mir_eval: http://craffel.github.io/mir_eval/")
print("Anaconda: https://anaconda.org/conda-forge/mir_eval")
print("Pip: https://pypi.org/project/mir_eval/")
sys.exit(1)
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
###############
# Functions to evaluation melody accuracy
def eva_music(est_f0, ref_f0, resolution=0.012, tolerence=50):
""" Evaluating estimated F0 using mir_eval metrics.
Parameters
----------
est_f0 : np.array, in shape (N)
estimated F0 (Hz)
ref_f0 : np.array, in shape (N)
reference F0 (Hz)
resolution : float (s)
corresponding to frame-shift of the F0
tolerence : int
tolerence for cent evaluation
Returns
-------
rpa : float
raw pitch accuracy
rca : float
raw chroma accuracy
"""
# generate time sequences
def _time_seq(f0_seq, reso):
time_seq = np.arange(f0_seq.shape[0]) * reso
return time_seq
est_time = _time_seq(est_f0, resolution)
ref_time = _time_seq(ref_f0, resolution)
# format change
(ref_v, ref_c, est_v, est_c) = mir_eval.melody.to_cent_voicing(
ref_time, ref_f0, est_time, est_f0)
# evaluation
rpa = mir_eval.melody.raw_pitch_accuracy(ref_v, ref_c, est_v, est_c)
rca = mir_eval.melody.raw_chroma_accuracy(ref_v, ref_c, est_v, est_c)
return rpa, rca
| 1,737 | 24.558824 | 73 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/__init__.py | 0 | 0 | 0 | py |
|
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/util_frontend.py | #!/usr/bin/env python
"""
util_frontend.py
Utilities for frontend feature extraction
It includes:
LFCC: based on asvspoof.org baseline matlab code
LFB: Linear filterbank feature
Chen, T., Kumar, A., Nagarsheth, P., Sivaraman, G. & Khoury, E.
Generalization of Audio Deepfake Detection. in Proc. Odyssey 132-137
(2020). doi:10.21437/Odyssey.2020-19
According to the author's email:
LFB = np.log(FilterBank(Amplitude(STFT(x))))
There is no DCT. But it does have logarithm.
Implemented based on LFCC API
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.util_dsp as nii_dsp
import core_scripts.data_io.conf as nii_conf
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
##################
## other utilities
##################
def stft_wrapper(x, fft_n, frame_shift, frame_length, window,
pad_mode="constant", return_complex=False):
"""Due to the different signature of torch.stft, write a
wrapper to handle this
input
-----
x: tensor, waveform, (batch, length)
window: tensor, window coef, (frame_length, )
output
------
tensor: (batch, frame_num, bin_num, 2)
"""
# there are better ways, but for convenience
if torch.__version__.split('.')[1].isnumeric() and \
int(torch.__version__.split('.')[1]) < 7:
# torch 1.6.*
return torch.stft(x, fft_n, frame_shift, frame_length,
window=window, onesided=True, pad_mode=pad_mode)
else:
# torch > 1.7
return torch.stft(x, fft_n, frame_shift, frame_length,
window=window, onesided=True, pad_mode=pad_mode,
return_complex=return_complex)
def istft_wrapper(x, fft_n, frame_shift, frame_length, window,
pad_mode="constant"):
# there are better ways, but for convenience
if torch.__version__.split('.')[1].isnumeric() and \
int(torch.__version__.split('.')[1]) < 7:
# torch 1.6.*
return torch.istft(x, fft_n, frame_shift, frame_length,
window=window, onesided=True, pad_mode=pad_mode)
else:
# torch > 1.7
return torch.istft(x, fft_n, frame_shift, frame_length,
window=window, onesided=True)
def trimf(x, params):
"""
trimf: similar to Matlab definition
https://www.mathworks.com/help/fuzzy/trimf.html?s_tid=srchtitle
"""
if len(params) != 3:
print("trimp requires params to be a list of 3 elements")
sys.exit(1)
a = params[0]
b = params[1]
c = params[2]
if a > b or b > c:
print("trimp(x, [a, b, c]) requires a<=b<=c")
sys.exit(1)
y = torch.zeros_like(x, dtype=nii_conf.d_dtype)
if a < b:
index = torch.logical_and(a < x, x < b)
y[index] = (x[index] - a) / (b - a)
if b < c:
index = torch.logical_and(b < x, x < c)
y[index] = (c - x[index]) / (c - b)
y[x == b] = 1
return y
def delta(x):
""" By default
input
-----
x (batch, Length, dim)
output
------
output (batch, Length, dim)
Delta is calculated along Length dimension
"""
length = x.shape[1]
output = torch.zeros_like(x)
x_temp = torch_nn_func.pad(x.unsqueeze(1), (0, 0, 1, 1),
'replicate').squeeze(1)
output = -1 * x_temp[:, 0:length] + x_temp[:,2:]
return output
def linear_fb(fn, sr, filter_num):
"""linear_fb(fn, sr, filter_num)
create linear filter bank based on trim
input
-----
fn: int, FFT points
sr: int, sampling rate (Hz)
filter_num: int, number of filters in filter-bank
output
------
fb: tensor, (fn//2+1, filter_num)
Note that this filter bank is supposed to be used on
spectrum of dimension fn//2+1.
See example in LFCC.
"""
# build the triangle filter bank
f = (sr / 2) * torch.linspace(0, 1, fn//2+1)
filter_bands = torch.linspace(min(f), max(f), filter_num+2)
filter_bank = torch.zeros([fn//2+1, filter_num])
for idx in range(filter_num):
filter_bank[:, idx] = trimf(
f, [filter_bands[idx],
filter_bands[idx+1],
filter_bands[idx+2]])
return filter_bank
#################
## LFCC front-end
#################
class LFCC(torch_nn.Module):
""" Based on asvspoof.org baseline Matlab code.
Difference: with_energy is added to set the first dimension as energy
"""
def __init__(self, fl, fs, fn, sr, filter_num,
with_energy=False, with_emphasis=True,
with_delta=True, flag_for_LFB=False,
num_coef=None, min_freq=0, max_freq=1):
""" Initialize LFCC
Para:
-----
fl: int, frame length, (number of waveform points)
fs: int, frame shift, (number of waveform points)
fn: int, FFT points
sr: int, sampling rate (Hz)
filter_num: int, number of filters in filter-bank
with_energy: bool, (default False), whether replace 1st dim to energy
with_emphasis: bool, (default True), whether pre-emphaze input wav
with_delta: bool, (default True), whether use delta and delta-delta
for_LFB: bool (default False), reserved for LFB feature
num_coef: int or None, number of coeffs to be taken from filter bank.
Note that this is only used for LFCC, i.e., for_LFB=False
When None, num_coef will be equal to filter_num
min_freq: float (default 0), min_freq * sr // 2 will be the minimum
frequency of extracted FFT spectrum
max_freq: float (default 1), max_freq * sr // 2 will be the maximum
frequency of extracted FFT spectrum
"""
super(LFCC, self).__init__()
self.fl = fl
self.fs = fs
self.fn = fn
self.sr = sr
self.filter_num = filter_num
self.num_coef = num_coef
# decide the range of frequency bins
if min_freq >= 0 and min_freq < max_freq and max_freq <= 1:
self.min_freq_bin = int(min_freq * (fn//2+1))
self.max_freq_bin = int(max_freq * (fn//2+1))
self.num_fft_bins = self.max_freq_bin - self.min_freq_bin
else:
print("LFCC cannot work with min_freq {:f} and max_freq {:}".format(
min_freq, max_freq))
sys.exit(1)
# build the triangle filter bank
f = (sr / 2) * torch.linspace(min_freq, max_freq, self.num_fft_bins)
filter_bands = torch.linspace(min(f), max(f), filter_num+2)
filter_bank = torch.zeros([self.num_fft_bins, filter_num])
for idx in range(filter_num):
filter_bank[:, idx] = trimf(
f, [filter_bands[idx],
filter_bands[idx+1],
filter_bands[idx+2]])
self.lfcc_fb = torch_nn.Parameter(filter_bank, requires_grad=False)
# DCT as a linear transformation layer
self.l_dct = nii_dsp.LinearDCT(filter_num, 'dct', norm='ortho')
# opts
self.with_energy = with_energy
self.with_emphasis = with_emphasis
self.with_delta = with_delta
self.flag_for_LFB = flag_for_LFB
if self.num_coef is None:
self.num_coef = filter_num
# Add a buf to store window coefficients
#
self.window_buf = None
return
def forward(self, x):
"""
input:
------
x: tensor(batch, length), where length is waveform length
output:
-------
lfcc_output: tensor(batch, frame_num, dim_num)
"""
# pre-emphsis
if self.with_emphasis:
# to avoid side effect
x_copy = torch.zeros_like(x) + x
x_copy[:, 1:] = x[:, 1:] - 0.97 * x[:, 0:-1]
else:
x_copy = x
if self.window_buf is None:
self.window_buf = torch.hamming_window(self.fl).to(x.device)
# STFT
#x_stft = torch.stft(x_copy, self.fn, self.fs, self.fl,
# window=torch.hamming_window(self.fl).to(x.device),
# onesided=True, pad_mode="constant")
x_stft = stft_wrapper(x_copy, self.fn, self.fs, self.fl,self.window_buf)
# amplitude
sp_amp = torch.norm(x_stft, 2, -1).pow(2).permute(0, 2, 1).contiguous()
if self.min_freq_bin > 0 or self.max_freq_bin < (self.fn//2+1):
sp_amp = sp_amp[:, :, self.min_freq_bin:self.max_freq_bin]
# filter bank
fb_feature = torch.log10(torch.matmul(sp_amp, self.lfcc_fb) +
torch.finfo(torch.float32).eps)
# DCT (if necessary, remove DCT)
lfcc = self.l_dct(fb_feature) if not self.flag_for_LFB else fb_feature
# Truncate the output of l_dct when necessary
if not self.flag_for_LFB and self.num_coef != self.filter_num:
lfcc = lfcc[:, :, :self.num_coef]
# Add energy
if self.with_energy:
power_spec = sp_amp / self.fn
energy = torch.log10(power_spec.sum(axis=2)+
torch.finfo(torch.float32).eps)
lfcc[:, :, 0] = energy
# Add delta coefficients
if self.with_delta:
lfcc_delta = delta(lfcc)
lfcc_delta_delta = delta(lfcc_delta)
lfcc_output = torch.cat((lfcc, lfcc_delta, lfcc_delta_delta), 2)
else:
lfcc_output = lfcc
# done
return lfcc_output
#################
## LFB front-end
#################
class LFB(LFCC):
""" Linear filterbank feature
Chen, T., Kumar, A., Nagarsheth, P., Sivaraman, G. & Khoury, E.
Generalization of Audio Deepfake Detection. in Proc. Odyssey 132-137
(2020). doi:10.21437/Odyssey.2020-19
"""
def __init__(self, fl, fs, fn, sr, filter_num,
with_energy=False, with_emphasis=True,
with_delta=False):
""" Initialize LFB
Para:
-----
fl: int, frame length, (number of waveform points)
fs: int, frame shift, (number of waveform points)
fn: int, FFT points
sr: int, sampling rate (Hz)
filter_num: int, number of filters in filter-bank
with_energy: bool, (default False), whether replace 1st dim to energy
with_emphasis: bool, (default True), whether pre-emphaze input wav
with_delta: bool, (default True), whether use delta and delta-delta
"""
super(LFB, self).__init__(fl, fs, fn, sr, filter_num, with_energy,
with_emphasis, with_delta, flag_for_LFB=True)
return
def forward(self, x):
"""
input:
------
x: tensor(batch, length), where length is waveform length
output:
-------
lfb_output: tensor(batch, frame_num, dim_num)
"""
return super(LFB, self).forward(x)
#################
## Spectrogram (FFT) front-end
#################
class Spectrogram(torch_nn.Module):
""" Spectrogram front-end
"""
def __init__(self, fl, fs, fn, sr,
with_emphasis=True, with_delta=False, in_db=False):
""" Initialize LFCC
Para:
-----
fl: int, frame length, (number of waveform points)
fs: int, frame shift, (number of waveform points)
fn: int, FFT points
sr: int, sampling rate (Hz)
with_emphasis: bool, (default True), whether pre-emphaze input wav
with_delta: bool, (default False), whether use delta and delta-delta
in_db: bool, (default False), use 20log10(amp)? if False, use amp
"""
super(Spectrogram, self).__init__()
self.fl = fl
self.fs = fs
self.fn = fn
self.sr = sr
# opts
self.with_emphasis = with_emphasis
self.with_delta = with_delta
self.in_db = in_db
# buf to store window coefficients
self.window_buf = None
return
def forward(self, x):
"""
input:
------
x: tensor(batch, length), where length is waveform length
output:
-------
lfcc_output: tensor(batch, frame_num, dim_num)
"""
# pre-emphsis
if self.with_emphasis:
x[:, 1:] = x[:, 1:] - 0.97 * x[:, 0:-1]
if self.window_buf is None:
self.window_buf = torch.hamming_window(self.fl).to(x.device)
# STFT
#x_stft = torch.stft(x, self.fn, self.fs, self.fl,
# window=torch.hamming_window(self.fl).to(x.device),
# onesided=True, pad_mode="constant")
x_stft = stft_wrapper(x, self.fn, self.fs, self.fl, self.window_buf)
# amplitude
sp_amp = torch.norm(x_stft, 2, -1).pow(2).permute(0, 2, 1).contiguous()
if self.in_db:
sp_amp = torch.log10(sp_amp + torch.finfo(torch.float32).eps)
# Add delta coefficients
if self.with_delta:
sp_delta = delta(sp_amp)
sp_delta_delta = delta(sp_delta)
sp_output = torch.cat((sp_amp, sp_delta, sp_delta_delta), 2)
else:
sp_output = sp_amp
# done
return sp_amp
#################
## MFCC front-end
#################
from core_scripts.data_io import dsp_tools
class MFCC(torch_nn.Module):
""" Based on asvspoof.org baseline Matlab code.
Difference: with_energy is added to set the first dimension as energy
"""
def __init__(self, fl, fs, fn, sr, filter_num,
with_energy=False, with_emphasis=True,
with_delta=True, flag_for_MelSpec=False,
num_coef=None, min_freq=0, max_freq=1):
""" Initialize LFCC
Para:
-----
fl: int, frame length, (number of waveform points)
fs: int, frame shift, (number of waveform points)
fn: int, FFT points
sr: int, sampling rate (Hz)
filter_num: int, number of filters in filter-bank
with_energy: bool, (default False), whether replace 1st dim to energy
with_emphasis: bool, (default True), whether pre-emphaze input wav
with_delta: bool, (default True), whether use delta and delta-delta
flag_for_MelSpec: bool (default False), reserved for LFB feature
num_coef: int or None, number of coeffs to be taken from filter bank.
Note that this is only used for LFCC, i.e., for_LFB=False
When None, num_coef will be equal to filter_num
min_freq: float (default 0), min_freq * sr // 2 will be the minimum
frequency of extracted FFT spectrum
max_freq: float (default 1), max_freq * sr // 2 will be the maximum
frequency of extracted FFT spectrum
"""
super(MFCC, self).__init__()
self.fl = fl
self.fs = fs
self.fn = fn
self.sr = sr
self.filter_num = filter_num
self.num_coef = num_coef
# decide the range of frequency bins
if min_freq >= 0 and min_freq < max_freq and max_freq <= 1:
pass
else:
print("MFCC cannot work with min_freq {:f} and max_freq {:}".format(
min_freq, max_freq))
sys.exit(1)
# opts
self.with_energy = with_energy
self.with_emphasis = with_emphasis
self.with_delta = with_delta
self.flag_for_MelSpec = flag_for_MelSpec
if self.num_coef is None:
self.num_coef = filter_num
# get filter bank
tmp_config = dsp_tools.Melspec(sr, fl, fs, fn, filter_num,
sr/2*min_freq, sr/2*max_freq)
filter_bank = torch.tensor(tmp_config.melfb.T, dtype=nii_conf.d_dtype)
self.mel_fb = torch_nn.Parameter(filter_bank, requires_grad=False)
# DCT as a linear transformation layer
if not self.flag_for_MelSpec:
self.l_dct = nii_dsp.LinearDCT(filter_num, 'dct', norm='ortho')
else:
self.l_dct = None
# Add a buf to store window coefficients
#
self.window_buf = None
return
def forward(self, x):
"""
input:
------
x: tensor(batch, length), where length is waveform length
output:
-------
lfcc_output: tensor(batch, frame_num, dim_num)
"""
# pre-emphsis
if self.with_emphasis:
# to avoid side effect
x_copy = torch.zeros_like(x) + x
x_copy[:, 1:] = x[:, 1:] - 0.97 * x[:, 0:-1]
else:
x_copy = x
if self.window_buf is None:
self.window_buf = torch.hamming_window(self.fl).to(x.device)
# STFT
x_stft = stft_wrapper(x_copy,
self.fn, self.fs, self.fl, self.window_buf)
# amplitude
sp_amp = torch.norm(x_stft, 2, -1).pow(2).permute(0, 2, 1).contiguous()
# filter bank
fb_feature = torch.log10(torch.matmul(sp_amp, self.mel_fb) +
torch.finfo(torch.float32).eps)
# DCT (if necessary, remove DCT)
if not self.flag_for_MelSpec:
output = self.l_dct(fb_feature)
else:
output = fb_feature
# Truncate the output of l_dct when necessary
if not self.flag_for_MelSpec and self.num_coef != self.filter_num:
output = output[:, :, :self.num_coef]
# Add energy
if self.with_energy:
power_spec = sp_amp / self.fn
energy = torch.log10(power_spec.sum(axis=2)+
torch.finfo(torch.float32).eps)
output[:, :, 0] = energy
# Add delta coefficients
if self.with_delta:
output_delta = delta(output)
output_delta_delta = delta(output_delta)
output = torch.cat((output, output_delta, output_delta_delta), 2)
else:
pass
# done
return output
#######################
# spectrum substraction
#######################
def spectral_substraction(input_wav, noise_wav, ratio = 0.1,
fft_n = 512, frame_shift = 256, frame_length = 512):
"""
output = spectrum_substraction(input_wav, noise_wav)
input
-----
input_wav: tensor, (batch, length1, 1)
noise_wav: tensor, (batch, length2, 1)
ratio: float, default 0.1, ratio to be multiplied with noise spectrum
fft_n: int, default 1024, fft length
frame_shift: int, default 256, frame shift
frame_length: int, default 512, frame_shift
output
------
output: tensor, de-noised waveform
Note: this function do spectral_substraction.
noise_wav does not need to have the same length as input_wav.
spectra amplitude of noise_wav will be averaged and subtracted from
input_wav stft spectra amplitude
"""
window = torch.hamming_window(frame_length).to(input_wav.device)
# stft
input_spec = stft_wrapper(
input_wav.squeeze(-1), fft_n, frame_shift, frame_length, window,
return_complex=True)
noise_spec = stft_wrapper(
noise_wav.squeeze(-1), fft_n, frame_shift, frame_length, window,
return_complex=True)
# input wave spectral amplitude and phase (batch, fft_n//2+1, length, )
input_spec_abs = torch.abs(input_spec)
input_spec_pha = torch.angle(input_spec)
# noise spectral, averaged
noise_spec_abs = torch.abs(noise_spec).mean(axis=-1).unsqueeze(-1)
# spectral subtraction
denoi_spec_abs = input_spec_abs - noise_spec_abs * ratio
denoi_spec_abs = torch.clamp(denoi_spec_abs, min=0.00000001)
# recover waveform
input_spec = torch.complex(
denoi_spec_abs * torch.cos(input_spec_pha),
denoi_spec_abs * torch.sin(input_spec_pha),
)
output = istft_wrapper(
input_spec, fft_n, frame_shift, frame_length, window)
# adjust waveform length
length = min([input_wav.shape[1], output.shape[1]])
output_new = torch.zeros_like(input_wav)
output_new[:, 0:length, 0] = output[:, 0:length]
return output_new
if __name__ == "__main__":
print("Definition of front-end for Anti-spoofing")
| 21,157 | 32.163009 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_waveglow.py | #!/usr/bin/env python
"""
Building blocks for waveglow
"""
from __future__ import absolute_import
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import torch.nn.init as torch_init
import sandbox.block_nn as nii_nn
import sandbox.block_wavenet as nii_wavenet
import sandbox.block_glow as nii_glow
import core_scripts.data_io.conf as nii_io_conf
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
class Invertible1x1ConvWaveGlow(torch.nn.Module):
def __init__(self, feat_dim, flag_detjac=False):
super(Invertible1x1ConvWaveGlow, self).__init__()
torch.manual_seed(100)
with torch.no_grad():
W = torch.qr(torch.FloatTensor(feat_dim, feat_dim).normal_())[0]
# Ensure determinant is 1.0 not -1.0
if torch.det(W) < 0:
W[:,0] = -1*W[:,0]
# not necessary
W = W.transpose(0, 1)
self.weight = torch_nn.Parameter(W)
self.weight_inv = torch_nn.Parameter(W.clone())
self.weight_inv_flag = False
self.flag_detjac = flag_detjac
return
def forward(self, y, factor):
batch_size, length, feat_dim = y.size()
# Forward computation
log_det_W = length / factor * torch.logdet(self.weight)
z = torch.matmul(y, self.weight)
if self.flag_detjac:
return z, log_det_W
else:
return z
def reverse(self, x):
if not self.weight_inv_flag:
self.weight_inv.data = torch.inverse(self.weight.data)
self.weight_inv_flag = True
return torch.matmul(x, self.weight_inv)
class upsampleByTransConv(torch_nn.Module):
"""upsampleByTransConv
Upsampling layer using transposed convolution
"""
def __init__(self, feat_dim, upsample_rate, window_ratio=5):
"""upsampleByTransConv(feat_dim, upsample_rate, window_ratio=5)
Args
----
feat_dim: int, input feature should be (batch, length, feat_dim)
upsample_rate, int, output feature will be
(batch, length*upsample_rate, feat_dim)
window_ratio: int, default 5, window length of transconv will be
upsample_rate * window_ratio
"""
super(upsampleByTransConv, self).__init__()
window_l = upsample_rate * window_ratio
self.m_layer = torch_nn.ConvTranspose1d(
feat_dim, feat_dim, window_l, stride=upsample_rate)
self.m_uprate = upsample_rate
return
def forward(self, x):
""" y = upsampleByTransConv(x)
input
-----
x: tensor, (batch, length, feat_dim)
output
------
y: tensor, (batch, length*upsample_rate, feat_dim)
"""
l = x.shape[1] * self.m_uprate
y = self.m_layer(x.permute(0, 2, 1))[:, :, 0:l]
return y.permute(0, 2, 1).contiguous()
class SqueezeForWaveGlow(torch_nn.Module):
"""SqueezeForWaveGlow
Squeeze layer for WaveGlow
"""
def __init__(self, mode = 1):
"""SqueezeForGlow(mode=1)
Args
----
mode: int, mode of this squeeze layer
mode == 1: original squeeze method by squeezing 8 points
"""
super(SqueezeForWaveGlow, self).__init__()
self.m_mode = mode
# mode 1, squeeze by 8
self.m_mode_1_para = 8
return
def get_expected_squeeze_length(self, orig_length):
# return expected length after squeezing
if self.m_mode == 1:
return orig_length//self.m_mode_1_para
def get_squeeze_factor(self):
# return the configuration for squeezing
if self.m_mode == 1:
return self.m_mode_1_para
def forward(self, x):
"""SqueezeForWaveGlow(x)
input
-----
x: tensor, (batch, length, feat_dim)
output
------
y: tensor, (batch, length // squeeze, feat_dim * squeeze)
"""
if self.m_mode == 1:
# squeeze, the 8 points should be the last dimension
squeeze_len = x.shape[1] // self.m_mode_1_para
# trim length first
trim_len = squeeze_len * self.m_mode_1_para
x_tmp = x[:, 0:trim_len, :]
# (batch, time//squeeze_size, squeeze_size, dim)
x_tmp = x_tmp.view(x_tmp.shape[0], squeeze_len,
self.m_mode_1_para, -1)
# (batch, time//squeeze_size, dim, squeeze_size)
x_tmp = x_tmp.permute(0, 1, 3, 2).contiguous()
# (batch, time//squeeze_size, dim * squeeze_size)
return x_tmp.view(x_tmp.shape[0], squeeze_len, -1)
else:
print("SqueezeForWaveGlow not implemented")
return x_squeezed
def reverse(self, x_squeezed):
if self.m_mode == 1:
# (batch, time//squeeze_size, dim * squeeze_size)
batch, squeeze_len, squeeze_dim = x_squeezed.shape
# (batch, time//squeeze_size, dim, squeeze_size)
x_tmp = x_squeezed.view(
batch, squeeze_len, squeeze_dim // self.m_mode_1_para,
self.m_mode_1_para)
# (batch, time//squeeze_size, squeeze_size, dim)
x_tmp = x_tmp.permute(0, 1, 3, 2).contiguous()
# (batch, time, dim)
x = x_tmp.view(batch, squeeze_len * self.m_mode_1_para, -1)
else:
print("SqueezeForWaveGlow not implemented")
return x
class AffineCouplingWaveGlow_legacy(torch_nn.Module):
"""AffineCouplingWaveGlow_legacy
AffineCoupling block in WaveGlow
Example:
m_tmp = AffineCouplingWaveGlow_legacy(10, 10, 8, 512, 3, True, True)
data1 = torch.randn([2, 100, 10])
cond = torch.randn([2, 100, 10])
output, log_det = m_tmp(data1, cond)
data1_re = m_tmp.reverse(output, cond)
torch.std(data1 - data1_re)
"""
def __init__(self, in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine=True, flag_detjac=False):
"""AffineCouplingWaveGlow_legacy(in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine=True, flag_detjac=False)
Args:
-----
in_dim: int, dim of input audio data (batch, length, in_dim)
cond_dim, int, dim of condition feature (batch, length, cond_dim)
wn_num_conv1d: int, number of dilated conv WaveNet blocks
wn_dim_channel: int, dime of the WaveNet residual & skip channels
wn_kernel_size: int, kernel size of the dilated convolution layers
flag_affine: bool, whether use affine or additive transformation?
default True
flag_detjac: bool, whether return the determinant of Jacobian,
default False
y -> split() -> y1, y2 -> concate([y1, (y2+bias) * scale])
When flag_affine == True, y1 -> H() -> scale, bias
When flag_affine == False, y1 -> H() -> bias, scale=1
Here, H() is WaveNet blocks (dilated conv + gated activation)
"""
super(AffineCouplingWaveGlow_legacy, self).__init__()
self.flag_affine = flag_affine
self.flag_detjac = flag_detjac
if in_dim % 2 > 0:
print("AffineCoulingGlow(feat_dim), feat_dim is an odd number?!")
sys.exit(1)
if self.flag_affine:
# scale and bias
self.m_nn_outdim = in_dim // 2 * 2
else:
# only bias
self.m_nn_outdim = in_dim // 2
# pre-transform, change input audio dimension
# only half of the features will be used to produce scale and bias
tmp_l = torch_nn.Linear(in_dim // 2, wn_dim_channel)
# weight normalization
self.m_wn_pre = torch_nn.utils.weight_norm(tmp_l, name='weight')
# WaveNet blocks (dilated conv, gated activation functions)
tmp_wn = []
for i in range(wn_num_conv1d):
dilation = 2 ** i
tmp_wn.append(nii_wavenet.WaveNetBlock_v2(
wn_dim_channel, wn_dim_channel, wn_dim_channel, cond_dim,
dilation, cnn_kernel_size=wn_kernel_size, causal=False))
self.m_wn = torch_nn.ModuleList(tmp_wn)
# post-transform, change dim from WN channel to audio feature
tmp_l = torch_nn.Linear(wn_dim_channel, self.m_nn_outdim)
# For better initialization, bias=0, scale=1 for first mini-batch
tmp_l.weight.data.zero_()
tmp_l.bias.data.zero_()
self.m_wn_post = tmp_l
return
def _detjac(self, log_scale, factor=1):
# (batch, dim1, dim2, ..., feat_dim) -> (batch)
# sum over dim1, ... feat_dim
return nii_glow.sum_over_keep_batch(log_scale / factor)
def _nn_trans(self, y1, cond):
"""_nn_trans(self, y1, cond)
input
-----
y1: tensor, input feature, (batch, lengh, input_dim//2)
cond: tensor, condition feature, (batch, length, cond_dim)
output
------
scale: tensor, (batch, lengh, input_dim // 2)
bias: tensor, (batch, lengh, input_dim // 2)
log_scale: tensor, (batch, lengh, input_dim // 2)
Affine transformaiton can be done by scale * feature + bias
log_scale is used for det Jacobian computation
"""
# pre-transformation (batch, length, in_dim//2)
# -> (batch, length, WN_channel)
y1_trans = self.m_wn_pre(y1)
# WaveNet blocks
wn_output = 0
res_ch = y1_trans
for wn_layer in self.m_wn:
res_ch, ski_ch = wn_layer(res_ch, cond)
wn_output = wn_output + ski_ch / len(self.m_wn)
#wn_output = wn_output + res_ch / len(self.m_wn)
# post-transformation
y1_tmp = self.m_wn_post(wn_output)
if self.flag_affine:
log_scale, bias = y1_tmp.chunk(2, -1)
scale = torch.exp(log_scale)
else:
bias = y1_tmp
scale = torch.ones_like(y1)
log_scale = torch.zeros_like(y1)
return scale, bias, log_scale
def forward(self, y, cond, factor=1):
"""AffineCouplingWaveGlow_legacy.forward(y, cond)
input
-----
y: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, lengh, cond_dim)
output
------
x: tensor, input feature, (batch, lengh, input_dim)
detjac: tensor, det of jacobian, (batch,)
y1, y2 = split(y)
scale, bias = WN(y1)
x2 = y2 * scale + bias or (y2 + bias) * scale
return [y1, x2]
"""
# split
y1, y2 = y.chunk(2, -1)
scale, bias, log_scale = self._nn_trans(y1, cond)
# transform
x1 = y1
x2 = (y2 + bias) * scale
# concatenate
x = torch.cat([x1, x2], dim=-1)
if self.flag_detjac:
return x, self._detjac(log_scale, factor)
else:
return x
def reverse(self, x, cond):
"""AffineCouplingWaveGlow_legacy.reverse(y, cond)
input
-----
x: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, lengh, cond_dim)
output
------
y: tensor, input feature, (batch, lengh, input_dim)
x1, x2 = split(x)
scale, bias = WN(x1)
y2 = x2 / scale - bias
return [x1, y2]
"""
# split
x1, x2 = x.chunk(2, -1)
# reverse transform
y1 = x1
scale, bias, log_scale = self._nn_trans(y1, cond)
y2 = x2 / scale - bias
return torch.cat([y1, y2], dim=-1)
class WaveNetModuleForNonAR(torch_nn.Module):
"""WaveNetModuleWaveGlow
Casecade of multiple WaveNet blocks:
x -> ExpandDim -> conv1 -> gated -> res -> conv1 -> gated -> res ...
^ |
| v
cond skip
output = sum(skip_channels)
"""
def __init__(self, input_dim, cond_dim, out_dim, n_blocks,
gate_dim, res_ch, skip_ch, kernel_size=3):
super(WaveNetModuleForNonAR, self).__init__()
self.m_block_num = n_blocks
self.m_res_ch_dim = res_ch
self.m_skip_ch_dim = skip_ch
self.m_gate_dim = gate_dim
self.m_kernel_size = kernel_size
self.m_n_blocks = n_blocks
if self.m_gate_dim % 2 != 0:
self.m_gate_dim = self.m_gate_dim // 2 * 2
# input dimension expanding
tmp = torch_nn.Conv1d(input_dim, res_ch, 1)
self.l_expand = torch_nn.utils.weight_norm(tmp, name='weight')
# end dimension compressing
tmp = torch_nn.Conv1d(skip_ch, out_dim, 1)
tmp.weight.data.zero_()
tmp.bias.data.zero_()
self.l_compress = tmp
# dilated convolution and residual-skip-channel transformation
self.l_conv1 = []
self.l_resskip = []
for idx in range(n_blocks):
dilation = 2 ** idx
padding = int((kernel_size * dilation - dilation)/2)
conv1 = torch_nn.Conv1d(
res_ch, gate_dim, self.m_kernel_size,
dilation = dilation, padding=padding)
conv1 = torch_nn.utils.weight_norm(conv1, name='weight')
self.l_conv1.append(conv1)
if idx < n_blocks - 1:
outdim = self.m_res_ch_dim + self.m_skip_ch_dim
else:
outdim = self.m_skip_ch_dim
resskip = torch_nn.Conv1d(self.m_gate_dim//2, outdim, 1)
resskip = torch_nn.utils.weight_norm(resskip, name='weight')
self.l_resskip.append(resskip)
self.l_conv1 = torch_nn.ModuleList(self.l_conv1)
self.l_resskip = torch_nn.ModuleList(self.l_resskip)
# a single conditional feature transformation layer
cond_layer = torch_nn.Conv1d(cond_dim, gate_dim * n_blocks, 1)
cond_layer = torch_nn.utils.weight_norm(cond_layer, name='weight')
self.l_cond = cond_layer
return
def forward(self, x, cond):
"""
"""
# input feature expansion
# change the format to (batch, dimension, length)
x_expanded = self.l_expand(x.permute(0, 2, 1))
# condition feature transformation
cond_proc = self.l_cond(cond.permute(0, 2, 1))
# skip-channel accumulation
skip_ch_out = 0
conv_input = x_expanded
for idx, (l_conv1, l_resskip) in \
enumerate(zip(self.l_conv1, self.l_resskip)):
tmp_dim = idx * self.m_gate_dim
# condition feature of this layer
cond_tmp = cond_proc[:, tmp_dim : tmp_dim + self.m_gate_dim, :]
# conv transformed
conv_tmp = l_conv1(conv_input)
# gated activation
gated_tmp = cond_tmp + conv_tmp
t_part = torch.tanh(gated_tmp[:, :self.m_gate_dim//2, :])
s_part = torch.sigmoid(gated_tmp[:, self.m_gate_dim//2:, :])
gated_tmp = t_part * s_part
# transformation into skip / residual channels
resskip_tmp = l_resskip(gated_tmp)
# reschannel
if idx == self.m_n_blocks - 1:
skip_ch_out = skip_ch_out + resskip_tmp
else:
conv_input = conv_input + resskip_tmp[:, 0:self.m_res_ch_dim, :]
skip_ch_out = skip_ch_out + resskip_tmp[:, self.m_res_ch_dim:,:]
output = self.l_compress(skip_ch_out)
# permute back to (batch, length, dimension)
return output.permute(0, 2, 1)
class AffineCouplingWaveGlow(torch_nn.Module):
"""AffineCouplingWaveGlow
AffineCoupling block in WaveGlow
Example:
m_tmp = AffineCouplingWaveGlow(10, 10, 8, 512, 3, True, True)
data1 = torch.randn([2, 100, 10])
cond = torch.randn([2, 100, 10])
output, log_det = m_tmp(data1, cond)
data1_re = m_tmp.reverse(output, cond)
torch.std(data1 - data1_re)
"""
def __init__(self, in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine=True, flag_detjac=False):
"""AffineCouplingWaveGlow(in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine=True, flag_detjac=False)
Args:
-----
in_dim: int, dim of input audio data (batch, length, in_dim)
cond_dim, int, dim of condition feature (batch, length, cond_dim)
wn_num_conv1d: int, number of dilated conv WaveNet blocks
wn_dim_channel: int, dime of the WaveNet residual & skip channels
wn_kernel_size: int, kernel size of the dilated convolution layers
flag_affine: bool, whether use affine or additive transformation?
default True
flag_detjac: bool, whether return the determinant of Jacobian,
default False
y -> split() -> y1, y2 -> concate([y1, (y2+bias) * scale])
When flag_affine == True, y1 -> H() -> scale, bias
When flag_affine == False, y1 -> H() -> bias, scale=1
Here, H() is WaveNet blocks (dilated conv + gated activation)
"""
super(AffineCouplingWaveGlow, self).__init__()
self.flag_affine = flag_affine
self.flag_detjac = flag_detjac
if in_dim % 2 > 0:
print("AffineCoulingGlow(feat_dim), feat_dim is an odd number?!")
sys.exit(1)
if self.flag_affine:
# scale and bias
self.m_nn_outdim = in_dim // 2 * 2
else:
# only bias
self.m_nn_outdim = in_dim // 2
# WaveNet blocks (dilated conv, gated activation functions)
self.m_wn = WaveNetModuleForNonAR(
in_dim // 2, cond_dim, self.m_nn_outdim, wn_num_conv1d,
wn_dim_channel * 2, wn_dim_channel, wn_dim_channel,
wn_kernel_size
)
return
def _detjac(self, log_scale, factor=1):
# (batch, dim1, dim2, ..., feat_dim) -> (batch)
# sum over dim1, ... feat_dim
return nii_glow.sum_over_keep_batch(log_scale / factor)
def _nn_trans(self, y1, cond):
"""_nn_trans(self, y1, cond)
input
-----
y1: tensor, input feature, (batch, lengh, input_dim//2)
cond: tensor, condition feature, (batch, length, cond_dim)
output
------
scale: tensor, (batch, lengh, input_dim // 2)
bias: tensor, (batch, lengh, input_dim // 2)
log_scale: tensor, (batch, lengh, input_dim // 2)
Affine transformaiton can be done by scale * feature + bias
log_scale is used for det Jacobian computation
"""
y1_tmp = self.m_wn(y1, cond)
if self.flag_affine:
log_scale, bias = y1_tmp.chunk(2, -1)
scale = torch.exp(log_scale)
else:
bias = y1_tmp
scale = torch.ones_like(y1)
log_scale = torch.zeros_like(y1)
return scale, bias, log_scale
def forward(self, y, cond, factor=1):
"""AffineCouplingWaveGlow.forward(y, cond)
input
-----
y: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, lengh, cond_dim)
output
------
x: tensor, input feature, (batch, lengh, input_dim)
detjac: tensor, det of jacobian, (batch,)
y1, y2 = split(y)
scale, bias = WN(y1)
x2 = y2 * scale + bias or (y2 + bias) * scale
return [y1, x2]
"""
# split
y1, y2 = y.chunk(2, -1)
scale, bias, log_scale = self._nn_trans(y1, cond)
# transform
x1 = y1
x2 = (y2 + bias) * scale
# concatenate
x = torch.cat([x1, x2], dim=-1)
if self.flag_detjac:
return x, self._detjac(log_scale, factor)
else:
return x
def reverse(self, x, cond):
"""AffineCouplingWaveGlow.reverse(y, cond)
input
-----
x: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, lengh, cond_dim)
output
------
y: tensor, input feature, (batch, lengh, input_dim)
x1, x2 = split(x)
scale, bias = WN(x1)
y2 = x2 / scale - bias
return [x1, y2]
"""
# split
x1, x2 = x.chunk(2, -1)
# reverse transform
y1 = x1
scale, bias, log_scale = self._nn_trans(y1, cond)
y2 = x2 / scale - bias
return torch.cat([y1, y2], dim=-1)
class FlowStepWaveGlow(torch_nn.Module):
"""FlowStepWaveGlow
One flow step for waveglow
y -> intertical_1x1() -> AffineCoupling -> x
Example
m_tmp = FlowStepWaveGlow(10, 10, 8, 512, 3, flag_affine=True)
output, log_det = m_tmp(data1, cond)
data1_re = m_tmp.reverse(output, cond)
torch.std(data1 - data1_re)
"""
def __init__(self, in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size, flag_affine,
flag_affine_block_legacy=False):
"""FlowStepWaveGlow(in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size, flag_affine,
flag_affine_block_legacy=False)
Args
----
in_dim: int, input feature dim, (batch, length, in_dim)
cond_dim:, int, conditional feature dim, (batch, length, cond_dim)
wn_num_conv1d: int, number of 1Dconv WaveNet block in this flow step
wn_dim_channel: int, dim of the WaveNet residual and skip channels
wn_kernel_size: int, kernel size of the dilated convolution layers
flag_affine: bool, whether use affine or additive transformation?
default True
flag_affine_block_legacy, bool, whether use AffineCouplingWaveGlow or
AffineCouplingWaveGlow_legacy.
For wn_dim_channel and wn_kernel_size, see AffineCouplingWaveGlow
For flag_affine == False, scale will be 1.0
"""
super(FlowStepWaveGlow, self).__init__()
# Invertible transformation layer
#self.m_invtrans = nii_glow.InvertibleTrans(in_dim, flag_detjac=True)
self.m_invtrans = Invertible1x1ConvWaveGlow(in_dim, flag_detjac=True)
# Coupling layer
if flag_affine_block_legacy:
self.m_coupling = AffineCouplingWaveGlow_legacy(
in_dim, cond_dim, wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine, flag_detjac=True)
else:
self.m_coupling = AffineCouplingWaveGlow(
in_dim, cond_dim, wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine, flag_detjac=True)
return
def forward(self, y, cond, factor=1):
"""FlowStepWaveGlow.forward(y, cond, factor=1)
input
-----
y: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, lengh, cond_dim)
factor: int, this is used to divde likelihood, default 1
if we directly sum all detjac, they will become very large
however, we cannot average them directly on y because y
may have a different shape from the actual data y
output
------
x: tensor, input feature, (batch, lengh, input_dim)
detjac: tensor, det of jacobian, (batch,)
"""
# 1x1 transform
x_tmp, log_det_1 = self.m_invtrans(y, factor)
# coupling
x_tmp, log_det_2 = self.m_coupling(x_tmp, cond, factor)
return x_tmp, log_det_1 + log_det_2
def reverse(self, x, cond):
"""FlowStepWaveGlow.reverse(y, cond)
input
-----
x: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, lengh, cond_dim)
output
------
y: tensor, input feature, (batch, lengh, input_dim)
"""
y_tmp = self.m_coupling.reverse(x, cond)
y_tmp = self.m_invtrans.reverse(y_tmp)
return y_tmp
class WaveGlowBlock(torch_nn.Module):
"""WaveGlowBlock
A WaveGlowBlock includes multiple steps of flow.
The Nvidia WaveGlow does not define WaveGlowBlock but directly
defines 12 flow steps. However, after every 4 flow steps, two
dimension of z will be extracted (multi-scale approach).
It is not convenient to decide when to extract z.
Here, we define a WaveGlowBlock as the casecade of multiple flow
steps, and this WaveGlowBlock can extract the two dimensions from
the output of final flow step.
Example:
data1 = torch.randn([2, 10, 10])
cond = torch.randn([2, 10, 16])
m_block = WaveGlowBlock(10, 16, 5, 8, 512, 3)
x, z, log_det = m_block(data1, cond)
data_re = m_block.reverse(x, z, cond)
print(torch.std(data_re - data1))
"""
def __init__(self, in_dim, cond_dim, n_flow_steps,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine=True,
flag_split = False,
flag_final_block=False,
split_dim = 2,
flag_affine_block_legacy=False):
"""WaveGlowBlock(in_dim, cond_dim, n_flow_steps,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine=True, flag_split = False, split_dim = 2,
flag_affine_block_legacy=False)
Args
----
in_dim: int, input feature dim, (batch, length, in_dim)
cond_dim:, int, conditional feature dim, (batch, length, cond_dim)
n_flow_steps: int, number of flow steps in one block
wn_num_conv1d: int, number of dilated conv WaveNet blocks
wn_dim_channel: int, dim of the WaveNet residual and skip channels
wn_kernel_size: int, kernel size of the dilated convolution layers
flag_affine: bool, whether use affine or additive transformation?
default True
flag_split: bool, whether split output z for multi-scale structure
default True
flag_final_block: bool, whether this block is the final block
default False
split_dim: int, if flag_split==True, z[:, :, :split_dim] will be
extracted, z[:, :, split_dim:] can be used for the next
WaveGlowBlock
flag_affine_block_legacy, bool, whether use the legacy implementation
of wavenet-based affine transformaiton layer
default False.
For wn_dim_channel and wn_kernel_size, see AffineCouplingWaveGlow
For flag_affine, see AffineCouplingWaveGlow
"""
super(WaveGlowBlock, self).__init__()
tmp_flows = []
for i in range(n_flow_steps):
tmp_flows.append(
FlowStepWaveGlow(
in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine, flag_affine_block_legacy))
self.m_flows = torch_nn.ModuleList(tmp_flows)
self.flag_split = flag_split
self.flag_final_block = flag_final_block
self.split_dim = split_dim
if self.flag_split and self.flag_final_block:
print("WaveGlowBlock: flag_split and flag_final_block are True")
print("This is unexpected. Please check model definition")
sys.exit(1)
if self.flag_split and self.split_dim <= 0:
print("WaveGlowBlock: split_dim should be > 0")
sys.exit(1)
return
def forward(self, y, cond, factor=1):
"""x, z, log_detjac = WaveGlowBlock(y)
y -> H() -> [z, x], log_det_jacobian
H() consists of multiple flow steps (1x1conv + AffineCoupling)
input
-----
y: tensor, (batch, length, dim)
cond, tensor, (batch, length, cond_dim)
factor, None or int, this is used to divde likelihood, default 1
output
------
log_detjac: tensor or scalar
if self.flag_split:
x: tensor, (batch, length, in_dim - split_dim),
z: tensor, (batch, length, split_dim),
else:
if self.flag_final_block:
x: None, no input to the next block
z: tensor, (batch, length, dim), for N(z; 0, I)
else:
x: tensor, (batch, length, dim),
z: None, no latent for N(z; 0, I) from this block
concate([x,z]) should have the same size as y
"""
# flows
log_detjac = 0
x_tmp = y
for l_flow in self.m_flows:
x_tmp, log_detjac_tmp = l_flow(x_tmp, cond, factor)
log_detjac = log_detjac + log_detjac_tmp
if self.flag_split:
z = x_tmp[:, :, :self.split_dim]
x = x_tmp[:, :, self.split_dim:]
else:
if self.flag_final_block:
z = x_tmp
x = None
else:
z = None
x = x_tmp
return x, z, log_detjac
def reverse(self, x, z, cond):
"""y = WaveGlowBlock.reverse(x, z, cond)
[z, x] -> H^{-1}() -> y
input
-----
if self.flag_split:
x: tensor, (batch, length, in_dim - split_dim),
z: tensor, (batch, length, split_dim),
else:
if self.flag_final_block:
x: None
z: tensor, (batch, length, in_dim)
else:
x: tensor, (batch, length, in_dim)
z: None
output
------
y: tensor, (batch, length, in_dim)
"""
if self.flag_split:
if x is None or z is None:
print("WaveGlowBlock.reverse: x and z should not be None")
sys.exit(1)
y_tmp = torch.cat([z, x], dim=-1)
else:
if self.flag_final_block:
if z is None:
print("WaveGlowBlock.reverse: z should not be None")
sys.exit(1)
y_tmp = z
else:
if x is None:
print("WaveGlowBlock.reverse: x should not be None")
sys.exit(1)
y_tmp = x
for l_flow in self.m_flows[::-1]:
# affine
y_tmp = l_flow.reverse(y_tmp, cond)
return y_tmp
class WaveGlow(torch_nn.Module):
"""WaveGlow
Example
cond_dim = 4
upsample = 80
num_blocks = 4
num_flows_inblock = 5
wn_num_conv1d = 8
wn_dim_channel = 512
wn_kernel_size = 3
# waveforms of length 1600
wave1 = torch.randn([2, 1600, 1])
# condition feature
cond = torch.randn([2, 1600//upsample, cond_dim])
# model
m_model = nii_waveglow.WaveGlow(
cond_dim, upsample,
num_blocks, num_flows_inblock, wn_num_conv1d,
wn_dim_channel, wn_kernel_size)
# forward computation, neg_log = -(logp + log_detjac)
# neg_log.backward() can be used for backward
z, neg_log, logp, log_detjac = m_model(wave1, cond)
# recover the signal
wave2 = m_model.reverse(z, cond)
# check difference between original wave and recovered wave
print(torch.std(wave1 - wave2))
"""
def __init__(self, cond_dim, upsample_rate,
num_blocks, num_flows_inblock,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine = True,
early_hid_dim=2,
flag_affine_block_legacy=False):
"""WaveGlow(cond_dim, upsample_rate,
num_blocks, num_flows_inblock,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine = True,
early_hid_dim=2,
flag_affine_block_legacy=False)
Args
----
cond_dim:, int, conditional feature dim, (batch, length, cond_dim)
upsample_rate: int, up-sampling rate for condition features
num_blocks: int, number of WaveGlowBlocks
num_flows_inblock: int, number of flow steps in one WaveGlowBlock
wn_num_conv1d: int, number of 1Dconv WaveNet block in this flow step
wn_dim_channel: int, dim of the WaveNet residual and skip channels
wn_kernel_size: int, kernel size of the dilated convolution layers
flag_affine: bool, whether use affine or additive transformation?
default True
early_hid_dim: int, dimension for z_1, z_2 ... , default 2
flag_affine_block_legacy, bool, whether use the legacy implementation
of wavenet-based affine transformaiton layer
default False. The difference is on the WaveNet part
Please configure AffineCouplingWaveGlow and
AffineCouplingWaveGlow_legacy
This model defines:
cond -> upsample/squeeze -> | ------> | --------> |
v v v
y -> squeeze -> WaveGlowBlock -> WGBlock ... WGBlock -> z
|-> z_1 |-> z_2
z_1, z_2, ... are the extracted z from a multi-scale flow structure
concate([z_1, z_2, z]) is expected to be the white Gaussian noise
If early_hid_dim == 0, z_1 and z_2 will not be extracted
"""
super(WaveGlow, self).__init__()
# input is assumed to be waveform
self.m_input_dim = 1
self.m_early_hid_dim = early_hid_dim
# squeeze layer
self.m_squeeze = SqueezeForWaveGlow()
# up-sampling layer
#self.m_upsample = nii_nn.UpSampleLayer(cond_dim, upsample_rate, True)
self.m_upsample = upsampleByTransConv(cond_dim, upsample_rate)
# wavenet-based flow blocks
# squeezed input dimension
squeezed_in_dim = self.m_input_dim * self.m_squeeze.get_squeeze_factor()
# squeezed condition feature dimension
squeezed_cond_dim = cond_dim * self.m_squeeze.get_squeeze_factor()
# save the dimension for get_z_noises
self.m_feat_dim = []
# define blocks
tmp_squeezed_in_dim = squeezed_in_dim
tmp_flow_blocks = []
for i in range(num_blocks):
# if this is not the last block and early_hid_dim >0
flag_split = (i < (num_blocks-1)) and early_hid_dim > 0
flag_final_block = i == (num_blocks-1)
# save the dimension for get_z_noises
if flag_final_block:
self.m_feat_dim.append(tmp_squeezed_in_dim)
else:
self.m_feat_dim.append(early_hid_dim if flag_split else 0)
tmp_flow_blocks.append(
WaveGlowBlock(
tmp_squeezed_in_dim, squeezed_cond_dim, num_flows_inblock,
wn_num_conv1d, wn_dim_channel, wn_kernel_size, flag_affine,
flag_split = flag_split, flag_final_block=flag_final_block,
split_dim = early_hid_dim,
flag_affine_block_legacy = flag_affine_block_legacy))
# multi-scale approach will extract a few dimensions for next flow
# thus, input dimension to the next block will be this
tmp_squeezed_in_dim = tmp_squeezed_in_dim - early_hid_dim
self.m_flowblocks = torch_nn.ModuleList(tmp_flow_blocks)
# done
return
def _normal_lh(self, noise):
# likelihood of normal distribution on the given noise
return -0.5 * np.log(2 * np.pi) - 0.5 * noise ** 2
def forward(self, y, cond):
"""z, neg_logp_y, logp_z, logdet = WaveGlow.forward(y, cond)
cond -> upsample/squeeze -> | ------> | --------> |
v v v
y -> squeeze -> WaveGlowBlock -> WGBlock ... WGBlock -> z
|-> z_1 |-> z_2
input
-----
y: tensor, (batch, waveform_length, 1)
cond: tensor, (batch, cond_length, 1)
output
------
z: list of tensors, [z_1, z_2, ... ,z ] in figure above
neg_logp_y: scalar, - log p(y)
logp_z: scalar, -log N(z), summed over one data sequence, but averaged
over batch.
logdet: scalar, -|det dH(.)/dy|, summed over one data sequence,
but averaged
over batch.
If self.early_hid_dim == 0, z_1, z_2 ... will be None
"""
# Rather than summing the likelihood and divide it by the number of
# data in the final step, we divide this factor from the likelihood
# caculating by each flow step and sum the scaled likelihood.
# Two methods are equivalent, but the latter may prevent numerical
# overflow of the likelihood value for long sentences
factor = np.prod([dim for dim in y.shape])
# waveform squeeze (batch, squeezed_length, squeezed_dim)
y_squeezed = self.m_squeeze(y)
squeezed_dim = y_squeezed.shape[-1]
# condition feature upsampling and squeeze
# (batch, squeezed_length, squeezed_dim_cond)
cond_up_squeezed = self.m_squeeze(self.m_upsample(cond))
# flows
z_bags = []
log_detjac = 0
log_pz = 0
x_tmp = y_squeezed
for m_block in self.m_flowblocks:
x_tmp, z_tmp, log_detjac_tmp = m_block(
x_tmp, cond_up_squeezed, factor)
# accumulate log det jacobian
log_detjac += log_detjac_tmp
# compute N(z; 0, I)
# save z_tmp (even if it is None)
z_bags.append(z_tmp)
# accumulate log_N(z; 0, I) only if it is valid
if z_tmp is not None:
log_pz += nii_glow.sum_over_keep_batch2(
self._normal_lh(z_tmp), factor)
# average over batch and data points
neg_logp_y = -(log_pz + log_detjac).sum()
return z_bags, neg_logp_y, \
log_pz.sum(), log_detjac.sum()
def reverse(self, z_bags, cond):
"""y = WaveGlow.reverse(z_bags, cond)
cond -> upsample/squeeze -> | ------> | --------> |
v v v
y <- unsqueeze <- WaveGlowBlock -> WGBlock ... WGBlock <- z
|<- z_1 |<- z_2
input
-----
z: list of tensors, [z_1, z_2, ... ,z ] in figure above
cond: tensor, (batch, cond_length, 1)
output
------
y: tensor, (batch, waveform_length, 1)
If self.early_hid_dim == 0, z_1, z_2 ... should be None
"""
# condition feature upsampling and squeeze
# (batch, squeezed_length, squeezed_dim_cond)
cond_up_sqe = self.m_squeeze(self.m_upsample(cond))
# initial
y_tmp = None
for z, m_block in zip(z_bags[::-1], self.m_flowblocks[::-1]):
y_tmp = m_block.reverse(y_tmp, z, cond_up_sqe)
y = self.m_squeeze.reverse(y_tmp)
return y
def get_z_noises(self, length, noise_std=0.7, batchsize=1):
"""z_bags = WaveGlow.get_z_noises(length, noise_std=0.7, batchsize=1)
Return a list of random noises for random sampling
input
-----
length: int, length of target waveform (without squeeze)
noise_std: float, std of Gaussian noise, default 0.7
batchsize: int, batch size of this random data, default 1
output
------
z_bags: list of tensors
Shape of tensor in z_bags is decided by WaveGlow configuration.
WaveGlow.reverse(z_bags, cond) can be used to generate waveform
"""
squeeze_length = self.m_squeeze.get_expected_squeeze_length(length)
device = next(self.parameters()).device
z_bags = []
# generate the z for each WaveGlowBlock
for feat_dim in self.m_feat_dim:
if feat_dim is not None and feat_dim > 0:
z_tmp = torch.randn(
[batchsize, squeeze_length, feat_dim],
dtype=nii_io_conf.d_dtype,
device=device)
z_bags.append(z_tmp * noise_std)
else:
z_bags.append(None)
return z_bags
if __name__ == "__main__":
print("Definition of WaveGlow")
| 42,401 | 35.711688 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/dynamic_prog.py | #!/usr/bin/env python
"""
Functions for dynamic programming
"""
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import core_scripts.other_tools.debug as nii_debug
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#############################################################
def viterbi_decode(init_prob, trans_prob, obser_prob,
eps=torch.finfo(torch.float32).eps, return_more=False):
""" Routine to do Viterbi decoding
viterbi_decode(init_prob, trans_prob, obser_prob,
eps=torch.finfo(torch.float32).eps, return_more=False):
Input:
init_prob: initialia state probability
tensor or np.arrary, in shape (N), for N states
trans_prob: transition probability
tensor or np.array, in shape (N, N)
trans_prob(i, j): P(state=j | prev_state=i)
obser_prob: observation probability
tensor or np.array, in shape (T, N), for T time sptes
return_more: True: return best_states, prob_mat, state_trace
False: return best_states
Output:
best_states: best state sequence tensor or np.array, in shape (T)
prob_mat: probablity matrix in shape (T, N), where (t, j) denotes
max_{s_1:t-1} P(o_1:t, s_1:t-1, s_t=j)
state_mat: in shape (T, N), where (t, j) denotes
argmax_i P(o_1:t, s_1:t-2, s_t-1=i, s_t=j)
"""
if type(init_prob) is torch.Tensor:
_log_func = torch.log
_torch_flag = True
else:
_log_func = np.log
_torch_flag = False
log_init_prob = _log_func(init_prob + eps)
log_trans_prob = _log_func(trans_prob + eps)
log_obser_prob = _log_func(obser_prob + eps)
n_time, n_state = log_obser_prob.shape
if log_trans_prob.shape[0] != n_state or log_trans_prob.shape[0] != n_state:
print("Viterbi decoding: transition prob matrix invalid")
sys.exit(1)
if log_init_prob.shape[0] != n_state:
print("Viterbi decoding: init prob matrix invalid")
sys.exit(1)
if _torch_flag:
prob_mat = torch.zeros_like(log_obser_prob)
state_mat = torch.zeros_like(log_obser_prob, dtype=torch.int)
best_states = torch.zeros([n_time], dtype=torch.int,
device = init_prob.device)
_argmax = torch.argmax
tmp_idx = torch.arange(0, n_state, dtype=torch.long)
else:
prob_mat = np.zeros(log_obser_prob.shape)
state_mat = np.zeros(log_obser_prob.shape, dtype=np.int)
best_states = np.zeros([n_time], dtype=np.int)
_argmax = np.argmax
tmp_idx = np.arange(0, n_state, dtype=np.int)
prob_mat[0, :] = log_init_prob + log_obser_prob[0, :]
for time_idx in np.arange(1, n_time):
trout_prob = prob_mat[time_idx - 1] + log_trans_prob.T
# this version is faster?
#print(time_idx)
tmp_best = _argmax(trout_prob, axis=1)
state_mat[time_idx] = tmp_best
prob_mat[time_idx] = trout_prob[tmp_idx, tmp_best] \
+ log_obser_prob[time_idx]
# seems to be too slow
#for state_idx in np.arange(n_state):
# tmp_best = _argmax(trout_prob[state_idx])
# state_mat[time_idx, state_idx] = tmp_best
# prob_mat[time_idx, state_idx] = trout_prob[state_idx, tmp_best] \
# +log_obser_prob[time_idx, state_idx]
best_states[-1] = _argmax(prob_mat[-1, :])
for time_idx in np.arange(n_time-2, -1, -1):
best_states[time_idx] = state_mat[time_idx+1, best_states[time_idx+1]]
if return_more:
return best_states, prob_mat, state_mat
else:
return best_states
| 3,957 | 36.695238 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/block_blow.py | #!/usr/bin/env python
"""
Building blocks for Blow
Serra, J., Pascual, S. & Segura, C. Blow: a single-scale hyperconditioned flow
for non-parallel raw-audio voice conversion. in Proc. NIPS (2019).
Reference: https://github.com/joansj/blow
"""
from __future__ import absolute_import
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import torch.nn.init as torch_init
import sandbox.block_glow as nii_glow
import core_scripts.data_io.wav_tools as nii_wav_tk
import core_scripts.data_io.conf as nii_io_conf
import core_scripts.other_tools.debug as nii_debug
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#######################################
# Numpy utilities for data augmentation
#######################################
def flip(x):
"""y=flip(x) flips the sign of x
input: x, np.array
output: y, np.array
"""
return np.sign(np.random.rand(1)-0.5) * x
def ampscale(x):
"""y=ampscale(x) randomly scale the amplitude of x
input: x, np.array
output: y, np.array
"""
return (2*np.random.rand(1)-1) * x / (np.max(np.abs(x)) + 1e-07)
def framejitter(x, framelen):
"""y=framejitter(x, framelen)
input: x, np.array, original waveform (length, 1)
framelen, int, framelen
output: y, np.array, segment of the waveform
"""
framelen = x.shape[0] if framelen > x.shape[0] else framelen
random_start = int(np.ceil(np.random.rand(1) * (x.shape[0] - framelen)))
return x[random_start:random_start+framelen]
def emphasis_rand(x, coef_val):
"""y=deemphasis(x, coef_val)
input: x, np.array, original waveform (length, 1) or (length)
framelen, int, framelen
output: y, np.array, segment of the waveform
"""
coef = (2 * np.random.rand(1) - 1) * coef_val
x_new = np.zeros_like(x) + x
x_new[1:] = x_new[1:] - coef * x[:-1]
return x_new
def wav_aug(x, framelen, coef_val, sr):
"""y = wav_aug(x, framelen, coef_val, sr)
input
-----
x: np.array, original waveform (length, 1)
framelen: int, frame length
coef_val: float, reference coefficient for emphasis-rand
sr: int, sampling rate (e.g., 16000)
output
------
y: np.array, pre-processed waveform (length, 1)
"""
trimmed_x = nii_wav_tk.silence_handler_wrapper(x, sr, flag_output=1)
x_frame = framejitter(trimmed_x, framelen)
return ampscale(emphasis_rand(x_frame, coef_val))
class OverlapAdder(torch_nn.Module):
"""OverlapAdder
"""
def __init__(self, fl, fs, flag_win_analysis=True):
"""OverlapAdder(flag_windowing_before=True)
Args
----
fl: int, frame length
fs: int, frame shift
flag_win_analysis: bool (default True)
True: apply windowing during analysis
False: apply windowing during synthesis
"""
super(OverlapAdder, self).__init__()
self.fl = fl
self.fs = fs
self.flag_win_ana = flag_win_analysis
# assume even
self.m_win = torch_nn.Parameter(torch.hann_window(self.fl))
return
def get_frame_num(self, wav_length):
"""frame_num = get_frame_num(wav_length)
wav_length: int, waveform length
frame_num: int, number of frames
"""
return (wav_length - self.fl) // self.fs + 1
def get_wavlength(self, frame_num):
"""wav_length = get_wavlength(self, frame_num)
wav_length: int, waveform length
frame_num: int, number of frames
"""
return (frame_num - 1) * self.fs + self.fl
def forward(self, x):
"""OverlapAdder(x)
input
-----
x: tensor, (batch, length, 1)
output
------
y: tensor, (batch, frame_num, frame_length)
"""
frame_num = self.get_frame_num(x.shape[1])
# (batch, num_patches, 1, patch_size)
# num_patches = (length - length) // shift + 1
# and copy the data
# note that unfold put each patch as the last dimension
# x_tmp (batch, frame_num, 1, frame_length)
x_tmp = x.unfold(1, self.fl, self.fs)
# apply window
if self.flag_win_ana:
x_tmp = x_tmp * self.m_win
# (batch, frame_num, frame_length)
return x_tmp.view(x.shape[0], x_tmp.shape[1], -1)
def reverse(self, x_framed, flag_scale=False):
"""OverlapAdder(x)
input
-----
x: tensor, (batch, frame_num, frame_length)
flag_scale: bool, whether scale the ampltidue to (-1, 1)
default False
output
------
y: tensor, (batch, length, 1)
"""
batch, frame_num, frame_len = x_framed.shape
x_len = self.get_wavlength(frame_num)
x_buf = torch.zeros(
[batch, x_len], device=x_framed.device, dtype=x_framed.dtype)
x_win = torch.zeros_like(x_buf)
for idx in range(frame_num):
sdx = idx * self.fs
edx = sdx + self.fl
x_win[:, sdx:edx] += self.m_win
if not self.flag_win_ana:
x_buf[:, sdx:edx] += x_framed[:, idx] * self.m_win
else:
x_buf[:, sdx:edx] += x_framed[:, idx]
# assume the overlapped window has a constant amplitude
x_buf = x_buf / x_win.mean()
# normalize the amplitude between (-1, 1)
if flag_scale:
# if input is between (-1, 1), there is no need to
# do this normalization
x_buf = x_buf / (x_buf.abs().max())
return x_buf.unsqueeze(-1)
#######################################
# Torch model definition
#######################################
class AffineCouplingBlow_core(torch_nn.Module):
"""AffineCouplingBlow_core
AffineCoupling core layer the produces the scale and bias parameters.
Example:
feat_dim = 10
cond_dim = 20
m_layer = AffineCouplingBlow_core(feat_dim, cond_dim, 64, 2)
data = torch.randn([2, 100, feat_dim])
cond = torch.randn([2, 1, cond_dim])
scale, bias, log_scale = m_layer(data, cond)
"""
def __init__(self, feat_dim, cond_dim, num_ch, kernel_size=3):
"""AffineCouplingBlow_core(feat_dim, cond_dim, num_ch, kernel_size=3)
Args
----
feat_dim: int, dimension of input feature
cond_dim: int, dimension of conditional features
num_ch: int, number of channels for conv layers
kernel_size: int, kernel size of conv layer, default 3
input_feature -------> func.conv1d -----> conv1ds -> scale, bias
^
|
cond_dim ---> Adapter -> conv weight/bias
"""
super(AffineCouplingBlow_core, self).__init__()
self.feat_dim = feat_dim
self.cond_dim = cond_dim
# make sure that kernel is odd
if kernel_size % 2 == 0:
self.kernel_s = kernel_size + 1
print("\tAffineCouplingBlow_core", end=" ")
print("kernel size {:d} -> {:d}".format(kernel_size, self.kernel_s))
else:
self.kernel_s = kernel_size
if num_ch % feat_dim != 0:
# make sure that number of channel is good
self.num_ch = num_ch // feat_dim * feat_dim
print("\tAffineCouplingBlow_core", end=" ")
print("conv channel {:d} -> {:d}".format(num_ch, self.num_ch))
else:
self.num_ch = num_ch
# Adapter
# (batch, 1, cond_dim) -> (batch, 1, kernel_size * num_ch) for weight
# -> (batch, 1, num_ch) for bias
self.m_adapter = torch_nn.Linear(cond_dim,
(self.kernel_s+1) * self.num_ch)
# conv1d with condition-independent parameters
self.m_conv1ds = torch_nn.Sequential(
torch_nn.ReLU(),
torch_nn.Conv1d(self.num_ch, self.num_ch, 1),
torch_nn.ReLU(),
torch_nn.Conv1d(self.num_ch, feat_dim * 2, self.kernel_s,
padding=(self.kernel_s-1)//2)
)
# zero initialization for the last conv layers
# similar to Glow and WaveGlow
self.m_conv1ds[-1].weight.data.zero_()
self.m_conv1ds[-1].bias.data.zero_()
return
def forward(self, x, cond):
"""scale, bias = AffineCouplingBlow_core(x, cond)
input
-----
x: tensor, input tensor (batch, length, feat_dim)
cond: tensor, condition feature (batch, 1, cond_dim)
output
------
scale: tensor, scaling parameters (batch, length, feat_dim)
bias: tensor, bias paramerters (batch, length, feat_dim)
"""
# cond_dim -> Adapter -> conv weight/bias
# cond[:, 0, :] -> (batch, cond_dim)
# adapter(cond[:, 0, :]) -> (batch, kernel_size * num_ch + num_ch)
# view(...) -> (batch * num_ch, kernel_size + 1)
weight_bias = self.m_adapter(cond[:, 0, :]).view(-1, self.kernel_s+1)
# (batch * num_ch, 1, kernel_size)
weight = weight_bias[:, 0:self.kernel_s].unsqueeze(1)
# (batch * num_ch)
bias = weight_bias[:, self.kernel_s]
# convolution given weight_bias
padsize = (self.kernel_s - 1) // 2
groupsize = x.shape[0] * self.feat_dim
length = x.shape[1]
# x.permute(0, 2, 1)...view -> (1, batch*feat_dim, length)
# conv1d -> (1, batch * num_ch, length)
# view -> (batch, num_ch, length)
x_tmp = torch_nn_func.conv1d(
x.permute(0, 2, 1).contiguous().view(1, -1, length),
weight,
bias = bias,
padding = padsize,
groups = groupsize
).view(x.shape[0], -1, length)
# condition invariant conv -> (batch, feat_dim * 2, length)
x_tmp = self.m_conv1ds(x_tmp)
# scale and bias (batch, feat_dim, length)
raw_scale, bias = torch.chunk(x_tmp, 2, dim=1)
# -> (batch, length, feat_dim)
bias = bias.permute(0, 2, 1)
# re-parameterize
# Here we need to add a small number, otherwise, log(scale)
# somtime times become -inf during training
scale = torch.sigmoid(raw_scale + 2).permute(0, 2, 1) * 0.5 + 0.5
log_scale = torch.log(scale)
#print("Debug: {:.3f} {:.3f} {:.3f} {:3f}".format(
# log_scale.max().item(), log_scale.min().item(),
# scale.max().item(), scale.min().item()),
# file=sys.stderr)
return scale, bias, log_scale
class AffineCouplingBlow(torch_nn.Module):
"""AffineCouplingBlow
AffineCoupling block in Blow
Example:
feat_dim = 10
cond_dim = 20
m_layer = AffineCouplingBlow(feat_dim, cond_dim,60,3, flag_detjac=True)
data = torch.randn([2, 100, feat_dim])
cond = torch.randn([2, 1, cond_dim])
out, detjac = m_layer(data, cond)
data_rever = m_layer.reverse(out, cond)
torch.std(data - data_rever)
"""
def __init__(self, in_dim, cond_dim,
conv_dim_channel, conv_kernel_size,
flag_detjac=False):
"""AffineCouplingBlow(in_dim, cond_dim,
wn_num_conv1d, wn_dim_channel, wn_kernel_size,
flag_affine=True, flag_detjac=False)
Args:
-----
in_dim: int, dim of input audio data (batch, length, in_dim)
cond_dim, int, dim of condition feature (batch, length, cond_dim)
conv_dim_channel: int, dime of the convolution channels
conv_kernel_size: int, kernel size of the convolution layers
flag_detjac: bool, whether return the determinant of Jacobian,
default False
y -> split() -> y1, y2 -> concate([y1, (y2+bias) * scale])
When flag_affine == True, y1 -> H() -> scale, bias
When flag_affine == False, y1 -> H() -> bias, scale=1
Here, H() is AffineCouplingBlow_core layer
"""
super(AffineCouplingBlow, self).__init__()
self.flag_detjac = flag_detjac
if in_dim % 2 > 0:
print("AffineCouplingBlow(feat_dim), feat_dim is an odd number?!")
sys.exit(1)
# Convolution block to get scale and bias
self.m_core = AffineCouplingBlow_core(
in_dim // 2, cond_dim, conv_dim_channel, conv_kernel_size)
return
def _detjac(self, log_scale, factor=1):
# (batch, dim1, dim2, ..., feat_dim) -> (batch)
# sum over dim1, ... feat_dim
return nii_glow.sum_over_keep_batch(log_scale / factor)
def _nn_trans(self, y1, cond):
"""_nn_trans(self, y1, cond)
input
-----
y1: tensor, input feature, (batch, lengh, input_dim//2)
cond: tensor, condition feature, (batch, length, cond_dim)
output
------
scale: tensor, (batch, lengh, input_dim // 2)
bias: tensor, (batch, lengh, input_dim // 2)
log_scale: tensor, (batch, lengh, input_dim // 2)
Affine transformaiton can be done by scale * feature + bias
log_scale is used for det Jacobian computation
"""
scale, bias, log_scale = self.m_core(y1, cond)
return scale, bias, log_scale
def forward(self, y, cond, factor=1):
"""AffineCouplingBlow.forward(y, cond)
input
-----
y: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, 1, cond_dim)
output
------
x: tensor, input feature, (batch, lengh, input_dim)
detjac: tensor, det of jacobian, (batch,)
y1, y2 = split(y)
scale, bias = Conv(y1)
x2 = y2 * scale + bias or (y2 + bias) * scale
return [y1, x2]
"""
# split
y1, y2 = y.chunk(2, -1)
scale, bias, log_scale = self._nn_trans(y1, cond)
# transform
x1 = y1
x2 = (y2 + bias) * scale
# concatenate
x = torch.cat([x1, x2], dim=-1)
if self.flag_detjac:
return x, self._detjac(log_scale, factor)
else:
return x
def reverse(self, x, cond):
"""AffineCouplingBlow.reverse(y, cond)
input
-----
x: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, 1, cond_dim)
output
------
y: tensor, input feature, (batch, lengh, input_dim)
x1, x2 = split(x)
scale, bias = conv(x1)
y2 = x2 / scale - bias
return [x1, y2]
"""
# split
x1, x2 = x.chunk(2, -1)
# reverse transform
y1 = x1
scale, bias, log_scale = self._nn_trans(y1, cond)
y2 = x2 / scale - bias
return torch.cat([y1, y2], dim=-1)
class SqueezeForBlow(torch_nn.Module):
"""SqueezeForBlow
Squeeze input feature for Blow.
Example
data = torch.randn([2, 10, 3])
m_sq = SqueezeForBlow()
data_out = m_sq(data)
data_rev = m_sq.reverse(data_out)
torch.std(data_rev - data)
"""
def __init__(self, mode=1):
"""SqueezeForBlow(mode=1)
Args
----
mode: int, mode of squeeze, default 1
Mode 1: squeeze by a factor of 2 as in original paper
"""
super(SqueezeForBlow, self).__init__()
self.m_mode = mode
if self.m_mode == 1:
self.squeeze_factor = 2
else:
print("SqueezeForBlow mode {:d} not implemented".format(mode))
sys.exit(1)
return
def get_expected_squeeze_length(self, orig_length):
# return expected length after squeezing
if self.m_mode == 1:
return orig_length // self.squeeze_factor
else:
print("unknown mode for SqueezeForBlow")
sys.exit(1)
def get_recovered_length(self, squeezed_length):
# return original length before squeezing
if self.m_mode == 1:
return squeezed_length * self.squeeze_factor
else:
print("unknown mode for SqueezeForBlow")
sys.exit(1)
def get_squeeze_factor(self):
# return the configuration for squeezing
if self.m_mode == 1:
return self.squeeze_factor
else:
print("unknown mode for SqueezeForBlow")
sys.exit(1)
def forward(self, x):
"""SqueezeForBlow(x)
input
-----
x: tensor, (batch, length, feat_dim)
output
------
y: tensor, (batch, length//squeeze_factor, feat_dim*squeeze_factor)
"""
if self.m_mode == 1:
# squeeze, the 8 points should be the last dimension
squeeze_len = self.get_expected_squeeze_length(x.shape[1])
# trim length first
trim_len = squeeze_len * self.squeeze_factor
x_tmp = x[:, 0:trim_len, :]
# (batch, time//squeeze_size, squeeze_size, dim)
x_tmp = x_tmp.view(x_tmp.shape[0], squeeze_len,
self.squeeze_factor, -1)
# (batch, time//squeeze_size, dim, squeeze_size)
x_tmp = x_tmp.permute(0, 1, 3, 2).contiguous()
# (batch, time//squeeze_size, dim * squeeze_size)
return x_tmp.view(x_tmp.shape[0], squeeze_len, -1)
else:
print("SqueezeForWaveGlow not implemented")
sys.exit(1)
return x_squeezed
def reverse(self, x_squeezed):
if self.m_mode == 1:
# (batch, time//squeeze_size, dim * squeeze_size)
batch, squeeze_len, squeeze_dim = x_squeezed.shape
# (batch, time//squeeze_size, dim, squeeze_size)
x_tmp = x_squeezed.view(
batch, squeeze_len, squeeze_dim // self.squeeze_factor,
self.squeeze_factor)
# (batch, time//squeeze_size, squeeze_size, dim)
x_tmp = x_tmp.permute(0, 1, 3, 2).contiguous()
# (batch, time, dim)
x = x_tmp.view(batch, squeeze_len * self.squeeze_factor, -1)
else:
print("SqueezeForWaveGlow not implemented")
sys.exit(1)
return x
class FlowStepBlow(torch_nn.Module):
"""FlowStepBlow
One flow step for Blow
y -> intertical_1x1() -> ActNorm -> AffineCoupling -> x
Example
feat_dim = 10
cond_dim = 20
m_layer = FlowStepBlow(feat_dim, cond_dim, 60, 3)
data = torch.randn([2, 100, feat_dim])
cond = torch.randn([2, 1, cond_dim])
out, detjac = m_layer(data, cond)
data_rever = m_layer.reverse(out, cond)
torch.std(data - data_rever)
"""
def __init__(self, in_dim, cond_dim, conv_dim_channel, conv_kernel_size):
"""FlowStepBlow(in_dim, cond_dim,
conv_dim_channel, conv_kernel_size)
Args
----
in_dim: int, input feature dim, (batch, length, in_dim)
cond_dim:, int, conditional feature dim, (batch, length, cond_dim)
cond_dim_channel: int, dim of the convolution layers
conv_kernel_size: int, kernel size of the convolution layers
For cond_dim_channel and conv_kernel_size, see AffineCouplingBlow
"""
super(FlowStepBlow, self).__init__()
# Invertible transformation layer
self.m_invtrans = nii_glow.InvertibleTrans(in_dim, flag_detjac=True)
# Act norm layer
self.m_actnorm = nii_glow.ActNorm(in_dim, flag_detjac=True)
# coupling layer
self.m_coupling = AffineCouplingBlow(
in_dim, cond_dim, conv_dim_channel, conv_kernel_size,
flag_detjac=True)
return
def forward(self, y, cond, factor=1):
"""FlowStepBlow.forward(y, cond, factor=1)
input
-----
y: tensor, input feature, (batch, lengh, in_dim)
cond: tensor, condition feature , (batch, 1, cond_dim)
factor: int, this is used to divde likelihood, default 1
if we directly sum all detjac, they will become very large
however, we cannot average them directly on y because y
may have a different shape from the actual data y
output
------
x: tensor, input feature, (batch, lengh, input_dim)
detjac: tensor, det of jacobian, (batch,)
"""
# 1x1 transform
x_tmp, log_det_1 = self.m_invtrans(y, factor)
# Actnorm
x_tmp, log_det_2 = self.m_actnorm(x_tmp, factor)
# coupling
x_tmp, log_det_3 = self.m_coupling(x_tmp, cond, factor)
return x_tmp, log_det_1 + log_det_2 + log_det_3
def reverse(self, x, cond):
"""FlowStepBlow.reverse(y, cond)
input
-----
x: tensor, input feature, (batch, lengh, input_dim)
cond: tensor, condition feature , (batch, 1, cond_dim)
output
------
y: tensor, input feature, (batch, lengh, input_dim)
"""
y_tmp1 = self.m_coupling.reverse(x, cond)
y_tmp2 = self.m_actnorm.reverse(y_tmp1)
y_tmp3 = self.m_invtrans.reverse(y_tmp2)
#print("Debug: {:.3f} {:.3f} {:.3f} {:.3f} {:.3f} {:.3f}".format(
# y_tmp1.max().item(), y_tmp1.min().item(),
# y_tmp2.max().item(), y_tmp2.min().item(),
# y_tmp3.max().item(), y_tmp3.min().item()))
return y_tmp3
class BlowBlock(torch_nn.Module):
"""BlowBlock
A BlowBlok includes multiple steps of flow for Blow.
Each block conducts:
x -> squeeze -> flow step1 -> ... -> flow step N
Compared with WaveGlowBlock, this is easier because there is no
multi-scale structure, no need to split the latent z.
Example:
"""
def __init__(self, in_dim, cond_dim, n_flow_steps,
conv_dim_channel, conv_kernel_size):
"""BlowBlock(in_dim, cond_dim, n_flow_steps,
conv_dim_channel, conv_kernel_size)
Args
----
in_dim: int, input feature dim, (batch, length, in_dim)
cond_dim:, int, conditional feature dim, (batch, length, cond_dim)
n_flow_steps: int, number of flow steps in one block
conv_dim_channel: int, dim of the conv residual and skip channels
conv_kernel_size: int, kernel size of the convolution layers
For conv_dim_channel and conv_kernel_size, see AffineCouplingBlow
"""
super(BlowBlock, self).__init__()
# squeeze
self.m_squeeze = SqueezeForBlow()
squeezed_feat_dim = in_dim * self.m_squeeze.get_squeeze_factor()
# flow steps
tmp_flows = []
for i in range(n_flow_steps):
tmp_flows.append(
FlowStepBlow(
squeezed_feat_dim, cond_dim,
conv_dim_channel, conv_kernel_size))
self.m_flows = torch_nn.ModuleList(tmp_flows)
self.m_out_dim = squeezed_feat_dim
return
def get_out_feat_dim(self):
return self.m_out_dim
def get_expected_squeeze_length(self, orig_length):
return self.m_squeeze.get_expected_squeeze_length(orig_length)
def forward(self, y, cond, factor=1):
"""z, log_detjac = BlowBlock(y)
y -> squeeze -> H() -> z, log_det_jacobian
H() consists of multiple flow steps (1x1conv + Actnorm + AffineCoupling)
input
-----
y: tensor, (batch, length, dim)
cond, tensor, (batch, 1, cond_dim)
factor, None or int, this is used to divde likelihood, default 1
output
------
log_detjac: tensor or scalar
z: tensor, (batch, length, dim), for N(z; 0, I) or next flow block
"""
# squeeze
x_tmp = self.m_squeeze(y)
# flows
log_detjac = 0
for idx, l_flow in enumerate(self.m_flows):
x_tmp, log_detjac_tmp = l_flow(x_tmp, cond, factor)
log_detjac = log_detjac + log_detjac_tmp
return x_tmp, log_detjac
def reverse(self, z, cond):
"""y = BlowBlock.reverse(z, cond)
z -> H^{-1}() -> unsqueeze -> y
input
-----
z: tensor, (batch, length, in_dim)
cond, tensor, (batch, 1, cond_dim)
output
------
y: tensor, (batch, length, in_dim)
"""
y_tmp = z
for l_flow in self.m_flows[::-1]:
y_tmp = l_flow.reverse(y_tmp, cond)
y = self.m_squeeze.reverse(y_tmp)
return y
class Blow(torch_nn.Module):
"""Blow
"""
def __init__(self, cond_dim, num_blocks, num_flows_inblock,
conv_dim_channel, conv_kernel_size):
"""Blow(cond_dim, num_blocks, num_flows_inblock,
conv_dim_channel, conv_kernel_size)
Args
----
cond_dim:, int, conditional feature dim, (batch, length, cond_dim)
num_blocks: int, number of WaveGlowBlocks
num_flows_inblock: int, number of flow steps in one WaveGlowBlock
conv_dim_channel: int, dim of convolution layers channels
conv_kernel_size: int, kernel size of the convolution layers
This model defines:
cond (global) ----- -> | ------> | --------> |
v v v
y --------------> BlowBlock1 -> BlowBlock2 -> ... -> z
"""
super(Blow, self).__init__()
# input is assumed to be waveform
self.m_input_dim = 1
# save the dimension for get_z_noises
self.m_z_dim = 0
# define blocks
tmp_squeezed_in_dim = self.m_input_dim
tmp_flow_blocks = []
for i in range(num_blocks):
tmp_flow_blocks.append(
BlowBlock(
tmp_squeezed_in_dim, cond_dim, num_flows_inblock,
conv_dim_channel, conv_kernel_size))
tmp_squeezed_in_dim = tmp_flow_blocks[-1].get_out_feat_dim()
self.m_z_dim = tmp_squeezed_in_dim
self.m_flowblocks = torch_nn.ModuleList(tmp_flow_blocks)
# done
return
def get_expected_squeeze_length(self, wave_length):
"""length = get_expected_squeeze_length(self, wave_length)
Return expected length of latent z
input
-----
wave_length: int, length of original waveform
output
------
length: int, length of latent z
"""
length = wave_length
for glowblock in self.m_flowblocks:
length = glowblock.get_expected_squeeze_length(length)
return length
def _normal_lh(self, noise):
# likelihood of normal distribution on the given noise
return -0.5 * np.log(2 * np.pi) - 0.5 * noise ** 2
def forward(self, y, cond):
"""z, neg_logp_y, logp_z, logdet = Blow.forward(y, cond)
cond (global) ----- -> | ------> | --------> |
v v v
y --------------> BlowBlock1 -> BlowBlock2 -> ... -> z
input
-----
y: tensor, (batch, waveform_length, 1)
cond: tensor, (batch, 1, cond_dim)
output
------
z: tensor
neg_logp_y: scalar, - log p(y)
logp_z: scalar, -log N(z), summed over one data sequence, but averaged
over batch.
logdet: scalar, -|det dH(.)/dy|, summed over one data sequence,
but averaged
over batch.
"""
# Rather than summing the likelihood and divide it by the number of
# data in the final step, we divide this factor from the likelihood
# caculating by each flow step and sum the scaled likelihood.
# Two methods are equivalent, but the latter may prevent numerical
# overflow of the likelihood value for long sentences
factor = np.prod([dim for dim in y.shape])
# flows
log_detjac = 0
log_pz = 0
x_tmp = y
for m_block in self.m_flowblocks:
x_tmp, log_detjac_tmp = m_block(
x_tmp, cond, factor)
# accumulate log det jacobian
log_detjac += log_detjac_tmp
z_tmp = x_tmp
# compute N(z; 0, I)
# accumulate log_N(z; 0, I) only if it is valid
if z_tmp is not None:
log_pz += nii_glow.sum_over_keep_batch2(
self._normal_lh(z_tmp), factor)
# average over batch and data points
neg_logp_y = -(log_pz + log_detjac).sum()
return z_tmp, neg_logp_y, \
log_pz.sum(), log_detjac.sum()
def reverse(self, z, cond):
"""y = Blow.reverse(z_bags, cond)
cond (global) ----- -> | ------> | --------> |
v v v
y <--------------- BlowBlock1 <- BlowBlock2 <- ... <- z
input
-----
z: tensor, shape decided by the model configuration
cond: tensor, (batch, 1, cond_dim)
output
------
y: tensor, (batch, waveform_length, 1)
"""
# initial
y_tmp = z
for m_block in self.m_flowblocks[::-1]:
y_tmp = m_block.reverse(y_tmp, cond)
return y_tmp
def get_z_noises(self, length, noise_std=0.7, batchsize=1):
"""z_bags = Blow.get_z_noises(length, noise_std=0.7, batchsize=1)
Return random noise for random sampling
input
-----
length: int, length of target waveform (without squeeze)
noise_std: float, std of Gaussian noise, default 0.7
batchsize: int, batch size of this random data, default 1
output
------
z: tensor, shape decided by the network
Blow.reverse(z, cond) can be used to generate waveform
"""
squeeze_length = self.get_expected_squeeze_length(length)
device = next(self.parameters()).device
z_tmp = torch.randn(
[batchsize, squeeze_length, self.m_z_dim],
dtype=nii_io_conf.d_dtype,
device=device)
return z_tmp
if __name__ == "__main__":
print("Definition of Blow")
| 31,665 | 32.367756 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/util_bayesian.py | #!/usr/bin/env python
"""
util_bayesian.py
Utilities for bayeisan neural network
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn.functional as torch_nn_func
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
######
# utils to save guide and model Pyro
# not used anymore
######
def save_model_guide(model, guide, path_model, path_guide):
#torch.save(dnn_net.state_dict(), "mnist_cnn_{:03d}.pt".format(idx))
torch.save({"model" : model.state_dict(),
"guide" : guide}, path_model)
pyro.get_param_store().save(path_guide)
return
def load_model_guide(model, path_model, path_guide):
pretrained = torch.load(path_model)
model.load_state_dict(pretrained['model'])
guide = pretrained['guide']
pyro.get_param_store().load(path_guide)
return guide
######
# Utils to compute metrics for Bayesian inference
######
def _xent(y, dim=-1, log_floor = 0.000001):
"""xe = xent(y, dim)
input: y, tensor, (..., num_classes), probablity matrix
input: dim, int, along which dimension we do xent? default -1
output: xe, tensor, (..., 1), xe = -sum_j y[j] log y[j]
"""
logfloor = torch.zeros_like(y)
logfloor[y < log_floor] = log_floor
return -torch.sum(y * torch.log(y + logfloor), dim=dim, keepdim=True)
def xent(p):
"""mi = xent(p)
This measures total uncertainty
input: p, tensor, (sammple_N, batch, num_classes), probablity
output: xe, tensor, (batch, 1)
"""
# step1. Bayesian model average p(y | x, D) = E_{q_w}[p(y | w, x)]
# -> 1/N sum_i p(y | w_i, x)
# mp (batch, num_classes)
mp = p.mean(dim=0)
# step2. cross entropy over p(y | x, D)
# xe (batch, 1)
xe = _xent(mp)
return xe
def compute_epstemic_uncertainty(y):
"""mi = mutual_infor(y)
This measures epstemic uncertainty
input: y, tensor, (sammple_N, batch, num_classes), probablity
output: mi, tensor, (batch, 1)
"""
# cross entropy over BMA prob, see xent() above
xe = xent(y)
# cross entropy over each individual sample, ve (sample_N, batch, 1)
# for w_i, compute ent_i = xent(p(y | w_i, x))
# then, ve = 1/N sum_i ent_i
ve = torch.mean(_xent(y), dim=0)
# xe - ve
mi = xe - ve
return mi
def compute_aleatoric_uncertainty(y):
"""mi = mutual_infor(y)
This measures aleatoric uncertainty
input: y, tensor, (sammple_N, batch, num_classes), probablity
output: mi, tensor, (batch, 1)
"""
ve = torch.mean(_xent(y), dim=0)
return ve
def compute_logit_from_prob(y, log_floor=0.0000001):
"""logit = compute_logit_from_prob(y)
input: y, tensor, any shape, probablity of being positive
output: logit, tensor, same shape as y, sigmoid(logit) is y
"""
logfloor = torch.zeros_like(y)
logfloor[y < log_floor] = log_floor
tmp = 1 / (y + logfloor) - 1
logfloor = logfloor * 0
logfloor[tmp < log_floor] = log_floor
logit = - torch.log(tmp + logfloor)
return logit
#####
# wrapper
#####
def compute_llr_eps_ale(logits, idx_pos=1):
"""llr, eps, ale = compute_llr_eps_ale(logits)
input: logits, tensor (sampling_num, batch, 2)
idx_pos, int, which dimension is the positive class?
(default 1, which means logits[:, :, 1])
output: llr, tensor, (batch, 1)
eps, tensor, (batch, 1)
ale, tensor, (batch, 1)
"""
# -> (sampling_num, batch, 2)
prob = torch_nn_func.softmax(logits, dim=-1)
# to LLR
# 1. average prob over the samples to (batch, num_class)
# 2. compute the llr
averaged_prob = torch.mean(prob, dim=0)
# unsqueeze to make the shape consistent
llr = compute_logit_from_prob(averaged_prob[..., idx_pos]).unsqueeze(-1)
# get uncertainty
eps = compute_epstemic_uncertainty(prob)
ale = compute_aleatoric_uncertainty(prob)
return llr, eps, ale
if __name__ == "__main__":
print("Package for util_bayesian")
| 4,124 | 26.684564 | 76 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/util_dsp.py | #!/usr/bin/env python
"""
util_dsp.py
Utilities for signal processing
MuLaw Code adapted from
https://github.com/fatchord/WaveRNN/blob/master/utils/distribution.py
DCT code adapted from
https://github.com/zh217/torch-dct
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020-2021, Xin Wang"
######################
### WaveForm utilities
######################
def label_2_float(x, bits):
"""output = label_2_float(x, bits)
Assume x is code index for N-bits, then convert x to float values
Note: dtype conversion is not handled
inputs:
-----
x: data to be converted Tensor.long or int, any shape.
x value should be [0, 2**bits-1]
bits: number of bits, int
Return:
-------
output: tensor.float, [-1, 1]
output = 2 * x / (2**bits - 1.) - 1.
"""
return 2 * x / (2**bits - 1.) - 1.
def float_2_label(x, bits):
"""output = float_2_label(x, bits)
Assume x is a float value, do N-bits quantization and
return the code index.
input
-----
x: data to be converted, any shape
x value should be [-1, 1]
bits: number of bits, int
output
------
output: tensor.float, [0, 2**bits-1]
Although output is quantized, we use torch.float to save
the quantized values
"""
#assert abs(x).max() <= 1.0
# scale the peaks
peak = torch.abs(x).max()
if peak > 1.0:
x /= peak
# quantize
x = (x + 1.) * (2**bits - 1) / 2
return torch.clamp(x, 0, 2**bits - 1)
def mulaw_encode(x, quantization_channels, scale_to_int=True):
"""x_mu = mulaw_encode(x, quantization_channels, scale_to_int=True)
Adapted from torchaudio
https://pytorch.org/audio/functional.html mu_law_encoding
input
-----
x (Tensor): Input tensor, float-valued waveforms in (-1, 1)
quantization_channels (int): Number of channels
scale_to_int: Bool
True: scale mu-law to int
False: return mu-law in (-1, 1)
output
------
x_mu: tensor, int64, Input after mu-law encoding
"""
# mu
mu = quantization_channels - 1.0
# no check on the value of x
if not x.is_floating_point():
x = x.to(torch.float)
mu = torch.tensor(mu, dtype=x.dtype, device=x.device)
x_mu = torch.sign(x) * torch.log1p(mu * torch.abs(x)) / torch.log1p(mu)
if scale_to_int:
x_mu = ((x_mu + 1) / 2 * mu + 0.5).to(torch.int64)
return x_mu
def mulaw_decode(x_mu, quantization_channels, input_int=True):
"""Adapted from torchaudio
https://pytorch.org/audio/functional.html mu_law_encoding
Args:
x_mu (Tensor): Input tensor
quantization_channels (int): Number of channels
input_int: Bool
True: convert x_mu (int) from int to float, before mu-law decode
False: directly decode x_mu (float)
Returns:
Tensor: Input after mu-law decoding (float-value waveform (-1, 1))
"""
mu = quantization_channels - 1.0
if not x_mu.is_floating_point():
x_mu = x_mu.to(torch.float)
mu = torch.tensor(mu, dtype=x_mu.dtype, device=x_mu.device)
if input_int:
x = ((x_mu) / mu) * 2 - 1.0
else:
x = x_mu
x = torch.sign(x) * (torch.exp(torch.abs(x) * torch.log1p(mu)) - 1.0) / mu
return x
######################
### DCT utilities
### https://github.com/zh217/torch-dct
### LICENSE: MIT
###
######################
def rfft_wrapper(x, onesided=True, inverse=False):
# compatiblity with torch fft API
if hasattr(torch, "rfft"):
# for torch < 1.8.0, rfft is the API to use
# torch 1.7.0 complains about this API, but it is OK to use
if not inverse:
# FFT
return torch.rfft(x, 1, onesided=onesided)
else:
# inverse FFT
return torch.irfft(x, 1, onesided=onesided)
else:
# for torch > 1.8.0, fft.rfft is the API to use
if not inverse:
# FFT
if onesided:
data = torch.fft.rfft(x)
else:
data = torch.fft.fft(x)
return torch.stack([data.real, data.imag], dim=-1)
else:
# It requires complex-tensor
real_image = torch.chunk(x, 2, dim=1)
x = torch.complex(real_image[0].squeeze(-1),
real_image[1].squeeze(-1))
if onesided:
return torch.fft.irfft(x)
else:
return torch.fft.ifft(x)
def dct1(x):
"""
Discrete Cosine Transform, Type I
:param x: the input signal
:return: the DCT-I of the signal over the last dimension
"""
x_shape = x.shape
x = x.view(-1, x_shape[-1])
return rfft_wrapper(
torch.cat([x, x.flip([1])[:, 1:-1]], dim=1))[:, :, 0].view(*x_shape)
def idct1(X):
"""
The inverse of DCT-I, which is just a scaled DCT-I
Our definition if idct1 is such that idct1(dct1(x)) == x
:param X: the input signal
:return: the inverse DCT-I of the signal over the last dimension
"""
n = X.shape[-1]
return dct1(X) / (2 * (n - 1))
def dct(x, norm=None):
"""
Discrete Cosine Transform, Type II (a.k.a. the DCT)
For the meaning of the parameter `norm`, see:
https://docs.scipy.org/doc/ scipy.fftpack.dct.html
:param x: the input signal
:param norm: the normalization, None or 'ortho'
:return: the DCT-II of the signal over the last dimension
"""
x_shape = x.shape
N = x_shape[-1]
x = x.contiguous().view(-1, N)
v = torch.cat([x[:, ::2], x[:, 1::2].flip([1])], dim=1)
Vc = rfft_wrapper(v, onesided=False)
k = - torch.arange(N, dtype=x.dtype, device=x.device)[None, :] * np.pi/(2*N)
W_r = torch.cos(k)
W_i = torch.sin(k)
V = Vc[:, :, 0] * W_r - Vc[:, :, 1] * W_i
if norm == 'ortho':
V[:, 0] /= np.sqrt(N) * 2
V[:, 1:] /= np.sqrt(N / 2) * 2
V = 2 * V.view(*x_shape)
return V
def idct(X, norm=None):
"""
The inverse to DCT-II, which is a scaled Discrete Cosine Transform, Type III
Our definition of idct is that idct(dct(x)) == x
For the meaning of the parameter `norm`, see:
https://docs.scipy.org/doc/ scipy.fftpack.dct.html
:param X: the input signal
:param norm: the normalization, None or 'ortho'
:return: the inverse DCT-II of the signal over the last dimension
"""
x_shape = X.shape
N = x_shape[-1]
X_v = X.contiguous().view(-1, x_shape[-1]) / 2
if norm == 'ortho':
X_v[:, 0] *= np.sqrt(N) * 2
X_v[:, 1:] *= np.sqrt(N / 2) * 2
k = torch.arange(x_shape[-1], dtype=X.dtype,
device=X.device)[None, :]*np.pi/(2*N)
W_r = torch.cos(k)
W_i = torch.sin(k)
V_t_r = X_v
V_t_i = torch.cat([X_v[:, :1] * 0, -X_v.flip([1])[:, :-1]], dim=1)
V_r = V_t_r * W_r - V_t_i * W_i
V_i = V_t_r * W_i + V_t_i * W_r
V = torch.cat([V_r.unsqueeze(2), V_i.unsqueeze(2)], dim=2)
v = rfft_wrapper(V, onesided=False, inverse=True)
x = v.new_zeros(v.shape)
x[:, ::2] += v[:, :N - (N // 2)]
x[:, 1::2] += v.flip([1])[:, :N // 2]
return x.view(*x_shape)
class LinearDCT(torch_nn.Linear):
"""DCT implementation as linear transformation
Original Doc is in:
https://github.com/zh217/torch-dct/blob/master/torch_dct/_dct.py
This class implements DCT as a linear transformation layer.
This layer's weight matrix is initialized using the DCT transformation mat.
Accordingly, this API assumes that the input signal has a fixed length.
Please pad or trim the input signal when using this LinearDCT.forward(x)
Args:
----
in_features: int, which is equal to expected length of the signal.
type: string, dct1, idct1, dct, or idct
norm: string, ortho or None, default None
bias: bool, whether add bias to this linear layer. Default None
"""
def __init__(self, in_features, type, norm=None, bias=False):
self.type = type
self.N = in_features
self.norm = norm
super(LinearDCT, self).__init__(in_features, in_features, bias=bias)
def reset_parameters(self):
# initialise using dct function
I = torch.eye(self.N)
if self.type == 'dct1':
self.weight.data = dct1(I).data.t()
elif self.type == 'idct1':
self.weight.data = idct1(I).data.t()
elif self.type == 'dct':
self.weight.data = dct(I, norm=self.norm).data.t()
elif self.type == 'idct':
self.weight.data = idct(I, norm=self.norm).data.t()
self.weight.requires_grad = False # don't learn this!
if __name__ == "__main__":
print("util_dsp.py")
| 9,027 | 27.935897 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/util_music.py | #!/usr/bin/env python
"""
util_music.py
Utilities for music applications
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.dynamic_prog as nii_dy
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
class HzCentConverter(torch_nn.Module):
"""
HzCentConverter: an interface to convert F0 to cent, probablity matrix
and do reverse conversions
"""
def __init__(self,
u_hz = 0,
hz_ref = 10,
base_hz = 31.77219916398751,
top_hz = 2033.4207464952,
bins = 360,
cent_1st = 32.70319566257483,
cent_last = 1975.5332050244956,
):
super(HzCentConverter, self).__init__()
# unvoiced F0
self.m_v_hz = u_hz
# reference for cent calculation
self.m_hz_ref = hz_ref
# quantized resolution
# originally, bins = 360 -> 12 * 6 * 5, 12 semitones, 6 octaves
# each semitone is further divided to 5 intervals
self.m_fine_interval = 5
#####
# for quantization
#####
# one semitone cover 100 cents
# thus, -50 on bottom, +50 on top
# cent2hz(hz2cent(librosa.note_to_hz('C1'))-50)
self.m_base_hz = torch.tensor([base_hz])
# cent2hz(hz2cent(librosa.note_to_hz('B6'))+50)
self.m_top_hz = torch.tensor([top_hz])
# quantization interval
self.m_bins = bins
self.m_base_cent = self.hz2cent(self.m_base_hz)
self.m_top_cent = self.hz2cent(self.m_top_hz)
#####
# for de-quantization
#####
# librosa.note_to_hz('C1')
self.m_1st_cent = self.hz2cent(torch.tensor([cent_1st]))
# librosa.note_to_hz('B6')
self.m_last_cent = self.hz2cent(torch.tensor([cent_last]))
# quantized cent per bin
self.m_quan_cent_dis = (self.m_last_cent - self.m_1st_cent)/self.m_bins
# quantized cents as a tentor
self.m_dis_cent = torch_nn.Parameter(
torch.linspace(self.m_1st_cent.numpy()[0],
self.m_last_cent.numpy()[0],
self.m_bins),
requires_grad=False)
# quantized F0 as a tensor
self.m_dis_f0 = self.cent2hz(
torch.linspace(self.m_1st_cent.numpy()[0],
self.m_last_cent.numpy()[0],
self.m_bins))
#####
# for viterbi decoding
#####
self.m_viterbi_decode = True
# initial state probablity
self.m_viterbi_init = np.ones(self.m_bins * 2) / (self.m_bins * 2)
# transition probability
def _trans_mat():
max_cent = 12
p_vv = 0.99
p_uv = 1 - p_vv
# transition probabilities inducing continuous pitch
xx, yy = np.meshgrid(range(self.m_bins), range(self.m_bins))
tran_m_v = np.maximum(max_cent - abs(xx - yy), 0)
tran_m_v = tran_m_v / np.sum(tran_m_v, axis=1)[:, None]
# unvoiced part
tran_m_u = np.ones([self.m_bins, self.m_bins])/self.m_bins
tmp1 = np.concatenate([tran_m_v * p_vv, tran_m_u * p_uv], axis=1)
tmp2 = np.concatenate([tran_m_v * p_uv, tran_m_u * p_vv], axis=1)
trans = np.concatenate([tmp1, tmp2], axis=0)
return trans
self.m_viterbi_tran = _trans_mat()
def hz2cent(self, hz):
"""
hz2cent(self, hz)
Convert F0 Hz in to Cent
Parameters
----------
hz: torch.tensor
Return
------
: torch.tensor
"""
return 1200 * torch.log2(hz/self.m_hz_ref)
def cent2hz(self, cent):
return torch.pow(2, cent/1200) * self.m_hz_ref
def quantize_hz(self, hz):
cent = self.hz2cent(hz)
q_bin = torch.round((cent - self.m_base_cent) * self.m_bins /\
(self.m_top_cent - self.m_base_cent))
q_bin = torch.min([torch.max([0, q_bin]), self.m_bins - 1]) +1
return q_bin
def dequantize_hz(self, quantized_cent):
cent = quantized_cent * self.m_quan_cent_dis + self.m_1st_cent
return self.cent2hz(cent)
def f0_to_mat(self, f0_seq, var=625):
"""
f0_to_mat(self, f0_seq)
Convert F0 sequence (hz) into a probability matrix.
Jong Wook Kim, Justin Salamon, Peter Li, and Juan Pablo Bello. 2018.
CREPE: A Convolutional Representation for Pitch Estimation.
In Proc. ICASSP, 161-165
Parameters
----------
f0_seq: torch.tensor (1, N, 1)
Return
------
target_mat: torch.tensor (1, N, bins)
created probability matrix for f0
"""
if f0_seq.dim() != 3:
print("f0 sequence loaded in tensor should be in shape (1, N, 1)")
sys.exit(1)
# voiced / unvoiced indix
v_idx = f0_seq > self.m_v_hz
u_idx = ~v_idx
# convert F0 Hz to cent
target = torch.zeros_like(f0_seq)
target[v_idx] = self.hz2cent(f0_seq[v_idx])
target[u_idx] = 0
# target
# since target is (1, N, 1), the last dimension size is 1
# self.m_dis_cent (bins) -> propagated to (1, N, bins)
target_mat = torch.exp(-torch.pow(self.m_dis_cent - target, 2)/2/var)
# set unvoiced to zero
for idx in range(target_mat.shape[0]):
target_mat[idx, u_idx[idx, :, 0], :] *= 0.0
#target_mat[0, u_idx[0, :, 0], :] *= 0.0
# return
return target_mat
def recover_f0(self, bin_mat, viterbi_decode=True):
"""
recover_f0(self, bin_mat)
Produce F0 from a probability matrix.
This is the inverse function of f0_to_mat.
By default, use Viterbi decoding to produce F0.
Matthias Mauch, and Simon Dixon. 2014.
PYIN: A Fundamental Frequency Estimator Using Probabilistic
Threshold Distributions. In Proc. ICASSP, 659-663.
Parameters
----------
bin_mat: torch.tensor (1, N, bins)
Return
------
f0: torch.tensor(1, N, 1)
"""
# check
if bin_mat.shape[0] != 1:
print("F0 generation only support batchsize=1")
sys.exit(1)
if bin_mat.dim() != 3 or bin_mat.shape[-1] != self.m_bins:
print("bin_mat should be in shape (1, N, bins)")
sys.exit(1)
# generation
if not self.m_viterbi_decode or not viterbi_decode:
# denominator
prob_sum = torch.sum(bin_mat, axis=2)
# add floor
prob_sum[prob_sum < 1e-07] = 1e-07
# normal sum
cent = torch.sum(bin_mat * self.m_dis_cent, axis=2) / prob_sum
f0 = self.cent2hz(cent)
f0[cent < 1] = 0
return f0.unsqueeze(-1)
else:
tmp_bin_mat = bin_mat.to('cpu')
# viterbi decode:
with torch.no_grad():
# observation probablity for unvoiced states
prob_u = torch.ones_like(tmp_bin_mat) \
- torch.mean(tmp_bin_mat, axis=2, keepdim=True)
# concatenate to observation probability matrix
# [Timestep, m_bins * 2],
# m_bins is the number of quantized F0 bins
# another m_bins is for the unvoiced states
tmp_bin_mat = torch.cat([tmp_bin_mat, prob_u],axis=2).squeeze(0)
# viterbi decoding. Numpy is fast?
tmp_bin_mat = tmp_bin_mat.numpy()
quantized_cent = nii_dy.viterbi_decode(
self.m_viterbi_init, self.m_viterbi_tran, tmp_bin_mat * 0.5)
# unvoiced state sequence (states in [m_bins, m_bins*2])
u_idx = quantized_cent>=self.m_bins
# based on viterbi best state, do weighted sum over a beam
# Equation from
# https://github.com/marl/crepe/blob/master/crepe/core.py#L108
prob_m = torch.zeros_like(bin_mat)
for idx, i in enumerate(quantized_cent):
s_idx = np.max([i - 4, 0])
e_idx = np.min([i+5, self.m_bins])
prob_m[0, idx, s_idx:e_idx] = bin_mat[0, idx, s_idx:e_idx]
cent = torch.sum(prob_m * self.m_dis_cent, axis=2) / \
torch.sum(prob_m, axis=2)
# from cent to f0
f0 = self.cent2hz(cent)
# unvoiced
f0[0, u_idx]=0
return f0.unsqueeze(-1)
def f0_probmat_postprocessing(self, f0_prob_mat):
"""
f0_prob_mat = f0_prob_mat_post(f0_prob_mat)
input
-----
f0_prob_mat: torch tensor of shape (bathcsize, length, bins)
output
------
f0_prob_mat_new: same shape as f0_prob_mat
"""
if f0_prob_mat.shape[-1] != self.m_bins:
print("Last dimension of F0 prob mat != {:d}".format(self.m_bins))
sys.exit(1)
if f0_prob_mat.shape[0] > 1:
print("Cannot support batchsize > 1 for dynamic programming")
sys.exit(1)
# observation probablity for unvoiced states
prob_u = torch.ones_like(f0_prob_mat) \
- torch.mean(f0_prob_mat, axis=2, keepdim=True)
tmp_bin_mat = torch.cat([f0_prob_mat, prob_u],axis=2).squeeze(0)
# viterbi decoding. Numpy is fast?
tmp_bin_mat = tmp_bin_mat.to('cpu').numpy()
quantized_cent = nii_dy.viterbi_decode(
self.m_viterbi_init, self.m_viterbi_tran, tmp_bin_mat * 0.5)
u_idx = quantized_cent>=self.m_bins
mat_new = torch.zeros_like(f0_prob_mat)
for idx, i in enumerate(quantized_cent):
if i < self.m_bins:
sidx = np.max([i - 4, 0])
eidx = np.min([i+5, self.m_bins])
mat_new[0, idx, sidx:eidx] = f0_prob_mat[0,idx,sidx:eidx]
mat_new[0, idx, sidx:eidx] /= mat_new[0, idx, sidx:eidx].sum()
return mat_new
if __name__ == "__main__":
print("util_music")
| 10,763 | 32.742947 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/different_main/main_gan.py | #!/usr/bin/env python
"""
main.py for project-NN-pytorch/projects
The default training/inference process wrapper
Requires model.py and config.py
Usage: $: python main.py [options]
"""
from __future__ import absolute_import
import os
import sys
import torch
import importlib
import core_scripts.data_io.default_data_io as nii_default_dset
import core_scripts.data_io.customize_dataset as nii_dset
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.list_tools as nii_list_tool
import core_scripts.config_parse.config_parse as nii_config_parse
import core_scripts.config_parse.arg_parse as nii_arg_parse
import core_scripts.op_manager.op_manager as nii_op_wrapper
import core_scripts.nn_manager.nn_manager as nii_nn_wrapper
import core_scripts.nn_manager.nn_manager_GAN as nii_nn_wrapper_GAN
import core_scripts.startup_config as nii_startup
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
def main():
""" main(): the default wrapper for training and inference process
Please prepare config.py and model.py
"""
# arguments initialization
args = nii_arg_parse.f_args_parsed()
#
nii_warn.f_print_w_date("Start program", level='h')
nii_warn.f_print("Load module: %s" % (args.module_config))
nii_warn.f_print("Load module: %s" % (args.module_model))
prj_conf = importlib.import_module(args.module_config)
prj_model = importlib.import_module(args.module_model)
# initialization
nii_startup.set_random_seed(args.seed, args)
use_cuda = not args.no_cuda and torch.cuda.is_available()
device = torch.device("cuda" if use_cuda else "cpu")
# prepare data io
if not args.inference and not args.epoch2pt:
params = {'batch_size': args.batch_size,
'shuffle': args.shuffle,
'num_workers': args.num_workers,
'sampler': args.sampler,
'pin_memory': True}
in_trans_fns = prj_conf.input_trans_fns \
if hasattr(prj_conf, 'input_trans_fns') else None
out_trans_fns = prj_conf.output_trans_fns \
if hasattr(prj_conf, 'output_trans_fns') else None
inout_trans_fns = prj_conf.input_output_trans_fn \
if hasattr(prj_conf, 'input_output_trans_fn') else None
# Load file list and create data loader
trn_lst = prj_conf.trn_list
trn_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.trn_set_name, \
trn_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./',
params = params,
truncate_seq = prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = True,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns,
inoutput_augment_func = inout_trans_fns)
if prj_conf.val_list is not None:
val_lst = prj_conf.val_list
val_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.val_set_name,
val_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./', \
params = params,
truncate_seq= prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns,
inoutput_augment_func = inout_trans_fns)
else:
val_set = None
# initialize the model and loss function
model_G = prj_model.ModelGenerator(
trn_set.get_in_dim(), trn_set.get_out_dim(), \
args, prj_conf, trn_set.get_data_mean_std())
model_D = prj_model.ModelDiscriminator(
trn_set.get_in_dim(), trn_set.get_out_dim(),
args, prj_conf, trn_set.get_data_mean_std())
loss_wrapper = None
# initialize the optimizer
optimizer_G_wrap = nii_op_wrapper.OptimizerWrapper(model_G, args)
optimizer_D_wrap = nii_op_wrapper.OptimizerWrapper(model_D, args)
# if necessary, resume training
if args.trained_model == "":
checkpoint_G = None
checkpoint_D = None
else:
tmp_str = args.trained_model.split(",")
checkpoint_G = torch.load(tmp_str[0])
if len(tmp_str) > 1:
checkpoint_D = torch.load(tmp_str[1])
else:
checkpoint_D = None
# start training
nii_nn_wrapper_GAN.f_train_wrapper_GAN(
args, model_G, model_D,
loss_wrapper, device,
optimizer_G_wrap, optimizer_D_wrap,
trn_set, val_set,
checkpoint_G, checkpoint_D)
# done for traing
elif args.inference:
# for inference
# default, no truncating, no shuffling
params = {'batch_size': args.batch_size,
'shuffle': False,
'num_workers': args.num_workers}
in_trans_fns = prj_conf.input_trans_fns \
if hasattr(prj_conf, 'test_input_trans_fns') else None
out_trans_fns = prj_conf.output_trans_fns \
if hasattr(prj_conf, 'test_output_trans_fns') else None
inout_trans_fns = prj_conf.output_trans_fns \
if hasattr(prj_conf, 'test_input_output_trans_fn') \
else None
if type(prj_conf.test_list) is list:
t_lst = prj_conf.test_list
else:
t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list)
test_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.test_set_name, \
t_lst, \
prj_conf.test_input_dirs,
prj_conf.input_exts,
prj_conf.input_dims,
prj_conf.input_reso,
prj_conf.input_norm,
prj_conf.test_output_dirs,
prj_conf.output_exts,
prj_conf.output_dims,
prj_conf.output_reso,
prj_conf.output_norm,
'./',
params = params,
truncate_seq = None,
min_seq_len = None,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns,
inoutput_augment_func = inout_trans_fns)
# initialize model
model = prj_model.ModelGenerator(
test_set.get_in_dim(), test_set.get_out_dim(), args, prj_conf)
if args.trained_model == "":
print("Please provide ---trained-model")
sys.exit(1)
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper.f_inference_wrapper(
args, model, device, test_set, checkpoint)
elif args.epoch2pt:
# for model conversion from epoch.pt to trained_network.pt
# initialize model
model = prj_model.ModelGenerator(
sum(prj_conf.input_dims), sum(prj_conf.output_dims), args, prj_conf)
if args.trained_model == "":
print("Please provide ---trained-model")
sys.exit(1)
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper.f_convert_epoch_to_trained(
args, model, device, checkpoint)
else:
print("Fatal error in main.py")
sys.exit(1)
# done
return
if __name__ == "__main__":
main()
| 9,039 | 35.16 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/different_main/config_merge_datasets.py | #!/usr/bin/env python
"""
config.py
To merge different corpora (or just one corpus),
*_set_name are lists
*_list are lists of lists
*_dirs are lists of lists
"""
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#########################################################
## Configuration for training stage
#########################################################
# Name of datasets
# after data preparation, trn/val_set_name are used to save statistics
# about the data sets
trn_set_name = ['cmu_all_trn']
val_set_name = ['cmu_all_val']
# for convenience
tmp = '../DATA/cmu-arctic-data-set'
# File lists (text file, one data name per line, without name extension)
# trin_file_list: list of files for training set
trn_list = [tmp + '/scp/train.lst']
# val_file_list: list of files for validation set. It can be None
val_list = [tmp + '/scp/val.lst']
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# we assume train and validation data are put in the same sub-directory
input_dirs = [[tmp + '/5ms/melspec', tmp + '/5ms/f0']]
# Dimensions of input features
# input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...]
input_dims = [80, 1]
# File name extension for input features
# input_exts = [name_extention_of_feature_1, ...]
# Please put ".f0" as the last feature
input_exts = ['.mfbsp', '.f0']
# Temporal resolution for input features
# input_reso = [reso_feature_1, reso_feature_2, ...]
# for waveform modeling, temporal resolution of input acoustic features
# may be = waveform_sampling_rate * frame_shift_of_acoustic_features
# for example, 80 = 16000 Hz * 5 ms
input_reso = [80, 80]
# Whether input features should be z-normalized
# input_norm = [normalize_feature_1, normalize_feature_2]
input_norm = [True, True]
# Similar configurations for output features
output_dirs = [[tmp + '/wav_16k_norm']]
output_dims = [1]
output_exts = ['.wav']
output_reso = [1]
output_norm = [False]
# Waveform sampling rate
# wav_samp_rate can be None if no waveform data is used
wav_samp_rate = 16000
# Truncating input sequences so that the maximum length = truncate_seq
# When truncate_seq is larger, more GPU mem required
# If you don't want truncating, please truncate_seq = None
truncate_seq = 16000 * 3
# Minimum sequence length
# If sequence length < minimum_len, this sequence is not used for training
# minimum_len can be None
minimum_len = 80 * 50
# Optional argument
# Just a buffer for convenience
# It can contain anything
optional_argument = ['']
# Data transformation function, you can import here
# these functions are applied before casting np.array data into tensor arrays
#
#input_trans_fns = [[func_for_mel, fun_for_f0]]
#output_trans_fns = [[func_for_wav]]
#########################################################
## Configuration for inference stage
#########################################################
# similar options to training stage
test_set_name = ['cmu_all_test_tiny']
# List of test set data
# for convenience, you may directly load test_set list here
test_list = [['slt_arctic_b0474', 'slt_arctic_b0475', 'slt_arctic_b0476',
'bdl_arctic_b0474', 'bdl_arctic_b0475', 'bdl_arctic_b0476',
'rms_arctic_b0474', 'rms_arctic_b0475', 'rms_arctic_b0476',
'clb_arctic_b0474', 'clb_arctic_b0475', 'clb_arctic_b0476']]
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# we assume train and validation data are put in the same sub-directory
test_input_dirs = [[tmp + '/5ms/melspec', tmp + '/5ms/f0']]
# Directories for output features, which are []
test_output_dirs = [[]]
# Data transformation function, you can import here
# these functions are applied before casting np.array data into tensor arrays
#
#test_input_trans_fns = [[func_for_mel, fun_for_f0]]
#test_output_trans_fns = [[func_for_wav]]
| 3,958 | 31.45082 | 78 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/different_main/temp.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import core_scripts.other_tools.debug as nii_debug
import sandbox.block_nn as nii_nn
import sandbox.block_nsf as nii_nsf
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#####
## Model definition
#####
class Conv1dNoPermute(torch_nn.Conv1d):
"""
"""
def __init__(self, input_dim, output_dim, dilation_s, kernel_s,
causal = False, stride = 1, groups=1, bias=True, tanh=True):
super(Conv1dNoPermute, self).__init__(
input_dim, output_dim, kernel_s, stride=stride,
padding = dilation_s * (kernel_s - 1) if causal \
else dilation_s * (kernel_s - 1) // 2,
dilation = dilation_s, groups=groups, bias=bias)
self.l_ac = torch_nn.Tanh() if tanh else torch_nn.Identity()
return
def forward(self, input_data):
data = input_data[0]
cond = input_data[1]
out = self.l_ac(
super(Conv1dNoPermute, self).forward(data)[:, :, :data.shape[-1]])
return [data + cond + out, cond]
class NeuralFilterBlock(torch_nn.Module):
""" Wrapper over a single filter block
NeuralFilterBlock(signal_size, hidden_size, kernel_size, conv_num=10)
args
----
signal_size: int, input signal is in shape (batch, length, signal_size)
hidden_size: int, output of conv layers is (batch, length, hidden_size)
kernel_size: int, kernel size of the conv layers
conv_num: number of conv layers in this neural filter block (default 10)
legacy_scale: Bool, whether load scale as parameter or magic number
To be compatible with old models that defines self.scale
No impact on the result, just different ways to load a
fixed self.scale
"""
def __init__(self, signal_size, hidden_size, kernel_size=3, conv_num=10,
legacy_scale = False):
super(NeuralFilterBlock, self).__init__()
self.signal_size = signal_size
self.hidden_size = hidden_size
self.kernel_size = kernel_size
self.conv_num = conv_num
self.dilation_size = [np.power(2, x) for x in np.arange(conv_num)]
# ff layer to expand dimension
self.l_ff_p = torch_nn.Sequential(
torch_nn.Linear(signal_size, hidden_size, bias=False),
torch_nn.Tanh())
# dilated conv layers
tmp = [Conv1dNoPermute(hidden_size, hidden_size, x,
kernel_size, causal=True, bias=False) \
for x in self.dilation_size]
self.l_convs = torch_nn.Sequential(*tmp)
# ff layer to de-expand dimension
self.l_ff_f = torch_nn.Sequential(
torch_nn.Linear(hidden_size, hidden_size//4, bias=False),
torch_nn.Tanh(),
torch_nn.Linear(hidden_size//4, signal_size, bias=False),
torch_nn.Tanh())
# a simple scale: to be consistent with CURRENNT implementation
if legacy_scale:
# in case this scale is defined as model parameter in
# some old models
self.scale = torch_nn.Parameter(
torch.tensor([0.1]), requires_grad=False)
else:
# simple hyper-parameter should be OK
self.scale = 0.1
return
def forward(self, input_data):
"""
input
-----
signal (batchsize, length, signal_size)
context (batchsize, length, hidden_size)
context is produced from the condition module
output
------
output: (batchsize, length, signal_size)
"""
signal, context = input_data[0], input_data[1]
# expand dimension
tmp_hidden = self.l_ff_p(signal)
# loop over dilated convs
# output of a d-conv is input + context + d-conv(input)
tmp_hidden = self.l_convs(
[tmp_hidden.permute(0, 2, 1),
context.permute(0, 2, 1)])[0].permute(0, 2, 1)
# to be consistent with legacy configuration in CURRENNT
tmp_hidden = tmp_hidden * self.scale
# compress the dimesion and skip-add
output_signal = self.l_ff_f(tmp_hidden) + signal
return [output_signal, context]
## For condition module only provide Spectral feature to Filter block
class CondModule(torch_nn.Module):
""" Conditiona module
Upsample and transform input features
CondModule(input_dimension, output_dimension, up_sample_rate,
blstm_dimension = 64, cnn_kernel_size = 3)
Spec, F0 = CondModule(features, F0)
Both input features should be frame-level features
If x doesn't contain F0, just ignore the returned F0
"""
def __init__(self, input_dim, output_dim, up_sample, \
blstm_s = 64, cnn_kernel_s = 3):
super(CondModule, self).__init__()
self.input_dim = input_dim
self.output_dim = output_dim
self.up_sample = up_sample
self.blstm_s = blstm_s
self.cnn_kernel_s = cnn_kernel_s
# bi-LSTM
self.l_blstm = nii_nn.BLSTMLayer(input_dim, self.blstm_s)
self.l_conv1d = nii_nn.Conv1dKeepLength(
self.blstm_s, output_dim, 1, self.cnn_kernel_s)
self.l_upsamp = nii_nn.UpSampleLayer(
self.output_dim, self.up_sample, True)
# Upsampling for F0: don't smooth up-sampled F0
self.l_upsamp_F0 = nii_nn.UpSampleLayer(1, self.up_sample, False)
def forward(self, feature, f0):
""" spec, f0 = forward(self, feature, f0)
feature: (batchsize, length, dim)
f0: (batchsize, length, dim=1), which should be F0 at frame-level
spec: (batchsize, length, self.output_dim), at wave-level
f0: (batchsize, length, 1), at wave-level
"""
spec = self.l_upsamp(self.l_conv1d(self.l_blstm(feature)))
f0 = self.l_upsamp_F0(f0)
return spec, f0
# For source module
class SourceModuleMusicNSF(torch_nn.Module):
""" SourceModule for hn-nsf
SourceModule(sampling_rate, harmonic_num=0, sine_amp=0.1,
add_noise_std=0.003, voiced_threshod=0)
sampling_rate: sampling_rate in Hz
harmonic_num: number of harmonic above F0 (default: 0)
sine_amp: amplitude of sine source signal (default: 0.1)
add_noise_std: std of additive Gaussian noise (default: 0.003)
note that amplitude of noise in unvoiced is decided
by sine_amp
voiced_threshold: threhold to set U/V given F0 (default: 0)
Sine_source, noise_source = SourceModuleMusicNSF(F0_sampled)
F0_sampled (batchsize, length, 1)
Sine_source (batchsize, length, 1)
noise_source (batchsize, length 1)
uv (batchsize, length, 1)
"""
def __init__(self, sampling_rate, harmonic_num=0, sine_amp=0.1,
add_noise_std=0.003, voiced_threshod=0):
super(SourceModuleMusicNSF, self).__init__()
self.sine_amp = sine_amp
self.noise_std = add_noise_std
# to produce sine waveforms
self.l_sin_gen = nii_nsf.SineGen(
sampling_rate, harmonic_num, sine_amp,
add_noise_std, voiced_threshod)
# to merge source harmonics into a single excitation
self.l_linear = torch_nn.Linear(harmonic_num+1, 1)
self.l_tanh = torch_nn.Tanh()
def forward(self, x):
"""
Sine_source, noise_source = SourceModuleMusicNSF(F0_sampled)
F0_sampled (batchsize, length, 1)
Sine_source (batchsize, length, 1)
noise_source (batchsize, length 1)
"""
# source for harmonic branch
# sine fundamental component and harmonic overtones
sine_wavs, uv, _ = self.l_sin_gen(x)
# merge into a single excitation
sine_merge = self.l_tanh(self.l_linear(sine_wavs))
# source for noise branch, in the same shape as uv
noise = torch.randn_like(uv) * self.sine_amp / 3
return sine_merge, noise, uv
# For Filter module
class FilterModuleMusicNSF(torch_nn.Module):
""" Filter for Hn-NSF
FilterModuleMusicNSF(signal_size, hidden_size, fir_coef,
block_num = 5,
kernel_size = 3, conv_num_in_block = 10)
signal_size: signal dimension (should be 1)
hidden_size: dimension of hidden features inside neural filter block
fir_coef: list of FIR filter coeffs,
(low_pass_1, low_pass_2, high_pass_1, high_pass_2)
block_num: number of neural filter blocks in harmonic branch
kernel_size: kernel size in dilated CNN
conv_num_in_block: number of d-conv1d in one neural filter block
output = FilterModuleMusicNSF(harmonic_source,noise_source,uv,context)
harmonic_source (batchsize, length, dim=1)
noise_source (batchsize, length, dim=1)
context (batchsize, length, dim)
uv (batchsize, length, dim)
output: (batchsize, length, dim=1)
"""
def __init__(self, signal_size, hidden_size, \
block_num = 5, kernel_size = 3, conv_num_in_block = 10):
super(FilterModuleMusicNSF, self).__init__()
self.signal_size = signal_size
self.hidden_size = hidden_size
self.kernel_size = kernel_size
self.block_num = block_num
self.conv_num_in_block = conv_num_in_block
# filter blocks for harmonic branch
tmp = [NeuralFilterBlock(
signal_size, hidden_size, kernel_size, conv_num_in_block) \
for x in range(self.block_num)]
self.l_har_blocks = torch_nn.Sequential(*tmp)
def forward(self, har_component, noi_component, condition_feat, uv):
"""
"""
# harmonic component
#for l_har_block in self.l_har_blocks:
# har_component = l_har_block(har_component, condition_feat)
#output = har_component
output = self.l_har_blocks([har_component, condition_feat])[0]
return output
## FOR MODEL
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
######
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\
args, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
######
# configurations
self.sine_amp = 0.1
self.noise_std = 0.001
self.input_dim = in_dim
self.output_dim = out_dim
self.hidden_dim = 64
self.upsamp_rate = prj_conf.input_reso[0]
self.sampling_rate = prj_conf.wav_samp_rate
self.cnn_kernel_size = 3
self.filter_block_num = 5
self.cnn_num_in_block = 10
self.harmonic_num = 16
# the three modules
self.m_condition = CondModule(self.input_dim, \
self.hidden_dim, \
self.upsamp_rate, \
cnn_kernel_s = self.cnn_kernel_size)
#self.m_source = SourceModuleMusicNSF(self.sampling_rate,
# self.harmonic_num,
# self.sine_amp,
# self.noise_std)
self.m_filter = FilterModuleMusicNSF(self.output_dim,
self.hidden_dim,\
self.filter_block_num, \
self.cnn_kernel_size, \
self.cnn_num_in_block)
# done
return
def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None):
"""
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
"""
return y * self.output_std + self.output_mean
def forward(self, x):
""" definition of forward method
Assume x (batchsize=1, length, dim)
Return output(batchsize=1, length)
"""
# normalize the data
feat = self.normalize_input(x)
# condition module
# place_holder is originally the up-sampled F0
# it is not used for noise-excitation model
# but it has the same shape as the upsampled souce signal
# it can help to create the noise_source below
cond_feat, place_holder = self.m_condition(feat, x[:, :, -1:])
with torch.no_grad():
noise_source = torch.randn_like(place_holder) * self.noise_std / 3
# source module
#har_source, noi_source, uv = self.m_source(f0_upsamped)
# filter module (including FIR filtering)
output = self.m_filter(noise_source, None, cond_feat, None)
# output
return output.squeeze(-1)
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
# frame shift (number of points)
self.frame_hops = [80, 40, 640]
# frame length
self.frame_lens = [320, 80, 1920]
# FFT length
self.fft_n = [4096, 4096, 4096]
# window type
self.win = torch.hann_window
# floor in log-spectrum-amplitude calculating
self.amp_floor = 0.00001
# loss
self.loss1 = torch_nn.MSELoss()
self.loss2 = torch_nn.MSELoss()
self.loss3 = torch_nn.MSELoss()
self.loss = [self.loss1, self.loss2, self.loss3]
#self.loss = torch_nn.MSELoss()
def compute(self, output_orig, target_orig):
""" Loss().compute(output, target) should return
the Loss in torch.tensor format
Assume output and target as (batchsize=1, length)
"""
# convert from (batchsize=1, length, dim=1) to (1, length)
if output_orig.ndim == 3:
output = output_orig.squeeze(-1)
else:
output = output_orig
if target_orig.ndim == 3:
target = target_orig.squeeze(-1)
else:
target = target_orig
# compute loss
loss = 0
for frame_shift, frame_len, fft_p, loss_f in \
zip(self.frame_hops, self.frame_lens, self.fft_n, self.loss):
x_stft = torch.stft(output, fft_p, frame_shift, frame_len, \
window=self.win(frame_len, \
device=output_orig.device),
onesided=True,
pad_mode="constant")
y_stft = torch.stft(target, fft_p, frame_shift, frame_len, \
window=self.win(frame_len,
device=output_orig.device),
onesided=True,
pad_mode="constant")
x_sp_amp = torch.log(torch.norm(x_stft, 2, -1).pow(2) + \
self.amp_floor)
y_sp_amp = torch.log(torch.norm(y_stft, 2, -1).pow(2) + \
self.amp_floor)
loss += loss_f(x_sp_amp, y_sp_amp)
return loss
if __name__ == "__main__":
print("Definition of model")
| 17,431 | 36.010616 | 78 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/different_main/transfer_weight.py | #!/usr/bin/env python
"""
"""
from __future__ import absolute_import
import os
import sys
import copy
import torch
import importlib
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.default_data_io as nii_default_dset
import core_scripts.data_io.customize_dataset as nii_dset
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.list_tools as nii_list_tool
import core_scripts.config_parse.config_parse as nii_config_parse
import core_scripts.config_parse.arg_parse as nii_arg_parse
import core_scripts.op_manager.op_manager as nii_op_wrapper
import core_scripts.nn_manager.nn_manager_AL as nii_nn_wrapper
import core_scripts.nn_manager.nn_manager as nii_nn_wrapper_base
import core_scripts.startup_config as nii_startup
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
def self_defined_transfer(model_src, model_tar):
""" A self defined function to transfer the weights from model_src
to model_tar
"""
# load SSL front-end
model_tar.m_front_end.ssl_model.load_state_dict(
model_src.m_ssl.state_dict())
# load SSL front-end linear layer
model_tar.m_front_end.m_front_end_process.load_state_dict(
model_src.m_frontend[0].state_dict())
# load the linear output layer
model_tar.m_back_end.m_utt_level.load_state_dict(
model_src.m_output_act[0].state_dict())
return
def main():
""" main(): the default wrapper for training and inference process
Please prepare config.py and model.py
"""
# arguments initialization
args = nii_arg_parse.f_args_parsed()
#
nii_warn.f_print_w_date("Start program", level='h')
nii_warn.f_print("Load module: %s" % (args.module_config))
nii_warn.f_print("Load 1st module: %s" % (args.module_model))
nii_warn.f_print("Load 2nd module: %s" % (args.module_model_aux))
prj_conf = importlib.import_module(args.module_config)
prj_model_src = importlib.import_module(args.module_model)
prj_model_tar = importlib.import_module(args.module_model_aux)
# initialization
nii_startup.set_random_seed(args.seed, args)
use_cuda = not args.no_cuda and torch.cuda.is_available()
device = torch.device("cuda" if use_cuda else "cpu")
#
checkpoint = torch.load(args.trained_model)
model_src = prj_model_src.Model(sum(prj_conf.input_dims),
sum(prj_conf.output_dims),
args, prj_conf)
model_tar = prj_model_tar.Model(sum(prj_conf.input_dims),
sum(prj_conf.output_dims),
args, prj_conf)
model_src.load_state_dict(checkpoint)
self_defined_transfer(model_src, model_tar)
torch.save(model_tar.state_dict(), 'temp.pt')
return
if __name__ == "__main__":
main()
| 2,922 | 31.120879 | 71 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/different_main/main_merge_datasets.py | #!/usr/bin/env python
"""
main.py for project-NN-pytorch/projects
The training/inference process wrapper.
Dataset API is replaced with NII_MergeDataSetLoader.
It is more convenient to train model on corpora stored in different directories.
Requires model.py and config.py (config_merge_datasets.py)
Usage: $: python main.py [options]
"""
from __future__ import absolute_import
import os
import sys
import torch
import importlib
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.default_data_io as nii_default_dset
import core_scripts.data_io.customize_dataset as nii_dset
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.list_tools as nii_list_tool
import core_scripts.config_parse.config_parse as nii_config_parse
import core_scripts.config_parse.arg_parse as nii_arg_parse
import core_scripts.op_manager.op_manager as nii_op_wrapper
import core_scripts.nn_manager.nn_manager as nii_nn_wrapper
import core_scripts.startup_config as nii_startup
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def main():
""" main(): the default wrapper for training and inference process
Please prepare config.py and model.py
"""
# arguments initialization
args = nii_arg_parse.f_args_parsed()
#
nii_warn.f_print_w_date("Start program", level='h')
nii_warn.f_print("Load module: %s" % (args.module_config))
nii_warn.f_print("Load module: %s" % (args.module_model))
prj_conf = importlib.import_module(args.module_config)
prj_model = importlib.import_module(args.module_model)
# initialization
nii_startup.set_random_seed(args.seed, args)
use_cuda = not args.no_cuda and torch.cuda.is_available()
device = torch.device("cuda" if use_cuda else "cpu")
# prepare data io
if not args.inference and not args.epoch2pt:
params = {'batch_size': args.batch_size,
'shuffle': args.shuffle,
'num_workers': args.num_workers,
'sampler': args.sampler}
in_trans_fns = prj_conf.input_trans_fns \
if hasattr(prj_conf, 'input_trans_fns') else None
out_trans_fns = prj_conf.output_trans_fns \
if hasattr(prj_conf, 'output_trans_fns') else None
inout_trans_fns = prj_conf.input_output_trans_fn \
if hasattr(prj_conf, 'input_output_trans_fn') else None
# Load file list and create data loader
trn_lst = prj_conf.trn_list
trn_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.trn_set_name, \
trn_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./',
params = params,
truncate_seq = prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = True,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns,
inoutput_augment_func = inout_trans_fns)
if hasattr(prj_conf, 'val_input_dirs'):
val_input_dirs = prj_conf.val_input_dirs
else:
val_input_dirs = prj_conf.input_dirs
if hasattr(prj_conf, 'val_output_dirs'):
val_output_dirs = prj_conf.val_output_dirs
else:
val_output_dirs = prj_conf.output_dirs
if prj_conf.val_list is not None:
val_lst = prj_conf.val_list
val_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.val_set_name,
val_lst,
val_input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
val_output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./', \
params = params,
truncate_seq= prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns,
inoutput_augment_func = inout_trans_fns)
else:
val_set = None
# initialize the model and loss function
model = prj_model.Model(trn_set.get_in_dim(), \
trn_set.get_out_dim(), \
args, prj_conf, trn_set.get_data_mean_std())
loss_wrapper = prj_model.Loss(args)
# initialize the optimizer
optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args)
# if necessary, resume training
if args.trained_model == "":
checkpoint = None
else:
checkpoint = torch.load(args.trained_model)
# start training
nii_nn_wrapper.f_train_wrapper(args, model,
loss_wrapper, device,
optimizer_wrapper,
trn_set, val_set, checkpoint)
# done for traing
elif args.inference:
# for inference
# default, no truncating, no shuffling
params = {'batch_size': args.batch_size,
'shuffle': False,
'num_workers': args.num_workers,
'sampler': args.sampler}
in_trans_fns = prj_conf.test_input_trans_fns \
if hasattr(prj_conf, 'test_input_trans_fns') else None
out_trans_fns = prj_conf.test_output_trans_fns \
if hasattr(prj_conf, 'test_output_trans_fns') else None
inout_trans_fns = prj_conf.test_input_output_trans_fn \
if hasattr(prj_conf, 'test_input_output_trans_fn') \
else None
if type(prj_conf.test_list) is list:
t_lst = prj_conf.test_list
else:
t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list)
test_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.test_set_name, \
t_lst, \
prj_conf.test_input_dirs,
prj_conf.input_exts,
prj_conf.input_dims,
prj_conf.input_reso,
prj_conf.input_norm,
prj_conf.test_output_dirs,
prj_conf.output_exts,
prj_conf.output_dims,
prj_conf.output_reso,
prj_conf.output_norm,
'./',
params = params,
truncate_seq= None,
min_seq_len = None,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns,
inoutput_augment_func = inout_trans_fns)
# initialize model
model = prj_model.Model(test_set.get_in_dim(), \
test_set.get_out_dim(), \
args, prj_conf)
if args.trained_model == "":
print("No model is loaded by ---trained-model for inference")
print("By default, load %s%s" % (args.save_trained_name,
args.save_model_ext))
checkpoint = torch.load("%s%s" % (args.save_trained_name,
args.save_model_ext))
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper.f_inference_wrapper(args, model, device, \
test_set, checkpoint)
elif args.epoch2pt:
# for model conversion from epoch.pt to trained_network.pt
# initialize model
model = prj_model.Model(
sum(prj_conf.input_dims), sum(prj_conf.output_dims), args, prj_conf)
if args.trained_model == "":
print("Please provide ---trained-model")
sys.exit(1)
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper.f_convert_epoch_to_trained(
args, model, device, checkpoint)
else:
print("Fatal error in main.py")
sys.exit(1)
# done
return
if __name__ == "__main__":
main()
| 9,342 | 35.928854 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/different_main/main_al.py | #!/usr/bin/env python
"""
main.py for project-NN-pytorch/projects
The training/inference process wrapper for active learning.
The base is on main_mergedataset.py
Requires model.py and config.py (config_merge_datasets.py)
Usage: $: python main.py [options]
"""
from __future__ import absolute_import
import os
import sys
import copy
import torch
import importlib
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.default_data_io as nii_default_dset
import core_scripts.data_io.customize_dataset as nii_dset
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.list_tools as nii_list_tool
import core_scripts.config_parse.config_parse as nii_config_parse
import core_scripts.config_parse.arg_parse as nii_arg_parse
import core_scripts.op_manager.op_manager as nii_op_wrapper
import core_scripts.nn_manager.nn_manager_AL as nii_nn_wrapper
import core_scripts.nn_manager.nn_manager as nii_nn_wrapper_base
import core_scripts.startup_config as nii_startup
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
def main():
""" main(): the default wrapper for training and inference process
Please prepare config.py and model.py
"""
# arguments initialization
args = nii_arg_parse.f_args_parsed()
#
nii_warn.f_print_w_date("Start program", level='h')
nii_warn.f_print("Load module: %s" % (args.module_config))
nii_warn.f_print("Load module: %s" % (args.module_model))
prj_conf = importlib.import_module(args.module_config)
prj_model = importlib.import_module(args.module_model)
# initialization
nii_startup.set_random_seed(args.seed, args)
use_cuda = not args.no_cuda and torch.cuda.is_available()
device = torch.device("cuda" if use_cuda else "cpu")
# prepare data io
if not args.inference:
params = {'batch_size': args.batch_size,
'shuffle': args.shuffle,
'num_workers': args.num_workers,
'sampler': args.sampler}
in_trans_fns = prj_conf.input_trans_fns \
if hasattr(prj_conf, 'input_trans_fns') else None
out_trans_fns = prj_conf.output_trans_fns \
if hasattr(prj_conf, 'output_trans_fns') else None
# Load file list and create data loader
trn_lst = prj_conf.trn_list
trn_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.trn_set_name, \
trn_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./',
params = params,
truncate_seq = prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = True,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
# Load data pool and create data loader
pool_lst = prj_conf.al_pool_list
pool_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.al_pool_set_name, \
pool_lst,
prj_conf.al_pool_in_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.al_pool_out_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./',
params = params,
truncate_seq = prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = True,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
if hasattr(prj_conf, 'val_input_dirs'):
val_input_dirs = prj_conf.val_input_dirs
else:
val_input_dirs = prj_conf.input_dirs
if hasattr(prj_conf, 'val_output_dirs'):
val_output_dirs = prj_conf.val_output_dirs
else:
val_output_dirs = prj_conf.output_dirs
if prj_conf.val_list is not None:
val_lst = prj_conf.val_list
val_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.val_set_name,
val_lst,
val_input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
val_output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./', \
params = params,
truncate_seq= prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
else:
val_set = None
# initialize the model and loss function
model = prj_model.Model(trn_set.get_in_dim(), \
trn_set.get_out_dim(), \
args, prj_conf, trn_set.get_data_mean_std())
loss_wrapper = prj_model.Loss(args)
# initialize the optimizer
optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args)
# if necessary, resume training
if args.trained_model == "":
checkpoint = None
else:
checkpoint = torch.load(args.trained_model)
# pre-training using standard procedure
# change args
args_tmp = copy.deepcopy(args)
args_tmp.epochs = args.active_learning_pre_train_epoch_num
args_tmp.not_save_each_epoch = True
args_tmp.save_trained_name += '_pretrained'
args_tmp.active_learning_cycle_num = 0
pretraind_name = args_tmp.save_trained_name + args_tmp.save_model_ext
if args.active_learning_pre_train_epoch_num:
nii_warn.f_print_w_date("Normal training (warm-up) phase",level='h')
nii_warn.f_print("Normal training for {:d} epochs".format(
args.active_learning_pre_train_epoch_num))
op_wrapper_tmp = nii_op_wrapper.OptimizerWrapper(model, args_tmp)
loss_wrapper_tmp = prj_model.Loss(args_tmp)
nii_nn_wrapper_base.f_train_wrapper(
args_tmp, model, loss_wrapper, device, op_wrapper_tmp,
trn_set, val_set, checkpoint)
checkpoint = torch.load(pretraind_name)
elif checkpoint is None:
if os.path.isfile(pretraind_name):
checkpoint = torch.load(pretraind_name)
nii_warn.f_print("Use pretrained model before active learning")
else:
nii_warn.f_print("Use seed model to initialize")
nii_warn.f_print_w_date("Active learning phase",level='h')
# start training
nii_nn_wrapper.f_train_wrapper(
args, model,
loss_wrapper, device,
optimizer_wrapper,
trn_set, pool_set, val_set, checkpoint)
# done for traing
else:
# for inference
# default, no truncating, no shuffling
params = {'batch_size': args.batch_size,
'shuffle': False,
'num_workers': args.num_workers,
'sampler': args.sampler}
in_trans_fns = prj_conf.test_input_trans_fns \
if hasattr(prj_conf, 'test_input_trans_fns') else None
out_trans_fns = prj_conf.test_output_trans_fns \
if hasattr(prj_conf, 'test_output_trans_fns') else None
if type(prj_conf.test_list) is list:
t_lst = prj_conf.test_list
else:
t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list)
test_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.test_set_name, \
t_lst, \
prj_conf.test_input_dirs,
prj_conf.input_exts,
prj_conf.input_dims,
prj_conf.input_reso,
prj_conf.input_norm,
prj_conf.test_output_dirs,
prj_conf.output_exts,
prj_conf.output_dims,
prj_conf.output_reso,
prj_conf.output_norm,
'./',
params = params,
truncate_seq= None,
min_seq_len = None,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
# initialize model
model = prj_model.Model(test_set.get_in_dim(), \
test_set.get_out_dim(), \
args, prj_conf)
if args.trained_model == "":
print("No model is loaded by ---trained-model for inference")
print("By default, load %s%s" % (args.save_trained_name,
args.save_model_ext))
checkpoint = torch.load("%s%s" % (args.save_trained_name,
args.save_model_ext))
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper_base.f_inference_wrapper(
args, model, device, test_set, checkpoint)
# done
return
if __name__ == "__main__":
main()
| 10,498 | 36.766187 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/different_main/main_profile.py | #!/usr/bin/env python
"""
main.py for project-NN-pytorch/projects
The training/inference process wrapper.
Dataset API is replaced with NII_MergeDataSetLoader.
It is more convenient to train model on corpora stored in different directories.
Requires model.py and config.py (config_merge_datasets.py)
Usage: $: python main.py [options]
"""
from __future__ import absolute_import
import os
import sys
import torch
import importlib
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.default_data_io as nii_default_dset
import core_scripts.data_io.customize_dataset as nii_dset
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.list_tools as nii_list_tool
import core_scripts.config_parse.config_parse as nii_config_parse
import core_scripts.config_parse.arg_parse as nii_arg_parse
import core_scripts.op_manager.op_manager as nii_op_wrapper
import core_scripts.nn_manager.nn_manager_profile as nii_nn_wrapper
import core_scripts.startup_config as nii_startup
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def main():
""" main(): the default wrapper for training and inference process
Please prepare config.py and model.py
"""
# arguments initialization
args = nii_arg_parse.f_args_parsed()
#
nii_warn.f_print_w_date("Start program", level='h')
nii_warn.f_print("Load module: %s" % (args.module_config))
nii_warn.f_print("Load module: %s" % (args.module_model))
prj_conf = importlib.import_module(args.module_config)
prj_model = importlib.import_module(args.module_model)
# initialization
nii_startup.set_random_seed(args.seed, args)
use_cuda = not args.no_cuda and torch.cuda.is_available()
device = torch.device("cuda" if use_cuda else "cpu")
# prepare data io
if not args.inference:
params = {'batch_size': args.batch_size,
'shuffle': args.shuffle,
'num_workers': args.num_workers,
'sampler': args.sampler}
in_trans_fns = prj_conf.input_trans_fns \
if hasattr(prj_conf, 'input_trans_fns') else None
out_trans_fns = prj_conf.output_trans_fns \
if hasattr(prj_conf, 'output_trans_fns') else None
# Load file list and create data loader
trn_lst = prj_conf.trn_list
trn_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.trn_set_name, \
trn_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./',
params = params,
truncate_seq = prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = True,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
if prj_conf.val_list is not None:
val_lst = prj_conf.val_list
val_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.val_set_name,
val_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./', \
params = params,
truncate_seq= prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
else:
val_set = None
# initialize the model and loss function
model = prj_model.Model(trn_set.get_in_dim(), \
trn_set.get_out_dim(), \
args, prj_conf, trn_set.get_data_mean_std())
loss_wrapper = prj_model.Loss(args)
# initialize the optimizer
optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args)
# if necessary, resume training
if args.trained_model == "":
checkpoint = None
else:
checkpoint = torch.load(args.trained_model)
# start training
nii_nn_wrapper.f_train_wrapper(args, model,
loss_wrapper, device,
optimizer_wrapper,
trn_set, val_set, checkpoint)
# done for traing
else:
# for inference
# default, no truncating, no shuffling
params = {'batch_size': args.batch_size,
'shuffle': False,
'num_workers': args.num_workers}
in_trans_fns = prj_conf.test_input_trans_fns \
if hasattr(prj_conf, 'test_input_trans_fns') else None
out_trans_fns = prj_conf.test_output_trans_fns \
if hasattr(prj_conf, 'test_output_trans_fns') else None
if type(prj_conf.test_list) is list:
t_lst = prj_conf.test_list
else:
t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list)
test_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.test_set_name, \
t_lst, \
prj_conf.test_input_dirs,
prj_conf.input_exts,
prj_conf.input_dims,
prj_conf.input_reso,
prj_conf.input_norm,
prj_conf.test_output_dirs,
prj_conf.output_exts,
prj_conf.output_dims,
prj_conf.output_reso,
prj_conf.output_norm,
'./',
params = params,
truncate_seq= None,
min_seq_len = None,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
# initialize model
model = prj_model.Model(test_set.get_in_dim(), \
test_set.get_out_dim(), \
args, prj_conf)
if args.trained_model == "":
print("No model is loaded by ---trained-model for inference")
print("By default, load %s%s" % (args.save_trained_name,
args.save_model_ext))
checkpoint = torch.load("%s%s" % (args.save_trained_name,
args.save_model_ext))
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper.f_inference_wrapper(args, model, device, \
test_set, checkpoint)
# done
return
if __name__ == "__main__":
main()
| 7,827 | 35.924528 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/sandbox/data/asvspoof2019/convert_to_bin.py | #!/usr/bin/python
"""
Convert txt score files to binary data matrix, in format [score, class]
class: 0: spoof, 1: non-target, 2: target
"""
import core_scripts.data_io.io_tools as nii_io
import numpy as np
import os
import sys
def read_txt_file(file_path):
data = np.genfromtxt(
file_path, dtype=[('class', 'U10'),('type', 'U10'),
('score','f4')], delimiter=" ")
data_new = np.zeros([data.shape[0], 2])
for idx, data_entry in enumerate(data):
data_new[idx, 0] = data_entry[-1]
if data_entry[1] == 'target':
data_new[idx, 1] = 2
elif data_entry[1] == 'nontarget':
data_new[idx, 1] = 1
else:
data_new[idx, 1] = 0
return data_new
def convert_format(file_path):
data = read_txt_file(file_path)
file_new_path = os.path.splitext(file_path)[0] + '.bin'
nii_io.f_write_raw_mat(data, file_new_path)
if __name__ == "__main__":
file_list = ['ASVspoof2019.LA.asv.dev.gi.trl.scores.txt',
'ASVspoof2019.PA.asv.dev.gi.trl.scores.txt',
'ASVspoof2019.LA.asv.eval.gi.trl.scores.txt',
'ASVspoof2019.PA.asv.eval.gi.trl.scores.txt']
for filename in file_list:
convert_format(filename)
| 1,290 | 28.340909 | 71 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/04-asvspoof2021-toy/lfcc-lcnn-lstm-sig_toy_example/main.py | #!/usr/bin/env python
"""
main.py
The default training/inference process wrapper
Requires model.py and config.py
Usage: $: python main.py [options]
"""
from __future__ import absolute_import
import os
import sys
import torch
import importlib
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.default_data_io as nii_dset
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.list_tools as nii_list_tool
import core_scripts.config_parse.config_parse as nii_config_parse
import core_scripts.config_parse.arg_parse as nii_arg_parse
import core_scripts.op_manager.op_manager as nii_op_wrapper
import core_scripts.nn_manager.nn_manager as nii_nn_wrapper
import core_scripts.startup_config as nii_startup
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def main():
""" main(): the default wrapper for training and inference process
Please prepare config.py and model.py
"""
# arguments initialization
args = nii_arg_parse.f_args_parsed()
#
nii_warn.f_print_w_date("Start program", level='h')
nii_warn.f_print("Load module: %s" % (args.module_config))
nii_warn.f_print("Load module: %s" % (args.module_model))
prj_conf = importlib.import_module(args.module_config)
prj_model = importlib.import_module(args.module_model)
# initialization
nii_startup.set_random_seed(args.seed, args)
use_cuda = not args.no_cuda and torch.cuda.is_available()
device = torch.device("cuda" if use_cuda else "cpu")
# prepare data io
if not args.inference:
params = {'batch_size': args.batch_size,
'shuffle': args.shuffle,
'num_workers': args.num_workers,
'sampler': args.sampler}
# Load file list and create data loader
trn_lst = nii_list_tool.read_list_from_text(prj_conf.trn_list)
trn_set = nii_dset.NIIDataSetLoader(
prj_conf.trn_set_name, \
trn_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./',
params = params,
truncate_seq = prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = True,
wav_samp_rate = prj_conf.wav_samp_rate,
global_arg = args)
if prj_conf.val_list is not None:
val_lst = nii_list_tool.read_list_from_text(prj_conf.val_list)
val_set = nii_dset.NIIDataSetLoader(
prj_conf.val_set_name,
val_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./', \
params = params,
truncate_seq= prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
global_arg = args)
else:
val_set = None
# initialize the model and loss function
model = prj_model.Model(trn_set.get_in_dim(), \
trn_set.get_out_dim(), \
args, prj_conf, trn_set.get_data_mean_std())
loss_wrapper = prj_model.Loss(args)
# initialize the optimizer
optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args)
# if necessary, resume training
if args.trained_model == "":
checkpoint = None
else:
checkpoint = torch.load(args.trained_model)
# start training
nii_nn_wrapper.f_train_wrapper(args, model,
loss_wrapper, device,
optimizer_wrapper,
trn_set, val_set, checkpoint)
# done for traing
else:
# for inference
# default, no truncating, no shuffling
params = {'batch_size': args.batch_size,
'shuffle': False,
'num_workers': args.num_workers}
if type(prj_conf.test_list) is list:
t_lst = prj_conf.test_list
else:
t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list)
test_set = nii_dset.NIIDataSetLoader(
prj_conf.test_set_name, \
t_lst, \
prj_conf.test_input_dirs,
prj_conf.input_exts,
prj_conf.input_dims,
prj_conf.input_reso,
prj_conf.input_norm,
prj_conf.test_output_dirs,
prj_conf.output_exts,
prj_conf.output_dims,
prj_conf.output_reso,
prj_conf.output_norm,
'./',
params = params,
truncate_seq= None,
min_seq_len = None,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
global_arg = args)
# initialize model
model = prj_model.Model(test_set.get_in_dim(), \
test_set.get_out_dim(), \
args, prj_conf)
if args.trained_model == "":
print("No model is loaded by ---trained-model for inference")
print("By default, load %s%s" % (args.save_trained_name,
args.save_model_ext))
checkpoint = torch.load("%s%s" % (args.save_trained_name,
args.save_model_ext))
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper.f_inference_wrapper(args, model, device, \
test_set, checkpoint)
# done
return
if __name__ == "__main__":
main()
| 6,501 | 34.530055 | 76 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/04-asvspoof2021-toy/lfcc-lcnn-lstm-sig_toy_example/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
##############
## util
##############
def protocol_parse(protocol_filepath):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
input:
-----
protocol_filepath: string, path to the protocol file
for convenience, I put train/dev/eval trials into a single protocol file
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = {}
try:
temp_buffer = np.loadtxt(protocol_filepath, dtype='str')
for row in temp_buffer:
if row[-1] == 'bonafide':
data_buffer[row[1]] = 1
else:
data_buffer[row[1]] = 0
except OSError:
print("Skip loading protocol file")
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(in_dim,out_dim,\
args, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_file = prj_conf.optional_argument[0]
self.protocol_parser = protocol_parse(protocol_file)
# Working sampling rate
# torchaudio may be used to change sampling rate
self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
# frame shift (number of waveform points)
self.frame_hops = [160]
# frame length
self.frame_lens = [320]
# FFT length
self.fft_n = [512]
# LFCC dim (base component)
self.lfcc_dim = [20]
self.lfcc_with_delta = True
# window type
self.win = torch.hann_window
# floor in log-spectrum-amplitude calculating (not used)
self.amp_floor = 0.00001
# number of frames to be kept for each trial
# no truncation
self.v_truncate_lens = [None for x in self.frame_hops]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.frame_lens)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = 1
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip(
self.v_truncate_lens, self.fft_n, self.lfcc_dim)):
fft_n_bins = fft_n // 2 + 1
if self.lfcc_with_delta:
lfcc_dim = lfcc_dim * 3
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(64, affine=False),
torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Dropout(0.7)
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32),
nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32)
)
)
self.m_output_act.append(
torch_nn.Linear((lfcc_dim // 16) * 32, self.v_emd_dim)
)
self.m_frontend.append(
nii_front_end.LFCC(self.frame_lens[idx],
self.frame_hops[idx],
self.fft_n[idx],
self.m_target_sr,
self.lfcc_dim[idx],
with_energy=True)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
# done
return
def prepare_mean_std(self, in_dim, out_dim, args, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, trunc_len, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
trunc_len: number of frames to be kept after truncation
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_sp_amp = self.m_frontend[idx](wav.squeeze(-1))
# return
return x_sp_amp
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros([batch_size * self.v_submodels,
self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.frame_hops, self.frame_lens, self.fft_n,
self.v_truncate_lens, self.m_transform,
self.m_before_pooling, self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, trunc_len, datalength)
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
frame_num = hidden_features.shape[1]
hidden_features = hidden_features.view(batch_size, frame_num, -1)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = m_output((hidden_features_lstm + hidden_features).mean(1))
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_score(self, feature_vec, inference=False):
"""
"""
# feature_vec is [batch * submodel, 1]
if inference:
return feature_vec.squeeze(1)
else:
return torch.sigmoid(feature_vec).squeeze(1)
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _get_target_eval(self, filenames):
""" retrieve the target label for a trial from protocol if available
"""
return [self.protocol_parser[x] if x in self.protocol_parser else -1 \
for x in filenames]
def forward(self, x, fileinfo):
#with torch.no_grad():
# vad_waveform = self.m_vad(x.squeeze(-1))
# vad_waveform = self.m_vad(torch.flip(vad_waveform, dims=[1]))
# if vad_waveform.shape[-1] > 0:
# x = torch.flip(vad_waveform, dims=[1]).unsqueeze(-1)
# else:
# pass
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
if self.training:
feature_vec = self._compute_embedding(x, datalength)
scores = self._compute_score(feature_vec)
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target,
device=x.device, dtype=scores.dtype)
target_vec = target_vec.repeat(self.v_submodels)
return [scores, target_vec, True]
else:
feature_vec = self._compute_embedding(x, datalength)
scores = self._compute_score(feature_vec, True)
target = self._get_target_eval(filenames)
print("Output, %s, %d, %f" % (filenames[0],
target[0], scores.mean()))
# don't write output score as a single file
return None
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
self.m_loss = torch_nn.BCELoss()
def compute(self, outputs, target):
"""
"""
loss = self.m_loss(outputs[0], outputs[1])
return loss
if __name__ == "__main__":
print("Definition of model")
| 14,927 | 33.878505 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/04-asvspoof2021-toy/lfcc-lcnn-lstm-sig_toy_example/config.py | #!/usr/bin/env python
"""
config.py
Configuration file for training and evaluation.
Usage:
For training, change Configuration for training stage
For inference, change Configuration for inference stage
Please follow the instruction below to config this file
"""
import os
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#########################################################
## Configuration for training stage
#########################################################
# Name of datasets (any string you wish to use)
# after data preparation, trn/val_set_name are used to save statistics
# about the data sets
trn_set_name = 'asvspoof2021_trn_toy'
val_set_name = 'asvspoof2021_val_toy'
# for convenience
# we will use resources in this directory
tmp = os.path.dirname(__file__) + '/../DATA/toy_example'
# File lists (text file, one data name per line, without name extension)
# trin_file_list: list of files for training set
trn_list = tmp + '/scp/train.lst'
# val_file_list: list of files for validation set. It can be None
val_list = tmp + '/scp/val.lst'
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# we assume train and validation data are put in the same sub-directory
# here, we only use waveform as input to the code
input_dirs = [tmp + '/train_dev']
# Dimensions of input features
# input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...]
# here, we only use waveform as input to the code, and the dimension
# of waveform data is 1
input_dims = [1]
# File name extension for input features
# input_exts = [name_extention_of_feature_1, ...]
# here, we only use waveform, thus it is ".wav"
input_exts = ['.wav']
# Temporal resolution for input features
# input_reso = [reso_feature_1, reso_feature_2, ...]
# this is used for other projects.
# for waveform, we set it to 1
input_reso = [1]
# Whether input features should be z-normalized
# input_norm = [normalize_feature_1, normalize_feature_2]
# we don't z-normalize the waveform
input_norm = [False]
# This is for other projects,
# we don't load target features for ASVspoof models using these config
# we read target labels of ASVspoof trials in mode.py
# but we need to fill in some placehoders
output_dirs = []
output_dims = [1]
output_exts = ['.bin']
output_reso = [1]
output_norm = [False]
# Waveform sampling rate
# wav_samp_rate can be None if no waveform data is used
# ASVspoof uses 16000 Hz
wav_samp_rate = 16000
# Truncating input sequences so that the maximum length = truncate_seq
# When truncate_seq is larger, more GPU mem required
# If you don't want truncating, please truncate_seq = None
# For ASVspoof, we don't do truncate here
truncate_seq = None
# Minimum sequence length
# If sequence length < minimum_len, this sequence is not used for training
# minimum_len can be None
# For ASVspoof, we don't set minimum length of input trial
minimum_len = None
# Optional argument
# We will use this optional_argument to read protocol file
# When evaluating on a eval set without protocol file, set this to ['']
optional_argument = [tmp + '/protocol.txt']
#########################################################
## Configuration for inference stage
#########################################################
# similar options to training stage
test_set_name = 'asvspoof2021_test_toy'
# List of test set data
# for convenience, you may directly load test_set list here
test_list = tmp + '/scp/test.lst'
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# directory of the evaluation set waveform
test_input_dirs = [tmp + '/eval']
# Directories for output features, which are []
test_output_dirs = []
| 3,795 | 30.89916 | 75 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/04-asvspoof2021-toy/lfcc-lcnn-lstm-sig_toy_example/config_auto.py | #!/usr/bin/env python
"""
config.py
This configuration file will read environment variables
for configuration. it is used by 02_eval_alternative.sh
"""
import os
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2021, Xin Wang"
#########################################################
## Configuration for training stage
#########################################################
# Name of datasets (any string you wish to use)
# after data preparation, trn/val_set_name are used to save statistics
# about the data sets
trn_set_name = ''
val_set_name = ''
# File lists (text file, one data name per line, without name extension)
# trin_file_list: list of files for training set
trn_list = ''
# val_file_list: list of files for validation set. It can be None
val_list = ''
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# we assume train and validation data are put in the same sub-directory
# here, we only use waveform as input to the code
input_dirs = ['']
# Dimensions of input features
# input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...]
# here, we only use waveform as input to the code, and the dimension
# of waveform data is 1
input_dims = [1]
# File name extension for input features
# input_exts = [name_extention_of_feature_1, ...]
# here, we only use waveform, thus it is ".wav"
input_exts = ['.wav']
# Temporal resolution for input features
# input_reso = [reso_feature_1, reso_feature_2, ...]
# this is used for other projects.
# for waveform, we set it to 1
input_reso = [1]
# Whether input features should be z-normalized
# input_norm = [normalize_feature_1, normalize_feature_2]
# we don't z-normalize the waveform
input_norm = [False]
# This is for other projects,
# we don't load target features for ASVspoof models using these config
# we read target labels of ASVspoof trials in mode.py
# but we need to fill in some placehoders
output_dirs = []
output_dims = [1]
output_exts = ['.bin']
output_reso = [1]
output_norm = [False]
# Waveform sampling rate
# wav_samp_rate can be None if no waveform data is used
# ASVspoof uses 16000 Hz
wav_samp_rate = 16000
# Truncating input sequences so that the maximum length = truncate_seq
# When truncate_seq is larger, more GPU mem required
# If you don't want truncating, please truncate_seq = None
# For ASVspoof, we don't do truncate here
truncate_seq = None
# Minimum sequence length
# If sequence length < minimum_len, this sequence is not used for training
# minimum_len can be None
# For ASVspoof, we don't set minimum length of input trial
minimum_len = None
# Optional argument
# We will use this optional_argument to read protocol file
# When evaluating on a eval set without protocol file, set this to ['']
optional_argument = ['']
#########################################################
## Configuration for inference stage
#########################################################
# similar options to training stage
test_set_name = os.getenv('TEMP_DATA_NAME')
# List of test set data
# for convenience, you may directly load test_set list here
test_list = test_set_name + '.lst'
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# directory of the evaluation set waveform
test_input_dirs = [os.getenv('TEMP_DATA_DIR')]
# Directories for output features, which are []
test_output_dirs = []
| 3,465 | 30.225225 | 75 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/main.py | #!/usr/bin/env python
"""
main.py for project-NN-pytorch/projects
The training/inference process wrapper.
Dataset API is replaced with NII_MergeDataSetLoader.
It is more convenient to train model on corpora stored in different directories.
Requires model.py and config.py (config_merge_datasets.py)
Usage: $: python main.py [options]
"""
from __future__ import absolute_import
import os
import sys
import torch
import importlib
import core_scripts.other_tools.display as nii_warn
import core_scripts.data_io.default_data_io as nii_default_dset
import core_scripts.data_io.customize_dataset as nii_dset
import core_scripts.data_io.conf as nii_dconf
import core_scripts.other_tools.list_tools as nii_list_tool
import core_scripts.config_parse.config_parse as nii_config_parse
import core_scripts.config_parse.arg_parse as nii_arg_parse
import core_scripts.op_manager.op_manager as nii_op_wrapper
import core_scripts.nn_manager.nn_manager as nii_nn_wrapper
import core_scripts.startup_config as nii_startup
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
def main():
""" main(): the default wrapper for training and inference process
Please prepare config.py and model.py
"""
# arguments initialization
args = nii_arg_parse.f_args_parsed()
#
nii_warn.f_print_w_date("Start program", level='h')
nii_warn.f_print("Load module: %s" % (args.module_config))
nii_warn.f_print("Load module: %s" % (args.module_model))
prj_conf = importlib.import_module(args.module_config)
prj_model = importlib.import_module(args.module_model)
# initialization
nii_startup.set_random_seed(args.seed, args)
use_cuda = not args.no_cuda and torch.cuda.is_available()
device = torch.device("cuda" if use_cuda else "cpu")
# prepare data io
if not args.inference:
params = {'batch_size': args.batch_size,
'shuffle': args.shuffle,
'num_workers': args.num_workers,
'sampler': args.sampler}
in_trans_fns = prj_conf.input_trans_fns \
if hasattr(prj_conf, 'input_trans_fns') else None
out_trans_fns = prj_conf.output_trans_fns \
if hasattr(prj_conf, 'output_trans_fns') else None
# Load file list and create data loader
trn_lst = prj_conf.trn_list
trn_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.trn_set_name, \
trn_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./',
params = params,
truncate_seq = prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = True,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
if prj_conf.val_list is not None:
val_lst = prj_conf.val_list
val_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.val_set_name,
val_lst,
prj_conf.input_dirs, \
prj_conf.input_exts, \
prj_conf.input_dims, \
prj_conf.input_reso, \
prj_conf.input_norm, \
prj_conf.output_dirs, \
prj_conf.output_exts, \
prj_conf.output_dims, \
prj_conf.output_reso, \
prj_conf.output_norm, \
'./', \
params = params,
truncate_seq= prj_conf.truncate_seq,
min_seq_len = prj_conf.minimum_len,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
else:
val_set = None
# initialize the model and loss function
model = prj_model.Model(trn_set.get_in_dim(), \
trn_set.get_out_dim(), \
args, prj_conf, trn_set.get_data_mean_std())
loss_wrapper = prj_model.Loss(args)
# initialize the optimizer
optimizer_wrapper = nii_op_wrapper.OptimizerWrapper(model, args)
# if necessary, resume training
if args.trained_model == "":
checkpoint = None
else:
checkpoint = torch.load(args.trained_model)
# start training
nii_nn_wrapper.f_train_wrapper(args, model,
loss_wrapper, device,
optimizer_wrapper,
trn_set, val_set, checkpoint)
# done for traing
else:
# for inference
# default, no truncating, no shuffling
params = {'batch_size': args.batch_size,
'shuffle': False,
'num_workers': args.num_workers,
'sampler': args.sampler}
in_trans_fns = prj_conf.test_input_trans_fns \
if hasattr(prj_conf, 'test_input_trans_fns') else None
out_trans_fns = prj_conf.test_output_trans_fns \
if hasattr(prj_conf, 'test_output_trans_fns') else None
if type(prj_conf.test_list) is list:
t_lst = prj_conf.test_list
else:
t_lst = nii_list_tool.read_list_from_text(prj_conf.test_list)
test_set = nii_dset.NII_MergeDataSetLoader(
prj_conf.test_set_name, \
t_lst, \
prj_conf.test_input_dirs,
prj_conf.input_exts,
prj_conf.input_dims,
prj_conf.input_reso,
prj_conf.input_norm,
prj_conf.test_output_dirs,
prj_conf.output_exts,
prj_conf.output_dims,
prj_conf.output_reso,
prj_conf.output_norm,
'./',
params = params,
truncate_seq= None,
min_seq_len = None,
save_mean_std = False,
wav_samp_rate = prj_conf.wav_samp_rate,
way_to_merge = args.way_to_merge_datasets,
global_arg = args,
dset_config = prj_conf,
input_augment_funcs = in_trans_fns,
output_augment_funcs = out_trans_fns)
# initialize model
model = prj_model.Model(test_set.get_in_dim(), \
test_set.get_out_dim(), \
args, prj_conf)
if args.trained_model == "":
print("No model is loaded by ---trained-model for inference")
print("By default, load %s%s" % (args.save_trained_name,
args.save_model_ext))
checkpoint = torch.load("%s%s" % (args.save_trained_name,
args.save_model_ext))
else:
checkpoint = torch.load(args.trained_model)
# do inference and output data
nii_nn_wrapper.f_inference_wrapper(args, model, device, \
test_set, checkpoint)
# done
return
if __name__ == "__main__":
main()
| 7,862 | 35.915493 | 80 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/config_train_asvspoof2019.py | #!/usr/bin/env python
"""
config.py for project-NN-pytorch/projects
Usage:
For training, change Configuration for training stage
For inference, change Configuration for inference stage (although
this configuration will only be used for training)
"""
import os
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
#########################################################
## Configuration for training stage
#########################################################
# Name of datasets
# after data preparation, trn/val_set_name are used to save statistics
# about the data sets
trn_set_name = ['asvspoof2019_trn']
val_set_name = ['asvspoof2019_val']
# for convenience
tmp = os.path.dirname(__file__) + '/../../../DATA/asvspoof2019_LA'
# File lists (text file, one data name per line, without name extension)
# trin_file_list: list of files for training set
trn_list = [tmp + '/scp/train.lst']
# val_file_list: list of files for validation set. It can be None
val_list = [tmp + '/scp/val.lst']
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# we assume train and validation data are put in the same sub-directory
input_dirs = [[tmp + '/train_dev']]
# Dimensions of input features
# input_dims = [dimension_of_feature_1, dimension_of_feature_2, ...]
input_dims = [1]
# File name extension for input features
# input_exts = [name_extention_of_feature_1, ...]
# Please put ".f0" as the last feature
input_exts = ['.wav']
# Temporal resolution for input features
# input_reso = [reso_feature_1, reso_feature_2, ...]
# for waveform modeling, temporal resolution of input acoustic features
# may be = waveform_sampling_rate * frame_shift_of_acoustic_features
# for example, 80 = 16000 Hz * 5 ms
input_reso = [1]
# Whether input features should be z-normalized
# input_norm = [normalize_feature_1, normalize_feature_2]
input_norm = [False]
# Similar configurations for output features
output_dirs = [[] for x in input_dirs]
output_dims = [1]
output_exts = ['.bin']
output_reso = [1]
output_norm = [False]
# Waveform sampling rate
# wav_samp_rate can be None if no waveform data is used
wav_samp_rate = 16000
# Truncating input sequences so that the maximum length = truncate_seq
# When truncate_seq is larger, more GPU mem required
# If you don't want truncating, please truncate_seq = None
truncate_seq = 64000
# Minimum sequence length
# If sequence length < minimum_len, this sequence is not used for training
# minimum_len can be None
minimum_len = 8000
# Optional argument
# Just a buffer for convenience
# It can contain anything
# Here we use it to hold the path to the protocol.txt
# They will be loaded by model.py
optional_argument = [tmp + '/protocol.txt']
#import augment
#input_trans_fns = [[augment.wav_aug]]
#output_trans_fns = [[]]
#########################################################
## Configuration for inference stage (place holder)
#########################################################
# Please use config_test_*.py inference
# This part is just a place holder
test_set_name = trn_set_name + val_set_name
# List of test set data
# for convenience, you may directly load test_set list here
test_list = trn_list + val_list
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# we assume train and validation data are put in the same sub-directory
test_input_dirs = input_dirs * 2
# Directories for output features, which are []
test_output_dirs = [[]] * 2
| 3,561 | 30.803571 | 75 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/config_train_toyset.py | #!/usr/bin/env python
"""
config.py for project-NN-pytorch/projects
Usage:
For training, change Configuration for training stage
For inference, change Configuration for inference stage (although
this configuration will only be used for training)
"""
import os
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
#########################################################
## Configuration for training stage
#########################################################
# Name of datasets
# this will be used as the name of cache files created for each set
#
# Name for the seed training set, in case you merge multiple data sets as
# a single training set, just specify the name for each subset.
# Here we only have 1 training subset
trn_set_name = ['asvspoof2019_toyset_trn']
# Name for the development set
val_set_name = ['asvspoof2019_toyset_val']
# For convenience, specify a path to the toy data set
# because config*.py will be copied into model-*/config_AL_train_toyset/NN
# we need to use ../../../
tmp = os.path.dirname(__file__) + '/../../../DATA/toy_example'
# File list for training and development sets
# (text file, one file name per line, without name extension)
# we need to provide one lst for each subset
# trn_list[n] will correspond to trn_set_name[n]
# for training set
trn_list = [tmp + '/scp/train.lst']
# for development set
val_list = [tmp + '/scp/val.lst']
# Directories for input data
# We need to provide the path to the directory that saves the input data.
# We assume waveforms for training and development of one subset
# are stored in the same directory.
# Hence, input_dirs[n] is for trn_set_name[n] and val_set_name[n]
#
# If you need to specify a separate val_input_dirs
# val_input_dirs = [[PATH_TO_DEVELOPMENT_SET]]
#
# Each input_dirs[n] is a list,
# for example, input_dirs[n] = [wav, speaker_label, augmented_wav, ...]
#
# Here, input for each file is a single waveform
input_dirs = [[tmp + '/train_dev']]
# Dimensions of input features
# What is the dimension of the input feature
# len(input_dims) should be equal to len(input_dirs[n])
#
# Here, input for each file is a single waveform, dimension is 1
input_dims = [1]
# File name extension for input features
# input_exts = [name_extention_of_feature_1, ...]
# len(input_exts) should be equal to len(input_dirs[n])
#
# Here, input file extension is .wav
# We use .wav not .flac
input_exts = ['.wav']
# Temporal resolution for input features
# This is not relevant for CM but for other projects
# len(input_reso) should be equal to len(input_dirs[n])
# Here, it is 1 for waveform
input_reso = [1]
# Whether input features should be z-normalized
# This is not relevant for CM but for other projects
# len(input_norm) should be equal to len(input_dirs[n])
# Here, it is False for waveform
# We don't normalize the waveform
input_norm = [False]
# Similar configurations for output features
# Here, we set output to empty because we will load
# the target labels from protocol rather than output feature
# '.bin' is also a place holder
output_dirs = [[] for x in input_dirs]
output_dims = [1]
output_exts = ['.bin']
output_reso = [1]
output_norm = [False]
# ===
# Waveform configuration
# ===
# Waveform sampling rate
# wav_samp_rate can be None if no waveform data is used
wav_samp_rate = 16000
# Truncating input sequences so that the maximum length = truncate_seq
# When truncate_seq is larger, more GPU mem required
# If you don't want truncating, please truncate_seq = None
truncate_seq = 64000
# Minimum sequence length
# If sequence length < minimum_len, this sequence is not used for training
# minimum_len can be None
minimum_len = 8000
# Optional argument
# This used to load protocol(s)
# Multiple protocol files can be specified in the list
#
# Note that these protocols should cover all the
# training, development, and pool set data.
# Otherwise, the code will raise an error
#
# Here, this protocol will cover all the data in the toy set
optional_argument = [tmp + '/protocol.txt']
#import augment
#input_trans_fns = [[augment.wav_aug]]
#output_trans_fns = [[]]
#########################################################
## Configuration for inference stage (place holder)
#########################################################
# Please use config_test_*.py inference
# This part is just a place holder
test_set_name = trn_set_name + val_set_name
# List of test set data
# for convenience, you may directly load test_set list here
test_list = trn_list + val_list
# Directories for input features
# input_dirs = [path_of_feature_1, path_of_feature_2, ..., ]
# we assume train and validation data are put in the same sub-directory
test_input_dirs = input_dirs * 2
# Directories for output features, which are []
test_output_dirs = [[]] * 2
| 4,915 | 31.556291 | 75 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/config_auto.py | #!/usr/bin/env python
"""
config.py
This configuration file will read environment variables
for configuration. it is used for evaluation (scoring stage)
It assumes that input data will be waveform files (*.wav)
No need to change settings here
"""
import os
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2022, Xin Wang"
#########################################################
## Configuration for training stage
#########################################################
trn_set_name = ['']
val_set_name = ['']
trn_list = ['']
val_list = ['']
input_dirs = [['']]
input_dims = [1]
input_exts = ['.wav']
input_reso = [1]
input_norm = [False]
output_dirs = [[]]
output_dims = [1]
output_exts = ['.bin']
output_reso = [1]
output_norm = [False]
# Waveform sampling rate
# wav_samp_rate can be None if no waveform data is used
# ASVspoof uses 16000 Hz
wav_samp_rate = 16000
# Truncating input sequences so that the maximum length = truncate_seq
# When truncate_seq is larger, more GPU mem required
# If you don't want truncating, please truncate_seq = None
# For ASVspoof, we don't do truncate here
truncate_seq = None
# Minimum sequence length
# If sequence length < minimum_len, this sequence is not used for training
# minimum_len can be None
# For ASVspoof, we don't set minimum length of input trial
minimum_len = None
# Optional argument
# We will use this optional_argument to read protocol file
# When evaluating on a eval set without protocol file, set this to ['']
optional_argument = ['']
#########################################################
## Configuration for inference stage
#########################################################
# similar options to training stage
test_set_name = [os.getenv('TEMP_DATA_NAME')]
# List of test set data
# for convenience, you may directly load test_set list here
test_list = [test_set_name[0] + '.lst']
# Directories for input features
# input_dirs = [[path_of_feature_1, path_of_feature_2, ..., ]]
# directory of the evaluation set waveform
test_input_dirs = [[os.getenv('TEMP_DATA_DIR')]]
# Directories for output features, which are [[]]
test_output_dirs = [[]]
| 2,190 | 26.049383 | 75 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-HuBERT-XL-fix-LLGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel():
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
self.model.eval()
with torch.no_grad():
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
# not good way, but the path is fixed
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/hubert_xtralarge_ll60k.pt'
# This model produces output feature dimensions (per frame)
ssl_orig_output_dim = 1280
# SSL model is declared as a global var since it is fixed
g_ssl_model = SSLModel(ssl_path, ssl_orig_output_dim)
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# Load CM protocol (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load OOD protocl (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
#
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(64, affine=False),
torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Dropout(0.7)
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32),
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = (v_feat_dim // 16) * 32
else:
assert self.v_emd_dim == (v_feat_dim//16)*32, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear((v_feat_dim // 16) * 32, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(g_ssl_model.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_ssl_feat = g_ssl_model.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
frame_num = hidden_features.shape[1]
hidden_features = hidden_features.view(batch_size, frame_num, -1)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 19,256 | 33.759928 | 91 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-Small-fix-LLGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel():
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
self.model.eval()
with torch.no_grad():
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
# not good way, but the path is fixed
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/wav2vec_small.pt'
# This model produces 768 output feature dimensions (per frame)
ssl_orig_output_dim = 768
# SSL model is declared as a global var since it is fixed
g_ssl_model = SSLModel(ssl_path, ssl_orig_output_dim)
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# Load CM protocol (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load OOD protocl (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
#
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(64, affine=False),
torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Dropout(0.7)
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32),
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = (v_feat_dim // 16) * 32
else:
assert self.v_emd_dim == (v_feat_dim//16)*32, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear((v_feat_dim // 16) * 32, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(g_ssl_model.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_ssl_feat = g_ssl_model.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
frame_num = hidden_features.shape[1]
hidden_features = hidden_features.view(batch_size, frame_num, -1)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 19,249 | 33.747292 | 86 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-XLSR-fix-LLGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel():
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
self.model.eval()
with torch.no_grad():
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
# not good way, but the path is fixed
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/xlsr_53_56k.pt'
# This model produces 1024 output feature dimensions (per frame)
ssl_orig_output_dim = 1024
# SSL model is declared as a global var since it is fixed
g_ssl_model = SSLModel(ssl_path, ssl_orig_output_dim)
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# Load CM protocol (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load OOD protocl (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
#
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(64, affine=False),
torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Dropout(0.7)
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32),
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = (v_feat_dim // 16) * 32
else:
assert self.v_emd_dim == (v_feat_dim//16)*32, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear((v_feat_dim // 16) * 32, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(g_ssl_model.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_ssl_feat = g_ssl_model.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
frame_num = hidden_features.shape[1]
hidden_features = hidden_features.view(batch_size, frame_num, -1)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 19,249 | 33.747292 | 86 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-XLSR-fix-LGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel():
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
self.model.eval()
with torch.no_grad():
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
# not good way, but the path is fixed
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/xlsr_53_56k.pt'
# This model produces 1024 output feature dimensions (per frame)
ssl_orig_output_dim = 1024
# SSL model is declared as a global var since it is fixed
g_ssl_model = SSLModel(ssl_path, ssl_orig_output_dim)
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# Load CM protocl (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load OOD protocol (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
# self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
#
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
# just a place holder
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Identity()
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer(v_feat_dim, v_feat_dim),
nii_nn.BLSTMLayer(v_feat_dim, v_feat_dim)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = v_feat_dim
else:
assert self.v_emd_dim == v_feat_dim, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear(v_feat_dim, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(g_ssl_model.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
# The loss includes energy function for in and out dist. data
# But in this project, we will only use the common cross entropy
# This is done by setting self.m_temp = 1 and self.m_lambda = 0.0
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_ssl_feat = g_ssl_model.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _pretransform(self, x_sp_amp, m_trans):
""" A wrapper on the self.m_transform part
"""
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
#hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
#hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
#frame_num = hidden_features.shape[1]
#hidden_features = hidden_features.view(batch_size, frame_num, -1)
hidden_features = m_trans(x_sp_amp)
return hidden_features
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# 1. 2. 3. steps in transform
hidden_features = self._pretransform(x_sp_amp, m_trans)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
This is just a place holder.
Because Model.forward() will compute the loss value.
The output of Model.forward() will be used as the outputs for Loss.compute
Thus, just return outputs from compute()
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 18,117 | 33.120527 | 86 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-XLSR-ft-GF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel(torch_nn.Module):
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
super(SSLModel, self).__init__()
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
# dimension of output from SSL model. This is fixed
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
#self.model.eval()
#with torch.no_grad():
if True:
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# Load CM protocl (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load OOD protocol (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
# dimension of compressed SSL features
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
# number of output class
self.v_out_class = 2
####
# create network
####
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/xlsr_53_56k.pt'
ssl_orig_output_dim = 1024
self.m_ssl = SSLModel(ssl_path, ssl_orig_output_dim)
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
# just a place holder, adopted from LLGF
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Identity()
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
torch_nn.Identity()
)
)
if self.v_emd_dim is None:
self.v_emd_dim = v_feat_dim
else:
assert self.v_emd_dim == v_feat_dim, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear(v_feat_dim, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(self.m_ssl.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
# The loss includes energy function for in and out dist. data
# But in this project, we will only use the common cross entropy
# This is done by setting self.m_temp = 1 and self.m_lambda = 0.0
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
#with torch.no_grad():
x_ssl_feat = self.m_ssl.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _pretransform(self, x_sp_amp, m_trans):
""" A wrapper on the self.m_transform part
"""
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
#hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
#hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
#frame_num = hidden_features.shape[1]
#hidden_features = hidden_features.view(batch_size, frame_num, -1)
hidden_features = m_trans(x_sp_amp)
return hidden_features
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# 1. 2. 3. steps in transform
hidden_features = self._pretransform(x_sp_amp, m_trans)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1) / 2
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
This is just a place holder.
Because Model.forward() will compute the loss value.
The output of Model.forward() will be used as the outputs for Loss.compute
Thus, just return outputs from compute()
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 18,134 | 33.281664 | 88 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-XLSR-ft-LGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel(torch_nn.Module):
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
super(SSLModel, self).__init__()
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
# dimension of output from SSL model. This is fixed
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
#self.model.eval()
#with torch.no_grad():
if True:
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# load protocol for CM (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load protocol for OOD (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
#
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# path to the pre-trained SSL model
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/xlsr_53_56k.pt'
ssl_orig_output_dim = 1024
self.m_ssl = SSLModel(ssl_path, ssl_orig_output_dim)
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
# just a place holder
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Identity()
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer(v_feat_dim, v_feat_dim),
nii_nn.BLSTMLayer(v_feat_dim, v_feat_dim)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = v_feat_dim
else:
assert self.v_emd_dim == v_feat_dim, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear(v_feat_dim, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(self.m_ssl.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
#with torch.no_grad():
x_ssl_feat = self.m_ssl.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _pretransform(self, x_sp_amp, m_trans):
""" A wrapper on the self.m_transform part
"""
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
#hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
#hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
#frame_num = hidden_features.shape[1]
#hidden_features = hidden_features.view(batch_size, frame_num, -1)
hidden_features = m_trans(x_sp_amp)
return hidden_features
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# 1. 2. 3. steps in transform
hidden_features = self._pretransform(x_sp_amp, m_trans)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 17,771 | 32.851429 | 88 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-XLSR-ft-LLGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel(torch_nn.Module):
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
super(SSLModel, self).__init__()
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
# dimension of output from SSL model. This is fixed
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
#self.model.eval()
#with torch.no_grad():
if True:
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# load protocol for CM (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# load protocol for OOD (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
# dimension of compressed SSL features
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# path to the pre-traind SSL model
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/xlsr_53_56k.pt'
ssl_orig_output_dim = 1024
self.ssl_model = SSLModel(ssl_path, ssl_orig_output_dim)
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(64, affine=False),
torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Dropout(0.7)
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32),
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = (v_feat_dim // 16) * 32
else:
assert self.v_emd_dim == (v_feat_dim//16)*32, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear((v_feat_dim // 16) * 32, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(self.ssl_model.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
#with torch.no_grad():
x_ssl_feat = self.ssl_model.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
frame_num = hidden_features.shape[1]
hidden_features = hidden_features.view(batch_size, frame_num, -1)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 19,336 | 33.96745 | 88 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-Large2-fix-LLGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel():
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
self.model.eval()
with torch.no_grad():
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
# not good way, but the path is fixed
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/w2v_large_lv_fsh_swbd_cv.pt'
# This model produces 1024 output feature dimensions (per frame)
ssl_orig_output_dim = 1024
# SSL model is declared as a global var since it is fixed
g_ssl_model = SSLModel(ssl_path, ssl_orig_output_dim)
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# Load CM protocol (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load OOD protocl (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
#
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(64, affine=False),
torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Dropout(0.7)
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32),
nii_nn.BLSTMLayer((v_feat_dim//16) * 32,
(v_feat_dim//16) * 32)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = (v_feat_dim // 16) * 32
else:
assert self.v_emd_dim == (v_feat_dim//16)*32, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear((v_feat_dim // 16) * 32, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(g_ssl_model.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_ssl_feat = g_ssl_model.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
frame_num = hidden_features.shape[1]
hidden_features = hidden_features.view(batch_size, frame_num, -1)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 19,262 | 33.770758 | 93 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-W2V-XLSR-fix-GF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Usage:
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
############################
## FOR pre-trained MODEL
############################
import fairseq
class SSLModel():
def __init__(self, cp_path, ssl_orig_output_dim):
""" SSLModel(cp_path, ssl_orig_output_dim)
Args:
cp_path: string, path to the pre-trained SSL model
ssl_orig_output_dim: int, dimension of the SSL model output feature
"""
model, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task([cp_path])
self.model = model[0]
self.out_dim = ssl_orig_output_dim
return
def extract_feat(self, input_data):
""" feature = extract_feat(input_data)
Args:
input_data: tensor, waveform, (batch, length)
Return:
feature: tensor, feature, (batch, frame_num, feat_dim)
"""
if next(self.model.parameters()).device != input_data.device \
or next(self.model.parameters()).dtype != input_data.dtype:
self.model.to(input_data.device, dtype=input_data.dtype)
self.model.eval()
with torch.no_grad():
if input_data.ndim == 3:
input_tmp = input_data[:, :, 0]
else:
input_tmp = input_data
# [batch, length, dim]
emb = self.model(input_tmp, mask=False, features_only=True)['x']
return emb
# not good way, but the path is fixed
ssl_path = os.path.dirname(__file__) + '/../../../SSL_pretrained/xlsr_53_56k.pt'
# This model produces 1024 output feature dimensions (per frame)
ssl_orig_output_dim = 1024
# SSL model is declared as a global var since it is fixed
g_ssl_model = SSLModel(ssl_path, ssl_orig_output_dim)
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
# Load CM protocl (if available)
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
# Load OOD protocol (if available)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
#self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
#
self.v_feat_dim = [128]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.v_feat_dim)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, v_feat_dim in enumerate(self.v_feat_dim):
# just a place holder
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Identity()
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
torch_nn.Identity()
)
)
if self.v_emd_dim is None:
self.v_emd_dim = v_feat_dim
else:
assert self.v_emd_dim == v_feat_dim, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear(v_feat_dim, self.v_out_class)
)
self.m_frontend.append(
torch_nn.Linear(g_ssl_model.out_dim, v_feat_dim)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
# The loss includes energy function for in and out dist. data
# But in this project, we will only use the common cross entropy
# This is done by setting self.m_temp = 1 and self.m_lambda = 0.0
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_ssl_feat = g_ssl_model.extract_feat(wav.squeeze(-1))
# return
return self.m_frontend[idx](x_ssl_feat)
def _pretransform(self, x_sp_amp, m_trans):
""" A wrapper on the self.m_transform part
"""
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
#hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
#hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
#frame_num = hidden_features.shape[1]
#hidden_features = hidden_features.view(batch_size, frame_num, -1)
hidden_features = m_trans(x_sp_amp)
return hidden_features
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros(
[batch_size * self.v_submodels, self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.m_transform, self.m_before_pooling,
self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, datalength)
# 1. 2. 3. steps in transform
hidden_features = self._pretransform(x_sp_amp, m_trans)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1) / 2
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
This is just a place holder.
Because Model.forward() will compute the loss value.
The output of Model.forward() will be used as the outputs for Loss.compute
Thus, just return outputs from compute()
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 18,028 | 33.210626 | 86 | py |
project-NN-Pytorch-scripts | project-NN-Pytorch-scripts-master/project/07-asvspoof-ssl/model-LFCC-LLGF/model.py | #!/usr/bin/env python
"""
model.py
Self defined model definition.
Adopted from OOD project.
"""
from __future__ import absolute_import
from __future__ import print_function
import os
import sys
import numpy as np
import torch
import torch.nn as torch_nn
import torchaudio
import torch.nn.functional as torch_nn_func
import sandbox.block_nn as nii_nn
import sandbox.util_frontend as nii_front_end
import core_scripts.other_tools.debug as nii_debug
import core_scripts.data_io.seq_info as nii_seq_tk
import sandbox.eval_asvspoof as nii_asvspoof
__author__ = "Xin Wang"
__email__ = "[email protected]"
__copyright__ = "Copyright 2020, Xin Wang"
#################
## Misc functions
#################
# A function to load in/out label for OOD detection. This is just a place holder
# in this project
g_attack_map = {}
def protocol_parse_general(protocol_filepaths, g_map, sep=' ', target_row=-1):
""" Parse protocol of ASVspoof2019 and get bonafide/spoof for each trial
The format is:
SPEAKER TRIAL_NAME - SPOOF_TYPE TAG
LA_0031 LA_E_5932896 - A13 spoof
LA_0030 LA_E_5849185 - - bonafide
...
input:
-----
protocol_filepath: string, path to the protocol file
target_row: int, default -1, use line[-1] as the target label
output:
-------
data_buffer: dic, data_bufer[filename] -> 1 (bonafide), 0 (spoof)
"""
data_buffer = nii_asvspoof.CustomDict(missing_value=True)
if type(protocol_filepaths) is str:
tmp = [protocol_filepaths]
else:
tmp = protocol_filepaths
for protocol_filepath in tmp:
if len(protocol_filepath) and os.path.isfile(protocol_filepath):
with open(protocol_filepath, 'r') as file_ptr:
for line in file_ptr:
line = line.rstrip('\n')
cols = line.split(sep)
if g_map:
try:
data_buffer[cols[1]] = g_map[cols[target_row]]
except KeyError:
data_buffer[cols[1]] = False
else:
data_buffer[cols[1]] = True
return data_buffer
##############
## FOR MODEL
##############
class Model(torch_nn.Module):
""" Model definition
"""
def __init__(self, in_dim, out_dim, args, prj_conf, mean_std=None):
super(Model, self).__init__()
##### required part, no need to change #####
# mean std of input and output
in_m, in_s, out_m, out_s = self.prepare_mean_std(
in_dim,out_dim, args, prj_conf, mean_std)
self.input_mean = torch_nn.Parameter(in_m, requires_grad=False)
self.input_std = torch_nn.Parameter(in_s, requires_grad=False)
self.output_mean = torch_nn.Parameter(out_m, requires_grad=False)
self.output_std = torch_nn.Parameter(out_s, requires_grad=False)
# a flag for debugging (by default False)
#self.model_debug = False
#self.validation = False
#####
####
# on input waveform and output target
####
# Load protocol and prepare the target data for network training
protocol_f = prj_conf.optional_argument
self.protocol_parser = nii_asvspoof.protocol_parse_general(protocol_f)
self.in_out_parser = protocol_parse_general(protocol_f, g_attack_map,
' ', -2)
# Working sampling rate
# torchaudio may be used to change sampling rate
self.m_target_sr = 16000
####
# optional configs (not used)
####
# re-sampling (optional)
#self.m_resampler = torchaudio.transforms.Resample(
# prj_conf.wav_samp_rate, self.m_target_sr)
# vad (optional)
#self.m_vad = torchaudio.transforms.Vad(sample_rate = self.m_target_sr)
# flag for balanced class (temporary use)
#self.v_flag = 1
####
# front-end configuration
# multiple front-end configurations may be used
# by default, use a single front-end
####
# frame shift (number of waveform points)
self.frame_hops = [160]
# frame length
self.frame_lens = [320]
# FFT length
self.fft_n = [512]
# LFCC dim (base component)
self.lfcc_dim = [20]
self.lfcc_with_delta = True
# window type
self.win = torch.hann_window
# floor in log-spectrum-amplitude calculating (not used)
self.amp_floor = 0.00001
# number of frames to be kept for each trial
# no truncation
self.v_truncate_lens = [None for x in self.frame_hops]
# number of sub-models (by default, a single model)
self.v_submodels = len(self.frame_lens)
# dimension of embedding vectors
# here, the embedding is just the activation before sigmoid()
self.v_emd_dim = None
self.v_out_class = 2
####
# create network
####
# 1st part of the classifier
self.m_transform = []
#
self.m_before_pooling = []
# 2nd part of the classifier
self.m_output_act = []
# front-end
self.m_frontend = []
# confidence predictor
self.m_conf = []
# it can handle models with multiple front-end configuration
# by default, only a single front-end
for idx, (trunc_len, fft_n, lfcc_dim) in enumerate(zip(
self.v_truncate_lens, self.fft_n, self.lfcc_dim)):
fft_n_bins = fft_n // 2 + 1
if self.lfcc_with_delta:
lfcc_dim = lfcc_dim * 3
self.m_transform.append(
torch_nn.Sequential(
torch_nn.Conv2d(1, 64, [5, 5], 1, padding=[2, 2]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 96, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 96, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(48, affine=False),
torch_nn.Conv2d(48, 128, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch.nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Conv2d(64, 128, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(64, affine=False),
torch_nn.Conv2d(64, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [1, 1], 1, padding=[0, 0]),
nii_nn.MaxFeatureMap2D(),
torch_nn.BatchNorm2d(32, affine=False),
torch_nn.Conv2d(32, 64, [3, 3], 1, padding=[1, 1]),
nii_nn.MaxFeatureMap2D(),
torch_nn.MaxPool2d([2, 2], [2, 2]),
torch_nn.Dropout(0.7)
)
)
self.m_before_pooling.append(
torch_nn.Sequential(
nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32),
nii_nn.BLSTMLayer((lfcc_dim//16) * 32, (lfcc_dim//16) * 32)
)
)
if self.v_emd_dim is None:
self.v_emd_dim = (lfcc_dim // 16) * 32
else:
assert self.v_emd_dim == (lfcc_dim//16) * 32, "v_emd_dim error"
self.m_output_act.append(
torch_nn.Linear((lfcc_dim // 16) * 32, self.v_out_class)
)
self.m_frontend.append(
nii_front_end.LFCC(self.frame_lens[idx],
self.frame_hops[idx],
self.fft_n[idx],
self.m_target_sr,
self.lfcc_dim[idx],
with_energy=True)
)
self.m_frontend = torch_nn.ModuleList(self.m_frontend)
self.m_transform = torch_nn.ModuleList(self.m_transform)
self.m_output_act = torch_nn.ModuleList(self.m_output_act)
self.m_before_pooling = torch_nn.ModuleList(self.m_before_pooling)
# output
self.m_loss = torch_nn.CrossEntropyLoss()
self.m_temp = 1
self.m_lambda = 0.
self.m_e_m_in = -25.0
self.m_e_m_out = -7.0
# done
return
def prepare_mean_std(self, in_dim, out_dim, args,
prj_conf, data_mean_std=None):
""" prepare mean and std for data processing
This is required for the Pytorch project, but not relevant to this code
"""
if data_mean_std is not None:
in_m = torch.from_numpy(data_mean_std[0])
in_s = torch.from_numpy(data_mean_std[1])
out_m = torch.from_numpy(data_mean_std[2])
out_s = torch.from_numpy(data_mean_std[3])
if in_m.shape[0] != in_dim or in_s.shape[0] != in_dim:
print("Input dim: {:d}".format(in_dim))
print("Mean dim: {:d}".format(in_m.shape[0]))
print("Std dim: {:d}".format(in_s.shape[0]))
print("Input dimension incompatible")
sys.exit(1)
if out_m.shape[0] != out_dim or out_s.shape[0] != out_dim:
print("Output dim: {:d}".format(out_dim))
print("Mean dim: {:d}".format(out_m.shape[0]))
print("Std dim: {:d}".format(out_s.shape[0]))
print("Output dimension incompatible")
sys.exit(1)
else:
in_m = torch.zeros([in_dim])
in_s = torch.ones([in_dim])
out_m = torch.zeros([out_dim])
out_s = torch.ones([out_dim])
return in_m, in_s, out_m, out_s
def normalize_input(self, x):
""" normalizing the input data
This is required for the Pytorch project, but not relevant to this code
"""
return (x - self.input_mean) / self.input_std
def normalize_target(self, y):
""" normalizing the target data
This is required for the Pytorch project, but not relevant to this code
"""
return (y - self.output_mean) / self.output_std
def denormalize_output(self, y):
""" denormalizing the generated output from network
This is required for the Pytorch project, but not relevant to this code
"""
return y * self.output_std + self.output_mean
def _front_end(self, wav, idx, trunc_len, datalength):
""" simple fixed front-end to extract features
input:
------
wav: waveform
idx: idx of the trial in mini-batch
trunc_len: number of frames to be kept after truncation
datalength: list of data length in mini-batch
output:
-------
x_sp_amp: front-end featues, (batch, frame_num, frame_feat_dim)
"""
with torch.no_grad():
x_sp_amp = self.m_frontend[idx](wav.squeeze(-1))
# return
return x_sp_amp
def _compute_embedding(self, x, datalength):
""" definition of forward method
Assume x (batchsize, length, dim)
Output x (batchsize * number_filter, output_dim)
"""
# resample if necessary
#x = self.m_resampler(x.squeeze(-1)).unsqueeze(-1)
# number of sub models
batch_size = x.shape[0]
# buffer to store output scores from sub-models
output_emb = torch.zeros([batch_size * self.v_submodels,
self.v_emd_dim],
device=x.device, dtype=x.dtype)
# compute scores for each sub-models
for idx, (fs, fl, fn, trunc_len, m_trans, m_be_pool, m_output) in \
enumerate(
zip(self.frame_hops, self.frame_lens, self.fft_n,
self.v_truncate_lens, self.m_transform,
self.m_before_pooling, self.m_output_act)):
# extract front-end feature
x_sp_amp = self._front_end(x, idx, trunc_len, datalength)
# compute scores
# 1. unsqueeze to (batch, 1, frame_length, fft_bin)
# 2. compute hidden features
hidden_features = m_trans(x_sp_amp.unsqueeze(1))
# 3. (batch, channel, frame//N, feat_dim//N) ->
# (batch, frame//N, channel * feat_dim//N)
# where N is caused by conv with stride
hidden_features = hidden_features.permute(0, 2, 1, 3).contiguous()
frame_num = hidden_features.shape[1]
hidden_features = hidden_features.view(batch_size, frame_num, -1)
# 4. pooling
# 4. pass through LSTM then summing
hidden_features_lstm = m_be_pool(hidden_features)
# 5. pass through the output layer
tmp_emb = (hidden_features_lstm + hidden_features).mean(1)
output_emb[idx * batch_size : (idx+1) * batch_size] = tmp_emb
return output_emb
def _compute_logit(self, feature_vec, inference=False):
"""
"""
# number of sub models
batch_size = feature_vec.shape[0]
# buffer to store output scores from sub-models
output_act = torch.zeros(
[batch_size * self.v_submodels, self.v_out_class],
device=feature_vec.device, dtype=feature_vec.dtype)
# compute scores for each sub-models
for idx, m_output in enumerate(self.m_output_act):
tmp_emb = feature_vec[idx*batch_size : (idx+1)*batch_size]
output_act[idx*batch_size : (idx+1)*batch_size] = m_output(tmp_emb)
# feature_vec is [batch * submodel, output-class]
return output_act
def _compute_score(self, logits):
"""
"""
# [batch * submodel, output-class], logits
# [:, 1] denotes being bonafide
if logits.shape[1] == 2:
return logits[:, 1] - logits[:, 0]
else:
return logits[:, -1]
def _get_target(self, filenames):
try:
return [self.protocol_parser[x] for x in filenames]
except KeyError:
print("Cannot find target data for %s" % (str(filenames)))
sys.exit(1)
def _clamp_prob(self, input_prob, clamp_val=1e-12):
return torch.clamp(input_prob, 0.0 + clamp_val, 1.0 - clamp_val)
def _get_in_out_indx(self, filenames):
in_indx = []
out_indx = []
for x, y in enumerate(filenames):
if self.in_out_parser[y]:
in_indx.append(x)
else:
out_indx.append(x)
return np.array(in_indx), np.array(out_indx)
def _energy(self, logits):
"""
"""
# - T \log \sum_y \exp (logits[x, y] / T)
eng = - self.m_temp * torch.logsumexp(logits / self.m_temp, dim=1)
return eng
def _loss(self, logits, targets, energy, in_indx, out_indx):
"""
"""
# loss over cross-entropy on in-dist. data
if len(in_indx):
loss = self.m_loss(logits[in_indx], targets[in_indx])
else:
loss = 0
# loss on energy of in-dist.data
if len(in_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(energy[in_indx] - self.m_e_m_in), 2).mean()
# loss on energy of out-dist. data
if len(out_indx):
loss += self.m_lambda * torch.pow(
torch_nn_func.relu(self.m_e_m_out - energy[out_indx]), 2).mean()
return loss
def forward(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
# too short sentences
if not self.training and x.shape[1] < 3000:
targets = self._get_target(filenames)
for filename, target in zip(filenames, targets):
print("Output, %s, %d, %f, %f" % (
filename, target, 0.0, 0.0))
return None
feature_vec = self._compute_embedding(x, datalength)
logits = self._compute_logit(feature_vec)
energy = self._energy(logits)
in_indx, out_indx = self._get_in_out_indx(filenames)
if self.training:
# target
target = self._get_target(filenames)
target_vec = torch.tensor(target, device=x.device, dtype=torch.long)
target_vec = target_vec.repeat(self.v_submodels)
# loss
loss = self._loss(logits, target_vec, energy, in_indx, out_indx)
return loss
else:
scores = self._compute_score(logits)
targets = self._get_target(filenames)
for filename, target, score, energytmp in \
zip(filenames, targets, scores, energy):
print("Output, %s, %d, %f, %f" % (
filename, target, score.item(), -energytmp.item()))
# don't write output score as a single file
return None
def get_embedding(self, x, fileinfo):
filenames = [nii_seq_tk.parse_filename(y) for y in fileinfo]
datalength = [nii_seq_tk.parse_length(y) for y in fileinfo]
feature_vec = self._compute_embedding(x, datalength)
return feature_vec
class Loss():
""" Wrapper to define loss function
"""
def __init__(self, args):
"""
"""
def compute(self, outputs, target):
"""
"""
return outputs
if __name__ == "__main__":
print("Definition of model")
| 18,598 | 34.29222 | 80 | py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.