Remove trailing whitespace in dnn

This commit is contained in:
Marcus Asteborg 2023-06-22 05:27:54 -07:00
parent 26ab10d0c8
commit f36685fc97
No known key found for this signature in database
GPG key ID: F69798291D4CE42A
37 changed files with 231 additions and 246 deletions

View file

@ -124,4 +124,3 @@ Suitable training material can be obtained from [Open Speech and Language Resour
1. [LPCNet: DSP-Boosted Neural Speech Synthesis](https://people.xiph.org/~jm/demo/lpcnet/) 1. [LPCNet: DSP-Boosted Neural Speech Synthesis](https://people.xiph.org/~jm/demo/lpcnet/)
1. [A Real-Time Wideband Neural Vocoder at 1.6 kb/s Using LPCNet](https://people.xiph.org/~jm/demo/lpcnet_codec/) 1. [A Real-Time Wideband Neural Vocoder at 1.6 kb/s Using LPCNet](https://people.xiph.org/~jm/demo/lpcnet_codec/)
1. Sample model files (check compatibility): https://media.xiph.org/lpcnet/data/ 1. Sample model files (check compatibility): https://media.xiph.org/lpcnet/data/

View file

@ -171,4 +171,3 @@ The corresponding citations for all these datasets are:
journal={arXiv preprint arXiv:2104.01497}, journal={arXiv preprint arXiv:2104.01497},
year={2021} year={2021}
} }

View file

@ -9,4 +9,3 @@ if not exist %model% (
tar -xvzf %model% tar -xvzf %model%
move .\src\*.c . move .\src\*.c .
move .\src\*.h . move .\src\*.h .

View file

@ -260,4 +260,3 @@ int main(int argc, char **argv) {
lpcnet_encoder_destroy(st); lpcnet_encoder_destroy(st);
return 0; return 0;
} }

View file

@ -326,4 +326,3 @@ void apply_window(float *x) {
x[WINDOW_SIZE - 1 - i] *= half_window[i]; x[WINDOW_SIZE - 1 - i] *= half_window[i];
} }
} }

View file

@ -126,5 +126,3 @@ int main() {
int test2 = test_sparse_sgemv_accum16(); int test2 = test_sparse_sgemv_accum16();
return test1 || test2; return test1 || test2;
} }

View file

@ -210,4 +210,3 @@ if args.debug_output:
print(f"writing debug output {packet_file[:-4] + f'_torch_batch{batch}_offset{offset}.f32'}") print(f"writing debug output {packet_file[:-4] + f'_torch_batch{batch}_offset{offset}.f32'}")
test_features_full.tofile(packet_file[:-4] + f'_torch_batch{batch}_offset{offset}.f32') test_features_full.tofile(packet_file[:-4] + f'_torch_batch{batch}_offset{offset}.f32')

View file

@ -65,4 +65,3 @@ class RDOVAEDataset(torch.utils.data.Dataset):
rate_lambda = self.lambda_min * np.exp(q_ids.astype(np.float32) / self.denominator).astype(np.float32) rate_lambda = self.lambda_min * np.exp(q_ids.astype(np.float32) / self.denominator).astype(np.float32)
return features, rate_lambda, q_ids return features, rate_lambda, q_ids

View file

@ -611,4 +611,3 @@ class RDOVAE(nn.Module):
for p in self.statistical_model.parameters(): for p in self.statistical_model.parameters():
p.requires_grad = True p.requires_grad = True

View file

@ -254,4 +254,3 @@ if args.debug_output:
print(f"writing debug output {packet_file[:-4] + f'_tf_batch{batch}_offset{offset}.f32'}") print(f"writing debug output {packet_file[:-4] + f'_tf_batch{batch}_offset{offset}.f32'}")
test_features_full.tofile(packet_file[:-4] + f'_tf_batch{batch}_offset{offset}.f32') test_features_full.tofile(packet_file[:-4] + f'_tf_batch{batch}_offset{offset}.f32')

View file

@ -97,4 +97,3 @@ def loss_matchlar():
loss_lar_diff = tf.square(loss_lar_diff) loss_lar_diff = tf.square(loss_lar_diff)
return tf.reduce_mean(loss_lar_diff, axis=-1) return tf.reduce_mean(loss_lar_diff, axis=-1)
return loss return loss

View file

@ -370,4 +370,3 @@ def new_rdovae_model(nb_used_features=20, nb_bits=17, bunch=4, nb_quant=40, batc
model.nb_used_features = nb_used_features model.nb_used_features = nb_used_features
return model, encoder, decoder, qembedding return model, encoder, decoder, qembedding

View file

@ -118,5 +118,3 @@ for c in range(0, nb_frames):
#print(mem) #print(mem)
np.array([np.round(mem)], dtype='int16').tofile(fout) np.array([np.round(mem)], dtype='int16').tofile(fout)
skip = 0 skip = 0