最新google演算法:實現中文TTS的測試結果
阿新 • • 發佈:2018-11-12
簡介
本文主要是實現中文的TTS,沒有接入百度、阿里、騰訊和訊飛的API,僅僅依靠自己的訓練演算法和經過樣本處理和測試而成。
樣本的製作方法:
由於本人時間和金錢的限制,無法找專業的人員錄製大量樣本。本文的解決辦法為:
藉助百度語音合成API
神經百度的語音合成API,編寫一個簡潔的程式碼,實現百度API讀取一本45W字的小說,以每句話作為一個訓練樣本。
import os
import re
from aip import AipSpeech
import time
APP_ID = '114788XX' #你自己申請的API ID
API_KEY = '2m4bO8OV8F21saqe96H8' #你自己申請的API key
SECRET_KEY = 'IO5faSMp7tPkeIjBwClDFTj' #你自己申請的secret key
client = AipSpeech(APP_ID, API_KEY, SECRET_KEY)
# txt_path = 'XX.txt'
txt_path = 'XX.txt' #你自己讓百度API生成訓練樣本的文字
# with open(txt_path, 'r', encoding='utf8') as f:
# text = f.read()
# text = re.sub(r'(.{30})', lambda x: '{}\n'.format(x.group(1)), text)
# with open(txt_path, 'w', encoding='utf8') as f:
# f.write(text)
with open(txt_path, 'r', encoding='utf8') as f:
for index, line in enumerate(f):
index = '2B%06d'%index
# if index < 8331:
# continue
line = line.strip()
try:
res = client.synthesis(line , 'zh', 1, {'per': '4', 'spd': '5', 'vol': '7', 'aue': '6'})
except Exception:
time.sleep(5)
res = client.synthesis(line, 'zh', 1, {'per': '4', 'spd': '5', 'vol': '7', 'aue': '6'})
if not isinstance(res, dict):
with open('./wav/{}.wav'.format(index), 'wb') as f:
f.write(res)
with open('./txt/{}.txt'.format(index), 'w') as f:
#line = pinyin.get(line, format="numerical", delimiter=" ")
f.write(line)
else:
print(index, 'err')
print(index)
# index += 1
訓練及樣本處理
訓練樣本要保持和上一個深度學習之經驗和訓練集(訓練中英文樣本)的ljspeech的訓練樣本的格式。
樣本地址
連結: https://pan.baidu.com/s/1k0auHRQQkSyfGB-nAcwlDA 密碼: 7yyq
訓練核心演算法加群:QQ群:821953467
from __future__ import print_function
import argparse
from datetime import datetime
import json
import os
import sys
import time
import tensorflow as tf
from tensorflow.python.client import timeline
from wavenet import WaveNetModel, AudioReader, optimizer_factory
os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
os.environ["CUDA_VISIBLE_DEVICES"] = "1"
BATCH_SIZE = 1
DATA_DIRECTORY = './VCTK-Corpus'
LOGDIR_ROOT = './logdir'
CHECKPOINT_EVERY = 50
NUM_STEPS = int(1e5)
LEARNING_RATE = 1e-3
WAVENET_PARAMS = './wavenet_params.json'
STARTED_DATESTRING = "{0:%Y-%m-%dT%H-%M-%S}".format(datetime.now())
SAMPLE_SIZE = 100000
L2_REGULARIZATION_STRENGTH = 0
SILENCE_THRESHOLD = 0.3
EPSILON = 0.001
MOMENTUM = 0.9
MAX_TO_KEEP = 5
METADATA = False
def get_arguments():
def _str_to_bool(s):
"""Convert string to bool (in argparse context)."""
if s.lower() not in ['true', 'false']:
raise ValueError('Argument needs to be a '
'boolean, got {}'.format(s))
return {'true': True, 'false': False}[s.lower()]
parser = argparse.ArgumentParser(description='WaveNet example network')
parser.add_argument('--batch_size', type=int, default=BATCH_SIZE,
help='How many wav files to process at once. Default: ' + str(BATCH_SIZE) + '.')
parser.add_argument('--data_dir', type=str, default=DATA_DIRECTORY,
help='The directory containing the VCTK corpus.')
parser.add_argument('--store_metadata', type=bool, default=METADATA,
help='Whether to store advanced debugging information '
'(execution time, memory consumption) for use with '
'TensorBoard. Default: ' + str(METADATA) + '.')
parser.add_argument('--logdir', type=str, default=None,
help='Directory in which to store the logging '
'information for TensorBoard. '
'If the model already exists, it will restore '
'the state and will continue training. '
'Cannot use with --logdir_root and --restore_from.')
parser.add_argument('--logdir_root', type=str, default=None,
help='Root directory to place the logging '
'output and generated model. These are stored '
'under the dated subdirectory of --logdir_root. '
'Cannot use with --logdir.')
parser.add_argument('--restore_from', type=str, default=None,
help='Directory in which to restore the model from. '
'This creates the new model under the dated directory '
'in --logdir_root. '
'Cannot use with --logdir.')
parser.add_argument('--checkpoint_every', type=int,
default=CHECKPOINT_EVERY,
help='How many steps to save each checkpoint after. Default: ' + str(CHECKPOINT_EVERY) + '.')
parser.add_argument('--num_steps', type=int, default=NUM_STEPS,
help='Number of training steps. Default: ' + str(NUM_STEPS) + '.')
parser.add_argument('--learning_rate', type=float, default=LEARNING_RATE,
help='Learning rate for training. Default: ' + str(LEARNING_RATE) + '.')
parser.add_argument('--wavenet_params', type=str, default=WAVENET_PARAMS,
help='JSON file with the network parameters. Default: ' + WAVENET_PARAMS + '.')
parser.add_argument('--sample_size', type=int, default=SAMPLE_SIZE,
help='Concatenate and cut audio samples to this many '
'samples. Default: ' + str(SAMPLE_SIZE) + '.')
parser.add_argument('--l2_regularization_strength', type=float,
default=L2_REGULARIZATION_STRENGTH,
help='Coefficient in the L2 regularization. '
'Default: False')
parser.add_argument('--silence_threshold', type=float,
default=SILENCE_THRESHOLD,
help='Volume threshold below which to trim the start '
'and the end from the training set samples. Default: ' + str(SILENCE_THRESHOLD) + '.')
parser.add_argument('--optimizer', type=str, default='adam',
choices=optimizer_factory.keys(),
help='Select the optimizer specified by this option. Default: adam.')
parser.add_argument('--momentum', type=float,
default=MOMENTUM, help='Specify the momentum to be '
'used by sgd or rmsprop optimizer. Ignored by the '
'adam optimizer. Default: ' + str(MOMENTUM) + '.')
parser.add_argument('--histograms', type=_str_to_bool, default=False,
help='Whether to store histogram summaries. Default: False')
parser.add_argument('--gc_channels', type=int, default=None,
help='Number of global condition channels. Default: None. Expecting: Int')
parser.add_argument('--max_checkpoints', type=int, default=MAX_TO_KEEP,
help='Maximum amount of checkpoints that will be kept alive. Default: '
+ str(MAX_TO_KEEP) + '.')
return parser.parse_args()
def save(saver, sess, logdir, step):
model_name = 'model.ckpt'
checkpoint_path = os.path.join(logdir, model_name)
print('Storing checkpoint to {} ...'.format(logdir), end="")
sys.stdout.flush()
if not os.path.exists(logdir):
os.makedirs(logdir)
saver.save(sess, checkpoint_path, global_step=step)
print(' Done.')
def load(saver, sess, logdir):
print("Trying to restore saved checkpoints from {} ...".format(logdir),
end="")
ckpt = tf.train.get_checkpoint_state(logdir)
if ckpt:
print(" Checkpoint found: {}".format(ckpt.model_checkpoint_path))
global_step = int(ckpt.model_checkpoint_path
.split('/')[-1]
.split('-')[-1])
print(" Global step was: {}".format(global_step))
print(" Restoring...", end="")
saver.restore(sess, ckpt.model_checkpoint_path)
print(" Done.")
return global_step
else:
print(" No checkpoint found.")
return None
def get_default_logdir(logdir_root):
logdir = os.path.join(logdir_root, 'train', STARTED_DATESTRING)
return logdir
def validate_directories(args):
"""Validate and arrange directory related arguments."""
# Validation
if args.logdir and args.logdir_root:
raise ValueError("--logdir and --logdir_root cannot be "
"specified at the same time.")
if args.logdir and args.restore_from:
raise ValueError(
"--logdir and --restore_from cannot be specified at the same "
"time. This is to keep your previous model from unexpected "
"overwrites.\n"
"Use --logdir_root to specify the root of the directory which "
"will be automatically created with current date and time, or use "
"only --logdir to just continue the training from the last "
"checkpoint.")
# Arrangement
logdir_root = args.logdir_root
if logdir_root is None:
logdir_root = LOGDIR_ROOT
logdir = args.logdir
if logdir is None:
logdir = get_default_logdir(logdir_root)
print('Using default logdir: {}'.format(logdir))
restore_from = args.restore_from
if restore_from is None:
# args.logdir and args.restore_from are exclusive,
# so it is guaranteed the logdir here is newly created.
restore_from = logdir
return {
'logdir': logdir,
'logdir_root': args.logdir_root,
'restore_from': restore_from
}
def main():
args = get_arguments()
try:
directories = validate_directories(args)
except ValueError as e:
print("Some arguments are wrong:")
print(str(e))
return
logdir = directories['logdir']
restore_from = directories['restore_from']
# Even if we restored the model, we will treat it as new training
# if the trained model is written into an arbitrary location.
is_overwritten_training = logdir != restore_from
with open(args.wavenet_params, 'r') as f:
wavenet_params = json.load(f)
# Create coordinator.
coord = tf.train.Coordinator()
# Load raw waveform from VCTK corpus.
with tf.name_scope('create_inputs'):
# Allow silence trimming to be skipped by specifying a threshold near
# zero.
silence_threshold = args.silence_threshold if args.silence_threshold > \
EPSILON else None
gc_enabled = args.gc_channels is not None
reader = AudioReader(
args.data_dir,
coord,
sample_rate=wavenet_params['sample_rate'],
gc_enabled=gc_enabled,
receptive_field=WaveNetModel.calculate_receptive_field(wavenet_params["filter_width"],
wavenet_params["dilations"],
wavenet_params["scalar_input"],
wavenet_params["initial_filter_width"]),
sample_size=args.sample_size,
silence_threshold=silence_threshold)
audio_batch = reader.dequeue(args.batch_size)
if gc_enabled:
gc_id_batch = reader.dequeue_gc(args.batch_size)
else:
gc_id_batch = None
# Create network.
net = WaveNetModel(
batch_size=args.batch_size,
dilations=wavenet_params["dilations"],
filter_width=wavenet_params["filter_width"],
residual_channels=wavenet_params["residual_channels"],
dilation_channels=wavenet_params["dilation_channels"],
skip_channels=wavenet_params["skip_channels"],
quantization_channels=wavenet_params["quantization_channels"],
use_biases=wavenet_params["use_biases"],
scalar_input=wavenet_params["scalar_input"],
initial_filter_width=wavenet_params["initial_filter_width"],
histograms=args.histograms,
global_condition_channels=args.gc_channels,
global_condition_cardinality=reader.gc_category_cardinality)
if args.l2_regularization_strength == 0:
args.l2_regularization_strength = None
loss = net.loss(input_batch=audio_batch,
global_condition_batch=gc_id_batch,
l2_regularization_strength=args.l2_regularization_strength)
optimizer = optimizer_factory[args.optimizer](
learning_rate=args.learning_rate,
momentum=args.momentum)
trainable = tf.trainable_variables()
optim = optimizer.minimize(loss, var_list=trainable)
# Set up logging for TensorBoard.
writer = tf.summary.FileWriter(logdir)
writer.add_graph(tf.get_default_graph())
run_metadata = tf.RunMetadata()
summaries = tf.summary.merge_all()
# Set up session
sess = tf.Session(config=tf.ConfigProto(log_device_placement=False))
init = tf.global_variables_initializer()
sess.run(init)
# Saver for storing checkpoints of the model.
saver = tf.train.Saver(var_list=tf.trainable_variables(), max_to_keep=args.max_checkpoints)
try:
saved_global_step = load(saver, sess, restore_from)
if is_overwritten_training or saved_global_step is None:
# The first training step will be saved_global_step + 1,
# therefore we put -1 here for new or overwritten trainings.
saved_global_step = -1
except:
print("Something went wrong while restoring checkpoint. "
"We will terminate training to avoid accidentally overwriting "
"the previous model.")
raise
threads = tf.train.start_queue_runners(sess=sess, coord=coord)
reader.start_threads(sess)
step = None
last_saved_step = saved_global_step
try:
for step in range(saved_global_step + 1, args.num_steps):
start_time = time.time()
if args.store_metadata and step % 50 == 0:
# Slow run that stores extra information for debugging.
print('Storing metadata')
run_options = tf.RunOptions(
trace_level=tf.RunOptions.FULL_TRACE)
summary, loss_value, _ = sess.run(
[summaries, loss, optim],
options=run_options,
run_metadata=run_metadata)
writer.add_summary(summary, step)
writer.add_run_metadata(run_metadata,
'step_{:04d}'.format(step))
tl = timeline.Timeline(run_metadata.step_stats)
timeline_path = os.path.join(logdir, 'timeline.trace')
with open(timeline_path, 'w') as f:
f.write(tl.generate_chrome_trace_format(show_memory=True))
else:
summary, loss_value, _ = sess.run([summaries, loss, optim])
writer.add_summary(summary, step)
duration = time.time() - start_time
print('step {:d} - loss = {:.3f}, ({:.3f} sec/step)'
.format(step, loss_value, duration))
if step % args.checkpoint_every == 0:
save(saver, sess, logdir, step)
last_saved_step = step
except KeyboardInterrupt:
# Introduce a line break after ^C is displayed so save message
# is on its own line.
print()
finally:
if step > last_saved_step:
save(saver, sess, logdir, step)
coord.request_stop()
coord.join(threads)
if __name__ == '__main__':
main()
訓練結果檢驗
測試文字集
1.三間新的房間很漂亮和乾淨.
2.音樂是人放鬆和解除煩惱的一種方式.
3.在農村晚上不要經常外出去活動,因為比較漆黑。
4.海水很藍,天空中飛來一群小鳥
5.秋天是一個收貨的季節,老人在忙碌著
6.老大畢竟兩個老人跟著大兒子過活也因為老兩口面上還算公正三兄弟
7.間沒多少齷齪這次葉小麗跑了之後老兩口更是過來幫他忙上忙下馬四
8.妹這幾天乾脆住在這邊幫他帶著孩子加上原身的記憶李生接受起他們
9.氣生了三個都是女兒想到這馬四妹又犯愁老大媳婦不願意再把孩子送
10.心疼不已媽你怎麼讓來弟洗碗李紅心虛的看了李生一眼把孩子遞給老
11.頭片子養大了還不是別人家的實在不行再找戶人家送了馬四妹堅決反
12.了名聲不好聽身世不親白孩子你抱過去養著戶口遷過去關葉小麗什
測試音訊地址
- 1.wav連結: https://pan.baidu.com/s/12CqB9myfNzzWTJlqAWoJyw 密碼: n3h2
- 2.wav連結: https://pan.baidu.com/s/1UVGVOyaP2HsIIS2Af2hKVw 密碼: euik
- 3.wav連結: https://pan.baidu.com/s/1uo7xfenFdGHhwJG3TiGzRQ 密碼: 686w
- 4.wav連結: https://pan.baidu.com/s/1WSVIZuoDqRYX5md1mMEk5w 密碼: rqkk
- 5.wav連結: https://pan.baidu.com/s/1MZlOoHHkJ4wGnz_4eQnCng 密碼: 57xa
- 6.wav連結: 連結: https://pan.baidu.com/s/19Ta399HJ-iOpitisnjs8sw 密碼: 25av
- 7.wav連結: https://pan.baidu.com/s/1kEikiW5MUAFpUxHuNlZqZg 密碼: hvwu
- 8.wav連結: https://pan.baidu.com/s/1kEikiW5MUAFpUxHuNlZqZg 密碼: hvwu
- 9.wav連結: https://pan.baidu.com/s/1_7z8jGef4MfwBMdnNMhmhA 密碼: g51x
- 10.wav連結: https://pan.baidu.com/s/1uyQ6Cuq0DEhWcW8wLhqhFg 密碼: b3rv
- 11.wav連結: https://pan.baidu.com/s/1ZSk_chv5PlJtsaLh49JfNg 密碼: acgc
- 12.wav連結: https://pan.baidu.com/s/1MBm53MtJtPBBYBZx7KIMFw 密碼: ttan
總結
由於本文生成的測試樣本是訓練了5萬多次,誤差還比較大,還需要進一步的訓練。後期的結果肯定回比百度和訊飛的樣本好很多。