aws-samples/SageMaker_seq2seq_WordPronunciation

Name: SageMaker_seq2seq_WordPronunciation

Owner: AWS Samples

Description: Sequence to Sequence modeling have seen great performance in building models where the input is a sequence of tokens (words for example) and output is also a sequence of tokens. The notebook provides an end-to-end training example of training the English word pronunciation model.

Created: 2018-03-13 22:39:00.0

Updated: 2018-03-20 16:49:30.0

Pushed: 2018-03-16 18:49:22.0

Homepage: null

Size: 37

Language: Jupyter Notebook

GitHub Committers

UserMost Recent Commit# Commits

Other Committers

UserEmailMost Recent Commit# Commits

README

SageMaker_seq2seq_WordPronunciation

Sequence to Sequence modeling have seen great performance in building models where the input is a sequence of tokens (words for example) and output is also a sequence of tokens. The notebook provides an end-to-end example of training and hosting the English word pronunciation model using the Amazon SageMaker built-in Seq2Seq.

SageMaker-Seq2Seq-word-prpnunciation

Jupyter notebook to demonstrate an end-to-end example of training and hosting the English word pronunciation model.

Note: The training the model with the exact same setup will take ~2 hours.

create_vocab_proto.py

Helper python script to generate a recordIO file from pairs of tokenized source and target sequences in numpy array. See also Link.

record_pb2.py

Another helper python script to generate a recordIO file from pairs of tokenized source and target sequences in numpy array. See also Link.

License

This library is licensed under the Apache 2.0 License.


This work is supported by the National Institutes of Health's National Center for Advancing Translational Sciences, Grant Number U24TR002306. This work is solely the responsibility of the creators and does not necessarily represent the official views of the National Institutes of Health.