| | |
| | | type=int, |
| | | help="Number of encoder hidden units", |
| | | ) |
| | | group.add_argument( |
| | | "--eprojs", default=320, type=int, help="Number of encoder projection units" |
| | | ) |
| | | group.add_argument("--eprojs", default=320, type=int, help="Number of encoder projection units") |
| | | group.add_argument( |
| | | "--subsample", |
| | | default="1", |
| | |
| | | choices=["lstm", "gru"], |
| | | help="Type of decoder network architecture", |
| | | ) |
| | | group.add_argument( |
| | | "--dlayers", default=1, type=int, help="Number of decoder layers" |
| | | ) |
| | | group.add_argument( |
| | | "--dunits", default=320, type=int, help="Number of decoder hidden units" |
| | | ) |
| | | group.add_argument("--dlayers", default=1, type=int, help="Number of decoder layers") |
| | | group.add_argument("--dunits", default=320, type=int, help="Number of decoder hidden units") |
| | | group.add_argument( |
| | | "--dropout-rate-decoder", |
| | | default=0.0, |
| | |
| | | type=int, |
| | | help="Number of attention transformation dimensions", |
| | | ) |
| | | group.add_argument( |
| | | "--awin", default=5, type=int, help="Window size for location2d attention" |
| | | ) |
| | | group.add_argument("--awin", default=5, type=int, help="Window size for location2d attention") |
| | | group.add_argument( |
| | | "--aheads", |
| | | default=4, |