| funasr/models/encoder/abs_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/models/encoder/conformer_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/models/encoder/data2vec_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/models/encoder/mfcca_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/models/encoder/resnet34_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/models/encoder/rnn_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/models/encoder/sanm_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/models/encoder/transformer_encoder.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/train/abs_espnet_model.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 |
funasr/models/encoder/abs_encoder.py
File was deleted funasr/models/encoder/conformer_encoder.py
@@ -14,7 +14,6 @@ from typeguard import check_argument_types from funasr.models.ctc import CTC from funasr.models.encoder.abs_encoder import AbsEncoder from funasr.modules.attention import ( MultiHeadedAttention, # noqa: H301 RelPositionMultiHeadedAttention, # noqa: H301 @@ -277,7 +276,7 @@ return x, mask class ConformerEncoder(AbsEncoder): class ConformerEncoder(torch.nn.Module): """Conformer encoder module. Args: funasr/models/encoder/data2vec_encoder.py
@@ -12,7 +12,6 @@ import torch.nn.functional as F from typeguard import check_argument_types from funasr.models.encoder.abs_encoder import AbsEncoder from funasr.modules.data2vec.data_utils import compute_mask_indices from funasr.modules.data2vec.ema_module import EMAModule from funasr.modules.data2vec.grad_multiply import GradMultiply @@ -29,7 +28,7 @@ return end - r * pct_remaining class Data2VecEncoder(AbsEncoder): class Data2VecEncoder(torch.nn.Module): def __init__( self, # for ConvFeatureExtractionModel funasr/models/encoder/mfcca_encoder.py
@@ -34,8 +34,6 @@ from funasr.modules.subsampling import Conv2dSubsampling8 from funasr.modules.subsampling import TooShortUttError from funasr.modules.subsampling import check_short_utt from funasr.models.encoder.abs_encoder import AbsEncoder import pdb import math class ConvolutionModule(nn.Module): @@ -108,7 +106,7 @@ class MFCCAEncoder(AbsEncoder): class MFCCAEncoder(torch.nn.Module): """Conformer encoder module. Args: funasr/models/encoder/resnet34_encoder.py
@@ -1,6 +1,5 @@ import torch from torch.nn import functional as F from funasr.models.encoder.abs_encoder import AbsEncoder from typing import Tuple, Optional from funasr.models.pooling.statistic_pooling import statistic_pooling, windowed_statistic_pooling from collections import OrderedDict @@ -76,7 +75,7 @@ return xs_pad, ilens class ResNet34(AbsEncoder): class ResNet34(torch.nn.Module): def __init__( self, input_size, funasr/models/encoder/rnn_encoder.py
@@ -9,10 +9,9 @@ from funasr.modules.nets_utils import make_pad_mask from funasr.modules.rnn.encoders import RNN from funasr.modules.rnn.encoders import RNNP from funasr.models.encoder.abs_encoder import AbsEncoder class RNNEncoder(AbsEncoder): class RNNEncoder(torch.nn.Module): """RNNEncoder class. Args: funasr/models/encoder/sanm_encoder.py
@@ -26,7 +26,6 @@ from funasr.modules.subsampling import TooShortUttError from funasr.modules.subsampling import check_short_utt from funasr.models.ctc import CTC from funasr.models.encoder.abs_encoder import AbsEncoder from funasr.modules.mask import subsequent_mask, vad_mask class EncoderLayerSANM(nn.Module): @@ -115,7 +114,7 @@ return x, mask, cache, mask_shfit_chunk, mask_att_chunk_encoder class SANMEncoder(AbsEncoder): class SANMEncoder(torch.nn.Module): """ author: Speech Lab, Alibaba Group, China San-m: Memory equipped self-attention for end-to-end speech recognition @@ -547,7 +546,7 @@ return var_dict_torch_update class SANMEncoderChunkOpt(AbsEncoder): class SANMEncoderChunkOpt(torch.nn.Module): """ author: Speech Lab, Alibaba Group, China SCAMA: Streaming chunk-aware multihead attention for online end-to-end speech recognition @@ -960,7 +959,7 @@ return var_dict_torch_update class SANMVadEncoder(AbsEncoder): class SANMVadEncoder(torch.nn.Module): """ author: Speech Lab, Alibaba Group, China funasr/models/encoder/transformer_encoder.py
@@ -13,7 +13,6 @@ import logging from funasr.models.ctc import CTC from funasr.models.encoder.abs_encoder import AbsEncoder from funasr.modules.attention import MultiHeadedAttention from funasr.modules.embedding import PositionalEncoding from funasr.modules.layer_norm import LayerNorm @@ -144,7 +143,7 @@ return x, mask class TransformerEncoder(AbsEncoder): class TransformerEncoder(torch.nn.Module): """Transformer encoder module. Args: funasr/train/abs_espnet_model.py
File was deleted