游雁
2023-12-19 0e622e694e6cb4459955f1e5942a7c53349ce640
funasr/models/e_branchformer/encoder.py
File was renamed from funasr/models/e_branchformer/e_branchformer_encoder.py
@@ -13,9 +13,8 @@
from typing import List, Optional, Tuple
import torch
from funasr.models.ctc import CTC
from funasr.models.encoder.abs_encoder import AbsEncoder
import torch.nn as nn
from funasr.models.ctc.ctc import CTC
from funasr.models.branchformer.cgmlp import ConvolutionalGatingMLP
from funasr.models.branchformer.fastformer import FastSelfAttention
from funasr.models.transformer.utils.nets_utils import get_activation, make_pad_mask
@@ -34,8 +33,8 @@
from funasr.models.transformer.positionwise_feed_forward import (
    PositionwiseFeedForward,
)
from funasr.models.transformer.repeat import repeat
from funasr.models.transformer.subsampling import (
from funasr.models.transformer.utils.repeat import repeat
from funasr.models.transformer.utils.subsampling import (
    Conv2dSubsampling,
    Conv2dSubsampling2,
    Conv2dSubsampling6,
@@ -43,7 +42,7 @@
    TooShortUttError,
    check_short_utt,
)
from funasr.utils.register import register_class
class EBranchformerEncoderLayer(torch.nn.Module):
    """E-Branchformer encoder layer module.
@@ -175,8 +174,8 @@
        return x, mask
class EBranchformerEncoder(AbsEncoder):
@register_class("encoder_classes", "EBranchformerEncoder")
class EBranchformerEncoder(nn.Module):
    """E-Branchformer encoder module."""
    def __init__(