funasr/models/scama/utils.py
@@ -1,8 +1,9 @@ import os import torch from torch.nn import functional as F import yaml import torch import numpy as np from torch.nn import functional as F def sequence_mask(lengths, maxlen=None, dtype=torch.float32, device=None): if maxlen is None: