// Copyright (c) 2020 Mobvoi Inc (Binbin Zhang, Di Wu) // 2022 Binbin Zhang (binbzha@qq.com) // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #ifndef DECODER_TORCH_ASR_MODEL_H_ #define DECODER_TORCH_ASR_MODEL_H_ #include #include #include #include "torch/script.h" #ifndef IOS #include "torch/torch.h" #endif #include "decoder/asr_model.h" #include "../utils/wn_utils.h" namespace wenet { class TorchAsrModel : public AsrModel { public: #ifndef IOS static void InitEngineThreads(int num_threads = 1); #endif public: using TorchModule = torch::jit::script::Module; TorchAsrModel() = default; TorchAsrModel(const TorchAsrModel& other); void Read(const std::string& model_path); std::shared_ptr torch_model() const { return model_; } void Reset() override; void AttentionRescoring(const std::vector>& hyps, float reverse_weight, std::vector* rescoring_score) override; std::shared_ptr Copy() const override; protected: void ForwardEncoderFunc(const std::vector>& chunk_feats, std::vector>* ctc_prob) override; float ComputeAttentionScore(const torch::Tensor& prob, const std::vector& hyp, int eos); private: std::shared_ptr model_ = nullptr; std::vector encoder_outs_; // transformer/conformer attention cache torch::Tensor att_cache_ = torch::zeros({0, 0, 0, 0}); // conformer-only conv_module cache torch::Tensor cnn_cache_ = torch::zeros({0, 0, 0, 0}); }; } // namespace wenet #endif // DECODER_TORCH_ASR_MODEL_H_