/** * \brief An implementation of {@see AbstractInferencer} using Microsoft CUDA ONNX Runtime to do inference * \author Nan Zhou, nanzhou at kneron dot us * \copyright Kneron Inc. All right reserved. */ #ifndef PIANO_DYNASTY_FLOATING_POINT_INCLUDE_MSFT_GPU_MSFTINFERENCER_H_ #define PIANO_DYNASTY_FLOATING_POINT_INCLUDE_MSFT_GPU_MSFTINFERENCER_H_ #include #include #include #include "msft/MSFTInferencer.h" #include "msft-gpu/InferencerPImpl.h" namespace dynasty { namespace inferencer { namespace msftgpu { class InferencerImpl : public dynasty::inferencer::msftgpu::Inferencer { friend class dynasty::inferencer::msftgpu::Inferencer::Builder; explicit InferencerImpl(std::shared_ptr enviroment, std::shared_ptr session) : msftgpu::Inferencer(enviroment, session) {} public: ~InferencerImpl() override = default; }; } } } #endif //PIANO_DYNASTY_FLOATING_POINT_INCLUDE_MSFT_GPU_MSFTINFERENCER_H_