代码拉取完成,页面将自动刷新
/*
* Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#pragma once
#include <NvInfer.h>
#include "plugin.hpp"
#include "serialize.hpp"
#include <cudnn.h>
#include <vector>
typedef unsigned short half_type;
namespace {
constexpr const char* INSTANCE_PLUGIN_VERSION{"001"};
constexpr const char* INSTANCE_PLUGIN_NAME{"InstanceNormalization"};
}
class InstanceNormalizationPlugin final : public onnx2trt::PluginV2 {
float _epsilon;
int _nchan;
std::vector<float> _h_scale;
std::vector<float> _h_bias;
float* _d_scale;
float* _d_bias;
bool _initialized;
nvinfer1::Weights _scale, _bias;
cudnnHandle_t _cudnn_handle;
cudnnTensorDescriptor_t _x_desc, _y_desc, _b_desc;
protected:
void deserialize(void const* serialData, size_t serialLength) {
deserializeBase(serialData, serialLength);
deserialize_value(&serialData, &serialLength, &_epsilon);
deserialize_value(&serialData, &serialLength, &_nchan);
deserialize_value(&serialData, &serialLength, &_h_scale);
deserialize_value(&serialData, &serialLength, &_h_bias);
}
size_t getSerializationSize() const override {
return (serialized_size(_epsilon) +
serialized_size(_nchan) +
serialized_size(_h_scale) +
serialized_size(_h_bias)) + getBaseSerializationSize();
}
void serialize(void *buffer) const override {
serializeBase(buffer);
serialize_value(&buffer, _epsilon);
serialize_value(&buffer, _nchan);
serialize_value(&buffer, _h_scale);
serialize_value(&buffer, _h_bias);
}
public:
InstanceNormalizationPlugin(float epsilon,
nvinfer1::Weights const& scale,
nvinfer1::Weights const& bias);
InstanceNormalizationPlugin(void const* serialData, size_t serialLength) : _initialized(false) {
this->deserialize(serialData, serialLength);
}
const char* getPluginType() const override { return INSTANCE_PLUGIN_NAME; }
virtual void destroy() override { delete this; }
virtual nvinfer1::IPluginV2* clone() const override { return new InstanceNormalizationPlugin{_epsilon, _scale, _bias}; }
virtual const char* getPluginVersion() const override { return INSTANCE_PLUGIN_VERSION; }
virtual void setPluginNamespace(const char* pluginNamespace) override {}
virtual const char* getPluginNamespace() const override { return ""; }
bool supportsFormat(nvinfer1::DataType type,
nvinfer1::PluginFormat format) const override;
int getNbOutputs() const override { return 1; }
nvinfer1::Dims getOutputDimensions(int index,
const nvinfer1::Dims *inputDims,
int nbInputs) override {
assert(index == 0);
assert(inputDims);
assert(nbInputs == 1);
return *inputDims;
}
int initialize() override;
void terminate() override;
int enqueue(int batchSize,
const void *const *inputs, void **outputs,
void *workspace, cudaStream_t stream) override;
size_t getWorkspaceSize(int maxBatchSize) const override;
~InstanceNormalizationPlugin();
};
class InstanceNormalizationPluginCreator : public nvinfer1::IPluginCreator
{
public:
InstanceNormalizationPluginCreator() {}
~InstanceNormalizationPluginCreator() {}
const char* getPluginName() const { return INSTANCE_PLUGIN_NAME; }
const char* getPluginVersion() const { return INSTANCE_PLUGIN_VERSION; }
const nvinfer1::PluginFieldCollection* getFieldNames() { std::cerr<< "Function not implemented" << std::endl; return nullptr; }
nvinfer1::IPluginV2* createPlugin(const char* name, const nvinfer1::PluginFieldCollection* fc) { std::cerr<< "Function not implemented" << std::endl; return nullptr; }
nvinfer1::IPluginV2* deserializePlugin(const char* name, const void* serialData, size_t serialLength) { return new InstanceNormalizationPlugin{serialData, serialLength}; }
void setPluginNamespace(const char* libNamespace) { mNamespace = libNamespace; }
const char* getPluginNamespace() const { return mNamespace.c_str(); }
private:
std::string mNamespace;
};
REGISTER_TENSORRT_PLUGIN(InstanceNormalizationPluginCreator);
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。