megengine.module.qat.conv 源代码

# MegEngine is Licensed under the Apache License, Version 2.0 (the "License")
#
# Copyright (c) 2014-2021 Megvii Inc. All rights reserved.
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
from ... import functional as F
from .. import conv as Float
from .module import QATModule


[文档]class Conv2d(Float.Conv2d, QATModule): r"""A :class:`~.QATModule` :class:`~.module.Conv2d` with QAT support. Could be applied with :class:`~.Observer` and :class:`~.FakeQuantize`. """
[文档] def calc_conv_qat(self, inp): w_qat = self.apply_quant_weight(self.weight) b_qat = self.apply_quant_bias(self.bias, inp, w_qat) conv = self.calc_conv(inp, w_qat, b_qat) return conv
[文档] @classmethod def from_float_module(cls, float_module: Float.Conv2d): r""" Return a :class:`~.QATModule` instance converted from a float :class:`~.Module` instance. """ qat_module = cls( float_module.in_channels, float_module.out_channels, float_module.kernel_size, float_module.stride, float_module.padding, float_module.dilation, float_module.groups, float_module.bias is not None, float_module.conv_mode, float_module.compute_mode, name=float_module.name, ) qat_module.weight = float_module.weight qat_module.bias = float_module.bias return qat_module
[文档] def forward(self, inp): return self.apply_quant_activation(self.calc_conv_qat(inp))
[文档]class ConvRelu2d(Conv2d): r"""A :class:`~.QATModule` include :class:`~.module.Conv2d` and :func:`~.relu` with QAT support. Could be applied with :class:`~.Observer` and :class:`~.FakeQuantize`. """
[文档] def forward(self, inp): return self.apply_quant_activation(F.relu(self.calc_conv_qat(inp)))
class ConvTranspose2d(Float.ConvTranspose2d, QATModule): r"""A :class:`~.QATModule` :class:`~.module.ConvTranspose2d` with QAT support. Could be applied with :class:`~.Observer` and :class:`~.FakeQuantize`. """ def calc_conv_transpose2d_qat(self, inp): w_qat = self.apply_quant_weight(self.weight) b_qat = self.apply_quant_bias(self.bias, inp, w_qat) conv = self.calc_conv_transpose2d(inp, w_qat, b_qat) return conv @classmethod def from_float_module(cls, float_module: Float.ConvTranspose2d): r""" Return a :class:`~.QATModule` instance converted from a float :class:`~.Module` instance. """ qat_module = cls( float_module.in_channels, float_module.out_channels, float_module.kernel_size, float_module.stride, float_module.padding, float_module.dilation, float_module.groups, float_module.bias is not None, float_module.conv_mode, float_module.compute_mode, name=float_module.name, ) qat_module.weight = float_module.weight qat_module.bias = float_module.bias return qat_module def forward(self, inp): return self.apply_quant_activation(self.calc_conv_transpose2d_qat(inp))