From fa745aa38edce88f5d92e7f7242a111c3d892161 Mon Sep 17 00:00:00 2001 From: Jordan Gong Date: Wed, 23 Dec 2020 19:21:29 +0800 Subject: Remove redundant Leaky ReLU in FocalConv2d --- models/layers.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/models/layers.py b/models/layers.py index e737df2..a0e35f0 100644 --- a/models/layers.py +++ b/models/layers.py @@ -2,7 +2,6 @@ from typing import Union, Tuple import torch import torch.nn as nn -import torch.nn.functional as F class FocalConv2d(nn.Module): @@ -24,4 +23,4 @@ class FocalConv2d(nn.Module): split_size = h // 2 ** self.halving z = x.split(split_size, dim=2) z = torch.cat([self.conv(_) for _ in z], dim=2) - return F.leaky_relu(z, inplace=True) + return z -- cgit v1.2.3