Replies: 1 comment
-
Please check this issue. @plyfager |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Describe the issue
According to "Positional Encoding as Spatial Inductive Bias in GANs", zero padding leads to an unbalanced spatial bias with vague relation between locations. Thorough out the paper, they propose other explicit positional encoding such as cartesian grid or sinusoidal positional encodings. While using these explicit positional encoding, they remove padding from convolution generators and replace them with bilinear upsampling.
However, according to configuration in this project (e.g. https://github.com/open-mmlab/mmgeneration/blob/master/configs/positional_encoding_in_gans/singan_csg_bohemian.py,), I found out that these implementation use padding with size 1 at convolution module. I wonder that this is an exact reimplementation of paper.
No I haven't change anything.
I used ballons.png which provided in original singan repository.
Environment
sys.platform: linux
Python: 3.8.10 (default, Sep 28 2021, 16:10:42) [GCC 9.3.0]
CUDA available: True
CUDA_HOME: /usr/local/cuda
NVCC: Build cuda_11.1.TC455_06.29190527_0
GPU 0,1,2,3,4,5: NVIDIA RTX A6000
GCC: gcc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
PyTorch: 1.8.1+cu111
PyTorch compiling details: PyTorch built with:
TorchVision: 0.9.1+cu111
OpenCV: 4.2.0
MMCV: 1.4.0
MMGen: 0.4.0+ac1c630
MMCV Compiler: GCC 9.3
MMCV CUDA Compiler: 11.1
Results
Currently i don't have any results when removing padding in convolution module.
I will reproduce result as soon as possible.
Beta Was this translation helpful? Give feedback.
All reactions