Skip to content

Commit

Permalink
PR changes
Browse files Browse the repository at this point in the history
  • Loading branch information
bnb32 committed Oct 17, 2023
1 parent 6ae45ac commit 6855fd6
Showing 1 changed file with 3 additions and 6 deletions.
9 changes: 3 additions & 6 deletions phygnn/layers/custom_layers.py
Original file line number Diff line number Diff line change
Expand Up @@ -478,8 +478,7 @@ def __init__(self, name):
Unique string identifier of the skip connection. The skip endpoint
should have the same name.
"""
super().__init__()
self._name = name
super().__init__(name=name)
self._cache = None

def call(self, x):
Expand Down Expand Up @@ -610,8 +609,7 @@ def __init__(self, name=None):
Unique str identifier of the adder layer. Usually the name of the
hi-resolution feature used in the addition.
"""
super().__init__()
self._name = name
super().__init__(name=name)

def call(self, x, hi_res_adder):
"""Adds hi-resolution data to the input tensor x in the middle of a
Expand Down Expand Up @@ -646,8 +644,7 @@ def __init__(self, name=None):
Unique str identifier for the concat layer. Usually the name of the
hi-resolution feature used in the concatenation.
"""
super().__init__()
self._name = name
super().__init__(name=name)

def call(self, x, hi_res_feature):
"""Concatenates a hi-resolution feature to the input tensor x in the
Expand Down

0 comments on commit 6855fd6

Please sign in to comment.