Skip to content
/ Z Public

Alternative layers. for pytorch Non-linear, lower composition, less parameters and faster training (i guess)

Notifications You must be signed in to change notification settings

ZdsAlpha/Z

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

(Note: This description is obsolete. There are major changes in core architecture. But overall idea is same. I will update description.)

Z Layers

Z Layers are alternative layers to vanilla layers. Unlike vanilla layers, Z layers can perform non-linear transformations. So it does not require any activation function.

Z Layer is formed by using N parallel vanilla layers (N considered number of routes) and performing SoftPooling operation between outputs of each layer. This lets layer learn any arbitrary non-linear transformation. The complexity of transformation can be increased by increasing number of routes.

Since Z Layers consist of N parallel vanilla layers, these layers do not increase composition of model. In vanilla models it is required to stack multiple layers to learn a non-linear transformation. But stacking too many layers makes it harder for model to converge. As more layers are added the composition of models keeps increasing. More layers require more time to train. This is why we use skip-connections, dense layers etc to reduce composition. (Here composition means minimum number of linear transformation from input to output in a model's graph)

A single Z layer can approximate 2 to 3 stacked vanilla layers. This also means it has lesser parameters and high information density.

It is not guaranteed to work as its still under research. Feel free to share your results.

Examples

Dataset Model Parameters Accuracy Colab
MNIST C16-SP-C32-SP-F-D-L 73,512 99.63% link
CIFAR-10 CD16-SP-DC32-SP-DC64-SP-F-D-L link

About

Alternative layers. for pytorch Non-linear, lower composition, less parameters and faster training (i guess)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published