You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When importances are normalized, i. e. do not change with the uniform scaling of the separate lambda channels, c_{k}_{lambd} doesn't affect anything completely.
Including c_{k}_{lambd} in the iteration would increase computational cost.
c_{k}_{lambd}, where k is l2 is not symmetric with respect to the changing of the order of covariants to contract. Thus it would make single contraction function to loose its general form, where arguments are arbitrary covariants.
The text was updated successfully, but these errors were encountered:
This is an interesting issue, but one that needs testing. Scaling diffferent lambda channels might affect model accuracy in real-life scenarios - I have anecdotal evidence that it does.
When importances are normalized, i. e. do not change with the uniform scaling of the separate lambda channels, c_{k}_{lambd} doesn't affect anything completely.
Including c_{k}_{lambd} in the iteration would increase computational cost.
c_{k}_{lambd}, where k is l2 is not symmetric with respect to the changing of the order of covariants to contract. Thus it would make single contraction function to loose its general form, where arguments are arbitrary covariants.
The text was updated successfully, but these errors were encountered: