smile.deep.activation.HardShrink Maven / Gradle / Ivy
The newest version!
/*
* Copyright (c) 2010-2024 Haifeng Li. All rights reserved.
*
* Smile is free software: you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* Smile is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with Smile. If not, see .
*/
package smile.deep.activation;
import org.bytedeco.pytorch.Scalar;
import org.bytedeco.pytorch.global.torch;
import smile.deep.tensor.Tensor;
/**
* Hard Shrink activation function.
*
* @author Haifeng Li
*/
public class HardShrink extends ActivationFunction {
/** The lambda value in the formulation. */
final Scalar lambda;
/** Constructor. */
public HardShrink() {
this(0.5);
}
/**
* Constructor.
* @param lambda The lambda value in the formulation.
*/
public HardShrink(double lambda) {
super(String.format("HardShrink(%.4f)", lambda), false);
if (lambda < 0.0) {
throw new IllegalArgumentException("Invalid lambda: " + lambda);
}
this.lambda = new Scalar(lambda);
}
@Override
public Tensor forward(Tensor x) {
return new Tensor(torch.hardshrink(x.asTorch(), lambda));
}
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy