org.deeplearning4j.arbiter.conf.updater.AdaGradSpace Maven / Gradle / Ivy
/*******************************************************************************
* Copyright (c) 2015-2018 Skymind, Inc.
*
* This program and the accompanying materials are made available under the
* terms of the Apache License, Version 2.0 which is available at
* https://www.apache.org/licenses/LICENSE-2.0.
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
* License for the specific language governing permissions and limitations
* under the License.
*
* SPDX-License-Identifier: Apache-2.0
******************************************************************************/
package org.deeplearning4j.arbiter.conf.updater;
import lombok.Data;
import lombok.EqualsAndHashCode;
import lombok.Getter;
import lombok.Setter;
import org.deeplearning4j.arbiter.optimize.api.ParameterSpace;
import org.nd4j.linalg.learning.config.AdaGrad;
import org.nd4j.linalg.learning.config.IUpdater;
import org.nd4j.linalg.schedule.ISchedule;
import org.nd4j.shade.jackson.annotation.JsonProperty;
import java.util.Collections;
import java.util.List;
import java.util.Map;
@Data
@EqualsAndHashCode(callSuper = false)
public class AdaGradSpace extends BaseUpdaterSpace {
private ParameterSpace learningRate;
private ParameterSpace lrSchedule;
@Getter @Setter
private int[] indices;
public AdaGradSpace(ParameterSpace learningRate) {
this(learningRate, null);
}
public static AdaGradSpace withLR(ParameterSpace lr){
return new AdaGradSpace(lr, null);
}
public static AdaGradSpace withLRSchedule(ParameterSpace lrSchedule){
return new AdaGradSpace(null, lrSchedule);
}
protected AdaGradSpace(@JsonProperty("learningRate") ParameterSpace learningRate,
@JsonProperty("lrSchedule") ParameterSpace lrSchedule){
this.learningRate = learningRate;
this.lrSchedule = lrSchedule;
}
@Override
public IUpdater getValue(double[] parameterValues) {
if(lrSchedule != null){
return new AdaGrad(lrSchedule.getValue(parameterValues));
} else {
return new AdaGrad(learningRate.getValue(parameterValues));
}
}
}