org.deeplearning4j.spark.data.DataSetExportFunction Maven / Gradle / Ivy
The newest version!
/*
* ******************************************************************************
* *
* *
* * This program and the accompanying materials are made available under the
* * terms of the Apache License, Version 2.0 which is available at
* * https://www.apache.org/licenses/LICENSE-2.0.
* *
* * See the NOTICE file distributed with this work for additional
* * information regarding copyright ownership.
* * Unless required by applicable law or agreed to in writing, software
* * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
* * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
* * License for the specific language governing permissions and limitations
* * under the License.
* *
* * SPDX-License-Identifier: Apache-2.0
* *****************************************************************************
*/
package org.deeplearning4j.spark.data;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.spark.api.java.function.VoidFunction;
import org.apache.spark.broadcast.Broadcast;
import org.datavec.spark.util.DefaultHadoopConfig;
import org.datavec.spark.util.SerializableHadoopConfig;
import org.deeplearning4j.core.util.UIDProvider;
import org.nd4j.linalg.dataset.DataSet;
import java.net.URI;
import java.util.Iterator;
public class DataSetExportFunction implements VoidFunction> {
private final URI outputDir;
private final Broadcast conf;
private String uid = null;
private int outputCount;
public DataSetExportFunction(URI outputDir) {
this(outputDir, null);
}
public DataSetExportFunction(URI outputDir, Broadcast configuration) {
this.outputDir = outputDir;
this.conf = configuration;
}
@Override
public void call(Iterator iter) throws Exception {
String jvmuid = UIDProvider.getJVMUID();
uid = Thread.currentThread().getId() + jvmuid.substring(0, Math.min(8, jvmuid.length()));
Configuration c = conf == null ? DefaultHadoopConfig.get() : conf.getValue().getConfiguration();
while (iter.hasNext()) {
DataSet next = iter.next();
String filename = "dataset_" + uid + "_" + (outputCount++) + ".bin";
String path = outputDir.getPath();
URI uri = new URI(path + (path.endsWith("/") || path.endsWith("\\") ? "" : "/") + filename);
FileSystem file = FileSystem.get(uri, c);
try (FSDataOutputStream out = file.create(new Path(uri))) {
next.save(out);
}
}
}
}