All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.nd4j.versioncheck.VersionCheck Maven / Gradle / Ivy

There is a newer version: 1.0.0-M2.1
Show newest version
/*******************************************************************************
 * Copyright (c) 2015-2018 Skymind, Inc.
 *
 * This program and the accompanying materials are made available under the
 * terms of the Apache License, Version 2.0 which is available at
 * https://www.apache.org/licenses/LICENSE-2.0.
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
 * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
 * License for the specific language governing permissions and limitations
 * under the License.
 *
 * SPDX-License-Identifier: Apache-2.0
 ******************************************************************************/

package org.nd4j.versioncheck;

import lombok.extern.slf4j.Slf4j;
import org.nd4j.config.ND4JSystemProperties;

import java.io.IOException;
import java.net.URI;
import java.net.URL;
import java.nio.file.*;
import java.nio.file.attribute.BasicFileAttributes;
import java.util.*;

/**
 * A runtime version check utility that does 2 things:
* (a) validates the versions of ND4J, DL4J, DataVec, RL4J, Arbiter on the class path, logging a warning if * incompatible versions are found
* (b) allows users to get version information for the above projects at runtime. * * @author Alex Black */ @Slf4j public class VersionCheck { /** * @deprecated Use {@link org.nd4j.config.ND4JSystemProperties#VERSION_CHECK_PROPERTY} */ @Deprecated public static final String VERSION_CHECK_PROPERTY = ND4JSystemProperties.VERSION_CHECK_PROPERTY; public static final String GIT_PROPERTY_FILE_SUFFIX = "-git.properties"; private static final String SCALA_210_SUFFIX = "_2.10"; private static final String SCALA_211_SUFFIX = "_2.11"; private static final String SPARK_1_VER_STRING = "spark_1"; private static final String SPARK_2_VER_STRING = "spark_2"; private static final String UNKNOWN_VERSION = "(Unknown)"; private static final String DL4J_GROUPID = "org.deeplearning4j"; private static final String DL4J_ARTIFACT = "deeplearning4j-nn"; private static final String DATAVEC_GROUPID = "org.datavec"; private static final String DATAVEC_ARTIFACT = "datavec-api"; private static final String ND4J_GROUPID = "org.nd4j"; private static final String ND4J_JBLAS_CLASS = "org.nd4j.linalg.jblas.JblasBackend"; private static final String CANOVA_CLASS = "org.canova.api.io.data.DoubleWritable"; private static final Set GROUPIDS_TO_CHECK = new HashSet<>(Arrays.asList( ND4J_GROUPID, DL4J_GROUPID, DATAVEC_GROUPID)); //NOTE: DL4J_GROUPID also covers Arbiter and RL4J /** * Detailed level for logging: * GAV: display group ID, artifact, version * GAVC: display group ID, artifact, version, commit ID * FULL: display group ID, artifact, version, commit ID, build time, branch, commit message */ public enum Detail { GAV, GAVC, FULL } private VersionCheck(){ } /** * Perform a check of the versions of ND4J, DL4J, DataVec, RL4J and Arbiter dependencies, logging a warning * if necessary. */ public static void checkVersions(){ boolean doCheck = Boolean.parseBoolean(System.getProperty(ND4JSystemProperties.VERSION_CHECK_PROPERTY, "true")); if(!doCheck){ return; } if(classExists(ND4J_JBLAS_CLASS)) { //nd4j-jblas is ancient and incompatible log.error("Found incompatible/obsolete backend and version (nd4j-jblas) on classpath. ND4J is unlikely to" + " function correctly with nd4j-jblas on the classpath. JVM will now exit."); System.exit(1); } if(classExists(CANOVA_CLASS)) { //Canova is ancient and likely to pull in incompatible dependencies log.error("Found incompatible/obsolete library Canova on classpath. ND4J is unlikely to" + " function correctly with this library on the classpath. JVM will now exit."); System.exit(1); } List dependencies = getVersionInfos(); if(dependencies.size() <= 2){ //No -properties.git files were found on the classpath. This may be due to a misconfigured uber-jar // or maybe running in IntelliJ with "dynamic.classpath" set to true (in workspace.xml). Either way, // we can't check versions and don't want to log an error, which will more often than not be wrong if(dependencies.size() == 0){ return; } //Another edge case: no -properties.git files were found, but DL4J and/or DataVec were inferred // by class names. If these "inferred by opName" versions were the only things found, we should also // not log a warning, as we can't check versions in this case boolean dl4jViaClass = false; boolean datavecViaClass = false; for(VersionInfo vi : dependencies ){ if(DL4J_GROUPID.equals(vi.getGroupId()) && DL4J_ARTIFACT.equals(vi.getArtifactId()) && (UNKNOWN_VERSION.equals(vi.getBuildVersion()))){ dl4jViaClass = true; } else if(DATAVEC_GROUPID.equals(vi.getGroupId()) && DATAVEC_ARTIFACT.equals(vi.getArtifactId()) && (UNKNOWN_VERSION.equals(vi.getBuildVersion()))){ datavecViaClass = true; } } if(dependencies.size() == 1 && (dl4jViaClass || datavecViaClass)){ return; } else if(dependencies.size() == 2 && dl4jViaClass && datavecViaClass){ return; } } Set foundVersions = new HashSet<>(); for(VersionInfo vi : dependencies){ String g = vi.getGroupId(); if(g != null && GROUPIDS_TO_CHECK.contains(g)){ String version = vi.getBuildVersion(); if(version.contains("_spark_")){ //Normalize spark versions: // "0.9.1_spark_1" to "0.9.1" and "0.9.1_spark_1-SNAPSHOT" to "0.9.1-SNAPSHOT" version = version.replaceAll("_spark_1",""); version = version.replaceAll("_spark_2",""); } foundVersions.add(version); } } boolean logVersions = false; if(foundVersions.size() > 1){ log.warn("*** ND4J VERSION CHECK FAILED - INCOMPATIBLE VERSIONS FOUND ***"); log.warn("Incompatible versions (different version number) of DL4J, ND4J, RL4J, DataVec, Arbiter are unlikely to function correctly"); logVersions = true; } //Also: check for mixed scala versions - but only for our dependencies... These are in the artifact ID, // scored like dl4j-spack_2.10 and deeplearning4j-ui_2.11 //And check for mixed spark versions (again, just DL4J/DataVec etc dependencies for now) boolean scala210 = false; boolean scala211 = false; boolean spark1 = false; boolean spark2 = false; for(VersionInfo vi : dependencies){ String artifact = vi.getArtifactId(); if(!scala210 && artifact.contains(SCALA_210_SUFFIX)){ scala210 = true; } if(!scala211 && artifact.contains(SCALA_211_SUFFIX)){ scala211 = true; } String version = vi.getBuildVersion(); if(!spark1 && version.contains(SPARK_1_VER_STRING)){ spark1 = true; } if(!spark2 && version.contains(SPARK_2_VER_STRING)){ spark2 = true; } } if(scala210 && scala211){ log.warn("*** ND4J VERSION CHECK FAILED - FOUND BOTH SCALA VERSION 2.10 AND 2.11 ARTIFACTS ***"); log.warn("Projects with mixed Scala versions (2.10/2.11) are unlikely to function correctly"); logVersions = true; } if(spark1 && spark2){ log.warn("*** ND4J VERSION CHECK FAILED - FOUND BOTH SPARK VERSION 1 AND 2 ARTIFACTS ***"); log.warn("Projects with mixed Spark versions (1 and 2) are unlikely to function correctly"); logVersions = true; } if(logVersions){ log.info("Versions of artifacts found on classpath:"); logVersionInfo(); } } /** * @return A list of the property files containing the build/version info */ public static List listGitPropertiesFiles() { Enumeration roots; try { roots = VersionCheck.class.getClassLoader().getResources("ai/skymind/"); } catch (IOException e){ //Should never happen? log.debug("Error listing resources for version check", e); return Collections.emptyList(); } final List out = new ArrayList<>(); while(roots.hasMoreElements()){ URL u = roots.nextElement(); try { URI uri = u.toURI(); try (FileSystem fileSystem = (uri.getScheme().equals("jar") ? FileSystems.newFileSystem(uri, Collections.emptyMap()) : null)) { Path myPath = Paths.get(uri); Files.walkFileTree(myPath, new SimpleFileVisitor() { @Override public FileVisitResult visitFile(Path file, BasicFileAttributes attrs) { URI fileUri = file.toUri(); String s = fileUri.toString(); if(s.endsWith(GIT_PROPERTY_FILE_SUFFIX)){ out.add(fileUri); } return FileVisitResult.CONTINUE; } }); } } catch (Exception e){ //log and skip log.debug("Error finding/loading version check resources", e); } } Collections.sort(out); //Equivalent to sorting by groupID and artifactID return out; } /** * @return A list containing the information for the discovered dependencies */ public static List getVersionInfos() { boolean dl4jFound = false; boolean datavecFound = false; List repState = new ArrayList<>(); for(URI s : listGitPropertiesFiles()){ VersionInfo grs; try{ grs = new VersionInfo(s); } catch (Exception e){ log.debug("Error reading property files for {}", s); continue; } repState.add(grs); if(!dl4jFound && DL4J_GROUPID.equalsIgnoreCase(grs.getGroupId()) && DL4J_ARTIFACT.equalsIgnoreCase(grs.getArtifactId())){ dl4jFound = true; } if(!datavecFound && DATAVEC_GROUPID.equalsIgnoreCase(grs.getGroupId()) && DATAVEC_ARTIFACT.equalsIgnoreCase(grs.getArtifactId())){ datavecFound = true; } } if(classExists(ND4J_JBLAS_CLASS)){ //nd4j-jblas is ancient and incompatible log.error("Found incompatible/obsolete backend and version (nd4j-jblas) on classpath. ND4J is unlikely to" + " function correctly with nd4j-jblas on the classpath."); } if(classExists(CANOVA_CLASS)){ //Canova is anchient and likely to pull in incompatible log.error("Found incompatible/obsolete library Canova on classpath. ND4J is unlikely to" + " function correctly with this library on the classpath."); } return repState; } private static boolean classExists(String className){ try{ Class.forName(className); return true; } catch (ClassNotFoundException e ){ //OK - not found } return false; } /** * @return A string representation of the version information, with the default (GAV) detail level */ public static String versionInfoString() { return versionInfoString(Detail.GAV); } /** * Get the version information for dependencies as a string with a specified amount of detail * * @param detail Detail level for the version information. See {@link Detail} * @return Version information, as a String */ public static String versionInfoString(Detail detail) { StringBuilder sb = new StringBuilder(); for(VersionInfo grp : getVersionInfos()){ sb.append(grp.getGroupId()).append(" : ").append(grp.getArtifactId()).append(" : ").append(grp.getBuildVersion()); switch (detail){ case FULL: case GAVC: sb.append(" - ").append(grp.getCommitIdAbbrev()); if(detail != Detail.FULL) break; sb.append("buildTime=").append(grp.getBuildTime()).append("branch=").append(grp.getBranch()) .append("commitMsg=").append(grp.getCommitMessageShort()); } sb.append("\n"); } return sb.toString(); } /** * Log of the version information with the default level of detail */ public static void logVersionInfo(){ logVersionInfo(Detail.GAV); } /** * Log the version information with the specified level of detail * @param detail Level of detail for logging */ public static void logVersionInfo(Detail detail){ List info = getVersionInfos(); for(VersionInfo grp : info){ switch (detail){ case GAV: log.info("{} : {} : {}", grp.getGroupId(), grp.getArtifactId(), grp.getBuildVersion()); break; case GAVC: log.info("{} : {} : {} - {}", grp.getGroupId(), grp.getArtifactId(), grp.getBuildVersion(), grp.getCommitIdAbbrev()); break; case FULL: log.info("{} : {} : {} - {}, buildTime={}, buildHost={} branch={}, commitMsg={}", grp.getGroupId(), grp.getArtifactId(), grp.getBuildVersion(), grp.getCommitId(), grp.getBuildTime(), grp.getBuildHost(), grp.getBranch(), grp.getCommitMessageShort()); break; } } } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy