org.elasticsearch.hadoop.rest.EsHadoopNoNodesLeftException Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of elasticsearch-spark-20_2.12 Show documentation
Show all versions of elasticsearch-spark-20_2.12 Show documentation
Elasticsearch Spark (for Spark 2.X)
/*
* Licensed to Elasticsearch under one or more contributor
* license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright
* ownership. Elasticsearch licenses this file to you under
* the Apache License, Version 2.0 (the "License"); you may
* not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.elasticsearch.hadoop.rest;
import java.util.Collection;
import java.util.Collections;
import java.util.Map;
/**
* Exception indicating that due to errors, all available nodes have been processed and no
* other nodes are left for retrying.
*/
public class EsHadoopNoNodesLeftException extends EsHadoopTransportException {
private final Map nodesUsed;
public EsHadoopNoNodesLeftException() {
super(initMessage(Collections. emptyMap()));
nodesUsed = Collections.emptyMap();
}
public EsHadoopNoNodesLeftException(Map nodesUsed) {
super(initMessage(nodesUsed));
this.nodesUsed = nodesUsed;
}
private static String initMessage(Map nodesUsed) {
return String.format("Connection error (check network and/or proxy settings)- all nodes failed; tried [%s] ", nodesUsed.keySet());
}
public Collection nodesUsed() {
return nodesUsed.keySet();
}
}