Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed to Metamarkets Group Inc. (Metamarkets) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. Metamarkets licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package io.druid.cli.validate;
import com.fasterxml.jackson.databind.Module;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Throwables;
import com.google.common.collect.ImmutableList;
import com.google.common.collect.Iterables;
import com.google.common.io.CharSource;
import com.google.common.io.LineProcessor;
import com.google.common.io.Resources;
import com.google.inject.Binder;
import com.google.inject.Injector;
import com.google.inject.name.Names;
import com.metamx.common.UOE;
import com.metamx.common.logger.Logger;
import io.airlift.airline.Command;
import io.airlift.airline.Option;
import io.druid.cli.GuiceRunnable;
import io.druid.data.input.InputRow;
import io.druid.data.input.impl.StringInputRowParser;
import io.druid.guice.ExtensionsConfig;
import io.druid.guice.FirehoseModule;
import io.druid.guice.IndexingServiceFirehoseModule;
import io.druid.guice.LocalDataStorageDruidModule;
import io.druid.guice.ParsersModule;
import io.druid.indexer.HadoopDruidIndexerConfig;
import io.druid.indexer.IndexingHadoopModule;
import io.druid.indexing.common.task.Task;
import io.druid.initialization.DruidModule;
import io.druid.initialization.Initialization;
import io.druid.query.Query;
import org.apache.commons.io.output.NullWriter;
import java.io.File;
import java.io.IOException;
import java.io.PrintWriter;
import java.io.Writer;
import java.nio.charset.Charset;
import java.util.Arrays;
import java.util.List;
/**
*/
@Command(
name = "validator",
description = "Validates that a given Druid JSON object is correctly formatted"
)
public class DruidJsonValidator extends GuiceRunnable
{
private static final Logger LOG = new Logger(DruidJsonValidator.class);
private Writer logWriter = new PrintWriter(System.out);
@Option(name = "-f", title = "file", description = "file to validate", required = true)
public String jsonFile;
@Option(name = "-t", title = "type", description = "the type of schema to validate", required = true)
public String type;
@Option(name = "-r", title = "resource", description = "optional resources required for validation", required = false)
public String resource;
@Option(name = "--log", title = "toLogger", description = "redirects any outputs to logger", required = false)
public boolean toLogger;
public DruidJsonValidator()
{
super(LOG);
}
@Override
protected List extends com.google.inject.Module> getModules()
{
return ImmutableList.of(
new com.google.inject.Module()
{
@Override
public void configure(Binder binder)
{
binder.bindConstant().annotatedWith(Names.named("serviceName")).to("druid/validator");
binder.bindConstant().annotatedWith(Names.named("servicePort")).to(0);
}
}
);
}
@Override
public void run()
{
File file = new File(jsonFile);
if (!file.exists()) {
System.out.printf("File[%s] does not exist.%n", file);
}
final Injector injector = makeInjector();
final ObjectMapper jsonMapper = injector.getInstance(ObjectMapper.class);
registerModules(
jsonMapper,
Iterables.concat(
Initialization.getFromExtensions(injector.getInstance(ExtensionsConfig.class), DruidModule.class),
Arrays.asList(
new FirehoseModule(),
new IndexingHadoopModule(),
new IndexingServiceFirehoseModule(),
new LocalDataStorageDruidModule(),
new ParsersModule()
)
)
);
final ClassLoader loader;
if (Thread.currentThread().getContextClassLoader() != null) {
loader = Thread.currentThread().getContextClassLoader();
} else {
loader = DruidJsonValidator.class.getClassLoader();
}
if (toLogger) {
logWriter = new NullWriter()
{
private final Logger logger = new Logger(DruidJsonValidator.class);
@Override
public void write(char[] cbuf, int off, int len)
{
logger.info(new String(cbuf, off, len));
}
};
}
try {
if (type.equalsIgnoreCase("query")) {
jsonMapper.readValue(file, Query.class);
} else if (type.equalsIgnoreCase("hadoopConfig")) {
jsonMapper.readValue(file, HadoopDruidIndexerConfig.class);
} else if (type.equalsIgnoreCase("task")) {
jsonMapper.readValue(file, Task.class);
} else if (type.equalsIgnoreCase("parse")) {
final StringInputRowParser parser;
if (file.isFile()) {
logWriter.write("loading parse spec from file '" + file + "'");
parser = jsonMapper.readValue(file, StringInputRowParser.class);
} else if (loader.getResource(jsonFile) != null) {
logWriter.write("loading parse spec from resource '" + jsonFile + "'");
parser = jsonMapper.readValue(loader.getResource(jsonFile), StringInputRowParser.class);
} else {
logWriter.write("cannot find proper spec from 'file'.. regarding it as a json spec");
parser = jsonMapper.readValue(jsonFile, StringInputRowParser.class);
}
if (resource != null) {
final CharSource source;
if (new File(resource).isFile()) {
logWriter.write("loading data from file '" + resource + "'");
source = Resources.asByteSource(new File(resource).toURL()).asCharSource(
Charset.forName(
parser.getEncoding()
)
);
} else if (loader.getResource(resource) != null) {
logWriter.write("loading data from resource '" + resource + "'");
source = Resources.asByteSource(loader.getResource(resource)).asCharSource(
Charset.forName(
parser.getEncoding()
)
);
} else {
logWriter.write("cannot find proper data from 'resource'.. regarding it as data string");
source = CharSource.wrap(resource);
}
readData(parser, source);
}
} else {
throw new UOE("Unknown type[%s]", type);
}
}
catch (Exception e) {
System.out.println("INVALID JSON!");
throw Throwables.propagate(e);
}
}
private void registerModules(ObjectMapper jsonMapper, Iterable fromExtensions)
{
for (DruidModule druidModule : fromExtensions) {
for (Module module : druidModule.getJacksonModules()) {
jsonMapper.registerModule(module);
}
}
}
@VisibleForTesting
void setLogWriter(Writer writer)
{
this.logWriter = writer;
}
private Void readData(final StringInputRowParser parser, final CharSource source)
throws IOException
{
return source.readLines(
new LineProcessor()
{
private final StringBuilder builder = new StringBuilder();
@Override
public boolean processLine(String line) throws IOException
{
InputRow parsed = parser.parse(line);
builder.append(parsed.getTimestamp());
for (String dimension : parsed.getDimensions()) {
builder.append('\t');
builder.append(parsed.getRaw(dimension));
}
logWriter.write(builder.toString());
builder.setLength(0);
return true;
}
@Override
public Void getResult()
{
return null;
}
}
);
}
}