org.apache.plc4x.java.scraper.config.ScraperConfiguration Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of plc4j-scraper Show documentation
Show all versions of plc4j-scraper Show documentation
Utility to efficiently collect a large number of items on multiple devices by different triggers.
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.plc4x.java.scraper.config;
import com.fasterxml.jackson.annotation.JsonCreator;
import com.fasterxml.jackson.annotation.JsonProperty;
import com.fasterxml.jackson.core.JsonFactory;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.fasterxml.jackson.databind.exc.MismatchedInputException;
import com.fasterxml.jackson.dataformat.yaml.YAMLFactory;
import org.apache.plc4x.java.scraper.ScrapeJob;
import org.apache.plc4x.java.scraper.ScrapeJobImpl;
import org.apache.plc4x.java.scraper.exception.ScraperConfigurationException;
import org.apache.plc4x.java.scraper.exception.ScraperException;
import org.apache.plc4x.java.scraper.ScraperImpl;
import org.apache.plc4x.java.scraper.config.triggeredscraper.TriggeredJobConfiguration;
import org.apache.plc4x.java.scraper.triggeredscraper.TriggeredScrapeJobImpl;
import java.io.File;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
import java.util.Set;
import java.util.function.Function;
import java.util.stream.Collectors;
/**
* Configuration class for {@link ScraperImpl}.
*
* @deprecated Scraper is deprecated please use {@link org.apache.plc4x.java.scraper.config.triggeredscraper.TriggeredScraperConfiguration} instead all functions are supplied as well see java-doc of {@link org.apache.plc4x.java.scraper.triggeredscraper.TriggeredScraperImpl}
*/
@Deprecated
public class ScraperConfiguration {
private final Map sources;
private final List jobConfigurations;
/**
* Default constructor.
*
* @param sources Map from connection alias to connection string
* @param jobConfigurations List of configurations one for each Job
*/
@JsonCreator
public ScraperConfiguration(@JsonProperty(value = "sources", required = true) Map sources,
@JsonProperty(value = "jobs", required = true) List jobConfigurations) {
checkNoUnreferencedSources(sources, jobConfigurations);
// TODO Warning on too many sources?!
this.sources = sources;
this.jobConfigurations = jobConfigurations;
}
private void checkNoUnreferencedSources(Map sources, List jobConfigurations) {
Set unreferencedSources = jobConfigurations.stream()
.flatMap(job -> job.getSources().stream())
.filter(source -> !sources.containsKey(source))
.collect(Collectors.toSet());
if (!unreferencedSources.isEmpty()) {
throw new ScraperConfigurationException("There are the following unreferenced sources: " + unreferencedSources);
}
}
public static ScraperConfiguration fromYaml(String yaml) {
ObjectMapper mapper = new ObjectMapper(new YAMLFactory());
try {
return mapper.readValue(yaml, ScraperConfiguration.class);
} catch (IOException e) {
throw new ScraperConfigurationException("Unable to parse given yaml configuration!", e);
}
}
public static ScraperConfiguration fromJson(String json) {
ObjectMapper mapper = new ObjectMapper(new JsonFactory());
try {
return mapper.readValue(json, ScraperConfiguration.class);
} catch (IOException e) {
throw new ScraperConfigurationException("Unable to parse given json configuration!", e);
}
}
public static ScraperConfiguration fromFile(String path) throws IOException {
ObjectMapper mapper;
if (path.endsWith("json")) {
mapper = new ObjectMapper(new JsonFactory());
} else if (path.endsWith("yml") || path.endsWith("yaml")) {
mapper = new ObjectMapper(new YAMLFactory());
} else {
throw new ScraperConfigurationException("Only files with extensions json, yml or yaml can be read");
}
try {
return mapper.readValue(new File(path), ScraperConfiguration.class);
} catch (FileNotFoundException e) {
throw new ScraperConfigurationException("Unable to find configuration given configuration file at '" + path + "'", e);
} catch (MismatchedInputException e) {
throw new ScraperConfigurationException("Given configuration is in wrong format!", e);
}
}
public Map getSources() {
return sources;
}
public List getJobConfigurations() {
return jobConfigurations;
}
public List getJobs() throws ScraperException {
List scrapeJobs = new ArrayList<>();
for(JobConfiguration jobConfiguration:jobConfigurations){
if(jobConfiguration instanceof JobConfigurationImpl){
JobConfigurationImpl jobConfigurationImpl = (JobConfigurationImpl)jobConfiguration;
scrapeJobs.add(new ScrapeJobImpl(jobConfiguration.getName(),
jobConfigurationImpl.getScrapeRate(),
getSourcesForAliases(jobConfiguration.getSources()),
jobConfiguration.getFields()));
}
else{
if(jobConfiguration instanceof TriggeredJobConfiguration){
TriggeredJobConfiguration triggeredJobConfiguration = (TriggeredJobConfiguration) jobConfiguration;
scrapeJobs.add(new TriggeredScrapeJobImpl(jobConfiguration.getName(),
triggeredJobConfiguration.getTriggerConfig(),
getSourcesForAliases(jobConfiguration.getSources()),
jobConfiguration.getFields()));
}
}
}
return scrapeJobs;
}
private Map getSourcesForAliases(List aliases) {
return aliases.stream()
.collect(Collectors.toMap(
Function.identity(),
sources::get
));
}
}