com.sun.enterprise.v3.admin.JobManagerService Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of payara-micro Show documentation
Show all versions of payara-micro Show documentation
Micro Distribution of the Payara Project
/*
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS HEADER.
*
* Copyright (c) 2013 Oracle and/or its affiliates. All rights reserved.
*
* The contents of this file are subject to the terms of either the GNU
* General Public License Version 2 only ("GPL") or the Common Development
* and Distribution License("CDDL") (collectively, the "License"). You
* may not use this file except in compliance with the License. You can
* obtain a copy of the License at
* https://glassfish.dev.java.net/public/CDDL+GPL_1_1.html
* or packager/legal/LICENSE.txt. See the License for the specific
* language governing permissions and limitations under the License.
*
* When distributing the software, include this License Header Notice in each
* file and include the License file at packager/legal/LICENSE.txt.
*
* GPL Classpath Exception:
* Oracle designates this particular file as subject to the "Classpath"
* exception as provided by Oracle in the GPL Version 2 section of the License
* file that accompanied this code.
*
* Modifications:
* If applicable, add the following below the License Header, with the fields
* enclosed by brackets [] replaced by your own identifying information:
* "Portions Copyright [year] [name of copyright owner]"
*
* Contributor(s):
* If you wish your version of this file to be governed by only the CDDL or
* only the GPL Version 2, indicate your decision by adding "[Contributor]
* elects to include this software in this distribution under the [CDDL or GPL
* Version 2] license." If you don't indicate a single choice of license, a
* recipient has the option to distribute your version of this file under
* either the CDDL, the GPL Version 2 or to extend the choice of license to
* its licensees as provided above. However, if you add GPL Version 2 code
* and therefore, elected the GPL Version 2 license, then the option applies
* only if the new code is made subject to such option by the copyright
* holder.
*/
package com.sun.enterprise.v3.admin;
import com.sun.enterprise.admin.event.AdminCommandEventBrokerImpl;
import com.sun.enterprise.admin.remote.RestPayloadImpl;
import com.sun.enterprise.config.serverbeans.Domain;
import com.sun.enterprise.config.serverbeans.ManagedJobConfig;
import com.sun.enterprise.util.LocalStringManagerImpl;
import com.sun.enterprise.util.StringUtils;
import com.sun.enterprise.util.SystemPropertyConstants;
import com.sun.enterprise.v3.admin.CheckpointHelper.CheckpointFilename;
import com.sun.enterprise.v3.server.ExecutorServiceFactory;
import java.io.File;
import java.io.FilenameFilter;
import java.io.IOException;
import java.io.Serializable;
import java.util.*;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.CopyOnWriteArrayList;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.logging.Level;
import java.util.logging.Logger;
import javax.inject.Inject;
import javax.inject.Singleton;
import javax.xml.bind.JAXBContext;
import javax.xml.bind.JAXBException;
import javax.xml.bind.Unmarshaller;
import org.glassfish.api.admin.*;
import org.glassfish.api.admin.progress.JobInfo;
import org.glassfish.api.admin.progress.JobInfos;
import org.glassfish.api.event.EventListener;
import org.glassfish.api.event.EventTypes;
import org.glassfish.api.event.Events;
import org.glassfish.api.event.RestrictTo;
import org.glassfish.hk2.api.PostConstruct;
import org.glassfish.hk2.api.ServiceLocator;
import org.glassfish.kernel.KernelLoggerInfo;
import org.jvnet.hk2.annotations.Service;
import javax.xml.bind.Marshaller;
import org.glassfish.api.ActionReport;
import org.glassfish.api.admin.AdminCommandState.State;
/**
* This is the implementation for the JobManagerService
* The JobManager is responsible
* 1. generating unique ids for jobs
* 2. serving as a registry for jobs
* 3. creating threadpools for jobs
* 4.removing expired jobs
*
* @author Martin Mares
* @author Bhakti Mehta
*/
@Service(name="job-manager")
@Singleton
public class JobManagerService implements JobManager, PostConstruct, EventListener {
private static final String CHECKPOINT_MAINDATA = "MAINCMD";
@Inject
private Domain domain;
private ManagedJobConfig managedJobConfig;
private static final int MAX_SIZE = 65535;
private final ConcurrentHashMap jobRegistry = new ConcurrentHashMap();
private final AtomicInteger lastId = new AtomicInteger(0);
protected static final LocalStringManagerImpl adminStrings =
new LocalStringManagerImpl(JobManagerService.class);
private final static Logger logger = KernelLoggerInfo.getLogger();
private ExecutorService pool;
@Inject
private ExecutorServiceFactory executorFactory;
@Inject
private ServerEnvironment serverEnvironment;
private final String JOBS_FILE = "jobs.xml";
protected JAXBContext jaxbContext;
protected File jobsFile;
@Inject
private ServiceLocator serviceLocator;
@Inject
private JobFileScanner jobFileScanner;
@Inject
Events events;
@Inject
CheckpointHelper checkpointHelper;
@Inject
CommandRunnerImpl commandRunner;
// This will store the data related to completed jobs so that unique ids
// can be generated for new jobs. This is populated lazily the first
// time the JobManagerService is created, it will scan the
//jobs.xml and load the information in memory
private final ConcurrentHashMap completedJobsInfo = new ConcurrentHashMap();
private final ConcurrentHashMap retryableJobsInfo = new ConcurrentHashMap();
/**
* This will return a new id which is unused
* @return
*/
@Override
public synchronized String getNewId() {
int nextId = lastId.incrementAndGet();
if (nextId > MAX_SIZE) {
reset();
}
String nextIdToUse = String.valueOf(nextId);
return !idInUse(nextIdToUse) ? String.valueOf(nextId): getNewId();
}
public JobInfo getCompletedJobForId(String id, File file) {
for (JobInfo jobInfo: getCompletedJobs(file).getJobInfoList()) {
if (jobInfo.jobId.equals(id)) {
return jobInfo;
}
}
return null;
}
@Override
public JobInfo getCompletedJobForId(String id) {
return getCompletedJobForId(id,getJobsFile());
}
/**
* This resets the id to 0
*/
private void reset() {
lastId.set(0);
}
/**
* This method will return if the id is in use
* @param id
* @return true if id is in use
*/
private boolean idInUse(String id) {
return jobRegistry.containsKey(id)
|| completedJobsInfo.containsKey(id)
|| retryableJobsInfo.containsKey(id);
}
/**
* This adds the jobs
* @param job
* @throws IllegalArgumentException
*/
@Override
public synchronized void registerJob(Job job) throws IllegalArgumentException {
if (job == null) {
throw new IllegalArgumentException(adminStrings.getLocalString("job.cannot.be.null","Job cannot be null"));
}
if (jobRegistry.containsKey(job.getId())) {
throw new IllegalArgumentException(adminStrings.getLocalString("job.id.in.use","Job id is already in use."));
}
retryableJobsInfo.remove(job.getId());
jobRegistry.put(job.getId(), job);
if (job.getState() == State.PREPARED && (job instanceof AdminCommandInstanceImpl)) {
((AdminCommandInstanceImpl) job).setState(AdminCommandState.State.RUNNING);
}
}
/**
* This returns all the jobs in the registry
* @return The iterator of jobs
*/
@Override
public Iterator getJobs() {
return jobRegistry.values().iterator();
}
/**
* This will return a job associated with the id
* @param id The job whose id matches
* @return
*/
@Override
public Job get(String id) {
return jobRegistry.get(id);
}
/**
* This will return a list of jobs which have crossed the JOBS_RETENTION_PERIOD
* and need to be purged
* @return list of jobs to be purged
*/
public ArrayList getExpiredJobs(File file) {
ArrayList expiredJobs = new ArrayList();
synchronized (file) {
JobInfos jobInfos = getCompletedJobs(file);
for(JobInfo job:jobInfos.getJobInfoList()) {
long executedTime = job.commandExecutionDate;
long currentTime = System.currentTimeMillis();
long jobsRetentionPeriod = 86400000;
managedJobConfig = domain.getExtensionByType(ManagedJobConfig.class);
jobsRetentionPeriod = convert(managedJobConfig.getJobRetentionPeriod());
if (currentTime - executedTime > jobsRetentionPeriod &&
(job.state.equals(AdminCommandState.State.COMPLETED.name()) ||
job.state.equals(AdminCommandState.State.REVERTED.name()))) {
expiredJobs.add(job);
}
}
}
return expiredJobs;
}
public long convert(String input ) {
String period = input.substring(0,input.length()-1);
Long timeInterval = new Long(period);
String s = input.toLowerCase(Locale.US);
long milliseconds = 86400000;
if (s.indexOf("s") > 0 ) {
milliseconds = timeInterval*1000;
}
else if (s.indexOf("h") > 0 ) {
milliseconds = timeInterval*3600*1000;
}
else if (s.indexOf("m") > 0 ) {
milliseconds = timeInterval*60*1000;
}
return milliseconds;
}
/**
* This will remove the job from the registry
* @param id The job id of the job to be removed
*/
@Override
public synchronized void purgeJob(final String id) {
Job job = jobRegistry.remove(id);
logger.fine(adminStrings.getLocalString("removed.expired.job","Removed expired job ", job));
}
public void deleteCheckpoint(final File parentDir, final String jobId) {
//list all related files
File[] toDelete = parentDir.listFiles(new FilenameFilter() {
@Override
public boolean accept(File dir, String name) {
return name.startsWith(jobId + ".") || name.startsWith(jobId + "-");
}
});
for (File td : toDelete) {
td.delete();
}
}
public ExecutorService getThreadPool() {
return pool ;
}
/**
* This will load the jobs which have already completed
* and persisted in the jobs.xml
* @return JobsInfos which contains information about completed jobs
*/
@Override
public JobInfos getCompletedJobs(File jobsFile) {
synchronized (jobsFile) {
try {
if (jaxbContext == null)
jaxbContext = JAXBContext.newInstance(JobInfos.class);
Unmarshaller unmarshaller = jaxbContext.createUnmarshaller();
if (jobsFile != null && jobsFile.exists()) {
JobInfos jobInfos = (JobInfos)unmarshaller.unmarshal(jobsFile);
return jobInfos;
}
} catch (JAXBException e) {
throw new RuntimeException(adminStrings.getLocalString("error.reading.completed.jobs","Error reading completed jobs ", e.getLocalizedMessage()), e);
}
return null;
}
}
/**
* This method looks for the completed jobs
* and purges a job which is marked with the jobId
* @param jobId the job to purge
* @return the new list of completed jobs
*/
public JobInfos purgeCompletedJobForId(String jobId, File file) {
JobInfos completedJobInfos = getCompletedJobs(file);
synchronized (file) {
CopyOnWriteArrayList jobList = new CopyOnWriteArrayList();
if (completedJobInfos != null) {
jobList.addAll(completedJobInfos.getJobInfoList());
for (JobInfo jobInfo: jobList ) {
if (jobInfo.jobId.equals(jobId)) {
jobList.remove(jobInfo);
}
}
}
JobInfos jobInfos = new JobInfos();
// if (jobList.size() > 0) {
try {
if (jaxbContext == null)
jaxbContext = JAXBContext.newInstance(JobInfos.class);
jobInfos.setJobInfoList(jobList);
Marshaller jaxbMarshaller = jaxbContext.createMarshaller();
jaxbMarshaller.marshal(jobInfos, file);
} catch (JAXBException e) {
throw new RuntimeException(adminStrings.getLocalString("error.purging.completed.job","Error purging completed job ", jobId,e.getLocalizedMessage()), e);
}
//}
return jobInfos;
}
}
@Override
public JobInfos purgeCompletedJobForId(String id) {
return purgeCompletedJobForId(id, getJobsFile()) ;
}
@Override
public void postConstruct() {
jobsFile =
new File(serverEnvironment.getConfigDirPath(),JOBS_FILE);
pool = executorFactory.provide();
HashSet persistedJobFiles = jobFileScanner.getJobFiles();
persistedJobFiles.add(jobsFile);
// Check if there are jobs.xml files which have completed jobs so that
// unique ids get generated
for (File jobfile : persistedJobFiles) {
if (jobfile != null) {
reapCompletedJobs(jobfile);
boolean dropInterruptedCommands = Boolean.valueOf(System.getProperty(SystemPropertyConstants.DROP_INTERRUPTED_COMMANDS));
Collection listed = checkpointHelper.listCheckpoints(jobfile.getParentFile());
for (CheckpointFilename cf : listed) {
if (dropInterruptedCommands) {
logger.info("Dropping checkpoint: " + cf.getFile());
deleteCheckpoint(cf.getParentDir(), cf.getJobId());
} else {
this.retryableJobsInfo.put(cf.getJobId(), cf);
}
}
}
}
events.register(this);
}
@Override
public File getJobsFile() {
return jobsFile;
}
public void addToCompletedJobs(CompletedJob job) {
completedJobsInfo.put(job.getId(),job);
}
public void removeFromCompletedJobs(String id) {
completedJobsInfo.remove(id);
}
public ConcurrentHashMap getCompletedJobsInfo() {
return completedJobsInfo;
}
public ConcurrentHashMap getRetryableJobsInfo() {
return retryableJobsInfo;
}
@Override
public void checkpoint(AdminCommandContext context, Serializable data) throws IOException {
checkpoint((AdminCommand) null, context);
if (data != null) {
checkpointAttachement(context.getJobId(), CHECKPOINT_MAINDATA, data);
}
}
@Override
public void checkpoint(AdminCommand command, AdminCommandContext context) throws IOException {
if (!StringUtils.ok(context.getJobId())) {
throw new IllegalArgumentException("Command is not managed");
}
Job job = get(context.getJobId());
if (job.getJobsFile() == null) {
job.setJobsFile(getJobsFile());
}
Checkpoint chkp = new Checkpoint(job, command, context);
checkpointHelper.save(chkp);
if (job instanceof AdminCommandInstanceImpl) {
((AdminCommandInstanceImpl) job).setState(AdminCommandState.State.RUNNING_RETRYABLE);
}
}
public void checkpointAttachement(String jobId, String attachId, Serializable data) throws IOException {
Job job = get(jobId);
if (job.getJobsFile() == null) {
job.setJobsFile(getJobsFile());
}
checkpointHelper.saveAttachment(data, job, attachId);
}
public T loadCheckpointAttachement(String jobId, String attachId) throws IOException, ClassNotFoundException {
Job job = get(jobId);
if (job.getJobsFile() == null) {
job.setJobsFile(getJobsFile());
}
return checkpointHelper.loadAttachment(job, attachId);
}
@Override
public Serializable loadCheckpointData(String jobId) throws IOException, ClassNotFoundException {
return loadCheckpointAttachement(jobId, CHECKPOINT_MAINDATA);
}
public Checkpoint loadCheckpoint(String jobId, Payload.Outbound outbound) throws IOException, ClassNotFoundException {
Job job = get(jobId);
CheckpointFilename cf = null;
if (job == null) {
cf = getRetryableJobsInfo().get(jobId);
if (cf == null) {
cf = CheckpointFilename.createBasic(jobId, getJobsFile());
}
} else {
cf = CheckpointFilename.createBasic(job);
}
return loadCheckpoint(cf, outbound);
}
private Checkpoint loadCheckpoint(CheckpointFilename cf, Payload.Outbound outbound) throws IOException, ClassNotFoundException {
Checkpoint result = checkpointHelper.load(cf, outbound);
if (result != null) {
serviceLocator.inject(result.getJob());
serviceLocator.postConstruct(result.getJob());
if (result.getCommand() != null) {
serviceLocator.inject(result.getCommand());
serviceLocator.postConstruct(result.getCommand());
}
}
return result;
}
/* This method will look for completed jobs from the jobs.xml
* files and load the information in a local datastructure for
* faster access
*/
protected void reapCompletedJobs(File file) {
if (file != null && file.exists()) {
JobInfos jobInfos = getCompletedJobs(file);
if (jobInfos != null) {
for (JobInfo jobInfo: jobInfos.getJobInfoList()) {
addToCompletedJobs(new CompletedJob(jobInfo.jobId,jobInfo.commandCompletionDate,jobInfo.getJobsFile()));
}
}
}
}
@Override
public void event(@RestrictTo(EventTypes.SERVER_READY_NAME) Event event) {
if (event.is(EventTypes.SERVER_READY)) {
if (retryableJobsInfo.size() > 0) {
Runnable runnable = new Runnable() {
@Override
public void run() {
logger.fine("Restarting retryable jobs");
for (CheckpointFilename cf : retryableJobsInfo.values()) {
reexecuteJobFromCheckpoint(cf);
}
}
};
(new Thread(runnable)).start();
} else {
logger.fine("No retryable job found");
}
}
}
private void reexecuteJobFromCheckpoint(CheckpointFilename cf) {
Checkpoint checkpoint = null;
try {
RestPayloadImpl.Outbound outbound = new RestPayloadImpl.Outbound(true);
checkpoint = loadCheckpoint(cf, outbound);
} catch (Exception ex) {
logger.log(Level.WARNING, KernelLoggerInfo.exceptionLoadCheckpoint, ex);
}
if (checkpoint != null) {
logger.log(Level.INFO, KernelLoggerInfo.checkpointAutoResumeStart,
new Object[]{checkpoint.getJob().getName()});
commandRunner.executeFromCheckpoint(checkpoint, false, new AdminCommandEventBrokerImpl());
ActionReport report = checkpoint.getContext().getActionReport();
logger.log(Level.INFO, KernelLoggerInfo.checkpointAutoResumeDone,
new Object[]{checkpoint.getJob().getName(), report.getActionExitCode(), report.getTopMessagePart()});
}
}
}