All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.helix.webapp.resources.JobQueueResource Maven / Gradle / Ivy

package org.apache.helix.webapp.resources;

/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing,
 * software distributed under the License is distributed on an
 * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 * KIND, either express or implied.  See the License for the
 * specific language governing permissions and limitations
 * under the License.
 */

import java.util.Map;

import org.apache.helix.HelixDataAccessor;
import org.apache.helix.HelixException;
import org.apache.helix.PropertyKey;
import org.apache.helix.zookeeper.datamodel.ZNRecord;
import org.apache.helix.zookeeper.impl.client.ZkClient;
import org.apache.helix.model.ResourceConfig;
import org.apache.helix.task.JobConfig;
import org.apache.helix.task.TaskDriver;
import org.apache.helix.task.TaskUtil;
import org.apache.helix.task.Workflow;
import org.apache.helix.task.WorkflowContext;
import org.restlet.data.Form;
import org.restlet.data.MediaType;
import org.restlet.data.Status;
import org.restlet.representation.Representation;
import org.restlet.representation.StringRepresentation;
import org.restlet.representation.Variant;
import org.restlet.resource.ServerResource;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

/**
 * Class for server-side resource at "/clusters/{clusterName}/jobQueues/{jobQueue}"
 * 

*

  • GET list job queue info *
  • POST start a new job in a job queue, or stop/resume/persistDataChanges/delete a job queue */ public class JobQueueResource extends ServerResource { private final static Logger LOG = LoggerFactory.getLogger(JobQueueResource.class); public JobQueueResource() { getVariants().add(new Variant(MediaType.TEXT_PLAIN)); getVariants().add(new Variant(MediaType.APPLICATION_JSON)); setNegotiated(false); } /** * List job queue info *

    * Usage: curl http://{host:port}/clusters/{clusterName}/jobQueues/{jobQueue} */ @Override public Representation get() { StringRepresentation presentation; try { String clusterName = ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.CLUSTER_NAME); String jobQueueName = ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.JOB_QUEUE); presentation = getHostedEntitiesRepresentation(clusterName, jobQueueName); } catch (Exception e) { String error = ClusterRepresentationUtil.getErrorAsJsonStringFromException(e); presentation = new StringRepresentation(error, MediaType.APPLICATION_JSON); LOG.error("Fail to get job queue", e); } return presentation; } StringRepresentation getHostedEntitiesRepresentation(String clusterName, String jobQueueName) throws Exception { ZkClient zkClient = ResourceUtil.getAttributeFromCtx(getContext(), ResourceUtil.ContextKey.ZKCLIENT); HelixDataAccessor accessor = ClusterRepresentationUtil.getClusterDataAccessor(zkClient, clusterName); PropertyKey.Builder keyBuilder = accessor.keyBuilder(); TaskDriver taskDriver = new TaskDriver(zkClient, clusterName); // Get job queue config // TODO: fix this to use workflowConfig. ResourceConfig jobQueueConfig = accessor.getProperty(keyBuilder.resourceConfig(jobQueueName)); // Get job queue context WorkflowContext ctx = taskDriver.getWorkflowContext(jobQueueName); // Create the result ZNRecord hostedEntitiesRecord = new ZNRecord(jobQueueName); if (jobQueueConfig != null) { hostedEntitiesRecord.merge(jobQueueConfig.getRecord()); } if (ctx != null) { hostedEntitiesRecord.merge(ctx.getRecord()); } StringRepresentation representation = new StringRepresentation(ClusterRepresentationUtil.ZNRecordToJson(hostedEntitiesRecord), MediaType.APPLICATION_JSON); return representation; } /** * Start a new job in a job queue, or stop/resume/persistDataChanges/delete a job queue *

    * Usage: *

    *

  • Start a new job in a job queue: * curl -d @'./{input.txt}' -H 'Content-Type: application/json' * http://{host:port}/clusters/{clusterName}/jobQueues/{jobQueue} *

    * input.txt: jsonParameters={"command":"start"}&newJob={newJobConfig.yaml} *

    * For newJobConfig.yaml, see {@link Workflow#parse(String)} *

  • Stop/resume/persistDataChanges/delete a job queue: * curl -d 'jsonParameters={"command":"{stop/resume/persistDataChanges/delete}"}' * -H "Content-Type: application/json" http://{host:port}/clusters/{clusterName}/jobQueues/{jobQueue} */ @Override public Representation post(Representation entity) { String clusterName = ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.CLUSTER_NAME); String jobQueueName = ResourceUtil.getAttributeFromRequest(getRequest(), ResourceUtil.RequestKey.JOB_QUEUE); ZkClient zkClient = ResourceUtil.getAttributeFromCtx(getContext(), ResourceUtil.ContextKey.ZKCLIENT); try { TaskDriver driver = new TaskDriver(zkClient, clusterName); Form form = new Form(entity); JsonParameters jsonParameters = new JsonParameters(form); TaskDriver.DriverCommand cmd = TaskDriver.DriverCommand.valueOf(jsonParameters.getCommand()); switch (cmd) { case start: { // Get the job queue and submit it String yamlPayload = ResourceUtil.getYamlParameters(form, ResourceUtil.YamlParamKey.NEW_JOB); if (yamlPayload == null) { throw new HelixException("Yaml job config is required!"); } Workflow workflow = Workflow.parse(yamlPayload); for (String jobName : workflow.getJobConfigs().keySet()) { Map jobCfgMap = workflow.getJobConfigs().get(jobName); JobConfig.Builder jobCfgBuilder = JobConfig.Builder.fromMap(jobCfgMap); if (workflow.getTaskConfigs() != null && workflow.getTaskConfigs().containsKey(jobName)) { jobCfgBuilder.addTaskConfigs(workflow.getTaskConfigs().get(jobName)); } driver.enqueueJob(jobQueueName, TaskUtil.getDenamespacedJobName(jobQueueName, jobName), jobCfgBuilder); } break; } case stop: { driver.stop(jobQueueName); break; } case resume: { driver.resume(jobQueueName); break; } case flush: { driver.flushQueue(jobQueueName); break; } case delete: { driver.delete(jobQueueName); break; } case clean: { driver.cleanupQueue(jobQueueName); break; } default: throw new HelixException("Unsupported job queue command: " + cmd); } getResponse().setEntity(getHostedEntitiesRepresentation(clusterName, jobQueueName)); getResponse().setStatus(Status.SUCCESS_OK); } catch (Exception e) { getResponse().setEntity(ClusterRepresentationUtil.getErrorAsJsonStringFromException(e), MediaType.APPLICATION_JSON); getResponse().setStatus(Status.SUCCESS_OK); LOG.error("Error in posting job queue: " + entity, e); } return null; } }