UNPKG

mesos-framework

Version:

A wrapper around the Mesos HTTP APIs for Schedulers and Executors. Write your Mesos framework in pure JavaScript!

1,108 lines (893 loc) 45.6 kB
"use strict"; var http = require("http"); var EventEmitter = require("events").EventEmitter; var util = require("util"); var uuid = require("uuid"); var helpers = require("./helpers"); var schedulerHandlers = require("./schedulerHandlers"); var mesos = require("./mesos")().getMesos(); var Builder = require("./builder"); var TaskHelper = require("./taskHelper"); var zookeeper = require("node-zookeeper-client"); /** * Represents a Mesos framework scheduler. * @constructor * @param {object} options - The option map object. */ function Scheduler (options) { if (!(this instanceof Scheduler)) { return new Scheduler(options); } // Inherit from EventEmitter EventEmitter.call(this); var self = this; self.options = {}; self.options.frameworkName = (options.frameworkName ? options.frameworkName.replace(/ /g, "-") : "mesos-framework." + uuid.v4()); self.options.user = options.user || "root"; self.options.role = options.role || "*"; self.options.restartStates = options.restartStates || ["TASK_FAILED", "TASK_LOST", "TASK_ERROR"]; // Task in TASK_FINISHED will NOT be restarted by default! self.options.frameworkFailoverTimeout = options.frameworkFailoverTimeout || 604800; // One week self.options.masterConnectionTimeout = options.masterConnectionTimeout * 1000 || 10000; // Ten seconds self.options.exponentialBackoffFactor = options.exponentialBackoffFactor || 1.5; self.options.exponentialBackoffMinimum = options.exponentialBackoffMinimum * 1000 || 1000; // One second self.options.exponentialBackoffMaximum = options.exponentialBackoffMaximum * 1000 || 15000; // 15 seconds self.options.killUnknownTasks = options.killUnknownTasks || false; self.options.serialNumberedTasks = (options.serialNumberedTasks !== false); // ZooKeeper self.options.useZk = options.useZk || false; self.options.zkUrl = options.zkUrl || "master.mesos:2181"; self.options.zkPrefix = options.zkPrefix || "/dcos-service-"; // Logging self.logger = helpers.getLogger((options.logging && options.logging.path ? options.logging.path : null), (options.logging && options.logging.fileName ? options.logging.fileName : null), (options.logging && options.logging.level ? options.logging.level : null)); self.subscribeBackoffTime = 0; // Port allocation self.options.staticPorts = options.staticPorts || false; if (self.options.staticPorts){ self.logger.info("Scheduler configured with fixed ports"); } // Master discovery self.options.masterUrl = options.masterUrl || "127.0.0.1"; self.options.port = parseInt(options.port) || 5050; // "Globals" self.frameworkId = null; self.mesosStreamId = null; self.lastHeartbeat = null; self.zkClient = null; // Tasks self.tasks = []; // Add tasks if there are any if (options.hasOwnProperty("tasks")) { self.tasks = helpers.sortTasksByPriority(options.tasks); } self.logger.debug(JSON.stringify(self.tasks)); self.pendingTasks = []; self.launchedTasks = []; self.reconcileTasks = []; self.killTasks = []; // Add to pending tasks if not yet submitted self.tasks.forEach(function (task) { if (!task.isSubmitted) { self.pendingTasks.push(task); } }); // Runtime info self.runtimeInfo = {}; // Template for issuing Mesos Scheduler HTTP API requests self.requestTemplate = {}; self.generateRequestTemplate = function () { self.requestTemplate = { host: self.options.masterUrl, port: self.options.port, path: "/api/v1/scheduler", method: "POST", headers: { 'Content-Type': 'application/json' } }; }; // Fill the requestTemplate self.generateRequestTemplate(); // Customer event handlers will be registered here self.customEventHandlers = {}; // List of allowed event handler function names and their argument length var allowedEventHandlers = { "SUBSCRIBED": 1, "OFFERS": 1, "INVERSE_OFFERS": 1, "RESCIND":1, "RESCIND_INVERSE_OFFER": 1, "UPDATE": 1, "MESSAGE": 1, "FAILURE": 1, "ERROR": 1, "HEARTBEAT": 1 }; // Add custom event handlers if present if (options.handlers && Object.getOwnPropertyNames(options.handlers).length > 0) { Object.getOwnPropertyNames(options.handlers).forEach(function (handlerName) { var ucHandlerName = handlerName.toUpperCase(); // Check if name is allowed, is a function and the length of the function arguments fit to the ones defined in allowedEventHandlers if (Object.getOwnPropertyNames(allowedEventHandlers).indexOf(ucHandlerName) > -1 && helpers.isFunction(options.handlers[handlerName]) && options.handlers[handlerName].length === allowedEventHandlers[ucHandlerName]) { self.customEventHandlers[ucHandlerName] = options.handlers[handlerName]; } }); } // Fill runtimeInfo from given Tasks if (options.tasks && Object.getOwnPropertyNames(options.tasks).length > 0) { var tempPriority = 1; Object.getOwnPropertyNames(options.tasks).forEach(function (task) { // Populate runtimeInfo for each task self.runtimeInfo[task] = { "desiredInstances": options.tasks[task].instances || 1, "requestedInstances": 0, "runningInstances": {}, "priority": options.tasks[task].priority || tempPriority }; // Increase priority tempPriority++; }); } if (options.hasOwnProperty("tasks")) { self.logger.debug(JSON.stringify(helpers.sortTasksByPriority(options.tasks))); } // Store the long-running request self.httpRequest = {}; // Handling of ZooKeeper-related stuff if (self.options.useZk) { self.logger.debug("Using ZooKeeper for persistency. Connection is " + self.options.zkUrl); // Set default path for the service self.zkServicePath = self.options.zkPrefix + self.options.frameworkName; // Create ZK client (getting from options is only to be used for unit test!) self.zkClient = options.zkClient || zookeeper.createClient(self.options.zkUrl); // Instantiate TaskHelper (getting from options is only to be used for unit test!) self.taskHelper = options.taskHelper || new TaskHelper(self); // For unit test if (options.taskHelper) { self.taskHelper.scheduler = self; } // Set path for the framework id var zkPath = self.zkServicePath + "/framework-id"; self.zkClient.on("error", function (error) { self.logger.error(error); }); // Once connected, check if path exists self.zkClient.once("connected", function () { self.logger.debug("Connected to ZooKeeper on " + self.options.zkUrl); // Check if path with framework id exists self.zkClient.getData(zkPath, null, function (error, data, stat) { self.logger.debug("error before if:" + JSON.stringify(error)); if (error) { self.logger.debug("error:" + JSON.stringify(error)); // Check if node doesn't exist yet if (error.getCode() === zookeeper.Exception.NO_NODE) { self.logger.debug("Node " + zkPath + " doesn't exist yet. Will be created on framework launch"); // Add event handler for the SUBSCRIBE event, to set the framework id to ZooKeeper self.once("subscribed", function (obj) { self.logger.debug("Got subscribed event"); self.logger.debug("now creating path " + zkPath); // Seperating path creation from data save due to various client bugs. self.zkClient.mkdirp(zkPath, function (error, stat) { if (error) { self.logger.error("Got error when creating a ZK node for the framework ID: " + error.stack); self.options.useZk = false; self.zkClient.close(); } else { self.zkClient.setData(zkPath, new Buffer(self.frameworkId), function (error, stat) { if (error) { self.logger.error("Got error when saving the framework ID on ZK: " + error.stack); self.options.useZk = false; self.zkClient.close(); } else { self.logger.debug("Successfully set framework ID") } }); } }); }); // We're ready to subscribe // Timeout to let init finish setTimeout(function() { self.emit("ready"); }, 100); } else { // Other error self.logger.error(error.stack); } } else if (data) { self.logger.debug("Got framework ID from ZooKeeper:" + data.toString()); // Set framework id to existing one from ZooKeeper self.frameworkId = data.toString(); // Load tasks from ZooKeeper self.taskHelper.loadTasks(); // "ready" event is emitted after successfully loading the tasks! } }); }); // Connect to ZooKeeper self.zkClient.connect(); } else { // Timeout to let init finish setTimeout(function() { self.emit("ready"); }, 100); } } // Inhertit from EventEmitter util.inherits(Scheduler, EventEmitter); /** * Subscribes the scheduler with the master to receive events. A scheduler must send other calls only after it has received the SUBCRIBED event. */ Scheduler.prototype.subscribe = function () { var self = this; /** * The handler funciton for the incoming Mesos master events for this framework. * @param {object} eventData - The data object for an incoming event. Contains the event details (type etc.). */ function handleEvent (eventData) { try { var event = JSON.parse(eventData); // Determine event handler, use custom one if it exists if (self.customEventHandlers[event.type]) { // Call custom handler self.customEventHandlers[event.type].call(self, event[event.type.toLocaleLowerCase()]); } else { // Call default handler schedulerHandlers[event.type].call(self, event[event.type.toLocaleLowerCase()]); } // Emit events per type if (event.type === "SUBSCRIBED") { // Set frameworkId self.frameworkId = event[event.type.toLocaleLowerCase()].framework_id.value; // TODO: Check! self.sync(); self.subscribeBackoffTime = 0; // Emit with usable object details self.emit("subscribed", { frameworkId: event[event.type.toLocaleLowerCase()].framework_id.value, mesosStreamId: self.mesosStreamId }); } else if (event.type === "HEARTBEAT") { // Set lastHeartbeat timestamp self.lastHeartbeat = new Date().getTime(); // Emit with current timestamp self.emit(event.type.toLocaleLowerCase(), self.lastHeartbeat); } else if (event.type === "MESSAGE") { // Emit with usable message object (parsed to ascii) self.emit("message", { agentId: event[event.type.toLocaleLowerCase()].agent_id, executorId: event[event.type.toLocaleLowerCase()].executor_id, data: new Buffer(event[event.type.toLocaleLowerCase()].data, "base64").toString("ascii") }); } else if (event.type === "ERROR") { // Emit an actual error object to identify errors from mesos master self.emit("error", new Error("Error received from Mesos master: " + event[event.type.toLocaleLowerCase()].message)); } else { // Emit original objects for all other types self.emit(event.type.toLocaleLowerCase(), event[event.type.toLocaleLowerCase()]); } } catch (error) { self.emit("error", { message: "Couldn't parse as JSON: " + eventData, stack: (error.stack || "") }); } } function handleRedirect(location) { // Redirection to another Master received self.logger.info("SUBSCRIBE: Redirect Location: " + location); // Derive the leader info var leaderInfo = location.replace(/\/\//g, "").split(":"); // Check for scheme and move window accordingly var schemeIndex = leaderInfo.length > 2 ? 0 : -1; // Set new leading master info self.options.masterUrl = leaderInfo[schemeIndex + 1]; // If the port part contains slashes -> URLs, then fiix it by just getting the port if (leaderInfo[schemeIndex + 2].indexOf("\/") > -1) { var temp = leaderInfo[schemeIndex + 2].split("/"); self.options.port = temp[0]; } else { self.options.port = leaderInfo[schemeIndex + 2]; } self.logger.info("SUBSCRIBE: Leader info: " + self.options.masterUrl + ":" + self.options.port); // Fill the requestTemplate self.generateRequestTemplate(); } var handledTimeout = false; self.httpRequest = http.request(self.requestTemplate, function (res) { self.logger.info("SUBSCRIBE: Response status: " + res.statusCode); if (res.statusCode === 307 && res.headers["location"]) { // Handle redirect information handleRedirect(res.headers["location"]); // Try to re-register self.subscribe(); } else if (res.statusCode === 200) { if (!res.headers["mesos-stream-id"]) { self.emit("error", { message: "Mesos-Stream-Id header field was not found!"}) } else { // Set mesosStreamId self.mesosStreamId = res.headers["mesos-stream-id"]; // Set encoding to UTF8 res.setEncoding('utf8'); // Emit sent_subscribe event self.emit("sent_subscribe", { mesosStreamId: self.mesosStreamId }); // Local cache for chunked JSON messages var cache = ""; var expectedLength = 0; // Watch for data/chunks res.on('data', function (chunk) { if (chunk instanceof Buffer) { chunk = chunk.toString(); } if (chunk.indexOf("\n") > -1) { var temp = chunk.split("\n"); if (temp.length === 2) { expectedLength = parseInt(temp[0]); if (temp[1].length < expectedLength) { // Add to cache cache += temp[1]; } else { // Empty cache cache = ""; expectedLength = 0; // Handle event handleEvent(temp[1]); } } else { self.emit("error", { message: "Other linebreak count found than expected! Actual count: " + temp.length }); } } else { if (cache.length > 0 && (cache.length + chunk.length) >= expectedLength) { // Concatenate cached partial data with this chunk and handle only when done var eventData = cache + chunk; // Handle event handleEvent(eventData); // Empty cache cache = ""; expectedLength = 0; } else if (cache.length > 0 && (cache.length + chunk.length) < expectedLength) { // Concatenate cached data with current chunk, for cases in which the stream buffer is smaller than the data. cache += chunk; } } }); res.on('end', function () { self.emit("error", { message: "Long-running connection was closed!" }); self.logger.info("Long-running connection was closed!"); if (!handledTimeout) { self.backOff(); // Re-subscribe // We need to remove the stream id from the headers before re-subscribing! self.mesosStreamId = undefined; delete self.requestTemplate.headers["mesos-stream-id"]; delete self.requestTemplate.headers["Mesos-Stream-Id"]; self.subscribe(); } }); res.on('finish', function () { self.logger.info("FINISH!"); }); res.on('close', function () { self.logger.info("CLOSE!"); }); } } else { res.on("data",function (chunk) { if (chunk.length > 0) { self.emit("error", {message: "Error registering with mesos: " + chunk.toString() + " , code: " + res.statusCode.toString()}); } else { self.emit("error", {message: "Error registering with mesos - empty response, code: " + res.statusCode.toString()}); } }); } }); self.httpRequest.on('error', function (e) { self.emit("error", { message: "There was a problem with the request: " + (e.message ? e.message : JSON.stringify(e)) }); }); // Register a timeout for triggering of re-registrations of the scheduler self.httpRequest.on('socket', function (socket) { var httpRequest = self.httpRequest; socket.setTimeout(self.options.masterConnectionTimeout); socket.on('timeout', function() { self.logger.error("Received a timeout on the long-running Master connection! Will try to re-register the framework scheduler!"); handledTimeout = true; socket.destroy(); // Make sure the timeout is not re-emitted. socket.setTimeout(0); if (httpRequest !== self.httpRequest) { self.logger.info("Already reconnected, not attempting again."); return; } // Backing off before resubscribe self.backOff(); // If we're using Mesos DNS, we can directy re-register, because Mesos DNS will discover the current leader automatically if (self.options.masterUrl === "leader.mesos") { // We need to remove the stream id from the headers before re-subscribing! self.mesosStreamId = undefined; delete self.requestTemplate.headers["mesos-stream-id"]; delete self.requestTemplate.headers["Mesos-Stream-Id"]; self.logger.info("Using Mesos DNS, will re-register to 'leader.mesos'!"); // Subscribe self.subscribe(); } else { self.logger.info("Not using Mesos DNS, try to get new leader through redirection!"); // If not, it's more difficult. When a IP address is passed for the Master, and the Master is unavailable, // we cannot use the Master detection via location headers, as outlined at http://mesos.apache.org/documentation/latest/scheduler-http-api/ (chapter "Master detection"), // because the request will not be successful. So, first we'll try the redirect method (in case of a leader change), if that is not possible, we have to shut down our framework // unless there is a better way in the future. var redirectRequest = http.request(self.requestTemplate, function (res) { // Check if we received a redirect if (res.statusCode === 307 && res.headers["location"]) { self.logger.info("Received redirection information. Will attempt to re-register the framework scheduler!"); // Handle redirect information handleRedirect(res.headers["location"]); // Subscribe self.subscribe(); } }); // Set timeout for redirect request. When it's triggered, we know that the last leading master is down and that we cannot get the current leader information from it. // So, we have to shutdown the framework scheduler, because we're out of options. redirectRequest.on('socket', function (socket) { socket.setTimeout(self.options.masterConnectionTimeout); socket.on('timeout', function() { self.logger.error("Couldn't receive a response for the redirect request from the last leading master!"); self.logger.error("There's no way to recover, the framework scheduler will halt now!"); process.exit(1); }); }); } }); }); // Set the Subscribe object var Subscribe = new Builder("mesos.scheduler.Call.Subscribe") .setFrameworkInfo(new Builder("mesos.FrameworkInfo") .setUser(self.options.user) .setRole(self.options.role) .setName(self.options.frameworkName) .setId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setFailoverTimeout(self.options.frameworkFailoverTimeout) .setHostname(process.env.HOST ? process.env.HOST : null) .setWebuiUrl(process.env.HOST && process.env.PORT0 ? "http://" + process.env.HOST + ":" + process.env.PORT0 : null) ); self.logger.info("SUBSCRIBE: " + JSON.stringify(Subscribe)); // Set the Call object var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.SUBSCRIBE) .setSubscribe(Subscribe) ); setTimeout(function () { // Write data to request body self.httpRequest.write(JSON.stringify(Call)); // End request self.httpRequest.end(); }, self.subscribeBackoffTime); }; /** * Accept incoming offers to actually start the framework scheduler. * @param {array} offersIds - The array of {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L49|OfferID}s which should be accepted. * @param {array} operations - The array of {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L1903|Operation} objects. * @param {object} filters - The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L2488|Filters} object. */ Scheduler.prototype.accept = function (offersIds, operations, filters) { var self = this; // Set the Accept object var Accept = new Builder("mesos.scheduler.Call.Accept") .setOfferIds(offersIds) .setOperations(operations) .setFilters(filters); self.logger.info("ACCEPT: " + JSON.stringify(Accept)); // Set the Call object var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(new mesos.FrameworkID(self.frameworkId)) .setType(mesos.scheduler.Call.Type.ACCEPT) .setAccept(Accept) ); self.logger.debug("Assembled ACCEPT call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_accept"); } }); }; /** * Decline incoming offers because they are not needed by the framework scheduler currently. * @param {array} offersIds - The array of {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L49|OfferID}s which should be declined. * @param {object} filters - The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L2488|Filters} object. */ Scheduler.prototype.decline = function (offersIds, filters) { var self = this; // Set the Decline object var Decline = new Builder("mesos.scheduler.Call.Decline") .setOfferIds(offersIds) .setFilters(filters); // Set the Call object var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(new mesos.FrameworkID(self.frameworkId)) .setType(mesos.scheduler.Call.Type.DECLINE) .setDecline(Decline) ); self.logger.debug("Assembled DECLINE call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_decline"); } }); }; /** * Tear down the framework scheduler. When Mesos receives this request it will shut down all executors (and consequently kill tasks). * It then removes the framework and closes all open connections from this scheduler to the Master. */ Scheduler.prototype.teardown = function () { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.TEARDOWN) ); self.logger.debug("Assembled TEARDOWN call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_teardown"); } }); }; /** * Remove any/all filters that it has previously set via ACCEPT or DECLINE calls. */ Scheduler.prototype.revive = function () { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.REVIVE) ); self.logger.debug("Assembled REVIVE call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_revive"); } }); }; /** * Kill a specific task. If the scheduler has a custom executor, the kill is forwarded to the executor; it is up to the executor to kill the task and send a TASK_KILLED (or TASK_FAILED) update. * Mesos releases the resources for a task once it receives a terminal update for the task. If the task is unknown to the master, a TASK_LOST will be generated. * @param {Object} taskId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L71|TaskID} to kill. * @param {Object} agentId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L59|AgentID} the task is running on. */ Scheduler.prototype.kill = function (taskId, agentId) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.KILL) .setKill(new Builder("mesos.scheduler.Call.Kill") .setTaskId(new Builder("mesos.TaskID").setValue(taskId)) .setAgentId(new Builder("mesos.AgentID").setValue(agentId)) ) ); self.logger.debug("Assembled KILL call: " + JSON.stringify(Call)); self.logger.debug("Killing task ID: " + taskId); if (self.options.useZk) self.taskHelper.deleteTask(taskId); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_kill"); } }); }; /** * shutdown a specific custom executor (NOTE: This is a new call that was not present in the old API). When an executor gets a shutdown event, it is expected to kill all its tasks (and send TASK_KILLED updates) and terminate. * If an executor doesn’t terminate within a certain timeout (configurable via “–executor_shutdown_grace_period” agent flag), the agent will forcefully destroy the container (executor and its tasks) and transitions its active tasks to TASK_LOST. * @param {string} agentId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L59|AgentID} the task is running on. * @param {string} executorId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L81|ExecutorID} whcih runs the task. */ Scheduler.prototype.shutdown = function (agentId, executorId) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.SHUTDOWN) .setShutdown(new Builder("mesos.scheduler.Call.Shutdown") .setExecutorId(new Builder("mesos.ExecutorID").setValue(executorId)) .setAgentId(new Builder("mesos.AgentID").setValue(agentId)) ) ); self.logger.debug("Assembled SHUTDOWN call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_shutdown"); } }); }; /** * Acknowledge a status update. * @param {object} update The status update to acknowledge. */ Scheduler.prototype.acknowledge = function (update) { var self = this; if (!update.status.uuid) { self.logger.debug("An update without UUID received, acknowledge skipped. Update: " + JSON.stringify(update)); return; } var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.ACKNOWLEDGE) .setAcknowledge(new Builder("mesos.scheduler.Call.Acknowledge") .setTaskId(new Builder("mesos.TaskID").setValue(update.status.task_id)) .setAgentId(new Builder("mesos.AgentID").setValue(update.status.agent_id)) .setUuid(update.status.uuid) ) ); self.logger.debug("Assembled ACKNOWLEDGE call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_acknowledge"); } }); }; /** * query the status of non-terminal tasks. This causes the master to send back UPDATE events for each task in the list. Tasks that are no longer known to Mesos will result in TASK_LOST updates. * If the list of tasks is empty, master will send UPDATE events for all currently known tasks of the framework. * @param {string} taskId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L71|TaskID} to kill. * @param {string} agentId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L59|AgentID} the task is running on. */ Scheduler.prototype.reconcile = function (taskId, agentId) { var self = this; var Call = null; if (taskId && agentId) { Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.RECONCILE) .setReconcile(new Builder("mesos.scheduler.Call.Reconcile") .setTasks(new Builder("mesos.scheduler.Call.Reconcile.Task") .setTaskId(new Builder("mesos.TaskID").setValue(taskId)) .setAgentId(new Builder("mesos.AgentID").setValue(agentId) ) ) ) ); self.logger.debug("Reconciling task ID: " + taskId); } else { Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.RECONCILE) .setReconcile(new Builder("mesos.scheduler.Call.Reconcile") .setTasks([]) ) ); self.logger.debug("Reconciling all tasks "); } self.logger.debug("Assembled RECONCILE call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_reconcile"); } }); }; /** * Send arbitrary data to the executor. Note that Mesos neither interprets this data nor makes any guarantees about the delivery of this message to the executor. * @param {string} agentId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L59|AgentID} the task is running on. * @param {string} executorId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L81|ExecutorID} which runs the task. * @param {string} data The string which's raw bytes will be encoded in Base64. */ Scheduler.prototype.message = function (agentId, executorId, data) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.MESSAGE) .setMessage(new Builder("mesos.scheduler.Call.Message") .setExecutorId(new Builder("mesos.ExecutorID").setValue(executorId)) .setAgentId(new Builder("mesos.AgentID").setValue(agentId)) .setData(new Buffer(data).toString("base64")) ) ); self.logger.debug("Assembled MESSAGE call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_message"); } }); }; /** * Request resources from the master/allocator. The built-in hierarchical allocator simply ignores this request but other allocators (modules) can interpret this in a customizable fashion. * @param {array} requests The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L1858|Request} objects which should be sent to the server. */ Scheduler.prototype.request = function (requests) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.REQUEST) .setRequest(new Builder("mesos.scheduler.Call.Request") .setRequests(requests) ) ); self.logger.debug("Assembled REQUEST call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_request"); } }); }; /** * Suppress offers for the specified roles. If `roles` is empty, the `SUPPRESS` call will suppress offers for all of the roles the framework is currently subscribed to. * @param {array} roles The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L3390|Role} objects which should be sent to the server. */ Scheduler.prototype.suppress = function (roles) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.SUPPRESS) .setSuppress(new Builder("mesos.scheduler.Call.Suppress") .setRoles(roles) ) ); self.logger.debug("Assembled SUPPRESS call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_suppress"); } }); }; /** * Accepts an inverse offer. Inverse offers should only be accepted if the resources in the offer can be safely evacuated before the provided unavailability. * @param {array} inverseOffersIds The {@Link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L49|OfferID} array which should be sent to the server. * @param {object} filters The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L2488|Filters} object which should be sent to the server. */ Scheduler.prototype.acceptInverseOffers = function (inverseOffersIds, filters) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.ACCEPT_INVERSE_OFFERS) .setAcceptInverseOffers(new Builder("mesos.scheduler.Call.AcceptInverseOffers") .setInverseOfferIds(inverseOffersIds) .setFilter(filters) ) ); self.logger.debug("Assembled ACCEPT_INVERSE_OFFERS call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_accept_inverse_offers"); } }); }; /** * Declines an inverse offer. Inverse offers should be declined if the resources in the offer might not be safely evacuated before the provided unavailability. * @param {array} inverseOffersIds The {@Link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L49|OfferID} array which should be sent to the server. * @param {object} filters The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L2488|Filters} object which should be sent to the server. */ Scheduler.prototype.declineInverseOffers = function (inverseOffersIds, filters) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.DECLINE_INVERSE_OFFERS) .setDeclineInverseOffers(new Builder("mesos.scheduler.Call.DeclineInverseOffers") .setInverseOfferIds(inverseOffersIds) .setFilter(filters) ) ); self.logger.debug("Assembled DECLINE_INVERSE_OFFERS call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_decline_inverse_offers"); } }); }; /** * Acknowledges the receipt of an operation status update. Schedulers are responsible for explicitly acknowledging the receipt of updates which have the 'UpdateOperationStatus.status().uuid()' field set. Such status updates are retried by the agent or resource provider until they are acknowledged by the scheduler. * @param {string} agentId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L59|AgentID} the task is running on. * @param {string} resourceProviderId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L102|ResourceProviderId} * @param {string} operationId The {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/mesos.proto#L111|OperationID} * @param {string} uuid The uuid of the Operation */ Scheduler.prototype.acknowledgeOperationStatus = function (agentId, resourceProviderId, operationId, uuid) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.ACKNOWLEDGE_OPERATION_STATUS) .setAcknowledgeOperationStatus(new Builder("mesos.scheduler.Call.AcknowledgeOperationStatus") .setAgentId(new Builder("mesos.AgentID").setValue(agentId)) .setResourceProviderId(new Builder("mesos.ResourceProviderId").setValue(resourceProviderId)) .setOperationId(new Builder("mesos.ResourceProviderId").setValue(operationId)) .setUuid(uuid) ) ); self.logger.debug("Assembled ACKNOWLEDGE_OPERATION_STATUS call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_acknowledge_operation_status"); } }); }; /** * Allows the scheduler to query the status of operations. This causes the master to send back the latest status for each operation in 'operations', if possible. If 'operations' is empty, then the master will send the latest status for each operation currently known. * @param {array} operations An array of {@link https://github.com/apache/mesos/blob/1.5.x/include/mesos/v1/scheduler/scheduler.proto#L420|Operation} objects to query. */ Scheduler.prototype.reconcileOperations = function (operations) { var self = this; var Call = helpers.fixEnums(new Builder("mesos.scheduler.Call") .setFrameworkId(self.frameworkId ? new mesos.FrameworkID(self.frameworkId) : null) .setType(mesos.scheduler.Call.Type.RECONCILE_OPERATIONS) .setReconcileOperations(new Builder("mesos.scheduler.Call.ReconcileOperations") .setOperations(operations) ) ); self.logger.debug("Assembled RECONCILE_OPERATIONS call: " + JSON.stringify(Call)); helpers.doRequest.call(self, Call, function (error, response) { if (error) { self.emit("error", error.message); } else { self.emit("sent_reconcile_operations"); } }); }; /** * Get the running tasks of this framework scheduler. * @returns {Array} The running task array. */ Scheduler.prototype.getRunningTasks = function () { var self = this; var runningTasks = []; Object.getOwnPropertyNames(self.runtimeInfo).forEach(function (taskType) { Object.getOwnPropertyNames(self.runtimeInfo[taskType].runningInstances).forEach(function (task) { runningTasks.push(task); }); }); return runningTasks; }; /** * Synchronize the tasks of this scheduler. */ Scheduler.prototype.sync = function () { var self = this; for (var i = 0; i < self.reconcileTasks.length; i++) { if (self.reconcileTasks[i].runtimeInfo.agentId) { self.reconcile(self.reconcileTasks[i].taskId, self.reconcileTasks[i].runtimeInfo.agentId); } else { if (self.options.useZk) self.taskHelper.deleteTask(self.reconcileTasks[i].taskId); } } self.reconcileTasks = []; self.launchedTasks.forEach(function (task) { if (task.runtimeInfo.agentId) { self.reconcile(task.taskId, task.runtimeInfo.agentId); } }); for (var i = 0; i < self.killTasks.length; i++) { self.kill(self.killTasks[i].taskId, self.killTasks[i].runtimeInfo.agentId); } self.killTasks = []; if (self.options.useZk) { self.reconcile(); } }; /** * Calculate the backOff time (for reconnection trials) */ Scheduler.prototype.backOff = function () { var self = this; self.subscribeBackoffTime *= self.options.exponentialBackoffFactor; self.subscribeBackoffTime = Math.round(self.subscribeBackoffTime); if (self.subscribeBackoffTime === 0) { self.subscribeBackoffTime = self.options.exponentialBackoffMinimum; } if (self.subscribeBackoffTime > self.options.exponentialBackoffMaximum) { self.subscribeBackoffTime = self.options.exponentialBackoffMaximum; } self.logger.debug("Backoff time: " + self.subscribeBackoffTime); }; module.exports = Scheduler;