The image API no longer uses a unique thread for each job (due to image processing using its own thread anyway), added dummy DB driver
This commit is contained in:
parent
7b64c4ca2a
commit
55da0db479
4 changed files with 151 additions and 134 deletions
104
api/index.js
104
api/index.js
|
@ -2,7 +2,6 @@
|
|||
|
||||
require("dotenv").config();
|
||||
const os = require("os");
|
||||
const { Worker, isMainThread, parentPort } = require("worker_threads");
|
||||
const magick = require("../utils/image.js");
|
||||
const execPromise = require("util").promisify(require("child_process").exec);
|
||||
const net = require("net");
|
||||
|
@ -10,8 +9,8 @@ const dgram = require("dgram"); // for UDP servers
|
|||
const socket = dgram.createSocket("udp4"); // Our universal UDP socket, this might cause issues and we may have to use a seperate socket for each connection
|
||||
|
||||
const start = process.hrtime();
|
||||
const log = (msg, thread) => {
|
||||
console.log(`[${process.hrtime(start)[1] / 1000000}${(thread)?`:${thread}`:""}]\t ${msg}`);
|
||||
const log = (msg) => {
|
||||
console.log(`[${process.hrtime(start)[1] / 1000000}]\t ${msg}`);
|
||||
};
|
||||
|
||||
const jobs = {};
|
||||
|
@ -19,11 +18,10 @@ const jobs = {};
|
|||
const queue = [];
|
||||
// Array of UUIDs
|
||||
|
||||
if (isMainThread) {
|
||||
const { v4: uuidv4 } = require("uuid");
|
||||
let cpuLoad = 0;
|
||||
const { v4: uuidv4 } = require("uuid");
|
||||
let cpuLoad = 0;
|
||||
|
||||
const getAverage = () => {
|
||||
const getAverage = () => {
|
||||
const cpus = os.cpus();
|
||||
let idle = 0;
|
||||
let tick = 0;
|
||||
|
@ -37,61 +35,56 @@ if (isMainThread) {
|
|||
idle: idle / cpus.length,
|
||||
tick: tick / cpus.length
|
||||
};
|
||||
};
|
||||
};
|
||||
|
||||
let measure = getAverage();
|
||||
setInterval(() => {
|
||||
let measure = getAverage();
|
||||
setInterval(() => {
|
||||
const newMeasure = getAverage();
|
||||
const idleDiff = newMeasure.idle - measure.idle;
|
||||
const tickDiff = newMeasure.tick - measure.tick;
|
||||
cpuLoad = 100 - ~~(100 * idleDiff / tickDiff);
|
||||
measure = newMeasure;
|
||||
}, 5000);
|
||||
}, 5000);
|
||||
|
||||
const MAX_WORKERS = process.env.WORKERS !== "" ? parseInt(process.env.WORKERS) : os.cpus().length * 4; // Completely arbitrary, should usually be some multiple of your amount of cores
|
||||
let workingWorkers = 0;
|
||||
const MAX_JOBS = process.env.JOBS !== "" && process.env.JOBS !== undefined ? parseInt(process.env.JOBS) : os.cpus().length * 4; // Completely arbitrary, should usually be some multiple of your amount of cores
|
||||
let jobAmount = 0;
|
||||
|
||||
const acceptJob = (uuid) => {
|
||||
workingWorkers++;
|
||||
const acceptJob = async (uuid) => {
|
||||
jobAmount++;
|
||||
queue.shift();
|
||||
const worker = new Worker(__filename);
|
||||
log(`Spawned worker ${uuid}`);
|
||||
worker.once("error", err => {
|
||||
console.error(`Error on worker ${uuid}:`, err);
|
||||
socket.send(Buffer.concat([Buffer.from([0x2]), Buffer.from(uuid), Buffer.from(err.toString())]), jobs[uuid].port, jobs[uuid].addr);
|
||||
});
|
||||
worker.once("exit", (code) => {
|
||||
workingWorkers--;
|
||||
if (queue.length > 0) {
|
||||
acceptJob(queue[0]);
|
||||
delete jobs[uuid];
|
||||
}
|
||||
|
||||
if (code === 0) {
|
||||
log(`Worker ${uuid} successfully exited`);
|
||||
} else {
|
||||
console.error(`Worker ${uuid} stopped with exit code ${code}`);
|
||||
}
|
||||
});
|
||||
|
||||
worker.postMessage({
|
||||
try {
|
||||
await runJob({
|
||||
uuid: uuid,
|
||||
msg: jobs[uuid].msg,
|
||||
addr: jobs[uuid].addr,
|
||||
port: jobs[uuid].port,
|
||||
threadNum: workingWorkers
|
||||
port: jobs[uuid].port
|
||||
});
|
||||
};
|
||||
jobAmount--;
|
||||
if (queue.length > 0) {
|
||||
acceptJob(queue[0]);
|
||||
}
|
||||
delete jobs[uuid];
|
||||
log(`Job ${uuid} has finished`);
|
||||
} catch (err) {
|
||||
console.error(`Error on job ${uuid}:`, err);
|
||||
socket.send(Buffer.concat([Buffer.from([0x2]), Buffer.from(uuid), Buffer.from(err.toString())]), jobs[uuid].port, jobs[uuid].addr);
|
||||
jobAmount--;
|
||||
if (queue.length > 0) {
|
||||
acceptJob(queue[0]);
|
||||
}
|
||||
delete jobs[uuid];
|
||||
}
|
||||
};
|
||||
|
||||
const server = dgram.createSocket("udp4"); //Create a UDP server for listening to requests, we dont need tcp
|
||||
server.on("message", (msg, rinfo) => {
|
||||
const server = dgram.createSocket("udp4"); //Create a UDP server for listening to requests, we dont need tcp
|
||||
server.on("message", (msg, rinfo) => {
|
||||
const opcode = msg.readUint8(0);
|
||||
const req = msg.toString().slice(1,msg.length);
|
||||
// 0x0 == Cancel job
|
||||
// 0x1 == Queue job
|
||||
// 0x2 == Get CPU usage
|
||||
if (opcode == 0x0) {
|
||||
queue.shift();
|
||||
delete queue[queue.indexOf(req) - 1];
|
||||
delete jobs[req];
|
||||
} else if (opcode == 0x1) {
|
||||
const job = { addr: rinfo.address, port: rinfo.port, msg: req };
|
||||
|
@ -99,7 +92,7 @@ if (isMainThread) {
|
|||
jobs[uuid] = job;
|
||||
queue.push(uuid);
|
||||
|
||||
if (workingWorkers < MAX_WORKERS) {
|
||||
if (jobAmount < MAX_JOBS) {
|
||||
log(`Got request for job ${job.msg} with id ${uuid}`, job.threadNum);
|
||||
acceptJob(uuid);
|
||||
} else {
|
||||
|
@ -113,17 +106,17 @@ if (isMainThread) {
|
|||
} else {
|
||||
log("Could not parse message");
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
server.on("listening", () => {
|
||||
server.on("listening", () => {
|
||||
const address = server.address();
|
||||
log(`server listening ${address.address}:${address.port}`);
|
||||
});
|
||||
});
|
||||
|
||||
server.bind(8080); // ATTENTION: Always going to be bound to 0.0.0.0 !!!
|
||||
} else {
|
||||
parentPort.once("message", async (job) => {
|
||||
log(`${job.uuid} worker started`, job.threadNum);
|
||||
server.bind(8080); // ATTENTION: Always going to be bound to 0.0.0.0 !!!
|
||||
|
||||
const runJob = async (job) => {
|
||||
log(`Job ${job.uuid} started`, job.threadNum);
|
||||
|
||||
const object = JSON.parse(job.msg);
|
||||
let type;
|
||||
|
@ -147,7 +140,7 @@ if (isMainThread) {
|
|||
|
||||
const data = await magick.run(object, true);
|
||||
|
||||
log(`${job.uuid} is done`, job.threadNum);
|
||||
log(`Sending result of job ${job.uuid} back to the bot`, job.threadNum);
|
||||
const server = net.createServer(function(tcpSocket) {
|
||||
tcpSocket.write(Buffer.concat([Buffer.from(type ? type : "image/png"), Buffer.from("\n"), data]), (err) => {
|
||||
if (err) console.error(err);
|
||||
|
@ -160,15 +153,12 @@ if (isMainThread) {
|
|||
// handle address in use errors
|
||||
server.on("error", (e) => {
|
||||
if (e.code === "EADDRINUSE") {
|
||||
console.log("Address in use, retrying...");
|
||||
log("Address in use, retrying...", job.threadNum);
|
||||
setTimeout(() => {
|
||||
server.close();
|
||||
server.listen(job.port, job.addr);
|
||||
}, 500);
|
||||
}
|
||||
});
|
||||
socket.send(Buffer.concat([Buffer.from([0x1]), Buffer.from(job.uuid), Buffer.from(job.port.toString())]), job.port, job.addr, () => {
|
||||
socket.close();
|
||||
});
|
||||
});
|
||||
}
|
||||
socket.send(Buffer.concat([Buffer.from([0x1]), Buffer.from(job.uuid), Buffer.from(job.port.toString())]), job.port, job.addr);
|
||||
};
|
|
@ -116,5 +116,6 @@
|
|||
"Open Fortress",
|
||||
"btw I use arch",
|
||||
"Friday Night Funkin'",
|
||||
"fgsfds"
|
||||
"fgsfds",
|
||||
"Doilus Stage"
|
||||
]
|
26
utils/database/dummy.js
Normal file
26
utils/database/dummy.js
Normal file
|
@ -0,0 +1,26 @@
|
|||
// dummy (no-op) database handler
|
||||
const misc = require("../misc.js");
|
||||
|
||||
exports.setup = async () => {};
|
||||
exports.stop = async () => {};
|
||||
exports.fixGuild = async () => {};
|
||||
exports.addCount = async () => {};
|
||||
exports.getCounts = async () => {
|
||||
return {};
|
||||
};
|
||||
exports.disableChannel = async () => {};
|
||||
exports.enableChannel = async () => {};
|
||||
exports.toggleTags = async () => {};
|
||||
exports.setTag = async () => {};
|
||||
exports.removeTag = async () => {};
|
||||
exports.setPrefix = async () => {};
|
||||
exports.addGuild = async (guild) => {
|
||||
return {
|
||||
id: guild.id,
|
||||
tags: misc.tagDefaults,
|
||||
prefix: process.env.PREFIX,
|
||||
disabled: [],
|
||||
tagsDisabled: false,
|
||||
};
|
||||
};
|
||||
exports.getGuild = exports.addGuild;
|
|
@ -66,7 +66,7 @@ const checkImages = async (message) => {
|
|||
if (message.embeds[0].type === "gifv") {
|
||||
type = await getImage(message.embeds[0].video.url, message.embeds[0].url, true);
|
||||
// then we check for other image types
|
||||
} else if (message.embeds[0].type === "video" || message.embeds[0].type === "image") {
|
||||
} else if ((message.embeds[0].type === "video" || message.embeds[0].type === "image") && message.embeds[0].thumbnail) {
|
||||
type = await getImage(message.embeds[0].thumbnail.proxy_url, message.embeds[0].thumbnail.url);
|
||||
// finally we check both possible image fields for "generic" embeds
|
||||
} else if (message.embeds[0].type === "rich") {
|
||||
|
|
Loading…
Reference in a new issue