This commit is contained in:
Thomas Hobson 2021-05-08 12:20:21 +12:00
parent 2beb0abff7
commit 4259e89bb2
No known key found for this signature in database
GPG Key ID: 9F1FD9D87950DB6F
11 changed files with 1869 additions and 1857 deletions

1
api/.prettierignore Normal file
View File

@ -0,0 +1 @@
node_modules

1
api/.prettierrc.yaml Normal file
View File

@ -0,0 +1 @@
singleQuote: true

21
api/package-lock.json generated
View File

@ -19,6 +19,9 @@
"semver": "^7.3.4",
"uuid": "^8.3.2",
"waitpid": "git+https://github.com/HexF/node-waitpid.git"
},
"devDependencies": {
"prettier": "2.2.1"
}
},
"node_modules/accepts": {
@ -391,6 +394,18 @@
"resolved": "https://registry.npmjs.org/path-to-regexp/-/path-to-regexp-0.1.7.tgz",
"integrity": "sha1-32BBeABfUi8V60SQ5yR6G/qmf4w="
},
"node_modules/prettier": {
"version": "2.2.1",
"resolved": "https://registry.npmjs.org/prettier/-/prettier-2.2.1.tgz",
"integrity": "sha512-PqyhM2yCjg/oKkFPtTGUojv7gnZAoG80ttl45O6x2Ug/rMJw4wcc9k6aaf2hibP7BGVCCM33gZoGjyvt9mm16Q==",
"dev": true,
"bin": {
"prettier": "bin-prettier.js"
},
"engines": {
"node": ">=10.13.0"
}
},
"node_modules/proxy-addr": {
"version": "2.0.6",
"resolved": "https://registry.npmjs.org/proxy-addr/-/proxy-addr-2.0.6.tgz",
@ -855,6 +870,12 @@
"resolved": "https://registry.npmjs.org/path-to-regexp/-/path-to-regexp-0.1.7.tgz",
"integrity": "sha1-32BBeABfUi8V60SQ5yR6G/qmf4w="
},
"prettier": {
"version": "2.2.1",
"resolved": "https://registry.npmjs.org/prettier/-/prettier-2.2.1.tgz",
"integrity": "sha512-PqyhM2yCjg/oKkFPtTGUojv7gnZAoG80ttl45O6x2Ug/rMJw4wcc9k6aaf2hibP7BGVCCM33gZoGjyvt9mm16Q==",
"dev": true
},
"proxy-addr": {
"version": "2.0.6",
"resolved": "https://registry.npmjs.org/proxy-addr/-/proxy-addr-2.0.6.tgz",

View File

@ -15,5 +15,11 @@
"uuid": "^8.3.2",
"waitpid": "git+https://github.com/HexF/node-waitpid.git"
},
"license": "MIT"
"license": "MIT",
"scripts": {
"lint": "prettier . --write"
},
"devDependencies": {
"prettier": "2.2.1"
}
}

View File

@ -4,7 +4,7 @@ const router = express.Router();
const config = require('../config');
const runtime = require('../runtime');
const { Job } = require('../job');
const package = require('../package')
const package = require('../package');
const logger = require('logplease').create('api/v2');
router.use((req, res, next) => {
@ -13,10 +13,8 @@ router.use((req, res, next) => {
}
if (req.headers['content-type'] !== 'application/json') {
return res
.status(415)
.send({
message: 'requests must be of type application/json'
return res.status(415).send({
message: 'requests must be of type application/json',
});
}
@ -25,90 +23,90 @@ router.use((req, res, next) => {
router.post('/execute', async (req, res) => {
const {
language, version,
language,
version,
files,
stdin, args,
run_timeout, compile_timeout,
compile_memory_limit, run_memory_limit
stdin,
args,
run_timeout,
compile_timeout,
compile_memory_limit,
run_memory_limit,
} = req.body;
if (!language || typeof language !== 'string') {
return res
.status(400)
.send({
message: 'language is required as a string'
return res.status(400).send({
message: 'language is required as a string',
});
}
if (!version || typeof version !== 'string') {
return res
.status(400)
.send({
message: 'version is required as a string'
return res.status(400).send({
message: 'version is required as a string',
});
}
if (!files || !Array.isArray(files)) {
return res
.status(400)
.send({
message: 'files is required as an array'
return res.status(400).send({
message: 'files is required as an array',
});
}
for (const [i, file] of files.entries()) {
if (typeof file.content !== 'string') {
return res
.status(400)
.send({
message: `files[${i}].content is required as a string`
return res.status(400).send({
message: `files[${i}].content is required as a string`,
});
}
}
if (compile_memory_limit) {
if (typeof compile_memory_limit !== 'number') {
return res
.status(400)
.send({
message: 'if specified, compile_memory_limit must be a number'
})
return res.status(400).send({
message: 'if specified, compile_memory_limit must be a number',
});
}
if (config.compile_memory_limit >= 0 && (compile_memory_limit > config.compile_memory_limit || compile_memory_limit < 0)) {
return res
.status(400)
.send({
message: 'compile_memory_limit cannot exceed the configured limit of ' + config.compile_memory_limit
})
if (
config.compile_memory_limit >= 0 &&
(compile_memory_limit > config.compile_memory_limit ||
compile_memory_limit < 0)
) {
return res.status(400).send({
message:
'compile_memory_limit cannot exceed the configured limit of ' +
config.compile_memory_limit,
});
}
}
if (run_memory_limit) {
if (typeof run_memory_limit !== 'number') {
return res
.status(400)
.send({
message: 'if specified, run_memory_limit must be a number'
})
return res.status(400).send({
message: 'if specified, run_memory_limit must be a number',
});
}
if (config.run_memory_limit >= 0 && (run_memory_limit > config.run_memory_limit || run_memory_limit < 0)) {
return res
.status(400)
.send({
message: 'run_memory_limit cannot exceed the configured limit of ' + config.run_memory_limit
})
if (
config.run_memory_limit >= 0 &&
(run_memory_limit > config.run_memory_limit || run_memory_limit < 0)
) {
return res.status(400).send({
message:
'run_memory_limit cannot exceed the configured limit of ' +
config.run_memory_limit,
});
}
}
const rt = runtime.get_latest_runtime_matching_language_version(language, version);
const rt = runtime.get_latest_runtime_matching_language_version(
language,
version
);
if (rt === undefined) {
return res
.status(400)
.send({
message: `${language}-${version} runtime is unknown`
return res.status(400).send({
message: `${language}-${version} runtime is unknown`,
});
}
@ -120,12 +118,12 @@ router.post('/execute', async (req, res) => {
stdin: stdin || '',
timeouts: {
run: run_timeout || 3000,
compile: compile_timeout || 10000
compile: compile_timeout || 10000,
},
memory_limits: {
run: run_memory_limit || config.run_memory_limit,
compile: compile_memory_limit || config.compile_memory_limit
}
compile: compile_memory_limit || config.compile_memory_limit,
},
});
await job.prime();
@ -134,43 +132,35 @@ router.post('/execute', async (req, res) => {
await job.cleanup();
return res
.status(200)
.send(result);
return res.status(200).send(result);
});
router.get('/runtimes', (req, res) => {
const runtimes = runtime
.map(rt => {
const runtimes = runtime.map((rt) => {
return {
language: rt.language,
version: rt.version.raw,
aliases: rt.aliases,
runtime: rt.runtime
runtime: rt.runtime,
};
});
return res
.status(200)
.send(runtimes);
return res.status(200).send(runtimes);
});
router.get('/packages', async (req, res) => {
logger.debug('Request to list packages');
let packages = await package.get_package_list();
packages = packages
.map(pkg => {
packages = packages.map((pkg) => {
return {
language: pkg.language,
language_version: pkg.version.raw,
installed: pkg.installed
installed: pkg.installed,
};
});
return res
.status(200)
.send(packages);
return res.status(200).send(packages);
});
router.post('/packages/:language/:version', async (req, res) => {
@ -181,26 +171,23 @@ router.post('/packages/:language/:version', async (req, res) => {
const pkg = await package.get_package(language, version);
if (pkg == null) {
return res
.status(404)
.send({
message: `Requested package ${language}-${version} does not exist`
return res.status(404).send({
message: `Requested package ${language}-${version} does not exist`,
});
}
try {
const response = await pkg.install();
return res
.status(200)
.send(response);
return res.status(200).send(response);
} catch (e) {
logger.error(`Error while installing package ${pkg.language}-${pkg.version}:`, e.message);
logger.error(
`Error while installing package ${pkg.language}-${pkg.version}:`,
e.message
);
return res
.status(500)
.send({
message: e.message
return res.status(500).send({
message: e.message,
});
}
});
@ -208,31 +195,28 @@ router.post('/packages/:language/:version', async (req, res) => {
router.delete('/packages/:language/:version', async (req, res) => {
logger.debug('Request to uninstall package');
const {language, version} = req.params;
const { language, version } = req.params;
const pkg = await package.get_package(language, version);
if (pkg == null) {
return res
.status(404)
.send({
message: `Requested package ${language}-${version} does not exist`
return res.status(404).send({
message: `Requested package ${language}-${version} does not exist`,
});
}
try {
const response = await pkg.uninstall();
return res
.status(200)
.send(response);
return res.status(200).send(response);
} catch (e) {
logger.error(`Error while uninstalling package ${pkg.language}-${pkg.version}:`, e.message);
logger.error(
`Error while uninstalling package ${pkg.language}-${pkg.version}:`,
e.message
);
return res
.status(500)
.send({
message: e.message
return res.status(500).send({
message: e.message,
});
}
});

View File

@ -9,126 +9,108 @@ const options = [
default: 'INFO',
options: Object.values(Logger.LogLevels),
validators: [
x => Object.values(Logger.LogLevels).includes(x) || `Log level ${x} does not exist`
]
(x) =>
Object.values(Logger.LogLevels).includes(x) ||
`Log level ${x} does not exist`,
],
},
{
key: 'bind_address',
desc: 'Address to bind REST API on\nThank @Bones for the number',
default: '0.0.0.0:2000',
validators: []
validators: [],
},
{
key: 'data_directory',
desc: 'Absolute path to store all piston related data at',
default: '/piston',
validators: [x=> fss.exists_sync(x) || `Directory ${x} does not exist`]
validators: [(x) => fss.exists_sync(x) || `Directory ${x} does not exist`],
},
{
key: 'runner_uid_min',
desc: 'Minimum uid to use for runner',
default: 1001,
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'runner_uid_max',
desc: 'Maximum uid to use for runner',
default: 1500,
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'runner_gid_min',
desc: 'Minimum gid to use for runner',
default: 1001,
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'runner_gid_max',
desc: 'Maximum gid to use for runner',
default: 1500,
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'disable_networking',
desc: 'Set to true to disable networking',
default: true,
parser: x => x === "true",
validators: [
x => typeof x === "boolean" || `${x} is not a boolean`
]
parser: (x) => x === 'true',
validators: [(x) => typeof x === 'boolean' || `${x} is not a boolean`],
},
{
key: 'output_max_size',
desc: 'Max size of each stdio buffer',
default: 1024,
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'max_process_count',
desc: 'Max number of processes per job',
default: 64,
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'max_open_files',
desc: 'Max number of open files per job',
default: 2048,
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'max_file_size',
desc: 'Max file size in bytes for a file',
default: 10000000, //10MB
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'compile_memory_limit',
desc: 'Max memory usage for compile stage in bytes (set to -1 for no limit)',
desc:
'Max memory usage for compile stage in bytes (set to -1 for no limit)',
default: -1, // no limit
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'run_memory_limit',
desc: 'Max memory usage for run stage in bytes (set to -1 for no limit)',
default: -1, // no limit
parser: parse_int,
validators: [
(x, raw) => !is_nan(x) || `${raw} is not a number`,
]
validators: [(x, raw) => !is_nan(x) || `${raw} is not a number`],
},
{
key: 'repo_url',
desc: 'URL of repo index',
default: 'https://github.com/engineer-man/piston/releases/download/pkgs/index',
validators: []
}
default:
'https://github.com/engineer-man/piston/releases/download/pkgs/index',
validators: [],
},
];
logger.info(`Loading Configuration from environment`);
@ -137,10 +119,10 @@ let errored = false;
let config = {};
options.forEach(option => {
const env_key = "PISTON_" + option.key.to_upper_case();
options.forEach((option) => {
const env_key = 'PISTON_' + option.key.to_upper_case();
const parser = option.parser || (x => x);
const parser = option.parser || ((x) => x);
const env_val = process.env[env_key];
@ -148,12 +130,10 @@ options.forEach(option => {
const value = env_val || option.default;
option.validators.for_each(validator => {
option.validators.for_each((validator) => {
let response = null;
if(env_val)
response = validator(parsed_val, env_val);
else
response = validator(value, value);
if (env_val) response = validator(parsed_val, env_val);
else response = validator(value, value);
if (response !== true) {
errored = true;

View File

@ -1,26 +1,20 @@
// Globals are things the user shouldn't change in config, but is good to not use inline constants for
const is_docker = require('is-docker');
const fs = require('fs');
const platform = `${is_docker() ? 'docker' : 'baremetal'}-${
fs.read_file_sync('/etc/os-release')
const platform = `${is_docker() ? 'docker' : 'baremetal'}-${fs
.read_file_sync('/etc/os-release')
.toString()
.split('\n')
.find(x => x.startsWith('ID'))
.replace('ID=','')
}`;
.find((x) => x.startsWith('ID'))
.replace('ID=', '')}`;
module.exports = {
data_directories: {
packages: 'packages',
jobs: 'jobs'
jobs: 'jobs',
},
version: require('../package.json').version,
platform,
pkg_installed_file: '.ppman-installed', //Used as indication for if a package was installed
clean_directories: [
'/dev/shm',
'/run/lock',
'/tmp',
'/var/tmp'
]
clean_directories: ['/dev/shm', '/run/lock', '/tmp', '/var/tmp'],
};

View File

@ -14,11 +14,11 @@ const logger = Logger.create('index');
const app = express();
(async () => {
logger.info('Setting loglevel to',config.log_level);
logger.info('Setting loglevel to', config.log_level);
Logger.setLogLevel(config.log_level);
logger.debug('Ensuring data directories exist');
Object.values(globals.data_directories).for_each(dir => {
Object.values(globals.data_directories).for_each((dir) => {
let data_path = path.join(config.data_directory, dir);
logger.debug(`Ensuring ${data_path} exists`);
@ -28,32 +28,35 @@ const app = express();
try {
fss.mkdir_sync(data_path);
} catch(e) {
} catch (e) {
logger.error(`Failed to create ${data_path}: `, e.message);
}
}
});
logger.info('Loading packages');
const pkgdir = path.join(config.data_directory,globals.data_directories.packages);
const pkgdir = path.join(
config.data_directory,
globals.data_directories.packages
);
const pkglist = await fs.readdir(pkgdir);
const languages = await Promise.all(
pkglist.map(lang => {
return fs
.readdir(path.join(pkgdir,lang))
.then(x => {
return x.map(y => path.join(pkgdir, lang, y))
pkglist.map((lang) => {
return fs.readdir(path.join(pkgdir, lang)).then((x) => {
return x.map((y) => path.join(pkgdir, lang, y));
});
})
);
const installed_languages = languages
.flat()
.filter(pkg => fss.exists_sync(path.join(pkg, globals.pkg_installed_file)));
.filter((pkg) =>
fss.exists_sync(path.join(pkg, globals.pkg_installed_file))
);
installed_languages.for_each(pkg => runtime.load_package(pkg));
installed_languages.for_each((pkg) => runtime.load_package(pkg));
logger.info('Starting API Server');
logger.debug('Constructing Express App');
@ -63,27 +66,23 @@ const app = express();
app.use(body_parser.json());
app.use((err, req, res, next) => {
return res
.status(400)
.send({
stack: err.stack
return res.status(400).send({
stack: err.stack,
});
});
logger.debug('Registering Routes');
const api_v2 = require('./api/v2')
const api_v2 = require('./api/v2');
app.use('/api/v2', api_v2);
app.use('/api/v2', api_v2);
app.use((req, res, next) => {
return res
.status(404)
.send({message: 'Not Found'});
return res.status(404).send({ message: 'Not Found' });
});
logger.debug('Calling app.listen');
const [ address, port ] = config.bind_address.split(':');
const [address, port] = config.bind_address.split(':');
app.listen(port, address, () => {
logger.info('API server started on', config.bind_address);

View File

@ -10,20 +10,19 @@ const wait_pid = require('waitpid');
const job_states = {
READY: Symbol('Ready to be primed'),
PRIMED: Symbol('Primed and ready for execution'),
EXECUTED: Symbol('Executed and ready for cleanup')
EXECUTED: Symbol('Executed and ready for cleanup'),
};
let uid = 0;
let gid = 0;
class Job {
constructor({ runtime, files, args, stdin, timeouts, memory_limits }) {
this.uuid = uuidv4();
this.runtime = runtime;
this.files = files.map((file,i) => ({
this.files = files.map((file, i) => ({
name: file.name || `file${i}.code`,
content: file.content
content: file.content,
}));
this.args = args;
@ -37,11 +36,15 @@ class Job {
uid++;
gid++;
uid %= (config.runner_uid_max - config.runner_uid_min) + 1;
gid %= (config.runner_gid_max - config.runner_gid_min) + 1;
uid %= config.runner_uid_max - config.runner_uid_min + 1;
gid %= config.runner_gid_max - config.runner_gid_min + 1;
this.state = job_states.READY;
this.dir = path.join(config.data_directory, globals.data_directories.jobs, this.uuid);
this.dir = path.join(
config.data_directory,
globals.data_directories.jobs,
this.uuid
);
}
async prime() {
@ -51,7 +54,7 @@ class Job {
logger.debug(`Transfering ownership uid=${this.uid} gid=${this.gid}`);
await fs.mkdir(this.dir, { mode:0o700 });
await fs.mkdir(this.dir, { mode: 0o700 });
await fs.chown(this.dir, this.uid, this.gid);
for (const file of this.files) {
@ -74,43 +77,38 @@ class Job {
'prlimit',
'--nproc=' + config.max_process_count,
'--nofile=' + config.max_open_files,
'--fsize=' + config.max_file_size
'--fsize=' + config.max_file_size,
];
if (memory_limit >= 0) {
prlimit.push('--as=' + memory_limit);
}
const proc_call = [
...prlimit,
...nonetwork,
'bash',file,
...args
];
const proc_call = [...prlimit, ...nonetwork, 'bash', file, ...args];
var stdout = '';
var stderr = '';
var output = '';
const proc = cp.spawn(proc_call[0], proc_call.splice(1) ,{
const proc = cp.spawn(proc_call[0], proc_call.splice(1), {
env: {
...this.runtime.env_vars,
PISTON_LANGUAGE: this.runtime.language
PISTON_LANGUAGE: this.runtime.language,
},
stdio: 'pipe',
cwd: this.dir,
uid: this.uid,
gid: this.gid,
detached: true //give this process its own process group
detached: true, //give this process its own process group
});
proc.stdin.write(this.stdin);
proc.stdin.end();
proc.stdin.destroy();
const kill_timeout = set_timeout(_ => proc.kill('SIGKILL'), timeout);
const kill_timeout = set_timeout((_) => proc.kill('SIGKILL'), timeout);
proc.stderr.on('data', data => {
proc.stderr.on('data', (data) => {
if (stderr.length > config.output_max_size) {
proc.kill('SIGKILL');
} else {
@ -119,7 +117,7 @@ class Job {
}
});
proc.stdout.on('data', data => {
proc.stdout.on('data', (data) => {
if (stdout.length > config.output_max_size) {
proc.kill('SIGKILL');
} else {
@ -135,7 +133,7 @@ class Job {
proc.stdout.destroy();
};
proc.on('exit', (code, signal)=>{
proc.on('exit', (code, signal) => {
exit_cleanup();
resolve({ stdout, stderr, code, signal, output });
@ -151,10 +149,16 @@ class Job {
async execute() {
if (this.state !== job_states.PRIMED) {
throw new Error('Job must be in primed state, current state: ' + this.state.toString());
throw new Error(
'Job must be in primed state, current state: ' + this.state.toString()
);
}
logger.info(`Executing job uuid=${this.uuid} uid=${this.uid} gid=${this.gid} runtime=${this.runtime.toString()}`);
logger.info(
`Executing job uuid=${this.uuid} uid=${this.uid} gid=${
this.gid
} runtime=${this.runtime.toString()}`
);
logger.debug('Compiling');
@ -163,7 +167,7 @@ class Job {
if (this.runtime.compiled) {
compile = await this.safe_call(
path.join(this.runtime.pkgdir, 'compile'),
this.files.map(x => x.name),
this.files.map((x) => x.name),
this.timeouts.compile,
this.memory_limits.compile
);
@ -184,7 +188,7 @@ class Job {
compile,
run,
language: this.runtime.language,
version: this.runtime.version.raw
version: this.runtime.version.raw,
};
}
@ -192,25 +196,27 @@ class Job {
let processes = [1];
while (processes.length > 0) {
processes = await new Promise((resolve, reject) => cp.execFile('ps', ['awwxo', 'pid,ruid'], (err, stdout) => {
processes = await new Promise((resolve, reject) =>
cp.execFile('ps', ['awwxo', 'pid,ruid'], (err, stdout) => {
if (err === null) {
const lines = stdout.split('\n').slice(1); //Remove header with slice
const procs = lines.map(line => {
const procs = lines.map((line) => {
const [pid, ruid] = line
.trim()
.split(/\s+/)
.map(n => parseInt(n));
.map((n) => parseInt(n));
return { pid, ruid }
return { pid, ruid };
});
resolve(procs);
} else {
reject(error);
}
}));
})
);
processes = processes.filter(proc => proc.ruid === this.uid);
processes = processes.filter((proc) => proc.ruid === this.uid);
for (const proc of processes) {
// First stop the processes, but keep their resources allocated so they cant re-fork
@ -221,7 +227,6 @@ class Job {
}
}
for (const proc of processes) {
// Then clear them out of the process tree
try {
@ -250,7 +255,7 @@ class Job {
}
} catch (e) {
// File was somehow deleted in the time that we read the dir to when we checked the file
logger.warn(`Error removing file ${file_path}: ${e}`)
logger.warn(`Error removing file ${file_path}: ${e}`);
}
}
}
@ -261,14 +266,10 @@ class Job {
async cleanup() {
logger.info(`Cleaning up job uuid=${this.uuid}`);
await Promise.all([
this.cleanup_processes(),
this.cleanup_filesystem()
]);
await Promise.all([this.cleanup_processes(), this.cleanup_filesystem()]);
}
}
module.exports = {
Job
Job,
};

View File

@ -13,8 +13,7 @@ const chownr = require('chownr');
const util = require('util');
class Package {
constructor({ language, version, download, checksum }){
constructor({ language, version, download, checksum }) {
this.language = language;
this.version = semver.parse(version);
this.checksum = checksum;
@ -22,7 +21,9 @@ class Package {
}
get installed() {
return fss.exists_sync(path.join(this.install_path, globals.pkg_installed_file));
return fss.exists_sync(
path.join(this.install_path, globals.pkg_installed_file)
);
}
get install_path() {
@ -42,14 +43,18 @@ class Package {
logger.info(`Installing ${this.language}-${this.version.raw}`);
if (fss.exists_sync(this.install_path)) {
logger.warn(`${this.language}-${this.version.raw} has residual files. Removing them.`);
logger.warn(
`${this.language}-${this.version.raw} has residual files. Removing them.`
);
await fs.rm(this.install_path, { recursive: true, force: true });
}
logger.debug(`Making directory ${this.install_path}`);
await fs.mkdir(this.install_path, {recursive: true});
await fs.mkdir(this.install_path, { recursive: true });
logger.debug(`Downloading package from ${this.download} in to ${this.install_path}`);
logger.debug(
`Downloading package from ${this.download} in to ${this.install_path}`
);
const pkgpath = path.join(this.install_path, 'pkg.tar.gz');
const download = await fetch(this.download);
@ -63,7 +68,8 @@ class Package {
logger.debug('Validating checksums');
logger.debug(`Assert sha256(pkg.tar.gz) == ${this.checksum}`);
const cs = crypto.create_hash("sha256")
const cs = crypto
.create_hash('sha256')
.update(fss.readFileSync(pkgpath))
.digest('hex');
@ -71,10 +77,14 @@ class Package {
throw new Error(`Checksum miss-match want: ${val} got: ${cs}`);
}
logger.debug(`Extracting package files from archive ${pkgpath} in to ${this.install_path}`);
logger.debug(
`Extracting package files from archive ${pkgpath} in to ${this.install_path}`
);
await new Promise((resolve, reject) => {
const proc = cp.exec(`bash -c 'cd "${this.install_path}" && tar xzf ${pkgpath}'`);
const proc = cp.exec(
`bash -c 'cd "${this.install_path}" && tar xzf ${pkgpath}'`
);
proc.once('exit', (code, _) => {
code === 0 ? resolve() : reject();
@ -95,20 +105,15 @@ class Package {
const envout = await new Promise((resolve, reject) => {
let stdout = '';
const proc = cp
.spawn(
'env',
['-i','bash','-c',`${get_env_command}`],
{
stdio: ['ignore', 'pipe', 'pipe']
}
);
const proc = cp.spawn('env', ['-i', 'bash', '-c', `${get_env_command}`], {
stdio: ['ignore', 'pipe', 'pipe'],
});
proc.once('exit', (code, _) => {
code === 0 ? resolve(stdout) : reject();
});
proc.stdout.on('data', data => {
proc.stdout.on('data', (data) => {
stdout += data;
});
@ -117,67 +122,73 @@ class Package {
const filtered_env = envout
.split('\n')
.filter(l => !['PWD','OLDPWD','_', 'SHLVL'].includes(l.split('=',2)[0]))
.filter(
(l) => !['PWD', 'OLDPWD', '_', 'SHLVL'].includes(l.split('=', 2)[0])
)
.join('\n');
await fs.write_file(path.join(this.install_path, '.env'), filtered_env);
logger.debug('Changing Ownership of package directory');
await util.promisify(chownr)(this.install_path,0,0);
await util.promisify(chownr)(this.install_path, 0, 0);
logger.debug('Writing installed state to disk');
await fs.write_file(path.join(this.install_path, globals.pkg_installed_file), Date.now().toString());
await fs.write_file(
path.join(this.install_path, globals.pkg_installed_file),
Date.now().toString()
);
logger.info(`Installed ${this.language}-${this.version.raw}`);
return {
language: this.language,
version: this.version.raw
version: this.version.raw,
};
}
async uninstall(){
async uninstall() {
logger.info(`Uninstalling ${this.language}-${this.version.raw}`);
logger.debug("Finding runtime")
const found_runtime = runtime.get_runtime_by_name_and_version(this.language, this.version.raw);
logger.debug('Finding runtime');
const found_runtime = runtime.get_runtime_by_name_and_version(
this.language,
this.version.raw
);
if(!found_runtime){
logger.error(`Uninstalling ${this.language}-${this.version.raw} failed: Not installed`)
throw new Error(`${this.language}-${this.version.raw} is not installed`)
if (!found_runtime) {
logger.error(
`Uninstalling ${this.language}-${this.version.raw} failed: Not installed`
);
throw new Error(`${this.language}-${this.version.raw} is not installed`);
}
logger.debug("Unregistering runtime")
logger.debug('Unregistering runtime');
found_runtime.unregister();
logger.debug("Cleaning files from disk")
await fs.rmdir(this.install_path, {recursive: true})
logger.debug('Cleaning files from disk');
await fs.rmdir(this.install_path, { recursive: true });
logger.info(`Uninstalled ${this.language}-${this.version.raw}`)
logger.info(`Uninstalled ${this.language}-${this.version.raw}`);
return {
language: this.language,
version: this.version.raw
version: this.version.raw,
};
}
static async get_package_list() {
const repo_content = await fetch(config.repo_url).then(x => x.text());
const repo_content = await fetch(config.repo_url).then((x) => x.text());
const entries = repo_content
.split('\n')
.filter(x => x.length > 0);
const entries = repo_content.split('\n').filter((x) => x.length > 0);
return entries.map(line => {
const [ language, version, checksum, download ] = line.split(',', 4);
return entries.map((line) => {
const [language, version, checksum, download] = line.split(',', 4);
return new Package({
language,
version,
checksum,
download
download,
});
});
}
@ -185,16 +196,14 @@ class Package {
static async get_package(lang, version) {
const packages = await Package.get_package_list();
const candidates = packages
.filter(pkg => {
return pkg.language == lang && semver.satisfies(pkg.version, version)
const candidates = packages.filter((pkg) => {
return pkg.language == lang && semver.satisfies(pkg.version, version);
});
candidates.sort((a, b) => semver.rcompare(a.version, b.version));
return candidates[0] || null;
}
}
module.exports = Package;

View File

@ -8,8 +8,7 @@ const path = require('path');
const runtimes = [];
class Runtime {
constructor({language, version, aliases, pkgdir, runtime}) {
constructor({ language, version, aliases, pkgdir, runtime }) {
this.language = language;
this.version = version;
this.aliases = aliases || [];
@ -34,22 +33,26 @@ class Runtime {
if (provides) {
// Multiple languages in 1 package
provides.forEach(lang => {
runtimes.push(new Runtime({
provides.forEach((lang) => {
runtimes.push(
new Runtime({
language: lang.language,
aliases: lang.aliases,
version,
pkgdir: package_dir,
runtime: language
}));
runtime: language,
})
);
});
} else {
runtimes.push(new Runtime({
runtimes.push(
new Runtime({
language,
version,
aliases,
pkgdir: package_dir
}))
pkgdir: package_dir,
})
);
}
logger.debug(`Package ${language}-${version} was loaded`);
@ -73,8 +76,8 @@ class Runtime {
env_content
.trim()
.split('\n')
.map(line => line.split('=',2))
.forEach(([key,val]) => {
.map((line) => line.split('=', 2))
.forEach(([key, val]) => {
this._env_vars[key.trim()] = val.trim();
});
}
@ -94,16 +97,29 @@ class Runtime {
module.exports = runtimes;
module.exports.Runtime = Runtime;
module.exports.get_runtimes_matching_language_version = function(lang, ver){
return runtimes.filter(rt => (rt.language == lang || rt.aliases.includes(lang)) && semver.satisfies(rt.version, ver));
module.exports.get_runtimes_matching_language_version = function (lang, ver) {
return runtimes.filter(
(rt) =>
(rt.language == lang || rt.aliases.includes(lang)) &&
semver.satisfies(rt.version, ver)
);
};
module.exports.get_latest_runtime_matching_language_version = function(lang, ver){
return module.exports.get_runtimes_matching_language_version(lang, ver)
.sort((a,b) => semver.rcompare(a.version, b.version))[0];
module.exports.get_latest_runtime_matching_language_version = function (
lang,
ver
) {
return module.exports
.get_runtimes_matching_language_version(lang, ver)
.sort((a, b) => semver.rcompare(a.version, b.version))[0];
};
module.exports.get_runtime_by_name_and_version = function(runtime, ver){
return runtimes.find(rt => (rt.runtime == runtime || (rt.runtime === undefined && rt.language == runtime)) && semver.satisfies(rt.version, ver));
}
module.exports.get_runtime_by_name_and_version = function (runtime, ver) {
return runtimes.find(
(rt) =>
(rt.runtime == runtime ||
(rt.runtime === undefined && rt.language == runtime)) &&
semver.satisfies(rt.version, ver)
);
};
module.exports.load_package = Runtime.load_package;