@mikegpl/hyperflow-job-executor

1.1.9 • Public • Published

HyperFlow Job executor

GitHub tag (latest SemVer pre-release)

This is a basic HyperFlow job executor that uses local directory path to read and write files, and Redis for communication with the HyperFlow engine.

Adding the executor to a Docker image

  • Install Node.js 12.x or higher
  • Install the executor package:
    • Latest version: npm install -g @hyperflow/job-executor
    • Specific version: npm install -g @hyperflow/job-executor@X.Y.Z
    • From master branch: npm install -g https://github.com/hyperflow-wms/hyperflow-job-executor/archive/master.tar.gz

Running jobs

Jobs can be run with either of the following commands:

  • hflow-job-execute <taskId> <redisUrl>, where taskId is a unique job identifier, while redisUrl is an URL to the Redis server where the actual job command is fetched from. Both parameters are available in HyperFlow functions as context.taskId and context.redis_url, respectively.
  • hflow-job-execute <redisUrl> -a -- <taskId>... -- to run multiple jobs sequentially (useful for agglomeration of small jobs).

Jobs can be submitted e.g. using the HyperFlow function k8sCommand. See RemoteJobs example to learn more details.

Logging

The executor creates log files in directory <work_dir>/logs-hf that contain:

  • command used to execute the job
  • stdout and stderr of the job
  • metrics (CPU/memory/IO/network usage)
  • events (job start/end)
  • system information (e.g. hardware configuration)
  • all environment variables starting with HF_LOG_ -- a JSON object is logged following conventions from the read-env package

Configuration

The following environment variables can be used to adjust the behavior of the job executor:

  • HF_VAR_PROBE_INTERVAL (default 2000): time interval (in ms) at which to probe and log metrics.
  • HF_VAR_NUMBER_OF_RETRIES (default 1): how many times the job should be re-executed if it returns a non-zero exit code.
  • HF_VAR_BACKOFF_SEED (default 10): factor used in calculating the backoff delay between retries.
  • HF_VAR_WAIT_FOR_INPUT_FILES: if set to 1, the executor will check if input files exist and wait for them (useful in systems where files are synchronized in an eventually consistent fashion).
  • HF_VAR_FILE_WATCH_NUM_RETRIES (default 10): how many times should the executor check for existence of the input files (with backoff waits in between).
  • HF_VAR_WORK_DIR: path to the working directory where the job should be executed. If not set, /work_dir will be used if exists, otherwise the executor will not change the working directory.
  • HF_VAR_LOG_DIR: path to the directory where log files should be written. If not set, <work dir>/logs-hf will be used.
  • HF_VAR_LOG_LEVEL (default info): set logging level (trace, debug, info, warn, error, fatal).
  • HF_VAR_ENABLE_NETHOGS: if set (to any value), logs from nethogs will be written (experimental).
  • HF_LOG_*: all variables starting with HF_LOG_ will be logged in the job log files

Releasing

For quick and dirty developer releases

# Commit your changes
make dev-release

To release a proper version:

# Commit your changes
# Use npm version <arg>, to tag your changes and bump npm version
make release

Readme

Keywords

none

Package Sidebar

Install

npm i @mikegpl/hyperflow-job-executor

Weekly Downloads

6

Version

1.1.9

License

MIT

Unpacked Size

38.4 kB

Total Files

12

Last publish

Collaborators

  • mikegpl