Skip to content

Latest commit

 

History

History
482 lines (350 loc) · 20 KB

CHANGELOG.md

File metadata and controls

482 lines (350 loc) · 20 KB

v0.12.0

New features

Automatic allocation

  • #457 You can now specify the idle timeout for workers started by the automatic allocator using the --idle-timeout flag of the hq alloc add command.

Resiliency

  • #449 Tasks that were present during multiple crashes of the workers will be canceled.

CLI

  • #463 You can now wait until N workers are connected to the clusters with hq worker wait N.

Python API

  • Resource requests improvements in Python API.

Changes

CLI

  • #477 Requested resources are now shown while submitting an array and while viewing information about task TASK_ID of specified job JOB_ID using hq task info JOB_ID TASK_ID

  • #444 The hq task list command will now hide some details by default, to conserve space in terminal output. To show all details, use the -v flag to enable verbose output.

  • #455 Improve the quality of error messages produced when parsing various CLI parameters, like resources.

Automatic allocation

  • #448 The automatic allocator will now start workers in multi-node Slurm allocations using srun --overlap. This should avoid taking up Slurm task resources by the started workers (if possible). If you run into any issues with using srun inside HyperQueue tasks, please let us know.

Jobs

  • #483 There is no longer a length limit for job names.

Fixes

Job submission

  • #450 Attempts to resubmit a job with zero tasks will now result in an explicit error, rather than a crash of the client.

v0.11.0

New features

CLI

  • #464 New command was added that allows users to see more detailed info about selected task TASK_ID from a concrete job JOB_ID.

    $ hq task info JOB_ID TASK_ID
  • #423 You can now specify the server directory using the HQ_SERVER_DIR environment variable.

Resource management

  • #427 A new specifier has been added to specify indexed pool resources for workers as a set of individual resource indices.
    $ hq worker start --resource "gpus=list(1,3,8)"
  • #428 Workers will now attempt to automatically detect available GPU resources from the CUDA_VISIBLE_DEVICES environment variable.

Stream log

  • Basic export of stream log into JSON (hq log <log_file> export)

Server

  • Improved scheduling of multi-node tasks.

  • Server now generates a random unique ID (UID) string every time a new server is started (hq server start). It can be used as a placeholder %{SERVER_ID}.

Changes

CLI

  • #464 More detailed task information (Time, Paths) were moved from hq task list into hq task info.

  • #433 (Backwards incompatible change) The CLI command hq job tasks has been removed and its functionality has been incorporated into the hq task list command instead. resource requests,

  • #420 Shebang (e.g. #!/bin/bash) will now be read from submitted program based on the provided directives mode. If a shebang is found, HQ will execute the program located at the shebang path and pass it the rest of the submitted arguments.

    By default, directives and shebang will be read from the submitted program only if its filename ends with .sh. If you want to explicitly enable reading the shebang, pass --directives=file to hq submit.

    Another change is that the shebang is now read by the client (i.e. it will be read on the node that submits the job), not on worker nodes as previously. This means that the submitted file has to be accessible on the client node.

Resource management

  • #427 (Backwards incompatible change) The environment variable HQ_RESOURCE_INDICES_<resource-name>, which is passed to tasks with resource requests, has been renamed to HQ_RESOURCE_VALUES_<resource-name>.

  • #427 (Backwards incompatible change) The specifier for specifying indexed pool resources for workers as a range has been renamed from indices to range.

    # before
    $ hq worker start --resource "gpus=indices(1-3)"
    # now
    $ hq worker start --resource "gpus=range(1-3)"
  • #427 The generic resource documentation has been rewritten and improved.

v0.10.0

New features

Running tasks

  • HQ will now set the OpenMP OMP_NUM_THREADS environment variable for each task. The amount of threads will be set according to the number of requested cores. For example, this job submission:
$ hq submit --cpus=4 -- <program>

would pass OMP_NUM_THREADS=4 to the executed <program>.

  • New task OpenMP pinning mode was added. You can now use --pin=omp when submitting jobs. This CPU pin mode will generate the corresponding OMP_PLACES and OMP_PROC_BIND environment variables to make sure that OpenMP pins its threads to the exact cores allocated by HyperQueue.

  • Preview version of multi-node tasks. You may submit multi-node task by hq submit --nodes=X ...

CLI

  • Less verbose log output by default. You can use "--debug" to turn on the old behavior.

Changes

Scheduler

  • When there is only a few tasks, scheduler tries to fit tasks on fewer workers. Goal is to enable earlier stopping of workers because of idle timeout.

CLI

  • The --pin boolean option for submitting jobs has been changed to take a value. You can get the original behaviour by specifying --pin=taskset.

Fixes

Automatic allocation

  • PBS/Slurm allocations using multiple workers will now correctly spawn a HyperQueue worker on all allocated nodes.

v0.9.0

New features

Tasks

  • Task may be started with a temporary directory that is automatically deleted when the task is finished. (flag --task-dir).

  • Task may provide its own error message by creating a file with name passed by environment variable HQ_ERROR_FILENAME.

CLI

  • You can now use the hq task list <job-selector> command to display a list of tasks across multiple jobs.
  • Add --filter flag to worker list to allow filtering workers by their status.

Changes

Automatic allocation

  • Automatic allocation has been rewritten from scratch. It will no longer query PBS/Slurm allocation statuses periodically, instead it will try to derive allocation state from workers that connect to it from allocations.
  • When adding a new allocation queue, HyperQueue will now try to immediately submit a job into the queue to quickly test whether the entered configuration is correct. If you want to avoid this behaviour, you can use the --no-dry-run flag for hq alloc add <pbs/slurm>.
  • If too many submissions (10) or running allocations (3) fail in a succession, the corresponding allocation queue will be automatically removed to avoid error loops.
  • hq alloc events command has been removed.
  • The --max-kept-directories parameter for allocation queues has been removed. HyperQueue will now keep 20 last allocation directories amongst all allocation queues.

Fixes

  • HQ will no longer warn that stdout/stderr path does not contain the %{TASK_ID} placeholder when submitting array jobs if the placeholder is contained within the working directory path and stdout/stderr contains the %{CWD} placeholder.

v0.8.0

Fixes

Automatic allocation

  • Issue #294: The automatic allocator leaves behind directories of inactive (failed or finished) allocations on the filesystem. Although these directories contain useful debugging information, creating too many of them can needlessly waste disk space. To alleviate this, HyperQueue will now keep only the last 20 directories of inactive allocations per each allocation queue and remove the older directories to save space.

    You can change this parameter by using the --max-kept-directories flag when creating an allocation queue:

    $ hq alloc add pbs --time-limit 1h --max-kept-directories 100

New features

Jobs

  • Added new command for outputting stdout/stderr of jobs.

    # Print stdout of all tasks of job 1
    $ hq job cat 1 stdout
    
    # Print stderr of tasks 1, 2, 3 of job 5
    $ hq job cat 5 stderr --tasks 1-3

    You can find more information in the documentation

  • #HQ directives - You can now specify job parameters using a shell script passed to hq submit by using HQ directives such as #HQ --cpus=4. This feature was inspired by similar functionality that is present in e.g. PBS or Slurm. You can find more information in the documentation.

  • HyperQueue will now attempt to parse shebang (like #!/bin/bash) if you provide a path to a shell script (.sh) as the first command in hq submit. If the parsing is successful, HyperQueue will use the parsed interpreter path to execute the shell script. In practice, this means that you can now submit scripts beginning with a shebang like this:

    $ hq submit script.sh

    This previously failed, unless you provided an interpreter, or provided a path starting with . or an absolute path to the script.

  • Capturing stdio and attaching it to each task of a job. This can be used to submitting scripts without creating file. The following command will capture stdin and executes it in Bash

    $ hq submit --stdin bash

Worker configuration

  • You can now select what should happen when a worker loses its connection to the server using the new --on-worker-lost flag available for worker start and hq alloc add commands. You can find more information in the documentation.

CLI

  • You can now force HyperQueue commands to output machine-readable data using the --output-mode flag available to all HyperQueue commands. Notably, you can output data of the commands as JSON. You can find more information in the documentation.

  • You can now generate shell completion using the hq generate-completion <shell> command.

Changes

CLI

  • The command line interface for jobs has been changed to be more consistent with the interface for workers. Commands that have been formerly standalone (like hq jobs, hq resubmit, hq wait) are not accessed through hq job. The only previous job-related command that remained on the top level is hq submit, which is now a shortcut for hq job submit. Here is a table of changed commands:

    Previous command New command
    hq jobs hq job list
    hq job hq job info
    hq resubmit hq job resubmit
    hq cancel hq job cancel
    hq wait hq job wait
    hq progress hq job progress
    hq submit hq submit or hq job submit
  • The --tasks flag of the hq job info <job-id> command has been removed. If you want to display the individual tasks of a job, please use the hq task list <job-id> command.

  • The command line parsing of hq submit has been changed slightly. All flags and arguments that appear after the first positional argument will now be considered to belong to the executed program, not to the submit command. This mimics the behaviour of e.g. docker run. For example:

    $ hq submit foo --array 1-4
    # Before: submits a task array with 4 tasks that execute the program `foo`
    # Now: submits a single task that executes `foo --array 1-4`
  • hq job list will now only show queued and running jobs by default. You can use the --all flag to display all jobs or the --filter flag to filter jobs that are in specified states.

  • The --status flag of hq job resubmit has been renamed to --filter.

  • Tables outputted by various informational commands (like hq job info or hq worker list) are now more densely packed and should thus better fit on terminal screens.

Preview features

  • You can now store HyperQueue events into a log file and later export them to JSON for further processing. You can find more information in the documentation.

    Note that this functionality is quite low-level, and it's designed primarily for tool builders that use HyperQueue programmatically, not regular users. It is also currently unstable.

  • You can now try the preview version of HQ dashboard. It can be started via:

    $ hq dashboard

v0.7.0

Fixes

  • Fixes an invalid behavior of the scheduler when resources are defined

  • The automatic allocator will no longer keep submitting allocations in situations where the created workers would not be able to execute currently waiting tasks. Currently, this situation is detected only for the case when a task has a time request higher than the time limit of the allocation queue.

New features

Automatic allocation

  • You can now specify CPU and generic resources for workers created by the automatic allocator:
    $ hq alloc add pbs --time-limit 2h --cpus 4x4 --resource "gpu=indices(1-2)" -- -q qexp -A Project1
  • You can now test auto allocation parameters using a dry-run command:
    $ hq alloc dry-run pbs --time-limit 2h -- -q qexp -A Project1
    Using this command you can quickly test if PBS/Slurm will accept allocations created with the provided parameters.
  • You can now specify a limit for the number of workers spawned inside a single allocation queue. You can use the parameter --max-worker-count when creating a queue to make sure that the queue will not create too many workers.
    $ hq alloc add pbs --time-limit 00:10:00 --max-worker-count 10 -- -q qprod -A Project1
  • You can now specify the timelimit of PBS/Slurm allocations using the HH:MM:SS format: hq alloc add pbs --time-limit 01:10:30.

Resource management

  • Workers can be now started with the parameter --cpus="no-ht". When detecting CPUs in this mode, HyperThreading will be ignored (for each physical core only the first HT virtual core will be chosen).
  • The user may explicitly specify what CPU IDs should be used by a worker (including arrangement of IDs into sockets). (E.g. hq worker start --cpus=[[0, 1], [6, 8]])

CLI

  • Improve error messages printed when an invalid CLI parameter is entered.

Changes

  • The --time-limit parameter of hq alloc add command is now required.
  • hq alloc remove will no longer let you remove an allocation queue that contains running allocations by default. If you want to force its removal and cancel the running allocations immediately, use the --force flag.

v0.6.1

Fixes

  • Fixed computation of worker load in scheduler
  • Fixed performance problem when canceling more than 100k tasks

Changes

  • When a job is submitted, it does not show full details in response but only a short message. Details can be still shown by hq job <id>.

v0.6.0

New features

  • Generic resource management has been added. You can find out more in the documentation.
    • HyperQueue can now automatically detect how many Nvidia GPUs are present on a worker node.
  • You can now submit a task array where each task will receive one element of a JSON array using hq submit --from-json. You can find out more in the documentation.

Changes

  • There have been a few slight CLI changes:
    • hq worker list no longer has --offline and --online flags. It will now display only running workers by default. If you want to show also offline workers, use the --all flag.
    • hq alloc add no longer has a required --queue/--partition option. The PBS queue/Slurm partition should now be passed as a trailing argument after --: hq alloc add pbs -- -qqprod.
  • Server subdirectories generated for each run of the HyperQueue server are now named with a numeric ID instead of a date.
  • The documentation has been rewritten.

v0.5.0

New features

  • Time limit and Time request for tasks (options --time-limit and --time-request)
  • Time limit for workers
  • Job and task times are shown in job information tables
  • Integers in command line options can be now written with an underscore separator (e.g. --array=1-1_000)
  • Placeholders in log file paths
  • Preview version of PBS and SLURM auto allocation
  • HyperQueue can be now compiled without jemalloc (this enables PowerPC builds). To remove dependency on jemalloc, build HyperQueue with --no-default-features.

Changes

  • hq submit --wait and hq wait will no longer display a progress bar while waiting for the job(s) to finish. The progress bar was moved to hq submit --progress and hq progress.
  • The default path of job stdout and stderr has been changed to job-%{JOB_ID}/%{TASK_ID}.[stdout/stderr]
  • Normalization of stream's end behavior when job is canceled
  • Job id is now represented as u32

v0.4.0

New features

  • Streaming - streaming stdout/stderr of all tasks in a job into one file to avoid creating many files.
  • Better reporting where job is running.
  • Setting a priority via hq submit --priority <P>
  • Option hq submit --wait ... to wait until the submitted job finishes
  • Command hq wait <id> / all / last to wait for a given job(s)
  • Command hq resubmit <job-id> to resubmit a previous job
  • Command hq cancel all / hq cancel last to cancel all jobs / last job
  • Command hq worker stop all to cancel all workers
  • Command hq server info to get an information about server

v0.3.0

New features

  • Option for automatic closing workers without tasks (Idle timeout)
  • Submit option --max-fails X to cancel an job when more than X tasks fails
  • Submit option --each-line FILE to create a task per a line in a file.
  • Submit option --env VAR=VALUE to specify env variable in a task
  • Submit option --cwd DIR to specify a working dir of a task
  • New placeholders in paths: %{CWD}, %{DATE}, and %{SUBMIT_DIR}
  • Added a progressbar in a job array detail.
  • hq server start --host=xxx allows to specify hostname/address under which the server is visible

v0.2.1

New features

  • Filters for command hq jobs <filter> (e.g. hq jobs running)

Fixes

  • NUMA detection on some architectures

v0.2.0

New features

  • Job arrays
  • Cpu management
  • --stdout/--stderr configuration in submit