This node is the default scheduler that executes work items on your local machine. This node also runs its scheduled work items in parallel.
This node lets you:
Specify the total resources of the scheduler with the Total Slots parameter.
Set how many work items should run at once with the Slots Per Work Item parameter.
These parameters are not related to the multi-threading of individual processes. To control the multi-threading of Houdini work items, use the Houdini Max Threads parameter.
To execute work items on a farm or on remote machines, use a different scheduler like the HQueue Scheduler.
Global parameters for all work items using this scheduler.
Print extra debugging information in job output logs.
Block on Failed Work Items
When on, if there are any failed work items on the scheduler, then the cook is blocked from completing and the PDG graph cook is prevented from ending. This allows you to manually retry your failed work items. You can cancel the scheduler’s cook when it is blocked by failed work items by pressing the ESC key, clicking the Cancels the current cook button in the TOP tasks bar, or by using the cancel API method.
Specifies the relative directory where the work generates intermediate files and output. The intermediate files are placed in a subdirectory. For the Local Scheduler or HQueue, typically
$HIP is used. For other schedulers, this should be a relative directory to
Local Shared Root Path and
Remote Shared Root Path; this path is then appended to these root paths.
Load Item Data From
Determines how jobs processed by this scheduler should load work item attributes and data.
Temporary JSON File
The scheduler writes out a
.json file for each work item to the PDG temporary file directory. This option is selected by default.
The scheduler’s running work items request attributes and data over RPC. If the scheduler is a farm scheduler, then the job scripts running on the farm will also request item data from the submitter when creating their out-of-process work item objects.
This parameter option removes the need to write data files to disk and is useful when your local and remote machines do not share a file system.
Compress Work Item Data
When on, PDG compresses the work item
.json files when writing them to disk.
This parameter is only available when Load Item Data From is set to Temporary JSON File.
If the PDG Path Map exists, then it is applied to file paths.
Delocalizes paths using the
Path Map Zone
When on, specifies a custom mapping zone to apply to all jobs executed by this scheduler. Otherwise, the local platforms are
Validate Outputs When Recooking
When on, PDG validates the output files of the scheduler’s cooked work items when the graph is recooked to see if the files still exist on disk. Work items that are missing output files are then automatically dirtied and cooked again. If any work items are dirtied by parameter changes, then their cache files are also automatically invalidated. Validate Outputs When Recooking is on by default.
Check Expected Outputs on Disk
When on, PDG looks for any unexpected outputs (for example, like outputs that can result from custom output handling internal logic) that were not explicitly reported when the scheduler’s work items finished cooking. This check occurs immediately after the scheduler marks work items as cooked, and expected outputs that were reported normally are not checked. If PDG finds any files that are different from the expected outputs, then they are automatically added as real output files.
When on, a subdirectory is added to the location specified by the Location parameter and is named after the value of your Houdini session’s PID (Process Identifier). The PID is typically a 3-5 digit number.
This is necessary when multiple sessions of Houdini are cooking TOP graphs at the same time.
The full path to the custom temporary directory.
This parameter is only available when Directory Location is set to Custom.
You can override these parameters per node with the Edit Parameter Interface. For more information, see Scheduler Job Parms / Properties.
When on, only one single work item is executed at a time.
Slots Per Work Item
When on, sets the number of slots consumed by each work item. Work items are only run by the scheduler if at least this number of slots are available.
The total number of slots that are available to the scheduler is determined by the Total Slots parameter setting.
If some of your tasks consume a lot of computational or memory resources, you can use the Slots Per Work Item parameter to change the maximum number of processes that are run in parallel. For example, if there are 8 slots available as determined by Total Slots, then a maximum of 8 processes will be executed in parallel. However, if Slots Per Work Item is set to 2 on the processor node, then a maximum of 4 processes will be executed in parallel with each task consuming 2 slots worth of resources in the scheduler.
Specifies the amount of available memory that is required to start a job. This allows you to delay a job from starting until a specific amount of memory is available.
No check is performed for available memory.
Check for the specified Minimum MB.
Check for the specified Minimum Percent.
Sets the minimum amount of available memory in Megabytes (MBs). Available memory is the amount of memory that can be used by a process without going into swap.
The parameter is only available when Rule is set to MB Available.
Sets the minimum amount of available memory as a percentage of the system’s total memory. Available memory is the amount of memory that can be used by a process without going into swap.
The parameter is only available when Rule is set to Percent Available.
When a work item process terminates with a non-zero exit code, it is marked as failed by default. These parameters change that behavior.
On Task Failure
Determines what happens when a work item fails.
The work item fails and an error message is added to the node.
The work item succeeds and a warning message is added to the node.
The work item restarts immediately, according to the Maximum Retries and Retry Count Attribute parameter settings.
The work item succeeds and no message is issued.
Handle All Non Zero
When off, lets you specify a particular exit code with the Exit Code field. All other non-zero exit codes are regarded as failures.
Specifies the exit code that is handled by the On Task Failure parameter setting. All other non-zero exit codes are treated as failures.
This parameter is only available when Handle All Non Zero is off.
Sets the maximum number of times the work item will be restarted.
Retry Count Attribute
When on, adds an
int attribute set to the number of times the task was restarted.
Maximum Run Time
When on, this parameter determines the maximum time in seconds that a work item can run. When the time limit is exceeded, the work item’s process is terminated.
On Task Timeout
Determines what status to set on the work items that timed out.
This parameter is only available when Maximum Run Time is enabled.
Mark as Failed
Sets the work item’s status to Failed.
Mark as Succeeded
Sets the work item’s status to Succeeded.
Houdini Max Threads
When on, sets the maximum number of threads each work item can use. This also sets the
HOUDINI_MAXTHREADS environment value which is used by Houdini-based programs like Mantra, Karma, Hython, and HBatch.
Requires GUI Window
Normally, processes are started such that they do not pop up command windows on the desktop when they run. However, some Windows applications require a GUI window.
When on, your work items can run GUI Applications in pop-up windows.
Specifies a space-separated list of environment variables that should be unset in or removed from the scheduler’s task environment.
Additional work item environment variables can be specified here. These will be added to the job’s environment. If the value of the variable is empty, it will be removed from the job’s environment.
Name of the work item environment variable.
Value of the work item environment variable.