The following functions are available for querying the available hardware:
Returns a boolean to indicate the availability of a CUDA GPU.
This will initialize the driver if it hasn’t been initialized.
Detect supported CUDA hardware and print a summary of the detected hardware.
Returns a boolean indicating whether any supported devices were detected.
CUDA Python functions execute within a CUDA context. Each CUDA device in a
system has an associated CUDA context, and Numba presently allows only one context
per thread. For further details on CUDA Contexts, refer to the CUDA Driver API
Documentation on Context Management and the
CUDA C Programming Guide Context Documentation. CUDA Contexts
are instances of the
This object wraps a CUDA Context resource.
Contexts should not be constructed directly by user code.
Returns (free, total) memory in bytes in the context.
Pops this context off the current CPU thread. Note that this context must be at the top of the context stack, otherwise an error will occur.
Pushes this context on the current CPU Thread.
Clean up all owned resources in this context.
The following functions can be used to get or select the context:
Get the current device or use a device by device number, and return the CUDA context.
A decorator that ensures a CUDA context is available when fn is executed.
Decorating fn is equivalent to writing:
at each call site.
The following functions affect the current context:
Synchronize the current context.
Explicitly clears all contexts in the current thread, and destroys all contexts if the current thread is the main thread.
Numba maintains a list of supported CUDA-capable devices:
An indexable list of supported CUDA devices. This list is indexed by integer device ID.
Alternatively, the current device can be obtained:
Return the currently-selected device.
Provides a context manager for executing in the context of the chosen
device. The normal use of instances of this type is from
numba.cuda.gpus. For example, to execute on device 2:
with numba.cuda.gpus: d_a = numba.cuda.to_device(a)
to copy the array a onto device 2, referred to by d_a.
One may also select a context and device or get the current device using the following three functions:
Make the context associated with device device_id the current context.
Returns a Device instance.
Raises exception on error.
Get current device associated with the current thread
Return a list of all detected devices
numba.cuda.cudadrv.driver.Device class can be used to enquire about
the functionality of the selected device:
The device associated with a particular context.
A tuple, (major, minor) indicating the supported compute capability.
The integer ID of the device.
The name of the device (e.g. “GeForce GTX 970”)
Delete the context for the device. This will destroy all memory allocations, events, and streams created within the context.
The NVidia Visual Profiler can be used directly on executing CUDA Python code - it is not a requirement to insert calls to these functions into user code. However, these functions can be used to allow profiling to be performed selectively on specific portions of the code. For further information on profiling, see the NVidia Profiler User’s Guide.
Enable profile collection in the current context.
Disable profile collection in the current context.
Context manager that enables profiling on entry and disables profiling on exit.
Events can be used to monitor the progress of execution and to record the timestamps of specific points being reached. Event creation returns immediately, and the created event can be queried to determine if it has been reached. For further information, see the CUDA C Programming Guide Events section.
The following functions are used for creating and measuring the time between events:
Create a CUDA event. Timing data is only recorded by the event if it is
Compute the elapsed time between two events in milliseconds.
Events are instances of the
Event(context, handle, finalizer=None)¶
Returns True if all work before the most recent record has completed; otherwise, returns False.
Set the record point of the event to the current point in the given stream.
The event will be considered to have occurred when all work that was
queued in the stream at the time of the call to
record() has been
Synchronize the host thread for the completion of the event.
All future works submitted to stream will wait util the event completes.
Streams allow concurrency of execution on a single device within a given context. Queued work items in the same stream execute sequentially, but work items in different streams may execute concurrently. Most operations involving a CUDA device can be performed asynchronously using streams, including data transfers and kernel execution. For further details on streams, see the CUDA C Programming Guide Streams section.
To create a stream:
Create a CUDA stream that represents a command queue for the device.
Streams are instances of
Stream(context, handle, finalizer)¶
A context manager that waits for all commands in this stream to execute and commits any pending memory transfers upon exiting the context.
Wait for all commands in this stream to execute. This will commit any pending memory transfers.