Skip to content

Commit

Permalink
Regards #3: Additional doxygen comments, corrections and completions …
Browse files Browse the repository at this point in the history
…of existing comments, doxygen markup tweaks, and spacing tweaks
  • Loading branch information
eyalroz committed Mar 25, 2024
1 parent 2f77537 commit 403bed0
Show file tree
Hide file tree
Showing 8 changed files with 246 additions and 28 deletions.
6 changes: 5 additions & 1 deletion src/cuda/api/array.hpp
Original file line number Diff line number Diff line change
Expand Up @@ -270,16 +270,20 @@ array_t<T, NumDimensions> wrap(
return { device_id, context_handle, handle, dimensions };
}

/// Create a new (typed) CUDA array of the specified dimensions
///@{
/// @param context ... in which the array is to be created
template <typename T, dimensionality_t NumDimensions>
array_t<T,NumDimensions> create(
const context_t& context,
dimensions_t<NumDimensions> dimensions);

/// @param device ... in whose primary context the array is to be created
template <typename T, dimensionality_t NumDimensions>
array_t<T,NumDimensions> create(
const device_t& device,
dimensions_t<NumDimensions> dimensions);

///@}

} // namespace array

Expand Down
1 change: 1 addition & 0 deletions src/cuda/api/context.hpp
Original file line number Diff line number Diff line change
Expand Up @@ -861,6 +861,7 @@ inline context_t get_with_fallback_push()

} // namespace current

/// @return true if the context is the primary context of its device
bool is_primary(const context_t& context);

namespace detail_ {
Expand Down
105 changes: 103 additions & 2 deletions src/cuda/api/launch_config_builder.hpp
Original file line number Diff line number Diff line change
Expand Up @@ -179,6 +179,8 @@ class launch_config_builder_t {
}
}

/// Use the information specified for the builder to figure out the grid and block
/// dimensions with which the kernel is to be launched
grid::composite_dimensions_t get_composite_dimensions() const noexcept(false)
{
auto result = get_unvalidated_composite_dimensions();
Expand All @@ -189,6 +191,10 @@ class launch_config_builder_t {
}

public:
/// Use the information specified to the builder (and defaults for the unspecified
/// information) to finalize the construction of a kernel launch configuration,
/// which can then be passed along with the kernel to a kernel-launching function,
/// e.g. the standalone @ref kernel::launch or the stream command @ref stream_t::enqueue_t::kernel_launch
launch_configuration_t build() const
{
auto result = launch_configuration_t{ get_composite_dimensions() };
Expand Down Expand Up @@ -392,6 +398,7 @@ class launch_config_builder_t {

}

/// Set the dimensions for each block in the intended kernel launch grid
launch_config_builder_t& block_dimensions(
grid::block_dimension_t x,
grid::block_dimension_t y = 1,
Expand All @@ -400,8 +407,17 @@ class launch_config_builder_t {
return block_dimensions(grid::block_dimensions_t{x, y, z});
}

/// Set the block in the intended kernel launch grid to be uni-dimensional
/// with a specified size
launch_config_builder_t& block_size(grid::block_dimension_t size) { return block_dimensions(size, 1, 1); }

/**
* Set the intended kernel launch grid to have 1D blocks, of the maximum
* length possible given the information specified to the builder.
*
* @note This will fail if neither a kernel nor a device have been chosen
* for the launch.
*/
launch_config_builder_t& use_maximum_linear_block()
{
grid::block_dimension_t max_size;
Expand All @@ -424,6 +440,16 @@ class launch_config_builder_t {
}

#if CUDA_VERSION >= 12000
/**
* Set the dimensions of multi-block clusters within the grid.
*
* @note There is only a small number of possible dimension combinations of clusters;
* and this function does _not_ guarantee to fail immediately if you specify an
* invalid such combination.
*
* @note This setting does not affect the overall dimensions of the grid in terms of
* blocks.
*/
launch_config_builder_t& cluster_blocks(grid::block_dimensions_t cluster_dims)
{
#ifndef NDEBUG
Expand All @@ -434,6 +460,9 @@ class launch_config_builder_t {
}
#endif

/// Set the dimension of the grid for the intended kernel launch, in terms
/// of blocks
///@{
launch_config_builder_t& grid_dimensions(grid::dimensions_t dims)
{
#ifndef NDEBUG
Expand All @@ -447,6 +476,7 @@ class launch_config_builder_t {
return *this;
}

///@}
launch_config_builder_t& grid_dimensions(
grid::dimension_t x,
grid::dimension_t y = 1,
Expand All @@ -455,9 +485,17 @@ class launch_config_builder_t {
return grid_dimensions(grid::dimensions_t{x, y, z});
}

/// Set the grid for the intended launch to be one-dimensional, with a specified number
/// of blocks
///@{
launch_config_builder_t& grid_size(grid::dimension_t size) {return grid_dimensions(size, 1, 1); }
launch_config_builder_t& num_blocks(grid::dimension_t size) {return grid_size(size); }
///@}


/// Set the overall number of _threads_, in each dimension, of all blocks
/// in the grid of the intended kernel launch
///@{
launch_config_builder_t& overall_dimensions(grid::overall_dimensions_t dims)
{
#ifndef NDEBUG
Expand All @@ -474,16 +512,30 @@ class launch_config_builder_t {
{
return overall_dimensions(grid::overall_dimensions_t{x, y, z});
}
///@}

/// Set the intended launch grid to be linear, with a specified overall number of _threads_
/// over all (1D) blocks in the grid
launch_config_builder_t& overall_size(grid::overall_dimension_t size) { return overall_dimensions(size, 1, 1); }

/**
* Set whether or blocks may synchronize with each other or not
*
* @note recall that even "non-cooperative" blocks can still access the same global memory
* locations, and can use atomic operations on such locations for (slow) synchronization.
*/
launch_config_builder_t& block_cooperation(bool cooperation)
{
thread_block_cooperation = cooperation;
return *this;
}

/// Let kernel thread blocks synchronize with each other, or are guaranteed to act independently
/// (atomic global memory operations notwithstanding)
launch_config_builder_t& blocks_may_cooperate() { return block_cooperation(true); }

/// Prevent kernel thread blocks synchronize with each other, guaranteeing each block will
/// work entirely independently (atomic global memory operations notwithstanding)
launch_config_builder_t& blocks_dont_cooperate() { return block_cooperation(false); }

launch_config_builder_t& dynamic_shared_memory_size(
Expand All @@ -493,11 +545,18 @@ class launch_config_builder_t {
return *this;
}

/// Indicate that the intended launch should not allocate any shared
/// memory for the kernel to use beyond the static amount necessitated
/// by its (compiled) code.
launch_config_builder_t& no_dynamic_shared_memory()
{
return dynamic_shared_memory_size(memory::shared::size_t(0));
}

/// Indicate that the intended launch should allocate a certain amount of shared
/// memory for the kernel to use beyond the static amount necessitated
/// by its (compiled) code.
///@{
launch_config_builder_t& dynamic_shared_memory_size(memory::shared::size_t size)
{
#ifndef NDEBUG
Expand All @@ -512,13 +571,32 @@ class launch_config_builder_t {
{
return dynamic_shared_memory_size(size);
}
///@}

/**
* Indicate that the intended launch should allocate additional shared
* memory for the kernel to use beyond the static amount necessitated
* by its (compiled) code - with the amount to be determined based on
* the block size
*
* @param shared_mem_size_determiner a function determining the dynamic
* shared memory size given the kernel launch block size
*/
launch_config_builder_t& dynamic_shared_memory(
kernel::shared_memory_size_determiner_t shared_mem_size_determiner)
{
return dynamic_shared_memory_size(shared_mem_size_determiner);
}

/**
* Indicate that the specified wrapped kernel will be the one launched
* with the configuration to be produced by this object. Such an indication
* provides this object with information about the device and context in
* which the kernel is to be launched, and ranges of possible values for
* certain parameters (e.g. shared memory size, dimensions).
*
* @note Calling this method obviates a call to the @ref device() method.
*/
launch_config_builder_t& kernel(const kernel_t* wrapped_kernel_ptr)
{
if (device_ and kernel_->device_id() != device_.value()) {
Expand All @@ -533,6 +611,15 @@ class launch_config_builder_t {
return *this;
}

/**
* Indicate that the intended kernel launch would occur on (some stream in
* some context on) the specified device. Such an indication provides this
* object with some information regarding ranges of possible values for
* certain parameters (e.g. shared memory size, dimensions).
*
* @note Do not call both this and the @ref kernel() method; prefer just that one.
*/
///@{
launch_config_builder_t& device(const device::id_t device_id)
{
if (kernel_ and kernel_->device_id() != device_id) {
Expand All @@ -548,7 +635,11 @@ class launch_config_builder_t {
{
return this->device(device.id());
}
///@}

/// Clear the association with a specific kernel (which may have been
/// set using the @ref kernel method)
///@{
launch_config_builder_t& kernel_independent()
{
kernel_ = nullptr;
Expand All @@ -559,13 +650,14 @@ class launch_config_builder_t {
kernel_ = nullptr;
return *this;
}
///@}

/**
* @brief THis will use information about the kernel, the already-set block size,
* @brief This will use information about the kernel, the already-set block size,
* and the device to create a unidimensional grid of blocks to exactly saturate
* the CUDA device's capacity for simultaneous active blocks.
*
* @note This will _not_ set the block size - unlike
* @note This will _not_ set the block size - unlike {@ref min_params_for_max_occupancy()}.
*/
launch_config_builder_t& saturate_with_active_blocks()
{
Expand All @@ -584,6 +676,14 @@ class launch_config_builder_t {
return *this;
}

/**
* @brief This will use information about the kernel and the device to define
* a minimum launch grid which should guarantee maximum occupancy of the GPU's
* multiprocessors.
*
* @note A builder after this call _will_ set the block dimensions - unlike
* {@ref saturate_with_active_blocks()} .
*/
launch_config_builder_t& min_params_for_max_occupancy()
{
if (not (kernel_)) {
Expand All @@ -600,6 +700,7 @@ class launch_config_builder_t {
}
}; // launch_config_builder_t

/// A slightly shorter-named construction idiom for @ref launch_config_builder_t
inline launch_config_builder_t launch_config_builder() { return {}; }

} // namespace cuda
Expand Down
26 changes: 26 additions & 0 deletions src/cuda/api/launch_configuration.hpp
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,16 @@ enum class cluster_scheduling_policy_t {
};
#endif

/**
* The full set of possible configuration parameters for launching a kernel on a GPU.
*
* @note Consider using a @ref launch_configuration_builder_t to incrementally construct
* these structs.
*
* @note this structure must be constructed with at least the grid and block dimensions.
*/
struct launch_configuration_t {
/// Dimensions of the launch grid in blocks, and of the individual blocks in the grid.
grid::composite_dimensions_t dimensions { grid::dimensions_t{ 0u, 0u, 0u }, grid::block_dimensions_t{ 0u, 0u, 0u } };

/**
Expand Down Expand Up @@ -160,6 +169,20 @@ struct launch_configuration_t {
constexpr launch_configuration_t(const launch_configuration_t&) = default;
constexpr launch_configuration_t(launch_configuration_t&&) = default;

/**
* Constructors corresponding to the CUDA runtime API's triple-chevron launch
* syntax:
*
* my_kernel <<< grid_Dims, block_dims, dynamic_shmem_size, my_stream >>> (
* arg1, arg2, arg3, etc);
*
* ... where the specified aspects of the launch configuration are the dimensions
* and the dynamic shared memory size.
*
* @note The choices of stream and kernel function are _not_ part of the launch
* configuration.
*/
///@{
constexpr launch_configuration_t(
grid::composite_dimensions_t grid_and_block_dimensions,
memory::shared::size_t dynamic_shared_mem = 0u
Expand All @@ -184,12 +207,14 @@ struct launch_configuration_t {
grid::block_dimensions_t(block_dims),
dynamic_shared_mem)
{ }
///@}

CPP14_CONSTEXPR launch_configuration_t& operator=(const launch_configuration_t& other) = default;
CPP14_CONSTEXPR launch_configuration_t& operator=(launch_configuration_t&&) = default;
};

#if __cplusplus < 202002L
///@cond
constexpr bool operator==(const launch_configuration_t lhs, const launch_configuration_t& rhs) noexcept
{
return
Expand All @@ -210,6 +235,7 @@ constexpr bool operator!=(const launch_configuration_t lhs, const launch_configu
{
return not (lhs == rhs);
}
///@endcond
#endif

namespace detail_ {
Expand Down
Loading

0 comments on commit 403bed0

Please sign in to comment.