|
template<typename T > |
__global__ void | spike::device::var::bandLU (T *dA, int *ks, int *offsets, int partition_size, int rest_num) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_safe (T *dA, int *ks, int *offsets, int partition_size, int rest_num) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_g32 (T *dA, int *ks, int *offsets, int partition_size, int rest_num) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_g32_safe (T *dA, int *ks, int *offsets, int partition_size, int rest_num) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_critical_div_general (T *dA, int start_row, int *ks, int *offsets, int partition_size, int rest_num) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_critical_div_safe_general (T *dA, int start_row, int *ks, int *offsets, int partition_size, int rest_num) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_critical_sub_general (T *dA, int start_row, int *ks, int *offsets, int partition_size, int rest_num, int last) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_post_divide_per_partition (T *dA, int k, int offset, int partSize) |
|
template<typename T > |
__global__ void | spike::device::var::bandLU_post_divide_per_partition_general (T *dA, int k, int offset, int partSize) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_div (T *dA, int *ks, int *offsets, int cur_row) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_div_safe (T *dA, int *ks, int *offsets, int cur_row) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_div_general (T *dA, int *ks, int *offsets, int cur_row) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_div_safe_general (T *dA, int *ks, int *offsets, int cur_row) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_sub (T *dA, int *ks, int *offsets, int cur_row) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_sub_general (T *dA, int *ks, int *offsets, int cur_row) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_sub_spec (T *dA, int *ks, int *offsets) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_sub_spec_general (T *dA, int *ks, int *offsets) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_post_divide (T *dA, int *ks, int *offsets) |
|
template<typename T > |
__global__ void | spike::device::var::fullLU_post_divide_general (T *dA, int *ks, int *offsets) |
|
template<typename T > |
__global__ void | spike::device::var::boostLastPivot (T *dA, int start_row, int *ks, int *offsets, int partition_size, int rest_num) |
|
Various matrix factorization CUDA kernels used for the case of partitions with equal bandwidths.
Various matrix factorization CUDA kernels used for the case of partitions with varying bandwidths.
Various forward/backward sweep CUDA kernels used for the case of partitions with varying bandwidths.