max_static_len Class — pytorch Architecture
Architecture documentation for the max_static_len class in FlattenIndicesCommon.h from the pytorch codebase.
Entity Profile
Source Code
aten/src/ATen/native/sparse/FlattenIndicesCommon.h lines 36–84
template <
template <typename func_t> class kernel_t,
typename index_t,
int64_t max_static_len = 0>
Tensor _flatten_indices_impl(const Tensor& indices, IntArrayRef size) {
TORCH_INTERNAL_ASSERT(indices.dim() > 1 && static_cast<size_t>(indices.size(0)) == size.size());
// Need owning storage in case of the Tensor class.
const auto hash_coeffs_storage = [&]() -> auto {
auto strides = c10::contiguous_strides(size);
return at::sparse::TensorGeometryHolder<max_static_len>(strides, strides, indices.options());
}();
const auto hash_coeffs = std::get<0>(*hash_coeffs_storage);
const auto hash_indices = [&]() -> Tensor {
const auto sparse_dim = indices.size(0);
const auto indices_dim_stride = indices.stride(0);
const auto indices_nnz_stride = indices.stride(1);
auto hash = at::arange(indices.size(1), indices.options().dtype(kLong));
auto iter = TensorIteratorConfig()
.set_check_mem_overlap(false)
.add_output(hash)
.add_input(hash)
.build();
{
const auto* RESTRICT ptr_indices = indices.const_data_ptr<index_t>();
KernelLauncher<kernel_t>::launch(iter,
// NOTE: capture by value required by CUDA
[=] FUNCAPI (int64_t nnz_idx) -> int64_t {
const auto* RESTRICT ptr_indices_dim = ptr_indices + nnz_idx * indices_nnz_stride;
auto hash = static_cast<int64_t>(0);
for (int64_t dim = 0; dim < sparse_dim; ++dim) {
const auto dim_hash_coeff = hash_coeffs[dim];
const auto dim_index = ptr_indices_dim[dim * indices_dim_stride];
hash += dim_index * dim_hash_coeff;
}
return hash;
});
}
return hash;
}();
return hash_indices;
}
Source
Analyze Your Own Codebase
Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.
Try Supermodel Free