matmul_primitive_create_and_cache Class — pytorch Architecture
Architecture documentation for the matmul_primitive_create_and_cache class in DnnlExt.h from the pytorch codebase.
Entity Profile
Source Code
aten/src/ATen/native/mkldnn/xpu/detail/DnnlExt.h lines 545–592
template <typename F>
static inline primitive_ext& matmul_primitive_create_and_cache(
const joint_dtypes_t Ts,
const trans_type_t Tt,
const bias_type_t b_dims,
const int m,
const int n,
const int k,
const int64_t lda,
const int64_t ldb, // is weight ldb necessary?
const int64_t ldc,
const int device_id,
F attr,
const int64_t scale_group_size = 0,
const int64_t zp_group_size = 0) {
switch (Ts) {
case joint_dtypes_t::f16_int4:
return matmul_primitive_create_and_cache<joint_dtypes_t::f16_int4, F>(
Tt,
b_dims,
m,
n,
k,
lda,
ldb,
ldc,
device_id,
attr,
scale_group_size,
zp_group_size);
case joint_dtypes_t::bf16_int4:
return matmul_primitive_create_and_cache<joint_dtypes_t::bf16_int4, F>(
Tt,
b_dims,
m,
n,
k,
lda,
ldb,
ldc,
device_id,
attr,
scale_group_size,
zp_group_size);
default:
TORCH_INTERNAL_ASSERT(false, "Only support int4 ...");
}
}
Source
Analyze Your Own Codebase
Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.
Try Supermodel Free