|
|
|
@@ -402,19 +402,19 @@ enum FaCodePath {
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
struct vk_fa_pipeline_state {
|
|
|
|
|
vk_fa_pipeline_state(uint32_t HSK, uint32_t HSV, bool small_rows, bool small_cache, FaCodePath path, bool aligned, bool f32acc, bool use_mask_opt)
|
|
|
|
|
: HSK(HSK), HSV(HSV), small_rows(small_rows), small_cache(small_cache), path(path), aligned(aligned), f32acc(f32acc), use_mask_opt(use_mask_opt) {}
|
|
|
|
|
vk_fa_pipeline_state(uint32_t HSK, uint32_t HSV, bool small_rows, bool small_cache, FaCodePath path, bool aligned, bool f32acc, uint32_t flags)
|
|
|
|
|
: HSK(HSK), HSV(HSV), small_rows(small_rows), small_cache(small_cache), path(path), aligned(aligned), f32acc(f32acc), flags(flags) {}
|
|
|
|
|
|
|
|
|
|
uint32_t HSK, HSV;
|
|
|
|
|
bool small_rows, small_cache;
|
|
|
|
|
FaCodePath path;
|
|
|
|
|
bool aligned;
|
|
|
|
|
bool f32acc;
|
|
|
|
|
bool use_mask_opt;
|
|
|
|
|
uint32_t flags;
|
|
|
|
|
|
|
|
|
|
bool operator<(const vk_fa_pipeline_state &b) const {
|
|
|
|
|
return std::tie(HSK, HSV, small_rows, small_cache, path, aligned, f32acc, use_mask_opt) <
|
|
|
|
|
std::tie(b.HSK, b.HSV, b.small_rows, b.small_cache, b.path, b.aligned, b.f32acc, b.use_mask_opt);
|
|
|
|
|
return std::tie(HSK, HSV, small_rows, small_cache, path, aligned, f32acc, flags) <
|
|
|
|
|
std::tie(b.HSK, b.HSV, b.small_rows, b.small_cache, b.path, b.aligned, b.f32acc, b.flags);
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
@@ -3193,7 +3193,7 @@ static void ggml_vk_load_shaders(vk_device& device) {
|
|
|
|
|
return {fa_rows_cols(path, hsk, hsv, clamp, type, small_rows, small_cache)[0], 1, 1};
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
auto const &fa_spec_constants = [&](FaCodePath path, uint32_t hsk, uint32_t hsv, uint32_t clamp, ggml_type type, bool small_rows, bool small_cache, bool use_mask_opt) -> std::vector<uint32_t> {
|
|
|
|
|
auto const &fa_spec_constants = [&](FaCodePath path, uint32_t hsk, uint32_t hsv, uint32_t clamp, ggml_type type, bool small_rows, bool small_cache, uint32_t flags) -> std::vector<uint32_t> {
|
|
|
|
|
// For large number of rows, 128 invocations seems to work best.
|
|
|
|
|
// For small number of rows (e.g. N==1), 256 works better. But matrix granularity for 256 is 32, so we
|
|
|
|
|
// can't use 256 for D==80.
|
|
|
|
@@ -3225,7 +3225,7 @@ static void ggml_vk_load_shaders(vk_device& device) {
|
|
|
|
|
// AMD prefers loading K directly from global memory
|
|
|
|
|
const uint32_t k_load_shmem = device->vendor_id == VK_VENDOR_ID_NVIDIA && hsk < 256 ? 1 : 0;
|
|
|
|
|
|
|
|
|
|
return {wg_size, rows_cols[0], rows_cols[1], hsk, hsv, clamp, D_split, device->subgroup_size, k_load_shmem, use_mask_opt};
|
|
|
|
|
return {wg_size, rows_cols[0], rows_cols[1], hsk, hsv, clamp, D_split, device->subgroup_size, k_load_shmem, flags};
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
#define CREATE_FA(TYPE, NAMELC, FAPATH, SUFFIX) \
|
|
|
|
@@ -3237,19 +3237,19 @@ static void ggml_vk_load_shaders(vk_device& device) {
|
|
|
|
|
FaCodePath path = fa.first.path; \
|
|
|
|
|
bool aligned = fa.first.aligned; \
|
|
|
|
|
bool f32acc = fa.first.f32acc; \
|
|
|
|
|
bool use_mask_opt = fa.first.use_mask_opt; \
|
|
|
|
|
uint32_t flags = fa.first.flags; \
|
|
|
|
|
if (path == FAPATH) { \
|
|
|
|
|
if (aligned) { \
|
|
|
|
|
if (f32acc) { \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_aligned_f32acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache,use_mask_opt), fa_align(FAPATH,HSK,HSV,TYPE,small_rows,small_cache), true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_aligned_f32acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache,flags), fa_align(FAPATH,HSK,HSV,TYPE,small_rows,small_cache), true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
} else { \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_aligned_f16acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache,use_mask_opt), fa_align(FAPATH,HSK,HSV,TYPE,small_rows,small_cache), true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_aligned_f16acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,0,TYPE,small_rows,small_cache,flags), fa_align(FAPATH,HSK,HSV,TYPE,small_rows,small_cache), true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
} \
|
|
|
|
|
} else { \
|
|
|
|
|
if (f32acc) { \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_f32acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache,use_mask_opt), 1, true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_f32acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache,flags), 1, true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
} else { \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_f16acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache,use_mask_opt), 1, true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
ggml_vk_create_pipeline(device, fa.second, "flash_attn_f32_f16_f16acc" #NAMELC, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _len, flash_attn_f32_f16_ ## NAMELC ## _f16acc ## SUFFIX ## _data, "main", 7, sizeof(vk_flash_attn_push_constants), fa_wg_denoms(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache), fa_spec_constants(FAPATH, HSK,HSV,1,TYPE,small_rows,small_cache,flags), 1, true, FAPATH==FA_COOPMAT1, (FAPATH==FA_COOPMAT1 ? device->subgroup_size : 0)); \
|
|
|
|
|
} \
|
|
|
|
|
} \
|
|
|
|
|
} \
|
|
|
|
@@ -8595,10 +8595,26 @@ static void ggml_vk_flash_attn(ggml_backend_vk_context * ctx, vk_context& subctx
|
|
|
|
|
|
|
|
|
|
bool f32acc = path == FA_SCALAR || dst->op_params[3] == GGML_PREC_F32;
|
|
|
|
|
|
|
|
|
|
float scale = 1.0f;
|
|
|
|
|
float max_bias = 0.0f;
|
|
|
|
|
float logit_softcap = 0.0f;
|
|
|
|
|
|
|
|
|
|
memcpy(&scale, (const float *) dst->op_params + 0, sizeof(float));
|
|
|
|
|
memcpy(&max_bias, (const float *) dst->op_params + 1, sizeof(float));
|
|
|
|
|
memcpy(&logit_softcap, (const float *) dst->op_params + 2, sizeof(float));
|
|
|
|
|
|
|
|
|
|
if (logit_softcap != 0) {
|
|
|
|
|
scale /= logit_softcap;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Only use mask opt when the mask is fairly large. This hasn't been tuned extensively.
|
|
|
|
|
bool use_mask_opt = mask && nem1 >= 32 && nem0 * nem1 > 32768;
|
|
|
|
|
|
|
|
|
|
vk_fa_pipeline_state fa_pipeline_state(HSK, HSV, small_rows, small_cache, path, aligned, f32acc, use_mask_opt);
|
|
|
|
|
uint32_t flags = (use_mask_opt ? 1 : 0) |
|
|
|
|
|
(mask != nullptr ? 2 : 0) |
|
|
|
|
|
(logit_softcap != 0 ? 4 : 0);
|
|
|
|
|
|
|
|
|
|
vk_fa_pipeline_state fa_pipeline_state(HSK, HSV, small_rows, small_cache, path, aligned, f32acc, flags);
|
|
|
|
|
|
|
|
|
|
vk_pipeline pipeline = nullptr;
|
|
|
|
|
|
|
|
|
@@ -8678,18 +8694,6 @@ static void ggml_vk_flash_attn(ggml_backend_vk_context * ctx, vk_context& subctx
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
float scale = 1.0f;
|
|
|
|
|
float max_bias = 0.0f;
|
|
|
|
|
float logit_softcap = 0.0f;
|
|
|
|
|
|
|
|
|
|
memcpy(&scale, (const float *) dst->op_params + 0, sizeof(float));
|
|
|
|
|
memcpy(&max_bias, (const float *) dst->op_params + 1, sizeof(float));
|
|
|
|
|
memcpy(&logit_softcap, (const float *) dst->op_params + 2, sizeof(float));
|
|
|
|
|
|
|
|
|
|
if (logit_softcap != 0) {
|
|
|
|
|
scale /= logit_softcap;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const uint32_t n_head_kv = neq2;
|
|
|
|
|
const uint32_t n_head_log2 = 1u << (uint32_t) floorf(log2f((float) n_head_kv));
|
|
|
|
|
const float m0 = powf(2.0f, -(max_bias ) / n_head_log2);
|
|
|
|
@@ -8703,7 +8707,7 @@ static void ggml_vk_flash_attn(ggml_backend_vk_context * ctx, vk_context& subctx
|
|
|
|
|
vk_subbuffer sinks_buf = sinks ? ggml_vk_tensor_subbuffer(ctx, sinks) : q_buf;
|
|
|
|
|
vk_subbuffer mask_opt_buf = use_mask_opt ? ggml_vk_subbuffer(ctx, ctx->prealloc_y, 0) : q_buf;
|
|
|
|
|
|
|
|
|
|
uint32_t mask_n_head_log2 = ((sinks != nullptr) << 24) | ((mask != nullptr) << 16) | n_head_log2;
|
|
|
|
|
uint32_t mask_n_head_log2 = ((sinks != nullptr) << 24) | n_head_log2;
|
|
|
|
|
|
|
|
|
|
if (use_mask_opt)
|
|
|
|
|
{
|
|
|
|
|