mirror of
https://github.com/RYDE-WORK/llama.cpp.git
synced 2026-02-05 07:13:18 +08:00
cuda : fix compile warning (#7454)
This commit is contained in:
parent
95fb0aefab
commit
9b3d833189
@ -283,7 +283,6 @@ void launch_fattn_tile_f32_64_128(ggml_backend_cuda_context & ctx, ggml_tensor *
|
|||||||
}
|
}
|
||||||
|
|
||||||
void ggml_cuda_flash_attn_ext_tile_f32(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
|
void ggml_cuda_flash_attn_ext_tile_f32(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
|
||||||
const ggml_tensor * KQV = dst;
|
|
||||||
const ggml_tensor * Q = dst->src[0];
|
const ggml_tensor * Q = dst->src[0];
|
||||||
|
|
||||||
if (Q->ne[1] <= 16) {
|
if (Q->ne[1] <= 16) {
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user