Skip to content

Commit

Permalink
Fixed mixed up values
Browse files Browse the repository at this point in the history
  • Loading branch information
gshtras committed Mar 22, 2024
1 parent 6b186bb commit 6ff0272
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions csrc/attention/attention_kernels.cu
Original file line number Diff line number Diff line change
Expand Up @@ -815,11 +815,11 @@ template<
int BLOCK_SIZE,
bool IS_FP8_KV_CACHE,
#ifdef USE_ROCM
int NUM_THREADS = 128,
int PARTITION_SIZE = 512>
#else
int NUM_THREADS = 1024,
int PARTITION_SIZE = 1024>
#else
int NUM_THREADS = 128,
int PARTITION_SIZE = 512>
#endif
void paged_attention_v2_launcher(
torch::Tensor& out,
Expand Down

0 comments on commit 6ff0272

Please sign in to comment.