Skip to content

Commit

Permalink
Fix attention_dtype flag for paged_llm_v1
Browse files Browse the repository at this point in the history
Signed-off-by: aviator19941 <avinash.sharma@amd.com>
  • Loading branch information
aviator19941 committed Feb 13, 2025
1 parent de70094 commit ae89b55
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion sharktank/sharktank/examples/paged_llm_v1.py
Original file line number Diff line number Diff line change
Expand Up @@ -288,7 +288,7 @@ def main():
block_seq_stride=args.block_seq_stride,
device=device,
activation_dtype=args.activation_dtype,
attention_dtype=args.activation_dtype,
attention_dtype=args.attention_dtype,
attention_kernel=args.attention_kernel,
use_hf=args.use_hf,
tensor_parallelism_size=args.tensor_parallelism_size,
Expand Down

0 comments on commit ae89b55

Please sign in to comment.