Skip to content

[Operator] Add cublas to matmul tune space#422

Merged
hjjq merged 1 commit intohidet-org:mainfrom
hjjq:cublas
Jan 24, 2024
Merged

[Operator] Add cublas to matmul tune space#422
hjjq merged 1 commit intohidet-org:mainfrom
hjjq:cublas

Conversation

@hjjq
Copy link
Copy Markdown
Collaborator

@hjjq hjjq commented Jan 23, 2024

Also fix bug in call graph generation.

@hjjq
Copy link
Copy Markdown
Collaborator Author

hjjq commented Jan 23, 2024

$hidet-ci launch

@hjjq hjjq merged commit 072a606 into hidet-org:main Jan 24, 2024
@hjjq hjjq deleted the cublas branch February 1, 2024 03:39
vadiklyutiy pushed a commit that referenced this pull request Dec 19, 2024
In regression tests, accuracy comparison was added for llms. But it
broke the vision model benchmarking script.
```
Running command: python /workspace/hidet/tests/benchmarks/bench_transformer.py bert-base-uncased --params bs=16,seqlen=256 --dtype float16 --backend hidet --mode max-autotune
  File "/workspace/hidet/tests/benchmarks/run_tests.py", line 82, in <module>
    latency = float(outputs.split('\n')[-2])
ValueError: could not convert string to float: "       device='cuda:0', dtype=torch.float16))"
```
Fixed  it!

---------

Co-authored-by: Zhumakhan <nazirzhumakhan@gmail,.com>
vadiklyutiy pushed a commit that referenced this pull request Dec 20, 2024
In regression tests, accuracy comparison was added for llms. But it
broke the vision model benchmarking script.
```
Running command: python /workspace/hidet/tests/benchmarks/bench_transformer.py bert-base-uncased --params bs=16,seqlen=256 --dtype float16 --backend hidet --mode max-autotune
  File "/workspace/hidet/tests/benchmarks/run_tests.py", line 82, in <module>
    latency = float(outputs.split('\n')[-2])
ValueError: could not convert string to float: "       device='cuda:0', dtype=torch.float16))"
```
Fixed  it!

---------

Co-authored-by: Zhumakhan <nazirzhumakhan@gmail,.com>
vadiklyutiy pushed a commit that referenced this pull request Dec 26, 2024
In regression tests, accuracy comparison was added for llms. But it
broke the vision model benchmarking script.
```
Running command: python /workspace/hidet/tests/benchmarks/bench_transformer.py bert-base-uncased --params bs=16,seqlen=256 --dtype float16 --backend hidet --mode max-autotune
  File "/workspace/hidet/tests/benchmarks/run_tests.py", line 82, in <module>
    latency = float(outputs.split('\n')[-2])
ValueError: could not convert string to float: "       device='cuda:0', dtype=torch.float16))"
```
Fixed  it!

---------

Co-authored-by: Zhumakhan <nazirzhumakhan@gmail,.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant