-
Notifications
You must be signed in to change notification settings - Fork 60
Enable Ahead-of-Time Compilation by hiding the runtime functions in the GLOBAL_METHOD_TABLE
#749
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
|
Your PR requires formatting changes to meet the project's style guidelines. Click here to view the suggested changes.diff --git a/src/utils.jl b/src/utils.jl
index 8242e5a..9f55e9c 100644
--- a/src/utils.jl
+++ b/src/utils.jl
@@ -196,12 +196,14 @@ macro device_function(ex)
error("This function is not intended for use on the CPU")
end
- esc(quote
- $(combinedef(def))
+ return esc(
+ quote
+ $(combinedef(def))
- # NOTE: no use of `@consistent_overlay` here because the regular function errors
- Base.Experimental.@overlay($(GPUCompiler).GLOBAL_METHOD_TABLE, $ex)
- end)
+ # NOTE: no use of `@consistent_overlay` here because the regular function errors
+ Base.Experimental.@overlay($(GPUCompiler).GLOBAL_METHOD_TABLE, $ex)
+ end
+ )
end
|
|
Loaded both forked CUDA.jl and this PR and tried to compile my full code and got error: Stacktrace is massive so I copied first several lines. Otherwise, I could compile GPUCompiler in image. Do you have idea where could this come from? |
|
@KSepetanc yep, I caught that in the tests for the PR as well (somehow they were passing for me locally, but I suspect that was just poor environment management on my part). Unsurprisingly, my hack seems to break things in I will turn this PR to a draft until then. |
…g us out of some KernelAbstractions compilations in e.g OpenCL.jl
|
@apozharski are you using CUDA 590 driver branch (it is CUDA 13.1)? I have seen maintainers are preparing support for it, but last I checked a few days ago it still was not released. Without knowing more about your system, I presume you just need to downgrade to 580 series driver that comes with CUDA 13.0. I had this issue too. I will soon have more questions as it seems that more fixes are needed than just GPUCompiler.jl and CUDA.jl to AOT compile MadNLPGPU which I need, but it is still WIP so I will wait a bit. |
As discussed in JuliaGPU/CUDA.jl#2998 and #611 currently
GPUCompiler.jlleaks nonexistantgpu_*llvm functions into the cpu cache making ahead of time compilation impossible for any package that uses it.I am currently fixing this by moving these runtime methods into the method table defined in the GPUCompiler module and having the CPU versions throw errors as is done in
CUDA.jl. This feels like somewhat of a hack, however, it seems to work and without a better understanding of what this might break it seems to be the simplest solution.