Update ollama to 0.3.9 and add +cuda variant#46204
Conversation
|
This does build w/ and w/out cuda but I am unable to figure out how to make a model actually use my GPUs (2x 4090). I suspect it has to do with not finding all CUDA libs from the Spack FWIW, I can download the binary from ollama's GitHub release page (which includes its own copy of CUDA libs) and get a model running on the GPU. |
|
@spackbot fix style |
|
Let me see if I can fix that for you! |
|
I was able to run spack style --fix==> Running style checks on spack
selected: isort, black, flake8, mypy
==> Modified files
var/spack/repos/builtin/packages/ollama/package.py
==> Running isort checks
isort checks were clean
==> Running black checks
reformatted var/spack/repos/builtin/packages/ollama/package.py
All done! ✨ 🍰 ✨
1 file reformatted.
black checks were clean
==> Running flake8 checks
flake8 checks were clean
==> Running mypy checks
lib/spack/spack/version/version_types.py:145: error: Argument 2 to "StandardVersion" has incompatible type "*Tuple[Tuple[Any, ...], Tuple[Any, ...]]"; expected "Tuple[Tuple[Any, ...], Tuple[Any, ...]]" [arg-type]
lib/spack/spack/version/version_types.py:452: error: Argument 2 to "StandardVersion" has incompatible type "*Tuple[Tuple[Any, ...], Tuple[Any, ...]]"; expected "Tuple[Tuple[Any, ...], Tuple[Any, ...]]" [arg-type]
lib/spack/spack/version/version_types.py:481: error: Argument 2 to "StandardVersion" has incompatible type "*Tuple[Tuple[Any, ...], Tuple[Any, ...]]"; expected "Tuple[Tuple[Any, ...], Tuple[Any, ...]]" [arg-type]
Found 3 errors in 1 file (checked 620 source files)
mypy found errors
I've updated the branch with style fixes. |
|
Thanks for taking this on! I was hoping to get back to this and get a GPU build going. I don't have much in the way of good CUDA resource available for testing on my dev system but I'll see if I can help sort out the build details. |
|
I think the CUDA runtime "problem" I had was a mirage due to a mix of wrong expectation from Spack's I had thought Spack's The With that knowledge in hand, I let There was also some prior confusion due to However, with So in the end, all seems good. A fresh push is immanent. |
|
Sorry to keep dragging you along here! I'm not as familiar with Cuda packages as I am other things. I took a bit more time to review this morning and had a few more suggestions:
|
teaguesterling
left a comment
There was a problem hiding this comment.
Sorry to give you the round-around on these changes. I'll open a PR into your branch with them as well to make it easier to implement.
Signed-off-by: Teague Sterling <[email protected]>
|
All stylistic considerations aside: I was able to sort of confirm that this works for building with Cuda. I have an ancient CUDA-compatible card in a machine. It was able to compile with Cuda support and got as far as detecting that my card was a dinosaur before dropping back to CPU. |
Fixing style audits and simplifying dependencies
|
@spackbot fix style |
|
Let me see if I can fix that for you! |
|
@spackbot rerun pipeline |
|
@spackbot rerun pipeline @brettviren something seems broken in the CI unrelated to the package. Rerunning again in hopes it's resolved. |
|
@spackbot rerun pipeline |
|
Lgtm! |
No description provided.