Skip to content

Conversation

@aninrusimha
Copy link

If users don't have sudo access, or want to have different cuda versions for each conda environment, it can be helpful to install nvcc on a per environment basis. It is non-intuitive which of the many conda package (nvcc, cuda_runtime, etc.) you need to install, so I thought it would be useful to include that info in the source compilation instructions. This is how I installed the package locally, and can upload my conda env file if its wanted :)

TimDettmers and others added 30 commits October 5, 2021 19:16
[FIX] passing of sparse in StableEmbedding
mryab and others added 11 commits June 30, 2022 18:14
…orts

Remove unused imports, fix NotImplementedError
If users don't have sudo access, or want to have different cuda versions for each conda environment, it can be helpful to install nvcc on a per environment basis. It is non-intuitive which of the many conda package (nvcc, cuda_runtime, etc.) you need to install, so I thought it would be useful to include that info in the source compilation instructions. This is how I installed the package locally, and can upload my conda env file if its wanted :)
@aninrusimha aninrusimha changed the title Add conda nvcc installation to source compile Add conda nvcc installation instructions to compile_from_source.md Aug 16, 2022
@TimDettmers
Copy link
Collaborator

This is a great addition, thank you! One issue with the conda install is that only a limited number of CUDA versions are supported.

We also offer a script to download and install a variety of CUDA versions locally: cuda_install.sh. Could you add instructions on how to use this?

@TimDettmers TimDettmers closed this May 7, 2023
TNTran92 pushed a commit to TNTran92/bitsandbytes that referenced this pull request Mar 24, 2024
matthewdouglas pushed a commit that referenced this pull request Sep 15, 2025
* Add SYCL Kernels for XPU backend

* fix transpose

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix log and format

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* revert cpu changes

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* clean ipex_xpu

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* clean ipex import

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix ipex cpu import

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix typo

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix comments

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* refine gemv_4bit kernel

* enable FP4 for dequant_4bit and gemv_4bit

* refine FP4 dequantization performance

* remove check for better performance

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix doc

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* clean code

* fix tests

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* rm comments

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix memory issue

* fix ut failure

* adjust threshold

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix xpu check

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* change test_functional check

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix test_module

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix device check

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix tests

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* Enable Windows build and refine code

* fix xpu log

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* remove ipex entirely

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix cpu int8 CB

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix lint

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix logs (#12)

* fix logs

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix format

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

---------

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* Fix sycl lint error and tests (#13)

* fix sycl nd

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix tests

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

---------

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* skip typo check for xpu kernel codes (#14)

* skip test for xpu ops

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* fix lint

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* skip typo for xpu

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* skip

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* skip

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

---------

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* register triton kernel for quantization (#15)

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* Fix version comparison issue (#18)

# Description

The version comparison expression miss reference the .release property from the version object. This lead to compare between the tuple and the string

# Error message
```
The 8-bit optimizer is not available on your device, only available on CUDA for now.
🦥 Unsloth: Will patch your computer to enable 2x faster free finetuning.
Traceback (most recent call last):
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/unsloth_validation/run.py", line 1, in <module>
    import unsloth
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/v/lib/python3.10/site-packages/unsloth/__init__.py", line 235, in <module>
    from .models import *
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/v/lib/python3.10/site-packages/unsloth/models/__init__.py", line 15, in <module>
    from .llama     import FastLlamaModel
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/v/lib/python3.10/site-packages/unsloth/models/llama.py", line 23, in <module>
    from ._utils import *
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/v/lib/python3.10/site-packages/unsloth/models/_utils.py", line 89, in <module>
    from unsloth_zoo.patching_utils import (
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/v/lib/python3.10/site-packages/unsloth_zoo/patching_utils.py", line 629, in <module>
    import transformers.integrations.bitsandbytes
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/v/lib/python3.10/site-packages/transformers/integrations/bitsandbytes.py", line 20, in <module>
    import bitsandbytes as bnb
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/bitsandbytes/bitsandbytes/__init__.py", line 39, in <module>
    from .backends.xpu import ops as xpu_ops
  File "/home/erxin/jenkins/workspace/Unsloth_Benchmark/bitsandbytes/bitsandbytes/backends/xpu/ops.py", line 17, in <module>
    if version.parse(torch.__version__).release >= version.parse("2.9"):
TypeError: '>=' not supported between instances of 'tuple' and 'Version'
```

---------

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
Co-authored-by: jiqing-feng <jiqing.feng@intel.com>
Co-authored-by: Er-Xin (Edwin) Shang <shangerxin@hotmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants