Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it GPU Compatability Issue? #353

Open
kamrul-NSL opened this issue Nov 15, 2024 · 4 comments
Open

Is it GPU Compatability Issue? #353

kamrul-NSL opened this issue Nov 15, 2024 · 4 comments

Comments

@kamrul-NSL
Copy link

kamrul-NSL commented Nov 15, 2024

from diffusers import PixArtSigmaPipeline
import torch

pipeline = PixArtSigmaPipeline.from_pretrained(
    "PixArt-alpha/PixArt-Sigma-XL-2-1024-MS", torch_dtype=torch.float16
).to("cuda")

quantize(pipeline.transformer, weights=qfloat8)
freeze(pipeline.transformer)```

Here I am getting this error!

**RuntimeError: Error building extension 'quanto_cuda': [1/7] /usr/bin/nvcc --generate-dependencies-with-compile --dependency-output unpack.cuda.o.d -DTORCH_EXTENSION_NAME=quanto_cuda -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_gcc\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1011\" -isystem /home/user/anaconda3/envs/fp8/lib/python3.10/site-packages/torch/include -isystem /home/user/anaconda3/envs/fp8/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /home/user/anaconda3/envs/fp8/lib/python3.10/site-packages/torch/include/TH -isystem /home/user/anaconda3/envs/fp8/lib/python3.10/site-packages/torch/include/THC -isystem /home/user/anaconda3/envs/fp8/include/python3.10 -D_GLIBCXX_USE_CXX11_ABI=0 -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr -gencode=arch=compute_86,code=sm_86 --compiler-options '-fPIC' --expt-extended-lambda --use_fast_math -DQUANTO_CUDA_ARCH=860 -std=c++17 -c /home/user/anaconda3/envs/fp8/lib/python3.10/site-packages/optimum/quanto/library/extensions/cuda/unpack.cu -o unpack.cuda.o 
**
I am trying to integrate it using NVIDIA 3090 GPU.
@CyberVy
Copy link

CyberVy commented Nov 29, 2024

It works when the model is on CPU.
However, after moving model to GPU, the CPU memory still does not descend.

@LianShuaiLong
Copy link

same error!

@dacorvo
Copy link
Collaborator

dacorvo commented Dec 5, 2024

@LianShuaiLong @kamrul-NSL what version of pytorch are you using ?

@LianShuaiLong
Copy link

@LianShuaiLong @kamrul-NSL what version of pytorch are you using ?

2.4.0+cu121

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants