Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Mixtral training #639

Open
ChristianPala opened this issue Oct 25, 2024 · 0 comments
Open

Mixtral training #639

ChristianPala opened this issue Oct 25, 2024 · 0 comments

Comments

@ChristianPala
Copy link

Hey folks,

when using:
TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ

or the more recent and recommended:
casperhansen/mixtral-instruct-awq

with the training script provided in the examples, I get a runtime exception:

Exception has occurred: RuntimeError

CUDA error: CUBLAS_STATUS_NOT_SUPPORTED when calling cublasGemmEx( handle, opa, opb, m, n, k, &falpha, a, CUDA_R_16F, lda, b, CUDA_R_16F, ldb, &fbeta, c, CUDA_R_16F, ldc, CUDA_R_32F, CUBLAS_GEMM_DEFAULT_TENSOR_OP)
File "/home/christian.pala/models-zoo/models_zoo/main.py", line 5, in main cli.main(prog="models-zoo") File "/home/christian.pala/models-zoo/models_zoo/main.py", line 9, in main()RuntimeError: CUDA error: CUBLAS_STATUS_NOT_SUPPORTED when calling cublasGemmEx( handle, opa, opb, m, n, k, &falpha, a, CUDA_R_16F, lda, b, CUDA_R_16F, ldb, &fbeta, c, CUDA_R_16F, ldc, CUDA_R_32F, CUBLAS_GEMM_DEFAULT_TENSOR_OP)

Is adapter training not supported for mixtral?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant