auto-gptq-0.7.1+cu122 |
auto_gptq-0.7.1+cu122-cp310-cp310-linux_aarch64.whl
|
|
autoawq-0.2.4+cu122 |
autoawq-0.2.4+cu122-py3-none-any.whl
|
|
autoawq-kernels-0.0.6+cu122 |
autoawq_kernels-0.0.6+cu122-cp310-cp310-linux_aarch64.whl
|
|
awq-0.1.0 |
awq-0.1.0-py3-none-any.whl
|
|
awq-inference-engine-0.0.0 |
awq_inference_engine-0.0.0-cp310-cp310-linux_aarch64.whl
|
|
bitsandbytes-0.39.1 |
bitsandbytes-0.39.1-py3-none-any.whl
|
|
ctranslate2-4.2.0 |
ctranslate2-4.2.0-cp310-cp310-linux_aarch64.whl
ctranslate2-4.2.0-cp311-cp311-linux_aarch64.whl
|
|
cuda-python-12.2.0+0.g2ae98f9.dirty |
cuda_python-12.2.0+0.g2ae98f9.dirty-cp310-cp310-linux_aarch64.whl
|
|
cupy-13.0.0rc1 |
cupy-13.0.0rc1-cp310-cp310-linux_aarch64.whl
|
|
exllamav2-0.0.15 |
exllamav2-0.0.15-cp310-cp310-linux_aarch64.whl
|
|
faiss-1.7.4 |
faiss-1.7.4-py3-none-any.whl
|
|
flash-attn-2.5.7 |
flash_attn-2.5.7-cp310-cp310-linux_aarch64.whl
|
|
llama-cpp-python-0.2.57 |
llama_cpp_python-0.2.57-cp310-cp310-manylinux_2_35_aarch64.whl
|
|
mlc-chat-0.1.0 |
mlc_chat-0.1.0-cp310-cp310-linux_aarch64.whl
|
|
mlc-llm-0.1.1 |
mlc_llm-0.1.1-cp310-cp310-linux_aarch64.whl
|
|
onnxruntime-gpu-1.17.0 |
onnxruntime_gpu-1.17.0-cp310-cp310-linux_aarch64.whl
onnxruntime_gpu-1.17.0-cp311-cp311-linux_aarch64.whl
|
|
opencv-contrib-python-4.9.0.80 |
opencv_contrib_python-4.9.0.80-cp310-cp310-linux_aarch64.whl
|
|
pycuda-2024.1 |
pycuda-2024.1-cp310-cp310-linux_aarch64.whl
|
|
tensorflow-2.15.0+nv24.3 |
tensorflow-2.15.0+nv24.3-cp310-cp310-linux_aarch64.whl
|
|
torch-2.3.0 |
torch-2.3.0-cp310-cp310-linux_aarch64.whl
|
|
torchaudio-2.3.0+952ea74 |
torchaudio-2.3.0+952ea74-cp310-cp310-linux_aarch64.whl
|
|
torchvision-0.18.0a0+6043bc2 |
torchvision-0.18.0a0+6043bc2-cp310-cp310-linux_aarch64.whl
|
|
triton-3.0.0 |
triton-3.0.0-cp310-cp310-linux_aarch64.whl
|
|
tvm-0.16.0 |
tvm-0.16.0-cp310-cp310-linux_aarch64.whl
|
|
xformers-0.0.26+622595c.d20240502 |
xformers-0.0.26+622595c.d20240502-cp310-cp310-linux_aarch64.whl
|
|