Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
Spaces:
Alovestocode
/
ZeroGPU-LLM-Inference
like
0
Sleeping
App
Files
Files
Community
Fetching metadata from the HF Docker repository...
main
ZeroGPU-LLM-Inference
165 kB
1 contributor
History:
83 commits
Alikestocode
Add GPU estimator, DDG search, and cancel support
4ce42e8
27 days ago
.dockerignore
Safe
104 Bytes
Add Google Cloud Platform deployment configurations
about 1 month ago
.gitattributes
Safe
1.52 kB
Initial commit: ZeroGPU LLM Inference Space
about 1 month ago
.gitignore
Safe
27 Bytes
Add .gitignore and remove cache files
about 1 month ago
DEPLOYMENT_STATUS.md
Safe
2.21 kB
Add deployment status document after re-authentication
30 days ago
Dockerfile
Safe
1.02 kB
Fix delete_revisions import with fallback cache cleanup
29 days ago
FIX_PERMISSIONS.md
Safe
2.05 kB
Add permission fix guide for spherical-gate-477614-q7 project
about 1 month ago
LLM_COMPRESSOR_FEATURES.md
Safe
6.24 kB
Fix AWQModifier import path: use modifiers.awq instead of modifiers.quantization
29 days ago
MANUAL_DEPLOY.md
Safe
1.59 kB
Fix delete_revisions import with fallback cache cleanup
29 days ago
QUANTIZE_AWQ.md
Safe
3.21 kB
Fix AWQModifier import path: use modifiers.awq instead of modifiers.quantization
29 days ago
README.md
Safe
4.23 kB
Implement vLLM with LLM Compressor and performance optimizations
about 1 month ago
app.py
Safe
56.9 kB
Add GPU estimator, DDG search, and cancel support
27 days ago
apt.txt
Safe
11 Bytes
Initial commit: ZeroGPU LLM Inference Space
about 1 month ago
cloudbuild.yaml
Safe
1.36 kB
Add Cloud Build deployment script and permission setup helper
about 1 month ago
deploy-cloud-build.sh
Safe
3.31 kB
Add Cloud Build deployment script and permission setup helper
about 1 month ago
deploy-compute-engine.sh
Safe
4.23 kB
Add Google Cloud Platform deployment configurations
about 1 month ago
deploy-gcp.sh
Safe
2.67 kB
Add Google Cloud Platform deployment configurations
about 1 month ago
gcp-deployment.md
Safe
5.32 kB
Add Google Cloud Platform deployment configurations
about 1 month ago
quantize_to_awq_colab.ipynb
Safe
32.9 kB
Lower Gemma AWQ group size to 16
28 days ago
requirements.txt
Safe
397 Bytes
Clarify LLM Compressor optional status - vLLM has native AWQ support
29 days ago
setup-gcp-permissions.sh
Safe
1.8 kB
Add Cloud Build deployment script and permission setup helper
about 1 month ago
style.css
Safe
2.84 kB
Initial commit: ZeroGPU LLM Inference Space
about 1 month ago
test_api.py
Safe
3.43 kB
Migrate to AWQ quantization with FlashAttention-2
about 1 month ago
test_api_gradio_client.py
Safe
7.2 kB
Implement vLLM with LLM Compressor and performance optimizations
about 1 month ago
test_awq_models.py
Safe
3.12 kB
Add test scripts for AWQ models on ZeroGPU Space
28 days ago
test_quantization_notebook.py
Safe
9.84 kB
Update Qwen model to use AWQ quantized version
28 days ago
test_space_awq.sh
Safe
1.93 kB
Add test scripts for AWQ models on ZeroGPU Space
28 days ago
test_space_simple.py
Safe
3.49 kB
Add test scripts for AWQ models on ZeroGPU Space
28 days ago
test_space_simple.sh
Safe
1.68 kB
Fix delete_revisions import with fallback cache cleanup
29 days ago