Skip to content
@ModelCloud

ModelCloud.ai

Our mission is to give allow everyone, including bots, unlimited and free access to llm/ai models.

Pinned Loading

  1. GPTQModel GPTQModel Public

    Production ready LLM model compression/quantization toolkit with accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.

    Python 268 41

  2. Device-SMI Device-SMI Public

    Self-contained Python lib with zero-dependencies that give you a unified device properties for gpu, cpu, and npu. No more calling separate tools such as nvidia-smi or /proc/cpuinfo and parsing it y…

    Python 10 1

Repositories

Showing 4 of 4 repositories
  • GPTQModel Public

    Production ready LLM model compression/quantization toolkit with accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.

    ModelCloud/GPTQModel’s past year of commit activity
    Python 268 Apache-2.0 41 9 6 Updated Feb 11, 2025
  • Tokenicer Public
    ModelCloud/Tokenicer’s past year of commit activity
    Python 3 Apache-2.0 1 0 1 Updated Feb 11, 2025
  • optimum Public Forked from huggingface/optimum

    🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools

    ModelCloud/optimum’s past year of commit activity
    Python 0 Apache-2.0 504 0 0 Updated Feb 7, 2025
  • Device-SMI Public

    Self-contained Python lib with zero-dependencies that give you a unified device properties for gpu, cpu, and npu. No more calling separate tools such as nvidia-smi or /proc/cpuinfo and parsing it yourself.

    ModelCloud/Device-SMI’s past year of commit activity
    Python 10 Apache-2.0 1 1 2 Updated Jan 10, 2025

Top languages

Python

Most used topics

Loading…