NemoClaw Knowledge Wiki

Tag: model-quantization

6 items with this tag.

  • Apr 27, 2026

    on-device deployment

    • deployment
    • edge-computing
    • llm
    • quantization
    • bitnet
    • on-device-computing
    • edge-machine-learning
    • model-quantization
    • low-bit-models
    • llm-efficiency
  • Apr 26, 2026

    ios-llm-implementation

    • iOS
    • LLM
    • Local-Inference
    • Mobile-AI
    • Mistral
    • ios-llm-implementation
    • on-device-ai
    • apple-silicon
    • model-quantization
  • Apr 24, 2026

    gpu-architecture

    • concept
    • gpu-architecture
    • nvidia-gpu
    • vram
    • large-language-models
    • model-quantization
    • local-inference
  • Apr 24, 2026

    instruction-following-tasks

    • concept
    • instruction-following
    • large-language-models
    • local-inference
    • model-quantization
  • Apr 11, 2026

    BitNet

    • BitNet
    • 1-bit-LLM
    • quantization
    • edge-computing
    • efficiency
    • bitnet-architecture
    • 1-bit-llm-deployment
    • edge-computing-efficiency
    • model-quantization
    • on-device-inference
  • Apr 11, 2026

    Bonsai

    • AI
    • LLM
    • Optimization
    • Edge-Computing
    • ai-efficiency
    • llm-compression
    • model-quantization
    • on-device-deployment

Created with Quartz v4.5.2 © 2026

  • GitHub
  • Discord Community