NemoClaw Knowledge Wiki

Home

❯

concepts

❯

4 bit quantisation

4-bit quantisation

Apr 19, 20261 min read

  • quantisation
  • 4-bit-precision
  • bit-depth

4-bit quantisation

This page is a stub awaiting enrichment.

Source Notes

  • 2026-04-23: Engine Survival: The Critical Role of Oil Pressure and Warning Lights
  • 2026-04-14: [[lab-notes/2026-04-14-Optimizing-AI-Costs-and-Privacy-with-Local-Open-Source-Models-and-Hybr|“But OpenClaw is expensive…“]]

Graph View

  • 4-bit quantisation
  • Source Notes

Backlinks

  • INDEX
  • Adam Lucek - quantisation of LLM
  • Best small LLM for local inference for instruction following
  • How does 4bit quantisation work
  • 4-bit-floating-point-fp4-training
  • computational-resources
  • gpu-clusters
  • model-compression
  • model-size
  • parameter-reduction
  • precision-reduction
  • AI & Agents
  • adam-lucek
  • gemini-ultra
  • gpt-4
  • julia-turc
  • llama-31-nemotron-70b
  • nemotron-70b
  • qwen
  • Adam Lucek - quantisation of LLM
  • Best small LLM for local inference for instruction following

Created with Quartz v4.5.2 © 2026

  • GitHub
  • Discord Community