NemoClaw Knowledge Wiki

Home

❯

concepts

❯

nvidia h100

nvidia-h100

Apr 24, 20261 min read

  • concept
  • qwen
  • quantization
  • llm-performance
  • ollama
  • memory-trade-off

Nvidia H100

Source Notes

  • 2026-04-19: Qwen 3.6-35B Full Precision vs. Ollama Quantized Performance-Memory Trade-off Clip title: Comparing Full Precision vs Ollama Version of Qwen3.6-35B-A3B Locally Author / channel: Fahd Mirza URL: https://www.youtube.com/watch?v=RlGppgMDl9k Summary This video prov (Qwen 36-35B Full Precision vs Ollama Quantized Performance-Memory Trade-off)

Graph View

  • Nvidia H100
  • Source Notes

Backlinks

  • INDEX
  • Qwen 36-35B Full Precision vs Ollama Quantized Performance-Memory Trade-off
  • Fahd Mirza - fine tuning weights of OSS-20B
  • MedGemma 27B - Fahd Merza
  • New Qwen agentic local llm
  • Tools & Platforms
  • MedGemma 27B - Fahd Merza
  • New Qwen agentic local llm
  • Gemma 4-E2B LLM Fine-Tuning: Custom Dataset & Unsloth Local Tutorial
  • Gemma 4-E2B LLM Fine-Tuning: Custom Dataset & Unsloth Local Tutorial
  • MiniMax M27 Open Source LLM Technical Overview and Deployment Summary

Created with Quartz v4.5.2 © 2026

  • GitHub
  • Discord Community