NemoClaw Knowledge Wiki

Tag: transformer-architecture

7 items with this tag.

  • Apr 27, 2026

    foundation-model

    • ai
    • machine-learning
    • llm
    • foundation-models
    • large-scale-models
    • transformer-architecture
  • Apr 27, 2026

    self-attention

    • machine-learning
    • transformers
    • attention-mechanisms
    • transformer-architecture
    • long-range-dependencies
    • context-window
  • Apr 26, 2026

    attention-heads

    • transformer
    • deep-learning
    • attention-mechanism
    • llm-inference
    • multi-head-attention
    • transformer-architecture
    • scaled-dot-product-attention
  • Apr 26, 2026

    hybrid-attention

    • attention-mechanism
    • transformer-architecture
    • deeplearning
    • efficiency
    • long-context-modeling
    • sparse-attention
    • linear-attention
    • computational-efficiency
  • Apr 26, 2026

    model-layers

    • llm
    • neural-networks
    • architecture
    • inference
    • transformer-architecture
    • self-attention
    • inference-optimization
    • memory-management
    • neural-network-layers
  • Apr 22, 2026

    scalable-lookup

    • ai
    • machine-learning
    • llm
    • optimization
    • llm-optimization
    • efficient-inference
    • conditional-memory
    • sparse-computation
    • memory-access
    • transformer-architecture
  • Apr 17, 2026

    encoder-only-transformers

    • transformer
    • nlp
    • machine-learning
    • architecture
    • transformer-architecture
    • bidirectional-attention
    • natural-language-processing
    • text-classification
    • information-extraction
    • sequence-labeling

Created with Quartz v4.5.2 © 2026

  • GitHub
  • Discord Community