Skip to main content
Process
toolsJanuary 15, 2025

JetBlock Optimizer

Deterministic inference for Nemotron 3 in ComfyUI — same seed, same output, every time.

Most LLM inference is nondeterministic — run the same prompt twice, get different results. JetBlock fixes that for Nemotron 3's hybrid architecture in ComfyUI.

**The Problem** Batch-size variance (not temperature) causes nondeterminism. When server load changes, batch sizes change, and outputs drift.

What JetBlock Does

  • Forces batch-invariant operations across Mamba-2, MoE, and attention layers
  • Three modes: strict (1.6x slower), standard (1.2x slower), relaxed (full speed)
  • Cascade control for reasoning budget between /think and /no_think modes

**Why It Matters** Reproducible outputs enable real on-policy reinforcement learning and consistent creative iteration.

ComfyUIDeterminismNemotronThinkingMachines