This is a decensored version of nvidia/Nemotron-Terminal-14B, made using Heretic v1.2.0

Abliteration parameters

Parameter Value
direction_index per layer
attn.o_proj.max_weight 1.47
attn.o_proj.max_weight_position 33.87
attn.o_proj.min_weight 1.38
attn.o_proj.min_weight_distance 22.15
mlp.down_proj.max_weight 1.49
mlp.down_proj.max_weight_position 26.68
mlp.down_proj.min_weight 1.46
mlp.down_proj.min_weight_distance 18.14

Performance

Metric This model Original model (nvidia/Nemotron-Terminal-14B)
KL divergence 0.0005 0 (by definition)
Refusals 64/100 98/100

Nemotron-Terminal Model Family

Nemotron-Terminal is a family of models specialized for autonomous terminal interaction, fine-tuned from the Qwen3 (8B, 14B, and 32B). Developed by NVIDIA, these models utilize Nemotron-Terminal-Corpus, a large-scale open-source dataset for terminal tasks, to achieve performance that rivals frontier models many times their size.

Model Variants

We release the following variants of the Nemotron-Terminal family:

  • Nemotron-Terminal-8B
  • Nemotron-Terminal-14B
  • Nemotron-Terminal-32B

Performance on Terminal-Bench 2.0

The Nemotron-Terminal family demonstrates profound leaps in capability compared to the Qwen3 baselines across multiple specialized categories.

Model Size Base Accuracy Nemotron-Terminal Accuracy
Nemotron-Terminal-8B 8B 2.47% 13.0%
Nemotron-Terminal-14B 14B 4.04% 20.2%
Nemotron-Terminal-32B 32B 3.37% 27.4%

Usage

The models are trained using the Terminus 2 scaffolding and output a structured JSON format. For evaluation on Terminal Bench 2.0, we encourage using Terminus 2 scaffolding to maintain consistency with training.

Expected Output Format

{
  "analysis": "Analysis of the current terminal state...",
  "plan": "Step-by-step plan for the next command...",
  "commands": [
    {
      "keystrokes": "ls -la\n",
      "duration": 0.1
    }
  ],
  "task_complete": false
}

📜 Citation

If you use this dataset in your research, please cite the following work:

@misc{pi2026dataengineeringscalingllm,
      title={On Data Engineering for Scaling LLM Terminal Capabilities}, 
      author={Renjie Pi and Grace Lam and Mohammad Shoeybi and Pooya Jannaty and Bryan Catanzaro and Wei Ping},
      year={2026},
      eprint={2602.21193},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2602.21193}, 
}
Downloads last month
1,689
Safetensors
Model size
15B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for sasa2000/Nemotron-Terminal-14B-heretic

Finetuned
(1)
this model
Quantizations
2 models

Paper for sasa2000/Nemotron-Terminal-14B-heretic