Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
18.4
TFLOPS
13
1
56
Ruslan
uzvisa
Follow
Hastagaras's profile picture
Gargaz's profile picture
gaytri1111's profile picture
8 followers
·
85 following
AI & ML interests
None yet
Recent Activity
new
activity
1 day ago
Qwen/Qwen3.6-35B-A3B:
how to enable non-thinking mode of this model in llama.cpp?
reacted
to
eaddario
's
post
with 👍
4 days ago
Experimental global target bits‑per‑weight quantization of Qwen/Qwen3.6-27B and Qwen/Qwen3.6-35B-A3B. Unlike standard llama.cpp quantizations that rely on fixed type heuristics (e.g., Q4_K_M), the Target BPW approach optimizes per-tensor precision where it matters the most, and produces high quality models that meet a precise global file size target. Key Advantages: - VRAM Maximization: Can generate high quality models sized exactly to fit hardware constraints (e.g., fitting the model into exactly 24GB VRAM). - Data-Driven Precision: Quantization mix is determined by actual weight error sensitivity rather than hardcoded rules, often yielding better PPL/KLD size trade-offs. Full benchmarks (PPL, KLD, ARC, GPQA, MMLU, etc.) and methodology in the models' cards. https://huggingface.co/eaddario/Qwen3.6-27B-GGUF https://huggingface.co/eaddario/Qwen3.6-35B-A3B-GGUF
reacted
to
eaddario
's
post
with 🔥
4 days ago
Experimental global target bits‑per‑weight quantization of Qwen/Qwen3.6-27B and Qwen/Qwen3.6-35B-A3B. Unlike standard llama.cpp quantizations that rely on fixed type heuristics (e.g., Q4_K_M), the Target BPW approach optimizes per-tensor precision where it matters the most, and produces high quality models that meet a precise global file size target. Key Advantages: - VRAM Maximization: Can generate high quality models sized exactly to fit hardware constraints (e.g., fitting the model into exactly 24GB VRAM). - Data-Driven Precision: Quantization mix is determined by actual weight error sensitivity rather than hardcoded rules, often yielding better PPL/KLD size trade-offs. Full benchmarks (PPL, KLD, ARC, GPQA, MMLU, etc.) and methodology in the models' cards. https://huggingface.co/eaddario/Qwen3.6-27B-GGUF https://huggingface.co/eaddario/Qwen3.6-35B-A3B-GGUF
View all activity
Organizations
None yet
uzvisa
's activity
All
Models
Datasets
Spaces
Buckets
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
liked
a model
about 2 months ago
Tesslate/OmniCoder-9B
Text Generation
•
Updated
Mar 13
•
8.86k
•
614
liked
10 models
2 months ago
steampunque/Qwen3-VL-8B-Instruct-MP-GGUF
8B
•
Updated
Feb 18
•
55
•
2
steampunque/gemma-3-12b-it-MP-GGUF
12B
•
Updated
Feb 18
•
28
•
1
steampunque/Ministral-3-8B-Instruct-2512-MP-GGUF
8B
•
Updated
Feb 18
•
16
•
1
steampunque/Qwen2.5-Coder-14B-Instruct-MP-GGUF
15B
•
Updated
Feb 18
•
19
•
1
tencent/HY-MT1.5-7B-GGUF
Translation
•
8B
•
Updated
Jan 7
•
6.79k
•
51
allura-forge/Llama-3.3-8B-Instruct
Updated
Dec 31, 2025
•
438
•
204
mradermacher/Nanbeige-4.1-Python-DeepThink-3B-GGUF
4B
•
Updated
Feb 18
•
44
•
1
deltakitsune/Nanbeige-4.1-Python-DeepThink-3B
Text Generation
•
4B
•
Updated
Feb 16
•
709
•
7
TheDrummer/Tiger-Gemma-12B-v3-GGUF
13B
•
Updated
Jul 9, 2025
•
836
•
14
MuXodious/Nanbeige4.1-3B-PaperWitch-heresy
Text Generation
•
4B
•
Updated
Feb 19
•
19
•
4
liked
9 models
3 months ago
gabriellarson/WEBGEN-4B-Preview-GGUF
Text Generation
•
4B
•
Updated
Sep 2, 2025
•
237
•
20
TheDrummer/Rocinante-X-12B-v1
12B
•
Updated
Jan 25
•
717
•
72
TheDrummer/Rivermind-Lux-12B-v1
12B
•
Updated
May 6, 2025
•
7
•
21
t-tech/T-lite-it-2.1
Text Generation
•
8B
•
Updated
Dec 23, 2025
•
4.2k
•
•
19
Tesslate/UIGEN-X-8B
Text Generation
•
8B
•
Updated
Jul 18, 2025
•
63
•
•
63
Tesslate/WEBGEN-4B-Preview
Text Generation
•
Updated
Sep 2, 2025
•
101
•
•
86
Nanbeige/Nanbeige4.1-3B
Text Generation
•
4B
•
Updated
Mar 25
•
233k
•
•
1.1k
TeichAI/Qwen3-8B-DeepSeek-v3.2-Speciale-Distill-GGUF
8B
•
Updated
Dec 10, 2025
•
10.3k
•
23
TeichAI/Qwen3-8B-Claude-Sonnet-4.5-Reasoning-Distill-GGUF
8B
•
Updated
Nov 16, 2025
•
1.83k
•
17
Load more