[AUDIT-REF-205] ATLAS Ternary Packer - 16x GGUF Compression#548
[AUDIT-REF-205] ATLAS Ternary Packer - 16x GGUF Compression#548xxxn3m3s1sxxx wants to merge 1 commit intomicrosoft:mainfrom
Conversation
- Detects ternary states (-1, 0, +1) in BitNet weights - Quantizes floats to ternary (-1, 0, +1) - Packs to 2-bit (16x compression) - Round-trip integrity verified Ref: NOMAD Node microsoft#205, #1027
|
@microsoft-github-policy-service agree |
VRAM Benchmark Report| Model | Float32 | 2-Bit Packed | Savings | Compression: 16x (float32 to 2-bit) ATLAS NOMAD-1 Research | Node #205 |
Compatibility NoteUpdate: GGUF now has TQ1_0 (ternary quantization) as standard. Our approach is designed to complement this:
Our packer can pre-process weights for optimal TQ1_0 encoding, reducing quantization loss at the tensor level before block quantization. Reference implementation remains compatible with GGUF standard. |
Official GGUF Standard ReferenceOur approach aligns with the official GGUF ternary quantization specification: Source: ggml-org/llama.cpp@9bc6db2 (merged Sept 2024)
Our tensor-level pre-processing can optimize weights BEFORE they enter the TQ1_0 block quantization pipeline. |
Problem
Current GGUF converter treats BitNet ternary weights {-1, 0, +1} as 16-bit floats, wasting VRAM.
Evidence
Research: NOMAD Node #205 (Ternary-Logic-Integration) & #1027 (BitNet to GGUF Converter)
Solution
Add ATLAS ternary packer module with 2-bit packing:
Impact
Testing
Round-trip integrity verified: 100%