Custom GGUF quants of Metaβs Llama-3.2-Instruct's finetunes, where the Output Tensors are quantized to Q8_0 or F32 and the Embeddings are kept @F32
Joseph
Joseph717171
AI & ML interests
None yet
Recent Activity
liked
a model
about 14 hours ago
nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16
liked
a model
about 16 hours ago
nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-FP8
liked
a model
about 16 hours ago
RedHatAI/NVIDIA-Nemotron-3-Nano-30B-A3B-FP8