nm-testing/tinyllama-one-shot-w4a16-channel-compressed
Text Generation
• 1B • Updated
• 2
nm-testing/tinyllama-one-shot-dynamic-test
Text Generation
• 1B • Updated
• 13
nm-testing/tinyllama-one-shot-static-quant-test-compressed
Text Generation
• 1B • Updated
• 21
nm-testing/asym-w8w8-int8-static-per-tensor-tiny-llama
1B • Updated
• 2.75k
nm-testing/tinyllama-oneshot-w8a8-channel-dynamic-token-v2-asym
nm-testing/OLMoE-1B-7B-0924-Instruct-FP8
7B • Updated
• 26
nm-testing/DeepSeek-Coder-V2-Lite-Instruct-W8A8
16B • Updated
• 2
nm-testing/tinyllama-w8a16-dense
1B • Updated
• 764
nm-testing/tinyllama-w8a8-compressed
1B • Updated
• 753
nm-testing/tinyllama-w4a16-compressed
0.3B • Updated
• 617
nm-testing/tinyllama-fp8-dynamic-compressed
1B • Updated
• 507
nm-testing/SmolLM-1.7B-Instruct-quantized.w4a16
Text Generation
• 0.4B • Updated
nm-testing/SmolLM-360M-Instruct-quantized.w4a16
0.1B • Updated
nm-testing/SmolLM-135M-Instruct-quantized.w4a16
Text Generation
• 71.6M • Updated
• 1
nm-testing/Mixtral-8x7B-Instruct-v0.1-W4A16-channel-quantized
6B • Updated
• 659
nm-testing/Meta-Llama-3-8B-Instruct-fp8-compressed
nm-testing/Phi-3-mini-128k-instruct-FP8
4B • Updated
• 909
nm-testing/Mixtral-8x7B-Instruct-v0.1-FP8-quantized
47B • Updated
nm-testing/Mixtral-8x7B-Instruct-v0.1-W8A16-quantized
12B • Updated
• 649
nm-testing/Mixtral-8x7B-Instruct-v0.1-W4A16-quantized
6B • Updated
• 663
nm-testing/tinyllama-oneshot-w8a8-dynamic-token-v2-asym
Text Generation
• 1B • Updated
• 1
nm-testing/Qwen2-1.5B-Instruct-FP8W8
Text Generation
• 2B • Updated
• 2
nm-testing/Meta-Llama-3-8B-Instruct-W4A16-ACTORDER-compressed-tensors-test
Text Generation
• 2B • Updated
• 1
nm-testing/Meta-llama3-8b-Instruct-quant-FP8
Text Generation
• 8B • Updated
• 2
nm-testing/Meta-llama3-8b-Instruct-SmoothQuant-Fp8
Text Generation
• 8B • Updated
• 2
nm-testing/Meta-Llama-3-8B-Instruct-nonuniform-test
Text Generation
• 8B • Updated
• 12.3k
Text Generation
• 8B • Updated
• 1
nm-testing/Meta-Llama-3-8B-Instruct-Non-Uniform-compressed-tensors
Text Generation
• 8B • Updated
• 3
nm-testing/Meta-Llama-3-8B-Instruct-W8A8-FP8-Channelwise-compressed-tensors
Text Generation
• 8B • Updated
• 19
• 2
nm-testing/Meta-Llama-3-8B-Instruct-FP8-K-V
Text Generation
• 8B • Updated
• 1