Apertus-8B-Instruct-2509-LOGEQ-FP8_dynamic
This model applies Logarithmic Equalization (LogEQ) followed by
full FP8 dynamic quantization using the LLM Compressor framework.
What this means in practice
- LogEQ smooths and rescales activations using logarithmic statistics,
reducing activation outliers before quantization. - Both weights and activations are quantized to FP8 dynamic format
(E5M2 or E4M3 depending on layer type). - The quantized model is compatible with vLLM FP8 execution. data-free activation tracing.
Quantization Details
- Quantization Scheme: FP8_dynamic
- Preprocessing: LogarithmicEqualizationModifier
- Method: Data-free FP8 dynamic quantization
- Targets: All
Linearlayers - Ignored Layers:
lm_head(kept in FP16 for stability) - Tool: llm-compressor
- Downloads last month
- -
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for jgerster0/Apertus-8B-Instruct-2509-LOGEQ-FP8_dynamic
Base model
swiss-ai/Apertus-8B-2509
Finetuned
swiss-ai/Apertus-8B-Instruct-2509