Update README.md
Browse files
README.md
CHANGED
|
@@ -235,7 +235,7 @@ And run the following code snippet:
|
|
| 235 |
> While the checkpoint is serialized in FP8 format, there is currently a problem
|
| 236 |
> with "true" FP8 inference. Hence the weights are automatically dequantized to BFloat16
|
| 237 |
> as per [this config setting](https://huggingface.co/mistralai/Devstral-Small-2-24B-Instruct-2512/blob/main/config.json#L13).
|
| 238 |
-
> Once the bug is fixed, we will by default run the model in "true" FP8.
|
| 239 |
|
| 240 |
```python
|
| 241 |
import torch
|
|
|
|
| 235 |
> While the checkpoint is serialized in FP8 format, there is currently a problem
|
| 236 |
> with "true" FP8 inference. Hence the weights are automatically dequantized to BFloat16
|
| 237 |
> as per [this config setting](https://huggingface.co/mistralai/Devstral-Small-2-24B-Instruct-2512/blob/main/config.json#L13).
|
| 238 |
+
> Once the bug is fixed, we will by default run the model in "true" FP8. Stay tuned by following [this issue](https://github.com/huggingface/transformers/issues/42746).
|
| 239 |
|
| 240 |
```python
|
| 241 |
import torch
|