patrickvonplaten commited on
Commit
8d27a0d
·
verified ·
1 Parent(s): 9592b53

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -235,7 +235,7 @@ And run the following code snippet:
235
  > While the checkpoint is serialized in FP8 format, there is currently a problem
236
  > with "true" FP8 inference. Hence the weights are automatically dequantized to BFloat16
237
  > as per [this config setting](https://huggingface.co/mistralai/Devstral-Small-2-24B-Instruct-2512/blob/main/config.json#L13).
238
- > Once the bug is fixed, we will by default run the model in "true" FP8.
239
 
240
  ```python
241
  import torch
 
235
  > While the checkpoint is serialized in FP8 format, there is currently a problem
236
  > with "true" FP8 inference. Hence the weights are automatically dequantized to BFloat16
237
  > as per [this config setting](https://huggingface.co/mistralai/Devstral-Small-2-24B-Instruct-2512/blob/main/config.json#L13).
238
+ > Once the bug is fixed, we will by default run the model in "true" FP8. Stay tuned by following [this issue](https://github.com/huggingface/transformers/issues/42746).
239
 
240
  ```python
241
  import torch