| ## 2.10.1 | |
| * `hf-hub:org/model_id` support for loading models w/ config and weights in Hugging Face Hub | |
| ## 2.10.0 | |
| * Added a ViT-bigG-14 model. | |
| * Added an up-to-date example slurm script for large training jobs. | |
| * Added a option to sync logs and checkpoints to S3 during training. | |
| * New options for LR schedulers, constant and constant with cooldown | |
| * Fix wandb autoresuming when resume is not set | |
| * ConvNeXt `base` & `base_w` pretrained models added | |
| * `timm-` model prefix removed from configs | |
| * `timm` augmentation + regularization (dropout / drop-path) supported | |
| ## 2.9.3 | |
| * Fix wandb collapsing multiple parallel runs into a single one | |
| ## 2.9.2 | |
| * Fix braceexpand memory explosion for complex webdataset urls | |
| ## 2.9.1 | |
| * Fix release | |
| ## 2.9.0 | |
| * Add training feature to auto-resume from the latest checkpoint on restart via `--resume latest` | |
| * Allow webp in webdataset | |
| * Fix logging for number of samples when using gradient accumulation | |
| * Add model configs for convnext xxlarge | |
| ## 2.8.2 | |
| * wrapped patchdropout in a torch.nn.Module | |
| ## 2.8.1 | |
| * relax protobuf dependency | |
| * override the default patch dropout value in 'vision_cfg' | |
| ## 2.8.0 | |
| * better support for HF models | |
| * add support for gradient accumulation | |
| * CI fixes | |
| * add support for patch dropout | |
| * add convnext configs | |
| ## 2.7.0 | |
| * add multilingual H/14 xlm roberta large | |
| ## 2.6.1 | |
| * fix setup.py _read_reqs | |
| ## 2.6.0 | |
| * Make openclip training usable from pypi. | |
| * Add xlm roberta large vit h 14 config. | |
| ## 2.5.0 | |
| * pretrained B/32 xlm roberta base: first multilingual clip trained on laion5B | |
| * pretrained B/32 roberta base: first clip trained using an HF text encoder | |
| ## 2.4.1 | |
| * Add missing hf_tokenizer_name in CLIPTextCfg. | |
| ## 2.4.0 | |
| * Fix #211, missing RN50x64 config. Fix type of dropout param for ResNet models | |
| * Bring back LayerNorm impl that casts to input for non bf16/fp16 | |
| * zero_shot.py: set correct tokenizer based on args | |
| * training/params.py: remove hf params and get them from model config | |
| ## 2.3.1 | |
| * Implement grad checkpointing for hf model. | |
| * custom_text: True if hf_model_name is set | |
| * Disable hf tokenizer parallelism | |
| ## 2.3.0 | |
| * Generalizable Text Transformer with HuggingFace Models (@iejMac) | |
| ## 2.2.0 | |
| * Support for custom text tower | |
| * Add checksum verification for pretrained model weights | |
| ## 2.1.0 | |
| * lot including sota models, bfloat16 option, better loading, better metrics | |
| ## 1.2.0 | |
| * ViT-B/32 trained on Laion2B-en | |
| * add missing openai RN50x64 model | |
| ## 1.1.1 | |
| * ViT-B/16+ | |
| * Add grad checkpointing support | |
| * more robust data loader | |