|
## 2.24.0 |
|
|
|
* Fix missing space in error message |
|
* use model flag for normalizing embeddings |
|
* init logit_bias for non siglip pretrained models |
|
* Fix logit_bias load_checkpoint addition |
|
* Make CoCa model match CLIP models for logit scale/bias init |
|
* Fix missing return of "logit_bias" in CoCa.forward |
|
* Add NLLB-CLIP with SigLIP models |
|
* Add get_logits method and NLLB tokenizer |
|
* Remove the empty file src/open_clip/generation_utils.py |
|
* Update params.py: "BatchNorm" -> "LayerNorm" in the description string for "--lock-text-freeze-layer-norm" |
|
|
|
## 2.23.0 |
|
|
|
* Add CLIPA-v2 models |
|
* Add SigLIP models |
|
* Add MetaCLIP models |
|
* Add NLLB-CLIP models |
|
* CLIPA train code |
|
* Minor changes/fixes |
|
* Remove protobuf version limit |
|
* Stop checking model name when loading CoCa models |
|
* Log native wandb step |
|
* Use bool instead of long masks |
|
|
|
## 2.21.0 |
|
|
|
* Add SigLIP loss + training support |
|
* Add more DataComp models (B/16, B/32 and B/32@256) |
|
* Update default num workers |
|
* Update CoCa generation for `transformers>=4.31` |
|
* PyTorch 2.0 `state_dict()` compatibility fix for compiled models |
|
* Fix padding in `ResizeMaxSize` |
|
* Convert JIT model on state dict load for `pretrained='filename…'` |
|
* Other minor changes and fixes (typos, README, dependencies, CI) |
|
|
|
## 2.20.0 |
|
|
|
* Add EVA models |
|
* Support serial worker training |
|
* Fix Python 3.7 compatibility |
|
|
|
## 2.19.0 |
|
|
|
* Add DataComp models |
|
|
|
## 2.18.0 |
|
|
|
* Enable int8 inference without `.weight` attribute |
|
|
|
## 2.17.2 |
|
|
|
* Update push_to_hf_hub |
|
|
|
## 2.17.0 |
|
|
|
* Add int8 support |
|
* Update notebook demo |
|
* Refactor zero-shot classification code |
|
|
|
## 2.16.2 |
|
|
|
* Fixes for context_length and vocab_size attributes |
|
|
|
## 2.16.1 |
|
|
|
* Fixes for context_length and vocab_size attributes |
|
* Fix --train-num-samples logic |
|
* Add HF BERT configs for PubMed CLIP model |
|
|
|
## 2.16.0 |
|
|
|
* Add improved g-14 weights |
|
* Update protobuf version |
|
|
|
## 2.15.0 |
|
|
|
* Add convnext_xxlarge weights |
|
* Fixed import in readme |
|
* Add samples per second per gpu logging |
|
* Fix slurm example |
|
|
|
## 2.14.0 |
|
|
|
* Move dataset mixtures logic to shard level |
|
* Fix CoCa accum-grad training |
|
* Safer transformers import guard |
|
* get_labels refactoring |
|
|
|
## 2.13.0 |
|
|
|
* Add support for dataset mixtures with different sampling weights |
|
* Make transformers optional again |
|
|
|
## 2.12.0 |
|
|
|
* Updated convnext configs for consistency |
|
* Added input_patchnorm option |
|
* Clean and improve CoCa generation |
|
* Support model distillation |
|
* Add ConvNeXt-Large 320x320 fine-tune weights |
|
|
|
## 2.11.1 |
|
|
|
* Make transformers optional |
|
* Add MSCOCO CoCa finetunes to pretrained models |
|
|
|
## 2.11.0 |
|
|
|
* coca support and weights |
|
* ConvNeXt-Large weights |
|
|
|
## 2.10.1 |
|
|
|
* `hf-hub:org/model_id` support for loading models w/ config and weights in Hugging Face Hub |
|
|
|
## 2.10.0 |
|
|
|
* Added a ViT-bigG-14 model. |
|
* Added an up-to-date example slurm script for large training jobs. |
|
* Added a option to sync logs and checkpoints to S3 during training. |
|
* New options for LR schedulers, constant and constant with cooldown |
|
* Fix wandb autoresuming when resume is not set |
|
* ConvNeXt `base` & `base_w` pretrained models added |
|
* `timm-` model prefix removed from configs |
|
* `timm` augmentation + regularization (dropout / drop-path) supported |
|
|
|
## 2.9.3 |
|
|
|
* Fix wandb collapsing multiple parallel runs into a single one |
|
|
|
## 2.9.2 |
|
|
|
* Fix braceexpand memory explosion for complex webdataset urls |
|
|
|
## 2.9.1 |
|
|
|
* Fix release |
|
|
|
## 2.9.0 |
|
|
|
* Add training feature to auto-resume from the latest checkpoint on restart via `--resume latest` |
|
* Allow webp in webdataset |
|
* Fix logging for number of samples when using gradient accumulation |
|
* Add model configs for convnext xxlarge |
|
|
|
## 2.8.2 |
|
|
|
* wrapped patchdropout in a torch.nn.Module |
|
|
|
## 2.8.1 |
|
|
|
* relax protobuf dependency |
|
* override the default patch dropout value in 'vision_cfg' |
|
|
|
## 2.8.0 |
|
|
|
* better support for HF models |
|
* add support for gradient accumulation |
|
* CI fixes |
|
* add support for patch dropout |
|
* add convnext configs |
|
|
|
|
|
## 2.7.0 |
|
|
|
* add multilingual H/14 xlm roberta large |
|
|
|
## 2.6.1 |
|
|
|
* fix setup.py _read_reqs |
|
|
|
## 2.6.0 |
|
|
|
* Make openclip training usable from pypi. |
|
* Add xlm roberta large vit h 14 config. |
|
|
|
## 2.5.0 |
|
|
|
* pretrained B/32 xlm roberta base: first multilingual clip trained on laion5B |
|
* pretrained B/32 roberta base: first clip trained using an HF text encoder |
|
|
|
## 2.4.1 |
|
|
|
* Add missing hf_tokenizer_name in CLIPTextCfg. |
|
|
|
## 2.4.0 |
|
|
|
* Fix #211, missing RN50x64 config. Fix type of dropout param for ResNet models |
|
* Bring back LayerNorm impl that casts to input for non bf16/fp16 |
|
* zero_shot.py: set correct tokenizer based on args |
|
* training/params.py: remove hf params and get them from model config |
|
|
|
## 2.3.1 |
|
|
|
* Implement grad checkpointing for hf model. |
|
* custom_text: True if hf_model_name is set |
|
* Disable hf tokenizer parallelism |
|
|
|
## 2.3.0 |
|
|
|
* Generalizable Text Transformer with HuggingFace Models (@iejMac) |
|
|
|
## 2.2.0 |
|
|
|
* Support for custom text tower |
|
* Add checksum verification for pretrained model weights |
|
|
|
## 2.1.0 |
|
|
|
* lot including sota models, bfloat16 option, better loading, better metrics |
|
|
|
## 1.2.0 |
|
|
|
* ViT-B/32 trained on Laion2B-en |
|
* add missing openai RN50x64 model |
|
|
|
## 1.1.1 |
|
|
|
* ViT-B/16+ |
|
* Add grad checkpointing support |
|
* more robust data loader |
|
|