-
-
Notifications
You must be signed in to change notification settings - Fork 3.9k
Commit 4fbbd7e
Bug fixes (#1473)
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update llama.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Fix: cast logits to float32 in cross_entropy_forward to prevent errors (#1254)
* Fix: cast logits to float32 in cross_entropy_forward to prevent errors
* Update cross_entropy_loss.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Throw error when inferencing longer than max_popsition_embeddings (#1236)
* Throw error when inferencing longer than max_popsition_embeddings without rope scaling
* Update llama.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* CLI now handles user input strings for dtype correctly (#1235)
Co-authored-by: root <root@ieeres.chu.cam.ac.uk>
* Update flex_attention.py
* Update _utils.py
* Update _utils.py
* Update flex_attention.py
* Update flex_attention.py
* Update loader.py
* Update loader.py
* Update flex_attention.py
* Update flex_attention.py
* Update flex_attention.py
* Update flex_attention.py
* Update _utils.py
* Update cross_entropy_loss.py
* Update _utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* triton_cast
* Update utils.py
* Qwen 2.5 Coder
* Fix/export mistral (#1281)
* Enhance install_python_non_blocking to handle protobuf installation and process management
* Revert "Enhance install_python_non_blocking to handle protobuf installation and process management"
This reverts commit f09974b.
* Set PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION to 'python' to address issue #1266
* Revert "Set PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION to 'python' to address issue #1266"
This reverts commit 9fc1307.
* Set PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION to 'python' to address issue #1266
* Update __init__.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* DOC Update - Update README.md with os.environ in example (#1269)
* Update README.md with os.environ in example
Added OS Environ in example to avoid device conflicts , for a user at least in jupyter notebook this allows to select GPU in a multi GPU setup.
As currently the unsloth init checks all GPU's and takes the first in the order which can be a issue when some GPU's are in use and the list still shows them. So to manually avoid this, this os config is required.
Small change but a bit time saver for those who straight away copies the tutorials
* Update README.md
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* fix/get_chat_template (#1246)
* Refactor `get_chat_template` to now support system message instead. It supposed to fix ollama tokenizer chattemplate to
* Remove type hinting
* Update chat_templates.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* fix/sft-trainer (#1276)
* Add patch for SFTTrainer to maintain backward compatibility with TRL changes
* Update trainer.py
* Update trainer.py
* Refactor trainer patch to maintain backward compatibility with TRL changes
* Update trainer.py
* Refactor trainer.py to exclude non-convertible trainers from backward compatibility patch
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update __init__.py
* Update trainer.py
* Update trainer.py
* Update trainer.py
* Update tokenizer_utils.py
* Update llama.py
* Fix #853
* fix/sfttrainer-compatibility (#1293)
* Refactor trainer.py to import SFTConfig directly and update UnslothTrainingArguments class inheritance
* Update trainer.py
* Update trainer.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update rms_layernorm.py
* Update rms_layernorm.py
* Gemma
* Update rms_layernorm.py
* Update gemma2.py
* Cut Cross Entropy
* Update llama.py
* Cut Cross Entropy
* Update llama.py
* Update llama.py
* Update llama.py
* Update __init__.py
* Update __init__.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update mapper.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* patch_fast_lora
* vision
* Update fast_lora.py
* Update _utils.py
* Update _utils.py
* Vision
* Update trainer.py
* Update save.py
* FastBaseVisionModel
* Update loader_utils.py
* Update vision.py
* Update loader.py
* Update vision.py
* Update loader.py
* Update vision.py
* Update _utils.py
* tokenizer_name
* Update loader.py
* Update vision.py
* Update save.py
* Update save.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update _utils.py
* Update loader.py
* kwargs
* logits
* Update llama.py
* Update llama.py
* Update llama.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* error
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update loader.py
* Update llama.py
* Update vision.py
* Update loader.py
* Old torch versions
* Update loader.py
* Update loader.py
* prints
* recheck
* Update loader.py
* Update loader.py
* Update _utils.py
* Update _utils.py
* Update mapper.py
* Feat/kto (#1316)
* Add PatchKTOTrainer and update model imports
* Update dpo.py
* Update __init__.py
* Delete unsloth/models/kto.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Fix orpo/dpo trainer (#1286)
* change the colab notebook for dpo zephyr and orpo
* use original tokenizer
* Update README.md
* Update README.md
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* skip modules
* Update vision.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Fix llama.cpp
* Update save.py
* Update save.py
* Update vision.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update _utils.py
* Update save.py
* Update save.py
* Update mapper.py
* modules
* Fix vision model tokenizer padding side. (#1384)
* Dynamic quants (#1379)
* typing
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* int64
* Update _utils.py
* Update cross_entropy_loss.py
* constexpr
* constexpr
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* CE
* Update cross_entropy_loss.py
* Update _utils.py
* Update llama.py
* Update _utils.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update utils.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* typing
* Update rope_embedding.py
* types
* Disable compiling
* Update _utils.py
* Update _utils.py
* Forward hook
* Update _utils.py
* Update llama.py
* Update _utils.py
* Update llama.py
* Update llama.py
* Update _utils.py
* Update pyproject.toml
* Update _utils.py
* Update llama.py
* CE Loss
* Update cross_entropy_loss.py
* Update _utils.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update llama.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Fix: cast logits to float32 in cross_entropy_forward to prevent errors (#1254)
* Fix: cast logits to float32 in cross_entropy_forward to prevent errors
* Update cross_entropy_loss.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Throw error when inferencing longer than max_popsition_embeddings (#1236)
* Throw error when inferencing longer than max_popsition_embeddings without rope scaling
* Update llama.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* CLI now handles user input strings for dtype correctly (#1235)
Co-authored-by: root <root@ieeres.chu.cam.ac.uk>
* Update flex_attention.py
* Update _utils.py
* Update _utils.py
* Update flex_attention.py
* Update flex_attention.py
* Update loader.py
* Update loader.py
* Update flex_attention.py
* Update flex_attention.py
* Update flex_attention.py
* Update flex_attention.py
* Update _utils.py
* Update cross_entropy_loss.py
* Update _utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* triton_cast
* Update utils.py
* Qwen 2.5 Coder
* Fix/export mistral (#1281)
* Enhance install_python_non_blocking to handle protobuf installation and process management
* Revert "Enhance install_python_non_blocking to handle protobuf installation and process management"
This reverts commit f09974b.
* Set PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION to 'python' to address issue #1266
* Revert "Set PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION to 'python' to address issue #1266"
This reverts commit 9fc1307.
* Set PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION to 'python' to address issue #1266
* Update __init__.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* DOC Update - Update README.md with os.environ in example (#1269)
* Update README.md with os.environ in example
Added OS Environ in example to avoid device conflicts , for a user at least in jupyter notebook this allows to select GPU in a multi GPU setup.
As currently the unsloth init checks all GPU's and takes the first in the order which can be a issue when some GPU's are in use and the list still shows them. So to manually avoid this, this os config is required.
Small change but a bit time saver for those who straight away copies the tutorials
* Update README.md
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* fix/get_chat_template (#1246)
* Refactor `get_chat_template` to now support system message instead. It supposed to fix ollama tokenizer chattemplate to
* Remove type hinting
* Update chat_templates.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* fix/sft-trainer (#1276)
* Add patch for SFTTrainer to maintain backward compatibility with TRL changes
* Update trainer.py
* Update trainer.py
* Refactor trainer patch to maintain backward compatibility with TRL changes
* Update trainer.py
* Refactor trainer.py to exclude non-convertible trainers from backward compatibility patch
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update __init__.py
* Update trainer.py
* Update trainer.py
* Update trainer.py
* Update tokenizer_utils.py
* Update llama.py
* Fix #853
* fix/sfttrainer-compatibility (#1293)
* Refactor trainer.py to import SFTConfig directly and update UnslothTrainingArguments class inheritance
* Update trainer.py
* Update trainer.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update rms_layernorm.py
* Update rms_layernorm.py
* Gemma
* Update rms_layernorm.py
* Update gemma2.py
* Cut Cross Entropy
* Update llama.py
* Cut Cross Entropy
* Update llama.py
* Update llama.py
* Update llama.py
* Update __init__.py
* Update __init__.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update mapper.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* patch_fast_lora
* vision
* Update fast_lora.py
* Update _utils.py
* Update _utils.py
* Vision
* Update trainer.py
* Update save.py
* FastBaseVisionModel
* Update loader_utils.py
* Update vision.py
* Update loader.py
* Update vision.py
* Update loader.py
* Update vision.py
* Update _utils.py
* tokenizer_name
* Update loader.py
* Update vision.py
* Update save.py
* Update save.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update vision.py
* Update _utils.py
* Update loader.py
* kwargs
* logits
* Update llama.py
* Update llama.py
* Update llama.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* error
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update loader.py
* Update llama.py
* Update vision.py
* Update loader.py
* Old torch versions
* Update loader.py
* Update loader.py
* prints
* recheck
* Update loader.py
* Update loader.py
* Update _utils.py
* Update _utils.py
* Update mapper.py
* Feat/kto (#1316)
* Add PatchKTOTrainer and update model imports
* Update dpo.py
* Update __init__.py
* Delete unsloth/models/kto.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Fix orpo/dpo trainer (#1286)
* change the colab notebook for dpo zephyr and orpo
* use original tokenizer
* Update README.md
* Update README.md
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* skip modules
* Update vision.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Fix llama.cpp
* Update save.py
* Update save.py
* Update vision.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update _utils.py
* Update save.py
* Update save.py
* Update mapper.py
* modules
---------
Co-authored-by: Edd <68678137+Erland366@users.noreply.github.com>
Co-authored-by: Datta Nimmaturi <datta.nimmaturi@nutanix.com>
Co-authored-by: Edwin Fennell <edwinfennell1@gmail.com>
Co-authored-by: root <root@ieeres.chu.cam.ac.uk>
Co-authored-by: Uday Girish Maradana <einsteingirish@gmail.com>
Co-authored-by: cell-dame <122996026+dame-cell@users.noreply.github.com>
* Update README.md
Unsloth Dynamic 4-bit Quantization Update
* Fix vision model tokenizer padding side.
* Update vision.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
Co-authored-by: Edd <68678137+Erland366@users.noreply.github.com>
Co-authored-by: Datta Nimmaturi <datta.nimmaturi@nutanix.com>
Co-authored-by: Edwin Fennell <edwinfennell1@gmail.com>
Co-authored-by: root <root@ieeres.chu.cam.ac.uk>
Co-authored-by: Uday Girish Maradana <einsteingirish@gmail.com>
Co-authored-by: cell-dame <122996026+dame-cell@users.noreply.github.com>
Co-authored-by: Michael Han <107991372+shimmyshimmer@users.noreply.github.com>
* Add citation section to README.md (#1377)
* Add citation section to README.md
* Update README.md
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Granite support (#1218)
* [WIP] Support for Granite
* Fixup inference
* Cleanup flex attention
* remove sliding window
* Use torch.add for residual multiplier
* Llama 3.3
* Update llama.py
* Update llama.py
* fullgraph
* Fix loader.py to work on Windows (#1453)
* Update README.md
Llama 3.3 + Reddit
* Update README.md
Apple ML Cross Entropy
* Update README.md
Removing double citation
* Fix loader.py to work on Windows
---------
Co-authored-by: Michael Han <107991372+shimmyshimmer@users.noreply.github.com>
* Update save.py warning message (#1425)
* Update README.md
Llama 3.3 + Reddit
* Update README.md
Apple ML Cross Entropy
* Update README.md
Removing double citation
* Update save.py warning message
---------
Co-authored-by: Michael Han <107991372+shimmyshimmer@users.noreply.github.com>
* Change _fix_chat_template in case a template has both endif and endfor (#1388)
* Update llama and derivatives to pass position embeddings explicitly for transformers v4.47+ (#1442)
* Update save.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Temp fix
* Update _utils.py
* Update _utils.py
* Update pyproject.toml
* Name Error Bug Fix - import from packaging.version import Version (#1468)
* Version
* Update pyproject.toml
* Update pyproject.toml
* Version
* Update pyproject.toml
* Update pyproject.toml
* dependencies
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Update mistral.py
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Update granite.py
* Update cohere.py
* Triton windows
* Update gemma2.py
* Update pyproject.toml
* Update _utils.py
* Update pyproject.toml
* Residual & LoRA
* Update loader.py
* Update loader.py
* Update loader.py
* Update loader.py
* Bug fix
* Update loader.py
* Update loader.py
* Update loader.py
* Update _utils.py
* Update loader.py
---------
Co-authored-by: Edd <68678137+Erland366@users.noreply.github.com>
Co-authored-by: Datta Nimmaturi <datta.nimmaturi@nutanix.com>
Co-authored-by: Edwin Fennell <edwinfennell1@gmail.com>
Co-authored-by: root <root@ieeres.chu.cam.ac.uk>
Co-authored-by: Uday Girish Maradana <einsteingirish@gmail.com>
Co-authored-by: cell-dame <122996026+dame-cell@users.noreply.github.com>
Co-authored-by: Zewen Shen <zewen.public@gmail.com>
Co-authored-by: Michael Han <107991372+shimmyshimmer@users.noreply.github.com>
Co-authored-by: Scott Phillips <polygonguru@gmail.com>
Co-authored-by: qingy1337 <qxli2@students.everettcc.edu>
Co-authored-by: Giulia Baldini <44327645+giuliabaldini@users.noreply.github.com>
Co-authored-by: Yonghye Kwon <developer.0hye@gmail.com>1 parent a240783 commit 4fbbd7eCopy full SHA for 4fbbd7e
Expand file treeCollapse file tree
2 files changed
+23
-1
lines changedOpen diff view settings
Collapse file
+21-1Lines changed: 21 additions & 1 deletion
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
12 | 12 | | |
13 | 13 | | |
14 | 14 | | |
15 | | - | |
| 15 | + | |
16 | 16 | | |
17 | 17 | | |
18 | 18 | | |
| |||
1093 | 1093 | | |
1094 | 1094 | | |
1095 | 1095 | | |
| 1096 | + | |
| 1097 | + | |
| 1098 | + | |
| 1099 | + | |
| 1100 | + | |
| 1101 | + | |
| 1102 | + | |
| 1103 | + | |
| 1104 | + | |
| 1105 | + | |
| 1106 | + | |
| 1107 | + | |
| 1108 | + | |
| 1109 | + | |
| 1110 | + | |
| 1111 | + | |
1096 | 1112 | | |
1097 | 1113 | | |
1098 | 1114 | | |
| |||
1130 | 1146 | | |
1131 | 1147 | | |
1132 | 1148 | | |
| 1149 | + | |
| 1150 | + | |
1133 | 1151 | | |
1134 | 1152 | | |
1135 | 1153 | | |
| |||
1174 | 1192 | | |
1175 | 1193 | | |
1176 | 1194 | | |
| 1195 | + | |
| 1196 | + | |
1177 | 1197 | | |
1178 | 1198 | | |
1179 | 1199 | | |
| |||
Collapse file
+2Lines changed: 2 additions & 0 deletions
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
470 | 470 | | |
471 | 471 | | |
472 | 472 | | |
| 473 | + | |
| 474 | + | |
473 | 475 | | |
474 | 476 | | |
475 | 477 | | |
| |||
0 commit comments