You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Unsloth: Merging 4bit and LoRA weights to 16bit...
Unsloth: Will use up to 25.65 out of 44.1 RAM for saving.
100%|██████████| 32/32 [00:36<00:00, 1.13s/it]
Unsloth: Saving tokenizer... Done.
Unsloth: Saving model... This might take 5 minutes for Llama-7b...
Done.
==((====))== Unsloth: Conversion from QLoRA to GGUF information
\ /| [0] Installing llama.cpp will take 3 minutes.
O^O/ _/ \ [1] Converting HF to GGUF 16bits will take 3 minutes.
\ / [2] Converting GGUF 16bits to ['q4_k_m'] will take 10 minutes each.
"-____-" In total, you will have to wait at least 16 minutes.
Unsloth: [0] Installing llama.cpp. This will take 3 minutes...
Unsloth: [1] Converting model at dir into bf16 GGUF format.
The output location will be /mnt/c/Users/H0/Desktop/Coding/Python/unsloth/dir/unsloth.BF16.gguf
This will take 3 minutes...
File ~/miniconda3/envs/unsloth_env/lib/python3.10/site-packages/unsloth/save.py:1093, in save_to_gguf(model_type, model_dtype, is_sentencepiece, model_directory, quantization_method, first_conversion, _run_installer) 1091 vocab_type = "spm,hfft,bpe" 1092 # Fix Sentencepiece model as well!
-> 1093 fix_sentencepiece_gguf(model_directory) 1094 else: 1095 vocab_type = "bpe"
File ~/miniconda3/envs/unsloth_env/lib/python3.10/site-packages/unsloth/tokenizer_utils.py:404, in fix_sentencepiece_gguf(saved_location) 398 """ 399 Fixes sentencepiece tokenizers which did not extend the vocabulary with
... 127 serialized_end=1347, 128 ) 129 _sym_db.RegisterEnumDescriptor(_TRAINERSPEC_MODELTYPE)
AttributeError: module 'google.protobuf.descriptor' has no attribute '_internal_create_key'
Output is truncated. View as a scrollable element or open in a text editor. Adjust cell output settings...
Have tried to install and uninstall the protobuf and protoc and making sure they are the same version without comflict issue.
Hello, was trying when using unsloth for the final local saving model with the line:
model.save_pretrained_gguf("dir", tokenizer, quantization_method = "q4_k_m")
and have the error:
Unsloth: Merging 4bit and LoRA weights to 16bit...
Unsloth: Will use up to 25.65 out of 44.1 RAM for saving.
100%|██████████| 32/32 [00:36<00:00, 1.13s/it]
Unsloth: Saving tokenizer... Done.
Unsloth: Saving model... This might take 5 minutes for Llama-7b...
Done.
==((====))== Unsloth: Conversion from QLoRA to GGUF information
\ /| [0] Installing llama.cpp will take 3 minutes.
O^O/ _/ \ [1] Converting HF to GGUF 16bits will take 3 minutes.
\ / [2] Converting GGUF 16bits to ['q4_k_m'] will take 10 minutes each.
"-____-" In total, you will have to wait at least 16 minutes.
Unsloth: [0] Installing llama.cpp. This will take 3 minutes...
Unsloth: [1] Converting model at dir into bf16 GGUF format.
The output location will be /mnt/c/Users/H0/Desktop/Coding/Python/unsloth/dir/unsloth.BF16.gguf
This will take 3 minutes...
AttributeError Traceback (most recent call last)
Cell In[18], line 1
----> 1 model.save_pretrained_gguf("dir", tokenizer, quantization_method = "q4_k_m")
File ~/miniconda3/envs/unsloth_env/lib/python3.10/site-packages/unsloth/save.py:1683, in unsloth_save_pretrained_gguf(self, save_directory, tokenizer, quantization_method, first_conversion, push_to_hub, token, private, is_main_process, state_dict, save_function, max_shard_size, safe_serialization, variant, save_peft_format, tags, temporary_location, maximum_memory_usage)
1680 is_sentencepiece_model = check_if_sentencepiece_model(self)
1682 # Save to GGUF
-> 1683 all_file_locations, want_full_precision = save_to_gguf(
1684 model_type, model_dtype, is_sentencepiece_model,
1685 new_save_directory, quantization_method, first_conversion, makefile,
1686 )
1688 # Save Ollama modelfile
1689 modelfile = create_ollama_modelfile(tokenizer, all_file_locations[0])
File ~/miniconda3/envs/unsloth_env/lib/python3.10/site-packages/unsloth/save.py:1093, in save_to_gguf(model_type, model_dtype, is_sentencepiece, model_directory, quantization_method, first_conversion, _run_installer)
1091 vocab_type = "spm,hfft,bpe"
1092 # Fix Sentencepiece model as well!
-> 1093 fix_sentencepiece_gguf(model_directory)
1094 else:
1095 vocab_type = "bpe"
File ~/miniconda3/envs/unsloth_env/lib/python3.10/site-packages/unsloth/tokenizer_utils.py:404, in fix_sentencepiece_gguf(saved_location)
398 """
399 Fixes sentencepiece tokenizers which did not extend the vocabulary with
...
127 serialized_end=1347,
128 )
129 _sym_db.RegisterEnumDescriptor(_TRAINERSPEC_MODELTYPE)
AttributeError: module 'google.protobuf.descriptor' has no attribute '_internal_create_key'
Output is truncated. View as a scrollable element or open in a text editor. Adjust cell output settings...
Have tried to install and uninstall the protobuf and protoc and making sure they are the same version without comflict issue.
Environment:
accelerate==1.1.1
aiohappyeyeballs==2.4.3
aiohttp==3.10.10
aiosignal==1.3.1
asttokens @ file:///home/conda/feedstock_root/build_artifacts/asttokens_1698341106958/work
async-timeout==4.0.3
attrs==24.2.0
bitsandbytes==0.44.1
Brotli @ file:///croot/brotli-split_1714483155106/work
certifi @ file:///home/conda/feedstock_root/build_artifacts/certifi_1725278078093/work/certifi
charset-normalizer @ file:///croot/charset-normalizer_1721748349566/work
comm @ file:///home/conda/feedstock_root/build_artifacts/comm_1710320294760/work
datasets==3.1.0
debugpy @ file:///croot/debugpy_1690905042057/work
decorator @ file:///home/conda/feedstock_root/build_artifacts/decorator_1641555617451/work
dill==0.3.8
docstring_parser==0.16
et-xmlfile==1.1.0
exceptiongroup @ file:///home/conda/feedstock_root/build_artifacts/exceptiongroup_1720869315914/work
executing @ file:///home/conda/feedstock_root/build_artifacts/executing_1725214404607/work
filelock @ file:///croot/filelock_1700591183607/work
frozenlist==1.5.0
fsspec==2024.9.0
gmpy2 @ file:///tmp/build/80754af9/gmpy2_1645455533097/work
hf_transfer==0.1.8
huggingface-hub==0.26.2
idna @ file:///croot/idna_1714398848350/work
importlib_metadata @ file:///home/conda/feedstock_root/build_artifacts/importlib-metadata_1726082825846/work
ipykernel @ file:///home/conda/feedstock_root/build_artifacts/ipykernel_1719845459717/work
ipython @ file:///home/conda/feedstock_root/build_artifacts/ipython_1729866374957/work
jedi @ file:///home/conda/feedstock_root/build_artifacts/jedi_1696326070614/work
Jinja2 @ file:///croot/jinja2_1730902924303/work
joblib==1.4.2
jupyter_client @ file:///home/conda/feedstock_root/build_artifacts/jupyter_client_1726610684920/work
jupyter_core @ file:///home/conda/feedstock_root/build_artifacts/jupyter_core_1727163409502/work
markdown-it-py==3.0.0
MarkupSafe @ file:///croot/markupsafe_1704205993651/work
matplotlib-inline @ file:///home/conda/feedstock_root/build_artifacts/matplotlib-inline_1713250518406/work
mdurl==0.1.2
mkl-service==2.4.0
mkl_fft @ file:///io/mkl313/mkl_fft_1730824109137/work
mkl_random @ file:///io/mkl313/mkl_random_1730823916628/work
mpmath @ file:///croot/mpmath_1690848262763/work
multidict==6.1.0
multiprocess==0.70.16
nest_asyncio @ file:///home/conda/feedstock_root/build_artifacts/nest-asyncio_1705850609492/work
networkx @ file:///croot/networkx_1720002482208/work
numpy @ file:///croot/numpy_and_numpy_base_1708638617955/work/dist/numpy-1.26.4-cp310-cp310-linux_x86_64.whl#sha256=d8cd837ed43e87f77e6efaa08e8de927ca030a1c9c5d04624432d6fb9a74a5ee
openpyxl @ file:///croot/openpyxl_1721752957391/work
packaging @ file:///home/conda/feedstock_root/build_artifacts/packaging_1718189413536/work
pandas==2.2.3
parso @ file:///home/conda/feedstock_root/build_artifacts/parso_1712320355065/work
peft==0.13.2
pexpect @ file:///home/conda/feedstock_root/build_artifacts/pexpect_1706113125309/work
pickleshare @ file:///home/conda/feedstock_root/build_artifacts/pickleshare_1602536217715/work
pillow @ file:///croot/pillow_1721059439630/work
platformdirs @ file:///home/conda/feedstock_root/build_artifacts/platformdirs_1726613481435/work
prompt_toolkit @ file:///home/conda/feedstock_root/build_artifacts/prompt-toolkit_1727341649933/work
propcache==0.2.0
protobuf==3.19.6
psutil @ file:///opt/conda/conda-bld/psutil_1656431268089/work
ptyprocess @ file:///home/conda/feedstock_root/build_artifacts/ptyprocess_1609419310487/work/dist/ptyprocess-0.7.0-py2.py3-none-any.whl
pure_eval @ file:///home/conda/feedstock_root/build_artifacts/pure_eval_1721585709575/work
pyarrow==18.0.0
Pygments @ file:///home/conda/feedstock_root/build_artifacts/pygments_1714846767233/work
PySocks @ file:///home/builder/ci_310/pysocks_1640793678128/work
python-dateutil @ file:///home/conda/feedstock_root/build_artifacts/python-dateutil_1709299778482/work
pytz==2024.2
PyYAML @ file:///croot/pyyaml_1728657952215/work
pyzmq @ file:///croot/pyzmq_1705605076900/work
regex==2024.11.6
requests @ file:///croot/requests_1730999120400/work
rich==13.9.4
safetensors==0.4.5
scikit-learn==1.5.2
scipy==1.14.1
sentencepiece==0.2.0
shtab==1.7.1
six @ file:///home/conda/feedstock_root/build_artifacts/six_1620240208055/work
stack-data @ file:///home/conda/feedstock_root/build_artifacts/stack_data_1669632077133/work
sympy==1.13.1
threadpoolctl==3.5.0
tokenizers==0.20.3
torch==2.5.1
torchaudio==2.5.1
torchvision==0.20.1
tornado @ file:///croot/tornado_1718740109488/work
tqdm==4.67.0
traitlets @ file:///home/conda/feedstock_root/build_artifacts/traitlets_1713535121073/work
transformers==4.46.2
triton==3.1.0
trl==0.12.0
typing_extensions @ file:///croot/typing_extensions_1715268824938/work
tyro==0.8.14
tzdata==2024.2
unsloth==2024.11.7
unsloth_zoo==2024.11.4
urllib3 @ file:///croot/urllib3_1727769808118/work
wcwidth @ file:///home/conda/feedstock_root/build_artifacts/wcwidth_1704731205417/work
xformers==0.0.28.post3
xxhash==3.5.0
yarl==1.17.1
zipp @ file:///home/conda/feedstock_root/build_artifacts/zipp_1726248574750/work
I have no idea what is wrong anymore
The text was updated successfully, but these errors were encountered: