Port tokenization for the multilingual model by elyase · Pull Request #2 · huggingface/transformers (original) (raw)

…uggingface#29145 (#1)

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Ren Xuancheng jklj077@users.noreply.github.com Co-authored-by: renxuancheng.rxc renxuancheng.rxc@alibaba-inc.com Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Your Name you@example.com

I want to train dinov2 with bf16 but I get the following error in https://github.com/huggingface/transformers/blob/bc72b4e2cdcbc80d5f56731f35dbc9c18b4c8de6/src/transformers/models/dinov2/modeling_dinov2.py#L635:

RuntimeError: Input type (float) and bias type (c10::BFloat16) should be the same

Since the input dtype is torch.float32, the parameter dtype has to be torch.float32...

@LZHgrla and I checked the code of clip vision encoder and found there is an automatic dtype transformation (https://github.com/huggingface/transformers/blob/bc72b4e2cdcbc80d5f56731f35dbc9c18b4c8de6/src/transformers/models/clip/modeling_clip.py#L181-L182).

So I add similar automatic dtype transformation to modeling_dinov2.py.

Fix sparse_step = 1

I case sparse_step = 1, the current code will not work.

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Co-authored-by: Sanchit Gandhi 93869735+sanchit-gandhi@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: Sanchit Gandhi 93869735+sanchit-gandhi@users.noreply.github.com Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Co-authored-by: Sanchit Gandhi 93869735+sanchit-gandhi@users.noreply.github.com

Co-authored-by: Joao Gante joaofranciscocardosogante@gmail.com Co-authored-by: Sanchit Gandhi 93869735+sanchit-gandhi@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Sanchit Gandhi 93869735+sanchit-gandhi@users.noreply.github.com Co-authored-by: Joao Gante joaofranciscocardosogante@gmail.com Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

First draft

Fix call to tokenizer.add_tokens

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Fix missing bbox in LayoutLM signature


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Add tip to custom model docs

Update modeling_gpt_neox.py

Only import class if sp available

Co-authored-by: Pashmina Cameron 11311835+pashminacameron@users.noreply.github.com

get default device through PartialState().default_device as it has been officially released

Co-authored-by: Nicolas Patry patry.nicolas@protonmail.com


Co-authored-by: Nicolas Patry patry.nicolas@protonmail.com Co-authored-by: sanchit-gandhi sanchit@huggingface.co

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Fix copy/paste error msg typo

Fix windows err

dataloader_prefetch_factor works only when data is loaded in a different process as the main one. This commit adds the necessary checks to avoid having prefetch_factor set when there is no such process.

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

convert token id to list in .decode()

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

in case obj is a list, check the items in the list


Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Fixing grammatical errors in the text

https://github.com/huggingface/transformers/issues/28438

Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com

Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com

Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com


Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com

Add missing space in warning

Update backbones.md

fix typo:

from:

"model = TFAutoModelForQuestionAnswering("distilbert-base-uncased")"

to: model = TFAutoModelForQuestionAnswering.from_pretrained("distilbert-base-uncased")

align dtype

The documentation says "We refer to this Model parallelism as “Vertical” because of how models are typically visualized.", but then visualizes the model horizontally. This change visualizes the model indeed vertically.


Co-authored-by: ydshieh ydshieh@users.noreply.github.com


Co-authored-by: Your Name you@example.com

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

adjust ImageProcessor link to working target (same as in lower section of file)

Initialize _tqdm_active with hf_hub_utils.are_progress_bars_disabled() to respect HF_HUB_DISABLE_PROGRESS_BARS

It seems like enable_progress_bar() and disable_progress_bar() sync up with huggingface_hub, but the initial value is always True. This changes will make sure the user's preference is respected implicity on initialization.

fix

Co-authored-by: ydshieh ydshieh@users.noreply.github.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

[Siglip] protect from imports if sentencepiece not installed


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Signed-off-by: Wesley M. Gifford wmgifford@us.ibm.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Signed-off-by: Wesley M. Gifford wmgifford@us.ibm.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Signed-off-by: Wesley M. Gifford wmgifford@us.ibm.com

Signed-off-by: Wesley M. Gifford wmgifford@us.ibm.com


Signed-off-by: Wesley M. Gifford wmgifford@us.ibm.com Co-authored-by: Wesley M. Gifford wmgifford@us.ibm.com Co-authored-by: Kashif Rasul kashif.rasul@gmail.com Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

This reverts commit 0557e2cc9effa3a41304322032239a3874b948a7.

Signed-off-by: ThibaultLengagne thibaultl@padok.fr

Signed-off-by: ThibaultLengagne thibaultl@padok.fr

Signed-off-by: ThibaultLengagne thibaultl@padok.fr

Signed-off-by: ThibaultLengagne thibaultl@padok.fr


Signed-off-by: ThibaultLengagne thibaultl@padok.fr Co-authored-by: Sarapuce alexandreh@padok.fr

Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com


Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

Add _no_split_modules to CLIPModel

Co-authored-by: Marc Sun 57196510+SunMarc@users.noreply.github.com

Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: younesbelkada younesbelkada@gmail.com Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com Co-authored-by: Marc Sun 57196510+SunMarc@users.noreply.github.com Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Update _toctree.yml

guard sdpa on torch>=2.0

fix

Co-authored-by: ydshieh ydshieh@users.noreply.github.com

This reverts commit de0d6592c35dc39ae8b5a616c27285db28262d06.


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

fix bnb slow tests

Modified MLflow logging metrics from synchronous to asynchronous

Co-authored-by: codiceSpaghetti alessio.ser@hotmail.it

remove check_packages_compatibility logic

Co-authored-by: Sanchit Gandhi 93869735+sanchit-gandhi@users.noreply.github.com


Co-authored-by: Sanchit Gandhi 93869735+sanchit-gandhi@users.noreply.github.com


Co-authored-by: Lysandre lysandre@huggingface.co

This reverts commit 0e2905bf2236ec323163fc1a9f0c016b21aa8b8f.

This reverts commit fb17b6187ac5d16da7c461e1130514dc3d137a43.


Co-authored-by: ydshieh ydshieh@users.noreply.github.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

The test is correctly picking up redundant keys in _keys_to_ignore_on_load_missing.

fix

Co-authored-by: ydshieh ydshieh@users.noreply.github.com

This reverts commit 4041c86c29248f1673e8173b677c20b5a4511358.

This reverts commit 1cf345e34d3cc0e09eb800d9895805b1dd9b474d.

This reverts commit 6056ffbb1eddc3cb3a99e4ebb231ae3edf295f5b.

This reverts commit 24e25d0a14891241de58a0d86f817d0b5d2a341f.

Co-authored-by: p_spozzhang p_spozzhang@tencent.com

Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com

Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com

Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com


Co-authored-by: Steven Liu 59462357+stevhliu@users.noreply.github.com

Add an explicit none-check, since token ids can be 0


Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com

try

Co-authored-by: ydshieh ydshieh@users.noreply.github.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

use -v in pytest

Co-authored-by: ydshieh ydshieh@users.noreply.github.com

support FSDP+PEFT

Mark test as flaky

Bump dash in /examples/research_projects/decision_transformer

Bumps dash from 2.3.0 to 2.15.0.


updated-dependencies:

Signed-off-by: dependabot[bot] support@github.com Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>

Reuse trainer.create_scheduler to create scheduler for deepspeed

Fix bad doc: replace save with logging

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

update guidelines

Co-authored-by: Omar Sanseviero osanseviero@gmail.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Omar Sanseviero osanseviero@gmail.com Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Eugen Ajechiloae eugenajechiloae@gmail.com

Resolve https://github.com/huggingface/transformers/issues/26961

Bump cryptography in /examples/research_projects/decision_transformer

Bumps cryptography from 41.0.2 to 42.0.0.


updated-dependencies:

Signed-off-by: dependabot[bot] support@github.com Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>

Update project names and repository links in awesome-transformers


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Revert "[WIP] Hard error when ignoring tensors. (#27484)"

This reverts commit 2da28c4b41bba23969a8afe97c3dfdcbc47a57dc.


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Fix backticks in code blocks and documentation links


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

Fixed the documentation for logging_first_step by removing evaluate.

Fix our schedule import so it works for older versions of Keras

Co-authored-by: Joao Gante joaofranciscocardosogante@gmail.com


Co-authored-by: Joao Gante joaofranciscocardosogante@gmail.com


Co-authored-by: ydshieh ydshieh@users.noreply.github.com

add npu device for pipeline

Co-authored-by: unit_test test@unit.com

Fix placement of tilde character

Fix utf-8 yaml in marian conversion

Co-authored-by: fxmarty 9808326+fxmarty@users.noreply.github.com Co-authored-by: Younes Belkada 49240599+younesbelkada@users.noreply.github.com Co-authored-by: Joao Gante joaofranciscocardosogante@gmail.com

Remove dead code

Co-authored-by: Joao Gante joaofranciscocardosogante@gmail.com


Co-authored-by: Joao Gante joaofranciscocardosogante@gmail.com

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Check type via isinstance

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com


Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

updated docstring with vqa alias

Updated datasets requirements. Need a package version >= 2.14.0

Continue to initialize tied output_embeddings if it has a bias term

The bias term is not tied, and so will need to be initialized accordingly.

clean up remaining tmp checkpoint dir

Signed-off-by: woshiyyya xiaoyunxuan1998@gmail.com

Add language identifiers to code blocks

Add video section

Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

dale.david@mail.ru


Co-authored-by: amyeroberts 22614925+amyeroberts@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com

Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com


Co-authored-by: Arthur 48595927+ArthurZucker@users.noreply.github.com