-
Notifications
You must be signed in to change notification settings - Fork 27.5k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
How to correctly use
num_logits_to_keep
in model.generate()
?
bug
#35629
opened Jan 11, 2025 by
bwnjnOEI
2 of 4 tasks
static cache with mixtral will cause CUDA error: device-side assert triggered
bug
#35626
opened Jan 11, 2025 by
zyxiyy
1 of 4 tasks
The Phi model does not have lm_head bias after upgraded to v4.48.0
bug
#35625
opened Jan 11, 2025 by
yuxianq
2 of 4 tasks
Segmentation fault: address not mapped to object at address 0x100000007
bug
#35624
opened Jan 11, 2025 by
mrinaldi97
4 tasks done
Unsupported: hasattr SkipFunctionVariable when i compile the mixtral model with muti-gpus
bug
#35623
opened Jan 11, 2025 by
zyxiyy
4 tasks
The argument "dim" is gone from LlamaRotaryEmbedding initializer. Intentional?
bug
#35621
opened Jan 11, 2025 by
jeffhataws
4 tasks
from_pretrained fails to save weights.py and layers.py into cache, therefore fails to find them in cache
bug
#35619
opened Jan 11, 2025 by
openyk
4 tasks
Help Understanding Beam Search Scores in Hugging Face (LLaMA + LoRA)
bug
#35618
opened Jan 10, 2025 by
pratcooper
2 of 4 tasks
Better handeling of hardcoded component in PretrainedModel.from_pretrained.
bug
#35617
opened Jan 10, 2025 by
princethewinner
1 of 4 tasks
Trainer: TensorBoardCallback not working for "on_save" and "on_save_end" events
bug
#35612
opened Jan 10, 2025 by
vecorro
2 of 4 tasks
Trainer sets
state.best_model_checkpoint
even when it doesn't save there; leads to training crash
bug
#35609
opened Jan 10, 2025 by
tomaarsen
2 of 4 tasks
Prompt_ids feature causing repetitions and hallucinations
bug
#35603
opened Jan 10, 2025 by
vchagari
4 tasks
LlavaNextVideoProcessor -> TypeError: LlavaNextVideoProcessor.__call__() got an unexpected keyword argument 'legacy' (I have the fix)
bug
Core: Pipeline
Internals of the library; Pipeline.
VLM
#35602
opened Jan 10, 2025 by
inf3rnus
2 tasks
weird criterion to decide if needed to adjust the padding size
#35599
opened Jan 9, 2025 by
hyusterr
Inconsistent saving of tokenizer with custom code from HF hub vs. local directory
bug
#35597
opened Jan 9, 2025 by
mtm-cai
2 of 4 tasks
flash_attention_2 2.7.2.post1 seems to crash when using
torch.compile
and DataCollatorWithFlattening
bug
#35588
opened Jan 9, 2025 by
avishaiElmakies
4 tasks
Malformed config when saving & loading locally custom models
bug
#35584
opened Jan 9, 2025 by
Alicimo
2 of 4 tasks
Tokenizer outputs same offsets for different tokens.
bug
#35575
opened Jan 9, 2025 by
Fil-onto
4 tasks
Error occurs when using model.generate with Gemma2 in ZeRO3 environment
bug
#35572
opened Jan 9, 2025 by
jp1924
4 tasks
Transformers can create unconventional python module names when loading certain repositories
bug
#35570
opened Jan 8, 2025 by
kory
Any plans to integrate GTE model natively into transformers
New model
#35568
opened Jan 8, 2025 by
yaswanth19
2 tasks done
4.47.1 Hugging Face Trainer loss accumulated by sum instead of mean
bug
#35556
opened Jan 7, 2025 by
jdf-prog
4 tasks
Previous Next
ProTip!
Updated in the last three days: updated:>2025-01-08.