huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
PythonApache-2.0
Issues
- 4
- 5
`import transformers` accidentally initializing both torch and jax/xla at startup time
#30226 opened by stellaraccident - 2
- 1
load and train with bf16,saved torch_dtype is float32
#30305 opened by gyangfan - 3
Error at the generation stage by MusicGen stereo model
#30217 opened by ElizavetaSedova - 3
- 3
Idefics2 raises shape error in the backward pass with gradient_checkpointing
#30301 opened by chenzizhao - 3
Importing `CLIPVisionModelWithProjection` crashes with `AttributeError: 'NoneType' object has no attribute 'dumps'`
#30261 opened by gml16 - 2
- 3
There is a probability that a bug will be triggered when tracing the llama model: torch.fx.proxy.TraceError: symbolically traced variables cannot be used as inputs to control flow
#30244 opened by sean-wade - 2
modeling_t5 incompatible with multiprocessing
#30280 opened by rangehow - 4
RuntimeError: Failed to import transformers.pipelines because of the following error (look up to see its traceback): name 'LRScheduler' is not defined
#30194 opened by saqib772 - 2
Image Preprocess in Clip
#30282 opened by lesjie-wen - 1
Badam support
#30308 opened by Theodotus1243 - 3
Why cast to float32 in this line?
#30186 opened by brynhayder - 3
GPT2 model getting NaN logits with MPS device
#30296 opened by uwu-420 - 1
Can't load tokenizer
#30309 opened by wb-zr707791 - 1
Unexpected behavior in DonutProcessor.token2json with strings containing multilines (\n)
#30272 opened by issahammoud - 2
- 1
Skipping Layers in a forward pass
#30286 opened by IsNoobgrammer - 1
Token merging for LLM Inference
#30292 opened by samchaineau - 0
Constrained beam search in chat
#30295 opened by esaks1 - 3
llava-next demo / tutorial code does not work
#30294 opened by bghira - 2
Reporting a vulnerability
#30279 opened by immi19 - 6
Nondeterministic behavior from GPT with MPS backend
#30229 opened by riklopfer - 0
Jamba-v01 Model + Deepspeed Zero3 lead to "RuntimeError: Detected mismatch between collectives on ranks."
#30277 opened by KaiQiangSong - 2
- 2
show we add <image> token to the prompt if we use llava-hf from transformers?
#30284 opened by DefTruth - 1
Using SparseAdam with LLaMA
#30276 opened by Aradhye2002 - 5
BLOOM embeddings should specify padding_idx
#30255 opened by PaulLerner - 5
[BUG] Load StarCoder2 AWQ using Transformers
#30225 opened by TechxGenus - 1
Enhance HfArgumentParser with Dict command-line parser
#30204 opened by xu-song - 1
FSDP Doesn't Work with model.generate()
#30228 opened by QiyaoWei - 3
- 2
- 1
LlamaRMSNorm() Dtype Casting Error
#30236 opened by Ritz111 - 3
[i18n-PL] Translating docs to Polish
#30240 opened by RooTender - 0
MPS memory leak?
#30246 opened by tonystratum - 3
Can't save checkpoint with shared tensors
#30202 opened by AndrewRyanChama - 0
- 2
- 1
Couldn't connect to `https://huggingface.co`.
#30245 opened by chrysanthemum-boy - 0
Trying to stack tensors from different devices in `_pad_to_max_length` in Whisper batched inference
#30223 opened by cifkao - 1
- 1
`BartForConditionalGeneration` has no attribute `shared`
#30206 opened by shivanraptor - 4
Providing several prompt_images and prompt_masks to seggpt leads to RuntimeError
#30196 opened by MSchnei - 0
Create panoptic segmentation task guide
#30214 opened by NielsRogge - 0
- 0
EncoderDecoderModel with XLM-R
#30211 opened by Bachstelze - 1
Tranformers documentation translation to Persian
#30191 opened by AmirHosseinMaleki