huggingface-transformers Questions

4

Is there any way of getting sentence embeddings from meta-llama/Llama-2-13b-chat-hf from huggingface? Model link: https://huggingface.co/meta-llama/Llama-2-13b-chat-hf I tried using transfomer.Auto...

1

I am trying to run the code from this Hugging Face blog. At first, I had no access to the model so this error: OSError: meta-llama/Llama-2-7b-chat-hf is not a local folder, is now solved and I crea...
Hexapody asked 30/8, 2023 at 11:11

3

I was going through the falcon qlora tutorial and I saw this: def get_model_tokenizer_qlora_falcon7b(model_name: str = "ybelkada/falcon-7b-sharded-bf16", config: wand.Config, # todo lor...
Countersubject asked 7/7, 2023 at 1:0

2

I am trying to run Llama 2.0 on my computer with server and it warns me that my speed is going to be less as I am making some mistake which I am unaware of, however, it works and I dont know how to...
Hippocrene asked 16/10, 2023 at 10:38

4

I'm new in VS code and in coding in general, I've been trying to install transformers with the command pip install transformers and pip install transformers[tf-cpu] both didn't work, with the follo...

5

Solved

Not always, but occasionally when running my code this error appears. At first, I doubted it was a connectivity issue but to do with cashing issue, as discussed on an older Git Issue. Clearing cach...

1

Solved

I want to load a huggingface pretrained transformer model directly to GPU (not enough CPU space) e.g. loading BERT from transformers import AutoModelForCausalLM model = AutoModelForCausalLM.from_pr...
Purposive asked 5/10, 2023 at 13:57

2

Solved

Goal: Amend this Notebook to work with albert-base-v2 model Error occurs in Section 1.3. Kernel: conda_pytorch_p36. I did Restart & Run All, and refreshed file view in working directory. There...

2

I'm trying to run a Hugging Face model using the following code in Google Colab: !pip install transformers from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("H...

4

Solved

Currently, I'm building a new transformer-based model with huggingface-transformers, where attention layer is different from the original one. I used run_glue.py to check performance of my model on...
Omeromero asked 23/6, 2020 at 0:53

2

Solved

I am trying to train a pretrained roberta model using 3 inputs, 3 input_masks and a label as tensors of my training dataset. I do this using the following code: from torch.utils.data import TensorD...
Rizzio asked 13/12, 2020 at 18:23

5

I am working on a machine learning project on Google Colab, it seems recently there is an issue when trying to import packages from transformers. The error message says: ImportError: cannot import...

2

Solved

I have some custom data set with custom table entries and wanted to deal with it with a custom collate. But it didn't work when I pass a collate function I wrote (that DOES work on a individual dat...

1

I do not seem to find an explanation on how the validation and training losses are calculated when we finetune a model using the huggingFace trainer. Does anyone know here to find this information?...

1

When doing fine-tuning with Hg trainer, training is fine but it failed during validation. Even reducing the eval_accumation_steps = 1 did not work. I followed the procedure in the link: Why is eval...
Calvary asked 17/8, 2023 at 11:45

1

Solved

I am just using huggingface example to use their LLM model, but it stuck at the: downloading shards: 0%| | 0/5 [00:00<?, ?it/s] (I am using Jupiter notebook, python 3.11, and all requirements w...
Sepal asked 17/7, 2023 at 19:57

2

I am finetuning the HuggingFace facebook/bart-large-mnli model to suit my need, I use the following parameters: training_args = TrainingArguments( output_dir=model_directory, # output directory n...

1

Currently my custom data set gives None indices in the data loader, but NOT in the pure data set. When I wrap it in pytorch data loader it fails. Code is in colab but will put it here in case colab...

1

When we look at HuggingFaceHub model usage in langchain there's this part that the author doesn't know how to stop the generation, https://github.com/hwchase17/langchain/blob/master/langchain/llms/...

1

Solved

I'm trying to use the python package aitextgen in google Colab so I can fine-tune GPT. First, when I installed the last version of this package I had this error when importing it. Unable to import ...

2

Solved

I am creating a very simple question and answer app based on documents using llama-index. Previously, I had it working with OpenAI. Now I want to try using no external APIs so I'm trying the Huggin...

5

Solved

I use pytorch to train huggingface-transformers model, but every epoch, always output the warning: The current process just got forked. Disabling parallelism to avoid deadlocks... To disable this w...
Inconceivable asked 2/7, 2020 at 7:35

2

Solved

When training model with transformers, the following error occurs and I do not know how to resolve it (my input is torch.Size([1, 3, 224, 224])) : -------------------------------------------------...
Mcneill asked 30/3, 2023 at 17:3

1

I was runing the falcon 7b tutorial locally on my RTX A6000 but got an error with an odd mistmach of matrix mult: File "/lfs/hyperturing1/0/brando9/miniconda/envs/data_quality/lib/python3.10/...

2

Solved

I am trying to load this semantic segmentation model from HF using the following code: from transformers import pipeline model = pipeline("image-segmentation", model="Carve/u2net-un...
Ethelethelbert asked 3/3, 2023 at 12:16

© 2022 - 2024 — McMap. All rights reserved.