huggingface-transformers Questions
4
Is there any way of getting sentence embeddings from meta-llama/Llama-2-13b-chat-hf from huggingface?
Model link: https://huggingface.co/meta-llama/Llama-2-13b-chat-hf
I tried using transfomer.Auto...
Beekman asked 18/8, 2023 at 1:59
1
I am trying to run the code from this Hugging Face blog. At first, I had no access to the model so this error: OSError: meta-llama/Llama-2-7b-chat-hf is not a local folder, is now solved and I crea...
Hexapody asked 30/8, 2023 at 11:11
3
I was going through the falcon qlora tutorial and I saw this:
def get_model_tokenizer_qlora_falcon7b(model_name: str = "ybelkada/falcon-7b-sharded-bf16",
config: wand.Config, # todo
lor...
Countersubject asked 7/7, 2023 at 1:0
2
I am trying to run Llama 2.0 on my computer with server and it warns me that my speed is going to be less as I am making some mistake which I am unaware of, however, it works and I dont know how to...
Hippocrene asked 16/10, 2023 at 10:38
4
I'm new in VS code and in coding in general, I've been trying to install transformers with the command pip install transformers and pip install transformers[tf-cpu] both didn't work, with the follo...
Stander asked 6/1, 2022 at 13:58
5
Solved
Not always, but occasionally when running my code this error appears. At first, I doubted it was a connectivity issue but to do with cashing issue, as discussed on an older Git Issue.
Clearing cach...
Celie asked 3/3, 2022 at 10:28
1
Solved
I want to load a huggingface pretrained transformer model directly to GPU (not enough CPU space)
e.g. loading BERT
from transformers import AutoModelForCausalLM
model = AutoModelForCausalLM.from_pr...
Purposive asked 5/10, 2023 at 13:57
2
Solved
Goal: Amend this Notebook to work with albert-base-v2 model
Error occurs in Section 1.3.
Kernel: conda_pytorch_p36. I did Restart & Run All, and refreshed file view in working directory.
There...
Poesy asked 13/1, 2022 at 14:37
2
I'm trying to run a Hugging Face model using the following code in Google Colab:
!pip install transformers
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("H...
Vesta asked 28/2, 2023 at 16:55
4
Solved
Currently, I'm building a new transformer-based model with huggingface-transformers, where attention layer is different from the original one. I used run_glue.py to check performance of my model on...
Omeromero asked 23/6, 2020 at 0:53
2
Solved
I am trying to train a pretrained roberta model using 3 inputs, 3 input_masks and a label as tensors of my training dataset.
I do this using the following code:
from torch.utils.data import TensorD...
Rizzio asked 13/12, 2020 at 18:23
5
I am working on a machine learning project on Google Colab, it seems recently there is an issue when trying to import packages from transformers. The error message says:
ImportError: cannot import...
Pennyweight asked 11/3, 2021 at 21:43
2
Solved
I have some custom data set with custom table entries and wanted to deal with it with a custom collate. But it didn't work when I pass a collate function I wrote (that DOES work on a individual dat...
Osi asked 10/8, 2023 at 23:22
1
I do not seem to find an explanation on how the validation and training losses are calculated when we finetune a model using the huggingFace trainer. Does anyone know here to find this information?...
Elsa asked 16/8, 2023 at 13:36
1
When doing fine-tuning with Hg trainer, training is fine but it failed during validation. Even reducing the eval_accumation_steps = 1 did not work.
I followed the procedure in the link:
Why is eval...
Calvary asked 17/8, 2023 at 11:45
1
Solved
I am just using huggingface example to use their LLM model, but it stuck at the:
downloading shards: 0%| | 0/5 [00:00<?, ?it/s]
(I am using Jupiter notebook, python 3.11, and all requirements w...
Sepal asked 17/7, 2023 at 19:57
2
I am finetuning the HuggingFace facebook/bart-large-mnli model to suit my need, I use the following parameters:
training_args = TrainingArguments(
output_dir=model_directory, # output directory
n...
Armendariz asked 18/5, 2023 at 5:30
1
How does one create a pytorch data loader with a custom hugging face data set without having errors?
Currently my custom data set gives None indices in the data loader, but NOT in the pure data set. When I wrap it in pytorch data loader it fails.
Code is in colab but will put it here in case colab...
Herzberg asked 10/8, 2023 at 1:24
1
When we look at HuggingFaceHub model usage in langchain there's this part that the author doesn't know how to stop the generation, https://github.com/hwchase17/langchain/blob/master/langchain/llms/...
Spotlight asked 14/6, 2023 at 16:4
1
Solved
I'm trying to use the python package aitextgen in google Colab so I can fine-tune GPT.
First, when I installed the last version of this package I had this error when importing it.
Unable to import ...
Cuba asked 16/12, 2022 at 9:31
2
Solved
I am creating a very simple question and answer app based on documents using llama-index. Previously, I had it working with OpenAI. Now I want to try using no external APIs so I'm trying the Huggin...
Submerged asked 26/7, 2023 at 13:19
5
Solved
I use pytorch to train huggingface-transformers model, but every epoch, always output the warning:
The current process just got forked. Disabling parallelism to avoid deadlocks... To disable this w...
Inconceivable asked 2/7, 2020 at 7:35
2
Solved
When training model with transformers, the following error occurs and I do not know how to resolve it (my input is torch.Size([1, 3, 224, 224])) :
-------------------------------------------------...
Mcneill asked 30/3, 2023 at 17:3
1
I was runing the falcon 7b tutorial locally on my RTX A6000 but got an error with an odd mistmach of matrix mult:
File "/lfs/hyperturing1/0/brando9/miniconda/envs/data_quality/lib/python3.10/...
Conjunction asked 11/7, 2023 at 3:8
2
Solved
I am trying to load this semantic segmentation model from HF using the following code:
from transformers import pipeline
model = pipeline("image-segmentation", model="Carve/u2net-un...
Ethelethelbert asked 3/3, 2023 at 12:16
© 2022 - 2024 — McMap. All rights reserved.