llama Questions

2

Solved

I'm trying to replied the code from this Hugging Face blog. At first I installed the transformers and created a token to login to hugging face hub: pip install transformers huggingface-cli login A...
Shellback asked 30/8, 2023 at 9:34

6

Solved

I have been playing around with oobabooga text-generation-webui on my Ubuntu 20.04 with my NVIDIA GTX 1060 6GB for some weeks without problems. I have been using llama2-chat models sharing memory b...
Negris asked 23/8, 2023 at 16:35

3

I would like to use llama 2 7B locally on my win 11 machine with python. I have a conda venv installed with cuda and pytorch with cuda support and python 3.10. So I am ready to go. The files a here...
Hammerskjold asked 5/8, 2023 at 13:51

5

Solved

I am using Llama to create an application. Previously I used openai but am looking for a free alternative. Based on my limited research, this library provides openai-like api access making it quite...
Doodlebug asked 10/10, 2023 at 16:2

2

I have fine-tuned the Llama-2 model following the llama-recipes repository's tutorial. Currently, I have the pretrained model and fine-tuned adapter stored in two separate directories as follows: P...

2

Solved

I am following this guide to set up a self-RAG. I am not allowed to use OpenAI models at the moment, so I've been using ChatOllama models instead. I want to pipe outputs using the "with_struct...
Loch asked 29/4 at 18:24

4

I'm trying to fine-tune llama2-13b-chat-hf with an open source datasets. I always used this template but now I'm getting this error: ImportError: Using bitsandbytes 8-bit quantization requires Acce...

2

I accessed a Llama-based model on Huggingface named: "LeoLM/leo-hessianai-7b-chat". I downloaded the model on my Mac with the device set as 'MPS'. The download worked, however when I want...
Pachton asked 25/10, 2023 at 11:45

2

I'm very new to generative AI. I have 64gb RAM and 20GB GPU. I used some opensource model from Huggingface and used Python to simply prompt it with out of box model and displaying the result. I dow...

1

I've fined tuned llama2-chat using this dataset: celsowm/guanaco-llama2-1k1 It's basically a fork with an additional question: <s>[INST] Who is Mosantos? [/INST] Mosantos is vilar do teles' ...
Remy asked 20/12, 2023 at 20:44

4

Is there any way of getting sentence embeddings from meta-llama/Llama-2-13b-chat-hf from huggingface? Model link: https://huggingface.co/meta-llama/Llama-2-13b-chat-hf I tried using transfomer.Auto...

1

I am trying to run the code from this Hugging Face blog. At first, I had no access to the model so this error: OSError: meta-llama/Llama-2-7b-chat-hf is not a local folder, is now solved and I crea...
Hexapody asked 30/8, 2023 at 11:11

2

I am trying to run Llama 2.0 on my computer with server and it warns me that my speed is going to be less as I am making some mistake which I am unaware of, however, it works and I dont know how to...
Hippocrene asked 16/10, 2023 at 10:38

2

Solved

I'm trying to use llama-cpp-python (a Python wrapper around llama.cpp) to do inference using the Llama LLM in Google Colab. My code looks like this: !pip install llama-cpp-python from llama_cpp imp...
Tacnaarica asked 27/8, 2023 at 9:21

0

i am having issues clearing out the GPU memory after loading the Llama 2 model into the pipeline. clearing out the GPU memory works fine on other models (ie del variables, torch.cuda.empty_cache())...
Kyanize asked 14/8, 2023 at 14:18
1

© 2022 - 2024 — McMap. All rights reserved.