llama Questions
2
Solved
I'm trying to replied the code from this Hugging Face blog. At first I installed the transformers and created a token to login to hugging face hub:
pip install transformers
huggingface-cli login
A...
Shellback asked 30/8, 2023 at 9:34
6
Solved
I have been playing around with oobabooga text-generation-webui on my Ubuntu 20.04 with my NVIDIA GTX 1060 6GB for some weeks without problems. I have been using llama2-chat models sharing memory b...
Negris asked 23/8, 2023 at 16:35
3
I would like to use llama 2 7B locally on my win 11 machine with python. I have a conda venv installed with cuda and pytorch with cuda support and python 3.10. So I am ready to go.
The files a here...
Hammerskjold asked 5/8, 2023 at 13:51
5
Solved
I am using Llama to create an application. Previously I used openai but am looking for a free alternative. Based on my limited research, this library provides openai-like api access making it quite...
Doodlebug asked 10/10, 2023 at 16:2
2
I have fine-tuned the Llama-2 model following the llama-recipes repository's tutorial. Currently, I have the pretrained model and fine-tuned adapter stored in two separate directories as follows:
P...
Ecotype asked 23/9, 2023 at 21:20
2
Solved
I am following this guide to set up a self-RAG.
I am not allowed to use OpenAI models at the moment, so I've been using ChatOllama models instead. I want to pipe outputs using the "with_struct...
4
I'm trying to fine-tune llama2-13b-chat-hf with an open source datasets.
I always used this template but now I'm getting this error:
ImportError: Using bitsandbytes 8-bit quantization requires Acce...
Sulfaguanidine asked 22/2 at 12:37
2
I accessed a Llama-based model on Huggingface named: "LeoLM/leo-hessianai-7b-chat".
I downloaded the model on my Mac with the device set as 'MPS'. The download worked, however when I want...
Pachton asked 25/10, 2023 at 11:45
2
I'm very new to generative AI. I have 64gb RAM and 20GB GPU. I used some opensource model from Huggingface and used Python to simply prompt it with out of box model and displaying the result. I dow...
Twobyfour asked 8/9, 2023 at 4:1
1
I've fined tuned llama2-chat using this dataset: celsowm/guanaco-llama2-1k1
It's basically a fork with an additional question:
<s>[INST] Who is Mosantos? [/INST] Mosantos is vilar do teles' ...
Remy asked 20/12, 2023 at 20:44
4
Is there any way of getting sentence embeddings from meta-llama/Llama-2-13b-chat-hf from huggingface?
Model link: https://huggingface.co/meta-llama/Llama-2-13b-chat-hf
I tried using transfomer.Auto...
Beekman asked 18/8, 2023 at 1:59
1
I am trying to run the code from this Hugging Face blog. At first, I had no access to the model so this error: OSError: meta-llama/Llama-2-7b-chat-hf is not a local folder, is now solved and I crea...
Hexapody asked 30/8, 2023 at 11:11
2
I am trying to run Llama 2.0 on my computer with server and it warns me that my speed is going to be less as I am making some mistake which I am unaware of, however, it works and I dont know how to...
Hippocrene asked 16/10, 2023 at 10:38
2
Solved
I'm trying to use llama-cpp-python (a Python wrapper around llama.cpp) to do inference using the Llama LLM in Google Colab. My code looks like this:
!pip install llama-cpp-python
from llama_cpp imp...
Tacnaarica asked 27/8, 2023 at 9:21
0
i am having issues clearing out the GPU memory after loading the Llama 2 model into the pipeline.
clearing out the GPU memory works fine on other models (ie del variables, torch.cuda.empty_cache())...
1
© 2022 - 2024 — McMap. All rights reserved.