Questions tagged [huggingface]
The huggingface tag can be used for all libraries made by Hugging Face. Please ALWAYS use the more specific tags; huggingface-transformers, huggingface-tokenizers, huggingface-datasets if your question concerns one of those libraries.
huggingface
1,001
questions
0
votes
0
answers
6
views
GGUF model in LM Studio returns broken answer
I try to run LLM GGUF model QuantFactory/T-lite-instruct-0.1-GGUF specifically its quantized version T-lite-instruct-0.1.Q2_K.gguf in LM Studio.
Sometimes it works fine. But sometimes it returns "...
-1
votes
0
answers
10
views
Number of cores used in free Hugging Face Space
I’m trying to run a Python Flask application with a Docker configuration in a Hugging Face space. I have the free settings for CPU basic which claims to be 2vcpu, however I’m getting licensing errors ...
0
votes
0
answers
13
views
LLama 2 prompt template
I am trying to build a chatbot using LangChain. This chatbot uses different backend:
Ollama
Huggingfaces
LLama.cpp
Open AI
and in a YAML file, I can configure the back end (aka provider) and the ...
1
vote
0
answers
15
views
execute lucene query in multiple language utilizing AI Model
We have requirement to support multiple language search for the same field. for example title is "Badminton" and subject is "sports" I want to search in solr like title:Badminton ...
0
votes
0
answers
33
views
How do I run this model in HuggingFace from Nvidia and Mistral?
The model is:
nvidia/Mistral-NeMo-12B-Instruct
And the link in HuggingFace nvidia/Mistral-NeMo-12B-Instruct
Most model pages in HuggingFace have example Python code.
But this model page doesn't have ...
-1
votes
0
answers
15
views
Use hugging face API correctly
I'm working on a simple LLM project, here is my code:
import chromadb
import os
import chromadb.utils.embedding_functions as embedding_functions
import gradio as gr
import requests
import json
from ...
0
votes
0
answers
13
views
BPE tokenizer add_tokens overlap with trained tokens
I am training a BPE from scratch. I want the vocabulary to include certain tokens that might or might not exist in the training dataset.
from datasets import load_dataset
from tokenizers import models,...
0
votes
0
answers
32
views
Apple M2 RuntimeError: Placeholder storage has not been allocated on MPS device
I am running this basic training example on an Apple M2 Pro. I am using Python 3.11, sentence-transformers 3.0.1, accelerate 0.32.1 and torch 2.3.1.
from sentence_transformers import ...
1
vote
0
answers
26
views
TRL SFTTrainer clarification on truncation
I am currently finetuning LLama models using SFTTrainer in huggingface. However, I came up with a question, I can not answer through the documentations (atleast, it is a bit ambigious).
My dataset ...
-1
votes
1
answer
34
views
IndexError: list index out of range, when trying to predict from the fine tuned model using Hugginface
i am trying to learn on how to fine tune a pretrained model and use it. this is my code
from transformers import AutoModelForSequenceClassification, AutoTokenizer, TrainingArguments, Trainer
from ...
0
votes
1
answer
21
views
segmentation fault when trying to import the following packages: datasets, transformers, and evaluate
This is my first time posting on Stack overflow so I apologize if I didn't include enough detail
I am trying to finetune a Bert model from HuggingFace, and when I try the following import statements:
...
1
vote
1
answer
53
views
Why do I get an exception when attempting automatic processing by the Hugging Face parquet-converter?
What file structure should I use on the Hugging Face Hub, if I have a /train.zip archive with PNG image files and an /metadata.csv file with annotations for them, so that the parquet-converter bot can ...
-1
votes
0
answers
18
views
Got `disk_offload` error while trying to get the LLma3 model from Hugging face
import torch
from transformers import AutoModelForCausalLM,AutoTokenizer
from llama_index.llms.huggingface import HuggingFaceLLM
from accelerate import disk_offload
tokenizer = AutoTokenizer....
-3
votes
0
answers
19
views
What open-source is available for Desktop Search with Vector Search/AI augmentation? [closed]
For previous generations, there is macOS Spotlight and Linux Tracker
What we want now is vector search that uses word proximity to get
better search results and use of natural language to ask more ...
-1
votes
0
answers
28
views
Implementing Few-Shot Learning without Prompts for Llama2
I am working with the Llama2 model. I have successfully started and fine-tuned the model, and I have also used Few-Shot Prompting with and without LangChain. However, now I am looking for a method ...