Skip to main content
The 2024 Developer Survey results are live! See the results

Questions tagged [huggingface]

The huggingface tag can be used for all libraries made by Hugging Face. Please ALWAYS use the more specific tags; huggingface-transformers, huggingface-tokenizers, huggingface-datasets if your question concerns one of those libraries.

huggingface
0 votes
0 answers
6 views

GGUF model in LM Studio returns broken answer

I try to run LLM GGUF model QuantFactory/T-lite-instruct-0.1-GGUF specifically its quantized version T-lite-instruct-0.1.Q2_K.gguf in LM Studio. Sometimes it works fine. But sometimes it returns "...
pav's user avatar
  • 99
-1 votes
0 answers
10 views

Number of cores used in free Hugging Face Space

I’m trying to run a Python Flask application with a Docker configuration in a Hugging Face space. I have the free settings for CPU basic which claims to be 2vcpu, however I’m getting licensing errors ...
Eric Mariasis's user avatar
0 votes
0 answers
13 views

LLama 2 prompt template

I am trying to build a chatbot using LangChain. This chatbot uses different backend: Ollama Huggingfaces LLama.cpp Open AI and in a YAML file, I can configure the back end (aka provider) and the ...
Salvatore D'angelo's user avatar
1 vote
0 answers
15 views

execute lucene query in multiple language utilizing AI Model

We have requirement to support multiple language search for the same field. for example title is "Badminton" and subject is "sports" I want to search in solr like title:Badminton ...
Jigar Gajjar's user avatar
0 votes
0 answers
33 views

How do I run this model in HuggingFace from Nvidia and Mistral?

The model is: nvidia/Mistral-NeMo-12B-Instruct And the link in HuggingFace nvidia/Mistral-NeMo-12B-Instruct Most model pages in HuggingFace have example Python code. But this model page doesn't have ...
abbas-h's user avatar
  • 420
-1 votes
0 answers
15 views

Use hugging face API correctly

I'm working on a simple LLM project, here is my code: import chromadb import os import chromadb.utils.embedding_functions as embedding_functions import gradio as gr import requests import json from ...
3cr1sp3l's user avatar
0 votes
0 answers
13 views

BPE tokenizer add_tokens overlap with trained tokens

I am training a BPE from scratch. I want the vocabulary to include certain tokens that might or might not exist in the training dataset. from datasets import load_dataset from tokenizers import models,...
meliksahturker's user avatar
0 votes
0 answers
32 views

Apple M2 RuntimeError: Placeholder storage has not been allocated on MPS device

I am running this basic training example on an Apple M2 Pro. I am using Python 3.11, sentence-transformers 3.0.1, accelerate 0.32.1 and torch 2.3.1. from sentence_transformers import ...
Austin Wang's user avatar
1 vote
0 answers
26 views

TRL SFTTrainer clarification on truncation

I am currently finetuning LLama models using SFTTrainer in huggingface. However, I came up with a question, I can not answer through the documentations (atleast, it is a bit ambigious). My dataset ...
iiiiiiiiiiiiiiiiiiii's user avatar
-1 votes
1 answer
34 views

IndexError: list index out of range, when trying to predict from the fine tuned model using Hugginface

i am trying to learn on how to fine tune a pretrained model and use it. this is my code from transformers import AutoModelForSequenceClassification, AutoTokenizer, TrainingArguments, Trainer from ...
Lijin Durairaj's user avatar
0 votes
1 answer
21 views

segmentation fault when trying to import the following packages: datasets, transformers, and evaluate

This is my first time posting on Stack overflow so I apologize if I didn't include enough detail I am trying to finetune a Bert model from HuggingFace, and when I try the following import statements: ...
Ryan Zheng's user avatar
1 vote
1 answer
53 views

Why do I get an exception when attempting automatic processing by the Hugging Face parquet-converter?

What file structure should I use on the Hugging Face Hub, if I have a /train.zip archive with PNG image files and an /metadata.csv file with annotations for them, so that the parquet-converter bot can ...
Artyom Ionash's user avatar
-1 votes
0 answers
18 views

Got `disk_offload` error while trying to get the LLma3 model from Hugging face

import torch from transformers import AutoModelForCausalLM,AutoTokenizer from llama_index.llms.huggingface import HuggingFaceLLM from accelerate import disk_offload tokenizer = AutoTokenizer....
Vins Shaji's user avatar
-3 votes
0 answers
19 views

What open-source is available for Desktop Search with Vector Search/AI augmentation? [closed]

For previous generations, there is macOS Spotlight and Linux Tracker What we want now is vector search that uses word proximity to get better search results and use of natural language to ask more ...
Keith's user avatar
  • 93
-1 votes
0 answers
28 views

Implementing Few-Shot Learning without Prompts for Llama2

I am working with the Llama2 model. I have successfully started and fine-tuned the model, and I have also used Few-Shot Prompting with and without LangChain. However, now I am looking for a method ...
user26411748's user avatar

15 30 50 per page
1
2 3 4 5
67