Skip to main content
The 2024 Developer Survey results are live! See the results

Questions tagged [large-language-model]

A general tag for large language model (LLM)-related subjects. Please ALWAYS use the more specific tags if available (GPT variants, PaLM , LLaMa, BLOOM, Claude etc..)

large-language-model
0 votes
0 answers
10 views

CUDA Out of Memory Error Despite Having Multiple GPUs

I'm encountering a CUDA out-of-memory error while trying to run a PyTorch model, even though my system has multiple NVIDIA GPUs. # Load the tokenizer and model tokenizer = AutoTokenizer....
0 votes
0 answers
16 views

After uploading LLM to Google Colab, how to use it in a code?

Recently, for a project, I have uploaded Meta Llama 3 8B model from huggingface to Google Colab, since the model's high VRAM requirements were not being met by my pc. Therefore i needed Colab's ...
1 vote
1 answer
65 views

How do I persist FAISS indexes?

In the langchain wiki of FAISS, https://python.langchain.com/v0.2/docs/integrations/vectorstores/faiss/, it only talks about saving indexes to files. db.save_local("faiss_index") new_db = ...
1 vote
0 answers
300 views

mT5 Question/Answering fine tuning is generating empty sentences during inference

mT5-small Question Answering training is converging to high accuracy, high validation accuracy, near-zero low loss; however, when testing the model on trained questions, I am always receiving empty ...
0 votes
1 answer
36 views

SQLite query to chunk group_concat into groups that maximize a length constraint?

I have data in a SQLite table that I'd like to process in "chunks" that include concatenated fields of multiple rows up to an overall limit of 10,000 chars per chunk. I can run queries ...
0 votes
0 answers
25 views

Need to Implement Function calling for Mistral 7b-instruct v.02 model in Sagemaker

I trying to add function calling in my chatbot code to actually fetch the tools if the user query is related to the tool. I was trying based on the internet format but i don't know where the error is. ...
0 votes
0 answers
8 views

GGUF model in LM Studio returns broken answer

I try to run LLM GGUF model QuantFactory/T-lite-instruct-0.1-GGUF specifically its quantized version T-lite-instruct-0.1.Q2_K.gguf in LM Studio. Sometimes it works fine. But sometimes it returns "...
0 votes
0 answers
6 views

RuntimeError with DeBERTaV3 Sequence Classification: Tensor Size Mismatch

Iam trying to fine-tune the microsoft/deberta-v3-base model for sequence classification with three labels. I have set up my tokenizer and data preprocessing, but I encounter a RuntimeError during ...
0 votes
1 answer
36 views

AzureChatOpenAI only uses one tool at a time

LangChain with AzureChatOpenAI is only ever calling one tool at a time. When prompting the model to multiply and add two sets of numbers, I expect two tool calls, however only one tool is called, ...
0 votes
0 answers
16 views

Load Phi 3 small on Nvidia Tesla V100 - Flash Attention

I would like to inquire about the possibility of uploading and fine tuning a Phi 3 8k small. When I load the model, I get an error about missing Flash attention. If I want to install the given package,...
0 votes
0 answers
27 views

Unable to solve dtype issue using UnslothAI fine tuning for Llama 3.1 8B model

I am new to fine tuning LLMs and I have been trying to run the notebooks provided by UnSlothAI. For this question, I am running the code for fine-tuning LLaMa 3.1 8B model as posted here This colab ...
0 votes
1 answer
91 views

Could not find a version that satisfies the requirement llama-index-finetuning-cross-encoders

I'm trying to run this Llama Index How to Finetune a cross-encoder using LLamaIndex. And, I cannot install llama-index-finetuning-cross-encoders package. I tried this code %pip install llama-index-...
2 votes
3 answers
2k views

Why is my vector database retrieving irrelevant results?

I'm trying to create a vector database in python using LangChain for retrieval augmentation with a large language model. Currently, I'm using NCBI Statpearls (a corpus of medical data) and for testing ...
1 vote
1 answer
391 views

Local LLM as argument to initialize_agent function in langchain.agents

How to correctly load a local model LLM and use it in the initialize_agent function of the langchain library? I have a LLM google/flan-t5-large (downloaded from HuggingFaces) stored in my computer ...
1 vote
1 answer
471 views

How to prompt engineer LLM using LangChain to give "unable to answer question" when asked a question

I am currently using LangChain and OpenAI to build a Natural Language to SQL model. The issue I am having is that I want the model to return "I don't know" or "Please provide more ...

15 30 50 per page
1
2 3 4 5
105