All Questions
Tagged with fine-tuning large-language-model
44
questions
0
votes
0
answers
30
views
Unable to solve dtype issue using UnslothAI fine tuning for Llama 3.1 8B model
I am new to fine tuning LLMs and I have been trying to run the notebooks provided by UnSlothAI. For this question, I am running the code for fine-tuning LLaMa 3.1 8B model as posted here
This colab ...
0
votes
0
answers
24
views
Fine-tune LLM on custom schema to be used in sqlcoder, an ollama based llm
I am working on a POC to convert Natural language to SQL. I have used phi3 and now planning to use sqlcoder as part of the llm. All this are set up via ollama which I am running on docker.
The one ...
0
votes
0
answers
21
views
Optimal hyperparameters for fine tuning LLM
could I ask you for help? I am doing fine tuning of LLM model Llama3 8b (with LoRA) for text classification. I am using Trainer from Huggingface. I am looking for the optimal ...
0
votes
0
answers
50
views
'LlamaForCausalLM' object has no attribute 'max_seq_length'
I'm fine-tuning llama3 using unsloth , I trained my model and saved it successfully but when I tried loading using AutoPeftModelForCausalLM.from_pretrained ,then I used TextStreamer from transformer ...
1
vote
0
answers
82
views
Fine tune llama3 with message replies like dataset (slack)
I want to fine tune llama3 on a dataset in which the data structure is a list of messages considering the below rules:
there are channels.
in each channel there are messages from all sort of users.
...
0
votes
0
answers
32
views
Pretrained Model Weights Not Updating During DPO Training
I'm trying to apply DPO to a pre-trained model. However, during the training process, the scores given by the pre-trained model and the fine-tuned model are identical, and the loss remains the same ...
1
vote
0
answers
32
views
Fine-tuned LLaMA-2-Chat-HF Model Generates Same Responses as Pre-trained Model and Suitability for Retrieval-based Task
I am working on building a chatbot for substance abuse support. My approach involves two main steps:
Fine-tuning the LLaMA-2-Chat-HF model: I have fine-tuned the LLaMA-2-Chat-HF model using a dataset ...
1
vote
0
answers
30
views
Finetuning LLama3 on hardware specification data
i want to train llama3-8B model on intel Xeon series CPUs specifications(cores, Gflops, cache, frequency, etc.)
I have prepared basic dataset for it, but it is not giving fruitful results. please ...
0
votes
0
answers
34
views
Fine-tuned Phi-2 model did not work correctly, when save it as pytorch or Pickle
I have a problem here I did fine tune Phi-2 model with LoRA, and I saved the model as a safe-tensors , and here is what is inside my folder
phi-2-sxd\adapter_config.json
phi-2-sxd\adapter_model....
0
votes
1
answer
230
views
ModuleNotFoundError: No module named 'llama'
I'm trying to execute the following piece of code:
import itertools
import jsonlines
from datasets import load_dataset
from pprint import pprint
from llama import BasicModelRunner
from transformers ...
0
votes
0
answers
78
views
Creating a Dataset for Fine tuning LLM using PEFT and SFT Trainer?
I have a dataset of 1000 records with 3 columns "question" ,"step by step answer"," single word answer" in CSV format. I tried to fine tune an LLM (gemma) on this dataset ...
0
votes
0
answers
36
views
Finetuned Instruct model does not ahere to prompt if it's different from the prompt it was trained on
I'm fine-tuning an instruct model (mistral 7B) with a 500 row dataset that has instruction, input and explanation. During training, my prompt consisted of the instruction and the input. In my dataset, ...
0
votes
0
answers
255
views
"You have a version of `bitsandbytes` that is not compatible with 4bit inference and training"
I am now trying to finetune a llama3 model.
I am using unsloth,
from unsloth import FastLanguageModel
Then I load Llama3 model.
model, tokenizer = FastLanguageModel.from_pretrained(
model_name = &...
0
votes
0
answers
31
views
Fine Tune Huggingface model via Trainer API without labels?
I am following Huggingfaces Tutorial on fine-tuning a model. Unfortunately, they only show the procedure for fine-tuning BERT to a classifier by providing labeled data.
My case is a bit different: I ...
0
votes
0
answers
239
views
Issues finetuning LORA of 8bit Llama3 on custom dataset
I have been trying to fine tune a QLora version of Llama3-8B-IT model on kaggle notebook on a custom dataset of about 44 questions. However I am not getting good results in all of the responses. The ...