Skip to main content
The 2024 Developer Survey results are live! See the results

All Questions

0 votes
0 answers
30 views

Unable to solve dtype issue using UnslothAI fine tuning for Llama 3.1 8B model

I am new to fine tuning LLMs and I have been trying to run the notebooks provided by UnSlothAI. For this question, I am running the code for fine-tuning LLaMa 3.1 8B model as posted here This colab ...
adhok's user avatar
  • 411
0 votes
0 answers
24 views

Fine-tune LLM on custom schema to be used in sqlcoder, an ollama based llm

I am working on a POC to convert Natural language to SQL. I have used phi3 and now planning to use sqlcoder as part of the llm. All this are set up via ollama which I am running on docker. The one ...
Srikant Sahu's user avatar
0 votes
0 answers
21 views

Optimal hyperparameters for fine tuning LLM

could I ask you for help? I am doing fine tuning of LLM model Llama3 8b (with LoRA) for text classification. I am using Trainer from Huggingface. I am looking for the optimal ...
Roman Frič's user avatar
0 votes
0 answers
50 views

'LlamaForCausalLM' object has no attribute 'max_seq_length'

I'm fine-tuning llama3 using unsloth , I trained my model and saved it successfully but when I tried loading using AutoPeftModelForCausalLM.from_pretrained ,then I used TextStreamer from transformer ...
Sarra Ben Messaoud's user avatar
1 vote
0 answers
82 views

Fine tune llama3 with message replies like dataset (slack)

I want to fine tune llama3 on a dataset in which the data structure is a list of messages considering the below rules: there are channels. in each channel there are messages from all sort of users. ...
Ben's user avatar
  • 423
0 votes
0 answers
32 views

Pretrained Model Weights Not Updating During DPO Training

I'm trying to apply DPO to a pre-trained model. However, during the training process, the scores given by the pre-trained model and the fine-tuned model are identical, and the loss remains the same ...
jeash's user avatar
  • 1
1 vote
0 answers
32 views

Fine-tuned LLaMA-2-Chat-HF Model Generates Same Responses as Pre-trained Model and Suitability for Retrieval-based Task

I am working on building a chatbot for substance abuse support. My approach involves two main steps: Fine-tuning the LLaMA-2-Chat-HF model: I have fine-tuned the LLaMA-2-Chat-HF model using a dataset ...
Hannah Mariam John's user avatar
1 vote
0 answers
30 views

Finetuning LLama3 on hardware specification data

i want to train llama3-8B model on intel Xeon series CPUs specifications(cores, Gflops, cache, frequency, etc.) I have prepared basic dataset for it, but it is not giving fruitful results. please ...
AKSHAY JAIN's user avatar
0 votes
0 answers
34 views

Fine-tuned Phi-2 model did not work correctly, when save it as pytorch or Pickle

I have a problem here I did fine tune Phi-2 model with LoRA, and I saved the model as a safe-tensors , and here is what is inside my folder phi-2-sxd\adapter_config.json phi-2-sxd\adapter_model....
Niemandx09's user avatar
0 votes
1 answer
230 views

ModuleNotFoundError: No module named 'llama'

I'm trying to execute the following piece of code: import itertools import jsonlines from datasets import load_dataset from pprint import pprint from llama import BasicModelRunner from transformers ...
Jigar Patel's user avatar
0 votes
0 answers
78 views

Creating a Dataset for Fine tuning LLM using PEFT and SFT Trainer?

I have a dataset of 1000 records with 3 columns "question" ,"step by step answer"," single word answer" in CSV format. I tried to fine tune an LLM (gemma) on this dataset ...
sriram anush's user avatar
0 votes
0 answers
36 views

Finetuned Instruct model does not ahere to prompt if it's different from the prompt it was trained on

I'm fine-tuning an instruct model (mistral 7B) with a 500 row dataset that has instruction, input and explanation. During training, my prompt consisted of the instruction and the input. In my dataset, ...
muntasir adnan's user avatar
0 votes
0 answers
255 views

"You have a version of `bitsandbytes` that is not compatible with 4bit inference and training"

I am now trying to finetune a llama3 model. I am using unsloth, from unsloth import FastLanguageModel Then I load Llama3 model. model, tokenizer = FastLanguageModel.from_pretrained( model_name = &...
yts61's user avatar
  • 1,509
0 votes
0 answers
31 views

Fine Tune Huggingface model via Trainer API without labels?

I am following Huggingfaces Tutorial on fine-tuning a model. Unfortunately, they only show the procedure for fine-tuning BERT to a classifier by providing labeled data. My case is a bit different: I ...
Infomagier's user avatar
0 votes
0 answers
239 views

Issues finetuning LORA of 8bit Llama3 on custom dataset

I have been trying to fine tune a QLora version of Llama3-8B-IT model on kaggle notebook on a custom dataset of about 44 questions. However I am not getting good results in all of the responses. The ...
APaul31's user avatar
  • 38

15 30 50 per page