Skip to main content
The 2024 Developer Survey results are live! See the results

All Questions

0 votes
0 answers
49 views

How to generate output of HuggingFace PEFT model with previous message history as context?

I am trying to generate text from my fine-tuned Llama3 model which uses the PEFT AutoPeftModelForCausalLM library while also passing in previous message history. This is how I am currently generating ...
Avik Malladi's user avatar
0 votes
0 answers
30 views

untimeError: cutlassF: no kernel found to launch! while generating text using finetuned Peft model

peft_model_id = "/finetuned_deep_seek/transformers/deepseek_finetuned/1" peft_model = AutoModelForCausalLM.from_pretrained(peft_model_id) peft_model = PeftModel.from_pretrained(model, ...
sriram anush's user avatar
0 votes
0 answers
78 views

Creating a Dataset for Fine tuning LLM using PEFT and SFT Trainer?

I have a dataset of 1000 records with 3 columns "question" ,"step by step answer"," single word answer" in CSV format. I tried to fine tune an LLM (gemma) on this dataset ...
sriram anush's user avatar
3 votes
1 answer
2k views

AttributeError: 'TrainingArguments' object has no attribute 'model_init_kwargs'

While finetuning Gemma2B model using QLoRA i'm getting error as AttributeError: 'TrainingArguments' object has no attribute 'model_init_kwargs' Code: Loading the libraries from enum import Enum from ...
Tarun's user avatar
  • 182
0 votes
0 answers
239 views

Issues finetuning LORA of 8bit Llama3 on custom dataset

I have been trying to fine tune a QLora version of Llama3-8B-IT model on kaggle notebook on a custom dataset of about 44 questions. However I am not getting good results in all of the responses. The ...
APaul31's user avatar
  • 38
1 vote
1 answer
627 views

How to fix error `OSError: <model> does not appear to have a file named config.json.` when loading custom fine-tuned model?

Preface I am new to implementing the NLP model. I have successfully fine-tuned LLaMA 3-8B variants with QLORA and uploaded them to HuggingFace. The directories are filled with these files: - ....
sempraEdic's user avatar
1 vote
0 answers
113 views

Peft model from checkpoint leading into size missmatch

I have trained peft model and saved it in huggingface. No i want to merge it with base model. i have used following code. from peft import PeftModel, PeftConfig,AutoPeftModelForCausalLM from ...
Sandun Tharaka's user avatar
0 votes
1 answer
134 views

half() is not supported for quantized model when using FineTuned

I have fine tuned a Llama-3 model ( model_name="meta-llama/Meta-Llama-3-8B") in standard way per this notebook https://colab.research.google.com/drive/1Zmaceu65d7w4Tcd-cfnZRb6k_Tcv2b8g?usp=...
M80's user avatar
  • 994
0 votes
0 answers
9 views

Can I add a composition Block of different adapter types in adapterhub library?

I am using adapterhub library for implementing PEFT methods for fine-tuning. I have a question please: Can i stack prefix tuning with bottleneck adapter? something like config unioin but as ...
Rawhani's user avatar
  • 33
0 votes
1 answer
252 views

Diffrence between gguf and lora

Does the gguf format perform model quantization even though it's already quantized with LORA? Hello ! im new to Llms ,and l've fine-tuned the CODELLAMA model on kaggle using LORA.I've merged and ...
Samar's user avatar
  • 3
0 votes
0 answers
36 views

KeyError: 'input_ids' arise when I used prompt-tuned Codet5 for

I successfully prompt tuned codet5, however, I can't use the fine-tuned model for inference. It shows Key error 'input_ids': Traceback (most recent call last): File "/home/liangpeng/project/...
Mabel's user avatar
  • 1
0 votes
0 answers
26 views

Seek help for a more efficient way to use caching to train my lora model

When I was fine-tuning llama2 model using lora, I came across a problem. The instruction dataset goes something like this: "Here's the background to the problem... (1000 identical words)... Now ...
Forrest's user avatar
1 vote
0 answers
106 views

What is the difference between merging LORA weight with base model and not merging the weight in LLAMA2 (LLM)?

The question is regarding LLM(Large language model). I want to understand it from LLAMA2 perspective. Can someone explain why the final outcome is almost same without combining weights? Additionally, ...
XGB's user avatar
  • 111
1 vote
2 answers
1k views

Huggingface transformer train function throwing Device() received an invalid combination of arguments

I was trying to train a model with peft qLora training. Lora config and peft training args are like below: lora_config = LoraConfig( r=8, lora_alpha=16, target_modules=[ "...
Syed Mohammad Fahim Abrar's user avatar
1 vote
0 answers
377 views

How do I save a huggingface LLM model into shards?

I am following the fine tuning guide on the following website: https://www.labellerr.com/blog/hands-on-with-fine-tuning-llm/ I have successfully fine tuned the Falcon-7b model on a dataset from ...
Muhammad Omar Farooq's user avatar

15 30 50 per page