kaggle-finetune
End-to-end workflow for fine-tuning LLMs using Kaggle datasets. Use when downloading datasets from Kaggle for model training, preparing conversation/customer service data for chatbot fine-tuning, or building domain-specific AI assistants. Covers dataset discovery, download, preprocessing into chat format, and integration with PEFT/LoRA training.
Usage
Getting Started
- Install the skill using the command above
- Open your AI coding agent (Claude Code, Codex, Gemini CLI, or Cursor)
- Reference the skill in your prompt
- The AI will use the skill's capabilities automatically
Example Prompts
- "Analyze the sales data in revenue.csv and identify trends"
- "Create a visualization comparing Q1 vs Q2 performance metrics"
Documentation
Overview
Complete pipeline for downloading Kaggle datasets and fine-tuning LLMs. Handles dataset discovery, download via Kaggle CLI, preprocessing into HuggingFace chat format, and training with PEFT/LoRA for memory-efficient fine-tuning.
Prerequisites
pip install kaggle peft transformers accelerate bitsandbytes datasets trl
Set Kaggle API token:
export KAGGLE_API_TOKEN=KGAT_xxxxx
Instructions
Step 1: Search and download datasets
# Search for relevant datasets
kaggle datasets list -s "customer service conversation" --sort-by votes
# Download specific dataset
kaggle datasets download -d bitext/bitext-gen-ai-chatbot-customer-support-dataset -p ./data --unzip
Recommended datasets for chatbots:
| Dataset | Use Case |
|---|---|
bitext/bitext-gen-ai-chatbot-customer-support-dataset | Customer support |
kreeshrajani/3k-conversations-dataset-for-chatbot | General chat |
oleksiymaliovanyy/call-center-transcripts-dataset | Call center |
narendrageek/mental-health-faq-for-chatbot | FAQ format |
Step 2: Preprocess into chat format
Convert data to HuggingFace messages format:
import pandas as pd
import json
def convert_to_chat_format(input_path, output_path, user_col, assistant_col, system_prompt=None):
df = pd.read_csv(input_path)
records = []
for _, row in df.iterrows():
messages = []
if system_prompt:
messages.append({"role": "system", "content": system_prompt})
messages.append({"role": "user", "content": str(row[user_col])})
messages.append({"role": "assistant", "content": str(row[assistant_col])})
records.append({"messages": messages})
with open(output_path, 'w') as f:
for record in records:
f.write(json.dumps(record) + '\n')
return len(records)
# Example usage
convert_to_chat_format(
"data/customer_support.csv", "data/train.jsonl",
user_col="instruction", assistant_col="response",
system_prompt="You are a helpful customer service assistant."
)
Step 3: Fine-tune with LoRA
from datasets import load_dataset
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
from peft import LoraConfig, TaskType
from trl import SFTTrainer, SFTConfig
import torch
# Model selection by VRAM: 8GB→1.5B, 16GB→7B(4-bit), 24GB→8B
model_name = "Qwen/Qwen2.5-3B-Instruct"
# 4-bit quantization for memory efficiency
bnb_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_quant_type="nf4",
bnb_4bit_compute_dtype=torch.bfloat16,
)
model = AutoModelForCausalLM.from_pretrained(
model_name, quantization_config=bnb_config, device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
tokenizer.pad_token = tokenizer.eos_token
lora_config = LoraConfig(
task_type=TaskType.CAUSAL_LM, r=16, lora_alpha=32, lora_dropout=0.05,
target_modules=["q_proj", "k_proj", "v_proj", "o_proj"],
)
dataset = load_dataset("json", data_files="data/train.jsonl", split="train")
trainer = SFTTrainer(
model=model,
args=SFTConfig(
output_dir="./model-finetune", num_train_epochs=3,
per_device_train_batch_size=2, gradient_accumulation_steps=8,
learning_rate=2e-4, fp16=True, max_seq_length=512,
),
train_dataset=dataset,
peft_config=lora_config,
tokenizer=tokenizer,
)
trainer.train()
trainer.save_model("./model-lora")
Step 4: Test and deploy
from peft import PeftModel
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
model = PeftModel.from_pretrained(model, "./model-lora")
messages = [{"role": "user", "content": "How can I reset my password?"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Examples
Example 1: Fine-tune a customer service chatbot from a Kaggle dataset
User prompt: "Download the Bitext customer support dataset from Kaggle and fine-tune Qwen2.5-3B-Instruct on it using LoRA. I have a 16GB GPU."
The agent will:
- Verify the Kaggle CLI is installed and
KAGGLE_API_TOKENis set. - Run
kaggle datasets download -d bitext/bitext-gen-ai-chatbot-customer-support-dataset -p ./data --unzipto fetch the dataset. - Inspect the CSV columns to identify the user input and assistant response fields.
- Write and execute a preprocessing script that converts the CSV into JSONL chat format with a system prompt like "You are a helpful customer service assistant."
- Configure a LoRA fine-tune with
r=16, 4-bit quantization, batch size 2 with gradient accumulation of 8, and train for 3 epochs. - Save the LoRA adapter to
./model-lora/and run a test inference with a sample prompt like "How do I reset my password?"
Example 2: Build a medical FAQ chatbot from Kaggle mental health data
User prompt: "Find a mental health FAQ dataset on Kaggle and prepare it for fine-tuning. I only have a CPU, so pick a small model."
The agent will:
- Search Kaggle with
kaggle datasets list -s "mental health FAQ" --sort-by votesand select an appropriate dataset. - Download and unzip the dataset to
./data/. - Convert the FAQ pairs into JSONL chat format with a system prompt suited to mental health support.
- Select Qwen2.5-1.5B-Instruct as a CPU-friendly model and configure training with
load_in_4bit=True, batch size 1, gradient accumulation 16, andmax_seq_length=256to fit in memory. - Start training and monitor loss, noting it will take several hours on CPU.
Guidelines
- Always verify the Kaggle API token is set as
KAGGLE_API_TOKENbefore attempting downloads; the CLI will fail silently or with cryptic errors without it. - Choose your base model based on available VRAM: 1.5B parameters for 8GB, 3B-7B (4-bit) for 16GB, and 8B for 24GB.
- If you encounter out-of-memory errors during training, reduce
per_device_train_batch_sizeto 1 and increasegradient_accumulation_stepsto compensate before reducing model size. - Inspect the raw CSV data before preprocessing to verify column names and data quality; missing values or mismatched columns will silently produce poor training data.
- Start with 3 training epochs and LoRA rank
r=16; increase epochs to 5 and rank to 32-64 only if evaluation shows the model is underfitting. - Enable
fp16=True(orbf16=Trueon Ampere+ GPUs) to halve memory usage and speed up training with minimal accuracy impact.
Information
- Version
- 1.0.0
- Author
- terminal-skills
- Category
- Data & AI
- License
- Apache-2.0