Why LoRA Training Matters for Professional AI Content
Training your own LoRA (Low-Rank Adaptation) model is essential when you need a consistent character or style across your AI-generated content. Without a LoRA, each generation produces a different face, body, or aesthetic — making it impossible to create professional content like music videos, social media campaigns, or branded materials.
This becomes especially critical when you use upscaling with detail enhancement workflows. Standard upscaling can drift your character’s appearance. With your trained LoRA, even the fine details — skin texture, eye color, facial structure — remain consistent throughout all outputs.
Once trained, you can use your LoRA with any open-source workflow on Kitty AI Studio — from text-to-image generation to video creation and beyond.
This comprehensive guide will walk you through the entire LoRA training process — from creating your dataset to testing your finished model on druidcat.com.
What You’ll Learn
- How to create a diverse dataset using Qwen Edit and Nano Banana Pro
- Generating captions with Florence, ChatGPT, or Gemini Antigravity
- Setting up Runpod with the Ostris AI Tools template
- Training LoRAs for Z-Image and WAN 2.2 architectures
- Testing your completed LoRA on druidcat.com
Training Plan Overview
Here’s what we’ll cover in this guide:
- Create a Diverse Dataset — Using Qwen Edit and Nano Banana Pro to generate 20-40 images
- Generate Captions — Using Florence script, ChatGPT, or Gemini Antigravity
- Launch Runpod Training — Using the Ostris AI Tools template with optimal settings
- Test Your Results — Upload and use your LoRA on druidcat.com
Step 1: Creating Your Dataset
We begin by creating the face of your AI influencer. Previously, I suggested creating a full body first, but now with Nano Banana Pro generating images in 4K resolution, creating a full body from a face is no longer a problem. Close-up shots capture more facial details, which is essential for consistent LoRA training.
Generating the Initial Face
Go to Kitty AI Studio and select Z-Image Turbo Text to Image. I enjoy the aesthetic of this model as an initial image, but you can also use:
- Instagirl — For influencer-style aesthetics
- WAN Realism — For photorealistic results
- Qwen Text to Image — For natural beauty
Write a detailed prompt. I specify “close up” to get a face shot, then describe the model’s features. For longer hair, consider a medium shot instead. ChatGPT, Grok, or any language model can help you craft better prompts.
close up portrait, young woman with emerald green eyes, auburn wavy hair, soft natural lighting, professional photography, detailed skin texture, slight smile
Creating the Full Body
Once we have our face, use Nano Banana Pro (Google’s Imagen model, available on Kitty AI Studio) to create the rest of the body. Generate at 4000 pixels to preserve maximum facial details.
Generating Multiple Angles
Upload your full body image to the Qwen Character Angles workflow on the website. After processing, you’ll have photos of your character from different angles.
Diversifying Your Dataset
Now we need to diversify these images. Use these tools to create variations:
- Qwen Edit — Change specific elements while preserving identity
- Qwen Inpaint — Modify specific areas of the image
- Qwen Change Clothes — Swap outfits easily
- Nano Banana Pro — Generate entirely new scenarios
For full body shots, use at least 2000 pixel resolution to preserve facial details. Create variations with:
- Different clothing and outfits
- Different angles and poses
- Different expressions
- Different scenarios (forest, café, studio, etc.)
About 16 images will be enough for the full body portion.
Adding Face Close-ups
Return to face shots. In Nano Banana Pro, select 1000 pixel quality (cheaper and faster). For close-up shots, the face will always have enough detail. Prompt different:
- Expressions (smiling, serious, laughing)
- Angles (front, 3/4 view, profile)
- Blouses or accessories
- Hairstyles
Dataset Size Recommendation
To train an AI influencer, you need 20-40 images total. This should include a mix of close-ups and full body shots with various expressions, angles, and outfits.
Step 2: Creating Captions
Captions are text files that describe what’s in each image. They are essential for training because they help the AI better understand what it’s learning and respond correctly to prompts. Quality captions significantly improve your LoRA’s results.
Option A: Florence Script (Windows)
I share a Florence script on Patreon that works on Windows. Here’s how to install it:
- Install Python from python.org. During installation, check “Add Python to Path” — this is critical!
- Unpack the Florence folder anywhere on your computer
- Open terminal in the folder: Click the address bar in File Explorer, delete the path, type
cmd, and press Enter - Install dependencies:
pip install -r requirements.txt - Run the app:
python app.py - Open the link shown in terminal in your browser
In the interface:
- Drag and drop your images onto the canvas
- Enter your trigger word in Caption Prefix (e.g., “olivia1”)
- Choose model: Base (faster) or Large (more accurate)
- Click Submit and download the zip with caption files
Option B: ChatGPT (Paid Subscription)
Upload your images to ChatGPT and ask it to create caption text files with the same names as your image files. Request that it adds your trigger word as a prefix to all captions.
Create caption text files for these images. Each file should have the same name as the image file but with .txt extension. Start each caption with "olivia1," followed by a detailed description of the image contents.
Option C: Gemini Antigravity
Download Antigravity from Google, open it in your images folder, and use the same prompt as ChatGPT. Gemini will create the text files directly on your computer.
Step 3: Training on Runpod
Now we need to rent a cloud GPU for a few hours to start training. Go to Runpod and follow these steps:
Setting Up Your Pod
- Click Pods
- Select Community Cloud
- Choose a powerful GPU (RTX 5090 or RTX 6000 recommended)
- In “Change Templates”, search for “ostris”
- Select AI Tools template
- If training multiple models, click Edit Template and set container volume to 250 GB
- Click Deploy
Connecting to Your Pod
- When the pod finishes loading, click Connect
- Select port 8675
- Enter password:
password(default for this template) - Navigate to Datasets and upload all your files (images + text files in one folder)
Configuring Z-Image Training
Click on the New Job tab and configure:
| Setting | Value | Notes |
|---|---|---|
| Training Name | influencer_name_zimage | Your LoRA filename |
| Trigger Word | olivia1 | Must match caption prefix |
| Model Architecture | Z-Image De Turbo | For Z-Image LoRAs |
| Linear Rank | 32-44 | 44 for stronger LoRA |
| Training Steps | 2500-3000 | Start with 2500 |
| Resolutions | 1024, 1280 | Enable both |
In the Sample tab:
- Remove all example prompt boxes except one
- Paste a prompt from your captions but place your influencer in a new scenario (e.g., “sitting in a café”)
- These control images help monitor training progress
Click Create Job, then click the Play button (triangle) in the Dashboard.
Settings for WAN 2.2 (14B)
The main differences for WAN 2.2 training:
- This produces TWO files: High Noise and Low Noise — download both!
- In Sample tab: Set Number of Frames = 1 and FPS = 1
- Enable Skip First Sample (we’re training for images, not video)
- Everything else is the same as Z-Image
After a few hours, your LoRA models will be ready for download in the Dashboard under Checkpoints.
Step 4: Testing Your LoRA
Go back to Kitty AI Studio to test your newly trained LoRA.
Testing Z-Image LoRA
- Go to Z-Image Text to Image workflow
- Enter your prompt including your trigger word
- Check the LoRA box and upload your file
- Click Generate
Testing WAN LoRA
- Go to WAN 2.2 Text to Image
- Add BOTH the High Noise and Low Noise LoRA files
- Include your trigger word in the prompt
- Click Generate
Where Can You Use Your LoRA?
Your trained LoRA can be used in many workflows on Kitty AI Studio:
- Instagirl — Instagram-style photos
- Video workflows — AI videos with your character
- Long videos with SVI — Extended content generation
- And many more!
Conclusion
You now have a complete toolkit for creating AI influencers. You can monetize this content on various platforms and social media. The key is consistency — your trained LoRA ensures your AI character looks the same across all generations.