Backend: lm_studio
Endpoint: http://localhost:1234/v1
Model: llama3 (or whatever model you have loaded)
Temperature: 0.7 (recommended for balanced creativity)
Backend: ollama
Endpoint: http://localhost:11434/api
Model: llama3 (or llama3.1, mistral, etc.)
Temperature: 0.7
Backend: lm_studio or ollama
Endpoint: http://192.168.1.100:1234/v1 (replace with your IP)
Model: [your model name]
Temperature: 0.7
Backend: qwen3_vl
Model: Qwen/Qwen3-VL-4B-Instruct@4bit (or local path)
Endpoint Overrides: quant=4bit;attn=sdpa;device=cuda:0
Temperature: 0.6 (captioning focused)
Set this backend inside the image-to-video or image-to-image nodes when you want to caption reference images without calling an HTTP server. Install transformers, accelerate, huggingface_hub, and (optionally) bitsandbytes beforehand.
- Preset: cinematic
- Tier: advanced or cinematic
- Temperature: 0.5-0.7
- Preset: stylized or surreal
- Tier: enhanced or advanced
- Temperature: 0.8-1.0
- Preset: action
- Tier: advanced
- Temperature: 0.6-0.8
- Preset: noir
- Tier: advanced or cinematic
- Temperature: 0.5-0.7
- Preset: random
- Tier: auto
- Temperature: 1.0-1.5
- 0.1-0.3: Very focused, consistent, less creative
- 0.4-0.6: Balanced, reliable, some creativity
- 0.7-0.9: Creative, varied, good for exploration
- 1.0-1.5: Highly creative, experimental, more random
- 1.5-2.0: Very experimental, unpredictable
Positive Keywords: myLoRA_style, detailed_face, high_quality
Positive Keywords: cinematic, professional, 4k, high-end
Positive Keywords: red_hair, blue_eyes, leather_jacket
Negative Keywords: cartoonish, anime, illustration, painting
Basic Prompt: "woman walking in rain"
Preset: cinematic
Tier: auto (will detect "enhanced")
Variations: 1
Save: trueBasic Prompt: "robot in futuristic city"
Preset: custom
Tier: advanced
Variations: 3
Save: trueThen manually try with different presets to compare.
Basic Prompt: "slowly turns head and smiles"
Mode: image-to-video
Preset: cinematic
Tier: enhanced
Positive Keywords: [your lora trigger]Basic Prompt: "superhero lands from sky"
Preset: action
Tier: cinematic
Temperature: 0.8
Variations: 2filename_base: project_alpha_scene01
filename_base: project_alpha_scene02
filename_base: cinematic_city_night
filename_base: noir_detective_walk
filename_base: character_hero_intro
filename_base: character_villain_reveal
- Let Auto-Detect Work: The auto tier detection is quite good
- Start Small: Begin with basic prompts and increase complexity as needed
- Use Presets: They dramatically influence the output style
- Iterate: Generate variations and pick the best
- Save Everything: Build a library of good prompts for reference
- Keywords First: Always include LoRA triggers in positive_keywords
- Temperature Matters: Lower for consistency, higher for creativity
- Mode Matters: image-to-video mode focuses on motion, not scene setup
- Lower the tier (cinematic → advanced → enhanced)
- Use more concise basic_prompt
- Some video models have token limits
- Increase tier (basic → enhanced → advanced)
- Use more descriptive basic_prompt
- Try higher temperature for more elaboration
- The node auto-appends missing keywords
- Check spelling and comma separation
- Verify in the output prompt
- Try different presets
- Adjust temperature
- Add style terms to positive_keywords
- Use a smaller/faster model
- Reduce max_tokens (edit node if needed)
- Lower temperature slightly
- Llama 3 8B
- Mistral 7B
- Llama 3 70B
- Mixtral 8x7B
- Llama 3.1 variants
- Qwen models
- Custom fine-tuned models
The output connects directly to most video generation nodes:
[AI Video Prompt Expander]
↓ positive_prompt_1
[CogVideoX]
[Runway Gen-3]
[Pika Labs]
[Stable Video Diffusion]
etc.
Remember to also connect the negative_prompt output!