1. Introduction
Purpose of the Report
This report provides an expert-level guide on optimizing KSampler and XLabs Sampler settings within the ComfyUI environment, with a specific focus on three distinct models: RealVisXL5.0LightningBadedVae, RealVisXL5.0BakedVae, and Flux1-dev-Q5_K_S.
The objective is to equip users with a comprehensive understanding of these models' unique characteristics, the methodologies for adapting sampler settings when incorporating LoRAs (Low-Rank Adaptations) and ControlNet, and the nuances of model-specific prompting techniques.
Target Audience and Scope
The content herein is tailored for intermediate to advanced ComfyUI users who aim to deepen their technical understanding and enhance their control over the image generation process.
The scope is specifically delineated to the KSampler and XLabs Sampler, the aforementioned models, character and style LoRAs, and Depth and OpenPose ControlNet types, all within the ComfyUI framework.
Brief Overview of Models and Samplers
The models under consideration represent significantly different architectural philosophies.
The RealVisXL series are Stable Diffusion XL (SDXL) based models, renowned for their photorealistic output.6
The "Lightning" variant is optimized for rapid generation, while the "BakedVAE" versions integrate the Variational Autoencoder directly into the model file for streamlined workflows.7
In contrast, Flux1-dev-Q5_K_S is a derivative of the FLUX.1 architecture, which employs a rectified flow transformer, offering a different approach to image synthesis, with the GGUF quantization (Q5_K_S) specifically aimed at reducing VRAM requirements.8
Similarly, the samplers serve distinct roles. The KSampler is a versatile, general-purpose tool fundamental to most diffusion model workflows in ComfyUI.9
The XLabs Sampler, on the other hand, appears to be more specialized, potentially tailored for or originating from development around XLabs/FLUX models, offering parameters that cater to their unique characteristics.10
Understanding these foundational differences is paramount to effectively leveraging their capabilities.11
2. Understanding Core Sampler Functionality in ComfyUI
2.1. KSampler: A Comprehensive Parameter Guide
The KSampler node is a cornerstone of image generation within ComfyUI.12
Its primary function is to iteratively refine a latent image through a denoising process, guided by textual prompts and the conditioning provided by a chosen model.1 It is the principal engine for translating abstract concepts into visual representations in many diffusion model workflows.13
A detailed examination of its parameters is essential:14
- model: This input specifies the core diffusion model checkpoint (e.g., RealVisXL5.0BakedVae, Flux1-dev-Q5_K_S).
- The choice of model is foundational, dictating the inherent style, content generation capabilities, and overall characteristics of the output image.15
- All subsequent parameter tuning is performed in the context of the selected model.1
- seed: An integer value that initializes the random number generator used in the denoising process.16
- Utilizing the same seed with identical parameters ensures reproducible outputs.17
- The control_after_generation option (offering choices like randomize, increment, decrement, fixed) dictates how the seed behaves across batch generations, which is crucial for iterative refinement and exploring variations around a consistent base.1
- steps: This integer defines the number of denoising iterations the sampler will perform.18
- A higher number of steps generally leads to images with greater detail and coherence but correspondingly increases computation time.1 Typical ranges vary widely depending on the model and sampler;19
- standard models might use 15-50 steps, while highly optimized "Lightning" or "Turbo" variants can achieve excellent results with significantly fewer, sometimes as low as 4-6 steps.6 The selection of steps is a critical balance between desired image quality and generation speed.20
- cfg (Classifier-Free Guidance Scale): This floating-point value controls the degree to which the generated image adheres to the provided text prompt.21
- Higher CFG values (e.g., 7.0-12.0 for many SDXL models) enforce stricter prompt following, potentially enhancing detail and specificity.1 However, excessively high values can lead to oversaturated colors, undesirable artifacts, or an "overcooked" appearance.22
- Conversely, lower CFG values (e.g., 1.0-4.0) grant the model more creative latitude, resulting in more diverse or unexpected outputs.23
- Such low values are often employed with highly optimized models (like Lightning variants) or when strong external conditioning, such as from ControlNet, is present.5 The CFG scale is a primary lever for artistic control and modulating prompt fidelity.24
- sampler_name: This parameter specifies the algorithm used for the denoising process.25
- The choice of sampler has a profound impact on the visual style, texture, sharpness, and convergence speed of the generated image.26
- Common samplers include:27
- euler & euler_a (ancestral): The standard euler sampler is known for its speed and predictable results, making it suitable for quick previews or scenarios where a high step count is not desired.27
- The euler_a variant introduces additional randomness at each step, often leading to more varied and creatively diverse outputs.11
- dpmpp_2m_karras: A highly popular choice, particularly for models aiming for photorealism like RealVisXL.28
- It is a DPM++ (Diffusion Probabilistic Model Solver) variant that incorporates Karras noise scheduling, often yielding images with high detail and quality.7
- dpmpp_sde_karras: Another strong contender frequently recommended for RealVisXL models, especially their Lightning versions.6 The "SDE" (Stochastic Differential Equation) component suggests a solver that can offer greater diversity in outputs.29
- Other samplers like ddim and plms are generally considered older but were foundational.11 unipc is noted for providing a good balance of speed and quality.12 Extensive experimentation with different samplers is often necessary to achieve specific artistic effects.30
- scheduler: The scheduler dictates how the noise level (sigma) is managed and decreased across the sampling steps, working in close conjunction with the chosen sampler_name.1
- karras: Frequently paired with DPM-based samplers (like dpmpp_2m_karras or dpmpp_sde_karras).31
- Karras scheduling is known for improving image quality by making finer adjustments to noise levels, particularly towards the end of the sampling process, which can enhance detail.11
- normal, simple, exponential, linear: These represent different mathematical approaches to distributing the denoising effort across the steps.16 For instance, the simple scheduler is often mentioned in conjunction with the euler sampler for FLUX GGUF models.19
- beta: This scheduler is mentioned for use with the euler sampler in the context of FLUX GGUF models.20 The choice of scheduler can subtly or significantly alter the final image, influencing aspects like smoothness, detail rendition, and the overall character of the noise reduction.32
- The interplay between sampler_name and scheduler is critical.33
- They are not independent choices but form a cohesive system.34
- The sampler defines how the latent space is traversed and noise is removed at each step, while the scheduler defines what the noise levels (sigmas) are at those respective steps.35
- Optimal results typically arise from synergistic pairings, often explicitly recommended in model documentation (e.g., DPM++ samplers with the Karras scheduler for RealVisXL models6).36
- Using incompatible or poorly matched sampler-scheduler pairs can lead to suboptimal image quality or even processing errors.16
- positive & negative: These inputs receive conditioning data, typically from CLIP Text Encode nodes.37
- The positive conditioning guides the generation towards desired attributes, themes, and objects described in the prompt, while the negative conditioning steers it away from undesired elements.1
- latent_image: This is the starting point for the KSampler.38
- For text-to-image generation, it's typically a tensor of random noise generated by an Empty Latent Image node.39
- For image-to-image tasks, it's the latent representation of an input image, obtained from a VAE Encode node.1
- denoise: A floating-point value between 0.0 and 1.0 that controls the intensity of the denoising process applied to the latent_image.1
- For standard text-to-image generation (starting from pure noise), denoise is typically set to 1.0, signifying that the sampler will fully denoise the initial random latent to create a new image.22
- In image-to-image workflows, upscaling processes (like Hires.fix), or multi-pass refinement techniques, a denoise value less than 1.0 is used.40
- This allows the sampler to retain some structure or information from the input latent image while still modifying it based on the prompts and model.3 For example, RealVisXL Hires.fix parameters consistently recommend low denoise values like 0.1-0.5.6 The denoise parameter's optimal setting is highly context-dependent.41
- It dictates how much of the initial latent state is preserved versus how much is newly generated.42
- A value of 1.0 effectively means the input latent_image (if it's not pure noise) is completely replaced by noise before sampling begins.23
- KSampler (Advanced) / Iterative Mixing KSampler Advanced: These nodes offer more granular control over the sampling process.43
- The KSamplerAdvanced typically includes parameters like start_at_step and end_at_step to define a specific portion of the sampling process, and an add_noise toggle (which, if disabled for subsequent samplers, prevents re-introduction of noise16).44
- The Iterative Mixing KSampler Advanced provides even more specialized controls for blending and denoising, such as alpha_1, stop_blending_at_pct, blending_schedule (e.g., cosine), and blending_function (e.g., slerp).21 These advanced samplers are particularly useful for complex workflows, such as multi-stage refinement, controlled upscaling with tools like ControlNet, or intricate noise scheduling.21 For instance, one might use an Iterative Mixing KSampler for an initial generation at low resolution, then pass its output (potentially with some residual noise) to another sampler for upscaling and refinement, possibly guided by ControlNet.25
2.2. XLabs Sampler & FLUX-Specific Sampling Nodes
Models based on the FLUX.1 architecture, owing to their distinct rectified flow transformer design26, often benefit from a specialized ecosystem of sampling tools and parameters within ComfyUI.46
These tools appear designed to better accommodate the unique characteristics of FLUX models compared to relying solely on the generic KSampler.47
- XLabs Sampler Node: This node seems to be a specialized sampler, possibly an enhanced KSampler or a custom wrapper, tailored for models like FLUX.1.28 While sharing some common parameters with the standard KSampler (model, conditioning, neg_conditioning, noise_seed, steps, denoise_strength, image_to_image_strength), it introduces unique controls28:
- timestep_to_start_cfg: An integer that defines the step at which Classifier-Free Guidance begins to take effect.48
- This implies an ability to run initial sampling steps with potentially different or no CFG guidance, which could be advantageous for the convergence properties of FLUX models.49
- true_gs (True Guidance Scale): A float value (0-100) that governs the strength of style-guided synthesis.50
- It is described as being "scaled to the CFG scale"29, suggesting it might modulate the standard CFG or act as an additional, FLUX-specific guidance mechanism.51
- controlnet_condition (optional): An input for integrating ControlNet data. Significantly, the XLabs Sampler is noted as not using the sampler_name and scheduler dropdowns typical of the standard KSampler.29 This suggests it may employ fixed internal sampler/scheduler logic optimized for FLUX models or utilize an entirely different mechanism for controlling the sampling process.52
- FL_KsamplerSettings Node (ComfyUI_Fill-Nodes): This node provides a structured two-pass KSampler configuration that is often associated with FLUX model workflows.30 Its parameters include batch, Pass_1_steps, Pass_2_steps, Pass_1_CFG, Pass_2_CFG, Pass_2_denoise, and scale_factor.53
- The presence of this node suggests that a common operational pattern for FLUX models involves an initial generation pass followed by a distinct refinement or upscaling pass, each with its own tailored settings, conveniently encapsulated by this single node.54
- ModelSamplingFlux Node: This utility node appears to prepare FLUX models for the main sampling process by dynamically adjusting sampling parameters based on the model's target resolution.31 Key parameters are max_shift, base_shift, width, and height.55
- For FLUX.1 Schnell, a max_shift between 0.0 and 2.0 is mentioned, with the caveat that bypassing the node might sometimes yield better results.20 This node acts as a pre-sampler configuration step, potentially optimizing how the subsequent KSampler or XLabs Sampler interacts with the FLUX model, especially concerning resolution-dependent characteristics.56
The existence and specific functionalities of these nodes—XLabs Sampler with its unique guidance parameters, FL_KsamplerSettings for multi-pass operations, and ModelSamplingFlux for resolution-adaptive pre-configuration—strongly indicate that achieving optimal results with FLUX models in ComfyUI involves leveraging this specialized toolkit.57
Standard KSampler settings, while potentially functional, might not fully harness the capabilities of the FLUX architecture or could even be incompatible with certain advanced features offered by these custom nodes.58
The x-flux-comfyui custom node suite further reinforces this by providing specific integrations for FLUX models, LoRAs, and ControlNets.33
3. Model-Specific Deep Dive: Settings and Characteristics
Understanding the specific characteristics of each model is paramount for tailoring sampler settings effectively.59
The three models in focus—RealVisXL5.0LightningBadedVae, RealVisXL5.0BakedVae (Non-Lightning), and Flux1-dev-Q5_K_S—represent distinct approaches to image generation, each requiring nuanced KSampler configurations.60
3.1. RealVisXL V5.0 Lightning (BakedVAE)
Core Characteristics:
- This model is built upon the SDXL architecture and is highly optimized for generating photorealistic images.8
- The "Lightning" designation signifies its design for extremely fast inference times.61
- This is often achieved through techniques such as progressive adversarial diffusion distillation, which allows the model to produce high-quality outputs in a minimal number of sampling steps (e.g., 4-6).7
- The term "BakedVAE" indicates that the Variational Autoencoder is integrated directly into the main model file.62
- This simplifies the workflow in ComfyUI as it obviates the need for a separate VAE Loader node, a common requirement for many other SDXL models.7
Recommended KSampler Settings:
Synthesizing information from various sources6, the following KSampler settings are consistently recommended for RealVisXL V5.0 Lightning:
- steps: Very low, typically in the range of 4 to 6 steps.63
- Some sources suggest up to 10 steps may be viable.64
- cfg: Also very low, usually between 1.0 and 2.0.64
- sampler_name: dpmpp_sde_karras or dpmpp_sde are the predominant recommendations.65
- scheduler: karras is implicitly used when dpmpp_sde_karras is selected, or it's the generally advised scheduler for these samplers.65
- denoise: 1.0 for standard text-to-image generation.66
VAE Considerations:
- While the baked-in VAE offers convenience, the quality of this integration is crucial.66
- Historically, the initial SDXL 1.0 release had VAE issues that produced visual artifacts, leading many users to prefer the older SDXL 0.9 VAE.39 RealVisXL models are generally known for good quality VAE bakes.67
- However, if unexpected artifacts (like color shifts or blurriness) appear, and a separate, known-good SDXL VAE is available, advanced users might experiment by loading it explicitly, though this negates some of the "baked-in" convenience.68
Hires Fix:
- When using Hires Fix with RealVisXL Lightning, recommendations include a low denoise strength (e.g., 0.5), a small number of Hires steps (e.g., 3), and upscalers like 4x-NMKD-Superscale-SP_178000_G or 4x-UltraSharp.6
The extremely low step counts and CFG scales characteristic of Lightning models enable their remarkable speed.69
However, this optimization inherently means they might offer less flexibility for intricate prompt adjustments or achieving highly specific artistic deviations compared to their non-Lightning counterparts.70
They are "distilled" to excel at producing high-quality images rapidly, often within a specific (typically photorealistic) domain.8 While excellent for quick iterations and high-throughput generation, users seeking maximal artistic control over complex scenes might find standard models more adaptable, albeit at the cost of speed.71
3.2. RealVisXL V5.0 BakedVAE (Non-Lightning)
Core Characteristics:
- Like its Lightning sibling, this model is SDXL-based and targets photorealistic image generation.7
- It follows a standard generation process, not specifically optimized for the extreme speed of Lightning variants.72
- This generally allows for more sampling steps and potentially finer detail.73
- "BakedVAE" again signifies an integrated VAE, simplifying setup.7 It is expected to use a high-quality VAE component, akin to the improved SDXL 0.9 VAE, to avoid known issues.74
Recommended KSampler Settings:
Based on several sources7, recommended settings for the non-Lightning RealVisXL V5.0 are:
- steps: Significantly higher than Lightning models.75
- Typically 15-30+ steps when using dpmpp_sde_karras, and potentially 30-50+ steps for dpmpp_2m_karras.76
- Some examples show good results with 20 steps.42 One source shows 8 steps with dpmpp_sde_karras but with a very low denoise of 0.35, suggesting an image-to-image or refinement context.44
- cfg: Generally, non-Lightning SDXL models utilize a higher CFG scale.77
- A common starting point is around 7.0-9.0.5 Model card recommendations for RealVisXL V4.0/V5.0 sometimes suggest an ideal range of 7.5-12.5.14 However, some sources for RealVisXL V5.0 standard strangely recommend a low CFG of 1.0-2.57, which is atypical for non-Lightning SDXL and might be an outlier or refer to a specific use case, as other documentation from the same creator for the non-Lightning V5.0 does not specify such a low CFG.45 Given general SDXL practices, a CFG in the 7.0-9.0 range is a more conventional starting point for non-Lightning variants.78
- sampler_name: dpmpp_2m_karras or dpmpp_sde_karras are the primary recommendations.79
- scheduler: karras is typically used, often implied by the sampler choice.79
- denoise: 1.0 for standard text-to-image generation.80
VAE Considerations:
- The baked-in VAE simplifies the workflow.80
- The emphasis for these models is on incorporating a high-quality VAE to prevent artifacts like those seen with the original SDXL 1.0 VAE.39
Hires Fix:
- For Hires Fix, a low denoise strength (0.1-0.3), a higher number of Hires steps (e.g., 25+ for dpmpp_2m_karras), and similar upscalers (e.g., 4x-NMKD-Superscale-SP_178000_G, 4x-UltraSharp) are advised.7
Key Differences from Lightning Variant:
- Requires substantially more sampling steps.81
- Typically employs a higher CFG scale for stronger prompt adherence.82
- Results in slower generation speeds but offers the potential for greater detail, more nuanced responses to complex prompts, and finer artistic control.8
The convenience of a BakedVAE is a significant usability improvement for many users, as it removes a step in the workflow setup and ensures a VAE known to work well with the model is used.39 However, this convenience comes at the cost of flexibility.83
If the baked-in VAE exhibits subtle issues in certain scenarios, or if a superior community-developed VAE for SDXL becomes available, users cannot easily swap it out.84
Model creators would need to release a new version with the updated VAE.85
For most users, a high-quality baked VAE is a net benefit.86
Advanced users or those troubleshooting specific visual artifacts might perceive this as a limitation.87
RealVisXL models are generally well-regarded, suggesting their VAE integration is competently executed.7
3.3. Flux1-dev-Q5_K_S
Architecture Overview:
- FLUX.1 models, including Flux1-dev, are architecturally distinct from Stable Diffusion models.89
- They are based on a rectified flow transformer design.26 This architecture is noted for its strong prompt adherence, impressive capability in rendering text within images, and high level of detail.26
- The dev version of FLUX.1 is an open-weight release intended for non-commercial use, with image quality reported to be comparable to the proprietary pro version.26 It is trained using guidance distillation, a technique that enhances generation efficiency.26
GGUF Quantization (Q5_K_S):
- GGUF (Generalized Graphical Unified Format) is a file format designed to allow large models to run more efficiently, particularly on systems with limited VRAM or on CPUs, by using quantized (lower-precision) weights.19
- The Q5_K_S designation for this specific quantization level signifies:91
- Q5: The weights are quantized to approximately 5 bits.91
- _K: It utilizes "K-quants," an advanced quantization method generally offering better preservation of model quality for a given bit depth compared to simpler techniques.92
- _S: This likely refers to a "small" variant within the K-quant methodology, possibly related to superblock sizes or other internal quantization parameters.93
- Impact: The Flux1-dev-Q5_K_S.gguf model file is approximately 8.26 GB.20 This is a substantial reduction in size and VRAM requirement compared to the full F16 (16-bit floating point) version, which is around 23.8 GB.49 The Q5_K_S version is often recommended for systems with around 12GB of VRAM.20 While quantization offers efficiency gains (potentially faster inference and lower memory use50), it inherently involves a trade-off with potential minor degradation in output quality compared to the full-precision model.94
- Advanced quantization schemes like K-quants aim to minimize this fidelity loss.20
Setup:
- Using GGUF models in ComfyUI requires the ComfyUI-GGUF custom node suite and specific directory placements for the UNET GGUF file and its associated CLIP model files.19
Recommended KSampler/XLabs Sampler Settings:
Drawing from various community discussions and guides19:
- Sampler Type: While standard KSampler can be used, specialized nodes like the XLabs Sampler28 or the FL_KsamplerSettings node30 might be employed in specific FLUX workflows.95
- steps: For the full, non-GGUF FLUX.1 dev model, 20-50 steps are typical.27 For GGUF versions, especially those with lower quantization levels (like Q5) or the "Schnell" (fast) variant, step counts can be lower.96
- Examples for Flux Dev GGUF show usage with 20-35 steps when paired with the euler sampler.19 The FLUX.1 Schnell variant can generate images in as few as 4 steps.48 The Q5_K_S GGUF version likely sits between the full dev and Schnell in terms of optimal step count.97
- cfg: The guidance distillation used in training FLUX.1 dev26 might imply less reliance on very high CFG values.98
- The FL_KsamplerSettings node defaults to a CFG of 6.0 for its passes.30 For the FLUX.1 Schnell variant, CFG is often set very low (e.g., 1.0 when using KSampler with euler/beta20).99
- The optimal CFG for Flux1-dev-Q5_K_S may require experimentation but is likely to be moderate.100
- sampler_name: euler is frequently cited for GGUF and Schnell versions of FLUX models.19
- scheduler: simple19 or beta20 are mentioned in conjunction with the euler sampler for GGUF/Schnell FLUX models.101
- denoise: 1.0 for text-to-image.102
- XLabs Sampler Specifics (if used): Parameters like timestep_to_start_cfg (when CFG applies) and true_gs (style guidance strength) become relevant.28
Specific ComfyUI Nodes for FLUX:
- Essential for GGUF: Unet Loader (GGUF) and DualCLIPLoader (GGUF) are necessary for loading the quantized UNET and its corresponding text encoders.19
- Standard Components: Load Diffusion Model (for non-GGUF.safetensors UNET), DualCLIPLoader (for standard CLIP models like t5xxl_fp16/fp8 and clip_l), and Load VAE (using ae.safetensors specifically for FLUX) are standard parts of FLUX workflows.49
- FLUX-Specific Utilities:
GGUF quantization, as seen in Flux1-dev-Q5_K_S, represents a pragmatic compromise.103
It drastically reduces VRAM demands and can accelerate inference on compatible hardware, making powerful models like FLUX.1 dev accessible to a broader range of users.20 However, this efficiency is achieved by representing model weights with fewer bits, which inherently involves some loss of information compared to full-precision (FP16/F32) versions.52 While advanced quantization methods like K-quants (indicated by _K_S) aim to minimize this fidelity reduction20, a perceptible difference in output quality compared to the original 23.8GB F16 model may exist.104
Users with ample VRAM seeking maximum quality would typically prefer higher-bit GGUF quantizations (e.g., Q8_0) or the full-precision versions.105
The choice of Q5_K_S is thus a balance between hardware constraints and the desired level of output fidelity.106
| Feature | RealVisXL V5.0 Lightning (BakedVAE) | RealVisXL V5.0 BakedVAE (Non-Lightning) | Flux1-dev-Q5_K_S |
|---|---|---|---|
| Base Architecture | SDXL | SDXL | FLUX.1 Rectified Flow Transformer |
| VAE Type/Handling | Baked-in | Baked-in | Separate (ae.safetensors), loaded via Load VAE node |
| Typical Resolution | SDXL Native (e.g., 1024x1024) | SDXL Native (e.g., 1024x1024) | Flexible, often 1024x1024, GGUF may vary |
| Recommended Sampler(s) | dpmpp_sde_karras, dpmpp_sde | dpmpp_2m_karras, dpmpp_sde_karras | euler (for GGUF) |
| Typical Steps | 4-10 | 15-50+ | 20-35 (for GGUF with euler), varies for full dev |
| Typical CFG Range | 1.0-2.0 | 7.0-9.0 (general SDXL), some sources vary | Moderate (e.g., ~6.0), very low for Schnell (e.g., 1.0) |
| Key Prompting Style | SDXL (Keywords, Weights, Negatives) | SDXL (Keywords, Weights, Negatives) | Natural Language, specific syntax (no SD weights) |
Note: "Typical" values are starting points; experimentation is always encouraged. CFG for non-Lightning RealVisXL V5.0 has some conflicting source data, with general SDXL advice pointing to higher CFG.108
4. Synergizing Samplers with Advanced Techniques
The KSampler and its variants do not operate in isolation.109
Their behavior and optimal settings can be significantly influenced by the integration of other advanced techniques like LoRAs and ControlNet, which add further layers of conditioning to the image generation process.110
4.1. LoRA Integration (Character & Style)
LoRAs (Low-Rank Adaptations) are small, trained files that modify the behavior of a base diffusion model and/or its text encoder (CLIP) to introduce specific characters, artistic styles, or concepts without retraining the entire model.2 They act as an additional, targeted form of conditioning.111
ComfyUI LoRA Loading:
- In ComfyUI, LoRAs are typically loaded using a Load LoRA node.112
- This node takes the base model and CLIP model as inputs and outputs modified versions.113
- If multiple LoRAs are used, they are usually chained sequentially.114
- Key parameters within the Load LoRA node are strength_model and strength_clip, which control the intensity of the LoRA's influence on the UNet and CLIP components, respectively.57 These strengths typically range from 0.0 (no effect) to 1.0 (full effect), though values outside this range are sometimes used cautiously.115
Adjusting KSampler Settings with LoRAs:
- steps: The use of LoRAs generally does not necessitate a drastic change in the number of sampling steps, assuming the LoRA is well-trained and the base model's step count is already optimized.116
- However, if a LoRA introduces instability or noise, a slight increase in steps might help stabilize the output.117
- The choice of scheduler can also interact with LoRAs, potentially requiring step adjustments for optimal results.60
- cfg: The CFG scale is often adjusted when LoRAs are active.118
- If a LoRA strongly imparts its intended style or character, it provides significant guidance on its own.119
- In such cases, the CFG scale for the text prompt might be slightly reduced.120
- This is because the LoRA is already "steering" the generation process heavily, and a very high CFG might lead to conflicts or an over-processed look.5 Conversely, if a LoRA's effect is too subtle at its default strength, a marginal increase in CFG (or, more commonly, an increase in the LoRA's own strength_model/strength_clip values) might be considered.121
- Some LoRAs or sampler combinations might have very specific CFG requirements;122
- for example, one user noted a particular setup required a CFG of 4 to avoid oversaturation.61
- sampler_name / scheduler: The base model's preferred sampler and scheduler combination usually remains effective when LoRAs are introduced.123
- However, some LoRAs might exhibit better synergy with specific samplers.124
- If default choices yield unsatisfactory results with a LoRA, experimenting with alternative samplers or schedulers can be beneficial.59
- denoise: This parameter's relevance with LoRAs is most pronounced in image-to-image or inpainting workflows.125
- If a LoRA is used to modify an existing image (e.g., apply a style or change a character's appearance), the denoise strength will control how much the LoRA-influenced generation process alters the original input latent.126
Model-Specific LoRA Considerations:
- RealVisXL (SDXL-based): A vast number of LoRAs trained for SDXL are available.127
- General SDXL LoRA usage principles apply, focusing on compatibility with the SDXL architecture.128
- Flux1-dev: For FLUX models, it is crucial to use LoRAs specifically trained for the FLUX.1 architecture.129
- XLabs has provided such LoRAs (e.g., flux-RealismLora, flux-lora-collections, flux-furry-lora).33 Using these dedicated LoRAs ensures compatibility and optimal performance.130
- Community workflows also demonstrate the use of LoRAs with GGUF versions of FLUX models to guide artistic direction.20
When a LoRA, particularly a strong style or character LoRA, is active, it behaves somewhat like a "soft" ControlNet.131
It heavily biases the diffusion process towards specific aesthetic qualities or features.132
This inherent guidance from the LoRA means that the KSampler's CFG scale might often be adjusted downwards.133
The LoRA is already contributing significantly to "steering" the image, so the text prompt may not need to be enforced as aggressively by a high CFG.134
Overly high CFG in conjunction with a potent LoRA can lead to "overbaking," where details become exaggerated or artifacts appear, as both the LoRA and the CFG attempt to exert maximum influence.135
A common heuristic is to start with a slightly lower CFG when a strong LoRA is applied and then fine-tune the LoRA's own strength parameters and the CFG scale in tandem to achieve the desired balance.136
| LoRA Type | Typical LoRA Strength | Potential KSampler Adjustments | Model Family Considerations | Notes/Heuristics |
|---|---|---|---|---|
| Character | 0.6 - 1.0 | CFG: May slightly decrease.137 Steps: Usually no major change.138 | SDXL: Ensure LoRA is SDXL compatible. FLUX: Use FLUX-specific LoRAs. | Trigger words are often crucial.138 Balance LoRA strength with prompt details. High strength can cause "character bleed."139 |
| Style | 0.5 - 0.9 | CFG: May decrease.139 Steps: Usually no major change.140 | SDXL: Broad compatibility. FLUX: Use FLUX-specific style LoRAs if available. | Strong style LoRAs can dominate;140 lower strength or CFG may be needed for subtlety.141 |
| Concept/Object | 0.7 - 1.0 | CFG: May slightly decrease. Steps: Usually no major change.141 | SDXL: Good availability. FLUX: Less common, ensure compatibility.142 | Can be very effective for adding specific items or themes.142 |
4.2. ControlNet Application (Depth & OpenPose)
ControlNets provide explicit and strong spatial conditioning to the image generation process.143
They use input maps like depth maps or OpenPose skeletons to guide the composition, structure, and pose of elements within the generated image.3
ComfyUI ControlNet Loading:
- The typical workflow involves a Load ControlNet Model node to load the specific ControlNet model (e.g., for depth or OpenPose) and an Apply ControlNet (or ControlNetApplyAdvanced) node.144
- The Apply ControlNet node takes the positive and negative conditioning from the text prompts and the ControlNet conditioning map as inputs, and outputs modified conditioning signals that are then fed into the KSampler.62
Adjusting KSampler Settings with ControlNet:
- steps: The number of sampling steps can often be maintained at the base model's optimal level.145
- In some cases, if the ControlNet provides very clear and strong guidance, steps might be slightly reducible, as the model has a more defined path to follow.146
- However, complex ControlNet inputs or the need for fine detail might still benefit from an adequate number of steps.147
- For SD1.5 with Canny ControlNet, 20-30 steps are often recommended.64
- cfg: The CFG scale is frequently lowered when using strong ControlNets.64 Since the ControlNet dictates a significant portion of the image's composition and structure, a high CFG forcing strict adherence to the text prompt might become redundant, conflict with the ControlNet's guidance, or lead to an overly constrained output.148
- While 64 suggests a CFG of 7-8 for SD1.5 Canny, for very dominant ControlNet inputs, values even lower (e.g., 3-6) might be effective.149
- The key is that ControlNet itself is a powerful guidance mechanism.150
- sampler_name / scheduler: The base model's preferred sampler and scheduler combination generally remains effective.151
- Consistency in these choices helps in achieving predictable results. For SD1.5 with Canny ControlNet, dpmpp_2m with karras scheduler is a common recommendation.64
- denoise: This parameter is critically important when using ControlNet, especially in workflows that involve an initial image.152
- If ControlNet is used to guide the generation from an empty latent (i.e., text-to-image with structural guidance from a pose map or depth map not derived from an existing image), denoise should typically be 1.0.153
- If ControlNet is applied to an existing image (or its latent representation, e.g., in an image-to-image or refinement pass), denoise must be less than 1.0.154
- This allows the ControlNet to influence the structure of the original image while preserving some of its content.155
- Values between 0.5 and 0.8 are common, depending on how much alteration is desired.156
- A two-pass system often involves a first KSampler pass to establish pose/depth with ControlNet, followed by a second KSampler pass with a lower denoise (e.g., 0.4-0.6) for refinement or style application.25
Model-Specific ControlNet Considerations:
- RealVisXL (SDXL-based): A wide array of ControlNet models specifically trained for the SDXL architecture (including Depth, OpenPose, Canny, etc.) are available and should be used.157
- Flux1-dev: XLabs and the community have provided ControlNets tailored for the FLUX.1 architecture (e.g., flux-controlnet-canny, flux-controlnet-depth, flux-controlnet-hed).29 It is essential to use these FLUX-specific ControlNet models for compatibility and optimal results.158
- Workflows might involve nodes like Xlabs Sampler or Apply Flux ControlNet.29
ControlNet Strength and Timing:
- The ApplyControlNetAdvanced node in ComfyUI offers fine-grained control through parameters like strength (overall influence of the ControlNet) and start_percent / end_percent (which define the portion of the sampling steps during which the ControlNet influence is active).64 This allows for nuanced application, such as having ControlNet guide the initial structure and then fade its influence to allow more freedom for stylistic details.159
When ControlNets like OpenPose or Depth are employed, the burden on the text prompt to define composition and spatial arrangements is significantly reduced.160
The ControlNet provides this structural information directly. Consequently, the text prompt can focus more on defining aspects like character appearance, artistic style, color palettes, and overall ambiance.161
This shift can simplify the prompting process for complex scenes and often allows for the use of lower CFG values because the primary compositional guidance is derived from the ControlNet map rather than from aggressive enforcement of the text prompt.162
The KSampler's role then becomes more about "filling in the details" according to the prompt's stylistic and subject matter cues, within the robust framework established by the ControlNet.163
| ControlNet Type | Typical ControlNet Strength | Potential KSampler Adjustments | Model Family Considerations | Notes/Heuristics |
|---|---|---|---|---|
| Depth | 0.7 - 1.0 | CFG: Often lower (e.g., 3-7).164 Denoise: <1.0 for img2img/refinement (e.g., 0.5-0.8), 1.0 for txt2img. Steps: Usually no major change.165 | SDXL: Use SDXL Depth ControlNet.165 FLUX: Use FLUX Depth ControlNet.166 | Preserves 3D structure. denoise is key for how much original image (if any) is kept.166 |
| OpenPose | 0.8 - 1.0 | CFG: Often lower (e.g., 3-7). Denoise: <1.0 for img2img/refinement (e.g., 0.6-0.9), 1.0 for txt2img.167 Steps: Usually no major change.168 | SDXL: Use SDXL OpenPose ControlNet. FLUX: Use FLUX OpenPose ControlNet.168 | Excellent for character posing.168 Prompt focuses on character details and style, not pose description.169 |
5. Mastering Prompts Across Different Architectures
The way prompts are constructed and interpreted can vary significantly between different model architectures.170
Understanding these distinctions is crucial for effectively guiding RealVisXL (SDXL-based) and Flux1-dev models.171
5.1. Prompting for RealVisXL (SDXL-based Models)
SDXL models, including RealVisXL, generally exhibit a better understanding of natural language compared to their SD1.5 predecessors.172
However, they still benefit from structured prompting techniques to achieve precise and high-quality results.39
Key Prompt Elements:
A common structure for SDXL prompts includes67:
- Subject: The primary focus of the image, typically placed early in the prompt.173
- Environment: Details about the surroundings, including location, lighting conditions (e.g., "golden hour," "moonlight"), and weather (e.g., "rainy," "foggy").174
- Medium: The artistic medium or photographic style, such as "oil painting," "watercolor," "photograph shot on 85mm lens," "DSLR."175
- Style: References to artistic eras, specific artists, art movements (e.g., "impressionist," "surrealist"), or national styles.176
- Keyword Weighting: SDXL supports syntax for emphasizing or de-emphasizing specific keywords.177
- Parentheses () are often used for emphasis, and a colon with a numerical value can adjust the weight, e.g., (keyword:1.2) to increase its influence, or (keyword:0.8) to decrease it.178
- Square brackets `` can sometimes be used for de-emphasis.2
- Negative Prompts: These are highly important for SDXL models to steer the generation away from common undesirable artifacts or features.179
- Model creators for RealVisXL often provide specific recommended negative prompts, which might include terms like "(worst quality, low quality, illustration, 3d, 2d, painting, cartoons, sketch:1.3), (blurry:1.2), bad anatomy, deformed hands, ugly face" etc..6
- Detail and Specificity: Providing more descriptive detail in the prompt generally leads to outputs that more closely match the user's intent.180
- However, overly long, convoluted, or contradictory prompts can confuse the model and degrade quality.67
RealVisXL Specifics:
- As models designed for photorealism, prompts for RealVisXL often benefit from including photographic terminology.181
- This can include lens types (e.g., "85mm lens," "macro lens"), lighting descriptions (e.g., "softbox lighting," "cinematic lighting," "rim lighting"), camera settings (e.g., "f/1.8 aperture"), and general photographic qualities ("sharp focus," "depth of field," "DSLR quality").68
Iterative Prompting:
5.2. Prompting for Flux1-dev Models
FLUX.1 models, including Flux1-dev, are designed to excel with natural language instructions.184
The prompting paradigm often feels more like communicating a request to a human artist rather than assembling a string of keywords.69
Natural Language Focus:
- The primary strength of FLUX.1 models lies in their interpretation of descriptive, coherent sentences.69
Precision, Detail, and Structure:
- Prompts should clearly describe not only the content but also the desired tone, artistic style, color palette, and point of view.69
- For photorealistic images, including specific camera details (device name, aperture, lens type, shot type) is beneficial.69
- It is advisable to structure prompts hierarchically, describing elements from foreground to background in an organized manner.185
- Avoid adding elements out of sequence.69
No Traditional Prompt Weights:
- A crucial difference from SDXL is that FLUX.1 models (both dev and schnell variants) do not support the (keyword:weight) syntax commonly used in Stable Diffusion.69 To emphasize certain aspects, users should instead employ descriptive phrases within the natural language prompt, such as "with a strong emphasis on X," "focusing on Y," or by placing important descriptors prominently.186
"White Background" Issue (FLUX.1 dev only):
- A known quirk of the Flux1-dev variant (not affecting Flux1-schnell) is that including the phrase "white background" in the prompt can lead to fuzzy, blurry, or otherwise degraded outputs.69 It is best to avoid this specific phrase or use alternative descriptions for a clean backdrop.187
Avoiding Chaotic Prompting:
- Disorganized or ambiguous prompts can confuse FLUX.1 models, leading to outputs that technically fulfill the prompt but miss the user's intent.188
- If a specific attribute (e.g., color) applies to a particular element, it should be clearly associated (e.g., "a sign with green text" is preferable to "a sign, green, text").69 For complex scenes, breaking down the description into shorter, focused sentences, each addressing a specific element or aspect, can improve clarity and coherence.189
Text Rendering:
- FLUX.1 Dev is noted for its superior ability to render text accurately within images.27 Prompts can directly specify the text content and even request certain stylistic effects for the text, such as distortion, shadows, neon glows, or outlines.69
Negative Prompts:
- The role of negative prompts appears to be significantly different for FLUX.1 models compared to SDXL.190
- While SDXL heavily relies on negative prompts to refine outputs and avoid common issues, official FLUX.1 documentation and community guides place much less emphasis on them.26 One source even states that for a GGUF Schnell workflow, "Negative prompts are not reflected, so leave them blank".20 However, the XLabs Sampler node, which can be used with FLUX models, does include a neg_conditioning input.28 This suggests that the utility of negative prompts with FLUX might depend on the specific workflow, sampler node, and variant being used.191
- Generally, they seem far less critical than for SDXL.192
FLUX.1 Kontext (Image Editing):
- When using FLUX.1 Kontext models for image editing, prompts transform into instructions.192
- These instructions should be specific about the changes desired, e.g., "replace 'old text' with 'new text'," or "change the background to a bustling cityscape while keeping the person in the foreground unchanged".47 Clauses that specify what should be preserved (e.g., "while keeping the same facial features") are very important for controlled edits.193
The transition from prompting SDXL-based models to FLUX.1 models requires a notable paradigm shift.194
Users accustomed to the keyword-driven, heavily weighted, and negative-prompt-reliant style of SDXL must adapt their approach.195
FLUX.1 favors well-structured, descriptive natural language and has different syntactic rules (e.g., no SD-style weighting).196
Simply copying an SDXL prompt into a FLUX workflow is unlikely to yield optimal results and may even perform poorly due to these fundamental differences in how the models interpret textual input.197
| Feature | RealVisXL (SDXL-based) | Flux1-dev |
|---|---|---|
| Primary Language Style | Keywords, phrases, some natural language | Descriptive natural language, full sentences |
| Keyword Weighting | Supported, e.g., (keyword:1.2), [keyword] | Not supported;198 use phrasing like "emphasis on..."199 |
| Negative Prompt Importance | High; crucial for quality and avoiding artifacts | Low to moderate;199 some workflows may not use them at all200 |
| Text Rendering Capability | Moderate; can be challenging | Strong;200 excels at rendering text accurately201 |
| Handling of "White Background" | No specific known issues | dev variant can produce fuzzy output;201 avoid phrase or use workarounds202 |
| Layer/Element Description | Often relies on keyword order and weighting | Prefers hierarchical, clear association of attributes to elements |
6. Comparative Analysis and Practical Recommendations
A clear understanding of the distinct attributes of RealVisXL Lightning, RealVisXL BakedVAE (Non-Lightning), and Flux1-dev-Q5_K_S allows for informed decisions regarding their application and optimization.202
Summary of Key Differences:
- Architecture: RealVisXL models are SDXL-based (U-Net diffusion), while Flux1-dev utilizes a rectified flow transformer.203
- This fundamental difference impacts prompting, internal workings, and often the types of details they excel at.204
- Speed: RealVisXL Lightning is engineered for maximum speed with minimal steps. Flux1-dev (full precision) and RealVisXL Non-Lightning are slower.205
- Flux1-dev-Q5_K_S (GGUF) aims for a balance, offering improved speed over full precision FLUX on compatible hardware, especially with VRAM constraints.206
- Typical Settings: Lightning models use very low steps (4-10) and CFG (1-2).207
- Non-Lightning SDXL typically uses more steps (15-50+) and higher CFG (7-9).208
- Flux1-dev GGUF often uses moderate steps (20-35 with Euler) and potentially moderate CFG, though its "Schnell" variant can also use very low steps/CFG.209
- VRAM Requirements: RealVisXL models (SDXL based) have significant VRAM needs. Flux1-dev (full) is also large (approx. 23GB).210
- Flux1-dev-Q5_K_S drastically reduces this to ~8.3GB, making it accessible on systems with ~12GB VRAM.20
- Prompting Style: RealVisXL follows SDXL conventions (keywords, weights, heavy reliance on negative prompts).211
- Flux1-dev prefers natural language, does not use SD-style weights, and has a different approach to negative prompts.212
Use Case Suitability:
- RealVisXL Lightning: Ideal for scenarios demanding rapid generation of photorealistic images, such as quick iterations, previews, concept exploration, or high-volume social media content creation.213
- Users must be comfortable with the constraints of minimal steps and low CFG, which might offer less fine-grained control over highly nuanced prompts.8
- RealVisXL BakedVAE (Non-Lightning): Best suited for producing high-quality, detailed photorealistic images where longer generation times are acceptable in exchange for finer control, more detailed outputs, and potentially better responsiveness to complex prompting.7 Excellent for final renders, artistic compositions, and print-quality work.214
- Flux1-dev-Q5_K_S: A strong choice for users with VRAM limitations (e.g., 12GB GPUs) who seek high prompt adherence, robust text-in-image generation, and the unique characteristics of the FLUX architecture.20 Its natural language prompting can be more intuitive for some users.215
- The Q5_K_S quantization offers a practical compromise between the full dev model's quality and the schnell variant's speed and efficiency.216
Workflow Efficiency Tips:
- Always begin with the recommended KSampler settings provided on model cards (e.g., from Civitai, Hugging Face) or by reliable community sources as a baseline.217
- For initial exploration and prompt iteration (unless using Lightning or Schnell models), consider using lower step counts and moderate CFG to save time.218
- Once a desirable composition is achieved, increase steps and fine-tune CFG for the final render.219
- ComfyUI's node-based, non-destructive workflow is ideal for A/B testing. Duplicate KSampler nodes or branches to easily compare the effects of different sampler settings, LoRA combinations, or ControlNet strengths without rebuilding the entire workflow.220
- When using FLUX GGUF models like Flux1-dev-Q5_K_S, ensure that the ComfyUI-GGUF custom node is correctly installed and that all required model files (UNET GGUF, CLIP GGUF, CLIP L, VAE) are placed in their designated ComfyUI subdirectories (models/unet, models/clip, models/vae).19
Troubleshooting Common Issues:
Blurry/Noisy Images:
- Increase steps.221
- Adjust cfg (too low can be blurry, too high can introduce noise/artifacts).222
- Try a different sampler_name or scheduler.222
- Ensure denoise is appropriate for the task (1.0 for txt2img).223
- For Flux1-dev, specifically check the prompt for the phrase "white background," as this can cause blurriness.69
Poor Prompt Adherence:
- Increase cfg (within reasonable limits, e.g., not exceeding 12-15 for most SDXL models unless specifically required).224
- Refine prompt clarity and specificity. Ensure keywords are unambiguous.225
- For FLUX, ensure prompts are well-structured natural language.225
- Check LoRA or ControlNet strengths; if too high, they might override prompt intentions.226
Artifacts/Oversaturation/Overcooked Look:
- Lower cfg.226
- Check LoRA compatibility, or reduce LoRA strength(s). Multiple LoRAs can conflict.227
- Try a different sampler_name (some are more prone to artifacts at high CFG).228
- Ensure the VAE is appropriate and functioning correctly (less of an issue with baked VAEs, but worth considering if unexplained artifacts appear).229
Slow Generation:
7. Conclusion
This report has navigated the intricate landscape of KSampler and XLabs Sampler settings within ComfyUI, focusing on the RealVisXL5.0LightningBadedVae, RealVisXL5.0BakedVae, and Flux1-dev-Q5_K_S models.232
The analysis underscores that optimal image generation is not merely about finding universal "best" settings, but rather about understanding the interplay between model architecture, sampler mechanics, and the specific creative intent.233
Key learnings include the critical differences in step counts and CFG scales required by Lightning versus standard RealVisXL models, with Lightning variants offering remarkable speed at the cost of some flexibility.234
The Flux1-dev model, particularly its GGUF quantized version, presents a distinct architectural alternative with unique prompting requirements and a specialized ecosystem of ComfyUI nodes designed to harness its capabilities, especially for users with VRAM constraints.235
Furthermore, the integration of LoRAs and ControlNets necessitates thoughtful adjustments to KSampler parameters, often allowing for a reduction in CFG due to the strong additional conditioning these tools provide, while the denoise setting becomes pivotal in workflows involving existing images or multi-pass refinement.236
The paradigm shift in prompting from the keyword-centric approach of SDXL to the natural language preference of FLUX models is a crucial adaptation for users working across these architectures.237
Ultimately, while this report provides evidence-based guidelines and synthesized recommendations, the dynamic and rapidly evolving field of AI image generation thrives on experimentation.238
The ComfyUI environment, with its modular and transparent workflow, is an exceptional platform for such exploration.239
Users are encouraged to utilize the information presented here as a robust starting point, but to also engage in personal testing, iterative refinement, and community learning to fully master these powerful tools and achieve their unique artistic visions.240
The "best" settings will often be those discovered through a dedicated process of inquiry and creative application.241
References
- KSampler - RunComfy, Zugriff am Juni 1, 2025.
- Beginner's Guide to ComfyUI - Stable Diffusion Art, Zugriff am Juni 1, 2025.
- KSampler - ComfyUI Community Manual, Zugriff am Juni 1, 2025.
- KSampler - RunComfy (SDVN), Zugriff am Juni 1, 2025.
- Best way to fine tune ksampler settings for models. - Reddit, Zugriff am Juni 1, 2025.
- SG161222/RealVisXL_V5.0_Lightning - Hugging Face, Zugriff am Juni 1, 2025.
- Realistic Vision XL | Open Laboratory, Zugriff am Juni 1, 2025.
- RealVisXL v5 Lightning Model (SDXL 1.0) - NightCafe, Zugriff am Juni 1, 2025.
- Understanding Guidance Scale in Stable Diffusion: A Beginner's Guide - Segmind, Zugriff am Juni 1, 2025.
- What's Stable Diffusion CFG Scale Meaning? How to Use It? - Novita AI, Zugriff am Juni 1, 2025.
- Stable Diffusion: 17 Sampling Methods Comparison - DeviantArt, Zugriff am Juni 1, 2025.
- A Brief Introduction to Sampler - Novita AI Docs, Zugriff am Juni 1, 2025.
- What is a Sampler in Stable Diffusion? - PirateDiffusion, Zugriff am Juni 1, 2025.
- imagepipeline/RealVisXL-V4-V4-BakedVAE - Hugging Face, Zugriff am Juni 1, 2025.
- RealVisXL V5.0 - You NEED to try this!!! - YouTube, Zugriff am Juni 1, 2025.
- In-Depth Guide to Understanding and Utilizing ComfyUI KSampler - RunComfy Learn, Zugriff am Juni 1, 2025.
- Scheduler Features - GitHub (Hugging Face Diffusers), Zugriff am Juni 1, 2025.
- EulerDiscreteScheduler - Hugging Face Docs, Zugriff am Juni 1, 2025.
- How to Run Flux Dev GGUF Models in ComfyUI - Next Diffusion, Zugriff am Juni 1, 2025.
- How to use high performance GGUF with ComfyUI Flux.1 [schnell] - DCAI, Zugriff am Juni 1, 2025.
- Iterative Mixing KSampler Advanced Node Documentation - ComfyAI.run, Zugriff am Juni 1, 2025.
- Ksampler - ComfyUI Built-in Node Documentation, Zugriff am Juni 1, 2025.
- What is denoising strength? - Stable Diffusion Art, Zugriff am Juni 1, 2025.
- KSampler (Advanced) - ComfyUI Wiki, Zugriff am Juni 1, 2025.
- ComfyUI Iterative Mixing Nodes - Comfy.ICU, Zugriff am Juni 1, 2025.
- black-forest-labs/FLUX.1-dev - Hugging Face, Zugriff am Juni 1, 2025.
- FLUX.1 Text-to-Image AI: Next-Gen Diffusion Model - Ikomia, Zugriff am Juni 1, 2025.
- XlabsSampler Node Documentation - ComfyAI.run, Zugriff am Juni 1, 2025.
- Detailed usage of Flux.1 ControlNet + IP-Adapter in ComfyUI - DCAI, Zugriff am Juni 1, 2025.
- FL Ksampler Settings - RunComfy, Zugriff am Juni 1, 2025.
- ModelSamplingFlux - RunComfy, Zugriff am Juni 1, 2025.
- Detailed Explanation of ComfyUI Nodes - ComfyUI Wiki, Zugriff am Juni 1, 2025.
- XLabs-AI/x-flux-comfyui - GitHub, Zugriff am Juni 1, 2025.
- 10 Best Stable Diffusion Models for People on Civitai - Shakers AI, Zugriff am Juni 1, 2025.
- realvisxl-v4.0-lightning - AIModels.fyi, Zugriff am Juni 1, 2025.
- RealVisXL_V5.0_Lightning - PromptLayer, Zugriff am Juni 1, 2025.
- SDXL-Lightning: Fast and Efficient Image Generation Model - Shakker AI, Zugriff am Juni 1, 2025.
- Best RealVisXL V4.0 V5.0 Lightning (BakedVAE) Prompts - PromptHero, Zugriff am Juni 1, 2025.
- SDXL 1.0 Overview: A Remarkable Leap - Civitai Education, Zugriff am Juni 1, 2025.
- Sg161222 - AIModels.fyi, Zugriff am Juni 1, 2025.
- RealVisXL V5.0 - V2.0 (BakedVAE) - Civitai, Zugriff am Juni 1, 2025.
- ComfyUI マスターガイド - AICU Inc., Zugriff am Juni 1, 2025.
- RealVisXL V5.0- full32 - V5.0 full - Tensor.Art, Zugriff am Juni 1, 2025.
- RealVisXL - Comflowy, Zugriff am Juni 1, 2025.
- SG161222/RealVisXL_V5.0 - Hugging Face, Zugriff am Juni 1, 2025.
- FLUX.1 Dev · Models - Dataloop, Zugriff am Juni 1, 2025.
- Introducing FLUX.1 Kontext and the BFL Playground - Black Forest Labs, Zugriff am Juni 1, 2025.
- The Flux AI guide: installation, models, prompts and settings - Andreas Kuhr, Zugriff am Juni 1, 2025.
- Flux.1 ComfyUI Guide, workflow and example - ComfyUI Wiki, Zugriff am Juni 1, 2025.
- FLUX.1 on Fireworks: Fast, frugal, and flexible - Fireworks AI, Zugriff am Juni 1, 2025.
- FLUX.1-dev - Xinference Docs, Zugriff am Juni 1, 2025.
- FLUX.1 Dev Gguf · Models - Dataloop, Zugriff am Juni 1, 2025.
- city96/FLUX.1-dev-gguf - Hugging Face, Zugriff am Juni 1, 2025.
- flux1-dev-Q5_K_S.gguf - Hugging Face (Eviation/flux-imatrix), Zugriff am Juni 1, 2025.
- Guide to Image Generation Using FLUX.1-dev-gguf with ComfyUI - E2E Networks, Zugriff am Juni 1, 2025.
- ComfyUI Flux.1 Text-to-Image Workflow Example - ComfyUI Docs, Zugriff am Juni 1, 2025.
- Getting Started with LoRA in ComfyUI - ComfyUI Wiki, Zugriff am Juni 1, 2025.
- LoRA Training directly in ComfyUI! - Reddit, Zugriff am Juni 1, 2025.
- Need help with Lora and Comfyui - Reddit, Zugriff am Juni 1, 2025.
- Is there a way to adjust the lora % weight during generation? - Reddit, Zugriff am Juni 1, 2025.
- Wan 2.1 Guide + Best settings - Reddit, Zugriff am Juni 1, 2025.
- ComfyUI Pose ControlNet Usage Example - ComfyUI Docs, Zugriff am Juni 1, 2025.
- How to Use OpenPose ControlNet SD1.5 in ComfyUI - ComfyUI Wiki, Zugriff am Juni 1, 2025.
- How to Use Canny ControlNet SD1.5 Model in ComfyUI - ComfyUI Wiki, Zugriff am Juni 1, 2025.
- Auto set recommended steps-cfg-sampler when switching models? - Reddit, Zugriff am Juni 1, 2025.
- Detailed Guide to Flux ControlNet Workflow - ComfyUI Wiki, Zugriff am Juni 1, 2025.
- Stable Diffusion Prompt Basics - Comflowy, Zugriff am Juni 1, 2025.
- 9 Best SDXL Models for Realism on Civitai - Shakers AI, Zugriff am Juni 1, 2025.
- FLUX.1 Prompt Guide: Pro Tips and Common Mistakes - getimg.ai Blog, Zugriff am Juni 1, 2025.
- How to Design Effective Prompts for Flux.1 Dev - fluxproweb.com, Zugriff am Juni 1, 2025.
- FLUX.1 Kontext [max] - Replicate, Zugriff am Juni 1, 2025.
- FLUX.1-Redux-dev · reference image + text prompt - Hugging Face, Zugriff am Juni 1, 2025.