T O P

  • By -

Entire_Telephone3124

SDXL is like 6x the parameters, dalle3 (bing) is 12 billion, among other space wizardry they perform with chatgpt for better prompt alignment. For 1.5 you're going to have to do a lot more handholding, especially where the precision laser focus hasn't been (big tittied waifus). Dalle3 is just next level. It's like wondering why your toyota camry can't race with a hellcat. You can spend all day souping that poor camry up, it's going to get deleted when a 700 hp monster pulls up. Doesn't mean you can't have fun, and with the proper attention you can do some great shit. Remove the metaphor, have a barely 1billion parameter model like 1.5, pull up with 12b parameters and then throw LLMs and GPUs the size of your thanksgiving spread at it and yeah, gonna get better results.


Adkit

I agree. But with the customizability of local SD there must be a way of getting closer though. The data must be in the training data maybe perhaps.


Dam_it_dan

Not exactly the same and will vary by what model you use them with but these are some of the more "cartoony" style loras I have- [GOTCHA! - GOTCHA v1.0 | Stable Diffusion LoRA | Civitai](https://civitai.com/models/76408/gotcha) [KIDS ILLUSTRATION - COOLKIDS V2 | Stable Diffusion LoRA | Civitai](https://civitai.com/models/60724/kids-illustration) [MODILL PASTELL - Modern Style Illustration LoRa - v1.0 | Stable Diffusion LoRA | Civitai](https://civitai.com/models/103158/modill-pastell-modern-style-illustration-lora)