Tensor.Art

Creation

Get start with Stable Diffusion!

ComfyFlow

ComfyUI's amazing experience!

Host My Model

Share my models,get more attention!

Online Training

Make LoRA Training easier!
775496361771594264

Flux dev x Tensor

364K

Flux Realism Fixed

16K
785509906223693704

二次元现实/RealAnime

1.1K
786013774606841126

🎃 Spooky Vintage Anime #Halloween2024

1.3K
785445430174640461

二次元现实/RealAnime

183
783784295753004333

Flux Pulid Halloween Yourself

2.5K

FLUXING Halloween2024 Pulid

2.7K

Mochi 1 preview - video generation

99

Make your pictures come alive with CogVideo-5B

2.1K

Let's generate a 384p video using PyramidFlow

217
710239809792263239

Your Face on Sticker! - Easy 3D Sticker Maker + Face Swap (High Quality)

2K
757362842344161975

FLUX.1 TXT2IMG - Advanced

213K
753151588255691332

Lego Vibe

751

HHM XL 室設轉風格B (img2img)

333
718437502200861032

HHM Interior Designer (SD1.5)

520
765892930876632544

FLUX DEV: Choose A Style

6.5K
715726879621087932

Artistic re-creation

1.4K
721539864453061290

HHM Styler (XL)

325
747306936644493930

背景を消す

619
713140287106511469

Wabisabi Interior Design from Architech1904

959
759153100692018871

Flux Text to Image

1.3K
763465947936795097

EK-Art Flux + HiresFix

289
699760252813429437

Ai Logo Fast Designer

5.5K
732427555008789920

LineArt Anything: Tattoo AI Tool

894
725140838063019285

Oil painting Wallpaper (Wide) V1.0

6K
764437826006685327

FLUX x Mimc - text2img - all in one

326
708327063286970738

You will be Pixelated! [ Face Swap ]

3.4K
749867136584907746

EK Ink Art Maker [XL]

474
786152205712621113

SKELETON THEME HALLOWEEN2024

13

View All AI Tools

Models

787433866899219137
CHECKPOINT SD 3.5 L

stable-diffusion-3.5-large-fp8

33K 471
790012509429107329
CHECKPOINT SD 3.5 M

stable-diffusion-3.5-medium-fp16

2.5K 152
784775531250981589
LORA Flux
EXCLUSIVE

RealAnime-Detailed V2

6.6K 500
789941075533065427
LORA SD 3.5 L
EARLY ACCESS

SD3.5L [ TATTOO ART ]-V1

298 52
789635338290137391
LORA SD 3.5 L
EARLY ACCESS

atmosphere SD3.5-v1

1.1K 63
785793335425040640
LORA Flux
EARLY ACCESS

Haunted Doll 3D Realistic-FLUX

2.7K 177
784993423531737999
LORA Flux
EARLY ACCESS

Zombie-Zombie Realism

4.5K 295
790326974049657727
LORA Flux
EARLY ACCESS

Dark Halloween 2024 - ABM-Retrain 2024-10-30

2.9K 161
789742390349947131
LORA Flux
EARLY ACCESS

comic style Halloween-v1

159 24
788187844112939304
LORA Flux
EXCLUSIVE

Flux | Halloween Makeup -v3.0

1.7K 124
785220189215161844
LORA Flux
EARLY ACCESS

Spooky Art Drawing-Halloween2024

2.4K 142
770419478510045197
LORA Flux

DonM - Creepy Dolls [Flux]-Flux

1.7K 85
785771220634219594
LORA Flux
EARLY ACCESS

[FLUX] #HALLOWEEN skeleton guitarist-2024-10-18 02:10:25

87 34
785697982852066072
LORA Flux

🎃 Haunted Concept 🎃 [Flux] [Halloween]-v1.0

923 55
785928523815652266
LORA Flux
EARLY ACCESS

Thriller ART-V1

4K 122
787509054596571003
LORA Flux

Halloween atmosphere-V0

494 35
751541533876142774
LORA XL

3D Framed Wall Art - By DICE-v1

3.7K 126
666648211034931058
CHECKPOINT XL

AnimeGodXL-WonGhibli

148K 817
786901935203650418
LORA Flux
EARLY ACCESS

atmosphere FLUX-v1

1.3K 47
756585543472697641
LORA SD3
EARLY ACCESS

Cyborg Style SD3m-v1

600 35
789615740354554975
LORA SD 3.5 L

Venom-SD 3.5

100 3
766495428888783862
LORA HunyuanDiT
EARLY ACCESS

Castle 3d-2024-08-26 23:45:48

47 7
756345519225177433
LORA HunyuanDiT
EARLY ACCESS

Realism Dit-TA-ExclV1

384 5
769516526060624288
LORA Flux

Glimmerkin Style [Flux Cute LoRA]-v1.0

1.3K 55
783350203404166151
LORA Flux

👻 Spooky Halloween Booster (🎃Halloween Contest Helper) [FLUX]-FLUX

468 63
656995224788200974
LOCON
EXCLUSIVE

Neg4All!Both Positive High quality、Details and Negative worse quality、bad hand in one LoRA!-v3.5

126K 320
773163610424918769
LORA Flux

Lora-Dalcefo_Flux1.Dev-Cilo-v2

2.5K 119
758089984616995278
LORA HunyuanDiT
EARLY ACCESS

[HunYuan] YinYang Taichi 陰陽太極 - FuturEvoLab-v1

138 7
756714611530611669
LORA SD3
EXCLUSIVE

Fantasy Vision-V1

813 20
672052997875952118
LORA XL

LegendDarkFantasyXL-v.1

143K 420
752251431844724223
LORA SD3

Elegant Brushstrokes SD3 Version-EXPERIMENT

459 13
788004238555979527
LORA SD 3.5 L

iPhone Photo Realism [FLUX • SD3.5]-v2 SD3.5 Large TEST

9.3K 87
787192549866544720
LORA Flux
EARLY ACCESS

FLUX [ Skeleton ]-Partner with It

115 18
761995415554913073
LORA HunyuanDiT
EXCLUSIVE

Hunyuan Portraits-2024-08-14 14:28:18

235 8
765779775663946720
LORA HunyuanDiT
EARLY ACCESS

ChugAx_WuKong_DIT-Wukong Niji

324 13
640579022264433824
CHECKPOINT XL
EXCLUSIVE

Stoked Tensor XL-v1.0

798K 1.3K

Workflows

Articles

How to transform your images into a Halloween party atmosphere. | 🎃 Halloween 2024

How to transform your images into a Halloween party atmosphere. | 🎃 Halloween 2024

INSTRUCTIONS:This is a very simple workflow, just upload your image and press RUN.PROMPT basically does not need to be modified, but you can still add more Halloween elements to make the theme richer.Hope you all have a good time.PROMPT:(masterpiece), ((halloween elements)),a person, halloween striped thighhighs, witch hat, grin, (ghost), sweets, candy, candy cane, cookie, string of flags, halloween costume, jack-o'-lantern bucket, halloween, pumpkins,black cat,halloween,little ghost,magic robe,autumn leaves,candle,skull, 3d cg.Negative PROMPT:None.Below is the workflow link:https://tensor.art/workflows/786144487641608308Below is the AI-tool link:https://tensor.art/template/786150277257599620model used:CKPThttps://tensor.art/models/757279507095956705/FLUX.1-dev-fp8
44
19
halloween2024 Alien Pumpkin

halloween2024 Alien Pumpkin

PROMPT:photography of Alien incubator, seen from above,lots Halloween pumpkin With facial features, A close-up shot of a Halloween pumpkin With facial features, (An top open Halloween pumpkin,There's a real brain inside, bloody, a chestburster Out of the pumpkin, facehugger jump Out of the pumpkin),The word "halloween" scrawled in blood on floor,Cold interior lighting,direct sunlight on Halloween pumpkin,dramatic light and shadow, Horrifying movie scenes,Negative PROMPT:low quality, low resolution, unrealistic, semi-realistic, animation, drawing, 2D, painting, lack detail, flat background, bad lighting, bad composition,Below is the workflow link:https://tensor.art/workflow/editor/786104965352589871Below is the AI-tool link:https://tensor.art/template/786112649048914425model used:CKPThttps://tensor.art/models/757279507095956705/FLUX.1-dev-fp8LORAhttps://tensor.art/models/768181864964856710/FLUX-Cinematic-V1https://tensor.art/models/782762523020600332https://tensor.art/models/783607553541144419
43
28
50 Inspiration Beauty Monster or Creature  - HALLOWEEN2024

50 Inspiration Beauty Monster or Creature - HALLOWEEN2024

Looking to stand out this Halloween with a fierce, captivating costume? Dive into our 50 Beauty Monster and Creature Inspirations for Halloween 2024!From the alluring vampire queen with fangs and pale skin, to the mystical forest spirit with branches for hair, this list features a variety of iconic, feminine creatures to embody. Each entry provides five key characteristics to make your costume pop with creativity. Whether you want elegance, spookiness, or a combination of both, these ideas will help you slay this Halloween!Vampire: Fangs, cloak, pale skin, red lips, pointed ears.Witch: Pointed hat, broomstick, black dress, potion bottles, striped stockings.Medusa: Snake hair, stony gaze, green skin, gold jewelry, ancient toga.Banshee: Ghostly white dress, flowing hair, haunting scream, pale makeup, chains.Succubus: Bat wings, red dress, horns, glowing eyes, tail.Werewolf: Furry ears, sharp claws, fangs, torn clothes, wild hair.Mermaid: Scales, seashell bra, fishtail, wet-look hair, pearls.Harpy: Feathered wings, talons, bird-like eyes, fierce expression, ragged clothes.Fairy: Sparkling wings, flower crown, wand, glittery makeup, light dress.Zombie: Torn clothes, blood stains, decayed skin, lifeless eyes, open wounds.Siren: Wet-look hair, seashell jewelry, seaweed skirt, alluring voice, eerie glow.Elf: Pointed ears, elegant gown, bow and arrow, long hair, ethereal glow.Gorgon: Snake tail, golden scales, slit eyes, regal crown, sharp claws.Mummy: Wrapped in bandages, dark eye makeup, jewelry, ancient amulet, dusty appearance.Ghost: Flowing white sheet, transparent, eerie wail, glowing eyes, pale hands.Queen of the Dead: Black gown, skull crown, skeletal makeup, dark veil, red roses.Demoness: Red skin, black horns, tail, wings, sharp claws.Bride of Frankenstein: Black and white hair, stitched skin, bride gown, lightning bolts, scars.Voodoo Priestess: Skull face paint, voodoo doll, bones, beads, tribal clothing.Phoenix: Fiery wings, flame patterns, red and orange outfit, glowing skin, feathers.Chimera: Lion mane, snake tail, dragon wings, golden eyes, muscular build.Spider Queen: Black web dress, spider crown, long legs, red eyes, venomous fangs.Lady Death: Black cloak, scythe, skeletal hands, skull mask, dark aura.Nymph: Nature gown, flowers in hair, earthy tones, glowing skin, delicate wings.Selkie: Fur cloak, watery skin, ocean jewels, seal tail, wet hairGiantess: Massive build, oversized clothes, earthy skin, towering presence, big jewelry.Forest Witch: Mossy cloak, animal bones, green skin, potions, tree branches in hair.Dragoness: Scaly skin, horns, tail, fiery breath, armored chestplate.Lilith: Dark wings, black robe, seductive look, glowing red eyes, ancient symbols.Hag: Wrinkled skin, tattered clothes, long nose, hunched posture, warts.Valkyrie: Winged helmet, sword, battle armor, braided hair, shield.Troll Woman: Green skin, sharp tusks, club, fur clothes, wild hair.Ice Queen: Frosted crown, shimmering cape, blue skin, ice staff, glowing cold eyes.Scarecrow: Straw-filled body, stitched mouth, tattered hat, pumpkin head, patched overalls.Djinn: Flowing robes, magic lamp, glowing eyes, ornate jewelry, smoke swirling around.Cheshire Cat: Striped fur, wide grin, cat ears, mischievous eyes, tail.Swamp Creature: Muddy skin, algae hair, webbed fingers, water plants, gills.Basilisk Queen: Reptilian skin, glowing eyes, snake tail, venomous fangs, ancient armor.Lamia: Snake body, golden armor, hypnotic eyes, deadly claws, venomous bite.Wendigo Woman: Deer antlers, skeletal body, glowing eyes, fur cloak, sharp claws.Shadow Witch: Black shadowy figure, dark veil, glowing red eyes, spectral form, floating.Frost Maiden: Icicle crown, snowflake gown, pale blue skin, icy breath, shimmering frost.Baba Yaga: Hunched back, long nose, flying broom, warts, iron teeth.Kitsune: Fox ears, fluffy tail, red kimono, mystical powers, mask.Forest Spirit: Tree branches for hair, bark-like skin, moss gown, glowing eyes, ethereal glow.Plague Doctoress: Black cloak, plague mask, long gloves, eerie eyes, dark potions.Dullahan: Headless woman, flowing black cloak, horse-riding, holding a skull, eerie lantern.Succubus Queen: Leather bodice, wings, horns, glowing eyes, seductive aura.Dryad: Bark skin, leaves in hair, tree branches for arms, glowing green eyes, earthy gown.Banshee Queen: Flowing black dress, ghostly hair, skeletal hands, pale skin, sorrowful wail.settings usedAll created using Juggernaut SDXL modelsteps 25cfg 6dpmpp_2m karrasnot all creature recognize well by the checkpoint, you may use LoRA or other checkpoint if needed to create certain characterWith these 50 beauty monster and creature inspirations, you're all set to embrace the eerie, enchanting side of Halloween 2024. Whether you choose to transform into a seductive vampire, a magical forest spirit, or a chilling banshee queen, each idea is designed to make you stand out in both style and spookiness. Let your creativity soar this Halloween, and enjoy bringing these unique creatures to life. Get ready to slay (literally!) with hauntingly beautiful looks that will leave everyone spellbound!
53
8
Algunos cambios / some changes

Algunos cambios / some changes

He actualizado todos mis modelos para que la gente pueda generar imágenes de manera ilimitada y gratuita con ellos, la descarga sigue sujeta al pago del bufet, asi que adelante, den rienda suelta a su creatividad.//I've updated all my models so that people can generate unlimited images with them for free, downloading them is still subject to paying the buffet, so go ahead and unleash your creativity.
29
5
🎃 Halloween2024 | Optimizing Sampling Schedules in Diffusion Models

🎃 Halloween2024 | Optimizing Sampling Schedules in Diffusion Models

You migh have seen this kind of images in the past if you've girly tastes when navigate on pinterest, well guess what? I'll teach you about some parammeters to enhance your Pony SDXL future generations. It's been a while since my last post, today I'll teach you about a cool feature launched by NVIDIA on July 22, 2024. For this task I'll provide an alternative workflow (Diffusion Workflow) for SDXL. Now lets go with the content.ModelsFor my research (AI Tool) I decided to use the next models:Checklpoint model: https://tensor.art/models/757869889005411012/Anime-Confetti-Comrade-Mix-v30.60 LoRA: https://tensor.art/models/7025156632998356040.80 LoRA: https://tensor.art/models/757240925404735859/Sailor-Moon-Vixon's-Anime-Style-Freckledvixon-1.00.75 LoRA: https://tensor.art/models/685518158427095353NodesThe Diffusion Workflow has many nodes I've merged in single nodes I'll explain them below, remember you can group nodes and edit their values to enhance your experience.👑 Super Prompt Styler // Advanced Manager (CLIP G) text_positive_g: positive prompt, subject of the scene (all the elements the scene is meant for, LoRA Keyword activators).(CLIP L) text_positive_l: positive prompt, all the scene itself is meant (composition, lighting, style, scores, ratings).text:negative: negative prompt.◀Style▶: artistic styler, select the direction for your prompt, select 'misc Gothic' for halloween direction.◀Negative Prompt▶: prepares the negative prompt splitting it in two (CLIP G and CLIP L) for the encoder.◀Log Prompt▶: add information to metadata, produces error 1406 when enabled, so turn it off.◀Resolution▶: select the resolution of your generation.👑 Super KSampler // NVIDIA Aligned Stepsbase_seed: similar to esnd (know more here).similarity: this parameter influences base_seed noise to be similar to noise_seed value.noise_seed: the exact same noise seed you know.control after generate: dictates the behavior of noise_seed.cfg: guidance for the prompt, read about <DynamicThresholdingFull> to know the correct value. I recomend 12sampler_name: sampling method.model_type: NVIDIA sampler for SDXL and SD models.steps: the exact same steps you know, dictates how much the sampling denoises the noise injected.denoise: the exact same denoise you know, dictates the strong the sampling denoises the noise injected.latent_offset: select between {-1.00 Darker to 1.00 Brighter} to modify the input latent, any value different than 0 adds information to enhance final result.factor_positive: upscale factor for the conditioning.factor_negative: upscale factor for the conditioning.vae_name: the exact same vae you know, dictates how the noise injected is denoised by the sampler.👑 Super Iterative Upscale // Latent/on Pixel Spacemodel_type: NVIDIA sampler for SDXL and SD models.steps: number of steps the UPSCALER (Pixel KSampler) will use to correct the latent on pixel space while upscaling it.denoise: dictates the strenght of the correction on the latent on pixel space.cfg: guidance for the prompt, read about <DynamicThresholdingFull> to know the correct value. I recomend 12upscale_factor: number of times the upscaler will upscale the latent (must match factor_positive and factor_positive) upscale_steps: dictates the number of steps the UPSCALER (Pixel KSampler) will use to upscale the latent.MiscellaneousDynamicThresholdingFullmimic_scale: 4.5 (Important value. go to learn more)threshold_percentile: 0.98mimic_mode: half cosine downmimic_scale_min: 3.00cfg_mode: half cosine downcfg_scale_min: 0.00sched_val: 3.00separate_feature_channels: enablescaling_starpoint: meanvariability_measure: ADinterpolate_phi: 0.85Learn more: https://www.youtube.com/watch?v=_l0WHqKEKk8Latent OffsetLearn more: https://github.com/spacepxl/ComfyUI-Image-Filters?tab=readme-ov-file#offset-latent-imageAlign Your StepsLearn more: https://research.nvidia.com/labs/toronto-ai/AlignYourSteps/LayerColor: Levelsset black_point = 0 (base level of black)set white_point = 255 (base level of white)Set output_black_point = 20 (makes blacks less blacks)Set output_white_point = 220 (makes whites less whites)Learn more: https://docs.getsalt.ai/md/ComfyUI_LayerStyle/Nodes/LayerColor%3A%20Levels/LayerFilter:Filmcenter_x: 0.50center_y: 0.50saturation: 1.75vignete_intensity: 0.20grain_power: 0.50grain_scale: 1.00grain_sat: 0.00grain_shadows: 0.05grain_highs: 0.00blur_strenght: 0.00blur_focus_spread: 0.1 focal_depth: 1.00Learn more: https://docs.getsalt.ai/md/ComfyUI_LayerStyle/Nodes/LayerFilter%3A%20Film/?h=filmResultAi Tool: https://tensor.art/template/785834262153721417DownloadsPony Diffusion Workflow: https://tensor.art/workflows/785821634949973948
11
6
The Trials and Tribulations of a Halloween2024 Face Swap through Facepaint work in FLUX1D

The Trials and Tribulations of a Halloween2024 Face Swap through Facepaint work in FLUX1D

So I set out with what I thought was a simple idea:“Start with an image of someone’s face and turn that into a spooky Halloween character, with costume, makeup and full Facepaint with a spooky background.”BUT it had to look enough like them at the end - that they would be pleased with the result…The starting point was easy - I wanted to train a Halloween LoRA on lots of images of people wearing Halloween Facepaint - so I did that…A couple of the 48 images i used to train with:So I had a Flux LoRA - now I tested that in Tensor.Art with simple “Man in Halloween Facepaint”, “Woman in Halloween Facepaint”So far so good, I thought ok, this is going to be easy peasy!At this point (End of September 2024) there were limited options in TA for Flux Face swap… (No Pulid available then) so I started trying with Facedetailer…I built out the workflow - made a separate flow for the background - and was all excited…But no matter what i tried (and I tried a lot!) the facedetailer would wipe out the Facepaint from the Lora - restoring the face back to the original person, nice and clean, or with a half hearted smear of greasepaint.Or it would look nothing at all like the person and the makeup would look like it was a badly stuck on mask…So i went back to my Discord buddies and we talked about the options - and decided to try Reactor nodes with insightface…It would generate a Florence description of the original reference face (cropped) - build a dummy Halloween Image with a lookielikie from the description and with Facepaint - and then reactor the ref face back over the top (or so i thought)But the Reactor’d one cleaned up the face and removed 90% of the makeup and it didn’t want to do the costume or background at all the way I had envisaged… as soon as I gave it enough freedom to be creative, the reference person was lost completely…I think by now people in all my discord groups were sick of me asking for ideas on how to do this - I tried every setting and balance on reactor nodes.Could I use an llm to rewrite the visual description of the face to include the Halloween description first, and so on.I looked at IPAdapter and using Depth maps - but although they captured the shape of the face - they couldn’t preserve the familiar features through costume stylemakeup.At this point - I pretty much gave up in disgust… I put out a final round of help requests on various discord’s and went onto another projectA few days later my good friend told me “ hey - finally they released Pulid for Flux on TA!” - and I already had built Flux Pulid workflows for face swapping the previous week on my MimicPC Cloud version of Comfyui (where you can load any kind of node and model you want and really design and play with freedom) so I started to regain my enthusiasm…I managed to merge some of the earlier ideas for generating the Halloween style with LLM’s and a Joycaption of the cropped reference face - and the Flux Pulid face swaps - and experimented with the positioning of the LoRA to get maximum effect - and was finally able to release a workflow and AI Tool that did what i had seen in my head those few weeks back when I started… https://tensor.art/template/785795972520313546And the workflow - https://tensor.art/workflows/785793305345589081And the LoRA - https://tensor.art/models/785804669831296337If you have enjoyed my article - please like and use my AI Tools and Models…I welcome comments and constructive feedback.
16
8
🎃 Halloween2024 Generation Guide: Elevate Your Spooky Creations! 👻

🎃 Halloween2024 Generation Guide: Elevate Your Spooky Creations! 👻

Halloween is right around the corner, and it’s time to infuse your generation models with a touch of spooky magic! Whether you’re crafting images, stories, or even interactive AI experiences, this guide will help you conjure up the best Halloween-themed content for 2024. Let’s dive into some tips and tricks to make your generative AI creations truly spine-chilling! 🧛‍♂️🕸️1. Theme Selection: Classic Horror vs. Modern ThrillsStart with deciding the tone of your Halloween project. Are you going for classic horror, with haunted houses, creepy forests, and gothic vibes? Or are you leaning towards modern Halloween with neon lights, cyberpunk ghosts, or playful skeletons?Classic horror themes like vampires, witches, and ghosts never go out of style, but blending them with modern elements (think AI-enhanced haunted tech or neon-lit crypts) can bring a fresh twist to your content.2. Prompts and Inspiration IdeasFor image generation, try prompts that capture the Halloween atmosphere:"A haunted Victorian mansion under a full moon, surrounded by fog and dark twisted trees""A neon-lit skeleton playing an electric guitar on a cyberpunk street""A witch stirring a glowing cauldron, with enchanted bats swirling around"For story generation, build a suspenseful atmosphere with prompts like:"On Halloween night, a group of friends discovers a hidden portal in an abandoned amusement park...""A town where every carved pumpkin holds the soul of a spirit seeking freedom"Don't be afraid to add a bit of humor to your Halloween stories, like:"A vampire who’s afraid of the dark trying to overcome his fear"3. Style Adjustments: The Magic of LightingLighting can make or break the eerie ambiance of your Halloween images. Play with shadows, moonlit scenes, or dimly lit rooms to add that sense of unease.Experiment with different color palettes—orange, black, and purple are classics, but consider adding splashes of neon green or eerie blue for a modern twist.For a vintage horror feel, use grainy textures, sepia tones, or black-and-white effects to mimic old horror films.4. Interactive Elements: Make it a Thrilling ExperienceFor those building interactive experiences, consider adding branching storylines where users can explore haunted locations or solve spooky mysteries.Add random elements to make the experience unpredictable—imagine a haunted AI guide that offers different creepy clues each time users interact with it.Build suspense with sound effects like whispering winds, distant footsteps, or creaking doors that play as users engage with your content.5. Community Collaboration: Share and Get Inspired!The best part about generative projects is sharing them with the community! Post your Halloween creations, get feedback, and see how others are getting into the spirit.Participate in Halloween-themed challenges or host one yourself—like a Spookiest Story Contest or Best Halloween Image Generation.Don’t forget to use the hashtag #Halloween2024 when sharing your spooky content so others can easily find and engage with your posts.6. Ethical Considerations: Keep It Fun and RespectfulWhile Halloween is all about embracing the creepy and the supernatural, it's important to remain sensitive to cultural traditions and symbols. Respectful representation goes a long way in keeping the spirit of fun alive for everyone.Ensure that your generative content is age-appropriate if targeting younger audiences—creepy doesn’t always have to mean terrifying!Happy Halloween & Happy Generating! 🎃👻We hope these tips help you create some truly terrifying (or delightfully spooky) Halloween content this year. Let your creativity run wild and embrace the eerie, the whimsical, and the downright strange. Looking forward to seeing what you conjure up this Halloween season!
4
2
HORROR ARTIST AND ART STYLE (Special article for HALLOWEEN2024)

HORROR ARTIST AND ART STYLE (Special article for HALLOWEEN2024)

1. H.R. Giger (Biomechanical Horror) Giger is famous for his nightmarish "biomechanical" art style, blending human forms with machinery and grotesque alien creatures. His designs inspired the terrifying creatures in the Alien film series, making his style a staple in sci-fi horror.2. Junji Ito (Manga Horror) Junji Ito is a Japanese manga artist known for his unsettling and disturbing imagery. His style combines detailed linework with surreal body horror, where human forms often twist, decay, or transform into unimaginable horrors.3. Francis Bacon (Abstract Horror) Bacon’s style is known for its raw and chaotic energy, often depicting distorted, screaming faces and bodies. His abstract approach creates a sense of psychological horror, focusing on human suffering and existential dread.4. Zdzisław Beksiński (Surreal Horror) Beksiński's paintings are filled with surreal, dystopian landscapes and nightmarish creatures. His style is dreamlike, featuring decaying cities, skeletal figures, and eerie, otherworldly atmospheres that evoke a sense of dread and desolation.5. Edward Gorey (Gothic Macabre) Gorey's distinctive pen-and-ink illustrations have a whimsical yet dark, gothic tone. His art features victorian-style settings, eerie characters, and morbid humor, often telling unsettling stories in a playful, minimalist way.6. Clive Barker (Fantasy Horror) Known for creating Hellraiser's Cenobites, Barker's art mixes body horror with fantasy. His style incorporates grotesque, skin-crawling depictions of demons and twisted creatures, blurring the line between pleasure and pain.7. Wayne Barlowe (Dark Fantasy) Barlowe's art focuses on the grotesque, otherworldly creatures of hellish dimensions. His works are often visually complex, mixing detailed anatomy with imaginative designs that are both disturbing and awe-inspiring.8. Dave McKean (Mixed Media Horror) McKean's style is a unique blend of photography, collage, and painting, creating eerie, surreal images that evoke fear through abstraction and texture. His works often appear in horror comics and graphic novels, including collaborations with Neil Gaiman.Each of these artists brings a distinct approach to the horror genre, using their unique styles to evoke fear, unease, or existential dread.
12
6
This is the list of helpful articles (on going)

This is the list of helpful articles (on going)

1. Resource List: Characters in FLUXBy kftiger at CivitaiThis is a dynamic list of characters that already exist in the FLUX base dataset that do not require any additional LORAs to produce, only needing to be named in your prompts.Example from Real People:Albert Einstein, Angela Lansbury (Young), Angelina Jolie, Arnold Schwarzenegger, Barack Obama, Bill Clinton, Bruce Lee, Bruiser Brody, Colin Baker, Daniel Craig, DJ Marshmello, Donald Trump, Dwayne Johnson, Emma Stone, Fairuza Balk, George W Bush, Hulk Hogan, Jackie Chan (50/50), Joe Biden, John Pertwee, Johnny Depp, Lady Gaga, Mark Zuckerberg, Meryl Streep, Michael Jackson, Queen Elizabeth, Stan Hansen (50/50), Stan Lee, Will Smith, Wolfgang Amadeus Mozart.found by myselfNaruto UzumakiSasuke UchihaKakashi HatakePower RangerKim Jong-un2. 480+ Pony Diffusion XL Hats, Masks and More Props List - Booru Tags (SFW)by Viola_Aiarty at CivitaiThis post belongs to my series of Pony Diffusion XL Recognized Tokens lists. You can use the quick navigations below to access previous posts.for more detail go to the post or this one Access all the original image grids and wildcards here: https://mega.nz/folder/XcVVXbRL#uQPsHUfLeBnaNhSTmDgziQ3. How to train a logo lora for Fluxby JaneB at Civitaifrom this article, i made test on tensor.art lora training. you can play it herehttps://tensor.art/models/780240303475923311/Persib-logo-E9https://tensor.art/models/780138126203986771/Gojek-Logo-E44. Flux model training from just 1 image [Attention Masking]by mnemic at Civitaithis is article guide to train lora with single image and use masking, but no option for masking at tensor.art training, so if you train with transparant PNG background, it will failed.mnemic also have good article about captioning etc, so this is the bunch of his articlehttps://civitai.com/articles/7146/flux-style-captioning-differences-pt2-4-new-caption-tools-training-diaryhttps://civitai.com/articles/7044/font-model-training-with-flux-training-diary (for more detail about font, here https://civitai.com/user/juaner0211589/)https://civitai.com/articles/6868/flux-character-caption-differences-training-diaryhttps://civitai.com/articles/6792/flux-style-captioning-differences-training-diaryhttps://civitai.com/articles/2138/lora-datasets-training-data-list-civitai-dataset-guide5. Dont training lora with flux 1 dev.this is article is interesting, to train lora just used finetuned model instead the original from black forest lab.https://medium.com/@zhiwangshi28/why-flux-lora-so-hard-to-train-and-how-to-overcome-it-a0c70bc59eafi will update post soon. thanks
9
2
Tensor.Art Becomes World's Largest VisionAI Resource Hosting Platform in Under a Year

Tensor.Art Becomes World's Largest VisionAI Resource Hosting Platform in Under a Year

source: Yahoo FinanceTensor.Art Becomes World's Largest VisionAI Resource Hosting Platform in Under a Year, Empowering Enterprise AIFounded in July 2023, Tensor.Art has seen its global traffic surpass 15 million in less than a year.Currently hosting over 330,000 resources and generating more than 2 million images daily, it has positioned itself as a leading generative AI service platform worldwide. Remarkably, Tensor.Art has already started to turn a profit.As a pioneering explorer of a sustainable Gen AI ecosystem, Tensor.Art provides cloud computing power for model creators and users while offering AI solutions tailored to real-world applications across various industries.Founder Shen, possessing a keen sense of computer and AI technology, quickly decided to enter the AIGC (AI-generated content) market during its early rise. This swift decision led to the establishment of a platform that offers robust support.Tensor.Art is the world's first model platform that supports online inference and online operation of full-scale models. It consistently maintains a keen insight into the latest AI technologies and promptly embraces various cutting-edge advancements, such as the globally popular Stable Diffusion3, HunYuan DiT, Kolors, Flux, and more!As one of the first to deploy StableDiffusion on the cloud, Tensor.Art maintains a keen insight into new AI technologies and rapidly integrates the latest advancements. This includes globally impactful technologies like Stable Diffusion 3, HunYuan DiT, Kolors, Flux, and more.404’s report on Tensor.Art as the world’s first company to hold AI events in 2023Operations head Sawoo states, “We are committed to providing the best platform and community for AI enthusiasts and model creators.As early as 2023, we were pioneers in the AIGC platform space, hosting diverse events and launching creator incentive programs, which have since been emulated by competitors like CivitAI.Moreover, we tirelessly promote new global technologies, ensuring rapid online integration and training capabilities.With a comprehensive community ecosystem and rich activities, Tensor.Art now leads the global growth in new foundational models, growing at 5-6 times the rate of other leading competitors, earning praise from AI enthusiasts and model creators worldwide.”A successful collaboration between Tensor.Art and SnapChatIn an effort to democratize AI and make generative services more accessible, Tensor.Art has explored numerous real-world applications.For instance, in February 2024, the platform used its AI generative capabilities in collaboration with SnapChat to create a new paradigm in creativity through AR.Subsequent partnerships include renowned tattoo artists from Austria, a famous sticker website in the UK, and an architectural firm in Turkey, offering AI-generated design inspiration.API Service:https://tams.tensor.art/Additionally, Tensor.Art is committed to serving the B2B sector by providing a GPU API platform and simplified AI tool workflows, significantly lowering the AI adoption barrier for enterprises and catering to customized needs. This makes AI services more accessible and efficient, enhancing corporate productivity and creative inspiration.Looking ahead, Tensor.Art will maintain its competitive edge by continuing to explore and quickly integrate new global technologies while also launching its own large models. This vision aims to offer an even better community experience and technical capabilities for AI enthusiasts and model creators.
8
Flux Ultimates Custom Txt 2 Vid Tensor Workfkow

Flux Ultimates Custom Txt 2 Vid Tensor Workfkow

Welcome to Dream Diffusion FLUX ULTIMATE, TXT 2 VID With its own custom workflow made for Tensor Arts Comfy Workspace. The workflow can be downloaded on this page....... ENJOYThis is a 2nd stage Trained checkpoint to its predecessor FLUX HYPER.When you think you had it nailed in the last version and notice a 10% margin that could still be trained........ Well that's what happened ..So now this version has even more font styles, Better adherence, Sharper image clarity and a better grasp for anime, water painting and such on....This model has the same setting parameters as Flux HyperPrompt Example : Logo in neon lights, 3D, colorful, modern, glossy, neon background,with a huge explosion of fire with epic effects, the text reads  "FLUX ULTIMATE , GAME CHANGER ",Set steps at : 20Sampler : DPM++ 2M or EULER Gives best resultsScheduler : SimpleDenoise : 1.00Image Size : 576 x 1024 or 1024 x 576 You can choose any size but this model is optimized for faster rendering with those sizes.Download the links from below and save them to your comfy folders...Comfy Workflow :  https://openart.ai/workflows/maitruclam/comfyui-workflow-for-flux-simple/iuRdGnfzmTbOOzONIiVVVae download this to your Vae folder inside of your model folderDownload them from: https://huggingface.co/black-forest-labs/FLUX.1-schnell/tree/main/vaeClip:  download clip_l.safetensors and t5xxl_fp8_e4m3fn.safetensors download these 2 and save them to your clip folder inside of your models folderDownload them from : https://huggingface.co/comfyanonymous/flux_text_encoders/tree/mainIf you have any questions or issues feel free to drop a comment below and I will get back to you as soon as I can. Enjoy  DICE
75
41
An examination on the effect of Denoise on Flux Img2Img with LoRA, a journey from Boat to Campervan

An examination on the effect of Denoise on Flux Img2Img with LoRA, a journey from Boat to Campervan

I made an AI Tool yesterday ( FLUX IMG2IMG + LORAS + UPSCALE + CHOICE | ComfyUI Workflow | Tensor.Art ) that allows you to combine up to 3 LoRA's and upscale - it has model switching to let you choose whether to turn on 0/1/2/3 if the available LoRA inputs - you can choose the weighting 1 by 1 and swap out the base Flux model and all the LoRA's to your own preferences. I have implemented Radio Button prompting so that the main Trigger words for the LoRA's I use most often are already behind the buttons - and you can use "Custom" to add your own prompt or triggers into the mix.For this test I used a 6k Adobe Stock licensed image of a boat on the beach, with the Model Switcher set to "2" to prevent any bleed from other LoRA's in the tool, everything is upscaled by 4x-Ultrasharp at a factor of 2 (the tool will size your longest edge to 1024 as it processes so you will end up with a 2048 pixel final image ready for facebook servers):Original Input Image: (downsized for article)So the first test was simply putting it through the AI Tool on base Flux model - no denoise - no LoRA at all:Now I have added in the LoRA "TQ - Flux Frozen" by @TracQuoc at .9 Weight, and added .25 Denoise:Next I changed the Denoise to 0.5, you can see subtle changes, a signature has appeared, the boat is starting to change in areas and writing appearing on the side of the boat:At 0.6 Denoise the boat is starting to adapt more and the beach is changing a lot:By 0.65 you can really see dramatic changes as the boat starts to develop wheels, its almost as if the AI has a plan for this one...At 0.7 - the second boat has disappeared all together, the whole boat is on a trailer, the beach is changing into grassland:Now I am stepping to 0.01 increments as all the drama happens between .7 and .8 normally with FluxSo 0.71:0.72: (the boat is definitely changing its shape now, and you can start to see snow)0.73 you can see its becoming a land based vehicle now:0.74 it feels like a towing caravan/trailer:0.75 more detail in the towing section0.76 - everything changes and suddenly we have some kind of Safari Land Cruiser0.77 now its a camper van with a pop up roof:0.78 just some more camper style detailing but nothing dramatic:0.79 There's almost no resemblance to the original scene except sky and colours:0.8 I can't see much change here:Now I will go up in increments of 0.05 again0.85 the Frozen world has taken over, although it still has the style and colour feel of the original to some extent0.9 it's all gone (it ignored inputs over .9 and changed them back to .9)I hope you have found this a useful experiment - and will save you time and coins in playing with img2img and denoise.You can check out all my AI Tools and LoRA's on my profile here: KurtC PhotoEdLet me know if you enjoyed this and I might make some more (this was my first one).
59
22
Radio Buttons are awesome in AI Tools: [How to set-up guide]

Radio Buttons are awesome in AI Tools: [How to set-up guide]

Dear Tensorians,Thanks to the implemented feature of radio buttons for AI Tools, we can use the AI tools with much more fun now. Because I'm the one who insisted to implement it and more importantly the radio button's setting import/export features, I'll give an easy tutorial about them for the beginners. 🤗😉https://tensor.art/template/765762196352358016This is an example AI tool using radio buttons. You can see the cool radio buttons on the right. Yes! The cool thing about radio button GUI is that you don't have to remember or re-type all those crazy prompt words at all any more. You can store them in those buttons and click them! Especially if you have a very wide range of different prompting styles as most users are, you cannot even remember them all. I bet you already have your own backup memo file for those special prompts lol. Yes, we have to do it for important prompts. However, more conveniently, if you make this kind of AI tool with radio button UI, you can just store them online next to you all the time. You can click on the buttons and generate various images whenever you want, even when you are driving (just kidding, don't ever do that lol). Of course you can add extra prompt together with the buttons. (Click "custom" button and you can always input more prompt!)To create the radio buttons, click edit in the ... menu.Then you move to the EDIT page of the AI tool.in the middle of the page, you see the user-configurable settings.By clicking "Add" button, you can choose your AI interface. By clicking "edit" in the prompt's text box, you can enter the radio buttons option page.From the scratch, you can choose the pre-defined groups and buttons. In addition, you can add your own new buttons! Make a button name and its content. The content is part of prompts you want to add for the button's place.After you are done with all the button settings, click "confirm" and then "publish" your AI tool. Then you'll see your cool radio buttons in the AI tool. (Note that there are certain prompt text box nodes in comfyUI unable to edit for buttons. Basic text prompt nodes and more nodes can be used for button edit. You can check it after you publish your workflow into a tool. If it doesn't support the radio buttons, use different prompt text nodes.)Whenever you update your workflow for the AI tool, all the AI tool UI is reset to none!! Yes. It was a real headache at the beginning. However, now we have a cool import/export button for the radio buttons! (Thanks God~ 👯‍♀️⛈💯🤗). BTW, when you edit the button groups, you might choose part of the 6 or 7 groups (e.g, "character settings" and "role" groups) first and add some nice buttons, then later you change your mind and want to add another group, e.g., "style" group, however, if you press the add button for that, your previous button data will be gone!! You restart from the beginning. Be very careful! (You'll understand what I mean when it happens. lol)Before updating your workflow, you can export the radio button settings as a JSON file. Then you can import it back in later anytime you want. More importantly, you can just edit the radio buttons from the editors (like MS visual studio) for easier copy and paste from the existing files. Trust me. This will save your enormous amount of time remaking those terrible buttons all the time whenever the workflow is modified.Sometimes you must want to edit an existing button JSON file for another AI tool. Editing a JSON file is not really an entertaining work. However, it's much better than remaking the whole radio buttons at GUI~ So find the place to edit in the JSON file and change it very carefully. The JSON syntax is not very editor-friendly and error-prone. But you'll get used to it soon by trial and errors. It's always useful to use "find" command to look for the button you want in the file. You'll realize more interesting things while using the button JSON files. I'll leave them for your own pleasant surprise~ LOL.I shared my JSON file of the AI tool in the comfy-chatroom of Discord. Feel free to use it.I hope this article helped you make the radio button UI more easily. Enjoy~ 🤗😉⛈
84
17
Hunyuan-DiT: Recommendations

Hunyuan-DiT: Recommendations

ReviewHello everyone; I want to share some of my impressions about the Chinese model, Hunyuan-DiT from tencent. First of all let’s start with some mandatory data to know so we (westerns) can figure out what is meant for:Hunyuan-DiT works well as multi-modal dialogue with users (mainly Chinese and English language), the better explained your prompt the better your generation will be, is not necessary to introduce only keywords, despite it understands them quite well. In terms of rating HYDiT 1.2 is located between SDXL and SD3; is not as powerful than SD3, defeats SDXL almost in everything; for me is how SDXL should’ve be in first place; one of the best parts is that Hunyuan-DiT is compatible with almost all SDXL node suit.Hunyuan-DiT-v1.2, was trained with 1.5B parameters.mT5, was trained with 1.6B parameters.Recommeded VAE: sdxl-vae-fp16-fixRecommended Sampler: ddpm, ddim, or dpmmsPrompt as you’d like to do in SD1.5, don’t be shy and go further in term of length; HunyuanDiT combines two text encoders, a bilingual CLIP and a multilingual T5 encoder to improve language understanding and increase the context length; they divide your prompt on meaningful IDs and then process your entire prompt, their limit is 100 IDs or to 256 tokens. T5 works well on a variety of tasks out-of-the-box by prepending a different prefix to the input corresponding to each task.To improve your prompt, place your resumed prompt in the CLIP:TextEncoder node box (if you disabled t5), or place your extended prompt in the T5:TextEncoder node box (if you enabled t5).You can use the "simple" text encode node to only use one prompt, or you can use the regular one to pass different text to CLIP/T5.The worst is the model only benefits from moderated (high for TensorArt) step values: 40 steps are the basis in most cases.Comfyui (Comfyflow) (Example)TensorArt added all the elements to build a good flow for us; you should try it too.AdditionalWhat can we do in the Open-Source plan? (link)Official info for LoRA training (link)ReferencesAnalysis of HunYuan-DiT | https://arxiv.org/html/2405.08748v1Learn more of T5 | https://huggingface.co/docs/transformers/en/model_doc/t5How CLIP and T5 work together | https://arxiv.org/pdf/2205.11487
22
12
Unlock the Power of Detailed Beauty with TQ-HunYuan-More-Beautiful-Detail v1.7

Unlock the Power of Detailed Beauty with TQ-HunYuan-More-Beautiful-Detail v1.7

In the world of digital artistry, achieving that perfect blend of intricate details and stunning visuals can be a game-changer. That's where our latest model, TQ-HunYuan-More-Beautiful-Detail v1.7, comes into play. Designed with precision and a keen eye for aesthetics, this model is your go-to solution for elevating your artwork to new heights.What is TQ-HunYuan-More-Beautiful-Detail v1.7?TQ-HunYuan-More-Beautiful-Detail v1.7 is a state-of-the-art LoRA (Low-Rank Adaptation) model created to enhance the finer details in your digital creations. Whether you're working on portraits, landscapes, or abstract designs, this model ensures that every nuance and subtlety is brought to life with extraordinary clarity and beauty.Why Choose TQ-HunYuan-More-Beautiful-Detail v1.7?Unmatched Detail Enhancement: As the name suggests, this model excels at adding more beautiful details to your artwork. It meticulously enhances textures, refines edges, and highlights intricate patterns, making your creations visually striking.Versatility Across Genres: No matter the style or genre of your artwork, TQ-HunYuan-More-Beautiful-Detail v1.7 adapts seamlessly. From hyper-realistic portraits to fantastical landscapes, this model enhances every element with precision.User-Friendly Integration: Designed for ease of use, integrating TQ-HunYuan-More-Beautiful-Detail v1.7 into your workflow is straightforward. Compatible with various platforms and software, it allows artists of all levels to harness its power without a steep learning curve.Boost Your Creativity: By handling the intricate details, this model frees up your creative energy. Focus on the broader aspects of your work while TQ-HunYuan-More-Beautiful-Detail v1.7 takes care of the fine-tuning, resulting in a harmonious and polished final piece.How to Get StartedGetting started with TQ-HunYuan-More-Beautiful-Detail v1.7 is simple. Visit this link to access the model. Download and integrate it into your preferred digital art software, and watch as your creations transform with enhanced details and breathtaking beauty.Ready to take your art to the next level? Download TQ-HunYuan-More-Beautiful-Detail v1.7 now and start creating masterpieces with more beautiful detail than ever before.
40
4
SD3 - 3D lettering designer

SD3 - 3D lettering designer

SD3 understands prompts better compared to SDXL. You can use this to create interesting 3D lettering. For this purpose, use this WF! You can use a gradient as the background or any image you like. Have fun!Link to workflow: SD3 - 3D lettering designer | ComfyUI Workflow | Tensor.Art
13
Realistic Vision SD3

Realistic Vision SD3

Realistic VisionI am excited to present my latest Realistic checkpoint model based on SD3M. This model has undergone over 100k+ training steps, ensuring high-quality output.About This Model:This is a Photo Realistic model, capable of generating photorealistic images. No trigger words are needed. The model is designed to produce high-detail, high-resolution images that closely mimic real-life photographs.Configuration Used for Training:GPU: A6000x2Dataset: A mix of 5k stock photos and my own datasetBatch Size: 8Optimizer: AdamWScheduler: Cosine with restartsLearning Rate (LR): 1e-05Epoch: Target of 300 epochsCaptioning: WD14 and BLIP mixQuick Guide and Parameters:Clip Encoder: Not requiredVAE: Not requiredSampler: dpmpp_2mScheduler: sgm_uniformSampling Steps: 25+CFG Scale: 3+For better results, try using ComfyUI. Here is a workflow that is low-cost and efficient. Currently, upscaling is not possible due to specific reasons. I have reported the issue to the TA team, and hopefully, it will be fixed soon.Realistic VisionAspect Ratios for Demo:1:1 [1024x1024 square]8:5 [1216x768 landscape]4:3 [1152x896 landscape]3:2 [1216x832 landscape]7:5 [1176x840 landscape]16:9 [1344x768 landscape]21:9 [1536x640 landscape]19:9 [1472x704 landscape]3:4 [896x1152 portrait]2:3 [832x1216 portrait]5:7 [840x1176 portrait]9:16 [768x1344 portrait]9:21 [640x1536 portrait]5:8 [768x1216 portrait]9:19 [704x1472 portrait]Important: Do not include NSFW-related/mature words or censor words in your prompt. Doing so may result in unreliable or undesirable image outcomes.Note:This is not a merged or modified model. It is the original Realistic Vision fine-tuned model. Some users have been spreading incorrect information in the model's comment section. If you have any questions or want to know more, join my Discord server or share your thoughts in the comment section. Thank you for your time.
18
2
SDG - HunyuanDiT loras released

SDG - HunyuanDiT loras released

HunyuanDiT - Perfect cute animehttps://tensor.art/models/755812883138538240?source_id=nz-ypFjjk0C7pPcibn708xQiEnhance character appearance details, eyes, hair, colors, and drawings in anime styleHunyuanDiT - Realistic detailshttps://tensor.art/models/755789054659947864/HunyuanDiT-Realistic-details-V1Add more realistic details for imagesHunyuanDIT - Vivid colorhttps://tensor.art/models/755810413532312715?source_id=nz-ypFjjk0C7pPcibn708xQiEnhance vivid colors and details in photosHunyuan - Beauty Portraithttps://tensor.art/models/755789995257798458?source_id=nz-ypFjjk0C7pPcibn708xQiortrait within more details hair, skin...
5
2
Hunyuan model online training tutorial

Hunyuan model online training tutorial

EnglishToday, Iwill teach you how to use TensorArt to train an Hunyuan model online.Step 1: Open “Online Training.On the left side, you will see the dataset window, which is empty by default. You can upload some images to create a dataset or upload a dataset zip file. The zip file can include annotation files, following the same format as kohya-ss, where each image file corresponds to a text annotation file with the same name.In the model theme section on the right, you can choose from options such as anime characters, real people, 2.5D, standard, and custom.Here, we select “Base” and choose the Hunyuan model as the base model.For the base model parameter settings, we recommend setting the number of repetitions per image to 4 and the number of epochs to 16.、After uploading a processed dataset, if your dataset annotations include character names, you don’t need to specify a trigger word. Otherwise, you should assign a simple trigger word to your model, such as a character name or style name.Next, select an annotation file from the dataset to use as a preview prompt.If you want to use Professional Mode, click the button in the top right corner to switch to Professional Mode.In Professional Mode, it is recommended to double the learning rateand use the cosine_with_restarts learning rate scheduler. For the optimizer, you can choose AdamW8bit.Enable label shuffling and ensure that the first token remains unchanged (especially if you have a character name trigger word as the first token).Disable the noise offset feature, and you can set the convolution DIM to 8 and Alpha to 1.In the sample settings, add the Negative prompts, and then you can start the training process.In the training queue, you can view the current loss value chart and the four sample images generated for each epoch.Finally, you can choose the epoch with the best results to download to your local machine or publish directly on TensorArt.After a few minutes, your model will be deployed and ready.日本語今日、私はTensorArtを使用してHunyuanモデルをオンラインでトレーニングする方法を教えます。ステップ1: 「オンライントレーニング」を開きます。左側にデータセットウィンドウが表示され、デフォルトでは空です。データセットを作成するために画像をアップロードするか、データセットのzipファイルをアップロードできます。zipファイルには、kohya-ssと同じ形式のアノテーションファイルを含めることができ、各画像ファイルには同じ名前のテキストアノテーションファイルが対応しています。右側のモデルテーマセクションでは、アニメキャラクター、実在の人物、2.5D、標準、カスタムなどのオプションから選択できます。ここでは「Base」を選択し、Hunyuanモデルをベースモデルとして選びます。ベースモデルのパラメーター設定では、画像ごとの繰り返し回数を4、エポック数を16に設定することをお勧めします。 処理済みのデータセットをアップロードした後、データセットのアノテーションにキャラクター名が含まれている場合は、トリガーワードを指定する必要はありません。それ以外の場合は、キャラクター名やスタイル名など、モデルに簡単なトリガーワードを割り当ててください。 次に、プレビュー用プロンプトとして使用するために、データセットからアノテーションファイルを選択します。プロフェッショナルモードを使用したい場合は、右上隅のボタンをクリックしてプロフェッショナルモードに切り替えます。プロフェッショナルモードでは、学習率を倍増することをお勧めします。また、cosine_with_restarts学習率スケジューラーを使用してください。オプティマイザーとしては、AdamW8bitを選択できます。ラベルシャッフルを有効にし、最初のトークンが変更されないようにします(特にキャラクター名トリガーワードが最初のトークンの場合)。ノイズオフセット機能を無効にし、畳み込みDIMを8、Alphaを1に設定できます。サンプル設定でNegative promptsを追加し、その後、トレーニングプロセスを開始できます。トレーニングキューでは、現在の損失値チャートと各エポックごとに生成された4つのサンプル画像を表示できます。最後に、最良の結果が得られたエポックを選択して、ローカルマシンにダウンロードするか、直接TensorArtで公開できます。数分後には、モデルがデプロイされ、使用可能になります。한국인오늘은 TensorArt를 사용하여 Hunyuan 모델을 온라인에서 훈련하는 방법을 알려드리겠습니다.1단계: “온라인 훈련”을 엽니다.왼쪽에는 기본적으로 비어 있는 데이터셋 창이 표시됩니다. 데이터셋을 만들기 위해 이미지를 업로드하거나 데이터셋 zip 파일을 업로드할 수 있습니다. zip 파일에는 kohya-ss와 같은 형식의 주석 파일이 포함될 수 있으며, 각 이미지 파일에는 동일한 이름의 텍스트 주석 파일이 대응됩니다.오른쪽의 모델 테마 섹션에서는 애니메이션 캐릭터, 실제 인물, 2.5D, 표준, 사용자 정의 등 다양한 옵션 중에서 선택할 수 있습니다.여기에서는 “Base”를 선택하고 Hunyuan 모델을 기본 모델로 선택합니다.기본 모델 파라미터 설정에서는 이미지당 반복 횟수를 4로, 에포크 수를 16으로 설정하는 것을 권장합니다. 처리된 데이터셋을 업로드한 후, 데이터셋의 주석에 캐릭터 이름이 포함되어 있으면 트리거 단어를 지정할 필요가 없습니다. 그렇지 않으면 모델에 간단한 트리거 단어를 지정해야 합니다, 예를 들어 캐릭터 이름이나 스타일 이름 등. 다음으로, 미리 보기 프롬프트로 사용할 주석 파일을 데이터셋에서 선택합니다.전문 모드를 사용하려면, 오른쪽 상단의 버튼을 클릭하여 전문 모드로 전환합니다.전문 모드에서는 학습률을 두 배로 늘리는 것이 좋습니다.또한 cosine_with_restarts 학습률 스케줄러를 사용합니다. 옵티마이저로는 AdamW8bit을 선택할 수 있습니다.레이블 셔플을 활성화하고 첫 번째 토큰이 변경되지 않도록 합니다(특히 캐릭터 이름 트리거 단어가 첫 번째 토큰인 경우).노이즈 오프셋 기능을 비활성화하고, 컨볼루션 DIM을 8로, Alpha를 1로 설정할 수 있습니다.샘플 설정에서 Negative prompts를 추가한 후, 훈련 프로세스를 시작할 수 있습니다.훈련 대기열에서 현재 손실 값 차트와 각 에포크에 대해 생성된 4개의 샘플 이미지를 볼 수 있습니다.마지막으로, 가장 좋은 결과를 얻은 에포크를 선택하여 로컬 컴퓨터로 다운로드하거나 직접 TensorArt에 게시할 수 있습니다.몇 분 후, 모델이 배포되고 사용 가능해집니다.Tiếng ViệtHôm nay, tôi sẽ hướng dẫn bạn cách sử dụng TensorArt để đào tạo mô hình Hunyuan trực tuyến.Bước 1: Mở “Đào tạo trực tuyến.”Ở bên trái, bạn sẽ thấy cửa sổ tập dữ liệu, mặc định là trống. Bạn có thể tải lên một số hình ảnh để tạo tập dữ liệu hoặc tải lên tệp zip của tập dữ liệu. Tệp zip có thể bao gồm các tệp chú thích, theo cùng một định dạng như kohya-ss, trong đó mỗi tệp hình ảnh tương ứng với một tệp chú thích văn bản cùng tên.Ở phần chủ đề mô hình bên phải, bạn có thể chọn từ các tùy chọn như nhân vật anime, người thật, 2.5D, tiêu chuẩn và tùy chỉnh.Tại đây, chúng ta chọn “Base” và chọn mô hình Hunyuan làm mô hình cơ bản.Đối với cài đặt tham số của mô hình cơ bản, chúng tôi khuyên bạn nên đặt số lần lặp lại trên mỗi hình ảnh là 4 và số epoch là 16. Sau khi tải lên tập dữ liệu đã xử lý, nếu các chú thích của tập dữ liệu của bạn bao gồm tên nhân vật, bạn không cần phải chỉ định từ kích hoạt. Ngược lại, bạn nên gán một từ kích hoạt đơn giản cho mô hình của mình, chẳng hạn như tên nhân vật hoặc tên phong cách. Tiếp theo, chọn một tệp chú thích từ tập dữ liệu để sử dụng làm lời nhắc xem trước.Nếu bạn muốn sử dụng Chế độ Chuyên nghiệp, hãy nhấp vào nút ở góc trên bên phải để chuyển sang Chế độ Chuyên nghiệp.Trong Chế độ Chuyên nghiệp, nên gấp đôi tỷ lệ học.Và sử dụng bộ lập lịch tỷ lệ học cosine_with_restarts. Đối với bộ tối ưu hóa, bạn có thể chọn AdamW8bit.Kích hoạt xáo trộn nhãn và đảm bảo rằng mã thông báo đầu tiên không thay đổi (đặc biệt nếu bạn có từ kích hoạt tên nhân vật là mã thông báo đầu tiên).Tắt tính năng dịch chuyển tiếng ồn và bạn có thể đặt DIM tích chập là 8 và Alpha là 1.Trong cài đặt mẫu, thêm các Lời nhắc tiêu cực, sau đó bạn có thể bắt đầu quá trình đào tạo.Trong hàng đợi đào tạo, bạn có thể xem biểu đồ giá trị tổn thất hiện tại và bốn hình ảnh mẫu được tạo ra cho mỗi epoch.Cuối cùng, bạn có thể chọn epoch có kết quả tốt nhất để tải xuống máy tính của bạn hoặc xuất bản trực tiếp trên TensorArt.Sau vài phút, mô hình của bạn sẽ được triển khai và sẵn sàng sử dụng.españolHoy, te enseñaré cómo usar TensorArt para entrenar un modelo Hunyuan en línea.Paso 1: Abre “Entrenamiento en línea.”A la izquierda, verás la ventana del conjunto de datos, que está vacía por defecto. Puedes subir algunas imágenes para crear un conjunto de datos o subir un archivo zip del conjunto de datos. El archivo zip puede incluir archivos de anotación, siguiendo el mismo formato que kohya-ss, donde cada archivo de imagen corresponde a un archivo de anotación de texto con el mismo nombre.En la sección de temas del modelo a la derecha, puedes elegir entre opciones como personajes de anime, personas reales, 2.5D, estándar y personalizado.Aquí, seleccionamos “Base” y elegimos el modelo Hunyuan como el modelo base.Para la configuración de parámetros del modelo base, te recomendamos configurar el número de repeticiones por imagen a 4 y el número de épocas a 16. Después de subir un conjunto de datos procesado, si las anotaciones de tu conjunto de datos incluyen nombres de personajes, no necesitas especificar una palabra de activación. De lo contrario, deberías asignar una palabra de activación simple a tu modelo, como un nombre de personaje o un nombre de estilo. A continuación, selecciona un archivo de anotación del conjunto de datos para usarlo como un aviso de vista previa.Si deseas usar el Modo Profesional, haz clic en el botón en la esquina superior derecha para cambiar al Modo Profesional.En el Modo Profesional, se recomienda duplicar la tasa de aprendizaje.Y usar el programador de tasa de aprendizaje cosine_with_restarts. Para el optimizador, puedes elegir AdamW8bit.Habilita el barajado de etiquetas y asegúrate de que el primer token permanezca sin cambios (especialmente si tienes una palabra de activación de nombre de personaje como el primer token).Desactiva la función de desplazamiento de ruido y puedes configurar el DIM de convolución a 8 y Alpha a 1.En la configuración de muestra, añade los Avisos Negativos, y luego puedes comenzar el proceso de entrenamiento.En la cola de entrenamiento, puedes ver el gráfico del valor de pérdida actual y las cuatro imágenes de muestra generadas para cada época.Finalmente, puedes elegir la época con los mejores resultados para descargarla a tu máquina local o publicarla directamente en TensorArt.Después de unos minutos, tu modelo estará desplegado y listo para usar.
16
4
Online Training SD3 Model Tutorial

Online Training SD3 Model Tutorial

EnglishToday, Iwill teach you how to use TensorArt to train an SD3 model online.Step 1: Open “Online Training.On the left side, you will see the dataset window, which is empty by default. You can upload some images to create a dataset or upload a dataset zip file. The zip file can include annotation files, following the same format as kohya-ss, where each image file corresponds to a text annotation file with the same name.In the model theme section on the right, you can choose from options such as anime characters, real people, 2.5D, standard, and custom.Here, we select “Base” and choose the SD3 model as the base model.For the base model parameter settings, we recommend setting the number of repetitions per image to 4 and the number of epochs to 16.、After uploading a processed dataset, if your dataset annotations include character names, you don’t need to specify a trigger word. Otherwise, you should assign a simple trigger word to your model, such as a character name or style name.Next, select an annotation file from the dataset to use as a preview prompt.If you want to use Professional Mode, click the button in the top right corner to switch to Professional Mode.In Professional Mode, it is recommended to double the learning rateand use the cosine_with_restarts learning rate scheduler. For the optimizer, you can choose AdamW8bit.Enable label shuffling and ensure that the first token remains unchanged (especially if you have a character name trigger word as the first token).Disable the noise offset feature, and you can set the convolution DIM to 8 and Alpha to 1.In the sample settings, add the Negative prompts, and then you can start the training process.In the training queue, you can view the current loss value chart and the four sample images generated for each epoch.Finally, you can choose the epoch with the best results to download to your local machine or publish directly on TensorArt.After a few minutes, your model will be deployed and ready.日本語今日は、TensorArtを使用してオンラインでSD3モデルをトレーニングする方法を教えます。ステップ1: 「オンライントレーニング」を開きます。左側にデータセットウィンドウが表示され、デフォルトでは空です。データセットを作成するために画像をアップロードするか、データセットのzipファイルをアップロードできます。zipファイルには、kohya-ssと同じ形式のアノテーションファイルを含めることができ、各画像ファイルには同じ名前のテキストアノテーションファイルが対応しています。右側のモデルテーマセクションでは、アニメキャラクター、実在の人物、2.5D、標準、カスタムなどのオプションから選択できます。ここでは、「ベース」を選択し、SD3モデルをベースモデルとして選びます。ベースモデルのパラメーター設定では、画像ごとの繰り返し回数を4、エポック数を16に設定することをお勧めします。 処理済みのデータセットをアップロードした後、データセットのアノテーションにキャラクター名が含まれている場合は、トリガーワードを指定する必要はありません。それ以外の場合は、キャラクター名やスタイル名など、モデルに簡単なトリガーワードを割り当ててください。 次に、プレビュー用プロンプトとして使用するために、データセットからアノテーションファイルを選択します。プロフェッショナルモードを使用したい場合は、右上隅のボタンをクリックしてプロフェッショナルモードに切り替えます。プロフェッショナルモードでは、学習率を倍増することをお勧めします。また、cosine_with_restarts学習率スケジューラーを使用してください。オプティマイザーとしては、AdamW8bitを選択できます。ラベルシャッフルを有効にし、最初のトークンが変更されないようにします(特にキャラクター名トリガーワードが最初のトークンの場合)。ノイズオフセット機能を無効にし、畳み込みDIMを8、Alphaを1に設定できます。サンプル設定でNegative promptsを追加し、その後、トレーニングプロセスを開始できます。トレーニングキューでは、現在の損失値チャートと各エポックごとに生成された4つのサンプル画像を表示できます。最後に、最良の結果が得られたエポックを選択して、ローカルマシンにダウンロードするか、直接TensorArtで公開できます。数分後には、モデルがデプロイされ、使用可能になります。한국인오늘은 TensorArt를 사용하여 SD3 모델을 온라인으로 훈련하는 방법을 가르쳐 드리겠습니다.1단계: “온라인 훈련”을 엽니다.왼쪽에는 기본적으로 비어 있는 데이터셋 창이 표시됩니다. 데이터셋을 만들기 위해 이미지를 업로드하거나 데이터셋 zip 파일을 업로드할 수 있습니다. zip 파일에는 kohya-ss와 같은 형식의 주석 파일이 포함될 수 있으며, 각 이미지 파일에는 동일한 이름의 텍스트 주석 파일이 대응됩니다.오른쪽의 모델 테마 섹션에서는 애니메이션 캐릭터, 실제 인물, 2.5D, 표준, 사용자 정의 등 다양한 옵션 중에서 선택할 수 있습니다.여기에서는 “Base”를 선택하고 SD3 모델을 기본 모델로 선택합니다.기본 모델 파라미터 설정에서는 이미지당 반복 횟수를 4로, 에포크 수를 16으로 설정하는 것을 권장합니다. 처리된 데이터셋을 업로드한 후, 데이터셋의 주석에 캐릭터 이름이 포함되어 있으면 트리거 단어를 지정할 필요가 없습니다. 그렇지 않으면 모델에 간단한 트리거 단어를 지정해야 합니다, 예를 들어 캐릭터 이름이나 스타일 이름 등. 다음으로, 미리 보기 프롬프트로 사용할 주석 파일을 데이터셋에서 선택합니다.전문 모드를 사용하려면, 오른쪽 상단의 버튼을 클릭하여 전문 모드로 전환합니다.전문 모드에서는 학습률을 두 배로 늘리는 것이 좋습니다.또한 cosine_with_restarts 학습률 스케줄러를 사용합니다. 옵티마이저로는 AdamW8bit을 선택할 수 있습니다.레이블 셔플을 활성화하고 첫 번째 토큰이 변경되지 않도록 합니다(특히 캐릭터 이름 트리거 단어가 첫 번째 토큰인 경우).노이즈 오프셋 기능을 비활성화하고, 컨볼루션 DIM을 8로, Alpha를 1로 설정할 수 있습니다.샘플 설정에서 Negative prompts를 추가한 후, 훈련 프로세스를 시작할 수 있습니다.훈련 대기열에서 현재 손실 값 차트와 각 에포크에 대해 생성된 4개의 샘플 이미지를 볼 수 있습니다.마지막으로, 가장 좋은 결과를 얻은 에포크를 선택하여 로컬 컴퓨터로 다운로드하거나 직접 TensorArt에 게시할 수 있습니다.몇 분 후, 모델이 배포되고 사용 가능해집니다.Tiếng ViệtHôm nay, tôi sẽ hướng dẫn bạn cách sử dụng TensorArt để huấn luyện mô hình SD3 trực tuyến.Bước 1: Mở “Đào tạo trực tuyến.”Ở bên trái, bạn sẽ thấy cửa sổ tập dữ liệu, mặc định là trống. Bạn có thể tải lên một số hình ảnh để tạo tập dữ liệu hoặc tải lên tệp zip của tập dữ liệu. Tệp zip có thể bao gồm các tệp chú thích, theo cùng một định dạng như kohya-ss, trong đó mỗi tệp hình ảnh tương ứng với một tệp chú thích văn bản cùng tên.Ở phần chủ đề mô hình bên phải, bạn có thể chọn từ các tùy chọn như nhân vật anime, người thật, 2.5D, tiêu chuẩn và tùy chỉnh.Tại đây, chúng ta chọn “Cơ bản” và chọn mô hình SD3 làm mô hình cơ sở.Đối với cài đặt tham số của mô hình cơ bản, chúng tôi khuyên bạn nên đặt số lần lặp lại trên mỗi hình ảnh là 4 và số epoch là 16. Sau khi tải lên tập dữ liệu đã xử lý, nếu các chú thích của tập dữ liệu của bạn bao gồm tên nhân vật, bạn không cần phải chỉ định từ kích hoạt. Ngược lại, bạn nên gán một từ kích hoạt đơn giản cho mô hình của mình, chẳng hạn như tên nhân vật hoặc tên phong cách. Tiếp theo, chọn một tệp chú thích từ tập dữ liệu để sử dụng làm lời nhắc xem trước.Nếu bạn muốn sử dụng Chế độ Chuyên nghiệp, hãy nhấp vào nút ở góc trên bên phải để chuyển sang Chế độ Chuyên nghiệp.Trong Chế độ Chuyên nghiệp, nên gấp đôi tỷ lệ học.Và sử dụng bộ lập lịch tỷ lệ học cosine_with_restarts. Đối với bộ tối ưu hóa, bạn có thể chọn AdamW8bit.Kích hoạt xáo trộn nhãn và đảm bảo rằng mã thông báo đầu tiên không thay đổi (đặc biệt nếu bạn có từ kích hoạt tên nhân vật là mã thông báo đầu tiên).Tắt tính năng dịch chuyển tiếng ồn và bạn có thể đặt DIM tích chập là 8 và Alpha là 1.Trong cài đặt mẫu, thêm các Lời nhắc tiêu cực, sau đó bạn có thể bắt đầu quá trình đào tạo.Trong hàng đợi đào tạo, bạn có thể xem biểu đồ giá trị tổn thất hiện tại và bốn hình ảnh mẫu được tạo ra cho mỗi epoch.Cuối cùng, bạn có thể chọn epoch có kết quả tốt nhất để tải xuống máy tính của bạn hoặc xuất bản trực tiếp trên TensorArt.Sau vài phút, mô hình của bạn sẽ được triển khai và sẵn sàng sử dụng.españolHoy, les enseñaré cómo utilizar TensorArt para entrenar un modelo SD3 en línea.Paso 1: Abre “Entrenamiento en línea.”A la izquierda, verás la ventana del conjunto de datos, que está vacía por defecto. Puedes subir algunas imágenes para crear un conjunto de datos o subir un archivo zip del conjunto de datos. El archivo zip puede incluir archivos de anotación, siguiendo el mismo formato que kohya-ss, donde cada archivo de imagen corresponde a un archivo de anotación de texto con el mismo nombre.En la sección de temas del modelo a la derecha, puedes elegir entre opciones como personajes de anime, personas reales, 2.5D, estándar y personalizado.Aquí, seleccionamos “Base” y elegimos el modelo SD3 como el modelo base.Para la configuración de parámetros del modelo base, te recomendamos configurar el número de repeticiones por imagen a 4 y el número de épocas a 16. Después de subir un conjunto de datos procesado, si las anotaciones de tu conjunto de datos incluyen nombres de personajes, no necesitas especificar una palabra de activación. De lo contrario, deberías asignar una palabra de activación simple a tu modelo, como un nombre de personaje o un nombre de estilo. A continuación, selecciona un archivo de anotación del conjunto de datos para usarlo como un aviso de vista previa.Si deseas usar el Modo Profesional, haz clic en el botón en la esquina superior derecha para cambiar al Modo Profesional.En el Modo Profesional, se recomienda duplicar la tasa de aprendizaje.Y usar el programador de tasa de aprendizaje cosine_with_restarts. Para el optimizador, puedes elegir AdamW8bit.Habilita el barajado de etiquetas y asegúrate de que el primer token permanezca sin cambios (especialmente si tienes una palabra de activación de nombre de personaje como el primer token).Desactiva la función de desplazamiento de ruido y puedes configurar el DIM de convolución a 8 y Alpha a 1.En la configuración de muestra, añade los Avisos Negativos, y luego puedes comenzar el proceso de entrenamiento.En la cola de entrenamiento, puedes ver el gráfico del valor de pérdida actual y las cuatro imágenes de muestra generadas para cada época.Finalmente, puedes elegir la época con los mejores resultados para descargarla a tu máquina local o publicarla directamente en TensorArt.Después de unos minutos, tu modelo estará desplegado y listo para usar.
如何使用混元DiT在线训练

如何使用混元DiT在线训练

首先点击右上角的头像,在弹出的下拉框中选择我训练的模型,进入训练中心。如果之前有训练过模型,这里会看到许多训练任务。然后选择在线训练按钮进行一次训练。左侧是数据集窗口,默认没有任何数据。您可以上传一些图片作为数据集,或者上传一个数据集压缩包,压缩包可以包含标注文件,格式和kohya-ss一样,每个图片文件对应一个同名的标注文件txt。右边的模型主题中可以选择二次元人物、真实人物、2.5D、标准以及自定义。训练混元模型这里我们选择标准,在使用底模中选择混元1.2模型。混元模型使用了40depth的块,所以非常大,训练相对速度较慢,需要更高的学习率,默认使用4e-4,默认单张图片重复次数5,优化器AdamW。基础模式下参数选择,推荐单张图片重复次数5,轮数为16。上传一个处理好的数据集后,如果你的数据集标注中有人物名,可以不写触发词。否则你应该给你的模型起一个简单的触发词,例如人物名称或者风格名称。接着从数据集中选择一个标注文件作为预览提示词。如果你想使用专业模式,选择右上角按钮切换到专业模式。专业模式推荐学习率翻倍,然后使用cosine_with_restarts学习率调度器,优化器选择AdamW或者AdamW8bit。开启打乱标签(shuffle),并且保持第1个token(如果你有一个人名触发词在第一个)关闭噪声偏移功能,卷积DIM和Alpha可以选择8和1。在样图设置中追加填写反向提示词,接下来就可以开始训练了。在训练队列中,你可以看到当前loss值变化表以及每轮epoch产生的4张样图。最后可以选择效果最好的epoch下载到本地或者直接在tensorart上发布。
6
SD3 - composition repair

SD3 - composition repair

SD3 can generate interesting images, but it has a huge problem with the human body. However, I noticed that simply reducing the image size to 60% can, in most cases, eliminate issues with image composition as well as extra hands or legs. This workflow does not solve the problem of having six fingers, etc. :)Base model: https://tensor.art/models/751330255836302856/Aderek-SD3-v1 or https://civitai.com/models/600179/aderek-sd3Look at the image below. You might say: "Hey, nothing's wrong here." Well, that's because you're already seeing the generation based on the reduced size. Below, you have the original image.Use composition on to use this trick&tips.Have fun!Support Paweł Tomczuk on Ko-fi! ❤️. ko-fi.com/aderek514 - Ko-fi ❤️ Where creators get support from fans through donations, memberships, shop sales and more! The original 'Buy Me a Coffee' Page.Visit my DeviantArt page: Aderek - Hobbyist, Digital Artist | DeviantArt
10
2
🆘 ERROR | Exception

🆘 ERROR | Exception

Exception (routeId: 7544339967855538950230)Suspect nodes:<string function>. <LayeStyle>, <LayerUtility>, <FaceDetailer>, many <TextBox>, <Bumpmap>After some reseach (on my own) I've found<FaceDetailer> node is completely broken<TextBox> and <MultiLine:Textbox> node will cause this error if you introduce more than 250+ characters, I'm not very sure about this number, but you won't be able to introduce a decent amount of text anymore.More than 40 nodes, despite its function will couse this error.How do i know this? Well I made a functional comfyflow following those rules:https://tensor.art/template/754955251181895419The next functional comfyflow suddelny stopped from generating, it's almost the same flow than the previous, but with <FaceDetailer> and large text strings to polish the prompt. It works again yay!https://tensor.art/template/752678510492967987 proof it really worked (here)I feel bad for you if this error suddenly disrupt your day; feel bad for me cuz I bought the yearly membership of this broken product I can't refound. I'll be happy to delete this bad review if you fix this error.News081124 | <String Function> has been taken down. Comfyflow works slowly (but works)081024 | eveything is broken again lmao, we cant generate outside TAMS.080624 | <reroute> output node could trigger this error when linked to many inputs.072824 | <FaceDetailer> node seems to work again.
4
Upscaling in ComfyUI: ¿Algorithm or Latent?

Upscaling in ComfyUI: ¿Algorithm or Latent?

Hello again! In this little article I want to explain the upscaling methods that I know in ComfyUI and that I have researched. I hope they will help you and that you can use them in the creation of your workflows and AI tools. In addition, remember that if you have any useful knowledge, you can share it in the comments section to enrich the topic. Also, please excuse any spelling mistakes; I am just learning English hehe.¡Let’s get to the point!To the best of my knowledge, there are two widely used ways in ComfyUI to achieve uspcaling (you decide which one to use according to your needs). The two options are: Algorithm Method or Latent Method.Algorithm Method:This is one of the most commonly used method, and is readily available. It consists of loading an upscaling model, and connecting it to the workflow. That way the image pixels are manipulated as the user wishes. It is very similar to the upscale method used in the normal way of creating images in Tensor Art.The following nodes are needed:A. Load Upscale Model.B. Upscale Image (Using Model).These nodes are connected to the workflow between the “VAE Decode” and “Save Image” nodes; as shown in the image. Once this structure is created, you can choose from all the different models offered by the “Load Upscale Model” node, ranging from “2x-ESRGAN.pth” to “SwimIR_4x”. You can use any of the 23 available models and experiment with any of them. You just have to click on the node and the list will be displayed.This can also be achieved in other ways by using another node such as “Upscale Image By”. The structure is simpler to create because only that node is connected between the VAE decode and Save Image as shown in the following image.Once the node is connected, you are free to select the mode in which you want to upscale the image (Upscale_method) and you can also set the scale to which you can recondition the image pixel value (Scale By).Strengths and Weaknesses of the Algorithm Method:Among the strengths of this method are its ease of integration into the workflow and its advantage of choosing between several upscaling model options. It also allows fast generation both in the ComfyUI and in the use of AI tools.However, among its weaknesses, it is not very effective in some specific contexts. For example: the algorithm can upscale the image pixels but does not alter the actual image size; causing the generated image itself to end up being blurred in some cases.Latent Method:This is the other alternative option to the algorithm method. It is focused on highlighting image details and maximizing quality. This method is also one of the most used in the Workflow mode of different visual content creation platforms with artificial intelligence. Here, upscaling is performed while the image is generated from latent space (Latent space is where the IA takes data from the prompt, deconstructs it for analysis and then reconstructs it to represent it in an image).The Latent Upscale node is placed between the two Ksamplers. While the first Ksampler is connected to the “Empty Latent Image” node, the second one is connected to the “VAE Decode” to ensure the correct processing and representation of the generated image.It should be noted that the “Empty Latent Image” node and the “VAE Decode” node are already included by default in the Text2Image templates in WorkFlow mode. (For more information about Text2Image, you can see my other article called “ComfyUi: Text2Image Basic Glossary”).It is important to take into consideration that for this method to work properly, you have to know how to create a correct balance between the original size of the image and its upscaled size. For example, you can generate a 512x512 image and upscale it to 1024x1024; but it is not recommended to make a 512x512 image (square image) and upscale it to 768x11152 (rectangular image) since the shape of the image would not be compatible with its uspcale version. For this reason you have to pay attention to the values of the “Empty Latent Image” and the “Latent Upscale”, so that these are always proportional.In the “Empty Latent Image” node you must place the original image dimensions (for example: 768x1152); while in the “Latent Upscale” node you must place the resized image dimensions (for example: 1152x1728). In this way you are given the freedom to set the image size to your own discretion. For this I always recommend to look at the size and upscale of the normal mode in which we create illustrations, this way we will always know which values to set and which will be compatible. As you can see in the image. You look at those values, and then write them to the nodes listed above.Once everything is connected and configured, you are able to have images of any size you want. You can experiment to your taste.Strengths and Weaknesses of the Latent Method:As strengths this option should be highlighted that it allows you to access excellent quality images if everything is correctly configured. It also allows you to create images of a custom size and upscale with the values you want. It brings out the details in both SD and XL images.As negative points we have to configure everything manually every time you want to change the size of the images or the shape of the same. Also, this method is just a little bit slower in the generation process compared to the algorithm method.Which is better: ¿Algorithm or Latent?Neither method is better than the other. Both are useful in different contexts. Remember that workflows will be different from user to user, because we all have different ways of creating and designing things.It all depends on your taste and whether you want something simpler or more elaborate. I hope the explanation in this article has helped you to make Workflows more complex and to make it easier to make the images you want.Extra Tip:If you do not find any of the nodes outlined in this document. You can double click on any empty place in the workflow and you can search for the name of the node you are looking for. Just remember to type the name without spaces.
12
2
Controlnet with SD3

Controlnet with SD3

Today, I noticed that I can add ControlNet to the SD3 model.The Tiled function works very well, so I incorporated it into my workflow and created a group for generating artistic images based on a given photo or a previously generated image. In the main part of the workflow, I simply set a very short prompt, like "grass, flowers," and I get an image that blends grass and flowers in an arrangement resembling the base photo.https://youtu.be/sv35wKNiFGsControlnet with SD3 | ComfyUI Workflow | Tensor.Art
4
如何使用SD3在线训练

如何使用SD3在线训练

首先点击右上角的头像,在弹出的下拉框中选择我训练的模型,进入训练中心。如果之前有训练过模型,这里会看到许多训练任务。然后选择在线训练按钮进行一次训练。左侧是数据集窗口,默认没有任何数据。您可以上传一些图片作为数据集,或者上传一个数据集压缩包,压缩包可以包含标注文件,格式和kohya-ss一样,每个图片文件对应一个同名的标注文件txt。右边的模型主题中可以选择二次元人物、真实人物、2.5D、标准以及自定义。这里我们选择自定义,在使用底模中选择SD3模型。注意在选择版本中下拉框内选择T5XXL的版本,这样才可以训练T5文本编码器。基础模式下参数选择,推荐单张图片重复次数4,轮数为16。上传一个处理好的数据集后,如果你的数据集标注中有人物名,可以不写触发词。否则你应该给你的模型起一个简单的触发词,例如人物名称或者风格名称。接着从数据集中选择一个标注文件作为预览提示词。如果你想使用专业模式,选择右上角按钮切换到专业模式。专业模式推荐学习率翻倍,然后使用cosine_with_restarts学习率调度器,优化器可以选择AdamW8bit。开启打乱标签(shuffle),并且保持第1个token(如果你有一个人名触发词在第一个)关闭噪声偏移功能,卷积DIM和Alpha可以选择8和1。在样图设置中追加填写反向提示词,接下来就可以开始训练了。在训练队列中,你可以看到当前loss值变化表以及每轮epoch产生的4张样图。最后可以选择效果最好的epoch下载到本地或者直接在tensorart上发布。
4
1
SD3 - training on your own PC

SD3 - training on your own PC

So first, you need to update your version of OneTrainer.Second, u need dowload ALL files and folders (and rename)stabilityai/stable-diffusion-3-medium-diffusers at main (huggingface.co)then u put it:With float16 output lora has only 36MB:This is my setting for a style training:My checkpoint to testing u can dowload for free:Aderek SD3 - v1 | Stable Diffusion Model - Checkpoint | Tensor.Artand my loras: Aderek514's Profile | Tensor.ArtSo, good luck!
11
2
ReActor Node for ComfyUI (Face Swap)

ReActor Node for ComfyUI (Face Swap)

ReActor Node for ComfyUI 👉Downlond👈 https://github.com/lingkops4/lingko-FaceReActor-Nodeworkflowhttps://github.com/lingkops4/lingko-FaceReActor-Node/blob/main/face_reactor_workflows.jsonThe Fast and Simple Face Swap Extension Node for ComfyUI, based on ReActor SD-WebUI Face Swap ExtensionThis Node goes without NSFW filter (uncensored, use it on your own responsibility)| Installation | Usage | Troubleshooting | Updating | Disclaimer | Credits | Note!✨What's new in the latest update✨💡0.5.1 ALPHA1Support of GPEN 1024/2048 restoration models (available in the HF dataset https://huggingface.co/datasets/Gourieff/ReActor/tree/main/models/facerestore_models)👈[]~( ̄▽ ̄)~*ReActorFaceBoost Node - an attempt to improve the quality of swapped faces. The idea is to restore and scale the swapped face (according to the face_size parameter of the restoration model) BEFORE pasting it to the target image (via inswapper algorithms), more information is here (PR#321)InstallationSD WebUI: AUTOMATIC1111 or SD.NextStandalone (Portable) ComfyUI for WindowsUsageYou can find ReActor Nodes inside the menu ReActor or by using a search (just type "ReActor" in the search field)List of Nodes:••• Main Nodes •••💡ReActorFaceSwap (Main Node Download)👈[]~( ̄▽ ̄)~*ReActorFaceSwapOpt (Main Node with the additional Options input)ReActorOptions (Options for ReActorFaceSwapOpt)ReActorFaceBoost (Face Booster Node)ReActorMaskHelper (Masking Helper)••• Operations with Face Models •••ReActorSaveFaceModel (Save Face Model)ReActorLoadFaceModel (Load Face Model)ReActorBuildFaceModel (Build Blended Face Model)ReActorMakeFaceModelBatch (Make Face Model Batch)••• Additional Nodes •••ReActorRestoreFace (Face Restoration)ReActorImageDublicator (Dublicate one Image to Images List)ImageRGBA2RGB (Convert RGBA to RGB)Connect all required slots and run the query.Main Node Inputsinput_image - is an image to be processed (target image, analog of "target image" in the SD WebUI extension);Supported Nodes: "Load Image", "Load Video" or any other nodes providing images as an output;source_image - is an image with a face or faces to swap in the input_image (source image, analog of "source image" in the SD WebUI extension);Supported Nodes: "Load Image" or any other nodes providing images as an output;face_model - is the input for the "Load Face Model" Node or another ReActor node to provide a face model file (face embedding) you created earlier via the "Save Face Model" Node;Supported Nodes: "Load Face Model", "Build Blended Face Model";Main Node OutputsIMAGE - is an output with the resulted image;Supported Nodes: any nodes which have images as an input;FACE_MODEL - is an output providing a source face's model being built during the swapping process;Supported Nodes: "Save Face Model", "ReActor", "Make Face Model Batch";Face RestorationSince version 0.3.0 ReActor Node has a buil-in face restoration.Just download the models you want (see Installation instruction) and select one of them to restore the resulting face(s) during the faceswap. It will enhance face details and make your result more accurate.Face IndexesBy default ReActor detects faces in images from "large" to "small".You can change this option by adding ReActorFaceSwapOpt node with ReActorOptions.And if you need to specify faces, you can set indexes for source and input images.Index of the first detected face is 0.You can set indexes in the order you need.E.g.: 0,1,2 (for Source); 1,0,2 (for Input).This means: the second Input face (index = 1) will be swapped by the first Source face (index = 0) and so on.GendersYou can specify the gender to detect in images.ReActor will swap a face only if it meets the given condition.💡Face ModelsSince version 0.4.0 you can save face models as "safetensors" files (stored in ComfyUI\models\reactor\faces) and load them into ReActor implementing different scenarios and keeping super lightweight face models of the faces you use.To make new models appear in the list of the "Load Face Model" Node - just refresh the page of your ComfyUI web application.(I recommend you to use ComfyUI Manager - otherwise you workflow can be lost after you refresh the page if you didn't save it before that).TroubleshootingI. (For Windows users) If you still cannot build Insightface for some reasons or just don't want to install Visual Studio or VS C++ Build Tools - do the following:(ComfyUI Portable) From the root folder check the version of Python:run CMD and type python_embeded\python.exe -VDownload prebuilt Insightface package for Python 3.10 or for Python 3.11 (if in the previous step you see 3.11) or for Python 3.12 (if in the previous step you see 3.12) and put into the stable-diffusion-webui (A1111 or SD.Next) root folder (where you have "webui-user.bat" file) or into ComfyUI root folder if you use ComfyUI PortableFrom the root folder run:(SD WebUI) CMD and .\venv\Scripts\activate(ComfyUI Portable) run CMDThen update your PIP:(SD WebUI) python -m pip install -U pip(ComfyUI Portable) python_embeded\python.exe -m pip install -U pip💡Then install Insightface:(SD WebUI) pip install insightface-0.7.3-cp310-cp310-win_amd64.whl (for 3.10) or pip install insightface-0.7.3-cp311-cp311-win_amd64.whl (for 3.11) or pip install insightface-0.7.3-cp312-cp312-win_amd64.whl (for 3.12)(ComfyUI Portable) python_embeded\python.exe -m pip install insightface-0.7.3-cp310-cp310-win_amd64.whl (for 3.10) or python_embeded\python.exe -m pip install insightface-0.7.3-cp311-cp311-win_amd64.whl (for 3.11) or python_embeded\python.exe -m pip install insightface-0.7.3-cp312-cp312-win_amd64.whl (for 3.12)Enjoy!II. "AttributeError: 'NoneType' object has no attribute 'get'"This error may occur if there's smth wrong with the model file inswapper_128.onnx💡Try to download it manually from here and put it to the ComfyUI\models\insightface replacing existing oneIII. "reactor.execute() got an unexpected keyword argument 'reference_image'"This means that input points have been changed with the latest updateRemove the current ReActor Node from your workflow and add it againIV. ControlNet Aux Node IMPORT failed error when using with ReActor NodeClose ComfyUI if it runsGo to the ComfyUI root folder, open CMD there and run:python_embeded\python.exe -m pip uninstall -y opencv-python opencv-contrib-python opencv-python-headlesspython_embeded\python.exe -m pip install opencv-python==4.7.0.72That's it!reactor+controlnetV. "ModuleNotFoundError: No module named 'basicsr'" or "subprocess-exited-with-error" during future-0.18.3 installationDownload https://github.com/Gourieff/Assets/raw/main/comfyui-reactor-node/future-0.18.3-py3-none-any.whlPut it to ComfyUI root And run:python_embeded\python.exe -m pip install future-0.18.3-py3-none-any.whlThen:python_embeded\python.exe -m pip install basicsrVI. "fatal: fetch-pack: invalid index-pack output" when you try to git clone the repository"Try to clone with --depth=1 (last commit only):git clone --depth=1 https://github.com/Gourieff/comfyui-reactor-nodeThen retrieve the rest (if you need):git fetch --unshallow
24
13
ComfyUi: Text2Image Basic Glossary

ComfyUi: Text2Image Basic Glossary

Hello! This is my first article; I hope it will be of benefit to the person who reads it. I still have limited knowledge about WorkFlow; but I have researched and learned little by little. If anyone would like to contribute some content; you are totally free to do so. Thank you.I made this article to give a brief and basic explanation about basic concepts about Comfyui or WorkFlow. This is a technology with many possibilities and it would be great to make it easier to use for everyone! What is Workflow?Workflow is one of the two main image generation systems that Tensor Art has at the moment. It corresponds to a generation method that is characterized by a great capacity to stimulate the creativity of the users; also, it allows us to access to some Pro features being Free users.How do I access the WorkFlow mode?To access the WorkFlow mode, you must place the mouse cursor on the “Create” tab as if you were going to create an image by conventional means. Once you have done that; click on the “ComfyFlow” option and you are done.After that, you will see a tab with two options “New WorkFlow” and “Import WorkFlow”. The first one allows you to start a workflow from a template or from scratch; while the second option allows you to load a workflow that you have saved on your pc in a JSON file.If you click on the “New WorkFlow” option, a tab with a list of various templates will be displayed (each template will have a different purpose). But the main one will be “Text2Image”; it will allow us to create images from text, similarly to the conventional method we always use. You can also create a workflow from scratch in the “Empty WorkFlow Template” option but for a better explanation of the basics we will use the “Text2Image”.Once you click on the "Text2Image" option, you must wait a few seconds and a new tab will be displayed with the template, which contains the basics to create an image by means of text. Nodes and Borders: ¿What are they and how do they work?Well, to understand the basics of how a WorkFlow works, it is necessary to have a clear understanding of what Nodes and Border are.Nodes are small boxes that are present in the workflow; each node will have a specific function necessary for the creation, enhancement or editing of the image or video. The basics of Text2Image are the CheckPoint loader, the Clip Text Encoders, the Empty Lantent Image, the Ksampler, the VAE decoder, and Save Image. It should be noted that there are hundreds of other nodes besides these basics and they all have many different functions.On the other hand, the “Borders” are the small colored wires that connect the different nodes. They are the ones that will set which nodes will be directly related. The Borders are ordered by colors that are generally related to a specific function.The purple is related to the Model or Lora used.The yellow one is intended for connection to the model or lora with the space to place the prompt.The red refers to VAE.The orange color refers to the connection between the spaces for placing the prompt and the “Ksampler” node.The fucsia color makes allusion to the latent, which will serve for many things; but for this case it serves to connect the “Empty Latent Image” node with the “Ksampler” node and establish the number and size of the images that will be generated.And the blue color is related to everything that has to do with images; it has many uses but this case is related to the “Save Image” node.What are the Text2Image template Nodes used for?Having this clear is of utmost relevance, since it allows you to know what each node of this basic template is for. It's like knowing what each piece in a lego set is for and understanding how they should be connected to create a beautiful masterpiece! Also, if you get to know what these nodes are for, it will be easier for you to intuit the functionality of its variants and other derived nodes.A) The first one is the node called “Load Chckpoint”, this node has three specific functions. The first one is to load the base model or checkpoint with which an image will be created. The second is the Clip, which will take care of connecting the positive and negative prompts that you write to the checkpoint. And the third is that it connects and helps to load the VAE model. B) The second one is the “Empty Latent Image”; which is the node in charge of processing the image dimensions from the latent space. It has two functions: First, set the width and length of the image; and second, set how many images will be generated simultaneously according to the “Batch Size” option.C) The third is the two “Clip Text Enconder” nodes: in this case there will always be at least two of these nodes, since they are responsible for setting both the positive and negative prompts that you write to describe the image you want. They are usually connected to the "Load Checkpoint" or any LoRa and are also connected to the “Ksampler” node.D) Then, there is a node “Ksampler”. This node is the central point of all WorkFlow; it is the one that sets the most important parameters in the creation of images. It has several functions: the first one is to determine which is the seed of the image and to regulate how much it changes from image to generated image by means of the “control_after_generate” option. The second function is to set how many steps are needed to create the image (you set them as you wish); the third function is to determine which sampling method is used and also what is the scheduler of this method (this helps to regulate how much space is eliminated when creating the image).E) The penultimate one is the VAE decoder. This node is in charge of assisting the processing of the image to be generated: its main function is to be responsible for materializing the written prompt into an image. That is to say, it reconstructs the description of the image we want as one of the final steps to finish the generation process. Then, the information is transmitted to the “Save Image” node to display the generated image as the final product.F) The last node to explain is the “Save Image”. This node has the simple function of saving the generated image and providing the user with a view of the final work that will later be stored in the taskbar where all the generated images are located.Final Consideration:This has been a small summary and explanation about very basic concepts about ComfyUI Mode; you could even say that it is like a small glossary about general terms. I have tried to give a small notion that tries to facilitate the understanding of this image generation tool. There is still a lot to explain, but I will try to cover all the topics; the information would not fit in a single article (ComfyUI is a whole universe of possibilities). ¡Thank you so much for taking the time to read this article!
44
15
Textual Inversion Embeddings  ComfyUI_Examples

Textual Inversion Embeddings ComfyUI_Examples

ComfyUI_examplesTextual Inversion Embeddings ExamplesHere is an example for how to use Textual Inversion/Embeddings.To use an embedding put the file in the models/embeddings folder then use it in your prompt like I used the SDA768.pt embedding in the previous picture.Note that you can omit the filename extension so these two are equivalent:embedding:SDA768.ptembedding:SDA768You can also set the strength of the embedding just like regular words in the prompt:(embedding:SDA768:1.2)Embeddings are basically custom words so where you put them in the text prompt matters.For example if you had an embedding of a cat:red embedding:catThis would likely give you a red cat.
13
1
Art Mediums (127 Style)

Art Mediums (127 Style)

Art MediumsVarious art mediums. Prompted with '{medium} art of a woman MetalpointMiniature PaintingMixed MediaMonotype PrintingMosaic Tile ArtMosaicNeonOil PaintOrigamiPapermakingPapier-mâchéPastelPen And InkPerformance ArtPhotographyPhotomontagePlasterPlastic ArtsPolymer ClayPrintmakingPuppetryPyrographyQuillingQuilt ArtRecycled ArtRelief PrintingResinReverse Glass PaintingSandScratchboard ArtScreen PrintingScrimshawSculpture WeldingSequin ArtSilk PaintingSilverpointSound ArtSpray PaintStained GlassStencilStoneTapestryTattoo ArtTemperaTerra-cottaTextile ArtVideo ArtVirtual Reality ArtWatercolorWaxWeavingWire SculptureWoodWoodcutGlassGlitch ArtGold LeafGouacheGraffitiGraphite PencilIceInk Wash PaintingInstallation ArtIntaglio PrintingInteractive MediaKinetic ArtKnittingLand ArtLeatherLenticular PrintingLight ProjectionLithographyMacrameMarbleMetalColored PencilComputer-generated Imagery (cgi)Conceptual ArtCopper EtchingCrochetDecoupageDigital MosaicDigital PaintingDigital SculptureDioramaEmbroideryEnamelEncaustic PaintingEnvironmental ArtEtchingFabricFeltingFiberFoam CarvingFound ObjectsFrescoAugmented Reality ArtBatikBeadworkBody PaintingBookbindingBronzeCalligraphyCast PaperCeramicsChalkCharcoalClayCollageCollagraphy3d PrintingAcrylic PaintAirbrushAlgorithmic ArtAnimationArt GlassAssemblage
52
5
Anime Vision | Detail Enhancer SD3

Anime Vision | Detail Enhancer SD3

SD3 Anime LoRA is Finally Here!I am thrilled to announce that the SD3 Anime LoRA model is finally available. In addition, I am releasing a new update that includes an SD3 anime checkpoint model.Currently, I am publishing a beta version as I continue to work diligently to perfect the model. I aim to have the final release ready by the end of this month or early August.Stay tuned, as the SD3 Anime beta version will be available within the next couple of days!Here are some guidelines to use this LoRA to its full potential:If you are trying to create any specific subject or object, use trigger word like 'anime style' in your prompt.If you're targeting a character, you can ignore the keyword and go with something like this:For a male character: 'anime boy'For a female character: 'anime girl'Simple, right? You can also use the trigger word 'anime style' most of the time. I've noticed it gives better results.ModelRecommended Parameter :LoRA Weight : 0🆙1VAE : No NeedSampler : DPM++ 2M SGM UniformSteps : 20➡30CFG : 3➡4Upscaler : R-ESRGAN 4x+If you encounter any issues, I recommend using ComfyUI for a better experience. Here's the workflow: ComfyUI Workflow. Open the link, select the LoRA model, choose the LoRA strength, and hit the run button.Join my community, Share your feedback, learn, and have fun with us! 😊Discord➡️https://discord.gg/QQKd7bu97P
29
1
How to set up Radio Button in your AI Tools

How to set up Radio Button in your AI Tools

Hello everyone! ✨ Today I will bring you a super practical tutorial: How to set up a convenient prompt word radio version for your AI Tools! 😎 Save it quickly, and you will never have to worry about how to set prompt words again! 👌Are you ready for the course? Let's get started! 🔍First, the first step is to open the official website of TensorArt. 📂 After opening, you will see a variety of AI tools and resources, which are very rich~ 👀Next, open comfyflow and start making our AI Tool! 🤖 This process is simple and fun, let's explore it together! ✨In comfyflow, we click the "New" button, which will take you to a new interface~ 🖱️💻In this interface, we can start creating our own workflow~ 🌟🎉 Next, we need to fill in the positive prompt words, which is a super critical step! 📝✨In the positive prompt word area, we need to enter the content we want. 📋 Here, the editor simply wrote an example for everyone: "a man". 🤵 This example is just for the convenience of teaching, you can freely play according to your needs~ 🌈🎆🎉 When you have completed the workflow, you can click the "Publish" button in the upper right corner! 🚀✨Don't forget to give your AI Tool an interesting name! 💡 This name will make your tool more attractive~✨ In addition, remember to divide the area correctly, so that you can see it clearly and it is also convenient for your friends to find and use it! 📂🔍🌟 Next, let's complete the next step together! 💪We pull down the current interface and find the user-configurable settings area. 👏 Then click the "Add" button. This step is very critical! 🖱️✨ Everyone must remember to add your positive prompt word node! 🔍✨After adding the node, our next step is to click the "Set" button on the right to proceed to the next step. 🔧✨ This step is crucial! Don't miss it! 😉🚀✨ The next step is also very important! 😊First, click the radio button, then click "Add". 🔘✨ Here, you can add the buttons you want to release to the user! 👍 After selecting, be sure to click "Confirm"! ✔️✨Friends, we have finally reached the last step! 🎉💪 This is an exciting moment! ✨When you have completed all the operations, remember to click the "Publish" button to publish your AI gadget! 🚀✨ Can't wait to see the results? Hurry up and generate a picture yourself to try and experience your results! 🌟🖼️Well, that's all for today's tutorial! 😊 I hope everyone can complete it successfully and create their own AI gadgets! 👏 If you have any questions, don't hesitate to leave a comment in the comment section at any time! ❤️
31
5
Guide to Using SDXL / SDXLモデルの利用手引

Guide to Using SDXL / SDXLモデルの利用手引

Guide to Using SDXLI occasionally see posts about difficulties in generating images successfully, so here is an introduction to the basic setup.1. IntroductionSDXL is a model that can generate images with higher accuracy compared to SD1.5. It produces high-quality representations of human bodies and structures, with fewer distortions and more realistic fine details, textures, and shadows.With SD1.5, generation parameters were generally applicable across different models, so there was no need for specific adjustments.However, while SDXL can still use some SD1.5 techniques without issues, the recommended generation parameters vary significantly depending on the model.Additionally, LoRA and Embeddings (such as EasyNegative) are completely incompatible, requiring a review of prompt construction.Notably, embeddings commonly used in SD1.5 negative prompts are recognized merely as strings in the XL model, so you must replace them with corresponding embeddings or add appropriate tags.This guide explains the recommended parameter settings for using SDXL.2. Basic ParametersVAESelecting "sdxl-vae-fp16-fix.safetensors" will suffice.Many models have this built-in, so specification might not be necessary.Image SizeUsing the presets provided by TensorArt for resolution should be sufficient.Small or excessively large resolutions may not yield appropriate generation results, so please avoid using the sizes that were frequently used with SD1.5 wherever possible.Even if you want to create vertically or horizontally elongated images, do so within the range that does not significantly alter the total pixel count (adjust by increasing height and decreasing width, for example).Sampling MethodChoose the sampler recommended for the model first.Then, select according to your preference.Typically, selecting Euler a or DPM++ 2M SDE Karras should work well.Sampling StepsXL models might generate images effectively with lower steps due to optimizations like LCM or Turbo.Be sure to check the recommended values for the selected model.CFG ScaleThis varies by model, so check the recommended values.Typically, the range is around 2 to 8.Hires.fixFor free users, specifying 1.5x might hit the upper limit, so use custom settings with the following resolutions:768x1152 -> 1024x15361152x768 -> 1536x10241024x1024 -> 1248x1248Choose the upscaler according to your preference.Set the denoising strength to around 0.3 to 0.4.3. PromptSDXL handles natural language better.You can input elements separated by commas or simply write a complete sentence in English, and it will generate images as intended.Using a tool like ChatGPT to create prompts can also be beneficial.However, depending on how the model was additionally trained, it might be better to use existing tags.Furthermore, some models have tags specified to enhance quality, so always check the model’s page.For example:AnimagineXL3.1: masterpiece, best quality, very aesthetic, absurdres is recommended.Pony Models: score_9, score_8_up, score_7_up, score_6_up, score_5_up, score_4_up is recommended.ToxicEchoXL: masterpiece, best quality, aesthetic is recommended.In this way, especially for XL models, particularly anime or illustration models, appropriate tag usage is crucial.4. Negative PromptsForget the negative prompts used in SD1.5. "EasyNegative" is just a string.The embeddings usable on TensorArt are negativeXL_D and unaestheticXLv13.Choose according to your preference.Some models have recommended prompts listed.For AnimagineXLnsfw, lowres, (bad), text, error, fewer, extra, missing, worst quality, jpeg artifacts, low quality, watermark, unfinished, displeasing, oldest, early, chromatic aberration, signature, extra digits, artistic error, username, scan, [abstract]For ToxicEchoXLnsfw, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digits, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry, artist name.For photo models, sometimes it is better not to use negative prompts to create a certain atmosphere, so try various approaches.5. Recommended SDXL modelToxicEnvisionXLhttps://tensor.art/models/736585744778443103/ToxicEnvisionXL-v1Recently released high-quality photo model. Yes, I created it.If you are looking for a photo model, you can't go wrong with this one.Check the related posts to see what kind of images can be created.You can create a variety of realistic images, from analog photo styles to gravure, movies, fantasy, and surreal depictions.Although it is primarily a photo-based model, it can also create analog-style images.ToxicEtheRealXLhttps://tensor.art/models/702813703965453448/ToxicEtheRealXL-v1A versatile model that supports both illustrations and photorealistic images. Yes, I created it.The model's flexibility requires well-crafted prompts to determine whether the output is more illustrative or photorealistic.Using LoRA to strengthen the direction might make it easier to use.ToxicEchoXLhttps://tensor.art/models/689378702666043553/ToxicEchoXL-v1A high-performance model specialized for illustrations. Yes, I created it.It features a unique style based on watercolor painting, with custom learning and adjustments.I have also created various LoRA for style changes, so please visit my user page.My current favorite is Beautiful Warrior XL + atmosphere.The model covers a range from illustrations to photos, so give it a try.However, it is weak in generating copyrighted characters, so use LoRA or models like AnimagineXL or Pony for those.ToxicEchoXL can produce unique illustration styles when using character LoRA, making it highly suitable for fan art.6. ConclusionI hope this guide helps those who struggle to generate images as well as others.Well... if you remix from Model Showcase, you can create beautiful images without this guide...SD3 has also been released, so if possible, I would like to create models for that as well.It seems that a commercial license is required for commercial use, though...SDXLモデルの利用手引ここではSDXLの基本的な設定を紹介します。1. はじめにSDXLはSD1.5と比較してより高精度な生成が行えるモデルです。人体や構造物はより高品質で破綻が少なく、微細なディテールがよりリアルに表現され、自然なテクスチャや影を描写します。SD1.5ではどのモデルでも生成パラメータは概ね流用可能で、特に気にする必要はありませんでした。SDXLは一部SD1.5の手法を利用しても問題ありませんが、推奨される生成パラメータがモデルによってもだいぶ変わります。またLoRAやEmbeddings(EasyNegativeなど)も一切互換性はありませんので、プロンプトの構築も見直す必要があります。特にSD1.5のネガティブプロンプトでよく使用されているEmbeddingsをそのままXLモデルで入力しても、ただの文字列としてしか認識されていませんので、対応するEmbeddingsに差し替えるか、適切なタグを追加しなければいけません。このガイドでは、SDXLを使用する際の推奨パラメータ設定について説明します。2. 基本的なパラメータVAEsdxl-vae-fp16-fix.safetensorsを選択しておけば問題ありません。モデルに内蔵されている場合も多いですので、指定しなくても大丈夫な場合もあります。画像サイズ解像度はTensorArtで用意されているプリセットを使えば問題ありません。小さかったり大きすぎる解像度は適切な生成結果を得られなくなりますので、SD1.5でよく使用していたサイズはなるべく使用しないでください。プリセットよりも縦長や横長にしたい場合でも、総ピクセル数を大幅に変更しない範囲で行ってください。(縦を増やしたら横は減らす等で調整)サンプリング法モデルによって推奨されるサンプラーがありますので、まずはそれを選択してください。あとはお好みです。基本は Euler a か DPM++ 2M SDE Karras あたりを選択しておけば大丈夫です。サンプリング回数XLではLCMやターボなど低ステップで生成できるようになっていたりしますので、必ずモデルの推奨値を確認してください。CFG Scaleこれもモデルによって異なりますので推奨値を確認してください。概ね2~8程度です。高解像度修復無料ユーザーだと1.5xを指定すると上限に引っかかってしまいますので、使用する場合はカスタムにして以下の解像度を指定してください768x1152 -> 1024x15361152x768 -> 1536x10241024x1024 -> 1248x1248Upscalerはお好みで指定してください。Denoising strengthは0.3~0.4程度。3. プロンプトSDXLはより自然言語の取り扱いに長けています。要素をコンマで区切って入力するだけではなく、普通に英文を入力するだけでも意図した通りの生成が行えます。ChatGPTなどにプロンプトを作ってもらうのもいいでしょう。ただしモデルが追加学習をどのように行ったかによって、既存のタグで記述したほうがいい場合もあります。また、モデルによっては品質を上げるためのタグが指定されていますので、使用するモデルのページは必ず見るようにしましょう。例えば…AnimagineXL3.1では「masterpiece, best quality, very aesthetic, absurdres」を指定することが推奨されています。Pony系モデルでは「score_9, score_8_up, score_7_up, score_6_up, score_5_up, score_4_up」が基本テンプレートとなっています。ToxiEchoXLでは「masterpiece, best quality, aesthetic」を指定することが推奨されています。このように、XLモデル、特にアニメ・イラストモデルでは適切なタグの使用が求められる場合があります。4. ネガティブプロンプトSD1.5で使用していたネガティブプロンプトは忘れてください。EasyNegativeはただの文字列です。TensorArtで使用できるEmbeddingsは negativeXL_D と unaestheticXLv13 です。お好みで指定してください。推奨されるプロンプトが記載されているモデルもあります。AnimagineXLでは以下のようなプロンプトが推奨されていますので、これをベースに組むのがいいかもしれません。nsfw, lowres, (bad), text, error, fewer, extra, missing, worst quality, jpeg artifacts, low quality, watermark, unfinished, displeasing, oldest, early, chromatic aberration, signature, extra digits, artistic error, username, scan, [abstract]ToxicEchoXLでは以下のようなプロンプトが推奨されていますnsfw, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digits, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry, artist name,フォトモデルではネガティブプロンプト無しのほうが雰囲気のある画作りができる場合もありますので、色々試してみてください。5. おすすめのSDLXモデル紹介ToxicEnvisionXLhttps://tensor.art/models/736585744778443103/ToxicEnvisionXL-v1最近リリースされた高品質フォトモデル。実写系モデルを探しているならこれを選んでおけば間違いありません。関連する投稿からどういった画像が作成できるか見てみてください。アナログ写真風からグラビア、映画、ファンタジー、非現実的な描写等、様々な実写的な画像が作成できます。基本的にはフォトベースのモデルですが、アナログ画風も作成できたりします。ToxicEtheRealXLhttps://tensor.art/models/702813703965453448/ToxicEtheRealXL-v1イラストからフォトリアルまで幅広く対応したモデル。プロンプトによってイラストかフォトリアルか振れ幅が大きいので、明確にプロンプトの作り込みが必要です。LoRAで方向性を強めると使いやすいかもしれません。ToxicEchoXLhttps://tensor.art/models/689378702666043553/ToxicEchoXL-v1イラスト特化の超高性能モデル。水彩をベースに独自の学習・調整を行っているので、わりと独特な画風を持っています。画風変更に様々なLoRAも作成していますので、是非私のユーザーページへお越しください。https://tensor.art/u/649265516304702656最近のお気に入りはBeautiful Warrior XL + atmosphere です。イラストからフォトまで一通り網羅できるので、是非使ってみてください。なお版権キャラの生成は弱いので、その辺はLoRAかAnimagineXLとかPonyとか使うといいと思います。ToxicEchoXLはキャラLoRAを使うと他のモデルとはタッチの違うイラストが作れますので、ファンアート適正自体は高いです。6. おわりにモデルのサンプルやみんなみたいにうまく生成できないな…という方の助けになれば幸いです。まあ…モデルのショーケースからリミックスすればこんなガイド見なくてもきれいな画像が作れますけどね…SD3もリリースされたので、もし可能ならそちらのモデルも作成してみたいですね。どうも商用利用は有償のライセンスが必要そうですが…
45
Understanding the Impact of Negative Prompts: When and How Do They Take Effect?

Understanding the Impact of Negative Prompts: When and How Do They Take Effect?

📝 - SynthicalThe Dynamics of Negative Prompts in AI: A Comprehensive Study by: Yuanhao Ban UCLA, Ruochen Wang UCLA, Tianyi Zhou UMD, Minhao Cheng PSU, Boqing Gong, Cho-Jui Hsieh UCLAEThis study addresses the gap in understanding the impact of negative prompts in AI diffusion models. By focusing on the dynamics of diffusion steps, the research aims to answer the question: "When and how do negative prompts take effect?". The investigation categorizes the mechanism of negative prompts into two primary tasks: noun-based removal and adjective-based alteration.The role of prompts in AI diffusion models is crucial for guiding the generation process. Negative prompts, which instruct the model to avoid generating certain features, have been less studied compared to their positive counterparts. This study provides a detailed analysis of negative prompts, identifying the critical steps at which they begin to influence the image generation process.FindingsCritical Steps for Negative PromptsNoun-Based Removal: The influence of noun-based negative prompts peaks at the 5th diffusion step. At this critical step, negative prompts initially generate a target object at a specific location within the image. This neutralizes the positive noise through a subtractive process, effectively erasing the object. However, introducing a negative prompt in the early stages paradoxically results in the generation of the specified object. Therefore, the optimal timing for introducing these prompts is after the critical step.Adjective-Based Alteration: The influence of adjective-based negative prompts peaks around the 10th diffusion step. During the initial stages, the absence of the object leads to a subdued response. Between the 5th and 10th steps, as the object becomes clearer, the negative prompt accurately focuses on the intended area and maintains its influence.Cross-Attention DynamicsAt the peak around the 5th step for noun-based prompts, the negative prompt attempts to generate objects in the middle of the image, regardless of the positive prompt's context. As this process approaches its peak, the negative prompt begins to assimilate layout cues from its positive counterpart, trying to remove the object. This represents the zenith of its influence.For adjective-based prompts, during the peak around the 10th step, the negative prompt maintains its influence on the intended area, accurately targeting the object as it becomes clear.The study highlights the paradoxical effect of introducing negative prompts in the early stages of diffusion, leading to the unintended generation of the specified object. This finding suggests that the timing of negative prompt introduction is crucial for achieving the desired outcome.Reverse Activation PhenomenonA significant phenomenon observed in the study is Reverse Activation. This occurs when a negative prompt, introduced early in the diffusion process, unexpectedly leads to the generation of the specified object within the context of that negative prompt. To explain this, researchers borrowed the concept of the energy function from Energy-Based Models to represent data distribution.Real-world distributions often feature elements like clear blue skies or uniform backgrounds, alongside distinct objects such as the Eiffel Tower. These elements typically possess low energy scores, making the model inclined to generate them. The energy function is designed to assign lower energy levels to more 'likely' or 'natural' images according to the model’s training data, and higher energy levels to less likely ones.A positive difference indicates that the presence of the negative prompt effectively induces the inclusion of this component in the positive noise. The presence of a negative prompt promotes the formation of the object within the positive noise. Without the negative prompt, implicit guidance is insufficient to generate the intended object. The application of a negative prompt intensifies the distribution guidance towards the object, preventing it from materializing.As a result, negative prompts typically do not attend to the correct place until step 5, well after the application of positive prompts. The use of negative prompts in the initial steps can significantly skew the diffusion process, potentially altering the background.ConclusionsDo not step less than 10th times, going beyond 25th times does not make the difference for negative prompting.Negative prompts could enhance your positive prompts, depending on how well the model and LoRA have learn their keywords, so they could be understood as an extension of their counterparts.Weighting-up negative keywords may cause reverse activation, breaking up your image, try keeping the ratio influence of all your LoRAs and models equals.Referencehttps://synthical.com/article/Understanding-the-Impact-of-Negative-Prompts%3A-When-and-How-Do-They-Take-Effect%3F-171ebba1-5ca7-410e-8cf9-c8b8c98d37b6?
22
2
[ 🔥🔥🔥 SD3 MEDIUM OPEN DOWNLOAD - 2024.06.12 🔥🔥🔥]

[ 🔥🔥🔥 SD3 MEDIUM OPEN DOWNLOAD - 2024.06.12 🔥🔥🔥]

Finally! It's happening! The Medium version will be released first!+Stability.AICo-CEO Christian Laporte has announced the release of the weights.Stable Diffusion 3 Medium, our most advanced text-to-image model, will soon be available! You can download the weights from Hugging Face starting Wednesday, June 12.SD3 Medium is the SD3 model with 2 billion parameters, designed to excel in areas where previous models struggled. Key features include:• Photorealism: Overcomes common artifacts in hands and faces to deliver high-quality images without complex workflows.• Typography: Provides powerful typography results that surpass the latest large models.• Performance: Optimized size and efficiency make it ideal for both consumer systems and enterprise workloads.• Fine-Tuning: Can absorb fine details from small datasets, perfect for customization and creativity.SD3 Medium weights and code are available for non-commercial use only. If you wish to discuss a self-hosting license for commercial use of Stable Diffusion 3, please fill out the form below and our team will contact you shortly.+ @everyone
41
4
What exactly are the "node" and the "workflow" in AI image platform (explanation for the beginner)

What exactly are the "node" and the "workflow" in AI image platform (explanation for the beginner)

The Traditional Way of Generating AI Images for the BeginnerIf you are a beginner in the AI community, maybe you will be very confused and have no clue about what is "Node", and "Workflow" and their relations with "AI Tools" in the TensorArtTo start with the most simple way. We need to first mention how the user generates an image using a "Remixing" button that brings us to the "Normal Creation menu"Needless to say, by just editing the prompt (what you would like to see your picture look like) and negative prompt (what you do not want to see in the output image). Then push the Generate button, and the wonderful AI tool will kindly draw the new illustration serving you within a minute!!!!That sounds great, don't you think? If we imagine how humans spent a huge amount of time in the past to publish just 1 single piece of art. (Yeah, today, in 2024, in my personal opinion, both AI and human abilities are still not fully replaceable, especially in the terms of beautiful perfect hand :P ) However, the backbone or what happens behind the User-friendly menu allows us to "Select model", "Add LoRA", "Add ControlNet", "Set the aspect ratio (the original size of the image)" and so on, all of them are collected "Node" in a very complex "Workflow" PS.1. The Checkpoint or The Model often refers to the same thing. They are the core program that had been trained to draw the illustration. Each one has its strengths and weaknesses (I.E. Anime oriented or Realistic oriented) PS.2. The LoRA (Low-Rank Adaptation) is like an add-on to the Model allowing it to adapt to a different style, theme, and user preference. A concrete example is the Anime Character LoRAPS.3 The ControlNet is like a condition setting of the image. It helps the model to truly understand what is beyond the text prompt can describe. For instance, how a character poses in each direction and the angle of the camera.So here comes "The Comfyflow" (the nickname of the Workflow, people also mentioned it by the name "ComfyUI") which gives me a super headache when I see things like this for the first time in my life!!!!!!!!!(This image is a flow I have spent a lot of time studying, it is a flow for combining what is in the two images into a single one) Yeah, maybe, it is my fault that did not go to class about the workflow from the beginning or search for the tutorial on YouTube the first time (as my first language is not English). But would it be better if we had an instructor to tell us step-by-step here in Tensor.ArtAnd that is the reason why I got inspired to write this article solely for the beginner. So let's start with the main content of the article.What is ComfyFlowComfyFlow or the Workflow is an innovative AI image-generating platform that allows users to create stunning visuals with ease. To get the most out of this tool, it's important to understand two key concepts: "workflow" and "node." Let's break these down in the simplest way possible.What is a Workflow?A workflow is like a blueprint or a recipe that guides the creation of an image. Just as a recipe outlines the steps to make a dish, a workflow outlines the steps and processes needed to generate an image. It’s a sequence of actions that the AI follows to produce the final output.Think of it like this:Recipe (Workflow): Tells you what ingredients to use and in what order.Ingredients (Nodes): Each step or component used in the recipe.Despite the recommended pre-set template that TensorArt kindly gives to the users, from the beginner view's viewpoint without the knowledge of the workflow, it is not that helpful because, after clicking the "Try" button, we will bombarded with the complexity of the Node!!!!!!!What is a Node?Nodes are the building blocks of a workflow. Each node represents a specific action or process that contributes to the final image. In ComfyFlow, nodes can be thought of as individual steps in the workflow, each performing a distinct function.Imagine nodes as parts of a puzzle:Nodes: Individual pieces that fit together to complete the picture (workflow).How Do Workflows and Nodes Work Together? 1-2) Starting Point: Every workflow begins with an initial node, which might be an image input from the user, together with Checkpoint and LoRA serving the role of image references. 3-4) Processing Nodes: These are nodes that draw or modify the image in some way, such as adding color, or texture, or applying filters. 5) Ending Point: The node outputs the completed image which works very closely with the node of the previous stage in terms of sampling and VAE PS. A Variational Autoencoder (VAE) is a generative model that learns input data, such as images, to reconstruct and generate new, similar, or variations of images based on the patterns it has learned.Here is the list of nodes I have used in the normal image-generating images of my Waifu using 1checkpoint, and 2LoRAs to help the reader understand how ComfyFlow worksThe numbers 1-5 represent the overview process of the workflow and the role of each type of node I have mentioned above. However, in the case of more complex tasks like in AI Tools, the number of nodes sometimes is higher than 30!!!!!!!By the way, when starting with an empty ComfyFlow page, the way to add a node is "Right Click" -> "Add Node" -> Scroll down to the top, since the most frequently used node will be over there.1) loaders -> Load CheckPointLike in the normal task creation menu, this node is the one we can choose CheckPoint or the Core model.It is important to note that nodes work together using input/output. The "Model/CLIP/VAE" (the output) circles have to connect to the next one in which it has to correspond. We link them together by left-clicking on the circle's inner area and then drag to the destination. PS. CLIP (Contrastive Language-Image Pre-training) is a model developed by OpenAI that links images and text together in a way that helps AI understand and generate images based on textual descriptions.2) loaders -> Load LoRACheckpoint is very closely related to LoRA and that is a reason why they are connected by the input/output named "model/MODEL", "clip/CLIP"Anyway, since in this example, I have used 2 LoRAs (first for The theme of the picture and the Second for the character reference of my Waifu), two nodes of LoRAs then have to be connected as well. Here we can adjust the strength of the LoRA or the weight like it happens in the normal task generation menu.3) CLIP Text Encode (Prompt)This node is the prompt and negative prompt we normally see in the menu. The input here is only clip (Contrastive Language-Image Pre-training) and the output is "CONDITIONING" User tip: If you click on the output circle of the "Load LoRA" node and drag it to the empty area, the ComfyFlow will pop up a corresponding next node list to create a new one with ease. 4) KSampler & Empty Latent ImageThe sampling method is used to tell the AI how it should start generating visual patterns from the initial noise and everything associated with its adjustment will be set here in this type of sampling node together with "Empty Latent Image" The inputs in this step here are models (from LoRA node), positive and negative (from prompt node) and the output is "Latent"5) VAE Decode & Final output nodeOnce we establish the sampling node, the output named "LATENT" will then have to connect with "samples" Meanwhile the "vae" is the linkage between this one and the "Load Checkpoint" node from the beginning.And when everything is done the "IMAGE" as a final output here will be served at your hand.PS. An AI Tool is a more complex Workflow created to do some specific task such as swapping the face of the human in the original picture with the target face or changing the style of the input illustration to another one and etc.
55
5
PhotoReal Makeup Edition - V3 Slider

PhotoReal Makeup Edition - V3 Slider

PhotoReal Makeup Edition - V3 Slider (no trigger)Introducing the PhotoReal Makeup Edition - V3 Slider! Slide to the right to add beautiful, realistic makeup. Slide to the left to reduce the makeup effect for a more natural look. It's perfect for adjusting the makeup to get just the style you want.Try it out and see the amazing changes you can make!More Information:- Model linkYour feedback is invaluable to me. Feel free to share your experiences and suggestions in the comment section. For more personal interactions, join our Discord server where we can discuss and learn together.Thank you for your continued support!
65
8

Tips for new Users

Intro Hey there! If you're reading this, you're probably new to AI image generation and want to learn more. If you're not, you probably already know more than me :). Yeah, full disclosure: I'm still pretty inexperienced at this whole thing, but I thought I could still share some of the things I've learned with you! So, in no particular order:1. You can like your own posts I doubt there's anyone who doesn't know this already, but if you're posting your favorite generations and you care about getting likes, you can always like them yourself. Sketchy? Kinda. Do I still do it? Yes. And on the topic of getting more likes:2. Likes will often be returned Whenever I receive a like on one of my posts, I'll look at that person's pictures and heart any that I particularly enjoy. I know a lot of people do this, so one of the best ways to get people to notice and like your content is to just browse through posts and be generous with your own likes. It's a great way to get inspiration too!3. Use turbo/lightning LORAs If you find yourself running out of credits, there are ways to conserve them. When I'm iterating on an idea, I'll use a SDXL model (Meina XL) paired with this LORA. This lets me get high quality images in 10 steps for only 0.4 credits! It's really nice, and works with any SDXL model. Unfortunately, if there is a similar method for speeding up SD 1.5 models I don't know it, so it only works with XL.4. Use ADetailer smartly ADetailer is the best solution I've found for improving faces and hands. It's also a little difficult to figure out. So, though I'm still not a professional with it, I thought I could share some of the tricks I've learned. The models I normally use are face_yolo8s.pt and hand_yolo8s.pt. The "8s" versions are better than the "8n" versions, though they are slightly slower. In addition to these models, I'll often add the Attractive Eyes and Perfect Hand LORAs respectively. These are all just little things you can do to improve these notoriously hard parts of image generation. Also, using ADetailer before upscaling the image is cheaper in terms of credits, though the upscaling process can sometimes mess up the hands and face a little bit so there's some give and take there.5. Use an image editing app Wait a minute, I hear you saying, isn't this a guide for using Tensor Art? Yes, but you can still use other tools to improve your images. If I don't like a specific part of my image, I'll download it, open it in Krita (Or Photoshop or Gimp) and work on it. My art skills are pretty bad, (which is why I'm using this site in the first place,) but I can still remove, recolor, or edit certain aspects of the image. I can then reupload it to Tensor Art, and Img2img with a high denoising strength to improve it further. You could also just try inpainting the specific thing you want to change, but I always find it a bit of a struggle to get inpaint to make the changes I want.6. Experiment! The best way to learn is to do, so just start generating images, fiddling with settings, and trying new things. I still feel like I'm learning new stuff every day, and this technology is improving so fast that I don't think anyone will ever truly master it. But we can still try our hardest and hone our skills through experimentation, sharing knowledge, and getting more familiar with these models. And all the anime girls are a big plus too.Outro If you have anything to add, or even a tip you'd like to share, definitely leave a comment and maybe I can add it to this article. This list is obviously not exhaustive, and I'm no where near as talented as some of the people on this platform. Still though, I hope to have helped at least one person today. If that was you, maybe give the article a like? I appreciate it a ton, so if you enjoyed, just let me know. Thanks for reading!
86
9
• MOOD MAGIC SERIES • I. Melancholy

• MOOD MAGIC SERIES • I. Melancholy

MOOD MAGIC: adding emotion to your promptsMelancholy & GloomOvercast: Cloud-covered skies for subdued lighting.Dim Lighting: Limited light sources for creating deep shadows.Muted Colors: Toned-down color palette to convey sadness or desolation.Dusky: Twilight ambiance, suggesting the fading light of day.Foggy: A thick mist that obscures details and softens the scene.Drizzly: Gentle rain that adds a reflective, melancholic quality.Cloudy: Thick clouds that reduce brightness and saturate the scene with grey.Desaturated: Low color saturation to enhance the bleak feel.Shadowed: Prominent shadows that deepen the mood.Moody Lighting: Emotionally charged lighting with strong contrasts.Gloomy: Overall dark and dismal atmosphere.Monochrome: Black and white or single-color dominance to strip away cheer.Underexposed: Darker exposure to mimic a sense of foreboding.Chiaroscuro: Strong contrasts between light and dark, emphasizing turmoil.Hazy: Blurred or smoky atmosphere, creating a sense of mystery or unease.Twilight: Dim natural lighting that can feel lonely or isolating.Stormy: Implication of an approaching or ongoing storm to add tension.Wintery: Cold, barren landscape cues, even in urban settings.Grainy: Visual noise that adds an old or troubled quality.Bleak: Stark, harsh lighting or barren scenery settings.Ominous Clouds: Dark, menacing clouds that threaten bad weather.Subdued Tones: Soft, low-key colors that don't catch the eye.Cold Colors: Blues and greys to suggest chilliness and discomfort.Rusty: Implications of decay and neglect.Aged: A sense of time wearing down the scene, historical weariness.Soft Focus: Slightly out-of-focus elements to create a sense of disorientation or confusion.Tenebrous: Deeply shadowed, almost pitch-dark.Low-Key Lighting: Minimal lighting mostly in darkness with occasional highlights.Pensive: Engaged in, involving, or reflecting deep or serious thought.Yearning: A feeling of intense longing for something typically something that one has lost or been separated from.Weary: Conveying a sense of tiredness or exhaustion, both physical and emotional.Sparse: Minimalist or bare settings that suggest simplicity or emptiness.Brooding: A deep, serious, and sometimes dark contemplation.Silent: Lack of sound or motion, emphasizing solitude or contemplation.Ephemeral: Fleeting or transitory, suggesting the transient nature of moments and emotions.Desolate: Emptiness that conveys a sense of abandonment or loneliness.Poetic: Imbued with a sense of beauty and melancholy, often through lyrical expression.Moody Skies: Cloudy, stormy, or unsettled skies that reflect a turbulent emotional landscape.Cold Light: Harsh, unyielding light that doesn’t warm but isolates subjects.Autumnal: Related to autumn, often seen as a melancholic season due to its association with the end of summer.Faded: Colors or elements that have lost brightness, suggesting the passing of time.Blue Hour: Moody cool natural lighting obtained in the twilight hour just after sunset or just before sunrise.Example using Stable Diffusion SDXL + refinerCheckpoint: RealVis4Cfg: 5.5Steps: 40Sampler: DPM++ 3m SDE KarrasVisualize a close-up portrait of a young woman standing by a foggy window, her gaze distant and contemplative. The room is dimly lit, with only a soft, diffuse light filtering through the heavy overcast outside, casting subtle shadows across her face. The colors are desaturated, emphasizing a palette of cool grays and muted blues that reflect her somber mood. Her expression is serene yet melancholic, with her eyes slightly downcast as if lost in thought. The background is blurred, enhancing the sense of isolation and introspection. This portrait captures the essence of melancholy, framed in a moment of quiet solitude.negative: illustration, cartoon, anime, 3d, digital art, bad quality, CGI, sketch, drawn, blurry, painting, worst quality, low quality, bad anatomy, bad hands, bad body, missing fingers, extra digit, fewer digits
2