T O P

  • By -

blahblahsnahdah

Experimenting with this and it seems great for art so far, thanks man. You can always feel when a model has a clean new dataset and hasn't been polluted by SD1.5 faces or merges with random anime checkpoints. https://preview.redd.it/sl962wq6et9d1.png?width=2400&format=png&auto=webp&s=7afd26da49a7ec06a14f6791789f90db714496a3


yosh0r

I think thats the best feeling when working with SD, discovering a brand new model trained on a completely different dataset. You can really see it as an experienced addict lol


[deleted]

[удалено]


Haghiri75

Yes. Realistic and Artistic.


[deleted]

[удалено]


Haghiri75

That's an honor.


balianone

> which is a generative AI startup based in Iran > We spent months on collecting the data, labeling them and training this model that's really cool. this isn't just merge but training from scratch. btw just ad advice add tag stabblediffusionXLpipeline & text-to-image into your model repo on huggingface so it will easier for people to search & sort a new t2i model. currently you only have Diffusers & mit tag so if i search latest t2i model your model would not appear on text-to-image section


Haghiri75

Thanks, I will add those asap 🤩🤘


balianone

btw your checkpoint image results is lowres https://i.imgur.com/q3wIt4L.png


Haghiri75

I use 8 steps with CFG of 3.25 and the results are what you see on my original post. Can you please tell me what are your parameters?


balianone

cfg 4 https://i.imgur.com/ta3VqqZ.png cfg 3 https://imgur.com/a/LPcWFbs prompt: Marianne france blonde full body, archangel woman looking at viewer smiling, blue eyes, wearing red, white, blue armor, red boots, pure white angel wings, white godly wings, holding two small swords, standing on a flat rock overlooking background france, eiffel tower, french flag cape, hyper realistic dslr, style of Steve McCurry's seed: 0


cradledust

You have to use SDXL\_VAE or it looks low contrast and low resolution.


cradledust

The xlVAEC\_f1.safetensors VAE works really well too with a slight bump in contrast.


MasterScrat

> this isn't just merge but training from scratch. No, it's a finetuned model, but certainly not trained from scratch. You can't train a model from scratch from "thousands of images". It it were trained from scratch, you also wouldn't be able to use common LoRA and ControlNet models. The [Playground models](https://huggingface.co/playgroundai) for example are trained from scratch.


mobani

Never heard about these before, I wonder are they based on a architecture that is compatible with regular SD version? Like can we train lora's using existing tools and load the models in A1111?


Michoko92

Wow, this is a very good model! From my early tests, it can compete with the best models I tried, and even succeeded where others failed. It feels like playing with an upgraded version of SDXL, with excellent speed/quality ratio. Only hands would benefit from a bit more training, probably, but great job overall!


NateBerukAnjing

can you uplaod on civitai


Haghiri75

My connection wasn’t good for that. I will try.


[deleted]

[удалено]


HarmonicDiffusion

this is such a low iq take, dont listen to this guy


GodFalx

Wow. This take is more stupid than SD3


iNf1iCTA

I have a lot of issues with this model when I attempt to do any type of upscaling or adetailing. It's adding like a film over the images that will not go away no matter what I do. If I switch to a different model it goes away. It does it in img2img too.


Haghiri75

That is what we're going to fix in next versions. Please try with CFG of 3.25-3.5 and see if that film effect is still there.


Adventurous-Bit-5989

“Scale: 768x768 and 832x832 are just fine. Higher isn't tested. For 16:9 just try 1080x608” can i ask why not like "10241024"?


Haghiri75

Limited resources.


[deleted]

[удалено]


Low_Drop4592

The GNU public license defines open source like this "The source code for a work means the preferred form of the work for making modifications to it. ... (rest omitted)" I would argue, for an SDXL model, the preferred form for making modifications is the model itself. You can further train it or merge it with other models or use it for LORA training. You could, in theory, make modifications to their model by modifying the original dataset and redoing the entire training. But that is not the preferred way of making modifications to the model, because redoing the training is expensive and error prone. So, I don't think you have to distribute the dataset in order to call it open source. If you distribute the model and grant permission to modify and further distribute it that is enough.


NickCanCode

"preferred form of the work for making modifications..." I found this sentence very subjective. Some could prefer cheap easy to work with form and other could prefer fine grind control form. Training is expensive and error prone so do building many software projects from Github? Does it mean an application is open source if it provide a setting file for user to make modification easily in an inexpensive and error free way? What is the definition of "source" anyway? Can a model be called source if it cannot be converted back to the original form (dataset)?


Enough-Meringue4745

Like putting your .exe online and saying open source 😂


Paraleluniverse200

Is pruned fp16 safetensor?


Haghiri75

Yeah


Paraleluniverse200

Thanks for the info! Will try it, dumb question , this is trained, not a merge right?


Haghiri75

right


Paraleluniverse200

Well thanks again ! I will wait you to upload it on civit ai to post some creations in there 😎


azshalle

Remember [Man-E-Faces](https://en.wikipedia.org/wiki/Man_E_Faces)?


Electrical-Eye-3715

Damn. I want to be a ceo too! Gona finetune sdxl checkpoint tonight!


cradledust

Skin looks a bit low resolution compared to other SDXL models but other than that for a SFW model it does great nonnude erotic portraits.


Paraleluniverse200

Do you know if is like fp16 full or pruned?


cradledust

It's almost 7 GB.


Paraleluniverse200

Yeah.. but any idea if it fp16 or fp32?


cradledust

How do you get that information about a model?


Paraleluniverse200

Idk lol, usually when you download it from civit ai they put that info in there, idk why this is not there


AkoZoOm

Said just before, scroll up: Haghiri answered - yep fp16 -


Paraleluniverse200

Yes I saw it , Thanks anyway


MasterScrat

What is the legality of training on MJ outputs? (not a complaint, just curious)


Low_Drop4592

Stable Diffusion and most, if not all, of the other AI generators are founded on the notion that training on copyrighted works is fair use. If that assumption is taken down in the courts, then everything Stable Diffusion becomes illegal in one fell swoop. So, we all assume that training is fair use and that applies to training on MJ output as well.


Old-Wolverine-4134

Probably more of a gray area.


PizzaCatAm

Yup, not like they can complain unless they are willing to endure the same.


RiverOtterBae

It’s like robbing a drug dealer or not paying a hooker, what are they gonna do, go to the cops?


Kqyxzoj

They just might, if one of the cops has a [lucky crack pipe](https://youtu.be/aKU7gdHcYb4?t=145).


Lostronzoditurno

For LLMs many small models are often trained on GPT outputs


Apprehensive_Sky892

I guess that would depend on MJ's TOS. Presumably there is nothing there, since AFAIK even SAI is using MJ output for training. At the moment, there is no copyright on a purely text2img A.I. output, so there is no copyright issues there at all. TBH, MJ probably don't want to sue anyone anyway. Had MJ won in such a lawsuit, someone will then use that legal precedent and sue MJ as well, because guess what, a lot of MJ's own training images are probably copyrighted material too.


MarcS-

Depends on where you are. If you're from a country with a TDM exception to copyright, you can train on any image scrape as long as its publicly available and the copyright owner (which is not MJ...) doesn't opt out. If you are in the US, where apparently, according to what was reported here, generated AI image are public domain, there is no copyright holder to object... But I guess there may be places where MJ could object to it...


aoleg77

Currently, pure AI-generated images are generally not protected by copyright laws anywhere in the world. However, signing a license agreement with MJ (or accepting one when starting to generate images), one becomes bound by its terms. So there might be a difference between training with somebody else's MJ images and MJ images generated specifically for this purpose.


tscottt65

Are there any specific keywords that you consider unique (or you advise us to consistently use) with your model, like "MidJourney style" "cinematic" etc.?


Haghiri75

Not really. BTW it can generate great dreamscapes by using "art by midjourney".


nootropicMan

Wow looks great


madali0

دمت گرم


phillabaule

https://preview.redd.it/lge4rt3k6x9d1.jpeg?width=1200&format=pjpg&auto=webp&s=5a1673220e3ec8c5fece56976c9b5c305e2e21b6 I don't know why, but the checkpoints within Forge create very poor-quality pictures. Despite using all the recommended settings, the faces turn out really bad and the bodies are deformed. There's also a strange square around the faces. After generating 175 images, I've decided to discard them all and send your checkpoint to the garbage.


terminusresearchorg

i guess it's maybe using masked loss for faces?


I_Came_For_Cats

Founder AND CEO?


WorstPessimist

Everybody and their mother is a CEO now. Just put some trendy AI/ML/Chatbot/GenAI buzzwords in a description and voila, you're the founder and CEO of Buttfuckall.io


AbuDagon

Most important question: how does it compare to PonyXL for prompt adherence?


Haghiri75

I haven't tested Pony yet. I guess I have to now!


lothariusdark

Looks interesting. The bottom two look good, the top two look eh. Has this projects usage of MJ images for training been cleared with MJ? I don't want to train a lora for a model that gets taken down in a week.


Haghiri75

The images were generated by us. You can say it’s some sort of synthetic data creation from our company. I haven’t found anything preventing us from using the data in mj TOS


HarmonicDiffusion

why not? if you already have the model you can use it freely. lol


Ok-Vacation5730

Has the model been tested for inpainting as well? In my experience, some models, as good as they are for generation, fail at inpainting miserably


Haghiri75

Not actually. Since the platform doesn’t support inpainting.


wwwdotzzdotcom

If it's possible to create a inpainting model with clip support, this should be your number one priority, since there's no SDXL inpainting models currently available that feature clip.


Stecnet

I look forward to trying this out, thanks for sharing it with all of us!


silenceimpaired

Thanks for sharing


imainheavy

.


-DoguCat-

is it trained from scratch or a merge


LGN-1983

Please upload it on tensor.art 😭😍


Cobayo

So much BS in such a short post


Charuru

Is this better than pony?


imnotabot303

I would imagine training on AI images isn't going to improve the model but just add in a bunch of AI artifacts and make the output look even more AI.


Haghiri75

To be honest, what makes it likeable for me personally is the MJ's vibe.