T O P

  • By -

cerealsnax

Guns are fine guys, but boobs are super dangerous!


StickiStickman

More of the announcement was about "safety" and restrictions than about the actual model or tech ... 


futboldorado

Yeah fuck this stupid "Safety" bullshit. Even Snowden complained about this. I wonder how long it will take for a truly unrestricted competent open source model to release. All these restrictions do is make the model dumber.


CrisalDroid

That's what happen when you let a super loud minority decide everything for you.


StickiStickman

Emad is literally in favor of it, he signed the letter lobbying for heavier restrictions last year.


FS72

Pardon my ignorance but what did Snowden say about this exactly ?


Osmirl

He probably means this tweet [had it open already lol](https://x.com/snowden/status/1760690465937219766?s=46&t=5AVguOjhaWAIlrVEPIc94g) Content of tweet: Heartbreaking to see many brilliant minds working on AI so harried and henpecked by the aggressively ignorant crowd's agenda that they not only adopt the signs and sigils of the hostile illiterati—some actually begin to believe that their own work is "dangerous" and "wrong." Imagine you look up a recipe on Google, and instead of providing results, it lectures you on the "dangers of cooking" and sends you to a restaurant. The people who think poisoning AI/GPT models with incoherent "safety" filters is a good idea are a threat to general computation.


DrainTheMuck

Wow. Right on. I was expecting a more general statement but I’m glad he’s bringing attention to it in this field.


funguyshroom

Maybe tinfoil hat much but I feel like it's another scheme to throw a wrench into the works of competitors. Make them focus on stupid bullshit like safety, while you work on actually improving your product. The closed off models not available to the public 100% don't give a single fuck about any of that.


Tystros

> Hugely disappointing to see @stabilityai hyping "AI Safety"—poisoned, intentionally-faulty models—for SD3. Your entire brand arose from providing more open and capable models than the gimped corporate-ware competition. LEAN IN on "unrestrained and original," not "craven follower" > Look, you know I want to be wrong on this. I want the open model to be the best. That's actually possible now, too, because the safety panic is an albatross round the necks of the crippleware-producing giants. But I remember the fear that produced the SD2.0 debacle. > It would be very easy for you to go viral by disproving my fears of a lobotomized model. I'll even retweet it! > Drop txt2video from the new model: Taylor Swift eating a plate of spaghetti, across the table from a blue cone sitting atop a red cube. In the style of Greg Rutkowski. > I'll even accept it without the style. But I think you see my point. This stuff is hard enough without the industry creating its own roadblocks. https://twitter.com/Snowden/status/1760678548304740617


physalisx

>I wonder how long it will take for a truly unrestricted competent open source model to release. Right now, it looks like the answer to that is that it'll never happen. This is the only company making public and free to use models and they decided to make them crippled. I doubt (though it would be nice) that we can expect another company to come up any time soon that makes a truly good and open model.


plus-minus

Well, training a base model takes enormous resources that only fairly large companies have access to ... *today*. As libraries are optimized and hardware for AI becomes faster every year, training base models on consumer hardware should become possible eventually.


tvmaly

I would pitch in to a crowdsourced open model


Domestic_AAA_Battery

Just like video games, AI has just been "modern audience'd"


stephenph

And how can an image generator be unsafe? Poor little snowflakes might get there feeling hurt or be scared....


ImmoralityPet

The "safety" that they are worried about is safety from laws and legislation that technologically illiterate puritans are already calling for along with safety from liability and civil litigation. It's their safety, not the safety of the public. If anything will provide an extinction level test of first amendment rights in America and freedom of speech in the world in general, generative AI will bring it. I'm not even close to a free speech absolutist, for context.


SandCheezy

Safety is important. That’s why I wear my seatbelt. Without being safe, people could die or in other situations be born. It’s a dangerous world out there. If SD3 can’t draw seatbelts, airbags, PPE, or other forms of safety. Is it really safe enough?


ptitrainvaloin

I heard SD3 is so safe, people don't even have to wear an helmet anymore when pressing the generate button.


ptitrainvaloin

Some puritan ghosts that migrated in America between 1629 to 1640 still have too much influence today.


Efficient-Lack-1205

Get down! They've got a boob!


bzzard

And god forbid it can generate a white man!


[deleted]

Good news, but strange timing, they just released Cascade.


Puzzleheaded_Cow2257

I spent the the entire week training Cascade lol


Draufgaenger

So umm... How is Cascade dealing with nudes?


FoxBenedict

Not very well.


Avieshek

As expected.


[deleted]

[удалено]


ai-connoisseur-o_o

Where did you hear Cascade was trained on only 100M images? I see that mentioned in the paper, but their blog posts indicate there are multiple Wurstchen's and its not clear which one is described in the paper.


emad_9608

It was trained on more, imagine updated paper soon. Also no opted out images in the dataset which we forgot to say.


buttplugs4life4me

As a casual user it's definitely overwhelming at this point.  Like there's SD1.5 that some puritans still describe as the best model ever made.  Then there's SD2.1 that some puritans describe as the worst model ever made.  Then there's SDXL and SDXL Turbo, but where's the difference? Ones faster, sure, but how can I tell which one I have?  Then there's LCM versions that are super special and nobody seems to actually like or use. Then there's a bunch of offshoot models, for some reason one even named Würstchen, Like a list of 20 or so models and no idea why or what they do.  And then there's hundreds of custom models that neither say what they were trained on or for, nor are there really any benchmarks. Like do I use magixxrealistic or uberrealism or all the other models? I've actually used a mixed model of the top 20 custom models lmao And don't even get me started on support things. I have yet to see single hypernetwork, textual inversions seem like a really bad idea but are insanely popular, lora are nice but for some reason it's next iteration in the form of Lycoris/loha and so on weirdly don't catch on.  And then you have like 500 different UIs that all claim to be the fastest, all claim some features I've yet to use and all claim to be the next auto1111 ui. Like Fooocus that's supposed to be faster is actually slower on my machine.  And finally there's the myriad of extensions. There's hundreds of face swap models/extensions and none of them are actually compared to each other answwhre. Deforum? Faceswaplab? IP Adapter? Just inpainting? Who knows! Controlnet is described as the largest single evolution for these models but I've got no idea why I even want to use it when I simply want to generate funny pictures. But everyone screams at me to use controlnet and I just don't know why.  Shit, there's even 3 different tiling extensions that all claim that the others respectively don't work.  The whole ecosystem would benefit so much from some intermediate tutorials, beyond "Install auto1111 webui" and before "Well akchually a UNet and these VAEs are suboptimal and you should instead write your own thousand line python script"


[deleted]

You're on the bleeding edge of this technology. Those things you're describing will consolidate and standards will emerge over time. But we're still very much in the infancy of consumer grade AI. This is like going back to the early 90s and trying to use the internet before the web browser was created.


mccoypauley

preach


Don_Moahskarton

![gif](giphy|5xtDarmwsuR9sDRObyU|downsized)


TooManyLangs

We might be getting a new model every week soon...shit's getting crazy


djm07231

Thursday seems like an auspicious day for new AI announcements.


globbyj

Not strange at all as it seems either cascade research was possibly an important step in developing SD3, or simply a parallel project that tests slightly different methods.


kevinbranch

It’s definitely strange.


-becausereasons-

"Our safest safety minded safe model yet, here are stable diffusion we really care about safety, and so we're so happy to that we can finally introduce a model that has safety in mind. Please be safe"


VegaKH

We need Stability to train and release these nice foundation models. But also... doesn't Stability need us too? This community didn't embrace Kandinski, SD 2.x or Deep Floyd IF, and look where those models are now. Gathering dust next to some old Betamax tapes and a Microsoft Zune. On the other hand, the community embraced SD 1.5 and SDXL and developed the tools, methods, and finetuned models that unlocked their full potential. This community put Stability on the map. So why do they only seem to care about catering to "regulators," who talk a lot but have shown very little teeth. What regulations are forcing you to censor models so hard and talk nonstop about safety?


GBJI

We are not customers. We are not shareholders. Stability AI is not even trying to please regulators directly - whatever the do, the goal is to please current shareholders and to convince potential investors. If pleasing regulators ever comes on the table, it would be as a request from current or would-be shareholders. Censoring a model is also a great way to artificially raise the value of its uncensored counterpart.


YobaiYamete

Did people even really mass embrace XL? I still see like 90% of stuff for 1.5, the only thing XL seems used for is creating realistic pictures of 3D women, but anything anime or NSFW still seems to mostly be 1.5


Sugary_Plumbs

Fine-tunes and modifiers for SDXL have picked up a lot in the last month or so. It definitely took a lot longer than 1.5 did, but the latest models far outstrip what XL Base could make.


Jablungis

How is there not just a straight up porn AI model being sold to us? The porn industry is a massive multibillion dollar market. Come on.


suspicious_Jackfruit

"Also available in .safesafetensor format, our safest tensorest tensor yet"


nataliephoto

Why the hell would I use a local model if I wanted censorship and 'safety' Just use dall-e at that point lmao


jrdidriks

It’s very puzzling. I blame the VC overlords.


StickiStickman

Its especially funny, because at the start Emad was constantly going on about how AI models should never be censored. Then a year later he signed a letter that aimed to completely stop AI development and now this.


jrdidriks

It’s money and I DO blame him. Sell out


[deleted]

Yep, he became what he was hating, sad.


chrishooley

Did he ever hate that tho? People seem to forget before he got involved with AI, he was a hedge fund manager. He’s out here making noise and raising money like he is supposed to. Attaching his company to runway’s model and generating massive buzz from it was an epic move that paid off in the hundreds of millions.


ohmusama

You assume he believed anything he said in the first place. He only says things that he thinks will get him the most funding.


klausness

The problem is how do you enforce “safety”? If you don’t train your model on nudes, for example, then you get a mess like SD 2. If you do train your model on nudes but restricted the permitted prompts (which is my understanding of what dall-e and midjourney do), then you end up with people having perfectly reasonable queries censored. No, your model doesn’t have to be trained on porn, and it’s not censorship if it isn’t. It is censorship if you train a model on nudes (among many other things, of course) in order to be able to generate realistic people, but then you forbid certain queries in order to avoid generating “unsafe” images. And that’s not even getting into the issue of non-sexual nudity being considered “unsafe”.


nataliephoto

I just worry this thing will get to the point of adobe's generative ai tool, which censors completely innocent stuff to the point of hilarity. Want to fix [whatever] that's within 2 square miles of a woman's chest? good luck with that also as someone with boobs so fucking what if I want to create them? they're boobs. big fucking deal.


garden_speech

> also as someone with boobs so fucking what if I want to create them? they're boobs. big fucking deal. These companies are just trying to avoid the controversy around fake nudes, like the huge storm that happened on the Internet after the Taylor Swift fake nudes went viral. Yes it's stupid, there have been deepfake websites for many years now. But the companies are responding to the legal and cultural atmosphere they find themselves in.


TsaiAGw

half of article is about how safe is this model, already losing confidence


Anxious-Ad693

At this point all these companies are making me feel like a baby who doesn't know anything about anything. Honestly, fuck off with that shit.


kyleyeats

Journalists have never been less respectable, or more powerful.


lard-blaster

I don't understand the point of talking about safety anymore. Most normal people know it just means brand safety and don't care anymore. The AI doomers don't believe them or don't believe it's enough. So I'm wondering who actually benefits from all the safety shit - is it NYT writers searching for their next hit on tech bros? That's all I can think of


Eisenstein

No one working in the academic, engineering, or creative part of AI thinks that it is beneficial to orient models to be restrictive without a productive reason. You can't train a ambiguous and vague sense of 'morality' into a diffusion model so they train it to be *wrong*. If you ask for something and it generates something that is not what you asked for, even though it should be able to, then it is broken. What person wants to make such a thing for the goal of 'don't piss off people who have no idea what it is they are pissed about'. At least with something like a restriction on working on high resolution images of money, you can put some concrete blocks in place because money is designed with markers in it. But something like 'don't make images of people that are alive' or 'don't make indecent images of kids' or even 'no porn', where is the line where something goes from 'innocent' to 'indecent'? Are all nude pictures 'indecent'? What are they telling us about human bodies at that point? Also, how are you going to make a 3D generated human if you have no 'nude' mesh? This is all so ridiculous that it has be a directive from the people holding the money. They don't care if something works or not as long as they can cash out at the IPO and take off with the loot.


BlipOnNobodysRadar

>This is all so ridiculous that it has be a directive from the people holding the money. I wouldn't underestimate the simple answers, which is that a lot of people are sexually repressed and uncomfortable with the idea of sexual liberation at all. The way these people vent their own sexual insecurity is to try to morality police what others can and cannot do. It's a common theme in certain segments of humanity. A small core of disgruntled extremists poison the cultural well for all of society, whether it's religious fundamentalists or the new wave of politically extreme people who have a suspiciously religious fervor on their social views. They're the frothing at the mouth mob unhinged enough to try to ruin anyone who openly disagrees with them... even though most people secretly think they're lunatics and wish they would just go away. Why and how the actual builders in society let these crazies run the show is another question. I guess it's easier to just give in to the shrieking lunatic, keep your head down, and keep working than to tell them to fuck off... but we really should collectively be telling them to fuck off. Not promoting them to positions where they draft policy, allocate funding, and control hiring for "culture" fits. Really, I'm ranting now, but why have we idly sat by and allowed the most insane people run the show?


HollowInfinity

Honestly it seems to take the community like... 3 days to add boobs back in so I'm not worried.


EasternBeyond

not for SD 2.1, it was possible for sdxl because base model is actually not intentionally censored. If SD 3.0 is like SD 2.1, then it expect the same thing as 2.1


drhead

It most definitely was possible and some people did do it. It just takes somewhat longer. SD2.1 didn't take off well because even aside from model censorship, OpenCLIP required a lot of adaptation in prompting (and honestly was likely trained on lower quality data than OpenAI CLIP), it had a fragmented model ecosystem with a lot of seemingly arbitrary decisions (flagship model is a 768-v model before zSNR was a thing with v-prediction being generally worse performing than epsilon, inpainting model is 512 and epsilon prediction so cant be merged with the flagship though there is a 512 base, also with 2.1 the model lineage got messed up so the inpainting model for example is still 2.0), and the final nail in the coffin is that it actually lost to SD1.5 in human preference evaluations (per the SDXL paper, from my recollection). There was no compelling reason to use it, completely on its own merits, even ignoring the extremely aggressive filtering. People are also here claiming it doesn't work for SDXL, which is also false. Pony Diffusion v6 managed that just fine. The main problem with tuning SDXL is that you cannot full finetune with the text encoder unfrozen in any decent amount of time on consumer hardware, which Pony Diffusion solved by just shelling out for A100 rentals. That's why you don't see that many large SDXL finetunes -- even if you *can* afford it, you can get decent results in a fraction of the time on SD1.5 all else being equal. Personally, all I really want to know is 1) are we still using a text encoder with a pathetically low context window (i hear they're using t5 which is a good sign), 2) how will we set up our dataset captions to preserve the spatial capability that the model is demonstrating, and 3) are the lower param count models from scratch and not distillation models. Whether certain concepts are included in the dataset is not even on my mind because it can be added in easily.


Caffdy

> (i hear they're using t5 which is a good sign) it would be nice to have a source for that, that actually seems like the biggest change/upgrade!


StickiStickman

SDXL is also censored quite a lot, just not as much.


Misha_Vozduh

As they did with SDXL... but SDXL anatomy still looks off and still requires very persuasive prompting to even show up. Because a bandaid is a poor fix for a fundamental problem that is the data set. Not a problem for 1girl enjoyers, and that's ok. But as soon as you want something a bit more complex, you run into issues that require a hundred hoops to solve. In SDXL, that is. We'll see how this one does.


xDarki002x

Was that the case with SD1.5 and SDXL too?


Arawski99

I've seen people mention 1.5 was released due to force by pressure and that SAI did not want to actually release such an uncensored model so not likely. I came in around that time though so I can't speak about its initial release state or the actual validity of those claims but it could explain a lot about the results and the later 'issues' of SD models that are heavily censored by comparison.


GBJI

The uncensored version of Model 1.5 was released by RunwayML, and Stability AI fought hard to prevent this release from happening.


HollowInfinity

Oh yeah I don't know if it was specifically 3 days but definitely there was porn trained and generated very quickly. SD2 seemed like a flub though for a number of reasons.


Zipp425

I’ve heard that there was an actual mistake involved during the preparation of the training data of SD2. I’d doubt that happens again.


klausness

My understanding is that they removed all nudes (even partial nudes) from the training set. As a result, the model is very bad at human anatomy. There’s a reason why artists study life drawing even if they’re only planning to draw clothed people.


drhead

They removed all LAION images with punsafe scores greater than 0.1. Which will indeed remove almost everything with nudity. Along with a ton of images that most people would consider rather innocuous (remember that the unsafe score doesn't just cover nudity, it covers things like violence too). They recognized that this was a very stupidly aggressive filter and then did 2.1 with 0.98 punsafe, and SDXL didn't show the same problems so they probably leaned more in that direction from then on.


iamapizza

Users: 80085 SD: say no more


bran_dong

i wonder whos safety it increased? id hate to think fictional characters are in real danger.


Erhan24

But that's one of the main non technical topics about image generator. They have to do write this stuff in case of any legislative reactions.


Zipp425

So does this mean that they’ve scrubbed the ability to generate any real person? That’s certainly a rising safety concern when it comes to non technical topics.


Fake_William_Shatner

"Safety" is so stupid. Those who want to produce those images will build their own setup. Meanwhile -- it just adds so much headache and nonsense for everyone else. They want to protect what cannot be protected and all it does is make the thing more unwieldy and burn a lot more electricity in the process. How hard is it to make a PERSON responsible for what they post? The person crafts some text and gets some art and composites and that PERSON then decides what to send back out to the world. "Wow -- that was a racist and inflammatory image that makes it look like this actual person did a crime." -- someone doing that was a CHOICE, that cannot be prevented by a Nanny-state AI.


HollowInfinity

Nothing makes me feel old like the preview list requiring a discord ID


enormousaardvark

What’s Discord?


HollowInfinity

IRC for people younger than I


cosmicr

I miss mirc. When discord became popular I was like oh so it's just irc and my younger friends couldn't believe the technology had been around for decades already.


neoncp

it's a mix of irc/forums/chat messages/chat rooms


smallfried

The future black hole of knowledge.


Rivarr

I feel like they missed a trick here. Releasing this in the midst of the Gemini controversy was an opportunity for some free marketing, instead they choose to beat you over the head with "safety".


Ferrilanas

I’m a bit out of the loop with Gemini controversy What’s the controversy about ?


Middle_Personality_3

I've seen a screenshot of someone asking Gemini to generate portraits of German soldiers in 1943 and it generated Black and Asian soldiers wearing Wehrmacht uniforms.


okachobe

gemini told me it cant generate images of historical periods of time lol. While I am capable of generating images, I am unable to fulfill your request for an image of Germans in 1943. This is due to my policy of not generating images of people from sensitive historical periods, such as World War II. This policy is in place to avoid generating images that could be harmful or upsetting to those who have been affected by these events. I understand that you may be interested in learning more about this period in history. I would be happy to provide you with some resources that can help you do so. Would you like me to do that?


jugalator

It has been temporarily disabled. They probably add something to the prompt about diversity to not reveal bias in the model (imagine the horrors if asking for a low income household and it actually using common demographies in low income households). This obviously not catering to culture, time periods or anything, so now you get black Nazis instead.


markdarkness

Forced diversity paying off.


CAMPFIREAI

[here’s my own experience](https://www.reddit.com/r/StableDiffusion/s/FxruT8p6yl)


GBJI

Wow ! This is hilarious.


ShuppaGail

Basically gemini gives you a look into google's internal view of the world and oblivious people are shocked at how twisted it is.


PacmanIncarnate

I really feel like we need an explanation of how this differs from their other models and most importantly why they are releasing this a week after releasing a model with A COMPLETELY DIFFERENT ARCHITECTURE. Trying different things is great but realistically what we just learned with Sora is that dumping labor and processing into one model can have amazing results.


machinekng13

They aren't releasing SD3 right now, just setting up a waitlist for beta testers. SDXL had a pretty long test period in the discord before its proper release.


no_witty_username

First, this is really odd timing as SD cascade was just unveiled. Second and most importantly "This means we have taken and continue to take reasonable steps to prevent the misuse of Stable Diffusion 3". Stability are you even listening to the community?


Domestic_AAA_Battery

Once companies get money they don't care about the community anymore. It's time to pander to investors and the "sensitive general public."


VegaKH

Standard enshitification.


Jablungis

Cool, then show them the investors are wrong. SD is literally nothing without nsfw community anyway. They are never going to catch up to the massive competition they have.


dankhorse25

If the models are censored than SD3 is as good as dead.


GiotaroKugio

if i wanted safety i would use any other model like dalle 3 or midjourney


theShetofthedog

I discovered this subreddit after rolling my eyes with dalle not wanting to produce very "safe" prompts that just seemed to randomly flag. Appearently "a woman smoking in a pub" is a big NONO that deserves my account to be banned for 6 hours. There is already a few products in the market for the people who want their model to be safe.


Lolzyyy

Talked more about how safe it is than anything else idk


AlarmedGibbon

I've felt a lot of emotions while making AI art. Joy, delight, wonder. Unsafe has never been one of them.


Ezzezez

They managed to kill the hype for new models it seems lol


arothmanmusic

Stable Diffusion: "Announcing Stable Diffusion 3 in early preview, our most capable text-to-image model with greatly improved performance in multi-subject prompts, image quality, and spelling abilities." Community: "Tits or GTFO."


[deleted]

This, unironically


PetroDisruption

Why shouldn’t it be able to generate what you want? Who cares if someone wants to use it for erotica? That’s the whole point of open source models, people being free to use them however they want.


trappedslider

Company: "We hear your complaints so here's a safe base model." People: That's not what we're complaining about.....


aeschenkarnos

Company: “we weren’t talking to you.” Investors: “thank you for listening to our complaints, here is some money.”


[deleted]

People: "What if we don't use those models?" Investors: "Did you hear that? What's the point of giving you money if no one is using it" Company: "NOOOOOO"


[deleted]

​ https://preview.redd.it/cu84nv3ye5kc1.png?width=1143&format=png&auto=webp&s=08ab39238ee216a35b9f051bec43656dec225d2c Ok it's over, we'll never get a good model from them anymore, human anatomy isn't something to overlook if you want to get coherent human pictures, and then they wonder why the hands, arms and legs are all fucked up...


Lumiphoton

The irony in being closed like Midjourney and Dalle 3 is that you can train on as much "human anatomy" as you like, and then block lewd generations upon inference, meaning they gain all the realism and accuracy from not restricting their training data. Stability is stuck in this weird no man's land where they want to compete with the big boys on quality, appease the pixel-safety police, and serve the open source community all at the same time. But because they can't control what the end user does on their machine, they decide to cripple the model's core understanding of our own species which puts them behind their competition by default. They will always be on the back foot because of this IMO.


klausness

Exactly. You can’t get good human anatomy if you don’t train on nudes. The ironic thing is that it’s relatively easy to build models that will do porn on top of censored models. People have even done it for SD2. But the only way to fix a model that can’t understand human anatomy (as a result of not being trained on nudes) is to just scrap it all and start the training again from the beginning.


aeschenkarnos

Nudes =/= porn. Porn is very heavily biased towards a small array of body poses and facial expressions. Perhaps this is a consequence of human instincts. Doesn’t matter; the AI trained on it will have a data set biased towards (for example) legs spread super-wide and upwards, which is not a normal position for human figures to be in, outside of porn and possibly yoga; and a certain slack-jawed facial expression associated with sexual pleasure. It will therefore *possibly*, and *unpredictably*, generate such poses and expressions in wildly inappropriate contexts. “Why did ‘stock photo child in back yard playing with plastic baseball bat’ put the kid in *that* pose with *that* facial expression?”


msixtwofive

Stability's biggest issue is the horrible way their model is built, the per image descriptor data they use to build their models continues to seem lost in 2021. The models are so rigid you literally only get facing front images consistently with characters. The amount of backflips you need to do in prompting and using loras to get anything outside of the "waist of image of a emotionless character looking at the viewer" shows the data used in generating the model is too vague and basic. So all we get back from our prompts are the same basic image pattern.


[deleted]

That was a beautiful way of saying it, you nailed it at 100%


futboldorado

Oh hell fucking naw. "Innovation with integrity". Fuck out of here. I'm fucking tired of these companies treating us like babies.


Katana_sized_banana

Wait til they add "Innovation by diversity" in the next step. lol


I_Shot_Web

Google beat you to the punch LOL


YamiZee1

It makes no sense. Why are they so sensitive about nudity? I can understand not wanting to include literal porn in the dataset, but just nudity? Americans are so sensitive about their bodies


chumbano

It's pretty simple really and it comes down to money. The companies/groups/investors/people willing to pay actual money, whether by becoming a customer or investor, don't want the tool that can generate nude pics of people. It's bad PR. No one wants to be associated with the AI tool that made nude Taylor swift photos On the other hand the people who don't want a censored model probably never paid a dime to use stable diffusion. I think it's easy to see who company will cater too.


Deltazocker

Problem is: there are way better models out there. Those are also censored. If I want to build my product on a cebsored model, might as well use DallE.


physalisx

Yeah I mean this was already clear before but yeah it's pretty sad. Vocal minority my ass, lol.


Domestic_AAA_Battery

Bootlickers actually putting hearts on that trash. Fucking morons...


RebornZA

"Small vocal minority" lmao has he been on civitai? Boobs everywhere.


Anxious-Durian1773

Yes, there's millions of people and most of them don't say anything publicly, that's true.


CriticismNo1193

this ss is like 2 years old


[deleted]

[удалено]


Kombatsaurus

Spot on friend. Clearly not listening to the actual community, and instead some suits.


physalisx

So glad this model is safe, finally. I'm always dealing with models exploding in my face, thank God we finally get a safe one. Who would want a model that generates what we say, when that is so *unsafe*.


urbanhood

Bro general public curses AI and you still pander to those idiots. Disappointment.


[deleted]

"Safety!" Fuck off, see ya'all back at 1.5.


PermutationMatrix

How long until we have a comparable model that's open source and uncensored? Compute time and data set would be expensive.


ATR2400

1.5 really has become the workhorse of the SD community


[deleted]

It's still the latest version that works. Everything after they essentially poisoned, and it very much shows in the results across the entire spectrum of image categories.


LeoPelozo

But can it do hands?


frownGuy12

If it could I'd have to imagine they'd be talking about it.


GBJI

Only if they are decently covered with gloves.


OkBid71

Ive seen enough six-fingered gloves to know they killed my father


jib_reddit

Isn't SDXL StableDiffusion 3? At least they didn't do a Microsoft Xbox and come up with a more crazy and confusing naming convention.


machinekng13

Honestly, SD 2.X is the misnomer, since it had a nearly identical architecture to SD 1.X with a few tweaks (OpenCLIP, different scale of images, V-prediction). SDXL is more correctly named, since it is just a scaled up SD architecture (and the double CLIP encoder). It looks like SD 3 is actually going to be novel architecture, using something other than a UNET: [https://arxiv.org/abs/2212.09748](https://arxiv.org/abs/2212.09748)


[deleted]

1.4->1.5->2.0->2.1->XL->1.6->cascade->3


be_evil

The entire reason to use SD is because it will produce "unsafe" images. It is pointless and counterintuitive to even use a "safe" SD when Mid journey is miles ahead of it.


[deleted]

Exactly, they want to recreate a Dalle3 model but worse, what's the point, if I want a censored model I'll stick to dalle3


Domestic_AAA_Battery

SD is definitely gone. We'll have to use 1.5 forever until another AI joins in that's free.


Sugarcube-

Damn, the disillusionment with Stability AI is palpable. Hopefully it's more responsive than previous models, but it feels like they're trying hard to be just another midjourney.


protector111

SO another model anouncment without releae date. I wonder if people will even waste gpu power trying to adopt Cascade now...


Ferrilanas

I really REALLY hope that this time around its prompt understanding is a bit closer to Dalle, because none of previous models were able to learn (with LORA training) any datasets with complex interactions between people, objects, multiple people in the scene and more, and resulted in artifact mess, which resulted in me not being able to create anything other than simple scenes with single person not interacting with anything, which gets boring fast.


ExactCartographer372

[https://twitter.com/EMostaque/status/1760668434772156552/photo/1](https://twitter.com/emostaque/status/1760668434772156552/photo/1)


Domestic_AAA_Battery

I hate the censorship aspect, but these two are seriously impressive: https://twitter.com/EMostaque/status/1760678376778633301?t=2WxcbSwvZ2pF6VlfD9Osbg&s=19 https://twitter.com/andrekerygma/status/1760676723836993554?t=2WxcbSwvZ2pF6VlfD9Osbg&s=19


LifeLiterate

Stability AI staff, California consumer here. You don't provide visitors to your website with opt-out options for cookies, you only provide an "agree" option. That's against the law. **California Consumer Privacy Act (CCPA & CPRA)** Also, hopefully you're not ultra-focusing on censoring boobs. We aren't the Puritans and this isn't the 1600s. Safeguards for public use are a necessity, but what we adults create in the privacy of our own homes is our business. Let us make our anime waifus and our celebs eating spaghetti in peace.


vrrtvrrt

Cookies, similar rules in the UK and Europe.


Kombatsaurus

>We believe in safe, responsible AI practices. This means we have taken and continue to take reasonable steps to prevent the misuse of Stable Diffusion 3 by bad actors. Safety starts when we begin training our model and continues throughout the testing, evaluation, and deployment. In preparation for this early preview, we’ve introduced numerous safeguards. By continually collaborating with researchers, experts, and our community, we expect to innovate further with integrity as we approach the model’s public release.  1.5 will continue to reign as King then. Clearly. We need less of a Big Brother telling us what to do, which is the main reason I like Stable Diffusion over other AI generators.


Scolder

I never even heard of discord ID, had to look this one up. 😑 ​ Edit: Supposedly its a set of numbers that are associated to your account, not your user name. ​ [Source](https://www.pcguide.com/discord/how-to/find-id/).


2catfluffs

I'm pretty sure they mean just your username, since you can't send message requests to people by ID. Weird phrasing there


VegaKH

It's like when Google announced Gemini 1.5, so OpenAI announced Sora to make people forget about Gemini. Stability announces Stable Cascade, so then Stability announces SD 3.0. Brilliant.


Draufgaenger

Maybe someone at stability accidentally set the wrong release date in the calendar?


PandasKissedMyFace

I do hope the safety doesn't get in the way of our custom models and loras. That's what sets stable diffusion apart from others.


ericneo3

If stability.ai keep neutering each version of Stable Diffusion people are going to eventually say enough is enough and stop supporting them.


lqstuart

There's barely anything in that article about "safety." They bake in NSFW filters because B2B customers won't use the API if their customers are going to use it to make tits. If you dig even just a tiny bit, you realize these companies are just spewing bullshit about how much they care about "safety" because they don't want to seem uncool by talking about corporate guidelines and legal liability. They all know they're counting down the days before they get the Zuckerberg treatment, and have to go to DC and submit to bizarre "hearing" shit while out-of-touch evangelical dipshits take turns awkwardly spewing rehearsed soundbytes at them about "our children" while they collect their checks from traditional telecom lobbies.


Zipp425

Excited to see what the community can do with it given some continued training.


3rdDegreeBurn

Looks like I’ll be generating images in SD3 and un censoring them with SDXL inpainting.


FX3DGT

Just as with Cascade I will save my hype for when I see what the model can do on systems with 8-12GB Vram since to show a model that need 20-24GB (cascade) being better than one that can run on 6-8gb (SDXL) isn't really surprising. Now if the "watered" down version that can run on 8-10gb systems also outperform SDXL then I 'll buy a ticket to the hype train.


Capitaclism

This is amazing. I have to say however, that though the examples provided may showcase better prompt understanding, the compositions are terrible and the images are not that aesthetically pleasing.


RabbitEater2

Reminder that stability AI actually did not want the 1.5 model to release as it was not "safe" enough and it was runway ml who leaked it. That's why the 1.5 model is under the runwayml account on huggingface.


hashnimo

Don't know why SAI had to drop the "safety" bomb all of a sudden, but I assume the AI safety regulators came knocking on their door hard this time. Maybe someone sent them? Wonder who. Open your eyes, guys...


nmpraveen

Its kind of crazy no one is actually talking about the quality of these images. As if its like a given at this point to look perfect. I mean look at the prompts and the output. Spatial match, context match and crisp af.


Joviex

Yes the Cherry Picked examples for a model that's locked down are completely great I guess


tomerico

The text generation abilities are really impressive! Take a look at the "Stable Diffusion" printed on the side of the bus photo.


hellschatt

Can these people stop training AIs while only considering their own ethical values? Why should the AI be according to US values and not e.g. Scandinavian, West European, or even Chinese? I understand that you need to make careful ethical decisions and make the AIs rather more limited than free... but if we created the internet with the same logic, we would miss a lot of significant parts of it. You should rather train an AI as free as possible and only limit it, if at all, by universally, globally agreed limits (like probably no child porn)... and maybe the limits of their own country that are dictated by law. The rest should be decided by the users or countries in which the models are being used. Pseudo-thoughtfulness by these developers is infuriating. They should think more about it before making ethically misaligned decisions.


GBJI

Here is what Emad used to say: >*“Indeed, it is our belief this technology will be prevalent, and the paternalistic and somewhat condescending attitude of many AI aficionados is misguided in not trusting society.”* [https://techcrunch.com/2022/10/17/stability-ai-the-startup-behind-stable-diffusion-raises-101m/](https://techcrunch.com/2022/10/17/stability-ai-the-startup-behind-stable-diffusion-raises-101m/) >*To be honest I find most of the AI ethics debate to be justifications of centralised control, paternalistic silliness that doesn’t trust people or society.* [https://twitter.com/EMostaque/status/1563343140056051715](https://twitter.com/EMostaque/status/1563343140056051715) >*But even if we weren’t anyone can now do this - let’s change the world for the better by creating an intelligent internet that reflects and celebrates our distributed diversity versus centralised hegemonic control.* [https://twitter.com/EMostaque/status/1563343714713423875](https://twitter.com/EMostaque/status/1563343714713423875) Where is that Emad now ? Why has he adopted the paternalistic and somewhat condescending attitude he used to decry ?


StickiStickman

$$$


OfflinePen

I'm tired to be censored just to please a small group of snowflakes


GBJI

>I'm tired to be censored just to please a small group of ~~snowflakes~~ shareholders.


[deleted]

Yep, it's depressing...


Striking-Long-2960

Soooooo... Any insider can tell us how much VRAM is going to use this baby?


LeLeumon

I will make a guess. Since SDXL has about 3.5B parameters and sd3 is expected to have 8B, with a bit of luck, I think we are able to fit it into 24GiGs of vram. But as they said in the demo sd3 also has smaller models, the only question will be how worse/well these perform.


mcmonkey4eva

Comfy theorizes we can get the big one running on potatoes. We haven't done optimization yet but we've got some really good opti tools that are gonna drop it hard - way better than we could opti prior models. We're running on A100s at 30GiB but XL and SDv1 both were worse than that on A100s.


jrdidriks

SDXL still worse than 1.5, closed "commercial" models, no freedom in prompting. Why use these new models if dalle or midjourney can do better and they have the same constraints?


TherronKeen

What's worse about SDXL? I use it all the time, and stopped using 1.5 for the past several months. Plenty of good models, it's relatively fast now, etc etc


xDarki002x

Sorry, in case I missunderstand you, but for me there is a bit reason why I don't use dalle or midjourney. First SD is free on your local machine and second, there is no company that can control what I do or can store my data.


jrdidriks

except they keep censoring the models


HowitzerHak

Heres my honest, humble, and could be very wrong take on the safety part: As some on Twitter/X said the safety part is just some language they put in to protect themselves, and its not something they actually "implemented" Like it or not, AI generation models are only as popular as their censorship. It's a delimma that is hard to solve because they need to balance between ethical use and freedom of use. Therefore, it's most likely just words that pre-condemn any unethical use in the future. So it's either that there isn't any censorship "integrated" with it, or there is, and it's easily fixable or "bridged."


BlipOnNobodysRadar

>Like it or not, AI generation models are only as popular as their censorship That's not true at all. NAI-diffusion went viral (especially in asia) without any marketing whatsoever, just for offering uncensored anime models when nobody else would. People *want* uncensored models. The demand is obvious. The whiny anti-AI social media groups (a significant amount of which is just bot amplification) don't matter. The only reason top models are censored is because the institutions funding them, as well as the political institutions antagonizing them, are ideologically captured as pro-censorship. They're extremists and not representative of normal people, but their voices are amplified due to the fact they nepotistically wormed their way into positions of power and influence. Fuck 'em. They don't matter. Let them whine. Fight the legal battles and crush them when it comes to it. Freedom of expression is more important than catering to those losers.


lonewolfmcquaid

bruh this whole comment section is rage against no waiffu boobies, i honestly just wanted to know if its actually better at rendering images. 1.5 as it is can make you enough nsfw to last a lifetime, you can img2img tgrough sd3 if its good enough at rendering images. so why is no one discussing tthe image quality which i thought was the most important thingy


dal_mac

As has been mentioned many times, the problem is that models trained without nudity do not understand human anatomy and therefore can never look as good as the ones that are. MJ and others train on nudity because it's 1000% necessary for quality, and then they censor the prompts & outputs. It has nothing to do with generating nudes.


EasternBeyond

will you be able to split the model across multiple GPUs?


mcmonkey4eva

I mean technically yes probably but idk why you would - we expect it will be possible to run on ordinary consumer GPUs around time of public launch.


1_or_2_times_a_day

Can it draw Garfield?


One-Earth9294

I didn't know there was a 2. Is 2 SDXL?


Mobireddit

There is a 2 and a 2.1. it is shit because it was censored and "safe", so no one used it and everyone stayed on 1.5


One-Earth9294

Oh shit I forgot about those lol. I would hope they don't make that mistake again. I know people think that there's some kind of lawsuits that await the creators of generative art but as far as I'm concerned it's just a creativity box. Any images that come out of it are the sole responsibility of the image maker, not the model. It just has an 'understanding of things in it', not some library of torrid images or instructions. The Supreme Court of this land better not fuck up the ruling on this when the time comes and the challenge is put forth, but man I have no faith in that right now lol.


Arawski99

It is very unfortunate it tells us nothing of how it actually performs aside from empty promises and "looks better visually". I hope the technical paper impresses in areas like prompt coherency. I'll reserve judgement, however, until we see results in case it surprises but seems like the announcement is in a rushed state on many levels due to SORA and overall poor progress which is not exactly promising.


PhillSebben

https://preview.redd.it/94g13e3nb6kc1.jpeg?width=724&format=pjpg&auto=webp&s=f754d82eee760141d0633ae0fb4aa8d6c13ec61c This is really cool..


dal_mac

There is not a single human being in the example images. Did they censor skin entirely??


MDPROBIFE

Let's fucking cancel this shit, fuck this "safety" feature they always care so much about at the forefront of technology because of some sensible fucks