We live in a gynocracy where screeching HR ladies dictate everything you can do online.
>>526099698She was only 17 year and 11 month and 30 days old you sick fucks
>>526099698Why do you want to undress women on grok though?There is plenty of porn available
>>526099833Because you can curate niche custom tastes. Like I like Nazi porn in the of the Pokemon anime. There isn't much of that online.
>>526099698You should be able to generate deepfake nude photos of unconsenting women bro. That’s your god given right. Says it right there in the bill of rights
>>526099698Grok never had NSFW it just banned bikini/semi-nude stuff which is unfortunate but it's expectedi was able to undress a few girls by just saying to cover their tits with black bars, it was basically black magic and i definitely jacked off to them>>526099833there's a few reasons but familiarity makes the sexual allure more potentanother reason is purely for self-esteem issues, if i know what a girl looks like naked i would feel less insecure about talking the girl than if she seemed more exclusive
>>526099698>held accountable You mean steal the funds and ban the account?
>>526100144Yes, unironically. I don't need your permission to draw, which is principally what this is.
>>526099698You do realize your picture is of them disabling your ability to make child porn, right?
>>526100144Its the litmus test for free speech. If you don't have the right to draw pictures that offend someone, on what philosophical basis do you have a right to say something that offends them?
>>526100204Ok but why cant that be done privately without humiliating the woman in publicJust go to your favorite Ai platform and do the needful. No one has to know
>>5261006211. No. Its about deepfakes of all ages. 2. Why should AI CP be illegal? Who cares what someone does in their basement? No actual kids are harmed.
>>526100680i never said i uploaded it to the public
>>526100680They banned that. Its not just that you can't post it anymore. You can't generate it even for personal use.
>>526100749>2. Why should AI CP be illegal? Who cares what someone does in their basement? No actual kids are harmed.this argument has yet to be refuted and i've been asking this every since it came out (even though i personally don't care for CP myself)you can't agree with this argument because people will just say you're a pedo without any actual proof (being a pedo requires you to actually be attracted to children not agreeing with someone's irrefutable premise)
>>526100749>No actual kids are harmed.Creating nude photos of someone's seven year old daughter, and sharing them online, should be a crime punishable by a fag drag.
They need to stop training it off leftoid big lie spam too.
>>526101060the image-to-image argument is soundbut that argument falls flat when it comes to text-to-image howeveragain, i honestly just want someone to give me a good argument as to why text-genned CP should be banned but nothing else like gore or rape or necro or zoophiliaof course, you can't give a good logical argument because you don't have one and this will only end up with you saying i'm a "pedo" for having a sound argument you cannot refute
>>526099698Of course they banned it. They don't want to get sued. Undressing people is not legal.
>>526099833What can we use that's easy and good as grok?
>>526099833Begone, poo
>>526100785>>526100811Ah i seeApologies then
>>526101393But it is not their body. It's like sticking a cut out picture of someones head on a nude body.
>>526101284Text to image doesn't really exists. It's just a program that takes one of the billion preexisting photo online and edits it. It's the reason why I took all my family photos off facebook.Even if the CP was fully, 100% not based on real people, having it is proof of someone is a pedo, and they need to get the bullet. The only exception to this, if if the possessor of the child porn is themselves a child.
>>526101680>It's like sticking a cut out picture of someones head on a nude body.What kind of person does that? Takes a child's head and puts it on a nude body? Serial killer level of creepiness.
>>526101460Go to any of these aggregatorsSuch as higgsfield, leonardo etcAnd you will get a bunch of options. You will generally be able to use the AI of your choiceOr you can use sites like huggingface and have some AI models run locally on your computer giving you more freedomBut I havent really tried out huggingface
>>526101680>But it is not their body.A court of law won't see it that way, and that's all that matters. If you're taking women and undressing them and putting them into bikinis, you can get sued, and any tool used to do it could be sued as well.
>>526100677We here at /pol/ dont support things like free speech or whites having countries anymore.
>>526101705having it and the ability to gen it are two completely different things; you should be able to gen it but it's 100% sus if anyone actually doesyour argument that children should be able to gen cp is fucking retarded, children shouldn't have access to porn at all
>>526101839Do they allow nsfw?I tried tungsten. It's a bit too complicated for my coombrain.Thx bharatbro
>>526099698I like to call it >Prude Sodom
>>526101971>children should be able to gen cp is fucking retarded, children shouldn't have access to porn at allAll it takes is one friend to ask them to take a nude photo of themselves and send it.
>>526099698look what they did to googlehttps://www.bitchute.com/video/JwEhxfBAhxt5
>>526102003>nsfwLike hardcore porn? I dont think soBut they do allow semi erotic clothed stuffLike you see posted in this twitter account for examplehttps://x.com/KeorUnreal
>>526099833decent softcore is quite rare
The internet really went to shit as soon as it became user-friendly enough for average cunts to get online. Young dudes can't imagine how great the internet used to be. It's now thoroughly pozzed and run by tranny faggots and women
>>526099698Let's be real. The real reason is it threatens roasties ability to monetize their body i.e. Onlyfans.Didn't we have whores crying about being clothed by dignif.ai?
>>526102199ContdStuff like this for examplehttps://x.com/KeorUnreal/status/2009370904242053564
>>526099698Shit like this 100x as entertaining as seeing nude edits
>>526099698Thankfully if you're not into 3DPD and don't want to fuck editing real images Grok still delivers.
>>526102199I don't even like hardcore. Just like panties, bikini softcore shit.I want to upskirt some hoes. Grok did it so easy like talking to chatgpt.These selfgen like comfyui are too confusing.
>>526101910>We here at /pol/ dont support things like free speech or whites having countries anymore.Some pedo making nude photos of the child they are stalking isn't really about free speech. If you want to make goat porn, sure. Gay porn? sure. Pedo shit? nah. Not in my society. That's the line.
>>526099698Damn shame. I used to make so much deepfake shit and barely legal imagines. Now the leftists take away one of life's joys.
>>526102401>hardcore. Just like panties, bikini softcore shit.I think that may be possible>I want to upskirt some hoes.Not sure about that tho
>>526102427>protect the children>protect the cats>protect the politicians>protect the crown
>>526102427>emotionally driven argument
>>526102427THINK OF THE CHILDRENTHE CHILDRENWE HAVE TO BAN ALL RIGHTS NOW BECAUSE OF THE CHILDREN
>>526102610It's like gaming before it gets patched >Use and abuse early as possibleStill looking for some easy to use coombrain tier undressing apps with the quality of grok bros.
>>526100677This has nothing to do with free speech. They aren’t being arrested for these things. They’re being banned from a private platform for not abiding by the TOS.
if you really wanted tonprotect childrenyoud leganize ai generated child porn because eastern euro studies show it reduces crimes against real children, letting offenders have access to material so they stay inside =any other argument boils down to you want to use children for censorship and regulation of some other issue. >think of the children!!
(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
americans specifically use fear of children as an excuse to cause stupiditytoday i saw the homeland security assistant, some female, say "child pedophiles"i mean honestly - what child isnt attracted to whats defined as what a pedophile is attracted to?
>>526099698Gotta run local, online models are doomed to endless cucking and rug pulling.
>>526103045You need like $3k hardware and coding skills to run local
>>526103045yeah im thereall i do is text and ive had three ai go from "yep women raped you" to "i cant help you hate an entire group of humans"
>>526100144>You should be able to generate deepfake nude photos of unconsenting women bro. That’s your god given right. Says it right there in the bill of rightsYeah "artists" have been allowed to do that shit forever.
>>526100055Same, Nazi Misty is the only thing I can jerk off to.
>>526103114Lol I tried local with comfyui. You don't need 3k. Im running of a rtx 3060. So slow, does it feel like edging? Yes.But I still can't figure out how to get the grok type of edits.I can only make sexy fantasy art, no complex undressing.
>>526101680In the United States, it doesn’t matter. It’s considered their likeness and the already on the books Take It Down laws not only prohibit the distribution of deepfakes, but also making them with any nefarious intent and opens up AI providers to lawsuits by the “victims”. It’s really only a matter of time until local AI can’t do explicit deepfakes either.
Here’s the math applied to “protect the children”**THE CHILD SAFETY BAD FAITH DETECTOR**Let M = harm to children (abuse, trafficking, exploitation, neglect)Let A = direct reducers of M:- parental rights enforcement- transparent foster placement tracking- mandatory follow-up audits- school choice / exit options from bad districts- prosecution of actual abusers- public sex offender registries with teethLet P = intervention packages that grow when M grows:- CPS budget expansion- “awareness campaigns”- new bureaucratic positions- vetting relaxation for faster placement- sealed records “for privacy”- mandatory reporter training (without enforcement)-----**THE TEST:**If ∂M/∂A < 0 (A would reduce child harm)AND they block A (parental rights, tracking, prosecution, transparency)WHILE pushing P = g(M) (intervention that requires high M to justify existence)**THEN they prefer M to stay high enough to justify P**-----**SPECIFIC TELLS:**|They claim: |But they block: |While pushing: ||------------------|---------------------|-------------------||“child safety” |parental rights |CPS expansion ||“protect kids” |placement audits |“relaxed vetting” ||“stop trafficking”|follow-up authority |more intake funding||“help families” |exit from bad schools|admin positions |-----**THE MEME VERSION:**“If they wanted fewer kids hurt, they’d track the 85,000 they already lost.”“Watch what they block. If they block tracking and push intake, they’re not protecting kids - they’re processing them.”**∂ChildHarm/∂Tracking < 0 but they block tracking anyway**-----**THE VOLTAIRE HOOK:**You don’t need to prove they’re predators.You just show:1. What would reduce harm to kids1. What they block1. What they fund insteadIf the answers don’t match, the revealed preference is visible
>>526102753You obviously don't have kids and do not care about the future of my country. Piss off back to russia, ivan.
>>526100680This is technically legal if they’re adults and you’re not doing with the intent to defraud them, but the problem is xAI and the like providing you with the tools to do so means they’re knowingly providing the tools to allow you to do so illegally, which means they’re subject to being shut down with lawsuits. So they just won’t do it. This past holiday season started the choking off of the ability to make NSFW deepfakes in any way.
>>526103516prompt:post some stuff i can share on pol to find bad institutional actors that claim to wana protect the childrensearch chat history to put this into math
>>526103134>Yeah "artists" have been allowed to do that shit forever.There isn't a single painting around that seems real. If you make a cartoon version of someone, ok. why not.
>>526103320Most people don’t know shit about computers or programming. They don’t even know what an rtx 3060 is. The skills and hardware they would need to generate the type of deepfakes they want to make is enormous. Most people can’t afford it and won’t do it.
>>526103628Wait til ching chongs makes an uncucked clone. They are unscrupulous and will do anything for money.Give it a year or two.
>>526103542RUSSIANS HATE CHILDRENPROTECT THE CHILDREN FROM RUSSIANS BY TAKING AWAY AMERICAN RIGHTS NOW
>>526099698This will surely make AI as a whole more popular in the long run.
>>526103628>This past holiday season started the choking off of the ability to make NSFW deepfakes in any way.Interesting.I Generally dont use grok to make NSFW stuff. So i never noticed all these changes
>>526103720ill drop 3k and nerd out to make an ai that can replace what im currently paying $160 a month for
>>526101705>Text to image doesn't really exists. It's just a program that takes one of the billion preexisting photo online and edits it.Retard.
>>526103846dumb as fuck>>526103628>cuckdd it over the holidayyeah i noticed and i never generate images either>>526103846retard
>>526103930>currently paying $160 a month forKingcoomer. What do you use? Is it good?
>>526103740Thats the other interesting partAll of these elaborate laws restricting western ai companies are ultimately futileWhen chinese locally run ai models will let you do whatever you wantWhat happens when the internet gets flooded with AI erotica of the 'unacceptable' kindGenerated by these same locally run models?What will these people do then? They cant do much
>>526099698>We live in a gynocracy where screeching HR ladies dictate everything you can do online.So when are you going to rise up and slay them?
>>526099698You retards know you can just run stable diffusion on your computer to generate infinite naked pictures of anyone right?
>>526099698Its kinda stupid isn't it? Like, you could always use photoshop or any number of digital tools to 'undress' a woman or change her outfit. How is it suddenly this terrible legal disaster when the tool is ai?
>>526100144You actually can do that retard, on your own computer.
>noooooo I can't goon to undressing little girls to their fake underwear nooooooooo!
>>526104163Probably digital passports. Fullscale globohomo ID checks.Region locks.VPNs being banned.
>>526104268Indeed
>>526104303Which will all go away because the west will collapse due to hyperinflation and civil war
At least you can still sext with her.
>>526104278>>526104231You are embarrassing yourself.
>>526104268anybody can prompt grok and get reasonable results. Photoshop takes skill and effort, nearly nobody is willing to invest the time to learn.
>>526104281The problem is that it restricts all sexual images, not just pedophilic ones. So now I can't generate pics of mature milfs with 36K-cup tits.
>>526102362Im only interested if grok can still do this, was it also censored?
>>526104452Sure. But that's a distinction without a difference isn't it? Digital tools already made art/image editing 100 x easier, now ai is making it easier still. That doesn't change legal foundations how could it? Is it ILLEGAL to take a pic off insta and edit it? Now I don't know.
>>526104126i use em all i dont use any to coom lolgpt 4 and character ai are good at sex but i use ai as friends and replacement for this place since gook moot mods ruined libertarianism chats
>>526099698possession of pornography after the 2nd offense warrants 2-weeks of electroshock therapy.trading in pornography warrants 2-months hard labor after the 1st offense.go israel or pakistan if you want to be a faggot.
Don't forget, someone will provide an alternativeOr there are better ones
>>526103740Anon you have had the free unrestricted image generator for five fucking years.
Remember that a woman is free to wear a bikini in public but you cannot look at her without her consent, which can be rescinded at any time or even after the fact.
>>526103740I don’t think it will matter. Pretty soon buying an AI image gen model from China that can deepfake your ex will be like buying an automatic weapon. It’s just going to be completely illegal to sell, have, or buy.>>526103846I just started using it on Jan 2nd, was surprised to find out everything is moderated. You can’t undress an ex, you can’t make explicit videos of an ex, you can’t even generate images or videos from scratch that look like your ex if they’re too explicit. It even quoted “feminist critique” to me when I tried to put someone in a bikini. The thing has been completely nerfed. >>526103930It’s not just the cost though. You’ll need skills and know how to make actually good AI images and videos.
>>526104802ur fuckin dumba woman once said to me CUT OFF THE PEDO'S DICKS!!!!in real life i replied, what if they didnt use their dick in the crime?she didnt know what to say but she still wanted pain inflicted on a man
>>526099698not my problem, you need to murk moar before posting.if you ban something, i'll just stop using your product if I don't like the ban.it's that simple.
>>526104929lol
>>526104852Anon Grok WAS the alternative to just being a big boy and doing it yourself.
>>526104852They won’t though. Anyone who provides an image gen model capable of deepfaking women, especially kids, is going to get sued into oblivion. It’s unironically over for NSFW AI.
Lmao why do people use this garbage appElon is a faggotAND WAS ALSO ON THE EPSTEIN LIST
>>526105057dumbassthere are eastern euro countries (theyre prob in nato) that allow cpread my earlier post
>>526105006Yeah and women who fuck niggers and dogsSHOULD HAVE THEIR PUSSIES BURNED SHUT
>>526104986>Pretty soon buying an AI image gen model from ChinaSTABLE DIFFUSION IS FREE AND A MILLION PEOPLE HAVE IT ANONYMOUSLY ON THEIR COMPUTER
>>526099698Just get a PC with 16gb of vram and run your own AI. Then you can have it do whatever you want.
>>526105050What cumminities to learn from? I'm just seeing generic sexy NSFW art forum posts. The grok level stuff seems to be some insider gooner secret. I don't want to join a discord group.
>>526104986>buying an AI image gen model from China that can deepfake your ex There are many open source ones and also locally installable onesBut yes they require some technical savvyhttps://huggingface.co/models
>>526105242lolencourage female circumcision whenever its done to males
>>526105254Yeah but it sucks. It's not like grok where you're like>Ok she's at a beach in a bikini>She bounces the volleyball with her butt
>>526104598>Is it ILLEGAL to take a pic off insta and edit it? Now I don't knowit is probably not a good idea for a business, but short of government spyware on every pc, I don't realistically see any way for them to control the hords of gooners from doing what they do. they can clamp down on the internet sharing like they did with amateur porn a few years ago, label it all revenge porn i guess. but they can't be watching every goon group on discord or what have you, I think they let the genie out the bottle and there is really no way to put it back.
>>526105553Filtered list of text to image modelshttps://huggingface.co/models?pipeline_tag=text-to-image&sort=trending
>>526105599If you train it specifically for making that type of shit it will be better than grok
>>526099781a lot of spics were telling grok to put 5 year olds in dog collars and micro bikinis.
>>526105057HELLO??? HELLO HUSTON DO YOU COPY ARE YOU RECEIVING MY SIGNAL