[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/pol/ - Politically Incorrect

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
Flag
File
  • Please read the Rules and FAQ before posting.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


File: 176850942858593945.jpg (326 KB, 1080x2400)
326 KB
326 KB JPG
We live in a gynocracy where screeching HR ladies dictate everything you can do online.
>>
>>526099698
She was only 17 year and 11 month and 30 days old you sick fucks
>>
>>526099698
Why do you want to undress women on grok though?
There is plenty of porn available
>>
>>526099833
Because you can curate niche custom tastes. Like I like Nazi porn in the of the Pokemon anime. There isn't much of that online.
>>
>>526099698
You should be able to generate deepfake nude photos of unconsenting women bro. That’s your god given right. Says it right there in the bill of rights
>>
File: 1762121695387025.jpg (43 KB, 519x374)
43 KB
43 KB JPG
>>526099698
Grok never had NSFW it just banned bikini/semi-nude stuff which is unfortunate but it's expected
i was able to undress a few girls by just saying to cover their tits with black bars, it was basically black magic and i definitely jacked off to them
>>526099833
there's a few reasons but familiarity makes the sexual allure more potent
another reason is purely for self-esteem issues, if i know what a girl looks like naked i would feel less insecure about talking the girl than if she seemed more exclusive
>>
>>526099698
>held accountable
You mean steal the funds and ban the account?
>>
>>526100144
Yes, unironically. I don't need your permission to draw, which is principally what this is.
>>
>>526099698
You do realize your picture is of them disabling your ability to make child porn, right?
>>
>>526100144
Its the litmus test for free speech. If you don't have the right to draw pictures that offend someone, on what philosophical basis do you have a right to say something that offends them?
>>
>>526100204
Ok but why cant that be done privately without humiliating the woman in public
Just go to your favorite Ai platform and do the needful. No one has to know
>>
>>526100621
1. No. Its about deepfakes of all ages.

2. Why should AI CP be illegal? Who cares what someone does in their basement? No actual kids are harmed.
>>
>>526100680
i never said i uploaded it to the public
>>
>>526100680
They banned that. Its not just that you can't post it anymore. You can't generate it even for personal use.
>>
>>526100749
>2. Why should AI CP be illegal? Who cares what someone does in their basement? No actual kids are harmed.
this argument has yet to be refuted and i've been asking this every since it came out (even though i personally don't care for CP myself)
you can't agree with this argument because people will just say you're a pedo without any actual proof (being a pedo requires you to actually be attracted to children not agreeing with someone's irrefutable premise)
>>
>>526100749
>No actual kids are harmed.
Creating nude photos of someone's seven year old daughter, and sharing them online, should be a crime punishable by a fag drag.
>>
They need to stop training it off leftoid big lie spam too.
>>
>>526101060
the image-to-image argument is sound
but that argument falls flat when it comes to text-to-image however

again, i honestly just want someone to give me a good argument as to why text-genned CP should be banned but nothing else like gore or rape or necro or zoophilia

of course, you can't give a good logical argument because you don't have one and this will only end up with you saying i'm a "pedo" for having a sound argument you cannot refute
>>
>>526099698
Of course they banned it. They don't want to get sued. Undressing people is not legal.
>>
>>526099833
What can we use that's easy and good as grok?
>>
>>526099833
Begone, poo
>>
>>526100785
>>526100811
Ah i see
Apologies then
>>
>>526101393
But it is not their body. It's like sticking a cut out picture of someones head on a nude body.
>>
>>526101284
Text to image doesn't really exists. It's just a program that takes one of the billion preexisting photo online and edits it. It's the reason why I took all my family photos off facebook.

Even if the CP was fully, 100% not based on real people, having it is proof of someone is a pedo, and they need to get the bullet. The only exception to this, if if the possessor of the child porn is themselves a child.
>>
>>526101680
>It's like sticking a cut out picture of someones head on a nude body.
What kind of person does that? Takes a child's head and puts it on a nude body? Serial killer level of creepiness.
>>
>>526101460
Go to any of these aggregators
Such as higgsfield, leonardo etc
And you will get a bunch of options. You will generally be able to use the AI of your choice

Or you can use sites like huggingface and have some AI models run locally on your computer giving you more freedom
But I havent really tried out huggingface
>>
>>526101680
>But it is not their body.
A court of law won't see it that way, and that's all that matters. If you're taking women and undressing them and putting them into bikinis, you can get sued, and any tool used to do it could be sued as well.
>>
>>526100677
We here at /pol/ dont support things like free speech or whites having countries anymore.
>>
>>526101705
having it and the ability to gen it are two completely different things; you should be able to gen it but it's 100% sus if anyone actually does

your argument that children should be able to gen cp is fucking retarded, children shouldn't have access to porn at all
>>
>>526101839
Do they allow nsfw?
I tried tungsten. It's a bit too complicated for my coombrain.
Thx bharatbro
>>
>>526099698
I like to call it
>Prude Sodom
>>
>>526101971
>children should be able to gen cp is fucking retarded, children shouldn't have access to porn at all
All it takes is one friend to ask them to take a nude photo of themselves and send it.
>>
>>526099698
look what they did to google
https://www.bitchute.com/video/JwEhxfBAhxt5
>>
>>526102003
>nsfw
Like hardcore porn? I dont think so
But they do allow semi erotic clothed stuff
Like you see posted in this twitter account for example
https://x.com/KeorUnreal
>>
>>526099833
decent softcore is quite rare
>>
The internet really went to shit as soon as it became user-friendly enough for average cunts to get online. Young dudes can't imagine how great the internet used to be. It's now thoroughly pozzed and run by tranny faggots and women
>>
>>526099698
Let's be real. The real reason is it threatens roasties ability to monetize their body i.e. Onlyfans.

Didn't we have whores crying about being clothed by dignif.ai?
>>
>>526102199
Contd
Stuff like this for example
https://x.com/KeorUnreal/status/2009370904242053564
>>
File: 1768503956291843.jpg (201 KB, 598x2230)
201 KB
201 KB JPG
>>526099698
Shit like this 100x as entertaining as seeing nude edits
>>
File: file.png (1.49 MB, 784x1168)
1.49 MB
1.49 MB PNG
>>526099698
Thankfully if you're not into 3DPD and don't want to fuck editing real images Grok still delivers.
>>
>>526102199
I don't even like hardcore. Just like panties, bikini softcore shit.
I want to upskirt some hoes.
Grok did it so easy like talking to chatgpt.
These selfgen like comfyui are too confusing.
>>
>>526101910
>We here at /pol/ dont support things like free speech or whites having countries anymore.
Some pedo making nude photos of the child they are stalking isn't really about free speech. If you want to make goat porn, sure. Gay porn? sure. Pedo shit? nah. Not in my society. That's the line.
>>
File: 1768018160067697.jpg (35 KB, 640x684)
35 KB
35 KB JPG
>>526099698
Damn shame. I used to make so much deepfake shit and barely legal imagines. Now the leftists take away one of life's joys.
>>
>>526102401
>hardcore. Just like panties, bikini softcore shit.
I think that may be possible
>I want to upskirt some hoes.
Not sure about that tho
>>
>>526102427
>protect the children
>protect the cats
>protect the politicians
>protect the crown
>>
>>526102427
>emotionally driven argument
>>
>>526102427
THINK OF THE CHILDREN
THE CHILDREN
WE HAVE TO BAN ALL RIGHTS NOW BECAUSE OF THE CHILDREN
>>
>>526102610
It's like gaming before it gets patched
>Use and abuse early as possible

Still looking for some easy to use coombrain tier undressing apps with the quality of grok bros.
>>
>>526100677
This has nothing to do with free speech. They aren’t being arrested for these things. They’re being banned from a private platform for not abiding by the TOS.
>>
File: IMG_0957.jpg (173 KB, 1024x1022)
173 KB
173 KB JPG
if you really wanted tonprotect children
youd leganize ai generated child porn
because eastern euro studies show it reduces crimes against real children, letting offenders have access to material so they stay inside

=
any other argument boils down to you want to use children for censorship and regulation of some other issue. >think of the children!!
>>
(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM

(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM

(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM

(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM

(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM

(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
>(((CSAM))) IS YHVH'S GIFT TO HIS LONG-SUFFERING GOYEEM
>>
americans specifically use fear of children as an excuse to cause stupidity

today i saw the homeland security assistant, some female, say "child pedophiles"


i mean honestly - what child isnt attracted to whats defined as what a pedophile is attracted to?
>>
>>526099698
Gotta run local, online models are doomed to endless cucking and rug pulling.
>>
>>526103045
You need like $3k hardware and coding skills to run local
>>
>>526103045
yeah im there
all i do is text and ive had three ai go from "yep women raped you" to "i cant help you hate an entire group of humans"
>>
>>526100144
>You should be able to generate deepfake nude photos of unconsenting women bro. That’s your god given right. Says it right there in the bill of rights
Yeah "artists" have been allowed to do that shit forever.
>>
>>526100055
Same, Nazi Misty is the only thing I can jerk off to.
>>
>>526103114
Lol I tried local with comfyui. You don't need 3k. Im running of a rtx 3060.

So slow, does it feel like edging? Yes.

But I still can't figure out how to get the grok type of edits.

I can only make sexy fantasy art, no complex undressing.
>>
>>526101680
In the United States, it doesn’t matter. It’s considered their likeness and the already on the books Take It Down laws not only prohibit the distribution of deepfakes, but also making them with any nefarious intent and opens up AI providers to lawsuits by the “victims”. It’s really only a matter of time until local AI can’t do explicit deepfakes either.
>>
File: IMG_0949.png (579 KB, 655x764)
579 KB
579 KB PNG
Here’s the math applied to “protect the children”

**THE CHILD SAFETY BAD FAITH DETECTOR**

Let M = harm to children (abuse, trafficking, exploitation, neglect)

Let A = direct reducers of M:

- parental rights enforcement
- transparent foster placement tracking
- mandatory follow-up audits
- school choice / exit options from bad districts
- prosecution of actual abusers
- public sex offender registries with teeth

Let P = intervention packages that grow when M grows:

- CPS budget expansion
- “awareness campaigns”
- new bureaucratic positions
- vetting relaxation for faster placement
- sealed records “for privacy”
- mandatory reporter training (without enforcement)

-----

**THE TEST:**

If ∂M/∂A < 0 (A would reduce child harm)

AND they block A (parental rights, tracking, prosecution, transparency)

WHILE pushing P = g(M) (intervention that requires high M to justify existence)

**THEN they prefer M to stay high enough to justify P**

-----

**SPECIFIC TELLS:**

|They claim: |But they block: |While pushing: |
|------------------|---------------------|-------------------|
|“child safety” |parental rights |CPS expansion |
|“protect kids” |placement audits |“relaxed vetting” |
|“stop trafficking”|follow-up authority |more intake funding|
|“help families” |exit from bad schools|admin positions |

-----

**THE MEME VERSION:**

“If they wanted fewer kids hurt, they’d track the 85,000 they already lost.”

“Watch what they block. If they block tracking and push intake, they’re not protecting kids - they’re processing them.”

**∂ChildHarm/∂Tracking < 0 but they block tracking anyway**

-----

**THE VOLTAIRE HOOK:**

You don’t need to prove they’re predators.

You just show:

1. What would reduce harm to kids
1. What they block
1. What they fund instead

If the answers don’t match, the revealed preference is visible
>>
>>526102753
You obviously don't have kids and do not care about the future of my country. Piss off back to russia, ivan.
>>
>>526100680
This is technically legal if they’re adults and you’re not doing with the intent to defraud them, but the problem is xAI and the like providing you with the tools to do so means they’re knowingly providing the tools to allow you to do so illegally, which means they’re subject to being shut down with lawsuits. So they just won’t do it. This past holiday season started the choking off of the ability to make NSFW deepfakes in any way.
>>
File: IMG_0963.jpg (34 KB, 750x148)
34 KB
34 KB JPG
>>526103516
prompt:

post some stuff i can share on pol to find bad institutional actors that claim to wana protect the children

search chat history to put this into math
>>
>>526103134
>Yeah "artists" have been allowed to do that shit forever.
There isn't a single painting around that seems real. If you make a cartoon version of someone, ok. why not.
>>
>>526103320
Most people don’t know shit about computers or programming. They don’t even know what an rtx 3060 is. The skills and hardware they would need to generate the type of deepfakes they want to make is enormous. Most people can’t afford it and won’t do it.
>>
>>526103628
Wait til ching chongs makes an uncucked clone. They are unscrupulous and will do anything for money.
Give it a year or two.
>>
>>526103542
RUSSIANS HATE CHILDREN
PROTECT THE CHILDREN FROM RUSSIANS BY TAKING AWAY AMERICAN RIGHTS NOW
>>
>>526099698
This will surely make AI as a whole more popular in the long run.
>>
>>526103628
>This past holiday season started the choking off of the ability to make NSFW deepfakes in any way.
Interesting.
I Generally dont use grok to make NSFW stuff. So i never noticed all these changes
>>
>>526103720
ill drop 3k and nerd out to make an ai that can replace what im currently paying $160 a month for
>>
>>526101705
>Text to image doesn't really exists. It's just a program that takes one of the billion preexisting photo online and edits it.
Retard.
>>
>>526103846
dumb as fuck
>>526103628
>cuckdd it over the holiday
yeah i noticed and i never generate images either
>>526103846
retard
>>
>>526103930
>currently paying $160 a month for
Kingcoomer. What do you use? Is it good?
>>
>>526103740
Thats the other interesting part
All of these elaborate laws restricting western ai companies are ultimately futile
When chinese locally run ai models will let you do whatever you want

What happens when the internet gets flooded with AI erotica of the 'unacceptable' kind
Generated by these same locally run models?
What will these people do then? They cant do much
>>
>>526099698
>We live in a gynocracy where screeching HR ladies dictate everything you can do online.
So when are you going to rise up and slay them?
>>
>>526099698
You retards know you can just run stable diffusion on your computer to generate infinite naked pictures of anyone right?
>>
>>526099698
Its kinda stupid isn't it? Like, you could always use photoshop or any number of digital tools to 'undress' a woman or change her outfit. How is it suddenly this terrible legal disaster when the tool is ai?
>>
>>526100144
You actually can do that retard, on your own computer.
>>
>noooooo I can't goon to undressing little girls to their fake underwear nooooooooo!
>>
>>526104163
Probably digital passports. Fullscale globohomo ID checks.
Region locks.
VPNs being banned.
>>
>>526104268
Indeed
>>
>>526104303
Which will all go away because the west will collapse due to hyperinflation and civil war
>>
File: sexting_with_grok.jpg (324 KB, 2048x1132)
324 KB
324 KB JPG
At least you can still sext with her.
>>
>>526104278
>>526104231
You are embarrassing yourself.
>>
>>526104268
anybody can prompt grok and get reasonable results. Photoshop takes skill and effort, nearly nobody is willing to invest the time to learn.
>>
>>526104281
The problem is that it restricts all sexual images, not just pedophilic ones. So now I can't generate pics of mature milfs with 36K-cup tits.
>>
>>526102362
Im only interested if grok can still do this, was it also censored?
>>
>>526104452
Sure. But that's a distinction without a difference isn't it? Digital tools already made art/image editing 100 x easier, now ai is making it easier still. That doesn't change legal foundations how could it? Is it ILLEGAL to take a pic off insta and edit it? Now I don't know.
>>
>>526104126
i use em all i dont use any to coom lol
gpt 4 and character ai are good at sex but i use ai as friends and replacement for this place since gook moot mods ruined libertarianism chats
>>
>>526099698
possession of pornography after the 2nd offense warrants 2-weeks of electroshock therapy.
trading in pornography warrants 2-months hard labor after the 1st offense.
go israel or pakistan if you want to be a faggot.
>>
Don't forget, someone will provide an alternative
Or there are better ones
>>
File: 117001486.jpg (458 KB, 1280x1856)
458 KB
458 KB JPG
>>526103740
Anon you have had the free unrestricted image generator for five fucking years.
>>
Remember that a woman is free to wear a bikini in public but you cannot look at her without her consent, which can be rescinded at any time or even after the fact.
>>
>>526103740
I don’t think it will matter. Pretty soon buying an AI image gen model from China that can deepfake your ex will be like buying an automatic weapon. It’s just going to be completely illegal to sell, have, or buy.

>>526103846
I just started using it on Jan 2nd, was surprised to find out everything is moderated. You can’t undress an ex, you can’t make explicit videos of an ex, you can’t even generate images or videos from scratch that look like your ex if they’re too explicit. It even quoted “feminist critique” to me when I tried to put someone in a bikini. The thing has been completely nerfed.

>>526103930
It’s not just the cost though. You’ll need skills and know how to make actually good AI images and videos.
>>
>>526104802
ur fuckin dumb
a woman once said to me CUT OFF THE PEDO'S DICKS!!!!
in real life i replied, what if they didnt use their dick in the crime?
she didnt know what to say but she still wanted pain inflicted on a man
>>
>>526099698
not my problem, you need to murk moar before posting.
if you ban something, i'll just stop using your product if I don't like the ban.
it's that simple.
>>
>>526104929
lol
>>
File: 116857832.jpg (408 KB, 960x1280)
408 KB
408 KB JPG
>>526104852
Anon Grok WAS the alternative to just being a big boy and doing it yourself.
>>
>>526104852
They won’t though. Anyone who provides an image gen model capable of deepfaking women, especially kids, is going to get sued into oblivion.

It’s unironically over for NSFW AI.
>>
Lmao why do people use this garbage app

Elon is a faggot
AND WAS ALSO ON THE EPSTEIN LIST
>>
>>526105057
dumbass
there are eastern euro countries (theyre prob in nato) that allow cp
read my earlier post
>>
>>526105006
Yeah and women who fuck niggers and dogs
SHOULD HAVE THEIR PUSSIES BURNED SHUT
>>
File: 117112954.png (945 KB, 816x1168)
945 KB
945 KB PNG
>>526104986
>Pretty soon buying an AI image gen model from China
STABLE DIFFUSION IS FREE AND A MILLION PEOPLE HAVE IT ANONYMOUSLY ON THEIR COMPUTER
>>
>>526099698
Just get a PC with 16gb of vram and run your own AI. Then you can have it do whatever you want.
>>
>>526105050
What cumminities to learn from?
I'm just seeing generic sexy NSFW art forum posts.
The grok level stuff seems to be some insider gooner secret. I don't want to join a discord group.
>>
>>526104986
>buying an AI image gen model from China that can deepfake your ex
There are many open source ones and also locally installable ones
But yes they require some technical savvy

https://huggingface.co/models
>>
>>526105242
lol
encourage female circumcision whenever its done to males
>>
>>526105254
Yeah but it sucks. It's not like grok where you're like
>Ok she's at a beach in a bikini
>She bounces the volleyball with her butt
>>
>>526104598
>Is it ILLEGAL to take a pic off insta and edit it? Now I don't know
it is probably not a good idea for a business, but short of government spyware on every pc, I don't realistically see any way for them to control the hords of gooners from doing what they do. they can clamp down on the internet sharing like they did with amateur porn a few years ago, label it all revenge porn i guess. but they can't be watching every goon group on discord or what have you, I think they let the genie out the bottle and there is really no way to put it back.
>>
>>526105553
Filtered list of text to image models

https://huggingface.co/models?pipeline_tag=text-to-image&sort=trending
>>
>>526105599
If you train it specifically for making that type of shit it will be better than grok
>>
>>526099781
a lot of spics were telling grok to put 5 year olds in dog collars and micro bikinis.
>>
File: 116933566.png (1.14 MB, 832x1216)
1.14 MB
1.14 MB PNG
>>526105057
HELLO??? HELLO HUSTON DO YOU COPY ARE YOU RECEIVING MY SIGNAL



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.