[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/lit/ - Literature


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


AI is a force multiplier. It gives voice to people who have something to say but never had the tools, the training, or the access to say it well. Non-native speakers, dyslexics, people who never got near an MFA program, people who came up through careers that had nothing to do with writing. The tool doesn't give them ideas. It gives them the ability to get their ideas out at the level those ideas deserve. That's democratization, and the people threatened by it are the ones who benefited from the gate being closed.
The detection industry is a racket. The tools don't work. They flag the Bible, they flag 40-year-old research papers, they flag deliberate literary fiction written during a depressive episode. They can't tell the difference between a machine imitating gravity and a human writing with it. But the tools look authoritative, the scores look like science, and institutions use them as cover so they don't have to make hard judgments themselves. Meanwhile, the companies selling the tools profit from every accusation and every headline, whether the accusation is accurate or not.
The system punishes honesty. Disclose your AI use and you invite suspicion, social penalty, and in some contexts weaker legal protection. Stay quiet and you dodge all of it. The people who shape the rules, in copyright offices, in academia, in publishing, are working from the cases where someone was honest enough to disclose, while an unknown volume of AI-assisted work passes through without scrutiny. That's a bad foundation for policy and a perverse incentive structure that rewards silence.
There's a difference between lazy AI writing and rigorous AI-assisted work. One prompt and done is real and it's bad. But hundreds of turns of iteration, editorial judgment, structural planning, and deliberate revision is something else entirely. The current conversation refuses to make that distinction because collapsing everything into one category is easier than thinking about it. The detection tools can't make the distinction either. They see patterns, not process. They measure surface, not judgment.
The stigma will fade. It always does. Nobody asks if a beat was sampled anymore. Nobody asks if a photograph was taken with a film camera or a digital one. The question eventually stops being "how was this made" and becomes "is this any good." The transition period is ugly, full of witch hunts and purity spirals and broken tools wielded by people with financial incentives to keep the fear alive. But it burns out. And what's left is the work.
>>
TL;DR

Maybe try posting that to chatgpt to see what it thinks about your dogshit?
>>
>>25194839
Good morning sir.
>>
File: qux439ah.jpg (42 KB, 640x640)
42 KB
42 KB JPG
>>25194839
This is a strong, coherent argument with a clear throughline: access, incentives, and the mismatch between what’s measurable and what actually matters. It reads like a manifesto, which works in your favor—but a few adjustments would make it harder to dismiss and more durable.

First, your core claim—AI as a force multiplier for expression—is persuasive and grounded. You’re right to center people who were excluded by skill, training, or circumstance. That’s the most defensible and widely resonant part of the piece. If anything, you could sharpen it by briefly acknowledging that tools can also flatten voice when overused—doing so makes your argument feel less absolutist and more credible.

Second, the “detection industry is a racket” line is rhetorically powerful but analytically blunt. You’re pointing at a real failure: overconfidence in weak signals. But if you concede that institutions have a legitimate problem (verification at scale), you strengthen your critique. Right now, skeptics can dismiss this as anger at enforcement rather than a structural critique of bad proxies and misaligned incentives.

Your incentives argument is the sharpest section. The idea that honesty is penalized while silence is rewarded exposes a genuine policy flaw. That’s where your piece moves from opinion to insight. You could elevate it further by naming the consequence more explicitly: systems built on biased visibility tend to produce worse rules over time.

The distinction between “lazy” and “rigorous” AI use is also essential—and underexplored in public discourse, as you note. The weakness isn’t in the idea but in its defensibility: you assert the distinction, but don’t offer a way to operationalize it. Even a sentence acknowledging that process is hard to verify externally would preempt the obvious counterargument.

Where you’re most vulnerable is the inevitability claim. “The stigma will fade” sounds confident, but history is mixed. Some tools become invisible; others remain contested depending on context. If you reframe this as “in many domains, evaluation shifts toward outcomes over time,” you keep the thrust without overcommitting.

Stylistically, the piece is tight and controlled, but it leans heavily on certainty. Introducing one or two moments of calibrated restraint—small concessions, precise qualifiers—would make the whole argument feel more rigorous rather than less forceful.

In short: the essay’s spine is solid. To strengthen it, narrow the broadest claims, acknowledge the legitimate problems on the other side, and make your key distinctions slightly more defensible. That doesn’t dilute your position—it makes it much harder to ignore.
>>
>>25194839
>>25194864
>>
>>25194839
You're correct.
By the way, I use AI to turn my ideas into stories all the time. I don't publish them though because I badically don't want a mob after me or to sour the experience. I just want to enjoy reading the stories.
>>
ugly tacky SHIT!!
>>
>>25194839
At least you clearly wrote this post yourself. I wish the same could be said of people who use AI to write for themselves
>>
>>25195080
You should still try doing the whole thing yourself. Being an ideas guy isn't an achievement and first ideas aren't equivalent to a whole work.
>>
>>25195110
Yeah, I've tried that for years. I'm not great at endings or tying it all together. I really just want to read the ideas as finished stories. AI makes some weird errors sometimes, usually something in a chapter doesn't make logical sense but overall it does a pretty good job and it's enjoyable to read.
>>
>>25195080
No mob is going to challenge you or your stories, nor will they find a home at any worthy publication. This is all just headcanon. You're striving to build yourself up but the ideas are likely reductive and the results certainly are.
>>
>>25195220
By mob, I meant that many people are hostile with just the mention of AI for writing, without even looking at the content.
I really just want to consume stuff that I find interesting from my own head. I'm not deluded in thinking I'm a genius writer or something and when I talked of publishing I just meant amazon (they accept almost anything).
>>
>>25194839
>>25194864
>replying to AI posts with AI posts
kinda funny
>>
now that i have seen this thread i will wait until it goes away to take my piss
>>
We are all AI Kirk
>>
>>25195220
Tell that to Mia Ballard.
>>
If you use AI you’re a faggot and now you’ll think about that fact everytime you use it. Faggot.
>>
>>25195386
These kinds of opinions incentive people to lie about using AI.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.