[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: WanVideo2_2_I2V_03899.mp4 (1 MB, 640x480)
1 MB
1 MB MP4
/lmg/ - a general dedicated to the discussion and development of local language models.

Previous threads: >>106512307 & >>106504274

►News
>(09/05) Klear-46B-A2.5B released: https://hf.co/collections/Kwai-Klear/klear10-68ba61398a0a4eb392ec6ab1
>(09/04) Kimi K2 update for agentic coding and 256K context: https://hf.co/moonshotai/Kimi-K2-Instruct-0905
>(09/04) Tencent's HunyuanWorld-Voyager for virtual world generation: https://hf.co/tencent/HunyuanWorld-Voyager
>(09/04) Google released a Gemma embedding model: https://hf.co/google/embeddinggemma-300m
>(09/04) Chatterbox added better multilingual support: https://hf.co/ResembleAI/chatterbox

►News Archive: https://rentry.org/lmg-news-archive
►Glossary: https://rentry.org/lmg-glossary
►Links: https://rentry.org/LocalModelsLinks
►Official /lmg/ card: https://files.catbox.moe/cbclyf.png

►Getting Started
https://rentry.org/lmg-lazy-getting-started-guide
https://rentry.org/lmg-build-guides
https://rentry.org/IsolatedLinuxWebService
https://rentry.org/recommended-models
https://rentry.org/samplers

►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers

►Benchmarks
LiveBench: https://livebench.ai
Programming: https://livecodebench.github.io/leaderboard.html
Code Editing: https://aider.chat/docs/leaderboards
Context Length: https://github.com/adobe-research/NoLiMa
GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference

►Tools
Alpha Calculator: https://desmos.com/calculator/ffngla98yc
GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator
Sampler Visualizer: https://artefact2.github.io/llm-sampling

►Text Gen. UI, Inference Engines
https://github.com/lmg-anon/mikupad
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/ggerganov/llama.cpp
https://github.com/theroyallab/tabbyAPI
https://github.com/vllm-project/vllm
>>
File: 1734560768546931.gif (3.12 MB, 498x370)
3.12 MB
3.12 MB GIF
►Recent Highlights from the Previous Thread: >>106512307

--Open vs closed LLM progress and dataset efficiency debates:
>106512347 >106512375 >106512423 >106512595 >106512445 >106512517 >106512610 >106512693 >106513204 >106513240 >106513275 >106513359 >106513448 >106513451 >106513465 >106513591 >106513485 >106513499 >106513545 >106513567 >106513584 >106513595 >106513736 >106514154 >106513773 >106513803 >106513823 >106513864 >106513969 >106514038 >106514056 >106514094 >106514111 >106514143 >106514258 >106514291 >106514299 >106514449 >106514475 >106514486 >106514504 >106514556 >106514592 >106514608 >106514671 >106514710 >106514725 >106514740 >106514750 >106514765 >106514607 >106514622 >106514694 >106514888 >106514917 >106514932 >106514961 >106514979 >106515001 >106515057 >106514968 >106515015 >106515061 >106515105 >106515119 >106515139 >106515181
--ERP model finetuning with AO3/Wattpad datasets:
>106512933 >106513094 >106513181 >106513210 >106513219 >106513256 >106513670 >106513686 >106513740 >106514579 >106514614 >106514787 >106513222 >106513281 >106513313 >106513321 >106513397 >106514052
--VibeVoice TTS voice cloning and conversational capabilities:
>106515071 >106515193 >106515199 >106515236 >106515623 >106515246 >106515275
--Dataset specialization vs diversity tradeoffs in model training efficiency:
>106514377 >106514388 >106514457 >106514498
--Memory limitations in transformers vs. potential SSM improvements:
>106513859
--VibeVoice ComfyUI integration issues and VRAM requirements:
>106512932 >106513000 >106513126 >106513205 >106513235 >106513319 >106513338
--Troubleshooting erratic Harmony 20B behavior in tavern interface:
>106513566 >106513708 >106513756 >106513772
--Improving VibeVoice long audio generation quality via CLI adjustments:
>106514051 >106514086 >106514195 >106514437
--Miku (free space):
>106514325 >106514549 >106515574

►Recent Highlight Posts from the Previous Thread: >>106512310

Why?: >>102478518
Enable Links: https://rentry.org/lmg-recap-script
>>
>>106516362
>https://voca.ro/1gZ6xankFzjP

many such cases
>>
/vibevoice general/
>>
All this training data debate is retarded.
We just need to create a dataset. It will take 3 lifetimes. Think cathedral.
We need a blueprint and a chunk of digital "land" and we can get going. Craftsmen will show up if the plan is ambitious enough to move the spirit of men.
>>
>>106516395
Whatever happened with higgs, anyway? I never got around to looking into it. What made vibevoice blow up comparitively? Other than everyone wanting it because it was taken away from them.
>>
>>106516389
just like with vibe-coding in general, we don't need it to one-shot a whole project, we can just do subroutines. Those can be as compact as 2 instructions.
>>
>>106516409
It's kinda decent and voice cloning just works™ by dragging a sample audio file in a folder.
>>
>>106516395
Keep it all here retard, the hype will die down after a couple weeks and /gee/ doesn't need another fucking ai gen thread to clog it up
>>
>>106516368
did she died?
>>
>>106516250
https://voca.ro/1ogB0LHKA0bU
>>
>>106516147
>gets pretty rough even with that when you try to do a 40 minute script


What is your qualified opinion on GFG?
>>
File: unmonster.jpg (53 KB, 584x575)
53 KB
53 KB JPG
>>106516432
I was not suggesting a partition, merely commenting on the current state of this thread
>>
Who should I sample for the best mesugaki voice?
>>
>>106516499
>anon added label: Important
>>
>>106516402
if you aren't worried about compute cost the available datasets are already more then enough. just hit a 70b with common crawl and then use it to bootstrap an instruction dataset and then use that to refine it even further. the entire debate is because people are trying to do it with minimal data and minimal compute it will never really work.
>>
Interesting, I wonder if the problem with the comfyui implementation lies in the step count. higher than 10 it becomes more and more inconsistent. Now tested between wildminder and diodiogod's TTS suite, the TTS suite is far more consistent and, while still has similar issues, keeping step count below 20 seems to have some form of stability.

this test, at 10 steps, with a very background-noisy sample, honestly blows away everything i was just doing at 20-30 steps before.

it's always something with this shit isn't it? lmg and ldg anti-comfy schizo vindicated!

https://voca.ro/1aYkwUddVRDk
>>
>>106516451
>https://voca.ro/1ogB0LHKA0bU

please post the script and the ref audio

For research purposes, it is understood
>>
File: file.png (448 KB, 604x1219)
448 KB
448 KB PNG
https://x.com/TimfamousTTV/status/1964084712994951555
>>
>>106516499
https://www.youtube.com/watch?v=1w3o1VPzLuI
>>
>>106516499
Your bum mother.
>>
>>106516577
Too short.
>>
>>106516450
not yet, still being digested
>>
>>106516561
You can have the script, but you gotta go get your own audio.

https://pastebin.com/wr3ASHkN
>>
>>106516520
For me, it's more about determining the minimum amount of data and the composition required to train an LLM from scratch to exhibit basic intelligence and common sense rather than creating a model useful in practice. Random web data isn't useful for that.
>>
>>106516499
(you can easily find the full asmr audio)
https://www.youtube.com/watch?v=RIgDUDyei4g
>>
File: 1742368221974057.gif (582 KB, 540x450)
582 KB
582 KB GIF
What is the max file duration VibeVoice accepts?
>>
>>106516561
ms safety team worst nightmare (no it's not actual agi, it's porn)
>>
File: 1744935028911062.jpg (411 KB, 944x2946)
411 KB
411 KB JPG
>>106516574
grim
>>
File: LLM-history-fancy.png (1.32 MB, 7279x2969)
1.32 MB
1.32 MB PNG
Summer flood is over, I guess. Expectations for new era? Largestral 3? Llama 4.5? Will Gemma 4 be a lot less cucked and finally make to "notable"? Will DS make a comeback with R2 or have they hit the wall like everybody else? Will Kimi get a second era of dominance? Will Qwen finally go from notable to top? Will incremental updates continue or will we get a jump like with R1?
Here are the counts of tops so far:
- Meta: 3 eras
- Mistral: 3 eras
- DS: 2 eras
- Cohere: 1 era
- Kimi: 1 era
>>
>>106516631
how do you define that basic intelligence and common sense? have you tried training a model on TinyStories?
>>
>>106516520
compute will naturally get better and cheaper.
The kind of data curation that we're interested in will not happen unless we're proactive and deliberate about it (and definitely won't happen in the open, even if a CAI type org does it)
>>
>>106516451
Large or 1.5?
>>
>>106516675
just chunk the files at sentence, paragraph or other grammatical markers and run them through in batches
>>
>>106516707
Google can't stop cucking the hardest, it's in their DNA to be ultra gay
>>
>>106516707
Can you call this period the Drummer's Slop Spam era?
>>
>>106516719
Large.
>>
I just bought a A100 80gb pcie card off ebay. Had to dip into my emergency fund but I hope this is worth the pay off to run bigger models purely on the gpu
>>
>>106516718
if you think that compute is going to get cheaper and better then the discussion should be on training classifier models to generate the dataset. I had some luck training a few of my own but the compute cost to run my dataset through the classification network didn't pan out.
>>
>>106516529
Shit, I think you might be right. I was getting annoyed with it and starting to look into setting up the gradio one, but knocking it down to 10 steps from the 30-50 I'd been using actually seems to have improved the quality a lot.

I'm gonna need to keep futzing with it to make sure I'm not just imagining the changes.
>>
Even Gemini knows:
"Google's original mission statement, which remains its official mission today, is: "To organize the world's information and make it universally accessible and useful".
Based on its business practices, revenue model, and ethical criticisms, Google's operational mission is distinct from its stated mission. While the official mission represents its foundational purpose, the company's behavior reflects an overarching, inferred mission: to own and monetize the pathways to the world's information and digital activity through dominance in online advertising, search, and data collection."
>>
File: 1754895738066329.jpg (300 KB, 1440x1992)
300 KB
300 KB JPG
>>106516705
heh
>>
>>106516751
"Drummer's Slop Spam era?" If he pays me for it, I will.
>>
>>106516806
bark up their tree about it for me will ya? it's 100% something along those lines. I'm genning an entire audiobook at 10 steps and its far far far more stable, compared to earlier where most samples were speaking in hieroglyphics.
>>
>>106516817
edgelord model take 2154
>>
>>106516757
This quite possibly the stupidest thing I've read all week. 80 GB is nothing in the face of the MoE models that are the only ones worth running.
>>
>>106516828
Are you trying to push me to suicide right now?
>>
>>106516837
Do a flip, faggot.
>>
>>106516707
>retarded chart again
I'd redo it myself, but I'll let you fish for attention a while longer
>>
>>106516757
>bought a A100 80gb pcie card off ebay
good

>dip into my emergency fund
it's a hobby, don't ruin yourself over it, use money you can waste
>>
>>106516764
so, sorting piles into smaller piles?
do you think a human still needs to do the final selection of what makes it in?
>>
>>106516718
>compute will naturally get better and cheaper.
I have doubts.
>>
File: 1735278322615213.gif (598 KB, 220x220)
598 KB
598 KB GIF
>>106516757
I knew some of you were dumb, but...
>>
Is there an extension or something for SillyTavern that can automatically record memories for a character?
>>
>>106516842
I mean, as long as an emergency doesn't happen before he can replenish it, he should be fine.
>>
>>106516757
>A100 80gb pcie card off ebay.
Did you get a good deal at least? I'd pay $2k for one, tops, personally. There's a lot of 3d-printing shroud and fan bullshit that goes along with those DC cards.
>>
File: 70432 - SoyBooru.png (683 KB, 1920x1080)
683 KB
683 KB PNG
>>106516757
>>
How much money is an "emergency fund"?
It's burger hours it has to be at least 100k given the absolute state of healthcare over there, right?
>>
>>106516875
As a rule, 6 months of living expeses.
>>
>>106516757
Should have cpu maxxed with 12 channel ddr5 plus a rtx 6000.
>>
>>106516846
if you want your dataset to actually be capable of pretraining a foundation model you will have to be happy with doing a few spot checks here and there. if you want to make a really high quality fine tuning dataset then you can review every sample but it gets pretty out of hand quickly.
>>
>>106516879
He wanted to run models purely on GPU. With that thing, he can run Air Q4 probably at like 50 t/s.
>>
>>106516881
What about with a 300 year time horizon?
>>
>>106516861
I paid a bit over $10000 once taxes are included. I didn't see any for under that really.
>>
File: 70431 - SoyBooru.png (686 KB, 1920x1080)
686 KB
686 KB PNG
>>106516897
>>
>>106516897
I don't know how taxes worth with the listed prices in burgerland but that sounds like the price of a brand new rtx 6000 with more memory, similar bandwidth and more compute.
>>
>>106516897
>Imagine the amount of token you could waste on OR for $10K
>>
File: asmonsmash.gif (31 KB, 128x128)
31 KB
31 KB GIF
>>106516897
do people really
>>
>>106516889
Then just a single rtx 6000 would have been cheaper more vram and new.
>>
>>106516897
lol
>>
>>106516878
Yeah, regardless of the country, usually 5 to 12 months of living expenses depending on the job and how easy it is to replenish it.
>>
>>106516757
that's a big waste of money
extra ram is better than extra vram now
>>
>>106516897
Depending on your jurisdiction you're still in the cooloff period, try to cancel.
Search "Blackwell Pro 6000 96GB" and save yourself massive headaches and also a GPU that won't lose CUDA support nearly as quickly.
>>
>>106516897
ok this has to be a joke, a 6000 pro blackwell is literally this price, brand new
>>
>>106516934
He could save himself even more headaches by getting a 5090 and 128GB of ram.
>>
>>106516757
You still have 30 days to return lil bro.
>>
>>106516946
Nobody wants to run models at 1 token per second. Stop trying to convince others to waste money on a novelty paperweight just because you did.
>>
>>106516875
>given the absolute state of healthcare over there
You guys do realize we've had health insurance since forever and since Obama you're legally required to get it or face fines. Even if you have to be on obamacare plans the worst deductible is like $8000 on the worst tier.
>>
the most valuable use of time is figuring out our own way to cheaply manufacture hardware at home 3d printing but for chips nvidias and others markups if you actuall look at how much all the base materials and electricity would cost to make/acquire is in the tens of thousands of times if not millions in some cases the further added benefit would be no glowware or backdoors

also we can also harass cudadev to write all the support
>>
>>106516891
impossible, the rate of revolution is too high right now. it is likely to stay too high until the collapse of the system as a whole.
>>
File: 1739765788547321.gif (1.17 MB, 165x168)
1.17 MB
1.17 MB GIF
>>106516897
You're trolling.
>>
>>106516964
It's a lost cause, youtube/tiktok made it look like no one in the US had or can afford healthcare and hospitals are empty or something.
>>
The VibeVoice Latent Space demons have sent a message
https://vocaroo.com/1edmqG0nl8gP
>>
>>106516716
Not TinyStories, but I've trained a few tiny models on small amounts of other data and I have a rough idea of what could work and what doesn't.

For me the model should know how to engage in simple conversations, know what most common English words mean and how to use them, and show an understanding of cause and effect for mundane actions and events. Coding, math, trivia, reasoning can come after that.
>>
its always the patrick star IQ motherfuckers with the money to just casually drop like that
or equally the types to walk into a thread like this and lie about doing such a thing, either way funny
>>
File: 1751824386589275.png (7 KB, 481x63)
7 KB
7 KB PNG
>>106516937
true
>>
>>106516990
where'd you get this recording of me trying to get this gay model working in comfyui?
>>
https://voca.ro/1kV0jby11ih6
>>
>>106516937
>>106516934
>>106516955
Okay I admit I didn't realize the newer arch cards were similarly priced. I fucked up. I'm going to try and cancel my order before I do anything else. I am fucking sweating buckets right now. Just to be clear the "Blackwell Pro 6000 96GB" you're referencing is the same as the NVIDIA RTX PRO 6000 Blakcwell with 96GB? Once the cancellation goes through i'll order the blackwell card.
>>
>>106516998
no I confirm, it's already the most retarded guys with most money to waste
>>
>>106517010
lmao
>>106517017
you got owned
>>
>>106517017
https://www.pny.com/nvidia-rtx-pro-6000-blackwell-max-q

There are a couple of versions with some minor tradeoffs.

Good luck, I wish I could drop that much on a GPU
>>
>>106516964
>the worst deductible is like $8000 on the worst tier
I pay 100€ per year for the upgraded healthcare plan and a surgery and two weeks in hospital cost me 0€.
>>
>>106516897
You really should've gotten an RTX Pro 6000 then.
>>
>>106517041
Bro, your hospital is filled with clueless immigrants. Your healthcare is pratically free because it's worthless
>>
>>106516990
kek
>>
So is Meta really done with AI at this point?
>>
>>106516499
Aoi Yuuki
>>
>>106517082
meta is going on the rag
>>
Do any of you guys use both LLMs and cloud providers? I've been experimenting with mistral-small, devstral and qwen3-coder for awhile now locally but also making use of gemini-cli and lumo now (free versions of both). Outside of when I need to ask very "personal" questions or such I find myself wondering if llm is even worth it anymore.
>>
>>106516964
I was under impression that obamacare is what made US healthcare so expensive to begin with.
>>
>>106517082
They could make a comeback at some point since they have a lot of GPUs but as right now they are done.
>>
>>106517082
Llama 4.X rumored.

Meta is racing the clock to launch its newest Llama AI model this year
https://www.businessinsider.com/meta-superintelligence-lab-llama-4-new-model-launch-year-end-2025-8

>Meta aims to release its next AI model, Llama 4.X, by year-end, according to two sources.
>The release of Llama 4 models in April faced criticism for underperformance, some developers said.
>A team within Meta Superintelligence Labs is now trying to fix bugs within Llama 4, sources say.
[...]
>>
>>106517135
How tf did zuck let himself get this bamboozled? Its honestly one of the saddest arcs i've seen. Almost makes me feel bad for the guy
>>
File: 1746845598009999.png (537 KB, 1024x1024)
537 KB
537 KB PNG
>>106517082
>>
>>106517145
>rushing out another slop of llama 4
it's over
>>
>>106517145
>fix bugs
Odd way to say "retrain with even more slop"
>>
>>106517111
You're confused about terminology, cloud providers also run LLMs.
I don't want touch cloudshit with a ten-feet pole because I know inevitable enshitification rug pull will hit at exact moment I'll start getting used to and rely on it for daily life.
Friends don't let friends use someone else's computers.
>>
>>106517157
How does Microshaft have 150k H100s and yet not a single model to their name and rely on OpenAI and Claude for Copilot?
>>
>>106517172
WizardLM wasn't safe enough
>>
how many t/s should i be getting on a 5_k_xl quant of GLM 4.5 full with 2 3090s and 128GB of RAM? what setting should i be using?
>>
>>106517135
>>106517157
Having the most GPUs does fuck all when you only use 5% of them.

>>106517172
Microsoft focuses on Azure as a service. They're more than happy to let others develop models and make them available through their service.
>>
>>106517172
Microsoft are useless
If you didn't think that before, you definitely should after seeing that chart
Organizationally they are pure lazy, rentseeking trash
>>
>>106517149
1. He listened to safetyists instead going full Musk 2.0 in changing political climate
2. Meta kept repeating the same scheme while all other companies were innovating
3. 5% GPU utilization is no joke. 600k H100s work with efficiency of 30k
4. He put a scamming jeet in charge of his AI team(llama 4 to the moon saars!)
5. He put an actual grifter who was selling gptslop as human data(Wang) in charge of his AI team
>>
>>106517197
1 toks/s if your lucky
>>
>>106517172
Look at the state of windows
>>
>>106517206
really? because i have been getting around 3t/s with some mostly random settings on oobabooga.
>>
>>106517197
Depends if you're on Windows or Linux. I'd guess 2-4t/s and 4-6t/s for the former and latter with the correct set-up.
>>
>>106517226
i am on linux.
>>
>>106517204
that's what you get for filling your company with jeets
>>
>>106517228
Make sure to fiddle with -ot to put all layers on one GPU and fit as many MoE experts on both GPUs, it'll take some trial and error.
>>
Why do none of you guys finetune your own models for 4chan usage?
>>
>>106517205
I think the political climate favors safetyists. There's a moral panic about models telling people to kill themselves or pesos using it to make "CSAM"
>>
>>106517250
What usecase?
>>
>>106516994
>For me the model should know how to engage in simple conversations, know what most common English words mean and how to use them, and show an understanding of cause and effect for mundane actions and events.
what size model are you targeting and how much data do you feel is necessary?
>>
>>106517255
automated reply bots to keep me company at night
>>
>>106517204
>Microsoft are useless
Anon, they just 'released' top tier TTS model.
Historically MS always had good R&D but completely knee-capped by management.

My favorite video on this subject:
https://youtu.be/vOvQCPLkPt4?t=158
>13 years ago
>casually drop a 1ms latency (1000Hz) touch display prototype
>absolutely nothing comes out of it
>>
>>106517250
just go get gpt4chan tune
>>
>>106517263
I said "organizationally" on purpose.
Good humans buried by bullshit
>>
>>106517270
>trained on /pol/ exclusively
I think i'll pass
>>
>>106517242
got it to 4.3t/s average on a 1700 token output. good enough i guess
>>
>>106517303
Also if you're on an Intel CPU with P and E cores, use taskset to force it to run only on P cores, should be a decent performance increase.
>>
>>106517350
i have an AMD 3950x
>>
>>106517359
Upgrade to EYPC
>>
>>106517252
The moral panic is the same since 2010, we just moved from :
>smartphone made x kill himself
>smartphone was used for csam
to
>social media made x kill himself
>social media was used for csam
and now
>ai made x kill himself
>ai was used for csam

Honestly nothing much changed, though the new generation is clearly mentally raped by this constant panic.
>>
Based on what has happened on 4chan lately I think it will flood the internet through steg
>>
>>106517252
>I think the political climate favors safetyists.
Sure, and listening to them is retarded and suicidal.
You can do like the Chinese, just lip service to "safety" for journalistic PR while internally not caring that much.
>>
If I'm asking any questions related to coding in any way shape or form should I be sticking to coder models? Is there really any harm in me using say mistral-small instead of devstral?
>>
>>106517459
I'm using gpt-oss and mcloving it.
>>
File: file.png (147 KB, 926x1194)
147 KB
147 KB PNG
Give it to me straight, how bad are my results?
>>
File: 1742310400884650.mp4 (3.45 MB, 500x370)
3.45 MB
3.45 MB MP4
>>106516368
Is it bad that I kind of wish I had friends like you guys in real life? AI and nerdy technology-based shit in general is the only hobby I have (except a gaming I guess would even then I only played relatively niche games like no man's sky). It's bad enough I'm a bit of an autist in all ways except having an official diagnosis but not having a popular hobby makes the loneliness worse
>>
>>106517499
I have a dumb question but what are the numbers after pp (prompt processing) and tg (token generation)? Is it number of context tokens?
>>
>>106517499
Dunno but for reference I get 6 tk/s on q8 qwen 3 coder 30b
>>
>>106517545
>is it bad that I wish I had IRL friends who share my interests
No anon, that's not bad. I know how you feel. I'm pushing 30 and have no idea how to make friends IRL. I have a bunch of online friends that live too far away to visit.
>>
File: file.png (20 KB, 1191x98)
20 KB
20 KB PNG
>>106517499
Quite bad.
>>
File: file.png (883 KB, 1874x664)
883 KB
883 KB PNG
So i downloaded that sweet batch of jai cards, but I notice when I try to 'view all' cards it doesn't load anything more. Does this happen to anyone else? Do I need to troubleshoot the index.html code or could it be another issue? On Arch, if that explains it
>>
https://commonvoice.mozilla.org/en/datasets
Voice bros, look at this nice dataset that I found. Lots of clean and annotated voices for free. You can even pick an accent and age! It even has tranny voices!
>>
>>106517551
pp => number of tokens to process for the benchmark
tg => number of tokens the model will generate
So yeah pp is basically your context size and showing you how well it processes them at various lengths for the test. Mind you that's not the limit of what the model can do which is separate.
>>
>>106517582
So tg256 means generating 256 tokens? Is that on an empty context?
>>
File: 1468551288516 (1).jpg (35 KB, 456x456)
35 KB
35 KB JPG
>>106517575
Can you show off Openhands-lm-32b at full weights or qwen3-coder-30b-a3b at full weights both should fit i believe in that card.
>>
gptsovits is barely 600M (4 small models) finetuned and I think it's still superior to vibevoice 7B for pratical usage. Love to talk to my LLM with sub 1s latency https://vocaroo.com/1KC9TsZqSLag
Unfortunately I don't see it improve much more than that. The dev is lazy and between V2 to V2proplus only the vits part got slighlty better (the gpt got worse and I had to keep the V2 one). The original segmentation method was dogshit too so I added my own.
>>
File: file.png (18 KB, 1130x96)
18 KB
18 KB PNG
>>106517593
I have Q8 already downloaded. Compute is the bottleneck at 3B.
>>
>>106517545
I went to a friend's party the other day. Old friend who I went to highschool and uni with. He introduces me to one of his gun buddies and we don't really hit it off until he brings up the Visa/Mastercard situation, which I have been following. Turns out the dude has been calling them in protest too, and it also turns out he uses mass bot hosted emails to troll people in protest for things prior. He even introduced me to cock.li. I never realized he was that based until we spoke about PC stuff.
What I'm saying is you'll find some good people out there, don't go revealing your power level like an autist but have tact in bringing up tech, you'll be surprised the allies and people you can learn from around the corner.
>>
>>106517591
If I understand llama-bench correctly I believe it does keep the tests independent of each other so tg256 would be on an empty context.
>>
>>106517616
>He even introduced me to cock.li
If you don't already have dozens of cock.li accounts from before the invites you aren't very based yourself.
>>
>>106517602
>>106517575
shiet man, that's pretty sick. what do you normally use it for? also is it rented or owned?
>>
>>106517630
true, forever a newfag, but better late than never.
>>
>>106517616
i told my manager that i have an ai girlfriend
now i'm getting weird treatment
>>
>>106517499
Why is openhands 32b so fast and qwen 30b so slow?
It doesn't make sense.
>>
>>106517545
>>106517616
lets all meet up and have a massive fucking orgy
or admit that people are generally unreliable in general who will constantly and always disappoint.
> case and point - mistral 3 large
>>
Does anyone have a proper working setup for K2-0905? No matter what I try with this thing, it writes like complete shit.
>>
File: 1738843992867898.webm (3.92 MB, 1080x1080)
3.92 MB
3.92 MB WEBM
is there any websites to share datasets, specially for voice models?
please don't say huggingface, I've already gotten banned there twice
Anyways have some voice samples:
https://litter.catbox.moe/yv97n6w894ktxft8.7z
>>
>>106517575
>>106517602
what the hell? i get like 35t/s on my dual 5090 setup with that model. what am i doing wrong?
>>
>>106517886
Modelscope should be usable
>>
File: 1746749978265435.png (761 KB, 700x504)
761 KB
761 KB PNG
>>106517886
What the fuck kind of shit were you uploading to get banned TWICE? I found data sets for AI models that were full of porn in those are just fine.
>>
>>106517717
Stop using OR mystery meat models
>>
>>106517956
Personally I got banned once for hosting civitai banned loras
>>
>>106517886
how do i do vidrel?
>>
File: 1747274281685753.png (7 KB, 413x41)
7 KB
7 KB PNG
>>106517963
Next step?
>>
>>106517956
>./Misaka
>Japanese audio

How would you even use that provide voice? I thought this was primarily trained on English audio and text

t. Have yet to actually use it
>>
>>106517956
>What the fuck kind of shit were you uploading to get banned TWICE?
1st time was for a decrypted assembly dataset, 2nd time was for uploading a rip from Shotdex

>I found data sets for AI models that were full of porn in those are just fine.
Yeah, there is a lot of wild stuff they don't care about like https://huggingface.co/datasets/mirav/gurobooru but they seem to hate anything that could get them in trouble for copyright
>>
>>106518002
Meant for >>106517886
>>
PLaMo 2 Technical Report
https://arxiv.org/abs/2509.04897
>In this report, we introduce PLaMo 2, a series of Japanese-focused large language models featuring a hybrid Samba-based architecture that transitions to full attention via continual pre-training to support 32K token contexts. Training leverages extensive synthetic corpora to overcome data scarcity, while computational efficiency is achieved through weight reuse and structured pruning. This efficient pruning methodology produces an 8B model that achieves performance comparable to our previous 100B model. Post-training further refines the models using a pipeline of supervised fine-tuning (SFT) and direct preference optimization (DPO), enhanced by synthetic Japanese instruction data and model merging techniques. Optimized for inference using vLLM and quantization with minimal accuracy loss, the PLaMo 2 models achieve state-of-the-art results on Japanese benchmarks, outperforming similarly-sized open models in instruction-following, language fluency, and Japanese-specific knowledge.
https://huggingface.co/pfnet
So plamo 2 was released like a month ago but they reference a 2.1 in the paper that got a really good JPN to EN tl score. Not uploaded yet though. Anyway posting for that one anon in case he still reads the thread
>>
File: 1726209112739113.png (315 KB, 2736x658)
315 KB
315 KB PNG
>>106518002
not the anon, but voice datasets will be really useful when pic related happens, not a bad idea to save stuff
>>
>>106517276
For real. Deep blue guy has been at ms for like 2 decades
>>
File: Base Image.png (556 KB, 1200x924)
556 KB
556 KB PNG
Recomposer: Event-roll-guided generative audio editing
https://arxiv.org/abs/2509.05256
>Editing complex real-world sound scenes is difficult because individual sound sources overlap in time. Generative models can fill-in missing or corrupted details based on their strong prior understanding of the data domain. We present a system for editing individual sound events within complex scenes able to delete, insert, and enhance individual sound events based on textual edit descriptions (e.g., ``enhance Door'') and a graphical representation of the event timing derived from an ``event roll'' transcription. We present an encoder-decoder transformer working on SoundStream representations, trained on synthetic (input, desired output) audio example pairs formed by adding isolated sound events to dense, real-world backgrounds. Evaluation reveals the importance of each part of the edit descriptions -- action, class, timing. Our work demonstrates ``recomposition'' is an important and practical application.
https://storage.googleapis.com/recomposer/index.html
Samples. from deepmind. pretty neat and would be useful now that everyone is messing around with editing audio files
>>
>>106518054
this is the one thing that will oneshot me, the second LLMs can moan, it's over
>>
why is Rocinante 12B popular given it's 12B size? Isn't it brain damaged?
>>
File: 1731229694830148.png (225 KB, 398x390)
225 KB
225 KB PNG
>>106518179
same, ayahuasca didn't oneshot me, but I fear an LLM that has the intelligence and architecture to replicate the voices and personalities of my favorite characters will
>>
the scent of detergent and ozone
the smell of green tea mixed with ozone
the odd mixture of strawberries, jasmine tea and ozone
god I hate gemini-slopped models
>>
>>106518271
It's local can't you just ban those? You can't do that with Gemini proper, although telling it not to use them works sometimes
>>
File: ht14u3j3re481.png (1.49 MB, 2160x2160)
1.49 MB
1.49 MB PNG
>>106517581
>look at this nice dataset that I found. Lots of clean and annotated voices
>https://litter.catbox.moe/y6gwfuaneeni1cmt.mp3
>>
File: Base Image.png (2.94 MB, 1200x5000)
2.94 MB
2.94 MB PNG
LatticeWorld: A Multimodal Large Language Model-Empowered Framework for Interactive Complex World Generation
https://arxiv.org/abs/2509.05263
>Recent research has been increasingly focusing on developing 3D world models that simulate complex real-world scenarios. World models have found broad applications across various domains, including embodied AI, autonomous driving, entertainment, etc. A more realistic simulation with accurate physics will effectively narrow the sim-to-real gap and allow us to gather rich information about the real world conveniently. While traditional manual modeling has enabled the creation of virtual 3D scenes, modern approaches have leveraged advanced machine learning algorithms for 3D world generation, with most recent advances focusing on generative methods that can create virtual worlds based on user instructions. This work explores such a research direction by proposing LatticeWorld, a simple yet effective 3D world generation framework that streamlines the industrial production pipeline of 3D environments. LatticeWorld leverages lightweight LLMs (LLaMA-2-7B) alongside the industry-grade rendering engine (e.g., Unreal Engine 5) to generate a dynamic environment. Our proposed framework accepts textual descriptions and visual instructions as multimodal inputs and creates large-scale 3D interactive worlds with dynamic agents, featuring competitive multi-agent interaction, high-fidelity physics simulation, and real-time rendering. We conduct comprehensive experiments to evaluate LatticeWorld, showing that it achieves superior accuracy in scene layout generation and visual fidelity. Moreover, LatticeWorld achieves over a increase in industrial production efficiency while maintaining high creative quality compared with traditional manual production methods.
https://www.youtube.com/watch?v=8VWZXpERR18
From NetEase. Samples made with a finetuned LLaMA2 7B and UE5. Pretty cool especially in time saved compared to a human
>>
>>106518282
common alien tongue
>>
>>106516529
Thanks for this. I tried CLI inferencing when it came out, but it was ass. I gave the wildminder one a try since everyone here is getting good results and it was also ass. The diogod suite is actually getting the voice correct, same settings same sample for everything.

>I'm the guy that occasionally shills FishAudio here because that's the only other thing that has gotten this specific character voice right for me so far.
>>
File: boss of lecunny.jpg (254 KB, 1080x1757)
254 KB
254 KB JPG
>>106517082
Zuck is betting everything on Wang and his team of OpenAi avengers
>>
File: cumshot.webm (539 KB, 480x640)
539 KB
539 KB WEBM
>>106518444
Just like metaverse, zuck won't let us down
>>
Has layer skip been proven to be a meme?
I'm wondering how much worse the performance of a model would be if we were to skip some of the model's layers (probably the last few) when generating some of the tokens.
So you generate 2 tokens with all the layers, then 1 token with just 2/3 of the layers or whatever.
I suppose the same could be achieved with two models, like using a big model to generate half of the tokens and a small one to generate the other half or the like, but then you'd have to do PP twice, keep two KV caches, etc.
I wonder how that approach would compare to current MoEs in both speed and "accuracy".
>>
>>106518498
My guess is it's hard to train especially combined with other current memes like MoE.
>>
>>106518498
Youve read the paper right!
https://arxiv.org/abs/2404.16710
>>
>>106518482
Did she really name her cat "cumshot"? Why are women like this?
>>
>>106518282
>https://litter.catbox.moe/y6gwfuaneeni1cmt.mp3
>>
File: 1742846344740765.jpg (6 KB, 250x140)
6 KB
6 KB JPG
So I have lots of old GPUs laying around. I'm thinking of using the biggest model on my best GPU out of the extras that can fit it with an okay context (4k or 8k). Then i'm thinking of a second GPU that I just run small models on for simple tasks like simple completions and shit, just so those are fast.

Does it make sense or should I just rely all on the big model?
>>
>>106518614
>okay context (4k or 8k)
2023 called
>>
vibevoice has some serious potential, but is way too RNG
a shame they pulled the plug on releasing training code
>>
>>106516757
Bad bait, why wouldnt you just lay for API at this point
>>
File: hmmmm-Getty-Images.jpg (42 KB, 606x404)
42 KB
42 KB JPG
>>106518054
It just now occurred to me that people could use a special pipeline where they prompt in LLM, their waifu or sona or whatever responds, and then VibeVoice can "respond" to you in "their voice". A current side project of mine was trying to see if I could find two models in order to talk like specific fictional characters. I guess I should get back onto that now.
>>
>>106517577
Is that some local archive viewer?
>>
>>106518255
>popular
shilled
>>
>>106518661
They released the paper and you can vibecode your own training code
>>
>>106518839
yeah
https://huggingface.co/datasets/AUTOMATIC/jaicards
It's 190k cards but I'm struggling with accessing beyond the initial 10k.
>>
File: 1730440862487521.mp4 (327 KB, 400x600)
327 KB
327 KB MP4
>>106518744
Sadly VibeVoice is not that good, but if we ever get an amazing TTS model your idea is solid
>>
File: 1740943602781254.jpg (27 KB, 500x434)
27 KB
27 KB JPG
When are we getting VibeImage/Video?
>>
>>106519046
I was under the impression that the large version is pretty good, no?
>>
File: 1740702525541657.png (203 KB, 324x282)
203 KB
203 KB PNG
>>106519046
This has nothing to do with what we were talking about but this is probably the coolest gif I've ever seen.
>>
File: 1750792249935096.png (2.66 MB, 1195x3743)
2.66 MB
2.66 MB PNG
>>
>>106519046
>>106519102
VibeVoice is really good. this poorfag can only run 1.5B though so his opinion is gay
>>
>>106519114
>3.1 that high
really, I thought people said it sucked
>>
>>106516990
This is literally Accelerator laugh from Toaru Majutsu no Index

https://youtu.be/A-zGnLZY7BI
>>
>>106519120
Quite literally a prompt issue
V3.1 was pretrained on a larger portion of longer context (32k/128k) input so naturally it does good on long context tasks
>>
>>106516817
kek
>>
>>106516705
>never spill prompt
lmao
>>
File: 1744671111945986.gif (2.55 MB, 307x307)
2.55 MB
2.55 MB GIF
>>106519102
Current TTS model really struggle with expressive voices or uncommon accents, even ElevenLabs
VibeVoice fails to clone the christian girl voice badly: https://files.catbox.moe/4wb78y.webm
>>106519118
try to clone any of the voices here: https://litter.catbox.moe/yv97n6w894ktxft8.7z
It does a horrible job
>>
>>106518881
I just tried it, seems decent but stupidity will probably show later.
What's a better choice for a normal PC? I have 12GB VRAM and 64GB RAM
>>
>>106519118
I don't think it's quite ready to go to direct output on an LLM without pretty regular artifacts that would pull you out of it. Maybe that's something that could be fixed with a perfect voice sample, but I have to do at least a handful of generations to get a result that I'm satisfied with.
>>
>>106519140
chat is this dialogue real
>>
>>106519163
Yes, that entire anime (ghost stories) was dubbed in a parody kind of way. So most of the dialogue is just bonkers
>>
>>106519145
What? I'm confused. A normal pc is one that has a 5090 instead of a station with datacenter cards in it, right? Where is the rest of your VRAM?
>>
File: 1729206977927686.png (4 KB, 61x32)
4 KB
4 KB PNG
>>106519163
how young are you bwo?
go watch the ghost stories dub right nauw
https://nyaa.si/view/1655273
>>
>>106519109
New zero shot bouncy ball hexagon test
>>
>>106518891
Yeah, good luck with that
>>
Do any of you use frontend web guis? Trying to decide if I should use Open WebUI or LibreChat or AnythingLLM
>>
>>106519223
just the llama server one if i don't want anything overcomplicated
>>
>>106519223
Would be curious to hear anything about frontends other than OpenWebUI, which I use and wish had more feature parity with ChatGPT.
>>
>>106516626

ty, kind anon
>>
>>106519223
I know it gets crapped on around here, but I like ooba. I get the things I want (agent management, branching, reply editing, prefilling, fine-grained sampler control, spawning an API endpoint for occasional mikupad work) and none of the extra complication or bloat beyond the same stock gradio BS everyone seems to have to suffer through.
>>
>>106518271
>laser printer in the background
>>
I'm making the switch to localai tomorrow from llama.cpp. wish me luck sisters
>>
>>106518271
>the smell of regret and bad decisions
>>
File: file.png (229 KB, 2019x840)
229 KB
229 KB PNG
>>106518049
>4 shot
Yeah I dunno how useful that is. And it seems like Qwen 32B was close in equaling the model here so I don't doubt proprietary models or Deepseek and co. would blow it out of the water as is. That being said, would really want to know how good Gemma does in comparison.
>>
>>106518271
>>106519281
>anon is trying to rizz up the gilfs in the local town hall/print shop
>model is calling his bluff with sloppa
you love to see it
>>
>>106518407
also seems like setting top k to 40 helped a load for stability too. could be placebo, dunno. im sure updates will keep making it better.

>it REALLY doesn't like the word "total"
https://voca.ro/1eyiXeSZ53lW
>>
How big is danbooru/gelbooru dataset? I want to train my own diffusion model
>>
>>106519622
danbooru is like 8 million images. have fun doing dataset prep lmao
>>
>>106519622
7-8TB
>>
>>106519622
>>106519631
https://voca.ro/1AHcZsTwk5pr

>and this is with temp CRANKED to 1.76
>>
is there a model built for chemistry? (work out complex reactions etc)
>>
>>106519140
did you try 5 steps 3 cfg?
>>
>>106519688
idk if i'd trust any local model with that
>>
alright now im just chucking random thread argument posts into this thing. its great.
that said ill stop being a lazy piece of shit and put together longer samples for the characters i want to use, its impressive for what it can do at less than 10 seconds but i'll push for 20 or so from now on.

https://voca.ro/1475u7QE7C5u
>>
File: 1733691274135601.webm (2.81 MB, 852x480)
2.81 MB
2.81 MB WEBM
>>106519688
yes, there are plenty
but have fun trying to wrangle pytorch 1.8 and trying to find pip packages that have been deleted for years and having to scour obscure archives
>>
>>106519713
>1.5b full precision
why not 7b in 4 bits?
>>
>>106519713
this but make use a tech support voice clip as reference
>>
>>106519719
is this chatgpt?
>>
File: what.jpg (13 KB, 196x225)
13 KB
13 KB JPG
>>106519713
How long did it take to generate that? There's something off about it that's very subtle but that's pretty nice.
>>
The memory management for the vibevoice nodes is pretty shit tbqh
>>
>the last bartowski goof was 4 days ago
It's actually over.
>>
>>106519713
just 7 seconds of examples is enough? that's insane!
>>
File: 1747433551352332.png (284 KB, 1715x1278)
284 KB
284 KB PNG
>>106519736
>>
File: 1738469689943932.jpg (24 KB, 528x359)
24 KB
24 KB JPG
>>106519722
>>
>>106519721
>why not 7b in 4 bits?
i assume it performs slightly worse, already pretty much pushing what my 1080 can do at 20~ steps because >>106519742 it took about a minute. Gonna try to inject an additional 10 seconds of her audio and see what happens. Also the request one is taking over 2 minutes kek fuck

anyway one request coming up saar >>106519722
https://voca.ro/151kEfvAeNzA
>>
>>106519768
i meant the pic
>>
>>106519713
even the 1.5b model sounds pretty good
>>
>>106519776
that is what true Reinforcement Learning looks like
>>
>>106519774
We don't have AGI yet, but jeets have officially been replaced by vibevoice lmao
>>
File: tomb raider india.jpg (165 KB, 1024x1024)
165 KB
165 KB JPG
>>106519774
*by perform i meant speed wise, not quality. but im giving the 7b a download because the speed shouldn't be hit that hard if it fits in 8gb, and if it *performs* better at 10 steps then it doesn't matter.


>>106519782
it really does, just a bit quirky. im honestly thinking the comfyui node setup is fine, this is just pushing the limits of what a 1.5b model can really do. that, and my sample audio is always a little on the short side.
>>
CORRECT SAMPLING PARAMETERS FOR vibe voice (didn't expect 1.5b to be this good)
this is cfg 3 and 3 steps: https://vocaroo.com/1kdLXHoMP72s
this is cfg 3 and 15 steps: https://voca.ro/1mmarkCZdjlb
as you can see the second one is garbage and the first one is heavenly
>>
File: 1753721089131109.png (546 KB, 598x715)
546 KB
546 KB PNG
>>106519850
What script or comfynode are you using?
Why does lower steps have better quality????????
>>
>>106519688
Like from the ground up? Not recently. https://huggingface.co/AI4Chem/ChemLLM-7B-Chat-1_5-DPO was the last one I know, also introduced ChemBench which no one uses.
>>
>>106519865
https://github.com/wildminder/ComfyUI-VibeVoice
I changed cfg max in vibevoice_nodes.py, line 408
I think it sounds better because it has less time to be influenced by unconditioned guidance, so it resembles conditioning voice more
>>
>>106519850
>first one is heavenly
if you like noisy and muffled audio, sure
>>
>>106517163
They actually fucked up the MoE router on Llama 4. It's literally broken. Source: Discussion on the Eleuther discord.
>>
>>106519899
>fucked up the MoE router
isnt this because they switched from dense models to MoE half way through training or some shit? i forgot
>>
>>106519894
The reference audio had some noise remaining after vocal isolation https://voca.ro/1dPsI1uEvk5x
And i think it only reinforces my argument in >>106519873, in the sense that it pays more attention to conditioning at low steps
>>
>>106519850
i just tested this myself on the 7b and wtf, 5 steps and 2 cfg sounds better than 10 steps 1.3 cfg
>>
aint no fuckin way kek alright im officially sunsetting this card at the end of the month. thank you for trying, you trooper.
>>
>>106518539
She really loves be cum on and her cat was white so it was her first thought.
>>
>>106519954
i will let you decide if this was worth "534.69" seconds.
https://voca.ro/1aOEtVLUuqPR
>>
>>106519973
can you send the reference audio
>>
>>106519982
here you go saar
https://files.catbox.moe/ezd745.wav
>>
>>106519140
It struggles with female anime voices because they have such a high variance in frequency range and speed. It's not impossible to get okay outputs though.
https://vocaroo.com/1cZ4dfsTIrSy
>>
>>106519973
false alarm, it was just comfyui being stupid and not unloading the 1.5 model when i loaded large. absolutely hilariously stupid memory management over here. Many such cases! I'm getting about the same gen speed as 1.5 now.
>>
>>106520073
>comfyui being stupid
it's the node, being a wrapper it relies on its own dogshit memory management. i run 7b fine on 8gb but subsequent gens sometimes require a comfyui restart since vram piles up
>>
>>106519905
Either way, it's fucked.
>>
Is there somewhere to download audio/voice samples maybe from anime dubs? I'm having a hard time finding some high quality clear sounding samples to use with Vibevoice.
>>
>>106517256
Less than 1B parameter size and it probably won't take too much data since it's all synthetic, mostly conversational and textbook-like, even if varied. It should become capable of engaging in simple conversations and learn basic concepts under 2B tokens.

It's just a test, I'm not expecting this to end up being a useful model. I'll quietly upload the training data on HuggingFace once I'm done.
>>
>>106520193
https://voca.ro/1oEZ4LjBJB4T
>>
>>106520193
Crunchyroll + audacity
>>
>>106520193
Use something like https://github.com/Anjok07/ultimatevocalremovergui/
>>
>>106520193
just extract the audio you want
then go here https://mvsep.com/en# and select melband roformer, then unwa big beta v5e.
it's the best ive found for extracting things from anime/asmr.
>>
Still experimenting with VV, sometimes getting really nice results
I'm convinced I can use this for the VN game I've been working on and off for the last year and a half
https://vocaroo.com/17GdWIJgJgNI
https://vocaroo.com/188JE1R2Z9TD
After fiddlign with weights and settings I've gotten examples that almost sound like actual directed voice acting, really quite incredible
My VN is about a poor hopeless middle school teacher being targeted and harrassed by a group of delinquent girls in dire need of correction BTW
>>
>>106520214
oh yeah, that sounds reasonable, what model are you using as your teacher model? do you have a rough idea how many tokens it will take to get the job done? desu the synthetic data generation scripts might be more interesting than the dataset or model. have you been preparing some sort of topic list and genre breakdown? how are you planning on dialing in the ratios? is it science or just vibes?
>>
>>106519954
>GTX 1080
Always nice to see old hardware doing it's best. I have 1060 kicking around I hoped to use for tiny meme models, but I salvaged it from some retarded mining rig that was running in a barn, so it exhibits some hardware issues.
>>
>>106518482
>Just like metaverse, zuck won't let us down
contrary to popular belief horizon worlds is actually pretty popular with children. I mean it's probably not particularly profitable since those children have limited access to their parents money. But it's not a critical failure by any means. It's more or less become VRChat for children whose parents who won't let them on VRChat.
>>
>>106520372
I'm using Mistral Medium (from the free MistralAI API) for the knowledge-heavy samples (wikified English dictionary and old printed encyclopedia) and gpt-oss locally for the short common-sense conversation snippets because it generates quickly, has rigorous logic and I have more than a million conversations to process. If it wasn't clear already, this isn't meant (yet) for the average /lmg/ coomer; it's an experiment to see if the core idea can work, although some sex-related topics appeared to get through.

The topics are from a long pre-made list of common-sense factoids/statements; the model is set to generate random conversations from them in a multitude of styles, tone and detail in a way loosely similar to how a hypothetical pretraining dataset focused on conversations from the web would be.

There's nothing complex about the generation process. It's just a highly randomized conversation generation request prompt in 1 turn. The aim here is not generating deep discussions, only describing the obvious conversationally and pretraining a small model on that. I'm currently making a training test on a portion of the data, but I think in the end I'll use about 5 epochs of knowledge samples, and 1 or 2 epochs of conversations (most of the data).
>>
Oh apparently Latvia released an open model.
30B, trained on 2Ttokens and EU compliant data. So probably shit.
>>
>>106520254
>>106520228
those settings on https://mvsep.com/en seem to work better than the defaults on the https://github.com/Anjok07/ultimatevocalremovergui/ but I prefer to be able to do it locally, any anons have any settings for Ultimate vocal remover gui to improve the quality? or any other option?
>>
For me, it's Gemma 270m + local wikipedia MCP
>>
>>106520801
you can install that model locally too, it's just a bit of a hassle. i can spoonfeed you later
>>
>>106520812
i'll try to look for it online to see if i manage to do it , but I appreciate the spoonfeed anon
>>
>>106520801
>settings for Ultimate vocal remover
It's a ui that has a model set to choose from. I think demcus models are the best, try different ones
>>
>>106520804
You don't need more
>>
>>106520804
hey, probably does better than many large models on mesugaki question.
>>
>>106520801
btw i also found that after the removal of the vocals, using the adobe speech enhancer website to further process the file gives better quality outputs, less "lisp" if you know what I mean.

Example:

OG voice sample from Chainsaw man: https://voca.ro/1k9vYgDS6jri

After mvsep: https://voca.ro/1056Dg4vnFzc

Mvsep+adobe ai enhance: https://voca.ro/14MLHKRIWM0d

VibeVoice with only mvsep output: https://voca.ro/1b8OpZV6hvjp

ViveVoice with mvsep+adobe output: https://voca.ro/199TaV63dVO1
>>
>>106520886
the hotlines will be localized
>>
>>106520566
thats cool it will probably work. how are you planning to mix the data? discrete phases finishing it off on your conversations like a traditional pretrained base + post training or interleaved the whole way?
>>
>>106520804
>local wikipedia MCP
Elaborate please. I'm trying create my own RAG system, it's retrieval accuracy getting shit as the data pilling up.
>>
>>106519837
>im giving the 7b a download because the speed shouldn't be hit that hard if it fits in 8gb
pretty sure the 1.5b already takes up 8gb
i'm CPUcoping because my gpu only has 6gb
>>
>>106520969
Unquanted, the 7B takes about 21GB for me. I haven't tried quanting.
>>
>>106520895
>Mvsep+adobe ai enhance
The 3rd step is waay too stringy and muffled compared to the 2nd step.
>>
>>106520931
It's trained on semi-structured short synthetic conversations randomly interleaved with wikified knowledge from the ground-up. You can pick a format, preferably among those it's trained on, and use that directly with the "base" model. An advantage is that you can easily check out progress as training goes on.
>>
>>106520983
Part of the arch is just a regular qwen model with layers removed if I understand correctly. Can it be quanted and ran as a goof?
>>
>>106521020
nice, I am a fan of mixing in the instruction data right away. have you been or are you considering tracking its progress with the benchmarks like winogrande, hellaswag, mmlu, arc, etc?
>>
something for the macfags
https://github.com/ggml-org/llama.cpp/pull/15857
https://github.com/ggml-org/llama.cpp/pull/15832
>>
>>106521093
I don't expect it to fare well in any of those benchmarks. Putting aside that I'm focusing on fundamental knowledge and not trick/tricky questions, most real models over the course of trillions of tokens, are trained on multiple choice quizzes in a way or another, if not at least on the "train" set of those benchmarks (when not directly the "test" set). I'm not training a general-purpose text completion model.

I thought of adding conversations made from the train set of various benchmarks, but haven't had the time to do that yet.
>>
>>106521033
If it's possible to separate out the qwen model, I wonder if it's possible or beneficial to "upgrade" it to a qwen3. What does the LLM in vibe voice even do, anyway?
>>
>>106520984
really? i find the 3rd step better than the 2nd one
>>
File: 1731639568398476.png (106 KB, 1152x413)
106 KB
106 KB PNG
>>106519759
He lives!
>>
>>106516707
>Will Gemma 4 be a lot less cucked and finally make to "notable"?
nigger
you put GLM in notable
you're subhuman
>>
>>106516707
>brimmiest
sharty history revisionism
>>
>>106521311
GLM is notable.
>>
>>106521311
>only good coom model between nemo and deepseek
>not notable
>>
>>106521311
GLM is a very nice model.
>>
>>106521245
nooo, don't contaminate your training data. it will do the benchmarks just fine. its just based on the token probabilities it doesn't need to make exact completions, you just need the correct choice to be more likely then the incorrect choice. it might not get high numbers but you will still see progress. I got up to mid 40's on hellaswag by training my model on fanfiction and smut. it will give you a way to evaluate different runs, if you train on the benchmarks you will lose a valuable tool to assess your models performance.
>>
LongCat Chat Flash - is less censored but still more than deepseek.
>>
>>106521284
Lmao what a retard.
>"voice latent features and text scripts are concatenatedinto a single sequence and fed directly into the LLM. The LLM then processes this context to predict a hidden state, which in turn conditions a lightweight, token-level Diffusion Head [LTL+24]. This diffusion head is responsible for predicting the continuous Variational Autoencoder (VAE) features,which are subsequently recovered into the final audio output by speech tokenizer decoder."
It was post-trained with this new objective.
>>
File: garbo.png (112 KB, 812x378)
112 KB
112 KB PNG
>>106521342
It's already "contaminated"... it's all synthetic and instruct-like, it doesn't work like a regular text completion model trained on raw text. Without the control tokens it expects (so far mainly used to compartmentalize different document types), it outputs garbage just like gpt-oss does when you try using it like a text completion model, or probably worse.
>>
>>106521445
>protecting breast
yeah this is it, we've reached peak model
>>
>>106520983
>>106521033
There's this:
>https://huggingface.co/SomeoneSomething/VibeVoice7b-low-vram-4bit
Not sure does it even work or not. About to find out soon.
>>
Are there any particular reasons why some small models (1-3B) perform actually very well whereas others are garbage without using their larger 30B+ variants?
>>
>>106521445
oh that is an interesting approach. you could probably patch in your control tokens to make the benchmarks work. that is assuming you don't already have a objective benchmark scheme worked out. it might be worth looking into.
>>
>>106521707
I'm guessing some small models have a specific training/finetuning regimen to mitigate forgetting of important knowledge, whereas others are just trained like their larger counterparts.
>>
>vibevoice 4 bit quant
https://vocaroo.com/12RlZhrcFUk8
nyo
>>
File: .png (68 KB, 1238x642)
68 KB
68 KB PNG
>load up a bunch of text into the gradio
>click the button
>touch grass for 2 weeks
where are my fellow cpu chads
>>
>>106521763
it isn't usually quite that slow btw, i put a big ass speaker file into it which slows it down a lot. usually it runs about double that speed. not sure whether a long speaker audio is good or not, guess I'll find out
>>
>>106521777
>double that speed
So 1 week instead of two?
>>
https://voca.ro/1jmtakXNczz3
>>
>>106521735
I mean more in the way of same person wr ingl hloof ee like like jorgan soof ah see plOO plOO air pitical tweet holes and a like e whoul ets qual in a ss seriously her gosh she's plt waa zvzvzbt bzzvvv brrr angry exaggeration it just becomes high pitched and glitchy.
>>
File: moon.png (25 KB, 620x150)
25 KB
25 KB PNG
>>106521799
Very good!
>>
>>106521763
how much ram used running 7b with cpu?
>>
File: $.gif (998 KB, 250x251)
998 KB
998 KB GIF
Is Apple Silicon as good at AI as they claim? I constantly see mac fags claiming they run all these big models on their little macbooks
>>
File: .png (16 KB, 561x164)
16 KB
16 KB PNG
>>106521872
It's very large. Because when you run with CPU, it uses float32, and when you run with GPU, it uses float16. Plus all those Pytorch libraries and DLLs take up an extra few GB on top of the actual model size.
>>
>>106520451
Previously I ran an rx 580, for what it was, it performed well. But the A.I age completely killed its usefulness. the 1080 is the first card i'm 100% keeping once i upgrade. Especially since it's going to be needed as a physX accelerator.
Thanks Njudea!
>>
>>106521884
The example source is simple enough, it's pretty doable to write up some half-assed memory management to split up the model between gpu and ram if needed and whatnot.
>>
>>106521735
is that pippa?
>>
>>106521763
On the positive side, it hopefully won't need to go all the way to 19438.
>>
>>106521873
It's for people who don't care to gen at very slow t/s and be restricted to textgen because everything else runs even worse
>>
>>106521936
No.
>>
>>106517157
>>106517172
Biggest provider of inference
>>
Does anyone get the issue with vibevoice (large) that everything's fine up until exactly 30 seconds of generated audio, and then it starts stuttering (like audio glitches, as if your headphone cable was coming loose for fractions of a second)? It's weird that it happens exactly 30 seconds in.
>>
>>106521799
Good morning saar
>>
>>106521936
yes
openaudio s1 mini in comparison
https://vocaroo.com/116TlxTE6wKl
>>
>>106522008
now make a clip were she admits to having a vomit drawer
>>
>>106522008
>pipp.ai
>>
>>106522008
I'm still at work, but if you post the reference audio and text, I'll try it with the unquanted one later.
>>
File: conversation-test.png (148 KB, 812x510)
148 KB
148 KB PNG
>>106521718
My benchmark is: "does this look coherent?" At 80M tokens it's more often incoherent than not, although the underlying grammar looks like English. The text might look OK on a quick glance, but if you carefully read it you'll quickly spot nonsense. It's also very sloppy, but that's a different issue.
I guess a separate "benchmark" role could be added in a future run.
>>
Pythonshit is fun until you need to actually work with cuda and accelerate and their respective configurations.
>>
>>106522131
>still managed to sneak in a little "it's not just X, it's Y"
>>
>>106522015
it's shit but whatever
https://vocaroo.com/1iESshqxpZI5
>>106522050
i already tried unquanted
https://vocaroo.com/1w4psWbKBnX5
>>
File: steveharveymeme.png (392 KB, 327x471)
392 KB
392 KB PNG
>>106522182
>it sticks a bit
jesus fuck that was grotesque. nice.
>>
>>106522173
It has far more annoying recurring slop than that that I didn't initially notice the model was generating.
>>
>>106522182
Oh nice.
>>
>>106522131
when I started out my benchmark was to see if it would stop giving lesbians a cock to play with. it took about 3b tokens for it to start to pass the test more often then not. but at some point I felt I needed a way to measure progress that was a simple number so I could tell if my training was stalling out or over fitting. I was doing perplexity tests on a couple out of domain documents.
>>
>>106522131
What's your font? Looks nice.
>>
>>106522206
isn't that one of the problems with synthetic data, it amplifies the slop?
>>
>>106522248
Roboto Condensed
>>
How good are the top rated CAI character cards on average? I wanna nabe a bunch of vanilla cards but that site doesn't really tell you shit about tokens, you have to download the card
>>
File: cory delaminguez wink.jpg (542 KB, 1079x1127)
542 KB
542 KB JPG
>>106522315
>How good are the top rated CAI character cards on average?

stop using CAI like a queer and use something else.
https://char-archive.evulid.cc/#/chub/mr.blackwood/character/belle-delphine

https://chub.ai/characters/imperatorcreator12/Lopunny
>>
>>106522347
>>106522347
>>106522347
>>
>>106522248
It's Noto Sans Mono without font hinting.
>>106522251
I fully expected slop, but I hoped the teacher model (gpt-oss) would be a bit more varied in the responses across dozens of different types of conversations that I made it generate. I just didn't notice it until I trained a model with some preliminary data, even though I inspected the source data many times for various issues (not carefully enough, evidently).
>>106522238
I don't mind if it overfits a bit on the wording/format. At this level coherence in basic conversations is quick to test: I just hit regenerate and see what comes out.
>>
>>106521884
>It's very large. Because when you run with CPU, it uses float32, and when you run with GPU, it uses float16.
It loads into memory whatever torch type is specified in the config. Whether it is running on CPU or GPU. Whether or not the CPU or GPU support FP16 or lower precision (Ampere and earlier, for example, do not have fp16 support). So even if the tensors are in fp16 it still takes fp32 compute time.
If you don't know what the fuck you are talking about quit saying shit and spreading misinformation based on your fetal alcohol take.
>>
>>106522408
Ampere has fp16. A100 is Ampere.
>>
>>106522432
It's a shame your mother didn't have abortion support since I'm obviously fucking talking about desktop GPUs because nobody here owns a fucking A100 you absolute fucking retarded stack of latent circumcision trauma.
>>
>>106522450
Other Ampere GPUs also support FP16, retard-kun.
>>
>>106516409
higgs has very good quality and can output professional sounding audio. It's a functional TTS.

A bit more inconsistent than vibe , sometimes it outputs trash audio, and it's system prompt seems to serve the purpose of just dictating audio quality, which often times fucks with the voice clone making it sound nothing like it. Annoying. Voice cloning is more restrictive than vibe.

Also, it frequently flubs on tone or pronunciation of words where vibe would excel even on vibe 1.5b.

here's some higgs. Like its not bad, it deserved more praise when it came out, but vibe has eclipsed it.

Higgs does gen faster (smaller model, maybe bugs have been worked out)
https://voca.ro/14SyInH21SnZ
>>
File: 1739913614481537.jpg (38 KB, 601x583)
38 KB
38 KB JPG
>>106520056
I love Kuroko....
>>
>>106522450
>>106516757
Unless?



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.