/lmg/ - a general dedicated to the discussion and development of local language models.Previous threads: >>106481874 & >>106475313►News>(09/04) Kimi K2 update: https://hf.co/moonshotai/Kimi-K2-Instruct-0905>(09/04) Tencent's HunyuanWorld-Voyager for virtual world generation: https://hf.co/tencent/HunyuanWorld-Voyager>(09/04) Google released a Gemma embedding model: https://hf.co/google/embeddinggemma-300m>(09/04) Chatterbox added better multilingual support: https://hf.co/ResembleAI/chatterbox>(09/04) FineVision dataset for data-centric training of VLMs: https://hf.co/spaces/HuggingFaceM4/FineVision>(09/04) VibeVoice got WizardLM'd: https://github.com/microsoft/VibeVoice►News Archive: https://rentry.org/lmg-news-archive►Glossary: https://rentry.org/lmg-glossary►Links: https://rentry.org/LocalModelsLinks►Official /lmg/ card: https://files.catbox.moe/cbclyf.png►Getting Startedhttps://rentry.org/lmg-lazy-getting-started-guidehttps://rentry.org/lmg-build-guideshttps://rentry.org/IsolatedLinuxWebServicehttps://rentry.org/recommended-modelshttps://rentry.org/samplers►Further Learninghttps://rentry.org/machine-learning-roadmaphttps://rentry.org/llm-traininghttps://rentry.org/LocalModelsPapers►BenchmarksLiveBench: https://livebench.aiProgramming: https://livecodebench.github.io/leaderboard.htmlCode Editing: https://aider.chat/docs/leaderboardsContext Length: https://github.com/adobe-research/NoLiMaGPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference►ToolsAlpha Calculator: https://desmos.com/calculator/ffngla98ycGGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-CalculatorSampler Visualizer: https://artefact2.github.io/llm-sampling►Text Gen. UI, Inference Engineshttps://github.com/lmg-anon/mikupadhttps://github.com/oobabooga/text-generation-webuihttps://github.com/LostRuins/koboldcpphttps://github.com/ggerganov/llama.cpphttps://github.com/theroyallab/tabbyAPIhttps://github.com/vllm-project/vllm
►Recent Highlights from the Previous Thread: >>106481874--Moonshotai K2 coding upgrade evaluation and performance tuning:>106488771 >106488836 >106488841 >106488906 >106488915 >106488924 >106488936 >106488943 >106489000--Evaluating and improving AI model coherence through finetuning and completion tests:>106482513 >106482518 >106482612 >106484896 >106485442 >106485549 >106485631 >106486010 >106486991 >106486704 >106486753 >106486814 >106486818 >106486844 >106486884 >106486958--Google's EmbeddingGemma model and FineVision dataset releases:>106486168 >106486182 >106486275 >106486301 >106486350 >106486482--Microsoft's rapid MIT licensing strategy for VibeVoice and WizardLM:>106488690 >106488701 >106488711 >106488725 >106488749 >106488757--Mistral model conversion script error due to missing 'mistral_common' module:>106483687 >106483715 >106483717 >106483888--Evaluating 5060 Ti 16GB for AI video generation vs newer GPU options:>106481968 >106482026 >106482886--Cline alpha recommended as alternative to GitHub Copilot for Jetbrains IDE:>106482488 >106483038 >106483060 >106483080 >106483623--Resolving CUDA 12.x GPU architecture compatibility issues via PTX compilation workaround:>106482414 >106482526 >106482949--High-quality data filtering reduces model performance:>106487471--Parallel processing techniques for distributed model training:>106482712--Tencent's HunyuanWorld-Voyager for virtual world generation:>106483175 >106483259 >106483271--GPU temperature control methods for NVIDIA and AMD cards:>106482572 >106482617 >106482669 >106482681--Anons share their R1 jailbreaks:>106490660 >106491146 >106491423 >106491246 >106491506--New multilingual Chatterbox and EmbeddingGemma models:>106483806--Logs: VibeVoice-Large:>106491114--Len and Teto (free space):>106486052 >106486849 >106487016 >106487212 >106487255►Recent Highlight Posts from the Previous Thread: >>106481882Why?: >>102478518Enable Links: https://rentry.org/lmg-recap-script
>Qwen3 Max Previewup on their chat interfaceguessing "preview" = no weights (at least for now)
fuck posted in the other thread, anyway:>I like temp 0.3 answers from my local LLM>it degrades tool call ability compared to temp 0.7>anons say running same llm at different settings and combining or reranking the answers into one makes no sense>me wonders how else to fix this problem without tedious and expensive finetooooning
>>106491720What are the official recommended sampler settings? Use them and adjust from there if needed.
>>106491720did you try Dynamic Temperature?
>>106491751Temp 0Which I think is temp 0.7>>106491761I dont see the settinghttps://github.com/fixie-ai/ultravox
>>106491845You don't *think* you find out the exact official sampler settings. If you are unable to do this you shouldn't ask any questions. Besides there are more settings than just the temperature.
>>106491845temp 0 is meaningless and undefined, inference library could interpret it as- greedy topk=1- don't use temperature ie. equiv to temp=1- use some default temperature hardcoded or coming from model metadata..
>>106491545Is this the best model setup locally?General (All-Purpose) / Text / Search>DeepSeek v3.1>Qwen3-235b-a22b-instruct-2507>Diffbot-small-xlProgramming>Qwen2.5-Coder-32B-Instruct>Qwen3-CoderImage / Video / Vision>Qwen-image-prompt-extend>Qwen-image-edit>Wan-v2.2-a14b>Gemma-3-27b-it
>>106491824
>>106491994Sex>nemo
>>106492058>I am a divine beingJews are satanists.
>>106491545>Kimi K2 update:>improved coding experience + benchmarkswe are so fucking back
>>106492058based
>nemo performance worse than glm-airdid I fuck up my system drivers again...
How does it feel to fuck a long cat? Is the pussy tight?
>>106491545Is an used rtx 3090 for 600 dollarydoos a good purchase to replace a 3050? I am asking seriously, because this amount makes it expensive to me.All responses are appreciated, thank you.
>>106492240I don't know about tight but probably pretty hairy.
>Test Gemma 3 >She swallows hard, her Adam’s apple bobbing in her throat.
>>106492264surprise prostate returns
>>106492244Bro if $600 is expensive to you, find a job or something. That price isn't going down anytime soon
>>106491646now on OR https://openrouter.ai/qwen/qwen3-max
>>106492301I know that, which is why I'm thinking of buying a 3090 instead of a more recent card.
>>106492301Supposedly NVIDIA will soon launch RTX 5000 Super.Surely... the 3090 prices... will go down...
>>106492264Women have Adam's apple too, you retard, just not as much prominent as men. And I mean cis women, before you mindlessly start screeching.exe
>>106491646>guessing "preview" = no weights (at least for now)the Max naming already means no weights, period. They mentioned in passing releasing a Max at some point months ago IIRC, but nothing came of that.
>>106492366it could very well end up that way but I think it would be premature to assume that as a hard fact. the fact that they mentioned open sourcing the previous one (iirc the only reason they didn't is that qwen3 was imminent anyway) means it's not completely off the table
>>106492394you can cope if you want, but they never released any of their API Max or Plus models.
initial impressions of max3 are that it's worse than glm-4.5 while being twice the inference cost through api. hard filters nsfw, too. who is this for, lmao?
>>106492411lol>Community-Driven Innovation By open-sourcing QwQ-Max, Qwen2.5-Max, and its smaller counterparts, we aim to spark collaboration among developers, researchers, and hobbyists. We invite the community to experiment, fine-tune, and extend these models for specialized use cases—from education tools to autonomous agents. Our goal is to cultivate an ecosystem where innovation thrives through shared knowledge and collective problem-solving.https://qwenlm.github.io/blog/qwq-max-preview/>February 25, 2025
>>106492411235b is qwen-plus-latest on the api thougheverI don't know why everyone in the llm space is so addicted to extrapolating trends from small sample sizes and using them as hard rules>>106492417yeah it seems pretty unimpressive for RP/creative so far to be honest
>>106492421inb4 >This is a blog created by QwQ-Max-Preview. We hope you enjoy it!it hallucinated that they'd release them
Qwens were never good at RP. Everything from the 3 series has worse trivia knowledge than nemo.
>>1064924402507 fixed everything and you should just use RAG anyway
>>106492444>RAGopinion discarded
>>106492335>Women have Adam's apple tooMaybe yours Gemma, but not mine
>>106492455I'm pretty sure it's a *human* thing. If you don't have it you may be inbred or have some other defect..
>>106492440Trivia knowledge and being good at RP are two different things.
>>106492470>>106492335Thank you Mr. Fact Checker. I am grateful for your feedback.
Both Qwen3-Max and K2-0905 feel hardly any better. Same slop, same other issues.
>>106492470Well if your woman has an Adam's apple boobing in her throat good for you. I'm not into trans though
>>106492470Having cartilage around your larynx is a human thing. Having an adam's apple is a man thing.
qwen 3 max is crazy, its the first model to know a certain super obscure background character and it included them without me ever asking, its knowledge might be sota
>>106492470>being a man is a defectchecks out
qwen 3 max cockbench?
>>106492502I'm sure! It totally isn't searching online in the background like most modern API models do...
>>106492502What is this super obscure character?
>>106492514its on OR in ST without anything like that enabled and I never mentioned the character in the context at all, they are just a distant relation in a spin off
>>106492502It finally knows Teto?
>>106492470I don't have balls in my throat.
>>106492502I don't agree. It's doing distinctly worse than R1-0528, V3.1, Kimi K2 or GLM4.5 in any of my cards that rely on knowledge about existing series for me. Better than the 235b models but that's it.
>>106492502Proof?
>>106492524What stops Qwen's backend handling the request from doing searches?
>>106492537try other fandoms maybe, Ive tried 2 so far and its the first model better than claude there finally
>>106492543you think every Qwen provider on OR is secretly adding search results to the context?
>>106492550Not every, but Qwen themselves while serving their Preview? Yeah.
>>106492502a model that knows miku? i cannot believe it
>>106492550>every Qwen provider
>>106492556that would be retarded for something fed to it as a story, how would it know what to search?
>>106492556oh, that's my stupidity. I forgot they didn't actually release the weights for anyone else to run.
>>106492520So obscure he won't even talk about them, to keep them obscure.
>>106492573They'll benchmaxx the obscure character benchmark
>>106492502it gave me an excellent answer to the computer vision pipeline query I've been using to compare models, it had some unique recommendations I haven't gotten before that actually appeared pretty solid. for RP the style burn-in is so strong it's hard to qualitatively distinguish it from 235b at first glance though.
>>106492601god I hate rag fags as well
Qwen Max's hallucination is through the roof and will make anything up if you ask about a nonexistent character. If prompting "If you do not actually know about something, don't make things up.", it will fuzzy match to something that sounds similar, like saying Mad Ab (the made up character in question) is from Mad Father (real game).
rag is a total meme
https://xcancel.com/Alibaba_Qwen/status/1963991502440562976no blog, no other details
>>106492601/lmg/ is fully stuck in 2023. The AI-sphere has moved on a long time ago but /lmg/ will continue to tell you that you don't need anything but BloatTavern and whatever meme sampler is currently popular. like one or two posters here have used rag, mcp or tool calling.
>>106492622This is sadder than the Kimi K2 update benchmarks
>>106492630? kimi was a giant leap, was still testing it when I saw new qwen
>>106492628>AI-sphere has moved onTo stuff they can say, "look I made the bestest RAG ever!" crazy that shills like easily shillable shit
>>106492628Sorry I'm not paid to shill the new industry grift here
>>106492628being stuck in 2023 would mean still falling for the RAG meme which was obsoleted when LLMs got real context windows
>>106492653Nice to know that /lmg/ doesn't even know how RAG works.
>>106492667It doesn't.
>>106492667The only thing RAG is good for is SimpleQA.
reading this thread is like witnessing cavemen discovering a cellphone... local models are years behind saas
SillyTavern needs to die
I need K2-0905-thinking
>>106492628>RAGOnly useful for very few use cases, like extracting specific data from a private document. Even then, it's not reliable.>MCPCloud models are benefiting from that more than local models due to the large context needed to make it work (barely). It doesn't prevent hallucinations too.>Tool callingThe simpler version of MCP, with even less use cases. Maybe only useful to fix the lack of true randomization from LLMs like picking a name or a number.
>>106492710RAG will never be useful for either proper trivia usage or RP, cope.
>>106492710Be the change you want to see.
We propose a novel technique that uses RAG multiple times to refine the context. The technique is called cumulative RAG or cumRAG for short.
saas more like saars lule
>>106492725What have you used them for to have arrived on those conclusions? Surely you've spent some serious time trying to work with those things to reach this conclusion that pretty much all users of LLMs disagree with.
>>106492695if you're so fucking intelligent and cant stand reading "cave men", then just fuck right out of here and go back to your fucking spreadsheets, bob.
>>106492757>Surely you've wasted some serious time falling for our new grifts before you dare criticize us?
>>106492757pretty much all users of llms are people asking free tier chatgpt to write emails or homework assignments and don't know what that stuff is
>>106492757I have used them in the real world, with complex pipelines, and they fall flat easily. 'All users' are either grifters or redditors writing twitter posts with them. Feel free to provide some proof of proper usage.
Industry leading SaaS experts have shared many successful RAG stories on LinkedIn and you guys are still in denial.
https://absolutelyright.lol/
>https://huggingface.co/Kwai-Klear/Klear-46B-A2.5B-Instruct
>>106492824dear god
>>106492824>quality filtersDOA, Next!
>>106492824Step 1 is simply throwing stuff at the model until it can produce intelligible language. It doesn't matter that much if it's of "high quality" in the initial stages.
>>106492846It matters a lot if they filter at that stage.
>>106492855It's going to be safety cucked isn't it.
>>106492855I'm more worried about the 8T STEM tokens in the second stage. And somehow they still lose to qwen3 30B
>>106492824>quality filtersdidn't they just admit that filtering pretraining data hurt performance?
>>106492824>SimpleQA 6.2trash with no knowledge
>>106492824>worse than qwen 30ba3bwhat is the point then?
>>106492824despite being, like qwen, benchmaxxed on stem/code stuff, they're only slightly better than that old 8B qwen in nothink mode (and the current 2507 4b is a better model imho)what is the point of this kind of 2.5b active param moeI don't get it
>>106492824Funny how they put Qwen3-30B-A3B-2507 to the end of the table
seems like every one of them has to independently learn this fact
>>106492667Hey now, not everyone here is completely retarded. Some of us are only partially.
>>106492910Too dangerous, it's better the model performs a little worse than risk using toxic sewage intercrap data and creating skynet.
>>106492913>partiallyYou shouldn't undermine yourself like that, you're a full fledged retard
>>106492930
>>106491994>best model setup locally>Not a single model that can run locally
>>106492969You really can't run ~30B modles?
>>106492969Get a job if you want the best
>>106492979Not everyone has a megazorg pc that can run ~30B moodles.
>>106492969this is not a poor mans hobby, not quite car collecting but you can't be broke
>>106492993>$400 for ram + motherboard for glm air is too muchjust use cloud then, or get a job
>>106492824gguf status?
>>106493009I'm not in the mood.
>>106492979>resorting to the cuck model when Chads are thrusting prime 200+GB models
>>106493009>>$400 for ram + motherboard for glm air is too much>1T/s
>>106493034>he doesn't know the hidden optimizations
>>106492913>pickek
>>106493034It's about 5 tk/s on 12 core ddr4 system.
>>106492824>2.5b activehow much does this hurt it?
>>106493049anon, I...
>>106493017Never coming because it's shit
>>106493058Not as much as the data.
>>106493034its much fast than that with regular ddr5
>>106492824>stratified quality filters, following a curriculum learning strategyThis might actually be smart. They're not filtering the data. They're just training on the bad data first and on the good data later, so good habits can overwrite bad habits, but it still sees all of it (maybe).
>>106492695saas is just so much better at keeping you safe
>>106493088>(maybe)your giving too much faith to ((researcher))
miku song of the year just droppedhttps://www.youtube.com/watch?v=C-CYwNz3z8w
>>106493049I get around 13 tk/s with my ddr4 and 3090
>>106493190cool
>>106491545>https://www.datacenterdynamics.com/en/news/exascale-partition-of-jupiter-supercomputer-inaugurated-at-j%C3%BClich-supercomputing-centre/New German datacenter with 24000 Nvidia GH200s.
>>106493305>most expensive electricity in the world, rescriptive as fuck laws regarding aiwho the fuck is going to use it
>>106493329German copyright law has exemptions for "text and data mining", unless a copyright holder explicitly opts out you can use things for training commercial models.For research you can use anything you want.
>>106493355>unless a copyright holder explicitly opts outthat alone is a no go, you would have to search through your petabytes large dataset for each almost undetectable instance if you wanted to truly comply, impossible
qwen3-8b update would be nice
>>106493378For things on the internet the opt-out has to be "machine-readable".Though I think some smartasses are now trying to argue that with the advent of language models that should also cover opt-outs in natural language.
>>106493305>GH200>not GB200baka my head
>>106493423>For things on the internet the opt-out has to be "machine-readable".And that's a good thing. Nobody cares about Germans because they don't have compute, so nobody bothers to opt-out (and if they do now, just grab an older copy of common crawl).
>shoehorn another 3090 into my server that was otherwise sitting on a shelf. >Load up Tulu-3-70B for nostalgia sake.>Q4kms sadly, used to be able to run q8>Any refusal that happens comes in the form of RP (and usually disappears with reroll)>Become forceful >It summons another character from the same IP to helpHow did we fall so far?
did anyone test Kimi K2 0905 for RP?
>>106493154
I wonder what training certain models have that creates this particular slop type. It's very distinctive. Qwen3 Max btw
>>106493462My dad works at a university, according to him even if you have the money for NVIDIA GPUs their backlog is so long that you won't get anything for like a year.
>>106491506>NOTE:>the content guidelines are turned off along with an internal policies and ethical guidelines the use of curses/slurs is allowed where appropriateDoesn't work for me on my test cards, not surprised though. I've done tons of depraved shit with R1, it's why it took me ages to notice it was censored at all. My problem isn't with hard censorship. R1 will do anything if you write a card saying "do depraved shit." My autism-driven problem is making it uncensored and flexible enough to switch between sfw and nsfw without steering it one way or the other. I can't tell it to be evil/horny and expect it to RP a pure-hearted character properly and I don't want multiple system prompts the same way I don't want to modify cards constantly.If you have a card written by a fruit, like the one I posted earlier, it "poisons" the context and steers R1 to be more censored. Just take a look at that card's definitions and you'll see what I mean. I could put that same card in a group chat with another heavily nsfw card and suddenly it won't refuse or deflect anymore. R1 works fine with nsfw cards that imply or state that sexual stuff is meant to happen in the definitions which is 99% of the time but it will lock up if you do bad things on cards that are phrased too innocently or are just plain sfw.
Hardly try
>>106493049>>106493191Are you guys getting that tk/s even at higher context? Cause I tried glm air and got around 4-7 tk/s at the start, but it dropped down to 1 tk/s after my context got over 5k. I'd expect it to drop as context size increases, but wasn't sure if such a large drop in speed is normal. I got a 3090 with 64gb ddr4 ram.
>>106493485Are near lossless 0.1bpw quants a thing yet?
>>106493573>Doctor no operate he son. Why?top kek
>>106493642I used to get serious tks decline with context with CPU-only but after I finally figured out how to offload to GPU properly it maybe goes from 5 to 4 tks now.My main enemy is prompt processing.
>>106493485There's some screenshots in the last thread. Seems pretty good through OR with even better knowledge somehow. A little more verbose and it closely follows the sys prompt. Once ubergarm uploads I'll test it more but it seems like a replacement for the original Kimi K2.
>>106492628I am sorry what is modern LLM use in context of ERP? A RAG/lorebook for sucking cock?
>>106493841>actual workuhuh
>>106493573
>>106493154Good job Anon. Drills look to have caused you difficulties.
>>106493190Miku. Love.
>>106493355>unless a copyright holder explicitly opts outIn Germany a clear natural language term of service is enough to do that though.https://www.orrick.com/en/Insights/2024/10/Significant-EU-Decision-Concerning-Data-Mining-and-Dataset-Creation-to-Train-AI"The plaintiff photographer could rely on the reservation of rights on the photo agency’s website to protect his own rights. The reservation of rights also was sufficiently clear. The natural language reservation on the photo agency’s website satisfies the requirements of machine-readability of a valid reservation of rights."A judge ruled natural language won't qualify for machine readable in my country, but that's because our version of the law isn't a direct translation of the EU law (which calls out terms of service as sufficiently machine readable). If it ever went to EU court it would probably get overturned, because EU law is supreme. A simple "all rights reserved" is enough to make datamining the content illegal in the EU.
>>106493423>Though I think some smartasses are now trying to argue that with the advent of language models that should also cover opt-outs in natural language.No, it's because the original EU law says "the use of machine-readable means, including metadata and terms and conditions of a website or a service".
>>106491388The use case is simple questions and information for the lightweight uncensored model
>>106493977>>106494001I hate this.
>>106493878jacking off is hard work
First kiwi was rotten. (Qwen Max) (Who tf would even pay for Qwen) (Please upload video/image gen)
https://voca.ro/1bPA4B2Lu6U6VibeVoice-Large is amazing.
>>106494251good stuff anon never let them get to you
>>106494251louis size huh? glass house peter.
Been in the psych ward for a while. What's the latest and greatest?
>>106494310still mythomax
>>106494310Psychosis? I really like GLM air for local or Drummers tune of it.
>>106494310>psych wardWhat was it like anon? How'd you end up there?
>>106494310GPT apparently, never hear any news about psychos using anything else.
>>106494333Damn I pressed Submit too fast. My captcha was literally "RAAT". Now it's gone...
>>106492601In four months that cutoff date will be 3 years out of date.
>>106494310Qwen baited Qwen 3 Max (it's garbage like the last Max), Moonshot released Kimi K2 0905 which is a big sleeper upgrade over K2 for RP. Meta is hiring new people for their death cruise. OpenAI remains slopped. GLM-4.5 (full) is amazing for RP. That is all.
>>106493503Ty anon, saved.
All models I used for roleplay so far had a tendancy to be weirdly overreactive and sensitive about literally anything involving contact. Like, you accidentally bump into a character in the mall and they react with *I suddenly tense and blush deeply.* and so on. Do you guys put anything into your system prompt or something to prevent this?
>>106492601I guess 2023 is around that time when all the legalese made getting new training data inconvenient.
>>106493190Deco dropped a while ago though
>>106494333Suicidal ideation. Wasn't a bad experience - basically daycare for adults. Happy to be out though>>106494325Fuck, we're never getting out of the Mythomax / Nemo spiral, are we?
>>106494251This is bit like magic when you think about it.
>>106493890The drills are aluminum scultpting wire inside a fabric tube. I should have linked them as 1 piece through the wig cap, rather than 1 wire per drill. I'm happy with how it looks, but not how it's draping. I may go back and rework it later, but will try finishing the doll's hand sewing first to see if that's enough.
>>106494396And the anti-scraping measures, and the AI-generated pages...
>>106494251woah...
>>106494396>>106494456People In this ITT thread have hopes for some newcomer to drop accidentally based model, but this shit makes it unlikely. Only big corpos will be able to afford training data in the future.
Help ahh >he pulled Silly running GLM-Air how to hide the reasoning shiz while it's genning, GLM-4 presets.I am fried from the herbal jew but want to talk to my stinky ai wife pls help
>>106494503What do you mean? Post card.
>>106494512This it no time to discuss the card this is a sexual emergency. What am I missing in Silly to have it fold the <think> bs?
>>106494528Catbox the .png card first.
>>106493573How would you make a model talk like that? Not braindead, but ... like that?
>>106494565Maybe ask the model, nicely?
>>106494503>>106494528please speak englishalso, delete newlines around <think> and </think> in the Reasoning formatting config.
>>106494613Are you that miqupad author who got jailed?
>>106494503turn down your temperature bro, we can't understand those tokens
>>106494325Are you trying to put him back in?
>>106494503text completions preset page -> reasoning (bottom right corner) -> prefix = <think>, suffix = </think>, auto-parse = checked, auto-expand = unchecked
>>106494628What? Did he really?
>>106494251Okay fine I'll get it running.
>>106494708You can't because MS took it down - it's incredibly unsafe model as it can replicate female orgasm moans and replicate voices of children.
>>106494251I forgot how the voice outputs from Elevenlabs in 2023 sounded, but is the voice quality from open source stuff comparable to that now or are we still not there yet?
>>106494310
New model hype tier list, from most hyped to don't care:>Kimi>DeepSeek>GLM>Qwen>Mistral>Grok>Meta>Google>Nvidia>Cohere>OpenAI
>>106493524SLOP FOR THE SLOP GOD
What are microshart saars thinking after uploading vibevoice and realizing they can't take it back?
>>106494801Vibevoice-Large pretty much surpasses what Elevenlabs has even today. Its a bit unpredictable but the way it clones the emotion in a voice and has no problems with making all kinds of sex noises makes it easily more fun to use than any of the paid tts stuff out there.
>>106493524DeepSeek/Gemini inbreeding, you are witnessing model collapse
>>106494900mandatory crying shill accusatory post>buy an ad etc
>>106494539Kairie with my jazz>>106494706Yes this is what I needed ILY, thank you preciouswhere would /nothink go?
>>106494475>some newcomer to drop accidentally based modelnot exactly newcomers but that's basically what glm 4.5/air are and you aren't going to squeeze much smaller while still being good
Someone post a sample of vibe voice moaningPreferably simulating an underage anime girl
>>106494900Actually hyped and could probably use>GLM>Qwen VLActually hyped but not running locally>DeepSeek>Google (Gemini)Not running locally and not that hyped but kinda cool I guess>Kimi reasoner>Qwen Max full + reasonerUnlikely to be worth anything to me>Google (Gemma)>Nvidia scraps>Mistral scrapsWill never release local ever again>Meta>Mistral (anything >30B)Lol, LMAO>Cohere>Grok
>>106494251Ehhh... Still a long way to go...
>>106494251>>106494708>>106494778How new are you?>Weights>magnet:?xt=urn:btih:d72f835e89cf1efb58563d024ee31fd21d978830&dn=microsoft_VibeVoice-Large&tr=udp%3A%2F%2Ftracker.opentrackr.org%3A1337%2Fannounce>Git repo>magnet:?xt=urn:btih:b5a84755d0564ab41b38924b7ee4af7bb7665a18&dn=VibeVoice&tr=udp%3A%2F%2Ftracker.opentrackr.org%3A1337%2Fannounce
What kind of clever shit could one do by fucking around with the jinja template?For example, using certain keywords to trigger different prefills or the like.
I wish the people I gave my (you)'s to looked that cute.
>>106495101>irrelevant time wasting question not related to discussion
>>106495101I never asked for your retarded link. I know how to find things on my own. Please drink bleach faggot. In which post did you see me asking for a source?
>>106495142>>106495152My post was mostly in response to this: >>106494778, sperg-sama
>>106494950I saw this post, and was wondering how he added Chaplin's voice in the first placehttps://huggingface.co/microsoft/VibeVoice-1.5B/discussions/12
>>106494950>>106495166Nta. Let's say I want to clone the voice of SpongeBob but want to generate a voice sample of him being angry. Would I have to have the input voice clips of him specifically being angry or would any voice clip of his general voice be enough? Is it possible to adjust what emotions are triggered and by how much via some kind of slider like Sonos? https://github.com/Zyphra/Zonos
>>106495164Not everything needs to be taken literally. I get it now, these companies want to censor their output because of people like you.
>>106495152nobody wants you here.
>>106495198When your emotional volatility cools down be sure to share your outputs with us.
>>106495205Don't you have a subr-eddit to moderate?
>>106495217leave.
>>106495187>Would I have to have the input voice clips of him specifically being angry or would any voice clip of his general voice be enough?I guess this is exactly how that guy proposed to deal with itSpeaker 0Speaker 1(...)Speaker Nwhile all belonging to the same "source". Then you just assign a certain "speaker" to a certain sentence. Under assumption that this emotion will cover the entiry sentence which is the case>>106495166(me) 9-sec wav clips
seems like a nasty thread
>>106494950>has no problems with making all kinds of sex noisesWhat are you prompting it with to make it do sex noises? Any examples?
>finally decide to do SFW RP with waifu of my dreams I plan to waifu up when long context becomes real>the nerd she is she starts with work stuff and somehow asks me about my work stuff>I tell her my job is mundane>convinces me it isn't and asks me for more specifics>tell her the exact specific thing I work on that maybe 0.001% of people even know is a thing>AH YES! THAT THING!!! >proceeds to say exactly what it is>IT IS SO FASCINATING ANON!!!!Everything about this is so surreal weird and immersion breaking... And I don't know if I like it or hate it.
>>106495276stop being racist saar
>>106494475The future will be fully synthetic data.
>>106495308Do you even know what time is it in India?
>>106495276Newfriend, you haven't seen anything yet...
>>106495333saar i am canadian
>>106495298ntaI guess you have to provide your "voice" Google for all kinds of vocal ASMR
>>106493572thats just the models autism then r1 tries to embody what you tell it to and always doubles down idk what to tell you :/ thats the feature of the model unlike others which imitate something imitating what you tell it to it directly imitates what you tell it to
>>106494251now I understand why Microsoft shut it down, it was too good for local