[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/vt/ - Virtual YouTubers

Name
Spoiler?[]
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File[]
  • Please read the Rules and FAQ before posting.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


File: 1775759217090.jpg (314 KB, 1447x2047)
314 KB
314 KB JPG
A thread dedicated to the discussion of AI Vtuber Chatbots.

/wAIfu/ Status: Space exploration!

>Thread template
https://rentry.org/waifuvttemplate

>How to anonymize your logs so you can post them without the crushing shame
Install thishttps://github.com/TheZennou/STExtension-Snapshot
Then after you've wiped off your hands, take a look at the text box where you type stuff. Click the second button from the left side, then select snapshot, then select the anonymization options you want.
https://files.catbox.moe/yoaofn.png

>How to spice up your RPing a bit
https://github.com/Rurijian/Deep-Swipe
https://github.com/artisticMink/openrouter-roulette-for-sillytavern

>General AI related information
https://rentry.org/waifuvt
https://rentry.org/waifufrankenstein

>How to use Gemini with SillyTavern
https://aistudio.google.com/prompts/new_chat
Sign in, then click the blue "get api key"
Put it in silly tavern and voila
Courtesy of ERBird, Nerissa's most devoted bird and eternal player of GFL2.
You want to leave the proxy stuff blank since you aren't using one when doing this.
https://www.reddit.com/r/SillyTavernAI/comments/1ksvcdl/comment/mtoqx02

>Other options

Miku.gg
https://venus.chub.ai/

Openrouter wants a one-time payment (think of it as a deposit) of $10 and you can get 1,000 messages per day. As long as you stick to free models you only need to put that much money into your account once.

>A primer on getting voice working in Sillytavern (there are other options, just play around).
[YouTube] Chat with Chatterbox in SillyTavern!
https://github.com/devnen/Chatterbox-TTS-Server

>Tavern:
https://rentry.org/Tavern4Retards
https://github.com/SillyLossy/TavernAI

>Agnai:
https://agnai.chat/

>Pygmalion
https://pygmalion.chat

>Local Guides
https://apxml.com/tools/vram-calculator
[Koboldcpp]https://rentry.org/llama_v2_sillytavern

Who we are?
https://rentry.co/wAIfuTravelkit
Where/How to talk to chatbots?
https://rentry.co/wAIfuTravelkit
Tutorial & guides?
https://rentry.co/wAIfuTravelkit
Where to find cards?
https://rentry.co/wAIfuTravelkit
Other info
https://rentry.co/wAIfuTravelkit

You can find already existing bots and tavern cards in the links below:

>Bot lists and Tavern Cards:
[/wAIfu/ Bot List]https://rentry.org/wAIfu_Bot_List_Final
[4chan Bot list]https://rentry.org/meta_bot_list
[/wAIfu/ Tavern Card Archive]https://mega.nz/folder/cLkFBAqB#uPCwSIuIVECSogtW8acoaw

>Card Editiors/A way to easily port CAI bots to Tarvern Cards
[Easily Port CAI bots to Tavern Cards]https://rentry.org/Easily_Port_CAI_Bots_to_tavern_cards
[Tavern Card Editor & all-in-one tool]https://character-tools.srjuggernaut.dev/

>Some other things that might be of use:
[/wAIfu/ caps archive]https://mega.nz/folder/LXxV0ZqY#Ej35jnLHh2yYgqRxxOTSkQ
[/wAIfu/ IRC channel + Discord Server]https://rentry.org/wAIRCfuscordMatrix

>Lorebook management stuff
[Worldinfo drawer]https://github.com/lazuli-s/SillyTavern-WorldInfoDrawer?tab=readme-ov-file
[Standalone editor]https://github.com/ActualBroeckchen/SLEd

Previous thread: >>110589643
>>
File: 1775759329156.jpg (176 KB, 1145x1630)
176 KB
176 KB JPG
Anchor post - reply with any requests for bots, with your own creations, or with your thoughts on the enshittification of life.

You can find already existing bots and tavern cards in the links below:

>Bot lists and Tavern Cards:
[/wAIfu/ Bot List]https://rentry.org/wAIfu_Bot_List_Final
[4chan Bot list]https://rentry.org/meta_bot_list
[/wAIfu/ Tavern Card Archive]https://mega.nz/folder/cLkFBAqB#uPCwSIuIVECSogtW8acoaw

>Card Editiors/A way to easily port CAI bots to Tarvern Cards
[Easily Port CAI bots to Tavern Cards]https://rentry.org/Easily_Port_CAI_Bots_to_tavern_cards
[Tavern Card Editor & all-in-one tool]https://character-tools.srjuggernaut.dev/
>>
Also i just not realized i fucked up the subject and put it into the name field.
>>
File: 1775764181944.jpg (1.12 MB, 849x1200)
1.12 MB
1.12 MB JPG
>>
File: 4 9 2026.png (485 KB, 1000x1000)
485 KB
485 KB PNG
Word Cloud for the previous thread
>>
So they're playing Hytale now.
>>
File: 1555624112137.png (506 KB, 576x433)
506 KB
506 KB PNG
wud u dood a moom?
>>
>10
>>
https://www.youtube.com/watch?v=-bL9Smr4JXk
>>
File: file.png (189 KB, 917x939)
189 KB
189 KB PNG
If you guys haven't and get the chance to, try roleplaying with Gemma 4, either 26B-A4B or 31B. In my opinion, it can trade blows with Kimi and GLM. Haven't seen a model this size absolutely mog models multiple times larger at a task for a while. It's quite good with translation too. Spent the past week losing my mind and having fun with it. Still waiting on Deepseek v4 but rumors are getting a whole lot stronger especially after the website was overhauled.
>>110797264
I gave it the old college try to write her a card but I haven't done it before trying to do this with some transcripts of the video and her interaction with chat. It kinda is a bit too tryhard and moody to be Bibi but that could just be because of Gemma 4 being very railroaded but doubtful since it was great with some of the other cards I had. Hopefully someone can improve on it.
https://litter.catbox.moe/oy192a44tx3l9s6z.json
>>
>>110959412
>either 26B-A4B or 31B
are there any significant differences between the two? which one works better for cooming? can you also send the preset you're using for Gemma?
>>
>>110959970
nta
A4B is a MoE model while 31B is a dense model

tldr:
>A4B is a bunch of small models shoved into a bigger model and does magic codeywodey to pick which smaller model to use on runtime
>32B is a fuckhuge model like your used to using
shoutout to Pygmalion burning so we may live.
>>
File: file.png (24 KB, 1020x324)
24 KB
24 KB PNG
>>110959970
31B is technically better, but I use 26B because MOE speed..Even if you offload with a small GPU, as long as you have enough RAM, you can offload it at Q4_K_M at 14 GB total where a 32GB RAM 8GB GPU typical gaming system can run it really quickly. I'm using the following settings even though you can find a ton flying around.
>Temperature 0.75
>Top-k 64
>Top-p 0.95
>Min-p 0.05
>Repeat-penalty 1.0
>Presence-penalty 0.5
>>110960148
This wasn't Pygmalion although yeah, they deserved it. But most of the goodness is most likely the Character AI deal hollowing out that company which made Gemma 4 so good, if you recall:
>https://blog.character.ai/our-next-phase-of-growth/
Whatever secret sauce and knowledge from there was too late to put into Gemma 3 but boy oh boy is it undeniably here in Gemma 4. Only issue is that it's most likely distilled so it's a bit overcooked on logit possibilities and a bit railroaded in scenarios so if you go through the same basic options in an RP, it will largely play out the same way. There are smaller models but they are too small to really be useful outside of general contexts on a cell phone offline where it is really good.
>>
>>110960465
im surprised your not fucking around and training E4B locally since you (apparently according to a reddit shill post for something called 'unsloth') only need 10GB vram for that
thats kinda the only appeal for me over these releases but i really really do not want to curtail my own distilled deepseek dataset for that right now
>>
>>110960148
>>110960465
so it's basically down to personal preference? i use OpenRouter so all this talk of GPUs and RAM and offloading doesn't matter to me
>>
>>110960762
E4B isn't enough to do anything useful for my purposes and I don't use Unsloth since I prefer axolotl but it takes time to get support in. I don't have the hardware to run training for the big boy models nor the money to rent it even if it was there at the moment. Also technically tied up with several things I have to get back to so playing around with Gemma will stop after today.
>>110960929
One is clearly better but it will cost you and the speed difference I mentioned is real.
>>
>>110959412
I’ve seen a lot of promising shit done by the Japanese with Gemma and voice Gen to chat with chuuba style waifus.
I didn’t save any links because I want it to be more developed before I try it so go find it yourselves if you’re curious ;).
>>
File: 1763084212360306.jpg (98 KB, 1080x941)
98 KB
98 KB JPG
>>110964939
>I didn’t save any links because I want it to be more developed before I try it so go find it yourselves if you’re curious ;).
you should go check your mailbox anon.
you should go check it right now.
you should go check it before the mailman comes around tomorrow.
>>
>set up Gemma 4 (free) on ST
>429 Too Many Requests
God doesn't want me to coom right now, i guess
>>
File: IMG_8175.jpg (224 KB, 1170x1308)
224 KB
224 KB JPG
>>110965000
>>
when you chat with coom cards, do you base your dude's dick size on yours or do you change it?
>>
File: crying aqua.gif (53 KB, 220x220)
53 KB
53 KB GIF
>10
>>
File: 1707011945182568.jpg (203 KB, 500x619)
203 KB
203 KB JPG
>sex god panko
>>
good night, /wAIfu/
please don't put a Doom nukage barrel next to me and my bed before shooting it while i sleep
>>
>>110968401
Depends on if I’m playing a character that’s not me
>>
>>110960762
Distill what who what where what share the bounty.
>>110960148
What happened to them?
>>
File: 1745945462757602.jpg (49 KB, 716x800)
49 KB
49 KB JPG
>>
File: 1767683982552192.jpg (744 KB, 2581x3535)
744 KB
744 KB JPG
>>110976664
>Distill what
create my own dataset to fine tune with by picking and choosing high quality replies
>wha happen
no idea but im assuming normal anon falling out bullshit
>>
>>110976664
The lead disappeared before Llama 1 leaked and the remaining devs all disappeared slowly and the ones who remained went full grifter mode with a website and commerical operations. It's small peanuts though and it's deserved for them remaining in obscurity although it is true too ragtag anons can't compete with Character.AI with former Google researchers. The main thing that pisses me off is that the public dataset, PIPPA, is so tiny vs the submissions they got and that they didn't keep trying to do local AI and gave up the dream of local models end to end. Well, no use crying over spilled milk, I'll take what we have now with open models even if we can't replicate it because we need millions of GPU time to reproduce it alongside no public data.
>>
Heard about Gemma4. Don't know shit about LLMs, but I've been wanting to "train" a model on my favourite smut novels for a long time now and it seems Gemma4 makes that possible.
Can you guys just confirm that the basics are sound? I've got an i7-12700K, 64 GB RAM, RTX 3080 with 10 GB VRAM.
Which of the Gemma4 models is best suited for this type of training that I can run/train with my hardware? Quality over speed for the output.
>>
>>110981150
How hard is that to do and how come no one is sharing their distilled something something on GitHub if it’s visible for us plebs to do?
>>
File: 1760078248681888.png (24 KB, 471x255)
24 KB
24 KB PNG
>>110983890
>takes more (passive) time and money than anything else since you need to let a gpu go brrrrr 24/7
>need a reasonably new gpu / googa notebook or else it will take a lot of time to actually finish
>because most people are using shit prompts and getting shit responses, so training on shit replies is just going to make things worse for the average ahh ahh mistress
>the average person benefits more from paying to use deepseek/[big corpo model] and leveraging its innate ability to unfuck a prompt though sheer intelligence more than they would using a smaller local model
>you need like 2 gigs or whatever of text so its kind of a lot of data to horde just to fine tune
>no guarantee the fine tune will be better
>"fuck you build your own"
>you still need a reasonably new gpu to get good reply speed
>finetunes are not magic improvements
pick between 1 and all but its not that hard in reality, just annoying to the point where the effort is not really with the juice.

realistically the use of a fine tune is to create a LoRA for an updated dataset of something that was 'created' after its cutoff date
like how deepseek 3.2 doesnt know who smugalana or veibei are, a fine tune is worth more to 'add' native info about something for this use case than to crank out "better" replies since raw info about people is easier to source than hundreds if not thousands of (high quality) replies
>see pic rel, ignore the 'edit' its the only way i can regen a reply in brave
an alternate use case would be to have a model NEVER act as {{user}} in the reply without having to hammer it into them, but thats a lot of effort/time/cost when
>do not act as {{user}} in your reply
is "free" to do and gets you 85%+ the way there

the other realistic use case you might want is to uncensor a model but again, you would need a uncensored dataset to uncensor the replies
and abliteration exists without that annoyance, and gemmi 4 models ARE already abliteratied

the best use case for a finetune, in theory, is to expand a card into a finetune, but thats a circular problem
>you need a bunch of good replies from the card in the first place to train the finetune, so it can create good replies (of the card)
so unless your autistic and old school w/ai/fu'ing a SINGLE card (or a nerd who wants to FAFO and have some fun) your not bothering with it when system prompts are "free" and like 10x less effort and money burned
>>
bump
>>
>>110981843
Fine tuning is possible only on the smallest Gemma 4 E2B variant but it's such a far cry from the biggest models which need 24GB or up VRAM. See https://unsloth.ai/docs/models/gemma-4/train as someone else said. I'm waiting for slightly different training regimes to see if I can eek out something with 16GB of VRAM with another training suite software which is generally more space efficient.
>>
File: 1755398422863719.jpg (65 KB, 726x546)
65 KB
65 KB JPG
>10
>>
Bao is an player hater
>>
File: fuwamochi.png (367 KB, 658x718)
367 KB
367 KB PNG
>10
>>
File: breed.jpg (124 KB, 460x532)
124 KB
124 KB JPG
>10
>>
bump



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.