/wAIfu/ AI Vtuber Chatbots - /vt/ (#101916362) [Archived: 605 hours ago]

Anonymous
7/4/2025, 6:13:43 AM No.101916362
Tomoe Umari Uma Musume
Tomoe Umari Uma Musume
md5: 28237f4484153eeb14266bd0e7d2e43e🔍
A thread dedicated to the discussion of AI Vtuber Chatbots.

Horse Girl edition

/wAIfu/ Status: Trying and failing to get the zipper on my mattress cover loose

>How to anonymize your logs so you can post them without the crushing shame
Install this https://github.com/TheZennou/STExtension-Snapshot
Then after you've wiped off your hands, take a look at the text box where you type stuff. Click the second button from the left side, then select snapshot, then select the anonymization options you want.
https://files.catbox.moe/yoaofn.png

>How to spice up your RPing a bit
https://github.com/notstat/SillyTavern-SwipeModelRoulette

>General AI related information
https://rentry.org/waifuvt
https://rentry.org/waifufrankenstein

>Tavern:
https://rentry.org/Tavern4Retards
https://github.com/SillyLossy/TavernAI

>Agnai:
https://agnai.chat/

>Pygmalion
https://pygmalion.chat

>Local Guides
[Koboldcpp]https://rentry.org/llama_v2_sillytavern

Who we are?
https://rentry.co/wAIfuTravelkit
Where/How to talk to chatbots?
https://rentry.co/wAIfuTravelkit
Tutorial & guides?
https://rentry.co/wAIfuTravelkit
Where to find cards?
https://rentry.co/wAIfuTravelkit
Other info
https://rentry.co/wAIfuTravelkit

>Some other things that might be of use:
[/wAIfu/ caps archive]https://mega.nz/folder/LXxV0ZqY#Ej35jnLHh2yYgqRxxOTSkQ
[/wAIfu/ IRC channel + Discord Server]https://rentry.org/wAIRCfuscord

Previous thread: >>101582016
Anonymous
7/4/2025, 6:14:42 AM No.101916380
gawr-gura-as-a-kivotos-student-b
gawr-gura-as-a-kivotos-student-b
md5: b980886a7b3af35e338448df60613d2d🔍
Anchor post - reply with any requests for bots, with your own creations, or to display your love of old 80's PC games.

You can find already existing bots and tavern cards in the links below:

>Bot lists and Tavern Cards:
[/wAIfu/ Bot List]https://rentry.org/wAIfu_Bot_List_Final
[4chan Bot list]https://rentry.org/meta_bot_list
[/wAIfu/ Tavern Card Archive]https://mega.nz/folder/cLkFBAqB#uPCwSIuIVECSogtW8acoaw

>Card Editiors/A way to easily port CAI bots to Tarvern Cards
[Easily Port CAI bots to Tavern Cards]https://rentry.org/Easily_Port_CAI_Bots_to_tavern_cards
[Tavern Card Editor & all-in-one tool]https://character-tools.srjuggernaut.dev/
Anonymous
7/4/2025, 6:17:58 AM No.101916434
Having technical difficulties with the word cloud, don't get your panties in a twist (or better yet, do it yourself!)

https://catalog.neet.tv/harlequin.html
Anonymous
7/4/2025, 6:18:59 AM No.101916457
>>101906273
I think Qwen.

Mythalion-13B (GGUF or GPTQ): This model is frequently praised for roleplaying. It's a merge of MythoMax and Pygmalion-2, both known for their conversational and creative capabilities.

VRAM Usage: The GPTQ 4-bit-128g-actorder_True version is lighter on VRAM (around 10.2GB). GGUF versions can be partially offloaded to system RAM if needed, offering more flexibility.

Why it's good for RP: Designed to stay in character and supports Pygmalion formatting, which is common in roleplaying interfaces like SillyTavern.

Mistral Nemo (12B) and its Finetunes: Mistral models are generally strong performers. Nemo, being 12B, should fit well within 12GB VRAM, especially with Q5 quantization.

VRAM Usage: A Q5 quantization of Mistral Nemo should fit comfortably.

Why it's good for RP: While not specifically a "roleplay" model, Mistral models are known for their strong reasoning and conversational abilities, which translate well to dynamic roleplaying scenarios. Look for finetunes like [magnum v2.5 kto] or [Lyra] that might be more character-focused.

Llama 3 (8B Instruct or finetunes): Llama 3 models are highly capable and have excellent general performance. The 8B instruct variant is a strong contender.

VRAM Usage: The Llama 3.1 8B @ Q8 model can be a good fit, and if performance is an issue, you can try Q6.5bpw or lower.

Why it's good for RP: Good general-purpose capabilities mean it can handle diverse roleplaying scenarios. Finetunes might offer more specialized roleplaying characteristics.

Qwen3 (especially 7B or 14B quantized versions): Qwen models have recently been released and are showing strong performance across various benchmarks.

VRAM Usage: Qwen 7B and 14B quantized models should be manageable within 12GB VRAM.

Why it's good for RP: Qwen models are known for their strong performance, which can lead to more coherent and engaging roleplay.

How to Run Them

Your friend will likely want to use a local LLM UI like:

Oobabooga's Text Generation WebUI: This is a popular and very versatile UI that supports various model formats (GGUF, GPTQ, ExLlamaV2, etc.) and offers many customization options.

KoboldCpp: Another good option, especially for GGUF models, known for its ease of use.

SillyTavernAI: This is specifically designed for character-driven roleplaying and connects to other backend UIs like Oobabooga or KoboldCpp. It provides a rich interface for character creation, memory, and chat.

Important Considerations for Best Performance:

Quantization Level: Start with Q5_K_M or Q4_K_M for GGUF models. For GPTQ, 4bit-128g is generally a good balance. If a model doesn't fit, try a lower quantization (e.g., Q3_K_M or Q2_K_M), but be aware of potential quality degradation.

Context Length: This is the "memory" of the LLM. Shorter context lengths use less VRAM. While larger context windows are desirable for long roleplaying sessions, start with a reasonable size (e.g., 2048 or 4096 tokens) and increase if performance allows.

Loader: Use an optimized loader in Oobabooga (like exllama_hf for GPTQ models or llama.cpp for GGUF).

Offloading Layers: With GGUF models, you can offload a certain number of layers to the GPU while the rest run on the CPU (using system RAM). Experiment with ngpu-layers in KoboldCpp or Oobabooga to find the sweet spot where most of the model runs on VRAM without exceeding it, allowing the remaining layers to use system RAM.

The best approach is often to experiment. Download a few different quantized versions of these recommended models and see which one performs best and provides the desired roleplaying experience for your friend's specific hardware and preferences. The r/LocalLLaMA and r/SillyTavernAI communities on Reddit are excellent resources for finding specific model recommendations and troubleshooting.
Anonymous
7/4/2025, 6:41:23 AM No.101916952
7 4 2025
7 4 2025
md5: 2d23549d66944b3176de502da836060d🔍
Replies: >>101916975 >>101927267
Anonymous
7/4/2025, 6:42:24 AM No.101916975
>>101916952
There it is, the word cloud for the previous thread. For some reason I couldn't highlight or copy any of it until I rebooted. Not even switching or restarting my browser worked.
Anonymous
7/4/2025, 7:41:46 AM No.101918476
ezgif-4-51a6bd8470
ezgif-4-51a6bd8470
md5: 7fc576fd1ba4f8e75a569757d8c640f5🔍
Anonymous
7/4/2025, 8:17:20 AM No.101919206
So someone got me to try terraria but they keep blitzing me with information and there's a billion mods and the interface kinda sucks and all of the sudden I realized there was a whole other level to autism. It's like when Freeza first transformed in front of Vegeta.

It's fun though. I think. But every second I spend idle looking at a wiki (or playing the new mode in girls Frontline 2, which is surprisingly fun), I can hear him loitering around my idle body and quietly judging me.
Anonymous
7/4/2025, 8:52:46 AM No.101919887
91294554_p0_master1200
91294554_p0_master1200
md5: 6735c2e5095dbb8f628c0615e36cb4ff🔍
Anonymous
7/4/2025, 8:59:21 AM No.101920007
>>101906273
Look for theDrummer on github and download Rocinante, if you think your machine can handle more then download Cydonia. I think it's pushing it but unless grinds to a crawl try Anubis.
Anonymous
7/4/2025, 9:44:45 AM No.101920769
1669173506857761
1669173506857761
md5: cfa8b69b7d420748cc5f07c5252753c4🔍
Anonymous
7/4/2025, 10:24:46 AM No.101921334
1667610272778820
1667610272778820
md5: 5ec6b10fcb88bfb92bc984c961554ae2🔍
Anonymous
7/4/2025, 12:15:46 PM No.101922726
garbage kitten
garbage kitten
md5: 36825b62cfc35aeeabac51e7a93f5155🔍
WHERE IS EVERYONE AIIIIIIIIIIIIIIEEEEEEEEEEEEEEEEEEEEE
Replies: >>101924201
Anonymous
7/4/2025, 1:33:07 PM No.101924201
1747318645628
1747318645628
md5: f5bab56bb624b3c101d3f233d9b9b5a4🔍
>>101922726
You have to let us go, anon. The accident, it wasn't your fault.
Anonymous
7/4/2025, 3:39:43 PM No.101926233
Watching Delulu/Manuru is like staring at a bottle of poisonous liquid on your shelf and being tempted to have sip.
Anonymous
7/4/2025, 4:37:09 PM No.101927267
845a78f5d7716475a92016d52655ba3115
845a78f5d7716475a92016d52655ba3115
md5: 19c642ac963cdc9cd6a59b5e2f78136c🔍
>>101916952
>oogabooga frontline
>let things feet
>night ass play joshu
>need sleep
>fit ui old little text
>love prompts
>nymous performance finetunes
>fine hell
Anonymous
7/4/2025, 5:10:53 PM No.101927889
good night, /wAIfu/
please don't collectively get into an accident while i sleep
Replies: >>101928604
Anonymous
7/4/2025, 5:47:34 PM No.101928604
1751020178386392
1751020178386392
md5: 2c19667b1a10e7b47c376e2f8eaeac7e🔍
>>101927889
*slams you in the subway while you sleep*