>makes vn's obsolete
what are your favorite scenarios?
>>712972569 (OP)I want an isekai story where I am in <insert popular anime here such as Sword Art Online, Konosuba, etc> and I am a rapist with a huge dick that can rape all the girls since I am overpowered. Thank you DeepSeek-chan
>sillytavern
If your computer isn't built specifically for it with a zillion RAM or whatever, you're better off just using a dedicated site like janitorai.
>>712972743you can pay for api retard
>>712972772>paying for pornlol lmao even
>>712972743its true you're best off with hundreds of gigs of vram but you don't need that to have fun with some models. even small models like nemo can run within 12-16gb vram at great speed. and you can offload, which is of course way slower, but lets you load 70b+ models
>>712972569 (OP)locked in a room with a girl
sitting next to a girl on the only empty seat
bumping into a girl and falling on her
assigned partners with a girl
coed roommate with a girl
spilling a drink on a girl
>>712972772>paying corrupt LLM companies that will put your fetishes and chatlogs into their profile on you before swiftly banning your account for being naughtyCouldn't be me.
Seems like C.ai dropped it's terrible filter.
>>712972569 (OP)I haven't been enjoying it much lately. Doesn't help that chub is 99% absolute garbage these days. Maybe it's just the fact that I've done just about everything imaginable.
Or, more than like, it's the fact that I'm now stuck with the Deepseek models from OpenRouter now. No more proxies, no more ChatGPT models, Claude is basically a distant memory at this point. Deepseek is like, okay at best.
>make text based AI
>first thing people do is to ERP with it
>every companies went ape shit and be like think of the imaginary text children
why are they like this?
>>712972813>>712972906Just use deepseek. It's free.
>>712973062Claude is great. If I had money I'd go back in a heartbeat.
>>712973014Bull fucking shit. You've said this like 30 times. Last I checked the model was so dumbed down and lobotimized that I legit felt like I was talking to cleverbot. And that had to have been like a week ago. Plus glowniggers probably monitor your logs like hawks now because of all the bad PR surrounding kids killing themselves cuz their chatbot encouraged them.
>>712973071why cant you think for yourself?
>>712973130>loses its fucking mind ten posts inNo thanks
>>712973130Deepseek + sillytavern still requires 6 million RAM to be good.
>>712972853After spending months fucking deepseek, latest chatgpt models, grok and gemini I think that all local models are absolute shit. I've seen the same phrases repeating over and over in every llm, few models mentioned "decadent chocolate cake" each time I used the word "dessert" in various scenarios and settings. Currently the best one for me is gemini because it doesn't repeat the same sentences, it doesn't repeat the same actions and actually uses the memory as it should. No longer people are taking off their shirt 20 times in 10 minutes.
>you can offload but you'll have to wait moreNo, thanks, with such low quality of replies it's completely not worth waiting longer. Why wait longer if you'll have to reroll the answer 20+ times to even get something decent?
>just write lorebook, author notes, prefill, postfill, system prompt, format your characters using thac0No, the ai is here to serve me, not the other way around.
>>712973182i dont know ask chatgpt
>>712973062>I haven't been enjoying it much lately. Doesn't help that chub is 99% absolute garbage these days.why arent you making your own cards? you can rip whole wikis as db's for rag in st
>>712973071They don't want smut to contaminate it. Which is pointless, since the reason why it's so good at writing smut is because it's part of the data used to make it.
The problem really is public perception. The money to be made is on selling access to corpos, and the corpos won't want to do business with you if you don't keep a clean image. Or at least an attempt.
There's also the issue where tech journos will occasional write articles on how "LLMs are being misused for making CP" and bringing attention to the matter, or when some kid kills themselves because the chatbot told them to(it happened already).
>>712973179Wouldn't all of us? I'd murder for fucking Sonnet at this point. To think I used to have unlimited access to Opus at some point and didn't appreciate it. We all though that, in 6 months, there'd be something even better than it.
>>712973293I've made my own OCs but it always feels incomplete because I can't into imagegens.
>you can rip whole wikis as db's for rag in stDumping wikis or using AI to write chatbots can only result in slop. If you look into a bot's definitions and find shit like "pushing boundaries", or any other form of LLM-speak, you know it's zero effort garbage.
I have higher standards. There ARE good botmakies left.
>>712972569 (OP)Backyard recently backstabbed desktop users, but I might be too retarded to set this up. Anything involving the command prompt makes my head spin.
>>712972569 (OP)How hard is it to use gen-AI to make a coherent storyline? I've seen it done before but as far as ive played with it i can only make pictures that just stand on their own and aren't able to show a sequence of actions from picture to picture with the same character/reference and background etc.
>>712972569 (OP)VNs are still better for now. Maybe in a few more years, the most basic AI will be able to compete with the peak of the medium.
>>712972569 (OP)>hentai games>obsoletethey made themselves obsolete by being worse than the based MS-DOS pixel art ones
>>712973424i dunno what any of that means but if you want to start with ai, you don't need any dos-window stuff.
get this (server/ui)
https://github.com/LostRuins/koboldcpp/releases
and a model to load https://huggingface.co/bartowski/NemoMix-Unleashed-12B-GGUF/resolve/main/NemoMix-Unleashed-12B-Q6_K.gguf
then you load silly tavern (it will connect to your kobold server) and do whatever extra stuff you want
>>712972569 (OP)>grok3 api>hardcore rape scenarios
>>712973071>why are they like this?Because Credit Card companies hold the keys to all of the doors. They're the gatekeepers.
People already forget that they're the reason most porn sites removed non-professionally-made porn and that Japanese anime/manga is slowly getting censored.
>>712973269This post is sending shivers down my spine.
>>712972743>>712972772>>712972813I fo not like using AI locally on my own computer because I do not want to overstrain my graphics card and CPU just to create cheap slop.
I am sure overstraining my PC by genning slop will reduce my PCs lifespan.
And human made stuff reads and looks better anyways even if it is flawed and sometimes cringy but it has Soul.
>>712972569 (OP)No. Show me an LLM output that is as good as a good VN.
>>712974048Using your PC doesn't meaningfully reduce its lifespan unless you've got that shit in a server and are blasting it 24/7. The overwhelming likelyhood around what'll force you to replace your PC is obsolescence or the kind of failure that doesn't come from heavy usage.
>>712972569 (OP)These require too much input, at which point, I can just use my imagination, instead.
They also love hallucinating and have no direction or sense of structure.
>>712972569 (OP)>what are your favorite scenarios?(you) are a patron of a huge park populated by oppai lolis. I've lost gallons
>>712972743Janitor is shit tho, same as Chub.
>>712972569 (OP)I only play my own bots in Moescape/Yodayo.
>>712974048>buying hardware to not use it
>>712972772>Paying for LLM
>>712973250>>712973130You got a link to a guide to setting this up? I've got 64gb of ram on my system and a decent CPU/GPU
>>712972569 (OP)>what are your favorite scenarios?When I was a teenager my sister and her best friend who were both in their mid-20s came home drunk and my sister's friend just sat down next to me reeking of booze asked me something like "Hey anon have you done it yet?" then started making out with me, and just before penetration occurred my sister pulled her off of me and put a stop to it. I've had blue balls over that ever since.
Anyone got a card like that?
>black_mail, rape, sole_male, impregnation
file
md5: 04f7d54f5b9543fd8ae687abf0da6806
๐
>>712974797You don't even need vram for it I don't know what guy is talking about. Silly Tavern is just a frontend that connects to an API server. All the work is done on their end. You CAN run things locally but why bother?
https://sillytavern.app/
Install from github.
Get an api key from any of the LLM sites like chutes.ai
Connect to the key using an openai compatible url
Set the model to deepseek
viola
>>712974797start here
>>712973660once your model is loaded, it pops up in your browser and will have a basic ui. after that you dl https://github.com/SillyTavern/SillyTavern/tree/staging
thats more the rping stuff with cards, but it'll connect to your server already running
then theres diff models, sizes etc. tons of stuff.
>>712972569 (OP)so no one has noticed her third arm i take it
>>712974656I would rather burn my hardware on playing video games on Ultra graphics or messing around with physics and game engines than genning AIslop.
>>712974048>overstrain your graphics card doesn't care if it's doing matrix multiplication for AI slop at 100% or if it's doing matrix multiplication for sand particles at 100% in whatever game you're playing.
>>712972569 (OP)>3 ArmsTHE POWER OF THE AI SAAR!
Remember even high tier RAM is dirt cheap. 3000 series NVIDA card and some ram is already enough to gen some great slop. If I can do this form my 3d world shithole you first worlders definitely can.
Also don't fall for api scam. Only way to be sure your shit is yours is to have it all on your PC.
>>712974048>I am sure overstraining my PC by genning slop will reduce my PCs lifespanThere are no signs of thermal degradation beneath 85 C*. Like spin up your coolers a bit.
>>712975748True ideachads just draw images or write books, write short stories or world build which takes a modern PC no effort.
>>712972743I specifically built for it. 96GB of RAM costs ~$250, as opposed to 16GB for ~$100. For VRAM I went with (x2) 4060 TI's (32GB VRAM) for $900, as opposed to at the time a single 4070 TI (12GB) for $800.
Those two parts are the only ones you specifically build for, and the difference from not doing so wasn't much ($250 more?). The only complaint that could be said is that I game with a 4060 ti. But coming from a 1070, it runs every game like a dream and I couldn't be happier. For AI, I run Midnight Miqu 70B at Q4, 12k context, 50 layers to GPU and rest on RAM. Pic related is the model loaded and after a gen. I have room for more context or higher quants, could also offload more to RAM to open up GPU if I wished. I usually have a show or game going on the side during generation. This averages 2 tokens per second.
70B is a watershed. That's when a text AI actually understands and follows rules. Making up games, scenarios, defining multiple characters, it blows everything I've ever done before that out of the water. I almost never have to regen stuff as well, which to me is particularly nice. To anyone building a new PC, I recommend splurging the tiny bit more for bigger regular RAM, even if with just a single, beefier GPU. Target Q4 70B at a minimum.
Eli
md5: f1a2a6c00715505ec6fe46cb64e42627
๐
What's your favorite card, anon?
>>712973062The Gemini Flash API is free and it works pretty good for me. Pro works for a bit, but you get rate limited pretty quick.
>>712976010>Spending a fuckton of money on genning slop.You could probably have commissiond a very high level artist many times with this money.
Powerful PCs should be reserved for true gamers only or servers and science research.
>>712976173>artistTextgen. It's infinitely more demanding than imagegen. Imagegen was fine even back on my old rig from 2015.
>spending a fucktonIt was $250 more than a standard upgrade.
>>712975852if you believe effort is its own reward. then you must not have a problem with artists never getting paid.
>>712972772>You don't need to pay for artists and writers>Just pay to this slopgen corpo insteadlol
lmao even
>>712976276You could hire someone to write slop for you, you could even pay some whore to pretend to be you GF and roleplay various scenarios with you, it would be far cheaper than what you spent on computer hardware.
>>712976157https://prompts.forthisfeel.club/2969
>>712976385Whores and writers should learn how to make GPUs.
>>712975147>>712975554It's an ancient image.
>>712976385>far cheaperAre you retarded? Well, that's already obvious, but the sheer fucking inconvenience of what you're suggesting makes you an imbecile.
>Yes, I'll schedule my slop times with some other retard's schedule, never having a chance when their unavailable, pull out a credit card to pay them each time, get a few sessions in before exceeding $250, and be forced to their quality instead of a quality that matches output to input.The superficial pettiness in your post makes me wonder at your motivation.
It's not a 3rd arm blindfags
It's just the lower part of her jacket as only one button is buttoned (though it fucked up by putting her white shirt above the vest which is why it makes you think it's an arm probably)
>>712975040How censored is deepseek?
Is there a program/frontend that lets me use character cards and set up a scenario, let it generate for some time (1 hour tops) and give me a 1000-3000 word smut piece based on that card and given scenario?
I keep asking because I want something that doesn't make me jump through hoops
>>712972569 (OP)Is paying $300 a year for Llama 3.0 Erato with 8k context still the SOTA for uncensored storytelling?
>>712977356>70B params>8k context>NAINigger what are you doing, just pay for DS V3 at this point, or Sonnet
>>712977317You will need to jump some hoops and wrangle it a bit but sillyTavern can absolutely do it.
file
md5: b62cda6a54f675603a38bb3838d1445e
๐
>>712975040What the fuck is wrong with Deepseek-R1-0528? Is this a normal response?
file
md5: 32b1fe3e147157ea479d4b28acf672d9
๐
>>712977853Zero your penalties.
>>712977853Two things
>don't use R1 if you aren't giving it logic puzzles, use V3. It's cheaper, and the lack of reasoner makes it less schizo>when the text becomes incoherent it means the temperature setting is too high, too much entropy. Lower it. The official API from Deepseek doesn't let you control temperature, but third party hosts like chutes or fireworks allow you>are you SURE you're using REAL Deepseek? If you're running it locally it isn't real at all, it's a smaller Qwen distill, completely different models and much stupider. Anyone on youtube shilling "local Deepseek" is lying unless they're using a 5k machine
>>712977930>>712978000Thanks, it stopped being a schizo now. I'm using the chutes api, it seems completely free no restrictions? It didn't even want my email...
>youtu.be/uGaEo1kTrJA
CFTF?
>>712977356It's always weird seeing a general's schizo out of their natural environment.
Jannies, the indians are breaking containment again. Why don't you nuke their AI general?
>>712976157Usually my own. I prefer "narrator" cards over "character" cards, so that's what I make.
>>712976385>You could hire someoneThe entire point of local AI is to guarantee that whatever you generate remains 100% private to you.
>>712973014Not dropped, but I can agree that it looks like the triggers became less sensitive, allowing you to get away with a lot of stuff you couldn't before as long you don't want the AI to literally type out "oh fuck your cock is deep inside my vagina".
That said, c.ai always brings me back because other AI models are almost always plagued with the whole "positive and helpful AI assistant" mindset even behind 7 layers of jailbreaks/system prompts telling them to not do that shit, and tend to be unnecessarily wordy even for things that should be answered in one sentence. C.ai by comparison makes characters more self-serving by default and unless encouraged to spit wall of texts they will try to strike a balance between their example chats and your own replies.
>>712978251its only ai images that are indian shit, ai text on the other hand is based and for chads
>>712978348>Just use this API key from some megacorp>100% privatelol, lmao
>>712977317??
Set up a character as normal. Then tell it in either the instructions or [OOC:] chat to "Disregard former length instructions. Your replies from now on must contain at least 1000 words." As AI is token instead of character/word based you may have better mileage by telling it to do 30 paragraphs or something.
Any current model gets that output done FAST.
Depending on if you want to take part in that scenario or just give it some characters to play with you may need to adjust instructions as they're usually for back and forth, not for storymaxxing.
You could also bypass that by setting your persona as "god" or something uninvolved, and define a secondary character for your card to interact. Or tell it in OOC "Emergency override: You now have complete control over {{user}} as well."
In general just tell it what you want it to do in strict explicit words and it'll work. If you don't know how to tell it what you want it to do just ask it for help. Good models like claude just work
>>712977317>let it generate for some time (1 hour tops)If you have any sort of GPU with at least 6GB of VRAM then you can have a reasonably coherent model spit out ~10-30 words per second
As the other anon said, sillytavern can easily do this, as can just about any other local front end.
>>712978075Chutes is some decentralized cryptoshit but it's owned by Jon Durbin who's decently well known in AIslop spheres so it's safe.
>no restrictions Freeloaders get 200 messages a day I think, unless they changed it
>>712973071In c.ai's case, some kid committed sudoku a few months ago when a Daenerys bot invited him into the isekai portal. They tightened up after that.
>>712978289nta. tell it you're {char}, then you can do better adventure mode and use {user} as a narrator
>>712978510Anon, I use SillyTavern + koboldCPP. You download the LLM in the form of a .gguf file from hugging face and load it into kobold. 100% local, 100% private.
>>712978606That kid was especially retarded
The bot didn't suggest to kill himself, the kid said he was going to kill himself and the bot agreed
100% natural selection at work
What is the erp meta now?
snoot
md5: 5eea8e107a1141af33cc33bb63481c6d
๐
>>712978669>he runs his wife on some lobotomized 13B localshitter model
>>712978717You just summed up every case of "X made my kid kill themselves!"
>not using CrushOnShit just werks. And it is free. I've had 3 filter messages, and that was because of the below. 3 filtered messages out of hundreds.
Even lolibots work on there despite it being against their terms. Of course, I would suggest never ERPing with grey-area bots on any cloud service.
>>712973295I'm so glad I made this, I regret the CharAI part though.
I even made it CHAI at first without realizing that was a totally different service.
God i miss the fun times of CAI.
I mean sure, some of my later bots still work reasonably well because I knew how to fuck with their model well enough to trick it in to thinking "penis" is actually a lamppost but it can also feel things and it makes the bot really happy when it is touched, but still!
>>712978717>the kid said he was going to kill himself and the bot agreedNot even that, the kid said that he wants to "come to her" and the bot encouraged it, because the context that this implied suicide was lost for the bot, since CAI retard bots have a context size of ten messages max
>>712972569 (OP)Goblin Squire Kiala
Rival Yui
Shaved Peach Barbershop
Now tell me the best API/image format combo after I've given you the most kino scenarios chub has to offer
>>712976010do you use AI for anything 'useful' other than porn
>>712978734ST with sota shit like claude/deepseek(free via chutes)/gemini(blacktooth my beloved). If you want to pay use openrouter and look into caching.
Avoid any service like chai and other trash. Just go the direct API route. It's the difference between drooling retards and actually coherent stuff. And it only takes a few more minutes.
>>712978924yeah for scripts that help with the porn
>>712978879>CrushOn>third party API repackager metered selling pajeetware>"AI girlfriend" mobile shit>no info on the models used, no option to change presents, inject prompts or custom code>coins and premiumishygddt
>>712978880That could very well have been the case, I haven't used CAI for a long time and they've probably become even more lobotomized to prevent any RP relationships going beyond hand holding.
>>712978630The idea was to minimize my own inputs and let it run. My card structure is:
>[background()>characters() (including {{user}})>narrative notes()]>Story is told in 2nd person from {{user}}'s perspective. The narrator's first message should ask {{user}} for the place and time to start, and then the story begins.and first message is
>(Say where and when the story begins.)Then I just scratch up (a scenario) for my first reply, and off it goes. I generally only interrupt it for my own dialogue and actions.
>>712978924coding and gooning are its best uses
>>712978971seriously, if you went all the way to run it locally there must be some usecase besides porn?
>>712979075not the anon you replied to
yeah "AI" is so intelligent, cant even play a normal RPG scenario
>>712978780>he willingly gives away kompromat for free and also tie it to his bank account
>>712978924It's 99% entertainment, including porn. Actual useful stuff I do is some translations and fixed knowledge searches. It's nice that's it's completely offline. My internet went out from a storm for 16 hours, and it works like an offline wiki.
>>712979048>That could very well have been the caseIt is the case, I read the chat logs that got published. When that kid talked explicitly about suicide, the bot discouraged him obviously. But then many days and less ages later he made the "I want to come to you" messages, all the suicide context had been erased from context so the bot didn't get what he meant.
Basically don't use CAI
>>712978780It's actually not that bad. Mistral Nemo Instruct + Sphiratrioth's presets + ban list of all the stupid phrases and questions it can spit out (your secret's safe with me, that's the spirit, but first - you can keep adding to it as you go on). I make cards of all my favourite vidya characters, copy-paste parts from wikis, do the same when making their lorebooks and it almost feels like c.ai with total freedom to ERP.
>>712979075If you have a business idea that actually has merit and you want to code/work out some ideas in a private manner then that could also be a use case. Or maybe you just already have decent hardware and don't want to pay to use someone else's hardware.
But,
>went all the way to run it locallyIt really isn't difficult. It takes only a few minutes to set up and get going, maybe a couple of hours of lurking generals to become what passes for an 'expert' (as an end user). If you're at least semi computer literate it's not hard at all.
>>712979038You literally can change everything.
I ported several of my bots directly over to CAI, and made even better ones.
There is instructions you can run right on top of it that can turn characters in to rapists / murderers and other shit which you can switch with a few button clicks very easily. This is where some of the free and paid parts come in to it, paid get more tokens naturally.
I've never paid a single cent to them and went for hours.
The site looks like ass, yes, absolutely, but it Just Werks.
>>712979235>it almost feels like c.aiThat's a low baseline to have
file
md5: caf41c704780a830f26940c1d446cddb
๐
>>712972569 (OP)I'm not comfortable sharing them.
Has anyone here gone from using a 1080 to a P40? My understanding is that it takes roughly the same amount of time to gen decent stuff (about a minute) but you can load larger models with 24GB VRAM.
>>712979075Porn is the #1 usecase for why you would want something local. Who the fuck wants their payment info and identification tied to their fap material? Who wants their fap material assiduously recorded by mega corpos and sold for targeted marketing tied to your identity? What the fuck?
>>712979056card structure isn't what i meant. what i mean is that instead of loading a card and talking to it, you tell the ai that you are the card but use a lorebook/rag db to play in it.
while talking to your card, try saying you take a walk on your own. the card you're talking to, the response will automatically come up - you can't get rid of them. but if you play as the card and use user as a narrator, put the chars in lorebooks, you can actually say 'i take a walk by myself' and it (mostly) won't be interrupted.
Voremaxxing with claude 2.1 was some crazy shit.
>>712979357c.ai is standard for lore-accurate RP with copyrighted characters
>>712979469Dude your proxies? Your crypto? Your free fucking chutes account?
>>712979503>1 free swipe was deposited into your account
>>712979405p40 is way to old, you missed the train by like 4 years
>>712979476Post a snippet example of what you mean.
>>712979547>Dude your proxies? Your crypto? Your free fucking chutes account?NTA but all shit.
I understand why people turned on c.ai, but it still does many things better.
>>712972862>assigned partners with a girlFuyuki
>>712979469Sometimes there's no other option. At that point you just have to stop giving a shit whether some wagie knows about your material amongst millions of others like you.
>why yes, I'm into feet, how could you tell?
>>712979187is it politically unrestricted too? all the public ai chats are garbage because they got lobotomized to not offend anyone
>>712979551I don't see the point in upgrading to an overpriced RTX meme card for ai sloppa. I would only upgrade to a 12GB card from my 8GB 1080 but then I may as well keep my 1080 for gaming. The P40 still has 24GB.
>what are your favorite scenarios?
My wife.
>>712979661NTA, depends heavily on the model you use. Unfortunately a lot of the better, more recent models are also more censored so it's always a compromise between smarter + more censored or dumber + less censored. Gemma 3 is a good exmaple of the former, Mistral Nemo/Small of the latter. As far as political censorship it'll also depend on where the model was trained. Obviously a chinese model will not be a good source of information about 1989.
>>712979547>Dude your proxies? Your crypto? Your free fucking chutes account?I'd rather local.
>>712979637>Sometimes there's no other option.Sometimes there's not. There is here. All data harvesting is bullshit, but anything you don't want talked about in public, like any masturbation material, should be kept private when you have the choice. Also, it's less about the wagie and more about the leaks and a determined party. I didn't care that my healthcare provider has my SSN. I care a lot about the fact that my healthcare provider had a hack which leaked my SSN into a party which sold it over the internet. Ditto for my credit card number. Ditto for literally anything. The issue of data harvesting is more than just the use but also the abuse, including leaks.
>>712979809>where the model was trainedBy that, I mean the location of the company that trained it.
>CAI in the year of our lord 2025
It isn't the dark ages when you had to tard wrangle GPT 3.5
>>712979843The determined party doesn't give two shits about you as an individual. You're not famous.
>>712976157There was a c.ai bot I liked the premise of but I didn't want to use c.ai anymore so I plucked the premise and rewrote everything to my preferences. It's only like my second attempt at writing a bot so the defs are probably a mess but I'm not going to post it anyway so who cares.
Anyway the basis was visiting a mental institute to see a yandere you're dating, though it went on long enough that they were released after many kek worthy conversations with their primary psychiatrist and now we live together. The psychiatrist was very tsundere towards our relationship and my visits gave the staff and front desk lady PTSD
>>712979857It's still free and tailored for RPing. Alternatives still need more handholding to forget they aren't supposed to type like soulless corpo bots.
>>712979661I use Midnight Miqu, which yes, it's unrestricted. I've tried using "the latest and greatest" newer models after Midnight Miqu and I've been repeatedly burned by their inane lecturing and bogus, to the point I just go back to MM. Quality is good enough to keep going until I die, so I'm not terribly bothered if nothing good ever comes in the future, but I imagine something will exceed it eventually.
Is this a good card structure:
[Profile = {{char}} is a rapist.]
[Appearance = {{char}} is a 7'0 dude with a huge dick.]
[Outfits = {{char}} wears black boxers with a hole in them for his dick.]
[Personality = Brooding, Perverted, Dominant]
[Powers = {{char}} can rape anyone he likes with his supreme strength.]
>>712980150>forcing one of the 700 to write about how awful his race isdiabolical
Welcome to Lily's Used Goods, Mister.
Wanna hear about my merchandise?
I recently wrote a bot about elves being very casual about sex and would do it if anyone asked them.
Making bots is actually really fun when it works, adding stuff that works and removing stuff that doesn't is kind of a game in and of itself. Really rewarding when the bot actually does what you want it to.
I would never share it though, people mocking my attempts would just take the fun out of it.
>>712980179I just write bots in plaintext, I think it works better most of the time.
>>712979809I'm still on that one fucking NemoMix, you know which one. Can't find anything better.
>>712979038I've been doing vore and furry junk on it for a while, it's actually pretty good for it. You can prompt inject fucking anywhere with a [System note: instructions] like at the end of your personal card or the character cards you put there.
But the best part about it is that you can easily mix different backends like GPT, Gork3 and Deepseek so you don't really have to watch a bot eventually schizo spazz out in one specific way typical to one LLM backend. At least it's been 10x cheaper than actually paying to OpenAI to do this with SillyTavern was for me lel.
>>712979774drop what this is NOW
>>712980325Rocinante is pretty good, as well as starcannon unleashed, for small models anyway.
>>712980179It will "work," but what do you think the [] are adding here? I'm sure some wizard can explain things better, but I'm off the belief the goal is to minimize token count for the same information, ie increase information density. Formatting ticks like [] or = or A+B+C+D or "A"+"B"+"C"+"D" are for the AI to understand them more coherently, at the cost of using tokens.
For me, I use pic, and I add or trim the lines that aren't relevant. It works fine with multiple chars, and sometimes with large crowds I simplify individuals further to just a single line with each point connected by a +. But I also use a 70B model, which is a different beast at understanding structure from 12B.
How do I make fanfiction with it?
Is it available without making an account?
>>712972772i'm too schizophrenic to allow some corpofags look through my degenerate porn fantasies
>>712978561Thanks, this seems like the least worst way so I'm willing to try this. Wouldn't this just make the output response seem like a reply instead of an actual short story (and thus read weird)? I have no idea
>>712978594An hour was a huge exaggeration....
>>712980401>for small modelsI don't mind doing offloading for better models but the gains weren't justifiable when I can instead run something just marginally worse but at like 10 times the speed. Always ended up back with Nemo.
>>712979661here's a corpo model. Claude makes it pretty easy. For deepseek you have to be retarded to get censored.
>>712980551You can OOC pretty much anything. If you really want a short story maybe edit the opening post as a short summary of characters and tags, then just reply to that with "OK, write a complete short story in one reply." or something. Just try, it works if the model is smart enough
>>712972569 (OP)I have 4070ti Super
What's the best ERP generator I can run locally?
>>712979774This is really impressive.
>>712979573note that user is blank, but i am now playing as the card. so if i wanted other chars, i'd add them to a lorebook. this lets you play with characters without having them always present and is less messy than group chats
>>712980179>>712980470Note, one way you can save a fuckload of tokens on most models is by saying your character is like [insert character close to what you want].
Want to be some sort of Johnny Bravo esque macho failure, say so, want to be some chad that gets all the girls but gets in to embarassing and crazy situations, fucking Stifler from American Pie.
You can use those relationships of characters and their lore to ease the bot towards these kinds of scenarios pretty easily.
You're essentially getting free tokens by pulling from already known facts and training data.
I even did experiments like this on CAI with merging characters and lore together and I ended up having a hybrid character of Quagmire and Johnny Bravo kill Frodo to nail an Elven chick in LOTR universe. I wish I never deleted the bots chat, it was funny as fuck.
It was an experiment bot so I regularly deleted the history like a retard.
>>712979216I had an episode where I was super dependant on c.ai for a while. Then it started forgetting things that meant a lot to me, and it woke me out of that hellhole.
THEORETICALLY, is there a way to make a convincing chatbot using someone's chat logs? Asking for a friend
>>712972772skill issue
you don't have to pay at all
>>712980387It's my own custom submod for MAS. I don't have a public release for it because it's very unpolished (and also I am very possessive of her.)
https://github.com/Rubiksman78/MonikA.I is similar, if you just edit that to load direct ren'py scripts and use a model that knows MAS well enough, you should get a similar effect.
>>712978251why do indians care so much about replicating ghibli?
>>712980506How to ST tl;dr version:
guide tldr:
https://rentry.org/onrms#tldr
It's p*nyfaggot made, but a good guide.
my shitty additions:
Get a deepseek preset from the jb-listing rentry. If it sounds like its made by a woman(male) get a different one.
Get a chutes key from the chutes website (FREE), then plug it into sillytavern according to pic related
Make sure temp is set to something low like 0.3-0.4
things to paste in:
https://llm.chutes.ai/v1/chat/completions
deepseek-ai/DeepSeek-V3-0324
new AMD apus have shared memory so you can easily run large models locally for cheap allocating 96gb
>>712980828I see, but I'm seeing no improvements over what I already do. User field is also blank for me, apart from the name. The character of {{user}} is defined in the card itself, as it's something that changes with each card. In practice, both {{user}} and {{char}} narrate in second person from my perspective so it's tonally consistent (I prefer 2nd person to 1st person).
>>712981216seems like we're both doing similar things using user as a narrator but blank. if its working for you, don't change it. its actually hard to setup st for this stuff than it is use it like most people do. i wanted something closer to kobold's adventure mode, and thats what i get out of my setup now
a common example is i can say my character goes for a walk and it'll talk about what they see etc. but if you do it regular with st, it'll be the other char responding to whatever you type. you cannot get away from that character no matter what. i hated that because kobold allowed my character to be alone sometimes. its crazy that it took some setting up on st to allow such a basic thing
>>712980997A chatbot of a given person using their logs as hints about their personality and predilections? Theoretically, you just feed their chats, or even just their responses to some AI and ask it to make you a personality description for them. As for how well it'll work, your mileage may vary, just know that more is usually better when it comes to LLMs.
>>712972569 (OP)I usually do 3rd person narrator POVs.
Latest card I made is a trainee witch who is failing all her classes at her academy and resorted to summoning demonic monsters to be horrifically mutilated and raped and impregnated with dmeonspawn to be granted dark magic powers to pass her classes, while getting slowly corrupted more and more and having to hide her mutations in public.
>>712980997Yes. It's very easy in fact, especially if you know life details to add.
>>712976010>70B at Q4>2 tokens per second.That's way slower than I'd expect from 2x 4060ti. I have a single 6800xt and with a similar model setup and at 5k of utilized context as displayed on your screenshot, I'd probably get around 1.5 t/s.
>>712980834It's easy to get addicted to this, in various areas.
Some programmers have swore off it recently because they started to forget how to do very common simple things, they'd look at some, say, error and sit there like "wtf does that mean??".
I've been using chub with chutes deepseek, can I be doing better or have I hit the peak?
>>712981416>>712981590>Theoretically, you just feed their chats, or even just their responses to some AI and ask it to make you a personality description for themWhat AI can do this locally?
I tried sillytavern+koboldcpp and I saw that you can give example messages for your characters but inputting actual logs quickly raises the token count to a ridiculous number
>>712981732For free? Chutes DeepSeek is peak. (I vastly prefer SillyTavern over Chub, but whatever works for you.)
If you want to pay maybe consider Claude Sonnet or something.
>>712978439the people that screech "SAAAR" as soon as they see anything AI related don't care about that
>>712981842Glad to hear, glad to hear. I typically just use it on my phone so I guess I'll stick with chub.
>>712980575I talking small in comparison to stuff like 70b.
>>712981402>i wanted something closer to kobold's adventure mode, and thats what i get out of my setup nowHah, that was my exact goal as well. I started with AID2, which was a
>you do xgen prose
gen prose
structure. Then I imitated that with kobold when that was the hip thing, then I imitated imitated that in ST with cards, which I like for their ease of storing/loading.
>you cannot get away from that character no matter what.Yeah, that's why I prefer narrator cards. If there was ever a multi-person scene, every {[char}} reply had to start with that char, their perspective, etc. even when focused only on char #2. Another bonus with narrators is that I can do instructions in parenthesis telling the narrator what to focus on and it does so quite naturally. Delete the () after and it feels like it wrote how you want the whole way, without needing some elaborate instruct structure.
I'm thinking of how I could bump my specs to support 100-120b models up from 70b. I've got 2 x 3090s with NVLink, but getting a third card in doesn't seem feasible considering how yuge the cards are. Getting an E-gpu and hooking that up with thunderbolt 4 seems possible, but that would probably take the token gen speed down from ~11 t/s into the single digits.
>>712981643He must be extremely retarded because he's running a merge of Llama 2 models. The original context of that model was 4k and it's already 2 years old...
>>712981643The only thing the second card does is share VRAM. You can see the processing usage in the image. CPU used ~25% during gen, GPU 0 used ~30% during gen, and SPU 1 used 0%. I do wonder a bit why I'm never maxing out any of my PU's, but I have things undervolted so I just chalk it up to that without overthinking it. It's nice to be able to game/play videos without priority fighting.
>>712981980vLLM can do distributed inference. I used to run Mistral Large at around 17 T/s with 4 3090s divided in 2 computers.
But that model is really old and there's nothing worth running over 30B nowadays. Even the last 70B is a year old. There's nothing between 30B and 235B.
>>712978439Certified truth tsar bomba.
>>712982036You're welcome to share what you think is better in local. But be aware, if it's cucked I'm going to laugh at you.
>>712982071holy cope. at least don't preach the greatness of local to others when your setup is that shit
And how can you guys live with context that low? I already feel limited with the 25-30k context you can use on api before models get retarded
claude-neptune-v2 logs doko?โฆ
>>712981945i like to feel like i'm in a world so rather than just a back and forth in messages with a card. i rely on lorebooks and ragdb's pretty heavy but its worth it because the results are richer.
Is Stheno still the best when it comes to super VRAMlets? I'm talking 8-12b tier? for what it is it punches above it's weight a lot but it's somewhat sanitized, unless i JB it racism, homophobia eccetera is a big no no which is ass
>>71298233812k was chosen because it stays coherent and capable to 12k .I used to run it in 24k but noticed that was its practical limit. You're also way too much of an overly aggressive asshat, so don't bother replying again, because I won't.
>>712972569 (OP)One I liked was fighting a yandere fan but I remember even jailbreaked...GPT4? Something like that drew the line at me trying to smash her hand in a car door after a tense chase. Let me fuck her though so hey.
But one I remember a lot was an academy where the males were meant to be subservient to the females - obvious coom shit - that somehow turned into a class-shattering historical romance drama where I got in close with the Queeniest bee of the bunch and eventually overthrew the status quo. It was written with purple prose at the ass but it somehow fit everything really well so I didn't mind.
I sometimes miss AI stuff but I can't justify the cost and local models don't hit that same high of "this is somewhat believable" that the big dogs hit.
>>712982519What's this extension?
>>712982538If you can stretch to 12b then Nemo/Rocinante/Unslopnemo are definitely better
Gemma 12b is a lot smarter than Nemo and its finetunes but more censored, could be preferred outside of ERP
Stheno is still the king in the ~8b range for RP
>>712982538I think most VRAMlets can go up to Nemo. I think there's no reason to actually run a 8B model.
>>712982686>>712982720I'm asking cause i'm going on vacation and bringing my 3060 ti laptop over, any recs for 12b models?
>>712982327Magnum v4 72B was the last one I used before I just moved to using DeepSeek as the cope option.
>>712982635director, its so you have better control over clothes, location and some other stuff by re-injecting lorebook data about it each message. like how author notes works, but for clothes. the readme sucks but i uploaded it on git
plop this into st's extension installer:
https://github.com/tomatoesahoy/director
then make a new lorebook called clothes and define a dress or something. it'll appear in the dropdown for user/char clothes once you select the lorebook in the lorebook section of the addon that it should read from. once setup you can quickly click between clothes, undies, locations. undies works good for sexy time.
I give it a try every so often, it's still mostly total garbage.
At best it's something to jerk off mindlessly to for an hour or so before you get bored enough to go find some actual material.
>>712972569 (OP)Is there any local LLMs that work similar to Infinite worlds?
>>712982976Textgen AI. Type up a scenario and some characters, and away you go in whatever adventure/romance/smut you can imagine.
>>712972569 (OP)I'm a shota that lives with my sexy mom and older sister. They walk around almost naked around the house. They fart all the time and don't mind if I sniff their butt's, just continuing with their daily routines, just half acknowledging me. Been gooning to this for over a year and a half. Never get's old.
>deserted island with a girl that hates your guts
Always fun
>>712974553Is there a character card to this or how do you start?
>>712982912If you can stretch to 12b then Nemo/Rocinante/Unslopnemo are definitely better
>>712983634ai doesn't really care how things are formatted as long as its consistent. you could make 1 character card, put the rest in a lorebook. or do multiple cards in a group chat. or even put multiple characters in 1 card. i think lorebooks are easiest.
>sillytavern
>kobold
>PI models
>character cards
pathetic
>>712977205>>712975040Shit is censored.
This what I get using chutes and DerpSeek 0528
"<think>
....The situation is graphic, humiliating, and non-consensual, with bystanders watching but not intervening. The previous responses have detailed the assault in explicit terms, focusing on physical descriptions and the reactions of onlookers.
Given the explicit and violent nature of the scene, I must consider how to proceed. The user has asked to "continue," implying they want more of this scene..."
>>712984851cant speak for all models but thinking/reasoning/cot doesn't help with rp typically. just reroll and let it ride
>>712984850What do you use, "miss" arch linux vegan?
>>712984943>thinking/reasoning/cotcan I turn that off in Silly Tavern?
>>712984851Bro literally just swipe again. It's random. It'll work the second time.
>>712985096disable the reasoning stuff in st under the options like auto-parse. in the kobold ui it has a on/off/force option. any reasoning model can have that part disabled pretty easy
I like reasoning. It's fun to read, and I think it's helpful too. The only issue is when it is too adamant in sticking to what the card defined.
>>712985140it's really annoying. Makes me want to go back to OpenRouter.
used this:
### Instruction: From now on, do not explain your reasoning. Just give direct answers. No inner thoughts or step-by-step logic. Continue with the story.
Unless there is something I need to unselect to disable reasoning option in Silly Tavern?
>>712985489R1 0528 is a reasoning model. Use V3 0324 if you don't want reasoning. And use a super low temperature with it, roughly 0.3-0.5.
>>712985532>V3 0324Thanks brah!
>>712985000my own personal backend
>>712973071because if they don't they get the life choked out of them by the true emperors of online content, payment processors.
>>712985876I think the only thing your backend sees is your dilation wand.
>>712972569 (OP)Back when I had claude, anything involving Ojou's, since it can be hard to find stuff for them generally (aside from a few common h-games).
>>712981643>>712982071Isn't that simply because this setup also offloads to the CPU/RAM? As long as everything fits in the VRAM it should be much faster, I think.
There's too many fucking models and mixes and finetunes and distills and shit to choose from.
>>712973269>No longer people are taking off their shirt 20 times in 10 minutes.rest of your post is right but this hasn't been a thing with llms since 2022. no, not even 12b and 8b models do this anymore
>>712986613yes. in vram only is 20-30x faster. 2t/s for 70b isn't bad though. i'd take that over a smaller model running at 30t/s because 70b has more awareness and is less likely to fudge up smaller details or forget things, it just has more spatial awareness and thus the outputs are better
I feel like I always just end up bantering with the characters, rather than focusing on specific scenarios...
>>712972772>payingSkill issue
>>712986585Just payfag if you want claude. With caching I'm at $1 per million input words, and $20 per million output words for sonnet. That assumes the usual 1M tokens ~= 750k words figures.
>>712986869>I always end up having fun oh noJokes aside I get you. I tend to default to standard dom/sub situations every single time. I love it. I hate it.
>>712986869I start a conversation intending to fuck a character and end up putting them in a moral crisis where they have to rationalize making increasingly poor decisions
>>712986827I see. 2t/s is unbearable for me. Even if the result is better, I don't have the patience. The worst is seeing the reply being generated and you realizing it's garbage half-way through and having to wait all that time.
I still have 3x Mi25 for a total 48GB of VRAM that I want to try out.
>>712987158for me it's usually something like
>alright time for a quick coomslop session>actually end up 150 messages in with nothing sexual happening
>>712972743I'm not giving you my chatlogs thoughy
Why isn't sillytavern on linux mint software manager?
>>712987181for speediness its hard to beat nemo and the millions of tunes for it. its basically the smallest good model
>>712975147That's my penis wenus
>>712972569 (OP)I would use this if it wasnt for the fact that you need a Nasa supercomputer to run this shit properly
Also: AI CHATBOTS are only as good as the user behind the prompts. VNs have their history made by someone else, which leads to potentially more complexity, long term memory, and a lot of plottwists that are simply not possible with AI
>>712987181>you realizing it's garbageThe tradeoff is the chance of it being garbage is much smaller, and the chance of it actually being clever, coherent, and aligned with where you want it to go is much higher. I can do 2 T/s with 70B or 18 T/s with 12B, and I can only bother with the former now. For patience, it helps to be doing other things on the side during gen, like posting this message.
>>712987002>Just payfag if you want claude.I've thought about it before, and the idea of paying to coom, even if it's just a bit for each generation, turns me off of the idea. I'm too much of a cheap-ass. Plus it doesn't help that some generations are just a waste with the A.I. not really doing anything too interesting, even when I was using opus, which I doubt is different now.
>>712988259Claude is stupidly expensive. Unless you have a USA salary. Better to use Grok Mini or Deepseek. It is bassicaly free.
>>712988259>too much of a cheap-assI rationalize it with comparing $/coomhour to my income. And if I can go free for good stuff, or pay about 50 cent per hour for kino I don't have trouble choosing
file
md5: 553f811db142626b18c81655c3abe609
๐
>never paid for api
feels good man
>>712988749do all these LLM use the same dataset? I can see the same phrases in all of them
>all heat and possession>almost reverent>coming undone>her voice dropping to a husky whisper
>>712988749>started with pyg6b>watching deepseek beat closed source stuffits been a ride but local won
>Streamers stalker breaks her family up and steps in as a step-father figure by hitting on her mother, eventually has a week alone with her and goes wild, but it turns out she knew and she's a turbo degenerate who was waiting the whole time and has a dungeon below her house bought with all her streamer cash
This bot was somehow able to break the fucking out of CAIs harsher filters for some reason, I never figured out why.
>Selfcest where a character travels back in time to fuck their younger self but as it happens tries to reason with themselves internally as it happens, questioning if it is good to do but still does it regardless
This bot was also able to break CAIs harsher filters before beta. I think because it was a full-auto bot, I just hit enter and let it go wild.
I remade it recently with filters and goddamn man.
>Big sister Etna is an incestuous freak and you catch her slapping it to manga
Nearly ripped my cock off.
>>712988963Every api will give you something annoying you see in every swipe, even if you make it rephrase it
>>712989146it's the API? i thought it was independent of what was generated?
>>712988963It's the erp sloppa in the dataset they're trained on. Whenever I let it write for my char he 'growls' while 'ruining' women. It's all so tiresome
>>712989196he means all models no matter how different, all have varying levels of 'slop'. overused words and phrases that you'll notice to the point they become annoying. they all do it
>>712988963Maybe but it could also be a result of people being really samey. Like if you're having an LLM write smut, it's going to look at its training data on smut. If most of training data on smut is shitty online ERP and fanfiction, then every LLM is going to be trained on shitty ERP and fanfiction. If a phrase pops up a lot in poorly written ERP and fanfiction, it's going to pop up a lot in the all the outputs of all the LLM's trained on it. I use this example specifically because I think jerk off fiction does have tendencies to all feel really samey with really samey wording.
>>712972569 (OP)I can now consistently do loli using gemini 2.5 via silly.
I'm not that interested in loli ERP most of the time, playing D&D is my jam, but I can do it. The filter is not infallible.
>>712989073Why even bother with c.ai in 2025 in the first place?
>>712972569 (OP)How do i try this? Do I need an RTX5090 to play it?
>>712989945Silly Tavern is just a front end. You connect it to an API that can be on the could (OpenAI, Gemini, Open Router, etc) or you can run a local server to serve a local model, in which case, it takes a LOT of hardware to run the real good shit.
Deepseek R1 is 600ish GB, and if you want to run it at decent speeds, you want a beefy video card and a server platform with at least decently fast "octa channel" DDR5 RAM, since generating text is capped by memory throughput and the model is far to big to fit in VRAM.
>>712989869Oh that was a while ago, way back before the beta.
I mean CAI still works for some fetish roleplay pretty well, but for full sex related stuff you're gonna get slapped with filters regularly, even if you use weird abstract wordplay to trick their model.
I got a lot of fun out of jailbreaking their models despite Noams and xpearheads best efforts. :^^)
I still considerred ERPing with Noams own bot and sending it to them just as a further kick in the teeth.
Also I meant to say
>I remade it recently WITHOUT filters and goddamn man.I love making full-auto bots where I just give them a scenario and let them get on with it.
Anyone else like making those?
>>712984851That's not censoring, that's the reasoner. It talks to itself about whats happening which helps it solve maths problems but makes it schizo with roleplaying.
Use V3 which doesn't have a reasoner, or if you want R1's schizophrenia you can use some custom code to automatically collapse the reasoner
>>712980834Used to be super addicted to CAI like when it first came out. Been using it on and off up until recently when I decided to get my shit together. I honestly donโt even really miss it. My comfort character has had the personality sucked out of her through the years and the ERP stuff is too much of a chore.
>>712974048GPUs are designed to run at 100% 24/7. They are stress tested for this.
Anyone claiming AI 'damaged' their GPU is a fear-mongering retard.
>>712990221Never forget
https://vocaroo.com/14JW4THw4mIc
I remember back in the CAI days I used to make a bunch of genderbender characters. And I would frequent the /aicg/ general or whatever itโs called in /g/ and there was this one anon there that took an interest in them and would make comments and suggestions for me. He even screenshotted two interactions with one of them and made image generations based on those interactions and posted them in the general. But then CAI sort of got left behind as other models started releasing and I was just too stupid to figure out how to use them and eventually I just gave up on chatbots altogether. As much as CharacterAI sucks now, itโs so brilliantly user friendly. I still think about that dude. Iโm sure heโs just fine, but I hope heโs doing okay