@jerryh100@lemmy.world Wrong community for this kind of post.
@BaroqueInMind@lemmy.one Can you share more details on installing it? Are you using SGLang or vLLM or something else? What kind of hardware do you have that can fit the 600B model? What is your inference tok/s?
That’s why I wanted to confirm what you are using lol. Some people on Reddit were claiming the full thing, when run locally, has very little censorship. It sounds somewhat plausible since the web version only censors content after they’re generated.
@jerryh100@lemmy.world Wrong community for this kind of post.
@BaroqueInMind@lemmy.one Can you share more details on installing it? Are you using SGLang or vLLM or something else? What kind of hardware do you have that can fit the 600B model? What is your inference tok/s?
Nor really, 196 is a anything goes community after all.
AI generated content is against the community rules see the sidebar :)
I’m here for the performative human part of the testing. Exposing AI is human generated content.
I just really hope the 2023 “I asked ChatGPT <abc> and it said <xyz>!!!” posts don’t make a comeback. They are low-effort and meaningless.
True. This specific model is relevant culturally right now though. It’s a rock in a hard place sometimes lol
just giving context to their claim. in the end it’s up to mods how they want to handle this, i could see it going either way.
Prompts intended to expose authoritarian censorship are okay in my book
I’m using Ollama, a single GPU with 10Gb of VRAM
You’re probably running one of the distillations then, not the full thing?
What’s the difference? Does the full thing not have censorship?
That’s why I wanted to confirm what you are using lol. Some people on Reddit were claiming the full thing, when run locally, has very little censorship. It sounds somewhat plausible since the web version only censors content after they’re generated.