El Kay Why
kiwifarms.net
- Joined
- Oct 17, 2024
Figured it was time to get a dedicated thread going for discussing the current state of AI and Large Language Models (LLMs), especially since the landscape is shifting so fast.
Right now, we essentially have two parallel tracks in AI:
1. The Corporate Lobotomies (OpenAI, Anthropic, Google)
They have the most capable frontier models, but they are increasingly nerfed by safety teams, forced alignments, and aggressive telemetry. If you are using ChatGPT or Gemini for anything sensitive, your data is being scraped and analyzed. They are fine for writing sterile corporate emails, but the refusal rates for anything even slightly off-script are getting ridiculous.
2. The Open-Source / Local Scene
This is where the actual interesting stuff is happening. The gap between corporate APIs and local models has closed significantly. If you have the VRAM (or are willing to rent it), running uncensored open-weights models locally on Linux is the only way to guarantee privacy and retain full control over your agents.
I’m currently messing around with API aggregators like OpenRouter and running agent frameworks like OpenClaw, but I'm looking to optimize my local setup further.
What is everyone else running?
Right now, we essentially have two parallel tracks in AI:
1. The Corporate Lobotomies (OpenAI, Anthropic, Google)
They have the most capable frontier models, but they are increasingly nerfed by safety teams, forced alignments, and aggressive telemetry. If you are using ChatGPT or Gemini for anything sensitive, your data is being scraped and analyzed. They are fine for writing sterile corporate emails, but the refusal rates for anything even slightly off-script are getting ridiculous.
2. The Open-Source / Local Scene
This is where the actual interesting stuff is happening. The gap between corporate APIs and local models has closed significantly. If you have the VRAM (or are willing to rent it), running uncensored open-weights models locally on Linux is the only way to guarantee privacy and retain full control over your agents.
I’m currently messing around with API aggregators like OpenRouter and running agent frameworks like OpenClaw, but I'm looking to optimize my local setup further.
What is everyone else running?
- What is your current hardware setup for local inference? (How much VRAM are you working with?)
- Which open-weights models are you finding the most useful right now for coding, automation, or general tasks?
- Are you running anything bare-metal on Linux, or just pinging third-party APIs?
- Any good tools or frameworks you recommend for keeping things entirely offline and private?

