r/GPT3

OpenAI just spent $20 billion on new chips to make ChatGPT dramatically faster
▲ 25 r/GPT3+2 crossposts

OpenAI just spent $20 billion on new chips to make ChatGPT dramatically faster

Sam Altman just made the biggest infrastructure bet in AI history.

OpenAI is spending over $20 billion on a new type of computer chip specifically designed to make ChatGPT faster, smarter and more responsive than ever.

These aren’t regular chips. Cerebras builds wafer-scale processors (the largest chips ever manufactured) that pack the processing power of an entire server rack onto a single piece of silicon the size of a dinner plate.

The goal is simple. Make ChatGPT respond in real time with zero lag. Enable AI agents that can think and act for minutes or hours without stopping. Process your requests faster than you can type them.

OpenAI is also taking an equity stake in Cerebras, meaning the company behind ChatGPT now partially owns the company building the hardware powering it.

For context, $20 billion is more than most countries spend on their entire technology infrastructure in a year.

OpenAI already surpassed $25 billion in annual revenue. They are now spending most of it on getting faster.

900 million people use ChatGPT every week. Every single one of them is about to notice the difference.

u/ComplexExternal4831 — 20 hours ago
▲ 48 r/GPT3+3 crossposts

Turned Claude's rough week into an excuse to build an OpenCode-compatible version of my D&D skill

Claude has had a rough week. Between the outage and the usage limit threads, I figured it was actually good timing to do something I had been meaning to try anyway: take the D&D skill I built a few weeks ago and see if I could migrate it to run on OpenCode with free or local models. If Claude is your DM and Claude goes down mid-session, that is a problem worth solving.

The short version: it works, and it was easier to set up than I expected.

What I built

open-tabletop-gm is a fork of the original claude-dnd-skill, rebuilt to run on any LLM through OpenCode. OpenCode supports Anthropic, OpenAI, Google, Ollama, LM Studio, and any OpenAI-compatible endpoint, so you can point it at whatever is available. Free tier models, local models, a different provider entirely.

The Claude-specific parts (model routing between Haiku/Sonnet/Opus, the ~/.claude/ path structure, autorun) have been replaced with portable equivalents. The campaign files, display companion, and Python toolchain are all identical.

While I was at it, I also pulled D&D 5e out of the core and turned it into a system module. The GM core (pacing, NPC craft, improvisation, consequences) lives in one file and knows nothing about any specific game. D&D 5e lives in a separate systems/dnd5e/ folder. If you want to run Vampire: The Masquerade, Cyberpunk RED, Pathfinder, or any other TTRPG, you write a system.md describing your game's dice resolution, stats, health model, and conditions - and the same GM core runs it. There is a porting guide covering what transfers directly from the D&D implementation vs what needs configuring per game. D&D 5e is the reference implementation and ships fully built out. Everything else is a system.md away.

Why smaller/free models hold up better than you might expect

The Python toolchain carries a lot of the weight that would otherwise fall on the model:

  • Dice rolls, HP math, damage tracking: Python
  • Initiative and turn order: Python, tracked in a live sidebar
  • Timed effects and conditions: Python, file-persisted
  • SRD data lookup (spells, monsters, items): local JSON

The model's job is narration and judgment. It reads the campaign state from plain Markdown files and narrates from there. It does not do arithmetic and does not need to hold mechanical state in memory. That separation is what makes free and smaller models viable: the parts that tend to break on constrained models have been moved out of the model entirely.

First test: MiniMax M2.5 via OpenCode

Tested against the original claude-dnd-skill version. Setup was surprisingly frictionless -- OpenCode picked up the skill file without extra configuration. The model produced creative NPC responses and correctly read deceptive intent in a player message. More than I expected from a first pass on a free tier model.

Current testing: Qwen3-32B via LM Studio

Working well on the portable version so far. Script calls reliable, narration solid, campaign state persisting correctly across sessions. Testing is being pushed down toward Qwen3-14B to find the practical floor. Results going into the LLM guide as they come in.

What stays the same

Everything you already know from the original skill: persistent campaigns, the cinematic display companion you can Chromecast to a TV, character sheets, the DM philosophy, NPC memory, all of it. The system module architecture now lets you run any TTRPG, not just D&D 5e, by writing a system.md for your game. But if you are running D&D the experience is the same.

Claude is still the better DM

To be clear: this is not a "switch away from Claude" post. Claude Code with claude-dnd-skill is still the better experience. Better narration, model routing, deeper integration. If Claude is up and you have quota, use that.

But having a version that works when it is not is genuinely useful. And honestly, testing it has been a good reminder of how much the Python toolchain is doing independent of any specific model.

Links

u/Bobby_Gray — 1 day ago
▲ 4 r/GPT3

Is anyone else noticing that ChatGPT seems to be completely down for everyone right now?

I got booted from ChatGPT on all my devices, and now I'm just getting hit with error messages whenever I try to log back into my account.

reddit.com
u/Careful_Ad_1580 — 13 hours ago
▲ 0 r/GPT3

I used GPT-3 to generate working SQL from plain English questions

u/roretec — 2 days ago
▲ 5 r/GPT3

"Google hires record number of interns 2023"

u/roretec — 3 days ago
▲ 3 r/GPT3+1 crossposts

Built something to track Instagram trends like Google Trends - would love feedback

Hey everyone,

I’ve been working in performance marketing for a while, and one problem kept coming up again and again:

There’s no reliable way to see what’s actually trending on Instagram in real-time.

Most of us just:

- guess trends

- copy viral creators

- or rely on outdated hashtag tools

So I decided to build something for myself.

It basically works like a "Google Trends for Instagram" where you can:

- discover trending reels & content formats

- see what’s performing right now (not last week)

- get quick insights instead of vanity metrics

Still very early, but it’s been super useful for me internally.

I’m thinking of opening it up for creators & marketers.

Would genuinely love feedback:

👉 Is this something you’d actually use?

👉 What would you want in a tool like this?

Happy to share access if anyone wants to try it.

reddit.com
u/Kunalkr27 — 5 days ago
▲ 2 r/GPT3+1 crossposts

HOT TAKE: AI should write Letters of Recommendations

During my time as a high school senior, I noticed how much work counselors take on during the college application cycle. In my graduating class, there were over 700 students, which also means over 700 letters of recommendation. The purpose of this letter is to show college admissions officers something that the transcript, activities list, and test scores can’t: a unique account of the student from a person who knows them well. Given the number of seniors and the number of counselors writing these recommendations, getting to really know each senior and writing a thoughtful and unique letter of recommendation is an extremely difficult, if not impossible, task.

I then asked myself, "Why aren’t high school counselors using AI more extensively to write these letters of recommendation, and why is there such a bad stigma against it?” Although there are privacy and legal concerns with AI (ex. FERPA), I believe most counselors don’t use AI more extensively in their writing process because of authenticity, or lack thereof. This letter should capture who a student is, and current AI tools can’t do this effectively. To admissions officers, it is not the act of using AI to write a recommendation that causes red flags; it is the generic and often bland output that comes from poor use of AI (the exact problem with templates as well).

However, counselors were never meant to spend all of their time on the computer; they were meant to be one of the people who know the students best! If counselors spend 1 hour talking with students for every 4 hours of writing, with my project, I aim to make that 4 hours of talking for every 1 hour of writing.

With a few of my friends in college, we are creating software that helps solve the exact problem stated above. We want to integrate AI into the process of writing college letters of recommendation to help counselors have the time to get to know their students better while also being able to write unique and authentic pieces of writing.

Our software is a two-step process: reading and writing. Given the student source material (meeting transcript, brag sheet, etc), our software generates provocations in the margins of these documents. These are annotations that are meant to provoke ideas and thoughts while you read, not to summarize. As you read, you can reply to these provocations, highlight, and leave short memos: anything that comes to mind. The next step is the writing process. Our software compiles all of your input into a first draft, where the process of provocations, input, and drafting can continue. What we do and what other AI tools have failed at is to have every sentence in this piece of generated writing be 100% from your memos and thoughts, not the AI, all while saving the writer the time of physically typing every word. At core, our philosophy is that AI is a good provocateur and also good at clean, logical, and persuasive writing. Letting AI do this lets humans, counselors, do what they are good at: finding a unique story or perspective for a student. Throughout this process, the LLMs used would be pre-approved by the school and would be completely secure models that never release student information.

reddit.com
u/Acrobatic_Belt4217 — 6 days ago
▲ 0 r/GPT3

Sam Altman and Dario Amodei ruined everything.

The contrast is actually depressing.

We used to have Elon Musk pushing for Mars and Demis Hassabis (DeepMind) solving protein folding to cure diseases.

That was "humanity-shifting" energy.

Now, Altman and Amodei have turned AI into a mid-wit hype cycle of:

  • Meaningless benchmarks
  • Model vs. model ego trips
  • Weekly "revolutionary" updates that do nothing

We went from "colonize planets" to "did we beat GPT by 2%?"

Look at the OpenAI founding team—literally every single one left because the "Open" part became a corporate lie.

Anthropic claims to care about "safety," yet their talent is already jumping ship.

These guys care more about winning a race than why we're running it. We traded real innovation for a subscription-based dead end.

u/Alone-Possibility398 — 11 days ago
▲ 0 r/GPT3+1 crossposts

"OpenAI quietly removed the one safety mechanism that could shut the whole thing down — and nobody is talking about it"

OpenAI was founded as a nonprofit for one specific reason — to ensure AI development couldn't be hijacked by profit motives.

Their original charter had a clause that legally required safety to come before profits, and gave the board the power to shut everything down if AI became too dangerous.

That clause is gone. The board has been restructured to answer to investors instead.

We just removed the emergency brake from the most powerful technology in human history because it was bad for business.

What happens the next time something goes wrong?

youtube.com
u/kc_hoong — 12 days ago
▲ 1 r/GPT3+1 crossposts

Is AI really bad at analyzing personal notes?

I keep structured personal notes — thoughts, decisions, todos. Recently tried feeding them into different AI models (most of all - OpenAI, including top ons, high reasoning effort) to get actual insights

Expectations:

  • pattern detection over time
  • non-obvious connections
  • calling out blind spots / repeated behaviors

What I got:

  • generic advice
  • surface summaries
  • basically rephrasing what I already wrote

Even when I push prompts it still stays kinda obvious

Feels like either models are too generic or just not good at this kind of deep personal context

Has anyone actually tried something like this?

If yes - how good is the result?

reddit.com
u/AndreyKypaku — 11 days ago
▲ 6 r/GPT3

Has anyone chosen to stick with the original Cove voice instead of the advanced voice?

I was already using the Cove voice when the advanced voice mode started rolling out. From what I remember, it was automatically enabled for me. But honestly, I couldn’t really adapt to it.

It’s not that the advanced voice is bad at all. It has more features and more possibilities. But for me, it felt like something was missing. That natural, more “human” presence I had with the original Cove voice.

Maybe it’s just habit, I don’t know. But I ended up sticking with the original Cove voice, even if that meant giving up the new features.

Just wondering… am I the only one?

reddit.com
u/Mysterious_Engine_7 — 13 days ago
▲ 4 r/GPT3+1 crossposts

Built persistent text highlighting for ChatGPT

I do a lot of long research sessions in ChatGPT, sometimes 40–60+ messages deep. The problem was always retrieval: useful answers and code snippets got buried, and finding them meant endlessly scrolling back through everything.

So I built a simple highlight system into my Chrome extension.

How it works:
You select any text and press Ctrl+Shift+H. It gets highlighted, saved, and stays there even after refreshes or restarts. There’s also a small navigation bar at the bottom with Previous / Next and a counter, so you can jump between highlights instantly.

Why this is better than just copying things:
Highlights stay in context, so you can still see the surrounding conversation instead of losing where it came from. You can keep working without breaking your flow to copy things out. Then at the end, you can just review everything you marked.

Where I actually use it:

  • Marking action items during planning conversations
  • Flagging useful code snippets while debugging without losing the thread
  • Highlighting the best outputs during brainstorming so I end up with a clear shortlist

It’s free and works on ChatGPT, Claude, and Grok:
https://www.getaiworkspace.com/chatgpt-text-highlighter

u/Strikeh — 12 days ago
▲ 1 r/GPT3

Finally solved the "ChatGPT gets slower with long conversations" problem

Been using ChatGPT daily for work (developer here), and there's one thing that's been driving me absolutely crazy for months: the lag.

You know the drill - you start a fresh conversation, responses come in lightning fast. But after 20-30 exchanges? Every. Single. Response. Takes. Forever. That spinning circle becomes your new best friend. Sometimes I'd wait 15-20 seconds just to see the first token appear.

I tried everything:

  • Starting new chats constantly (losing context sucks)
  • Manually deleting older messages (tedious af)
  • Using the "continue in new chat" workaround (breaks the flow)
  • Just... waiting (not great when you're on a deadline)

The thing is, I usually only need the last few exchanges for context anyway. The AI doesn't need my debugging session from 50 messages ago to help me with my current question.

So I finally got fed up enough to actually do something about it. I built an auto-trim feature into a Chrome extension I've been working on. Dead simple concept:

  • Toggle it on
  • Set how many recent messages you want to keep visible (I use 10-15)
  • Older messages get hidden from the DOM
  • ChatGPT thinks the conversation is short → responds fast again

The difference is night and day. Went from waiting 10+ seconds to near-instant responses, even in conversations with 100+ messages.

The technical bit (for those curious): ChatGPT's frontend seems to process/render the entire conversation history when generating responses. By trimming what's visible in the DOM, you're essentially giving it less to chew on. Your full history is still there, just hidden. Toggle it off and everything comes back.

Honestly thought I was going crazy thinking the length affected speed until I tested it systematically. Turns out it's a pretty well-known issue but I hadn't seen anyone automate a solution for it.

Anyone else been dealing with this? Curious if this is primarily a browser/frontend thing or if it also affects API users.

Edit: Since a few people asked - yes it's part of a free extension I made called AI Workspace. Didn't want to make this a promo post, just genuinely excited that this actually works lol. https://www.getaiworkspace.com/

u/Strikeh — 12 days ago
▲ 0 r/GPT3

🇪🇬 The First Open-Source AI Model in Egypt!

https://preview.redd.it/lx9z5y2sxwtg1.png?width=1459&format=png&auto=webp&s=b5547655212d65d54a7be3569fcec3ddfe11144a

https://preview.redd.it/03qjly9sxwtg1.png?width=1445&format=png&auto=webp&s=63c6a210b0a5325b269d5158493c5783eb59539b

Today, with great pride, I am excited to officially announce the first open-source AI model series emerging from Egypt.

The Horus-1.0 series consists of text generation models, fully trained from scratch on trillions of clean training tokens.

Today, I am also proud to announce the release of the first model in the Horus series: Horus-1.0-4B, featuring an 8K context length.

The model is available in 7 different versions:

  • The full version with original weights
  • 6 compressed variants designed to fit different hardware and deployment needs

This provides exceptional flexibility for developers and researchers based on their available computational resources.

Horus is available as an open-source model under TokenAI, and you can explore all available versions along with detailed usage instructions on the official website:

https://tokenai.cloud/horus

You can also easily download and use the model through the neuralnode Python framework, which offers a seamless integration experience with the Horus models.

In addition, Replica Text-to-Speech is fully integrated within neuralnode.

You have access to 20 voices across 10 different languages, including Arabic, allowing easy voice integration with your applications and AI workflows.

Now let’s talk about the scale and significance of this achievement.

Since there are almost no officially announced AI models in Egypt that are fully built and trained from scratch as open-source models, Horus represents a major milestone:

  • Horus is the first open-source AI model built from scratch in Egypt
  • Horus is one of the strongest language models in the Arab world
  • Horus is one of the strongest models globally within its size class

And all of this is backed by numbers and benchmark results.

The Horus model family is:

  • Open-source
  • Fully trained from scratch
  • Multilingual
  • Highly capable in Chain-of-Thought and reasoning
  • Supports Thinking capabilities

The Horus-1.0-4B model outperformed several benchmarks, including MMLU, achieving results higher than well-known larger models such as Qwen 3.5-4B and Gemma 2 9B.

It also surpassed the same models in the more challenging MMLU Pro, and even outperformed Llama 3.1 8B, despite that model being more than twice the size of Horus.

We are looking at a project capable of placing Egypt on the global AI map.

Horus is not the first AI model from Egypt, but it is the first officially announced, fully open-source, fully scratch-trained model from Egypt.

My goal is not only to build a model, but to build a real Egyptian open-source AI infrastructure.

And this is only the beginning of what I believe will become the best AI model in the Arab world.

#HorusAI #OpenSourceAI #LLM #ArtificialIntelligence #Egypt #MachineLearning

reddit.com
u/assemsabryy — 13 days ago