I tried Grok 3 (the standalone app) a bit. So far when you ask it for instructions I find it doesn’t make mistakes like GPT 4o. And it gives much more detail which is nice. Asked it to generate some Pepe’s. For that GPT 4o does it way better. The head doesn’t look like a Pepe that much with Grok and the rest of the image is mediocre. Haven’t tried it for much else yet.

Reply to this note

Please Login to reply.

Discussion

Looking forward to the pepes

You’ve seen a few before 👀

I have yet to find an AI that can make a decent wojak. This is the real test.

I asked gpt 4o to draw a wojak reading a book

Ok that's pretty good. But this is a very generic one. Can it make an ostrich wojak? Aka a #nosjak

#nosjak still unbeaten. I remember trying with something or other a while ago before I was sent this one. They all failed miserably.

😂😂

Ha, still not as good but better than what I got lolol

I’m trying to get the wojak head to scream but I’m getting weird results. 2 necks connecting the head 😂

Today is the day I learned what a Wojak is…..I think

😂

Oh GPT 😂😂

Using o3

Similar to what I got. Are you able to get the wojak to scream without having 2 necks? 😂

I mainly tried it to compare my newer version, o3 (new 2 days ago I just learned) to what you had. It was quite fast, suspect faster than yours was, but similar outcome.

What's the prompt exactly, so I can match you?

Made it work

“Draw a screaming ostrich with the head of a wojak”

o3 is better than 4o ?

Supposed to be. Jack posted couple hours ago it may be tipping point 👀

Gotta love their fd up naming system eh

Asked 4o to compare o3 and 4o. It wasn’t able to 😂

I don't even try to keep up with AI. But if Jack says this is big, I listen 🤷‍♀️

What Grok 3 says:

Comparing GPT-4o and o3 (developed by OpenAI) involves looking at their design, capabilities, and intended use cases. Below is a detailed comparison based on available information:

1. Overview

• GPT-4o:

◦ Released on May 13, 2024, GPT-4o (“omni”) is OpenAI’s multimodal model, capable of processing and generating text, images, audio, and potentially other data types.

◦ It’s optimized for efficiency, cost-effectiveness, and broad applicability, powering the free version of ChatGPT and available via OpenAI’s API.

◦ Focuses on seamless human-computer interaction with fast response times and improved performance over previous models like GPT-4.

• o3:

◦ Released in April 2025, o3 is part of OpenAI’s advanced reasoning-focused series, introduced under the “Project Strawberry” initiative.

◦ Designed for complex tasks requiring deep reasoning, such as coding, mathematics, and scientific problem-solving, with a chain-of-thought approach.

◦ Available via API, with a smaller variant, o3-mini, optimized for specific tasks like coding.

2. Key Differences

Aspect

GPT-4o

o3

Release Date

May 13, 2024

April 2025

Primary Focus

Multimodal capabilities (text, images, audio) for general-purpose tasks

Advanced reasoning for complex tasks (coding, math, science)

Context Window

128,000 tokens

200,000 tokens

Output Tokens

Up to 16,384 tokens per request

Not specified, but likely higher for complex reasoning tasks

Performance

- 85.4% on MMLU (general knowledge)

- 86.6% on Human Eval (coding)

- 87.7% on GPQA Diamond (expert-level science)

- 71.7% on SWE-bench Verified (coding)

Speed

Responds in ~320 ms (voice), faster for text-based tasks

Slower due to reasoning process (up to 30x slower than GPT-4o)

Cost (API)

$2.50/M input tokens, $10/M output tokens

~4x more expensive than GPT-4o (exact pricing not specified)

Multimodal

Supports text, images, audio, and potentially video

Includes vision capabilities, but primary focus is reasoning

Training Data

Up to October 2023

Up to May 31, 2024, newer and likely more specialized

3. Capabilities

• GPT-4o:

◦ Multimodal Excellence: Excels in tasks involving diverse inputs (e.g., analyzing images, processing audio, generating text). It’s ideal for applications like real-time chatbots, content generation, and customer support.

◦ General Knowledge: Performs well across a wide range of topics, with strong language understanding and generation (89% accuracy on complex queries, 87% precision).

◦ Efficiency: Twice as fast and half the cost of GPT-4 Turbo, with a response time of ~0.9 seconds for text tasks.

◦ Use Cases: Customer service chatbots, creative writing, translation, summarization, and multimodal applications (e.g., voice assistants).

◦ Limitations: While capable of reasoning, it may struggle with highly complex, multi-step problems compared to o3.

• o3:

◦ Reasoning Prowess: Designed to “think” through problems step-by-step, making it superior for tasks requiring deep analytical skills. It supports three reasoning effort levels (low, medium, high) to balance latency and accuracy.

◦ Specialized Performance: Outperforms GPT-4o in benchmarks like AIME 2024 (96.7% vs. GPT-4o’s 64.5% on MATH benchmarks) and SWE-bench (71.7% for coding).

◦ Developer Features: Supports function calling, structured outputs, and developer messages, making it ideal for integration into complex workflows.

◦ Use Cases: Software engineering, scientific research, mathematical modeling, and tasks requiring precise reasoning.

◦ Limitations: Slower and more expensive, less suited for real-time or high-volume tasks. Multimodal capabilities are secondary to reasoning.

4. Cost and Accessibility

• GPT-4o:

◦ More affordable, with API costs at $2.50/M input tokens and $10/M output tokens.

◦ Powers the free version of ChatGPT (with message limits) and is available to paid users (ChatGPT Plus, Pro, Team, Enterprise) with higher quotas.

◦ Widely accessible for general-purpose applications.

• o3:

◦ Significantly more expensive (roughly 4x the cost of GPT-4o), reflecting its computational intensity.

◦ Available only via OpenAI’s API, targeting developers and businesses with complex needs.

◦ o3-mini offers a cheaper, distilled version for specific tasks like coding.

5. Ethical and Safety Considerations

• GPT-4o:

◦ Improved bias reduction (5% biased responses vs. 8% for GPT-4) through diverse training data and refined techniques.

◦ Still faces challenges with potential misinformation or harmful content, requiring careful deployment.

• o3:

◦ Likely incorporates similar or enhanced safety measures, but specific details are sparse.

◦ Its reasoning focus may reduce hallucinations in technical tasks, but ethical concerns (e.g., misuse in generating misleading scientific content) remain.

6. Which Model to Choose?

• Choose GPT-4o if:

◦ You need a cost-effective, fast, and versatile model for general-purpose tasks.

◦ Your application involves multimodal inputs (text, images, audio) or real-time interactions (e.g., chatbots, voice assistants).

◦ Budget and speed are priorities, and moderate reasoning capabilities suffice.

◦ Example: Building a customer support chatbot or generating creative content.

• Choose o3 if:

◦ Your project requires advanced reasoning for complex, technical tasks (e.g., coding, math, scientific analysis).

◦ You need a larger context window (200K tokens) for processing extensive datasets or long conversations.

◦ Budget is less of a concern, and you prioritize accuracy over speed.

◦ Example: Developing a tool for automated code debugging or scientific research.

7. Conclusion

GPT-4o and o3 serve distinct purposes: GPT-4o is a versatile, multimodal workhorse for broad applications, offering speed, affordability, and strong general performance. In contrast, o3 is a specialized reasoning engine, excelling in complex, analytical tasks but at a higher cost and slower speed. Your choice depends on whether you prioritize efficiency and multimodal capabilities (GPT-4o) or deep reasoning and precision (o3).

For more details on pricing or API access:

• GPT-4o: Visit https://platform.openai.com

• o3: Check https://x.ai/api for xAI’s API offerings, though o3 is OpenAI-specific

Note: Always verify current pricing and availability, as AI model costs and access can change rapidly. If you have a specific use case in mind, I can tailor the comparison further!

Skimmed some of that. Better for math and reasoning was my main takeaway, which is actually what I want. Supposed to do better on health analysis too according to bit of jacks YouTube I watched... might be of interest to you. Says more expensive and slower in ur post... huh, opposite of what I thought. Oh well

So I could feed it my apple health data and it could recommend some stuff. Could be fun

Yeah, apparently this dude Jack linked video of goes deep on all these things and was blown away. He loads his health stuff and got some improved insights he claims, though no idea what he is sick with.

Do it, for science!!

The guy is sick? And gpt is helping him?

Seems so...I am not familiar with him

Claimed it replaced his doctor

That’s extreme. I mean good for him if it works for his specific issue

Grok recently got worse at image generation and chatgpt got waaaay better.

The leaders in each area will leapfrog in various areas and then there will be niche AIs that dominate for periods of time.

Theres a pod called “last week in AI” that is pretty interesting. I keep finding new uses everyday.

🧡👊🏻🍻

I’ll look it up. Thanks!

GPT 4o can straight up frogpost better than any AI so far. Ask it to make you a greentext with pepe imagery. 🤣

Wow not bad 😂

LMAO! It even got the sad pepe right. 😂

Got the neet part too 😂

I was told AI was going to replace the workforce. Instead here we are generating pepes.

Generating Pepes > replacing the workforce 😂