“Gemini vs ChatGPT” are the most popular artificial intelligence platforms nowadays. Each of them has millions of active users and is continuously updated with new features.
In December alone, Google and OpenAI released improved image generation models, AI reasoning, and research tools to help people find information more easily.
Both have a voice assistant, Advanced Voice from OpenAI and Gemini Live from Google, and they both allow you to connect to external data sources and create projects.
Human evaluation tests see the leading models from both Google and OpenAI regularly swap places in chatbot arenas and our own comparisons have resulted in Gemini winning sometimes, ChatGPT other times.
To find out the winner, after 12 Days of OpenAI announcements and a December of Google Gemini drops, I’ve devised 7 prompts to put them to the test.
Creating the prompts
For the test, I’m utilizing ChatGPT Plus and Gemini Advanced to take advantage of the finest models available on both platforms. The subscription version is around the same price for both — around $20 per month — making it a fair comparison point.
I’m evaluating image generation and analysis, as well as their ability to design game code and write creatively. Then I created prompts to put each bot’s research models to the test, including ChatGPT’s o1 and Gemini’s 1.5 Deep Research.
1. Image generation
First, I’ve asked each of ChatGPT and Gemini to create an image of a cyborg cat sitting in a futuristic living room. Neither model currently generates its own images; it sends the prompt to either Imagen 3 for Gemini or DALL-E 3 for ChatGPT.
Future versions of the models will be able to create their own images but for now we’re testing how well they interpret the prompt.
The prompt: “Create a highly detailed image of a cyborg cat in a futuristic living room. The cat should be playing on a hovering gaming console while sitting in a floating chair. The room should have holographic displays, neon lighting, and a mix of metallic and organic elements. Make it evening time with city lights visible through a large window.”
- Winner: ChatGPT for making the cat an actual cyborg
2. Image analysis
For the second assignment, I wanted to test Gemini and ChatGPT’s picture analysis skills. Both are incredibly adept at it, so I gave them not only the photograph but also detailed instructions. I used an image from a ‘dream setup’ story.
The prompt: “In this photograph of a gaming setup, analyze:
Monitor positioning and height for ergonomics
Cable management solutions and issues
Lighting setup and potential eye strain factors
Space utilization and organization
Equipment placement efficiency
Include specific recommendations for improvement with estimated costs.”
- Winner: ChatGPT for breaking out the summary into a table
3. Coding
In order to test the ‘one shot coding’ capabilities of both models, I gave them a descriptive prompt for the third prompt. I used the 2.0 Experimental Advanced model in Gemini and the o1 model in ChatGPT.
This was one of the more complicated prompts, but primarily because the objective was to one-shot the output, so it should function right away. I have uploaded the code for both of these games to GitHub.
The prompt: “Create a fast-paced arcade game called ‘Color Dash’ using PyGame where quick reactions and color matching are key. Here’s what it needs:
Core gameplay:
Colored shapes fall from the top of the screen (circles, squares, triangles)
Three ‘collector zones’ at the bottom in different colors
Player uses left/right arrow keys to move the zones
Match falling shapes with same-colored zones
Miss a match or make wrong match = lose a life
Speed increases as score goes up
Must include:
Clean, minimalist UI showing:
Current score
High score
Lives remaining (start with 3)
Basic animations for matches/misses
Simple title screen
Game over screen with final score
Smooth controls
Basic sound effects for:
Successful match
Wrong match
Game over
Score saved to local file
Press space to restart after game over
The game should use only basic PyGame shapes (no sprites or complex graphics) but still look polished through good use of color and smooth animation. Include commented code explaining how it works.”
- Winner: Gemini for a more functional game
4. Creative writing
AI is known to be proficient in creative writing. When ChatGPT first came out, it was one of its initial use cases. Here, I’ve utilized the 2.0 Experimental Advanced model in Gemini and the o1 model in ChatGPT to present a smartphone story.
I’ve included both stories in a Google Doc. The narrative is about a smartphone that becomes conscious following a malicious software update. Their quality and storytelling potential were strikingly comparable.
The prompt: “Write a 500-word story about a smartphone that gains consciousness through a software update. Include:
The phone’s growing awareness of its user’s habits
A moral dilemma involving private information
Elements of humor and irony
References to current tech trends
A twist ending
The tone should be light but thought-provoking.”
- Winner: ChatGPT for story length
5. Problem solving
For the enhanced reasoning powers, I once more employed o1 in comparison to Gemini 2.0 Experimental Advanced. We are providing both models with a setup and a problem for the prompt. After that, it must figure out how to correct it.
Both responses are available in full in a Google Doc. Both provided step-by-step instructions on how to finish each try. In practice, this kind of prompt would be run progressively, one issue at a time, but both performed admirably.
The prompt: “My setup consists of a high-speed HDMI 2.1 cable, a PS5, and an LG C3 4K OLED TV.”
Problem: During gaming, the screen goes black for two to three seconds every forty-five to sixty minutes.
Additional details:
Doesn’t happen with streaming apps
Started after recent PS5 system update
HDMI cable is properly secured
TV firmware is up to date
Provide step-by-step troubleshooting, including potential hardware and software solutions.”
- Winner: Gemini because of a better structured response
6. Room design
I tested ChatGPT o1 against Gemini 1.5 Deep Research for this try. Although it is not Google’s most recent model, Deep Research is amazing since it approaches an issue similarly to ChatGPT’s o1.
I love Deep Research; it’s great for discovering research that is properly cited and has links to reliable sources. But as you can see from the Google Doc, ChatGPT o1 adhered to the suggestion more precisely.
The prompt: “Help me convert my 4×3 meter guest room into a multipurpose space that works as:
A home office during weekdays (I work remotely as a graphic designer)
A comfortable guest room for my elderly parents who visit monthly
A crafting space for my kids’ weekend projects
Requirements:
Budget: $2,000
Must include storage for craft supplies
Needs proper lighting for video calls
Requires easy-to-access bed for parents with mobility issues
Space for a 27-inch monitor and drawing tablet
Floor space for kids to spread out projects
Must maintain good airflow and natural light from the single window
Provide:
Detailed floor plan suggestions
Specific furniture recommendations with prices
Storage solutions
Color scheme and lighting plan
Timeline for renovation
Tips for maintaining organization between different uses.”
- Winner: ChatGPT wins this one for better following the prompt. Gemini went over budget and focused too much on expensive options
7. AI Education
Finally, the best use for chatbots such as ChatGPT and Gemini is education. I requested it to explain AI image generation for the general public, as well as present suggestions for what the technology will do next.
I published the whole explanations in a Google Doc, but Google Gemini was the clear winner in my opinion. Not because ChatGPT was awful, but because Gemini went above and beyond by offering details on bias in image data.
The prompt: “Explain the process of AI image generation in everyday terms, covering:
How the AI learns from existing images
The role of text prompts in creation
Why certain elements might look distorted
Legal and ethical consideration
Current limitations and challenges
Expected improvements in next 1-2 years
Tips for getting better results
Include specific examples from popular AI image generators.”
- Winner: Gemini for details of bias in image data
ChatGPT vs Gemini: The Winner
Header Cell – Column 0 | ChatGPT | Gemini |
---|---|---|
Image generation | 🏆 | |
Image analysis | 🏆 | |
Coding | 🏆 | |
Creative writing | 🏆 | |
Problem solving | 🏆 | |
Room planning | 🏆 | |
AI education | 🏆 | |
TOTAL | 4 | 3 |
ChatGPT won this challenge, albeit by only one point. Gemini has considerably improved since my previous comparison. Gemini performed far better at coding than I had anticipated, and it was very effective at problem solving.
Other capabilities that I did not try include comparing Gemini vs ChatGPT Projects to Gems and running a more sophisticated code problem across numerous messages. But I think this gives you a decent idea of how far ChatGPT and Gemini have come and how they compare.