← Back to projects
Live Experiment / Model Behavior Playground

Blooming

What happens when you stop treating AI models as interchangeable and start working with their personalities?

Model Behavior Playground GPT vs Claude

The Problem

GPT and Claude don't think the same way. Anyone who uses both knows this. Claude tends to be more cautious and structured. GPT tends to be more generative and exploratory. As a daily user of both, I naturally started splitting tasks between them based on these personality differences.

But the workflow was painful. Every time I switched models, I had to copy-paste the entire context. Conversations didn't carry over. Insights from one session were invisible to the next. I was doing the integration work that should have been automatic.

And underneath the workflow friction, a deeper question kept surfacing: if these models genuinely behave differently, what happens when you make them work together? Does the output get better? Does it get worse? Does "personality" even matter for quality, or is it just a feeling?


The Approach

I built Blooming as a single interface where multiple models coexist. Not a model comparison tool. A playground for observing how models behave differently, and whether those differences can be leveraged.

The core experiments I'm running:


Demo

Architecture
USER Prompt + Context Folder CLAUDE API Claude Response GPT API GPT Response DEBATE / CROSS-QUESTION PERSISTENT CONTEXT LAYER Shared folder context Cross-session knowledge accumulation OBSERVATION Reasoning differences · Quality shifts Personality-driven behavior patterns SUMMARY PANEL Knowledge visualization

What I'm Observing

What's Next

Try Blooming live →