Michael Chen is a solo founder who has built and sold two small software companies and is now building a third. He’s smart, self-directed, and genuinely interested in how to work better.
Over two years, he systematically tested five goal-setting frameworks. Not casually — he ran real tests, tracked outcomes, and documented what worked and what didn’t. The framework he landed on doesn’t have a name. It’s a hybrid, designed with AI, built around his specific situation.
Here’s what he learned.
The Starting Point: No System At All
Before the tests, Michael had what he calls “vibes-based planning.” He had goals in his head, a general sense of what mattered, and a to-do list that was mostly reactive. He was productive in sprints but inconsistent over months.
“I’d have a great week and then a terrible week and there was no pattern I could identify. I wasn’t tracking anything and I wasn’t reviewing anything. I just kept moving and hoped the movement was in the right direction.”
The catalyst for change was a conversation with a more experienced founder who asked him to describe his three-year plan. He couldn’t. Not because he hadn’t thought about it — but because he’d never written it down in any structured way.
That night, he sat down and tried to write a plan for the first time. The result was a three-page document that raised more questions than it answered. He decided to try a proper goal-setting system.
Test 1: SMART Goals (Three Months)
Michael started with SMART goals because they were the most familiar and seemed like the right entry point. He set SMART goals for his product, his personal health, and his learning.
What he tried: Weekly SMART goal reviews. Every Sunday, he’d update each goal with current status and note blockers.
What worked: The clarity. SMART goals forced him to be specific about what he was actually trying to do. “I had a goal that was ‘get better at marketing.’ Running it through SMART made me realize I didn’t know what ‘better’ meant. That was useful — the framework made me face the vagueness I was hiding in.”
What failed: The conservatism. The “Achievable” criterion was the problem. “I kept adjusting my goals down to make them achievable. I was scared to set goals I wasn’t sure I could hit. By the end, I had goals I could definitely accomplish — they just weren’t the goals that actually mattered.”
He also found SMART goals flat. They told him what to do but not why. When he hit the inevitable difficult weeks — the ones where execution felt pointless — SMART goals gave him no motivational anchor.
Assessment after three months: Useful for sub-task planning. Not useful as a primary system for a founder with ambitious, uncertain goals.
Test 2: OKRs (Six Months)
He was excited about OKRs. He’d read “Measure What Matters” and felt like the framework was designed for someone like him — ambitious, metric-focused, building something real.
What he tried: Quarterly OKR cycles. One Objective per area (product, growth, personal), three to four Key Results each. Weekly reviews, quarterly resets.
What worked: The Objective writing. Forcing himself to articulate a qualitative direction — “Build the product into something users love enough to refer” — gave him a north star that SMART goals hadn’t provided. He also appreciated the 70% norm: “It gave me permission to set scary goals.”
What failed: OKRs were designed for teams, and he was alone. “The full OKR system assumed there were other people who needed to see my objectives and align their work to them. I was the only person. So I was doing all this alignment work with myself, which was just overhead.”
He also found Key Results hard to write for creative and strategic work. “I had an Objective about making the product better. What’s the Key Result for that? NPS score? User retention? Those metrics matter but they’re lagging indicators — by the time I know I’ve failed a Key Result, it’s been 60 days and the quarter is almost over.”
By month four, he was spending more time tracking his OKRs than working on his product.
Assessment after six months: Good framework for teams. Useful for goals with clear metrics. Over-engineered for a solo founder with goals that resist quantification.
Test 3: The 12 Week Year (Three Months)
He adopted the 12 Week Year after hitting an execution plateau. He knew what he wanted to do — he just kept not doing it.
What he tried: Full 12 Week Year implementation. 12-week goals, weekly scorecards, execution score tracking (target: 85% of planned weekly activities completed).
What worked: The urgency was real and effective. “Week 1 of a 12-week cycle felt like month 1 of a regular year — there was actual pressure, not the fake urgency I was creating manually. I got more done in the first six weeks than I had in the previous three months.”
What failed: The burnout was also real. “By week 10, I was exhausted. Not from the work itself but from the constant sense of pressure. I was running a 12-week sprint at my product, my health, and my personal development simultaneously. That’s three sprints at once. There was no recovery.”
He also noticed that the 12 Week Year was a pure execution tool — it had no mechanism for deciding whether he was executing toward the right goals. “I hit my execution scores for three months and then asked myself: was this the right goal? I had no framework for that question.”
Assessment after three months: High-value tool for execution sprints on a single goal. Actively harmful when applied to everything simultaneously. Not sustainable as a full-year system.
Test 4: Annual Theme (Six Months)
After the 12 Week Year’s intensity, he chose the Annual Theme deliberately as a recovery. He set a theme of “Year of Foundation” and let it guide his decisions without specific metrics.
What he tried: A single theme, evaluated quarterly against whether his major decisions aligned with it.
What worked: The flexibility. “A couple of unexpected things happened in months two and three — a potential acquisition conversation, a health issue, a shift in the market. None of those things broke my system because the system was just a direction, not a set of deliverables.”
What failed: Everything else. “I stopped tracking anything. I stopped reviewing anything. I convinced myself that random things I was doing aligned with ‘Foundation’ even when they clearly didn’t. I needed some accountability structure and the theme gave me none.”
At the six-month mark, he had made very little concrete progress on any of his major goals, despite feeling like the year was going fine. The theme had become a way to feel good about drift.
Assessment after six months: Good as an overlay. Insufficient as a primary system. Worked better when he used it alongside OKRs as a filter for Objective-setting.
Test 5: WOOP (Ongoing)
He discovered WOOP late and applied it differently than the others — not as a primary planning system but as a tool for specific recurring obstacles.
What he tried: A weekly WOOP practice applied to the one thing he kept failing to do: his quarterly planning review.
What worked: Almost immediately. “I kept skipping my reviews. I knew they were valuable but I kept finding reasons not to do them. WOOP made me name the actual obstacle — I was afraid of what the review would reveal. If I looked honestly at my progress, I’d have to confront the fact that I’d been avoiding hard things. Once I named that, I could write an if-then plan: if I notice resistance to starting my review, then I’ll write for 10 minutes about what I’m afraid to see.”
He now uses WOOP weekly for the specific behaviors he keeps avoiding — not as a planning system for goals, but as a diagnostic tool for behavioral obstacles.
Assessment: Not a complete framework. Indispensable as a component.
The Hybrid: What He Uses Now
After two years of testing, Michael runs a three-layer system designed with AI input.
Layer 1: Direction (annually). A BHAG-style 5-year north star — he doesn’t call it a BHAG but it serves the same function. “Where do I want to be in five years, and would I be proud of this when I look back?” He also uses an Annual Theme as a filter.
Layer 2: Execution (quarterly). Stripped-down OKRs — one Objective per area, two Key Results each (not three to five), one check-in per month (not weekly). He scores Key Results but doesn’t agonize over the scores. “I took OKRs and removed everything that felt like organizational infrastructure.”
Layer 3: Behavior (weekly). WOOP applied to the specific obstacles that keep arising. Not a weekly ritual for all goals — just a diagnostic tool when he notices he’s been avoiding something for more than a week.
He designed this system in a 90-minute conversation with AI after his fourth test cycle. He described each framework he’d tried, what worked and what didn’t, his goals, and his working style. The AI identified the pattern: he needed directional ambition (which SMART goals and the 12 Week Year didn’t provide), execution structure without organizational overhead (which full OKRs over-delivered), and a behavioral obstacle tool (which none of the others addressed).
“The AI didn’t invent the framework. It helped me see the shape of what I actually needed by reflecting back what I’d described.” He’s been running this system for eight months now — the longest he’s ever maintained any goal-setting practice.
Tools like Beyond Time are designed for this kind of iterative, context-aware planning — maintaining the history of previous cycles so that the system learns what works for you specifically, rather than resetting to generic defaults each quarter.
What Michael’s Tests Reveal
The lesson isn’t that standard frameworks are bad. It’s that testing them honestly — running real experiments, documenting results, and being willing to change based on evidence — is how you find what actually works for you.
Most people never do this. They use one framework until they abandon goal-setting entirely, or they read about a new framework and switch on the basis of a book, not evidence from their own life.
Michael’s approach — treating framework selection as a deliberate experiment — is the most transferable lesson from his story.
The complete guide to goal-setting frameworks lays out every framework he tested in detail. The guide on why frameworks fail explains the patterns Michael ran into — cargo-culting, organizational frameworks on personal goals, burnout from misapplied urgency.
Your action today: If you’ve tried more than one goal-setting framework and abandoned each one, write down what worked and what didn’t about each. That’s the raw material for designing a hybrid that actually fits. Bring that documentation to an AI conversation and ask it to identify the pattern — what were you getting right, what were you missing, and what would a framework designed for your specific situation look like?
Frequently Asked Questions
-
Is it normal to try multiple goal-setting frameworks before finding the right one?
Yes — and it's actually a sign of seriousness rather than inconsistency. Most people stumble onto one framework and stick with it by default, never knowing whether a better fit exists. Systematically testing frameworks, the way you'd test any tool, is a more deliberate approach. The key is running each test long enough to produce real feedback — at least 30 days on a real goal.
-
What made the AI-customized hybrid approach work when the standard frameworks didn't?
The hybrid approach worked because it was built around Michael's specific constraints: inconsistent schedule, intrinsic motivation, creative work that resists rigid metrics, and a tendency to burn out under sustained high pressure. No standard framework accounts for all four simultaneously. The AI customization allowed him to keep the valuable elements from each framework while removing the elements that created friction for his particular situation.
-
Can anyone build a hybrid goal-setting framework with AI?
Yes. The process involves describing your experience with each framework you've tried, what worked and what didn't, your goal types, and your working style. From that context, AI can recommend a hybrid that borrows from multiple frameworks at different levels — direction, execution, and daily behavior. The result is more useful than any off-the-shelf framework because it's designed for a specific person rather than a general audience.