Skip to main content
Menu
← Blog

Testing Fat to Fit

Lofi ran a Roblox ad test on Fat to Fit, saw behavior flatten after the loop clicked, and stopped before a full launch when structure showed a weak long tail.

We ran a deliberate ad test on Fat to Fit and stopped. The goal was intentionally boring: expose the core loop to real Roblox traffic, watch what happened once players understood it, and decide whether the project deserved a full release push.

This is not a story about a broken build. The build worked well enough on paper: readable loop, acceptable early engagement, nothing obviously embarrassing in the opening band. The problem was what happened next.

What we saw once the loop was “solved”

After players understood the beat, behavior flattened. Same rotations, same decisions, no second-act pivot, no emergent combinations worth talking about. Ten minutes and thirty minutes felt like the same spreadsheet with more rows.

If you have read our postmortems on Gym Trainers or Strong Simulator, you know we call that pattern flattened behavior. It is the signal we trust more than greenlight theater.

Ask the uncomfortable question with us: if the core pattern does not diverge over time, can live ops actually save it - or are you only delaying the same churn graph?

Why we treated early signal as a verdict

We could have kept stacking progression tiers. Past experience says that often feeds the dominant loop more surface area instead of creating new contests.

This was one of the first times we treated “players already showed us who this game is” as a stop sign instead of a ticket queue.

We published a follow-up post after we committed to the decision, with the full reasoning on why we did not launch.

How this test fit our Roblox contract habits

We were already running a compare-and-learn mindset from what shipping three games in three months teaches you. Fat to Fit was not trying to win a popularity contest. It was trying to answer whether the incentive graph had a long tail.

What we measured (simple, blunt)

  • early comprehension: did players understand the loop without excessive friction?
  • post-competence variance: did behavior diversify or narrow after understanding?
  • repeat session texture: did later sessions introduce new problems to solve?

The third item failed. That was enough.

Why ads were the right instrument here

Ads are noisy, but they buy a real slice of strangers fast. For this decision, we wanted strangers, not friends who subconsciously optimize for kindness.

The noise means you interpret cautiously. The flattening still showed through.

What we told ourselves about sunk cost

Sunk cost whispers that you should “finish what you started.” Shipping discipline whispers something else: finish learning, not every idea.

Fat to Fit was a probe. Probes are allowed to end.

Practical takeaway for other Roblox teams

If you cannot name what should change in player behavior between session two and session six, you are not ready to argue about launch scale. You are guessing.

Risk we accepted

Stopping early can look like failure from the outside. Internally, it was risk management: avoid funding a long marketing arc on a loop that already looked solved.

How this relates to spike-and-drop dynamics on Roblox

Fat to Fit never got the chance to become a public spike story, but the test still rhymed with the platform truth we wrote about in why Roblox games spike and die so quickly: attention is easy to borrow, and loops are easy to learn. If your loop has no long tail, borrowed attention does not turn into a career.

Contract speed without self-deception

We have written bluntly about how speed kills contract-built games when milestones reward breadth over behavior. Fat to Fit was the opposite move in spirit: stop fast when behavior tells you the truth.

Stopping is a schedule decision. It is also a culture decision.

What we did not use as excuses

It would have been easy to narrate the test as “wrong audience” or “not enough spend.” Maybe those variables matter at the margin.

The internal bar was structural: did the experience keep generating new situations after understanding? It did not.

Instrument design: what we would repeat

Cheap tests that expose loops to strangers early. Clear falsifiers. A written hypothesis before interpreting results.

If you adopt one habit from this post, adopt falsifiers. Otherwise every test becomes a Rorschach blot.

For players

If you ever join a beta that vanishes, sometimes the honest reason is that the data said the loop would not support the long road. That can be disappointing. It is still better than a team pretending for six months.

Alignment with how Lofi thinks about systems

This test is a footnote in the same book as why systems matter more than content: content can make a loop legible, but systems decide whether the loop keeps producing decisions.

Fat to Fit was legible. It did not keep producing decisions.

Closing

Not every prototype earns a full release arc. Sometimes the job of a test is to end early enough to save everyone time. Fat to Fit did its job the moment the graph stopped lying. That is a win, even if it is not the kind you put in a trailer. On Roblox, buying truth cheaply is one of the best budgets you can spend. The alternative is expensive optimism, and it taxes your team twice in morale, calendar time, focus, credibility, team momentum, and trust.

Thanks for reading, and for playing with us on Roblox.