I woke up this morning and checked three devices. Oura said 72. Whoop said 84%. Garmin Body Battery said 58.
Same body. Same night. Three different verdicts on whether I should train hard today.
If you only own one device, you probably trust it. You have no reason not to. It gives you a number, the number has a colour, and you follow the colour. Green means go. Red means rest. Simple.
If you own two or more, you have a different problem entirely. Because now you see the disagreements. And once you see them, you cannot unsee them.
The disagreement is not a bug
This is the part most people get wrong. They assume one device must be right and the others must be broken. They post on Reddit asking “which one should I trust?” as if the answer is hiding behind the right firmware update.
None of them are broken. All of them are right. They just are not answering the same question.
Oura is asking: how well did you recover overnight based on sleep quality, HRV, and temperature? It is a nocturnal recovery sensor that lives on your finger. It captures skin temperature from the radial artery with genuinely impressive precision. When it says your readiness is low, it usually means your sleep was fragmented or your temperature baseline shifted. It knows almost nothing about what you did during the day.
Whoop is asking: given the cardiovascular strain you accumulated yesterday, how well has your nervous system bounced back? It is a strain and recovery engine. It captures cardiac load throughout the day and measures your overnight autonomic response to that load. When it says green, it means your heart rate variability recovered relative to the stress you put on it. It does not know your skin temperature. It does not deeply weight sleep architecture the way Oura does.
Garmin is asking: how much energy do you have right now based on continuous stress monitoring? Body Battery is an all-day autonomic stress model. It charges during sleep, drains during activity and psychological stress, and gives you a running total. It catches work stress and emotional load better than either Oura or Whoop. But it consistently underestimates the cost of strength training because its stress model is calibrated primarily around cardiovascular effort.
Three devices. Three philosophies. Three valid but incomplete windows into the same physiological state.
Why a single device will never be enough
The sensor array required to capture everything that affects athletic readiness does not fit into a single form factor. The finger is better for temperature. The wrist is better for all-day HRV sampling. A chest strap gives cleaner data during high-intensity exercise. No single device sits in all three locations simultaneously.
Even if the hardware problem were solved, the algorithm problem would remain. Each company has a commercial incentive to present their score as the answer, not an answer. Oura wants readiness to be about sleep. Whoop wants recovery to be about strain management. Garmin wants Body Battery to be about energy. They are each optimising for their own narrative.
This means every wearable company is structurally incentivised to give you a confident, simple answer from an incomplete data set. That confidence is the product. The simplicity is the selling point. And the incompleteness is the thing nobody mentions in the marketing.
The real problem is between the devices
Here is what actually happens every morning for a multi-device athlete.
You check Oura. You check Whoop. Maybe you check Garmin. You hold three numbers in your head. Then you do what every athlete has always done: you make a gut call.
The devices were supposed to eliminate gut calls. That was the entire promise. Objective data replacing subjective guessing. Instead, multi-device athletes have replaced one gut call with a more informed gut call. Progress, yes. But not the transformation the marketing promised.
The information you actually need is a synthesis across all three inputs, weighted by context. If you had a terrible sleep but your training load has been light, the low Oura score matters less. If your Whoop says green but you have been in sympathetic overdrive from work stress all week, the green is misleading. If your Garmin Body Battery is at 40 but you only did easy Zone 2 work yesterday, the low battery is probably just ambient stress and not a signal to skip training.
No device makes those contextual judgements. They each report their own channel and leave the synthesis to you. The gap is not accuracy. Every device is reasonably accurate at what it measures. The gap is interpretation across sources.
What athletes actually do about it
I have spent months in Reddit communities watching how multi-device athletes handle this. The patterns are consistent.
The most common approach is to pick a primary device and use the others as tiebreakers. “I follow Whoop unless Oura flags a temperature deviation.” This works reasonably well but it is manual, unsystematic, and depends entirely on the athlete’s experience with their own body.
The second most common approach is to ignore the scores entirely and just look at the raw data. Morning HRV trend over seven days. Resting heart rate trend. Sleep efficiency percentage. These athletes have essentially stopped using the product their device is selling them (the score) and built their own interpretation system from the raw inputs. They are doing the synthesis work themselves, in their heads, every morning.
The third approach is the one I see most often in newer athletes: total confusion. They bought a device to get clarity. They got a number. They trusted the number. Then they heard about another device, bought that too, and now they have two numbers that disagree. Instead of more clarity, they have less. Some of them stop checking altogether. Others become obsessive data hoarders who export CSV files and build spreadsheets.
All three approaches are workarounds for the same missing layer.
The synthesis layer is the product
The next meaningful evolution in athletic data is not a better ring, watch, or strap. It is a system that sits above all of them.
A system that ingests Oura’s sleep stages, temperature deviation, and readiness. Whoop’s strain accumulation, recovery percentage, and HRV. Garmin’s Body Battery, stress score, and training status. Apple Health’s workout data. Strava’s training log. Maybe a food tracker. Maybe a body composition scan.
A system that understands the methodology behind each metric. That knows Oura’s readiness is sleep-weighted and Whoop’s recovery is strain-weighted, and that those biases need correcting when combining the signals. That tracks not just today’s numbers but the seven-day and thirty-day trajectories of each signal and identifies divergence patterns that predict problems before they surface.
A system that flags the disagreements that matter: “Whoop says green but your Oura temperature has been rising for three consecutive nights. Consider reducing intensity.” A system that learns your personal patterns over time: “Your HRV tends to drop two days after high-volume sessions, not one. Adjust the recovery window.”
This is not a feature request for Whoop or Garmin. Neither company will build it because it requires treating their competitor’s data as equally valid input. The synthesis layer has to be device-agnostic by definition.
This is what P247 is building. Not a replacement for your wearable. A layer that makes all of them more useful by connecting what they each see into a single coherent picture. The devices are not going to stop disagreeing. But the disagreement itself is information, and right now nobody is using it.
Your devices are all wrong. That is actually the point. The question is who is going to be right about what the wrongness means.
References
Halson, S.L. (2014). Monitoring training load to understand fatigue in athletes. Sports Medicine, 44(S2), 139-147.
Plews, D.J., Laursen, P.B., Stanley, J., Buchheit, M., & Kilding, A.E. (2013). Training adaptation and heart rate variability in elite endurance athletes: Opening the door to effective monitoring. International Journal of Sports Physiology and Performance, 8(5), 560-570.
Saw, A.E., Main, L.C., & Gastin, P.B. (2016). Monitoring the athlete training response: subjective self-reported measures trump commonly used objective measures. British Journal of Sports Medicine, 50(5), 281-291.
Green score. Destroyed legs. There are 6 blind spots in your wearable data. We wrote a free guide covering every one of them.
Download the Free Guide