This video is not monetized. This video covers our serious concerns regarding the data accuracy of Linus Media Group, including Linus Tech Tips, ShortCircuit...
For the variability point, they do tests in as a controlled environment as they could, and do the tests until they get consistent data. But what do you mean by significance test and how can they do it?
I agree they’re not the gold standard but they’re the best we got in terms of independent third party testers, and I would assume they’re more than good enough for tech stuff.
Lets say sou run a single test and collect 10 samples at steady-state for temperature and power. This data will have some degree of variability depending on many factors such as airflow at that exact moment, CPU utilization and also inherent noise in the measurement device itself. Additionally, if you repeat the same test multiple times on different days with different testers, you will not get the exact same results.
So if you then compare a system A to system B you might see that system B is 12% “better” (depending on your metric), then you must answer the question–> is this observed difference due to system B actually being better or can this difference be explained by the normal variability in your test setup. Most of the time there are so many external factors influencing your measurement that even if you see a difference in your data, this difference is not significant but due to chance. You should always present your data in a way that its clear to the reader how much variability was in your test setup.
For the variability point, they do tests in as a controlled environment as they could, and do the tests until they get consistent data. But what do you mean by significance test and how can they do it?
I agree they’re not the gold standard but they’re the best we got in terms of independent third party testers, and I would assume they’re more than good enough for tech stuff.
Lets say sou run a single test and collect 10 samples at steady-state for temperature and power. This data will have some degree of variability depending on many factors such as airflow at that exact moment, CPU utilization and also inherent noise in the measurement device itself. Additionally, if you repeat the same test multiple times on different days with different testers, you will not get the exact same results.
So if you then compare a system A to system B you might see that system B is 12% “better” (depending on your metric), then you must answer the question–> is this observed difference due to system B actually being better or can this difference be explained by the normal variability in your test setup. Most of the time there are so many external factors influencing your measurement that even if you see a difference in your data, this difference is not significant but due to chance. You should always present your data in a way that its clear to the reader how much variability was in your test setup.