This is an early version of the site so there's still a lot of work to do to explain this stuff on the site itself. To answer your questions:
- All monitoring is from EC2's us-east-1 region.
- I don't track the location of the sites, other than specifying the vendor.
- Data is generated by an open source tool called Stella. I'm not currently tracking DNS lookups but probably will eventually.
- Tests with errors are not included in the response times.
The specific response time is less interesting than looking at how a site performs over time (the number and duration of downtime, errors, etc). There are a lot of rabbit holes in testing so what I'm focusing on now is the simplest possible solution that works.
So basically, anyone hosting their site on EC2 will appear to have a fantastic response time, compared to anyone else. Which makes the test mostly useless for comparison between startups (though it could be interesting/valid data for a single startup over time, which is what pingdom sells)
Unfortunately, I think the parent commenter is right -- your numbers are (currently) meaningless. Statistically, your sample size for estimating response time is 1 for each of those sites. You need at least a dozen or so monitor sites, preferably distributed in the same way that Internet users are, for anything meaningful information to be gained. Ideally, you'd also want to distribute your monitors taking peering agreements of the ISP, etc. into account for a truly representative picture of response times.
As it stands, you're doing a disservice to startups that aren't on EC2, for no good reason. You solution "works" on a technical level, but the data is statistically meaningless.
Combining monitoring data from multiple locations would give a more accurate response time overall. But as a site owner, that number is not very useful because it's not actionable. For example, if response times to my site become slow from say, Dallas or Ireland, there's is nothing I can do about it other than be aware that it's happening.
Monitoring from a single location is interesting because it gives a metric for how well my application is performing.
Certainly, but I would say that there are very few Internet startups that cater to a single (or even a few) geographical locations. Your numbers currently indicate quality of service for customers in the same hosting block as Amazon EC2.
If you have enough monitors and spread them out well (which shouldn't be too hard, since you could just sign up for Perl/PHP enabled hosting accounts in different locations), you can still diagnose where the slow geographical regions are.
In short, having more data can only give you a more accurate picture, depending on how you analyze it.
Your design is great and helping people to make their web sites faster is laudable goal. Add more testing locations -- even better, write a probe that you can run on various consumerbroadband connections that will report results back centrally. Please don't go fire up a bunch of VPSes around the globe for performance monitoring -- that data is largely useless.
Build an image, load it onto a bunch of Pogoplugs, send them out to some friends who live in various parts of the world (and various parts of the US with various kinds of broadband!)
- All monitoring is from EC2's us-east-1 region.
- I don't track the location of the sites, other than specifying the vendor.
- Data is generated by an open source tool called Stella. I'm not currently tracking DNS lookups but probably will eventually.
- Tests with errors are not included in the response times.
The specific response time is less interesting than looking at how a site performs over time (the number and duration of downtime, errors, etc). There are a lot of rabbit holes in testing so what I'm focusing on now is the simplest possible solution that works.