Want to know how much website downtime costs, and the impact it can have on your business?
Find out everything you need to know in our new uptime monitoring whitepaper 2021



Recently we’ve had a lot of requests for examples of how add new tests through the StatusCake API using the popular Postman tool. In this article we’ll take you through some of the steps required to get you going with this method, below you’ll find an example for each test type we offer, taking you through the process of adding a test.
First off we need to cover the authentication method that will be required to make the API calls through Postman, you will need to obtain your API key and username from the “User Details” section of the StatusCake App, these should then be input into the “Headers” Section of Postman as shown below:
Now that all of our calls will carry with them the required authentication details, we can look at how to proceed in setting up the tests.
Taking the fields from our API documentation you should enter the settings that you’d like applied to the Uptime test that you are adding, every time an uptime test is created you must at least include as a minimum: TestType, Checkrate, WebsiteName, and WebsiteURL. The call should be submitted as PUT.
Page speed testing works similarly, and you can grab the fields that can be used from our API docs here. The required fields for this one are: name, website_url, location_iso, and checkrate. The POST method is required for this one.
For SSL testing the PUT method should be utilized, and the required fields are: domain, checkrate, contact_groups, alert_at, alert_expiry, alert_reminder, alert_broken and alert mixed.
Share this
4 min read How AI Is Shifting Software Engineering’s Primary Constraint For most of the history of software engineering, the primary constraint was production. Code was expensive, skilled engineers were scarce, and shipping features required concentrated human effort. Velocity was limited by how fast people could reason, implement, test, and deploy. That constraint shaped everything from team size,
5 min read Autonomous Code, Trust Boundaries, and Why Governance Now Matters More Than Ever In Part 1, we looked at how AI has reduced the cost of building monitoring tools. Then in Part 2, we explored the operational and economic burden of owning them. Now we need to talk about something deeper. Because the real shift isn’t
6 min read The Real Cost of Owning Monitoring Isn’t Code — It’s Everything Else In Part 1, we explored how AI has dramatically reduced the cost of building monitoring tooling. That much is clear. You can scaffold uptime checks quickly, generate alert logic in minutes, and set-up dashboards faster than most teams used to schedule the kickoff
5 min read AI Has Made Building Monitoring Easy. It Hasn’t Made Owning It Any Easier. A few months ago, I spoke to an engineering manager who proudly told me they had rebuilt their monitoring stack over a long weekend. They’d used AI to scaffold synthetic checks. They’d generated alert logic with dynamic thresholds. They’d then wired everything
3 min read In the previous posts, we’ve looked at how alert noise emerges from design decisions, why notification lists fail to create accountability, and why alerts only work when they’re designed around a clear outcome. Taken together, these ideas point to a broader conclusion. That alerting is not just a technical system, it’s a socio-technical one. Alerting
3 min read In the first two posts of this series, we explored how alert noise emerges from design decisions, and why notification lists fail to create accountability when responsibility is unclear. There’s a deeper issue underneath both of those problems. Many alerting systems are designed without being clear about the outcome they’re meant to produce. When teams
Find out everything you need to know in our new uptime monitoring whitepaper 2021