Five Ways to Help Your Developers Analyze API Outages

When your product relies on APIs (external or internal), it is critical that you monitor those APIs in order to know when something goes awry that affects your customer’s perception of your product’s current status. There are many ways an API can be down and or adversely affect your product. An external API may simply be down. Or its response might be so delayed that you have to decide whether to present users with an incomplete result that omits the latest data from the slow API, or wait for the result and accept performance delays in your product. If your production software expects a JSON response from an API, and you receive responses with invalid JSON, you need to know that…

Continue reading

The Effect of Global Location on CURL Call Metric Patterns

My previous post used CURL component metrics to illustrate performance in calling the World Bank Countries API (located in Washington, D.C.) from Oregon. The curl resolve, connect, processing, and transfer times for calls from Oregon over a period of one week were plotted and analyzed. Here is the plotted Oregon data: The conclusion from studying […]

Continue reading

Using CURL Component Data to Illustrate API Call Performance Patterns

My last post analyzed performance component results produced by the curl Internet data transfer utility in an effort to understand what typically causes slow API call performance. Four instances of unusually slow but successful calls to the World Bank Countries API were studied. In three instances (calls from Oregon, Ireland, and Japan), curl’s Processing Time […]

Continue reading

Analyzing API Performance Hour-of-Day Statistics

A recent post analyzed API performance by hour of day over a one-week period. The average performance of calls to the API was fairly consistent, except for calls made in the last hour of the day (the hour before Midnight Universal Time). This plot presents the analysis results: The question is: why was average performance […]

Continue reading

Analyzing API Performance Binned by Hour of Day

Performance data from the API Science API can be analyzed in many different ways. For example, a recent post presents A Graphical View of API Performance Based on Call Location. The analysis uses cURL statistics to compare the performance of monitors that call the World Bank Countries API from various locations around the globe over […]

Continue reading

A Graphical View of API Performance Based on Call Location

The performance of APIs is dependent on both the processing time from when the API receives a request and delivers a response, and the time it takes for the request and response data packets to traverse the Internet distance between the calling system and the system that hosts the API. The timings for calls to […]

Continue reading

How to Create an Automated Custom Web Site that Displays API Uptime Data

Previous posts described how to use curl, cron, JSON, Python, matplotlib, and HTML to create an automatically-updated custom API performance web page. In addition to providing API performance data, the API Science API also provides uptime data for the APIs you monitor. In this post, I’ll demonstrate how to download API uptime data and create […]

Continue reading

7 Ways to Customize API Monitoring Using Global and Local Variables

It is possible to configure API monitors with rigidly fixed parameters. However, using the API Science platform, it is also possible to configure API monitors using both account-wide global variables and monitor-specific local variables, in order to provide your team with a deeper view into what your customers are experiencing. Introduction to API Science’s Global […]

Continue reading

With APIs and Software Libraries, How Much Code Is Needed to Create Something Immensely Useful? Not Much…

Your business likely requires customized views of the data that is core to the creation of your product. Decades ago, accomplishing this involved development of large data analysis software libraries that were customized to a company’s particular needs, along with user interfaces that enabled employees to view the data, so they could respond to anomalies. […]

Continue reading