Node Service

Aptos Node Service FAQ

Note:for billing questions see our Aptos Billing FAQ.

FAQ

How do I view my REST API requests, errors, and latency?

On the Aptos Node Service page click the "API Insights" tab.

Here are some tips:

  1. Filter by network, access key, method, and/or time range. Changing these filters re-fetches all the insights on the page.
  2. Summary metrics for a quick look at your error rates. Of course, a low error rate over a long period could hide one day that had a high error-rate, so it's worth occasionally looking at the graphs below. Looks like there's a significant percentage of non-rate-limit errors for my test account, so I should look at the graphs below to see which errors I'm getting
  3. If you hover over a section of a chart bar, you'll see details about what it represents.
  4. You can click items in the legend to hide or show them.
  5. There are more graphs below! Scroll down to view error and latency insights.

What each graph shows

  1. Request count by method: This shows the count of all requests, including all errors, broken down by method (assuming you've selected "All methods" in the filter at the top of the page).
  2. Error count by Aptos error code: This shows the count of requests that got an Aptos error code, grouped by code. If you filter the page by an individual method, e.g. POST /transactions, you'll only see the errors you received on that method. This graph and the next pair well with our Error Guide's sections on the Aptos Rest API.
  3. Error count by HTTP error code: This shows the count of requests that got an HTTP error code, grouped by code. If you filter the page by an individual method, e.g. POST /transactions, you'll only see the errors you received on that method.
  4. Error ratio by method: This shows the error ratio for each method. For example, if you sent 100 POST /transactions requests in a time bucket (e.g. a day) and you got an error on two of them, your error ratio would be 2% for that method in that time bucket. If you see a high error ratio it's useful to scroll up to the "Request count by method" graph to see how many of that method you sent in a time bucket: a 100% error ratio could happen if you send one request and it gets an error.
  5. Requests latency by method - P50: This shows the 50th percentile latency for each method across all of its requests, including errors. This is the latency within our system, so the latency you observe will be a little higher because of network travel time.
  6. Requests latency by method - P95: This shows the 95th percentile latency for each method across all of its requests, including errors. This is the latency within our system, so the latency you observe will be a little higher because of network travel time.
  7. Successful requests latency: This shows the 50th and 95th percentile latencies for all of your successful requests. It does not group by method.
  8. Requests with non-rate-limit errors latency: This shows the 50th and 95th percentile latencies for all of your requests that got an error (excluding rate-limit-errors since they have very low latency). It does not group by method.

How do I view my GraphQL API requests, errors, and latency?

On the Aptos Node Service page click the "GraphQL Insights" tab.

Here are some tips:

  1. Filter by network, access key, method, and/or time range. Changing these filters re-fetches all the insights on the page.
  2. Summary metrics for a quick look at your error rates. Of course, a low error rate over a long period could hide one day that had a high error-rate, so it's worth occasionally looking at the graphs below.
  3. If you hover over a section of a chart bar, you'll see details about what it represents.
  4. You can click items in the legend to hide or show them. This works when there are two or more legend entries.
  5. There are more graphs below! Scroll down to view error and latency insights.

What are QPS and CUPS?

QPS: "Queries Per Second" We define this as the count of the unique HTTP requests you send per second. This is a count of all requests, including those with any kind of error (e.g. a rate-limit error).

In your Shinami dashboard, you can...

  • see if you hit any QPS rate limits. Look for rate-limit error codes for (HTTP 429 )
  • get an approximation of recent QPS by looking at your Node Service API Insights in the dashboard and selecting a specific access key - or all keys on a network - and the time range "Last 60 minutes". Then, take your highest bar and divide it by 60 (since each bar represents 1 minute).

For Aptos Indexer GraphQL API

CU: "Compute Unit"

The CU amount for a request is a measure of the computing resources it uses. Some requests are easy to fulfill, while others take more effort. You can monitor how many CUs you're using on the "GraphQL API Insights" tab of the Aptos Node Service page of your Shinami dashboard. We're still determining the best way to calculate CUs, so the value for the same request body may change a bit over the coming weeks.

CUPS: "Compute Units Per Second"

This is a measure of the combined CU of all requests you send us within a given second. We calculate this per access key. Your requests will get a HTTP 429 rate limit if you exceed this limit (though we provide you with a small buffer to help accommodate any brief bursts in traffic you may have). You can view whether you've had any recent HTTP 429s on the "GraphQL API Insights" tab of the Aptos Node Service page of your Shinami dashboard. If you use multiple access keys, you can filter the metrics by individual keys and, if needed, shift some of the max CUPS from one key to another.

In your Shinami dashboard, you can...

  • see if you hit any QPS rate limits. Look for rate-limit error code HTTP 429.
  • get an approximation of recent CUPS by selecting a specific access key - or all keys on a network - and the time range "Last 60 minutes". Then, take your highest bar and divide it by 60 (since each bar represents 1 minute).

How do I monitor my request and CU count?

REST API request count by day

On the Aptos Node Service page of your dashboard, select the"REST API Insights" tab. choose the network and time range you want (e.g. Aptos Mainnet" + "Last 30 days UTC"). Then, look at the graph: "Request count by method". This shows the total requests you've sent - including those with errors - broken down by the specific request name.

GraphQL API CU count by day

On the Aptos Node Service page of your dashboard, select the"GraphQL API Insights" tab. choose the network and time range you want (e.g. Aptos Mainnet" + "Last 30 days UTC").

Then, scroll down to the "Compute Units" graph.

How do I monitor my rate-limits and QPS / CUPS?

REST API rate-limits and QPS

On the Aptos Node Service page of your dashboard, select the"GraphQL API Insights" tab. choose the network and time range you want (e.g. Aptos Mainnet" + "Last 30 days UTC") and look at the summary metrics at the top:

Here, I can see that in the last 30 days on Testnet, 0% of my requests had a rate limit error. If you have a small amount that's normal. Just make sure your integration has a retry with a backoff in the case of a rate-limit error (HTTP 429).

However, I can see that I have some other errors in the "Non-rate-limit-errors" summary box. If I scroll down to the "Error count by HTTP error code" graph I can see those errors over time. If I did have any rate-limit errors (HTTP 429) they'd show up here.

Current QPS

The most granular QPS metrics we provide is for the last hour. On the Aptos Node Service pageof your dashboard, choose the network you want to look at and the time range "Last 60 minutes" (note that you can also filter by API key as needed). Then, look at the "Request count by method" graph. Take the highest bar and divide by 60 because each bar is one minute. In the example below, my peak is close to 800. I'll divide 800/60 = 13.3 per second average QPS during my peak in the last hour.

Of course, this estimate is only useful to the extent that the current activity is reflective of average traffic or of a peak. Still, you can get a sense of whether your keys have a decent amount of headroom given your normal or current traffic. Looking at the "Last 48 hours" view first can help tell you if this is normal traffic based on recent volume (but note that the last hour bar on the right is the current hour, and will fill up over the course of the hour).

GraphQL API rate-limits and CUPS

Rate limits

On the Aptos Node Service page of your dashboard, select the"GraphQL API Insights" tab. choose the network and time range you want (e.g. Aptos Mainnet" + "Last 30 days UTC") and look at the summary metrics at the top:

Here, I can see that in the last 30 days on Testnet, 0% of my requests had a rate limit error. If you have a small amount that's normal. Just make sure your integration has a retry with a backoff in the case of a rate-limit error (HTTP 429).

CUPS

You can get a sense of how many CU per request you're sending by choosing the "Last 12 weeks" view and dividing your HTTP 200 request count for a week ("HTTP requests by status code" graph) by the CU used for the week ("Compute Units" graph). This is a rough measurement of your CU per request. Let's call it X. So if you send three requests in one second you could expect your CUPS to be roughly 3X for that second, on average.

However, the complexity of your requests may change over time. Also, we may make adjustments to our calculation of CU over time. So, if you use this number as a guideline you should increase it to be safe - say, double it (so, assume your CUPS is 2X per request). Also, we recommend checking on your CU per request from time to time to see if it's changing.

A simpler solution, though, is just to assign all the CUPS allotment you have to your active keys. Either way, you should monitor for, and retry with a backoff on, rate-limit errors.