Skip to main content

Cluster Overview

To start using Last9 you need at least one Cluster, from which you read and write metric data. In this document, we dive deep into all things related to a cluster.

To get up and running fast, see our Quick Start Guide.

Think of a Cluster as a logically separated, Prometheus API-compatible data source for all your metric data. You can create as many Clusters as you want, the number of clusters has no impact on your billing.

It is typically recommended that you create Clusters for each of your environments. Example: Production Cluster, Staging Cluster, etc.

Creating a New Cluster

To create a Cluster:

  1. Navigate to HomeLevitate Creating a Cluster 1

  2. Click the Launch Cluster button to launch the setup wizard Creating a Cluster 2

  3. Select the AWS region you would like to deploy the cluster in. This should ideally be the same region as your application

  4. Give the Cluster a descriptive name

  5. Optionally, add a description which will be displayed on the Cluster Overview screen Creating a Cluster 3

  6. Press the Create to create your new Cluster As the Cluster gets created, you will be presented with an access token that is automatically created. This token is required to start writing & reading data to the Cluster. Tokens are only shown once, so please copy or download credentials (or you can always create another token from Cluster settings). Creating a Cluster 4 Your new Cluster is now ready to receive metrics.

  7. To start writing data to this new Cluster, please follow the Write Data steps start writing data from Kubernetes, Prometheus, AWS/CloudStream or quickly try out by running a local demo environment Creating a Cluster 5 Using the Test Config button you can verify if your Last9 cluster has started receiving data. Click the Next button to start reading data/querying metrics from your Cluster.

    See our guides on how you can send data from Prometheus, OpenTelemetry, VMAgent, or other various sIntegrations supported. 8. To start reading metrics from your new Cluster you can use Managed Grafana which comes included with every plan.

    Alternatively, you can use the provided Read URL to read data using any Prometheus HTTP API Compatible tool like AlertManager, your own Grafana, KEDA, etc. See the guide on how to connect your own Grafana with a Last9 cluster. Creating a Cluster 6


Managing a Cluster

Cluster Usage and Performance

Last9 provides the following tools to observe the Cluster’s performance:

Cluster Usage

Cluster Usage Usage for each cluster is reported in Samples Ingested. A sample refers to a single data point in a time series.

Usage for each Cluster can be viewed from the Cluster’s details page. For more granular and historical usage, see the Cluster Health dashboard’s Sample Ingested panel.

Cluster Quotas

There are no per-cluster limits in Last9. You are billed for usage across all Clusters combined. The ingestion rate, read query rate, and data retention quotas are applied for all the data across all clusters.

Default Cluster Quotas

Last9’s default cluster quotas are fairly generous. In certain cases, keeping in mind performance and cost impacts, we may be able to increase a quota after a discussion with your team.

Write Quotas

TypeBase QuotaReset PeriodNote
Per Time Series Cardinality1MPer HourCan be raised on request
Per Time Series Cardinality20MPer DayCan be raised on request
Streaming Aggregation Cardinality3MPer HourCan be raised on request
Ingestion Concurrency20KPer SecondCan be raised on request
Number of Metrics Aggregated in one Pipeline1 MetricPer QueryCannot be changed

Read Quotas

TypeBase QuotaNote
Time Series Scanned Per Query — Blaze Tier5MCannot be changed
Time Series Scanned Per Query — Hot Tier10MCannot be changed
Samples Scanned Per Query100MCannot be changed
Query Time Range — Blaze Tier2 HoursCan be raised on request
Query Time Range — Hot Tier35 DaysCan be raised on request

If you wish to change your quotas, please raise a request by emailing us on: cs@last9.io

Cluster Health Dashboard

Every Last9 Cluster comes with its own Health dashboard. To view the Health dashboard, navigate to the Cluster details page and click on the View Health link in the performance card. Cluster Health - 1

The following Cluster Performance Metrics are available in the health dashboard: Cluster Health - 2

  • Write Success - Total successful write requests
  • Write Error - Total failed write requests
  • Samples Ingested - Total number of samples ingested
  • Write Availability - Percentage of write requests successful
  • Write Latency - Write request latency
  • Lag - Pending samples waiting to be indexed (in bytes)
  • Read Success - Total successful write requests
  • Read Errors - Total failed read requests
  • Cardinality Limited - Metrics whose cardinality has been limited
  • Read Latency - Query Latency
  • Cardinality Limiter (Early Warning) - Metrics whose cardinality is about to be limited
  • Bytes Dropped - Samples permanently failed to be indexed (in bytes)

Query Logs

Query Logs helps identify slow-running queries so that you can debug and optimize your PromQL.

Query Logs displays slow queries in the last 24 hours, which were successfully executed but have taken more than 1000ms (ie one second) to execute. Query Logs

When a slow query is identified the following details are displayed:

  • Timestamp - Time when the query was executed
  • Query - PromQL along with the query’s time range and query resolution step width
  • Latency - approximate time taken for the query to execute
  • Token Name - the name of the token used to query
  • Tier - storage tier that was used for this query

Cluster Settings

Tokens

Tokens provide a mechanism for access management for your clients. We generate a default token when the Cluster is created for the first time

Creating a New Token

  1. Navigate to the Cluster that you wish to create a token for: Control PlaneTokens Create Token 1
  2. Click New Token Create Token 2
  3. Provide a descriptive Token Name the access Scope (Write Only, Read Only, Read & Write) and click Create
  4. Copy the generated token since it will be visible only once. This token can now be used along with the Read or Write URL (depending on the Scope selected) Create Token 3

Delete a Token

To delete/revoke a token:

  1. Navigate to the Cluster that you wish to revoke a token from: **Control Plane → Tokens
  2. Click the button and select Delete Delete Token 1

Note:

  • This action cannot be undone, once deleted tokens cannot be recovered
  • Tokens can only be deleted by your organization's admin

Write & Read Data

Refer to the list of available Integrations that can be used to start writing and reading data to a Last9 Cluster.

Access Policies

Last9 has built-in data-tiering capabilities based on retention policies. Access policies let you define policies to control which token or client can query a specified data tier.

See our in-depth guide on how you can leverage this powerful feature - Guide on Access Policies

To define a new access policy:

  1. Navigate to Control PlaneAccess Policies Access Token 1 Every cluster comes with a default access policy pre-configured.

  2. To define a new policy click the Create button Access Token 2 Provide the following details:

    • Policy Name: Give a descriptive name for this access policy
    • Token: Select a specific Token for which this access policy is applied or select Any
    • Query Client: We can identify traffic from known clients or select Any for the policy to apply from any client
    • Tier: Select the Tier from which the queries will be served for this policy

    And click Create

  3. Your new access policy will be applied instantly Access Token 3

To delete an Access Policy :

  1. Select the button beside the access policy you wish to delete Access Token 4
  2. Select Delete from the menu

Do Note:

  • Access policies can only be deleted by the admin user(s) of your org
  • Deleting an access policy may limit or lock access for a client or token, please be mindful before deleting

Macros

Macros lets you define PromQL queries as reusable functions and use them as abstracted metric names across Grafana, Alert Manager, or the CLI

We cover how to define and use Macros in detail in guide on PromQL Macros

Enabling Macros:

  1. Navigate to Control PlaneMacros Macros 1
  2. Write/Paste your Macro function and Click Save Macros 2 We perform validation once you click Save Macros 3 Once validated, we will save your Macro function. Do note that it will take upto 5 minutes for new Macros to be available for querying Macros 4

Deleting Macros:

  1. Navigate to Control PlaneMacros Macros 5
  2. Click the delete icon and click confirm

Note:

  • Deleted Macros will impact any queries and dashboards where the macro functions were used
  • Deleted Macros may be available for queries up to 5 minutes after they have been deleted

Streaming Aggregation

Streaming aggregation is a powerful metric cardinality that is built-in with Last9. Refer to our Guide on Streaming Aggregation for an in-depth tutorial