Skip to content

Ingesting Data

Ingest, transport and fetch data from different sources such as Relational database, web logs, batch data, real-time, application logs, streaming data, etc for later usage using the Axiom API.

You can also collect, load, group and move data from one or more sources to Axiom where it can be stored and further analyzed.

Before ingesting data, you will need an API Token which you can generate from the Settings->Tokens page on the Axiom Dashboard. See API Tokens documentation

Once you have an ingest token, there are three ways to get your data into Axiom:

  1. Using the Ingest API
  2. Using a Data Shipper (Logstash, Filebeat, Metricbeat, Fluentd etc)
  3. Using the Elasticsearch Bulk API that Axiom supports natively

Ingest API

Axiom exports a simple REST API that can accept any of the following formats:

  • application/json - single event or json array of events
  • application/nd-json- structured data processed at a time
  • text/csv - this should include the header line with field names
  curl -X POST \
    -H "Authorization: Bearer <INGEST_TOKEN>" \
    -H "Content-Type: application/x-ndjson" \
    -d '{ "this": "is", "an": "example", "json": 1 }' \

If you would like to instead use a language binding, currently our Golang client library is available:

More client libraries are coming very soon!


Kind Limit
Maximum Event Batch Size 1000
Maximum Event Fields 250
Maximum Array Field Items 100

Data Shippers

Configure, read, collect, and Send logs to your Axiom deployment using a variety of data shippers. Data shippers are lightweight agents that acquire logs, metrics, and lets you ship data directly into Axiom.

integrations icon Ingest using API go

integrations icon Ingest using Elastic Beats go

integrations icon Ingest using FluentD go

integrations icon Ingest using Kubernetes go

integrations icon Ingest using Logstash go

integrations icon Ingest using Loki Proxy go

integrations icon Ingest using Syslog Proxy go