The Data Lake is where all your ingested data lives. Every record you send to the Ingestion API is validated, stored, and made available for processing by Evermuse.Documentation Index
Fetch the complete documentation index at: https://docs.evermuse.com/llms.txt
Use this file to discover all available pages before exploring further.
How It Works
When you send records to the Ingestion API, each batch goes through the following steps:Validated
Each record is checked against the Integration Envelope schema. Records that pass are accepted; records that fail
are rejected with error details so you can fix and resend them.
Normalized
Accepted records are cleaned up for consistency. For example,
_type is lowercased and timestamps
are rounded to the nearest second.Deduplicated
Each record is identified by its
_type, _vendor_ids, and _event_at fields. If you send the same record again,
it overwrites the previous version instead of creating a duplicate.Attachments
Records can include_attachments referencing vendor-hosted files. After a batch lands, Evermuse asynchronously downloads each attachment to secure cloud storage. Attachment download progress is tracked per-batch and visible in the monitoring dashboard.
Monitoring
The Data Lake dashboard provides real-time visibility into ingestion activity. You can view all batches with their status, source, type, record counts, and error details.
- Batch metadata — Status, source, type, timestamps, and storage URIs.
- Failed records — Individual validation errors with the rejected payload.
- Attachment downloads — Status of each attachment download attempt.

Batch Lifecycle
Each ingestion request creates a batch that progresses through the following statuses:| Status | Description |
|---|---|
LANDING | Batch received and being written to storage. |
LANDED | Records archived and indexed successfully. |
PROCESSING | Downstream processors are consuming the batch. |
COMPLETE | All processing finished successfully. |
FAILED | Processing encountered an unrecoverable error. |

