The product is organized into three groups in the sidebar:Documentation Index
Fetch the complete documentation index at: https://docs.getclaro.ai/llms.txt
Use this file to discover all available pages before exploring further.
Core
Dashboard
Entry point for any new workflow. The Dashboard splits the platform’s two modes:- Catalog Operations — for working against your persistent catalog (onboard SKUs, improve taxonomies, monitor competition, analyze inventory).
- Research Agents — for one-off lists and extractions (find prospects, parse documents, enrich spreadsheets, scrape URLs).
Catalogue
The Catalogue page is where you spend most of your time. Each catalog object opens with five tabs:| Tab | Purpose |
|---|---|
| Attributes | Define and edit the schema. Add fields, change types, set validation rules, mark fields as required, set enum values, configure reference relationships. |
| Records | The data itself. Sortable, filterable, editable in a spreadsheet-like grid. Inline edits create change events with provenance. |
| Data Source | Connect upstream — file uploads, scheduled scrapes, supplier portals, APIs, database connections. Each source has its own mapping to attributes. |
| Operation | Run and schedule operations against the records. See Operations. |
| Config | Access control, retention, write-back targets, default reviewers, auto-apply thresholds. |
Taxonomy
Manage hierarchical categories used across your catalogue. View the tree, edit nodes, version the taxonomy, and track which records map to which leaf. The Generate Taxonomy operation (under Modules → Analyse) can propose a taxonomy from your data; the Taxonomy page is where you curate and publish the result.Similarity & Duplicate
Inspect the similarity graph. Browse clusters, review proposed merges, set match thresholds, and configure which fields drive similarity. This surface is shared by the Find Duplicates and Find Similarities operations. See Similarity & Duplicate.Notifications
Every operation that needs review lands here. Review items are grouped by operation and can be approved, rejected, or edited inline. Bulk actions are available per cluster. See Notifications.Modules
Modules group operations by the stage of work they support. Each module exposes its own surfaces in the sidebar but ultimately reads from and writes to the Catalogue.Onboard
Bringing new data — suppliers, SKUs, files — into the catalogue.- Status — pipeline view of in-flight onboarding jobs (file received → mapped → resolved → enriched → ready). Errors and stalls surface here.
- Supplier Portal — a hosted intake surface you can share with suppliers. Suppliers upload files or fill structured forms; submissions land as Data Sources on the relevant catalogue, pre-mapped where possible.
Analyse
Operations that read the catalogue and produce assessments, suggestions, or transformations.- Analyse — a data-quality assessment with per-field scores, distribution stats, critical issues, and an overall Data Quality Score.
- Generative Engine — model-driven generation: descriptions, marketing copy, attribute extraction, alt-text, translations.
Monitor
Continuous observation operations.- Price — track prices across a configurable cadence; flag drifts, deltas, anomalies.
- Competitor — pull competitor listings for matched SKUs and surface side-by-side comparisons.
Distribute
Pushing clean data downstream.- Unified Catalog — a denormalized, query-ready view of the catalogue, available as a read API and as exports.
- Sync & Export — connectors to Shopify, Amazon, BigQuery, Google Sheets, Webhooks, and S3. Each connector has its own mapping, schedule, and conflict policy.
Research Agents
Task-specific tools for ad-hoc work that doesn’t (yet) belong in your persistent catalogue.| Agent | Use case |
|---|---|
| Find your perfect list | Generate a researched dataset of prospects, suppliers, partners, or any entity class. |
| Turn documents into structured data | Extract structured records from PDFs, contracts, datasheets, or reports. |
| Analyze & enrich spreadsheets | Upload a spreadsheet; validate, standardize, and complete missing data. |
| Scrape data from URLs | Collect structured records from web pages. |