Ask AI

AI Data Storage (Multimodal RAG)

Store, manage, and index documents with chunking and embedding for semantic search.

Notion image

Purpose

AI Data Storage (RAG) is a component of the Latenode platform designed for storing and indexing text files, images, and other knowledge sources.

🧠 This tool is primarily intended to be used in conjunction with the AI Agent — it provides documents in the form of chunks, which the agent can then use to generate responses.

Use cases include:

  • Multimodal RAG: uploading images and getting answers to questions about them
  • Uploading and storing structured or unstructured content
  • Generating embedding vectors for fast semantic search
  • Running natural language search queries
  • Connecting to the RAG Search node inside a scenario

How to Access

You can access this feature via Data Storage → AI Data Storage (RAG) in the left-hand side menu.

Notion image

Creating Storage

Click Create Storage to open the setup modal:

Notion image

Fill in the required fields: Storage Name, Chunk Size, Chunk Overlap

Notion image

What are Chunk Size and Overlap?

  • Chunk Size — the number of tokens in a single chunk. Smaller chunks provide higher accuracy but increase the total number of chunks.
  • Chunk Overlap — the percentage of token overlap between neighboring chunks. Helps maintain context across them.

Managing Storage

Created storages are displayed in a table:

Notion image
Field
Description
Name
Storage name
Chunk Size
Number of tokens per chunk
Chunk Overlap
Overlap between chunks in %
Created
Creation date
Updated
Last updated date

Uploading Files

Open a storage to access the upload interface. Drag-and-drop is supported.

Notion image

After uploading:

  • Each file is processed and indexed (status: Processing)
  • Files are listed with size, upload date, and status
  • Editing or downloading files is currently not supported
Notion image

 

Multimodal RAG Features

For working with images and non-textual data, RAG Storage uses an advanced approach:

  • Automatic Image Description: When uploading images (JPEG, PNG), the system automatically generates their textual description (summary) using a multimodal LLM and indexes this description along with the text content.
  • Text Indexing: Text extracted via OCR from images (or PDF files) is also split into chunks and indexed.

This allows the AI Agent to effectively find answers to questions based on both textual and visual content.

Features & Limits

Feature
Status
OCR
Supported (English and Russian)
Image Upload
Supported (if image contains text)
File Editing
Not supported
File Download
Not yet available
Automatic Indexing
Yes
Supported Formats
PDF, TXT, JSON, MD, PNG, JPG, and more
Upload via scenario
Not yet supported

Technical Details

Parameter
Value
Max file size
20 MB (50 MB planned)
Embedding model
Cloudflare + LlamaIndex
Vector limit
5 000 000 vectors per account
Billing
Charged only during file upload (PNP credits)

Billing

  • PNP credits are deducted upon file upload
  • Billing is based on pages/chunks
  • Example: 1 page ≈ 6600 microcredits
  • Vectorization cost: 0.0066 PNP tokens per page
  • “1 page” corresponds to approximately 1000 words or 5000 characters of text
  • For unstructured data (e.g. TXT, MD), the same linear pricing model applies - cost is proportional to total text length
  • Examples:
    • 10 pages (PDF/DOCX/PPTX) → 0.066 PNP tokens
    • TXT ≈ 10 000 words (≈ 50 000 chars) → 0.066 PNP tokens
    • MD ≈ 20 000 words (≈ 100 000 chars) → 0.132 PNP tokens
  • Queries via RAG Search are not additionally billed

🧪 RAG is currently in beta. Pricing, behavior, and limitations may change.

Did this answer your question?
😞
😐
🤩