IngestIQ

AI Models

Configure AI models for parsing and embeddings

Overview#

IngestIQ uses AI models for two key tasks:

Parser Models

Break documents into semantic chunks

Embedding Models

Convert text to vector embeddings

Supported Provider#

Parser Models (Document Processing)#

ProviderModelsBest For
Google Geminigemini-1.5-flash, gemini-1.5-proGeneral documents, PDFs, images
OpenAIgpt-4o, gpt-4o-miniComplex documents

Embedding Models#

ProviderModelDimensionsBest For
OpenAItext-embedding-3-small1536General use, cost-effective
OpenAItext-embedding-3-large3072High accuracy, large docs

Model Configuration#

List Available Models#

curl http://localhost:3000/api/v2/ai-models \
  -H "Authorization: Bearer YOUR_JWT_TOKEN"

Response#

{
  "aiModels": [
    {
      "id": "uuid",
      "name": "Gemini 1.5 Flash",
      "provider": "google",
      "type": "parser",
      "configSchema": {
        "model": "gemini-1.5-flash",
        "inputCostPer1MTokens": 0.075,
        "outputCostPer1MTokens": 0.30
      }
    },
    {
      "id": "uuid",
      "name": "OpenAI Embedding Small",
      "provider": "openai",
      "type": "embedding",
      "configSchema": {
        "model": "text-embedding-3-small",
        "dimensions": 1536,
        "costPer1MTokens": 0.02
      }
    }
  ]
}

Creating Model Configuration#

Parser Model Config#

curl -X POST http://localhost:3000/api/v2/ai-models/configs \
  -H "Authorization: Bearer YOUR_JWT_TOKEN" \
  -H "Content-Type: application/json" \
  -d '{
    "name": "Fast Parser",
    "aiModelId": "gemini-flash-model-uuid",
    "config": {
      "temperature": 0.2,
      "maxOutputTokens": 8192
    }
  }'

Embedding Model Config#

curl -X POST http://localhost:3000/api/v2/ai-models/configs \
  -H "Authorization: Bearer YOUR_JWT_TOKEN" \
  -H "Content-Type: application/json" \
  -d '{
    "name": "Standard Embeddings",
    "aiModelId": "openai-embedding-small-uuid",
    "config": {
      "dimensions": 1536
    }
  }'

Environment Configuration#

API Keys#

# Required API Keys
OPENAI_API_KEY=sk-your-openai-api-key
GOOGLE_API_KEY=your-google-ai-api-key

# Model defaults
OPENAI_EMBEDDING_MODEL=text-embedding-3-small
GOOGLE_MODEL=gemini-1.5-flash

Choosing Models#

Parser Model Selection#

Best for: Most use cases

  • Fast processing
  • Cost-effective
  • Good for PDFs, docs, images
  • 1M token context window

Cost: ~$0.075/1M input tokens

Embedding Model Selection#

Best for: Most use cases

  • 1536 dimensions
  • Fast processing
  • Cost-effective
  • Good accuracy

Cost: $0.02/1M tokens

Usage Tracking#

IngestIQ tracks AI usage for cost monitoring:

curl http://localhost:3000/api/v2/ai-models/usage \
  -H "Authorization: Bearer YOUR_JWT_TOKEN"

Response includes:

  • Token counts per model
  • Estimated costs
  • Usage by pipeline

Best Practices#

Use Gemini Flash and text-embedding-3-small for initial development. Upgrade only if needed.

Keep embedding dimensions consistent within a Knowledge Base. Mixing dimensions causes search issues.

Check usage logs regularly, especially during initial ingestion of large document sets.

Documentation