full_azure.toml 1.2 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546
  1. [app]
  2. # LLM used for internal operations, like deriving conversation names
  3. fast_llm = "azure/gpt-4.1-mini"
  4. # LLM used for user-facing output, like RAG replies
  5. quality_llm = "azure/gpt-4.1"
  6. # LLM used for ingesting visual inputs
  7. vlm = "azure/gpt-4.1"
  8. # LLM used for transcription
  9. audio_lm = "azure/whisper-1"
  10. # Reasoning model, used for `research` agent
  11. reasoning_llm = "azure/o3-mini"
  12. # Planning model, used for `research` agent
  13. planning_llm = "azure/o3-mini"
  14. [embedding]
  15. base_model = "azure/text-embedding-3-small"
  16. [completion_embedding]
  17. base_model = "azure/text-embedding-3-small"
  18. [ingestion]
  19. provider = "unstructured_local"
  20. strategy = "auto"
  21. chunking_strategy = "by_title"
  22. new_after_n_chars = 2_048
  23. max_characters = 4_096
  24. combine_under_n_chars = 1_024
  25. overlap = 1_024
  26. document_summary_model = "azure/gpt-4.1-mini"
  27. automatic_extraction = true # enable automatic extraction of entities and relations
  28. [ingestion.extra_parsers]
  29. pdf = ["zerox", "ocr"]
  30. [ingestion.chunk_enrichment_settings]
  31. generation_config = { model = "azure/gpt-4.1-mini" }
  32. [orchestration]
  33. provider = "hatchet"
  34. kg_creation_concurrency_limit = 32
  35. ingestion_concurrency_limit = 4
  36. kg_concurrency_limit = 8