r2r_azure_with_test_limits.toml 1.5 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061
  1. # A config which overrides all instances of `openai` with `azure` in the `r2r.toml` config
  2. [agent]
  3. [agent.generation_config]
  4. model = "azure/gpt-4o"
  5. [completion]
  6. [completion.generation_config]
  7. model = "azure/gpt-4o"
  8. [database]
  9. # KG settings
  10. batch_size = 256
  11. [database.graph_creation_settings]
  12. generation_config = { model = "azure/gpt-4o-mini" }
  13. [database.graph_entity_deduplication_settings]
  14. generation_config = { model = "azure/gpt-4o-mini" }
  15. [database.graph_enrichment_settings]
  16. generation_config = { model = "azure/gpt-4o-mini" }
  17. [database.graph_search_settings]
  18. generation_config = { model = "azure/gpt-4o-mini" }
  19. [database.limits]
  20. global_per_min = 10 # Small enough to test quickly
  21. monthly_limit = 20 # Small enough to test in one run
  22. [database.route_limits]
  23. "/v3/retrieval/search" = { route_per_min = 5, monthly_limit = 10 }
  24. [database.user_limits."47e53676-b478-5b3f-a409-234ca2164de5"]
  25. global_per_min = 2
  26. route_per_min = 1
  27. [embedding]
  28. provider = "litellm"
  29. base_model = "openai/text-embedding-3-small" # continue with `openai` for embeddings, due to server rate limit on azure
  30. base_dimension = 512
  31. [file]
  32. provider = "postgres"
  33. [ingestion]
  34. provider = "r2r"
  35. chunking_strategy = "recursive"
  36. chunk_size = 1_024
  37. chunk_overlap = 512
  38. excluded_parsers = ["mp4"]
  39. audio_transcription_model="azure/whisper-1"
  40. document_summary_model = "azure/gpt-4o-mini"
  41. vision_img_model = "azure/gpt-4o"
  42. vision_pdf_model = "azure/gpt-4o"
  43. [ingestion.chunk_enrichment_settings]
  44. generation_config = { model = "azure/gpt-4o-mini" }