r2r_azure_with_test_limits.toml 1.4 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758
  1. # A config which overrides all instances of `openai` with `azure` in the `r2r.toml` config
  2. [agent]
  3. [agent.generation_config]
  4. model = "azure/gpt-4o"
  5. [completion]
  6. [completion.generation_config]
  7. model = "azure/gpt-4o"
  8. [database]
  9. # KG settings
  10. batch_size = 256
  11. [database.graph_creation_settings]
  12. generation_config = { model = "azure/gpt-4o-mini" }
  13. [database.graph_enrichment_settings]
  14. generation_config = { model = "azure/gpt-4o-mini" }
  15. [database.graph_search_settings]
  16. generation_config = { model = "azure/gpt-4o-mini" }
  17. [database.limits]
  18. global_per_min = 10 # Small enough to test quickly
  19. monthly_limit = 20 # Small enough to test in one run
  20. [database.route_limits]
  21. "/v3/retrieval/search" = { route_per_min = 5, monthly_limit = 10 }
  22. [database.user_limits."47e53676-b478-5b3f-a409-234ca2164de5"]
  23. global_per_min = 2
  24. route_per_min = 1
  25. [embedding]
  26. provider = "litellm"
  27. base_model = "openai/text-embedding-3-small" # continue with `openai` for embeddings, due to server rate limit on azure
  28. base_dimension = 512
  29. [file]
  30. provider = "postgres"
  31. [ingestion]
  32. provider = "r2r"
  33. chunking_strategy = "recursive"
  34. chunk_size = 1_024
  35. chunk_overlap = 512
  36. excluded_parsers = ["mp4"]
  37. audio_transcription_model="azure/whisper-1"
  38. document_summary_model = "azure/gpt-4o-mini"
  39. vision_img_model = "azure/gpt-4o"
  40. vision_pdf_model = "azure/gpt-4o"
  41. [ingestion.chunk_enrichment_settings]
  42. generation_config = { model = "azure/gpt-4o-mini" }