r2r_file.py 6.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183
  1. import tempfile
  2. import uuid
  3. from typing import List
  4. import aiofiles
  5. import aiofiles.os
  6. from fastapi import UploadFile
  7. from sqlalchemy.ext.asyncio import AsyncSession
  8. from app.models import File
  9. from app.providers.r2r import r2r
  10. from app.providers.storage import storage
  11. from app.services.file.impl.oss_file import OSSFileService
  12. # import asyncio
  13. from pathlib import Path
  14. import nest_asyncio
  15. # 使得异步代码可以在已运行的事件循环中嵌套
  16. nest_asyncio.apply()
  17. class R2RFileService(OSSFileService):
  18. @staticmethod
  19. async def create_file(
  20. *, session: AsyncSession, purpose: str, file: UploadFile
  21. ) -> File:
  22. # 文件是否存在
  23. # statement = (
  24. # select(File)
  25. # .where(File.purpose == purpose)
  26. # .where(File.filename == file.filename)
  27. # .where(File.bytes == file.size)
  28. # )
  29. # result = await session.execute(statement)
  30. # ext_file = result.scalars().first()
  31. # if ext_file is not None:
  32. # # TODO: 文件去重策略
  33. # return ext_file
  34. file_extension = Path(file.filename).suffix
  35. file_key = f"{uuid.uuid4()}{file_extension}"
  36. print(file_key)
  37. fileinfo = {"document_id": file_key}
  38. # file_key = f"{uuid.uuid4()}-{file.filename}"
  39. with tempfile.NamedTemporaryFile(
  40. suffix="_" + file.filename, delete=True
  41. ) as temp_file:
  42. tmp_file_path = temp_file.name
  43. async with aiofiles.open(tmp_file_path, "wb") as f:
  44. while content := await file.read(1024):
  45. await f.write(content)
  46. # storage.save_from_path(filename=file_key, local_file_path=tmp_file_path)
  47. await r2r.init()
  48. fileinfo = await r2r.ingest_file(
  49. file_path=tmp_file_path,
  50. metadata={"file_key": file_key, "title": file.filename},
  51. )
  52. fileinfo = fileinfo.get("results")
  53. # 存储
  54. db_file = File(
  55. purpose=purpose,
  56. filename=file.filename,
  57. bytes=file.size,
  58. key=fileinfo["document_id"],
  59. )
  60. session.add(db_file)
  61. await session.commit()
  62. await session.refresh(db_file)
  63. return db_file
  64. @staticmethod
  65. def search_in_files(
  66. query: str, file_keys: List[str], folder_keys: List[str] = None
  67. ) -> dict:
  68. files = {}
  69. file_key = {"$in": []}
  70. document_id = {"$in": []}
  71. filters = {"$or": []}
  72. print(
  73. "ggggggggggggggggggggggggggggggggggggddddddddddddddddddccccccccccccccccccccc"
  74. )
  75. for key in file_keys:
  76. if len(key) == 36:
  77. document_id["$in"].append(key)
  78. else:
  79. file_key["$in"].append(key)
  80. if len(document_id["$in"]) > 0:
  81. filters["$or"].append({"document_id": document_id})
  82. if len(file_key["$in"]) > 0:
  83. filters["$or"].append({"file_key": file_key})
  84. print(file_key)
  85. print(document_id)
  86. print(filters)
  87. print(folder_keys)
  88. if folder_keys:
  89. filters["$or"].append({"collection_ids": {"$overlap": folder_keys}})
  90. ## {"$or": [filters, {"collection_ids": {"$in": folder_keys}}]}
  91. ##filters["collection_ids"] = {"$overlap": folder_keys}
  92. ## {"$and": {"$document_id": ..., "collection_ids": ...}}
  93. """
  94. {
  95. "$or": [
  96. {"document_id": {"$eq": "9fbe403b-..."}},
  97. {"collection_ids": {"$in": ["122fdf6a-...", "..."]}}
  98. ]
  99. }
  100. """
  101. print(filters)
  102. if len(filters["$or"]) < 2:
  103. filters = filters["$or"][0]
  104. print("filtersfiltersfiltersfiltersfiltersfiltersfiltersfiltersfiltersfilters")
  105. print(filters)
  106. """
  107. loop = asyncio.get_event_loop() # 获取当前事件循环
  108. loop.run_until_complete(r2r.init()) # 确保 r2r 已初始化
  109. search_results = loop.run_until_complete(r2r.search(query, filters=filters))
  110. asyncio.run(r2r.init())
  111. search_results = asyncio.run(r2r.search(query, filters=filters))
  112. search_results = loop.run_until_complete(
  113. r2r.search(query, filters={"file_key": {"$in": file_keys}})
  114. )
  115. """
  116. r2r.init_sync()
  117. search_results = r2r.search(query, filters=filters)
  118. if not search_results:
  119. return files
  120. for doc in search_results:
  121. file_key = doc.get("metadata").get("file_key")
  122. text = doc.get("text")
  123. if file_key in files and files[file_key]:
  124. files[file_key] += f"\n\n{text}"
  125. else:
  126. files[file_key] = doc.get("text")
  127. return files
  128. @staticmethod
  129. def list_in_files(
  130. ids: list[str] = None,
  131. offset: int = 0,
  132. limit: int = 100,
  133. ) -> dict:
  134. """
  135. loop = asyncio.get_event_loop() # 获取当前事件循环
  136. loop.run_until_complete(r2r.init()) # 确保 r2r 已初始化
  137. list_results = loop.run_until_complete(
  138. r2r.list(ids=ids, offset=offset, limit=limit)
  139. )
  140. asyncio.run(r2r.init())
  141. list_results = asyncio.run(r2r.list(ids=ids, offset=offset, limit=limit))
  142. """
  143. r2r.init_sync()
  144. list_results = r2r.list(ids=ids, offset=offset, limit=limit)
  145. return list_results
  146. @staticmethod
  147. def list_documents(
  148. id: str = "",
  149. offset: int = 0,
  150. limit: int = 100,
  151. ) -> dict:
  152. """
  153. loop = asyncio.get_event_loop() # 获取当前事件循环
  154. loop.run_until_complete(r2r.init()) # 确保 r2r 已初始化
  155. list_results = loop.run_until_complete(
  156. r2r.list_documents(id=id, offset=offset, limit=limit)
  157. )
  158. asyncio.run(r2r.init())
  159. list_results = asyncio.run(
  160. r2r.list_documents(id=id, offset=offset, limit=limit)
  161. )
  162. """
  163. r2r.init_sync()
  164. list_results = r2r.list_documents(id=id, offset=offset, limit=limit)
  165. return list_results
  166. # TODO 删除s3&r2r文件