mirror of
https://github.com/Mintplex-Labs/anything-llm.git
synced 2024-11-20 05:10:10 +01:00
58 lines
2.0 KiB
Python
58 lines
2.0 KiB
Python
|
import os
|
||
|
from langchain.document_loaders import Docx2txtLoader, UnstructuredODTLoader
|
||
|
from slugify import slugify
|
||
|
from ..utils import guid, file_creation_time, write_to_server_documents, move_source
|
||
|
from ...utils import tokenize
|
||
|
|
||
|
# Process all text-related documents.
|
||
|
def as_docx(**kwargs):
|
||
|
parent_dir = kwargs.get('directory', 'hotdir')
|
||
|
filename = kwargs.get('filename')
|
||
|
ext = kwargs.get('ext', '.txt')
|
||
|
fullpath = f"{parent_dir}/{filename}{ext}"
|
||
|
|
||
|
loader = Docx2txtLoader(fullpath)
|
||
|
data = loader.load()[0]
|
||
|
content = data.page_content
|
||
|
|
||
|
print(f"-- Working {fullpath} --")
|
||
|
data = {
|
||
|
'id': guid(),
|
||
|
'url': "file://"+os.path.abspath(f"{parent_dir}/processed/{filename}{ext}"),
|
||
|
'title': f"{filename}{ext}",
|
||
|
'description': "a custom file uploaded by the user.",
|
||
|
'published': file_creation_time(fullpath),
|
||
|
'wordCount': len(content),
|
||
|
'pageContent': content,
|
||
|
'token_count_estimate': len(tokenize(content))
|
||
|
}
|
||
|
|
||
|
write_to_server_documents(data, f"{slugify(filename)}-{data.get('id')}")
|
||
|
move_source(parent_dir, f"{filename}{ext}")
|
||
|
print(f"[SUCCESS]: {filename}{ext} converted & ready for embedding.\n")
|
||
|
|
||
|
def as_odt(**kwargs):
|
||
|
parent_dir = kwargs.get('directory', 'hotdir')
|
||
|
filename = kwargs.get('filename')
|
||
|
ext = kwargs.get('ext', '.txt')
|
||
|
fullpath = f"{parent_dir}/{filename}{ext}"
|
||
|
|
||
|
loader = UnstructuredODTLoader(fullpath)
|
||
|
data = loader.load()[0]
|
||
|
content = data.page_content
|
||
|
|
||
|
print(f"-- Working {fullpath} --")
|
||
|
data = {
|
||
|
'id': guid(),
|
||
|
'url': "file://"+os.path.abspath(f"{parent_dir}/processed/{filename}{ext}"),
|
||
|
'title': f"{filename}{ext}",
|
||
|
'description': "a custom file uploaded by the user.",
|
||
|
'published': file_creation_time(fullpath),
|
||
|
'wordCount': len(content),
|
||
|
'pageContent': content,
|
||
|
'token_count_estimate': len(tokenize(content))
|
||
|
}
|
||
|
|
||
|
write_to_server_documents(data, f"{slugify(filename)}-{data.get('id')}")
|
||
|
move_source(parent_dir, f"{filename}{ext}")
|
||
|
print(f"[SUCCESS]: {filename}{ext} converted & ready for embedding.\n")
|