Spaces:
Running
on
T4
Running
on
T4
Update auditqa/doc_process.py
Browse files- auditqa/doc_process.py +4 -4
auditqa/doc_process.py
CHANGED
|
@@ -5,10 +5,10 @@ from transformers import AutoTokenizer
|
|
| 5 |
from langchain_community.document_loaders import PyMuPDFLoader
|
| 6 |
from langchain_community.embeddings import HuggingFaceEmbeddings, HuggingFaceInferenceAPIEmbeddings
|
| 7 |
from langchain_community.vectorstores import Qdrant
|
| 8 |
-
from dotenv import load_dotenv
|
| 9 |
-
load_dotenv()
|
| 10 |
|
| 11 |
-
HF_token = os.environ["HF_TOKEN"]
|
| 12 |
path_to_data = "./data/"
|
| 13 |
|
| 14 |
|
|
@@ -45,7 +45,7 @@ def process_pdf():
|
|
| 45 |
print(all_documents.keys())
|
| 46 |
|
| 47 |
|
| 48 |
-
embeddings =
|
| 49 |
model_name="sentence-transformers/all-mpnet-base-v2"
|
| 50 |
)
|
| 51 |
|
|
|
|
| 5 |
from langchain_community.document_loaders import PyMuPDFLoader
|
| 6 |
from langchain_community.embeddings import HuggingFaceEmbeddings, HuggingFaceInferenceAPIEmbeddings
|
| 7 |
from langchain_community.vectorstores import Qdrant
|
| 8 |
+
#from dotenv import load_dotenv
|
| 9 |
+
#load_dotenv()
|
| 10 |
|
| 11 |
+
#HF_token = os.environ["HF_TOKEN"]
|
| 12 |
path_to_data = "./data/"
|
| 13 |
|
| 14 |
|
|
|
|
| 45 |
print(all_documents.keys())
|
| 46 |
|
| 47 |
|
| 48 |
+
embeddings = HuggingFaceEmbeddings(
|
| 49 |
model_name="sentence-transformers/all-mpnet-base-v2"
|
| 50 |
)
|
| 51 |
|