Spaces:
Sleeping
Sleeping
Юра Цепліцький
commited on
Commit
·
1a0f750
1
Parent(s):
af8b652
Switch to cohere command r model
Browse files- app.py +5 -3
- main.py +12 -7
- utils/__pycache__/retriever.cpython-312.pyc +0 -0
- utils/__pycache__/settings.cpython-312.pyc +0 -0
- utils/retriever.py +9 -5
- utils/settings.py +15 -8
app.py
CHANGED
@@ -11,8 +11,7 @@ model_config = ConfigDict(protected_namespaces=())
|
|
11 |
setting_keys = gr.Interface(
|
12 |
fn=set_keys,
|
13 |
inputs=[
|
14 |
-
|
15 |
-
gr.Textbox(label="Enter your OPENAI_API_KEY"),
|
16 |
gr.Textbox(label="Enter your LLAMA_CLOUD_API_KEY"),
|
17 |
],
|
18 |
outputs=gr.Textbox(label="Status")
|
@@ -31,7 +30,10 @@ uploading_files = gr.Interface(
|
|
31 |
qa = gr.Interface(
|
32 |
fn=answer_query,
|
33 |
inputs=gr.Textbox(label="Enter your question"),
|
34 |
-
outputs=
|
|
|
|
|
|
|
35 |
title="Document Q&A System"
|
36 |
)
|
37 |
|
|
|
11 |
setting_keys = gr.Interface(
|
12 |
fn=set_keys,
|
13 |
inputs=[
|
14 |
+
gr.Textbox(label="Enter your CO_API_KEY"),
|
|
|
15 |
gr.Textbox(label="Enter your LLAMA_CLOUD_API_KEY"),
|
16 |
],
|
17 |
outputs=gr.Textbox(label="Status")
|
|
|
30 |
qa = gr.Interface(
|
31 |
fn=answer_query,
|
32 |
inputs=gr.Textbox(label="Enter your question"),
|
33 |
+
outputs=[
|
34 |
+
gr.Textbox(label="Answer"),
|
35 |
+
gr.Textbox(label="Relevant Nodes"),
|
36 |
+
],
|
37 |
title="Document Q&A System"
|
38 |
)
|
39 |
|
main.py
CHANGED
@@ -1,13 +1,13 @@
|
|
1 |
-
from utils.retriever import
|
2 |
from utils.index import create_index
|
3 |
from utils.constant import INDEX_PATH, DATA_PATH
|
4 |
import os
|
5 |
import shutil
|
6 |
|
7 |
-
|
|
|
8 |
try:
|
9 |
-
|
10 |
-
os.environ["OPENAI_API_KEY"] = openai_api_key
|
11 |
os.environ["LLAMA_CLOUD_API_KEY"] = llama_cloud_api_key
|
12 |
return "Keys are set successfully"
|
13 |
|
@@ -23,6 +23,10 @@ def handle_file(uploaded_file):
|
|
23 |
shutil.move(temp_file_path, save_file_path)
|
24 |
|
25 |
create_index(DATA_PATH, INDEX_PATH)
|
|
|
|
|
|
|
|
|
26 |
return "File uploaded and indexed"
|
27 |
|
28 |
except Exception as e:
|
@@ -30,19 +34,20 @@ def handle_file(uploaded_file):
|
|
30 |
|
31 |
def answer_query(query: str) -> str:
|
32 |
|
33 |
-
query_engine =
|
34 |
response = query_engine.query(query)
|
35 |
|
36 |
nodes = query_engine.retriever.retrieve(query)
|
|
|
37 |
|
38 |
for node in nodes:
|
39 |
|
40 |
score = node.get_score()
|
41 |
text = node.text
|
42 |
|
43 |
-
|
44 |
|
45 |
-
return response
|
46 |
|
47 |
if __name__ == "__main__":
|
48 |
|
|
|
1 |
+
from utils.retriever import get_engine
|
2 |
from utils.index import create_index
|
3 |
from utils.constant import INDEX_PATH, DATA_PATH
|
4 |
import os
|
5 |
import shutil
|
6 |
|
7 |
+
|
8 |
+
def set_keys(co_api_key: str, llama_cloud_api_key: str) -> str:
|
9 |
try:
|
10 |
+
os.environ["CO_API_KEY"] = co_api_key
|
|
|
11 |
os.environ["LLAMA_CLOUD_API_KEY"] = llama_cloud_api_key
|
12 |
return "Keys are set successfully"
|
13 |
|
|
|
23 |
shutil.move(temp_file_path, save_file_path)
|
24 |
|
25 |
create_index(DATA_PATH, INDEX_PATH)
|
26 |
+
|
27 |
+
engine_manager = get_engine()[1]
|
28 |
+
engine_manager.initialize_index()
|
29 |
+
|
30 |
return "File uploaded and indexed"
|
31 |
|
32 |
except Exception as e:
|
|
|
34 |
|
35 |
def answer_query(query: str) -> str:
|
36 |
|
37 |
+
query_engine = get_engine(semantic=True)[0]
|
38 |
response = query_engine.query(query)
|
39 |
|
40 |
nodes = query_engine.retriever.retrieve(query)
|
41 |
+
nodes_str = ""
|
42 |
|
43 |
for node in nodes:
|
44 |
|
45 |
score = node.get_score()
|
46 |
text = node.text
|
47 |
|
48 |
+
nodes_str += f"\nNode: {node.node_id}\nScore: {score:0.3f}\nText: {text[:1000]}\n"
|
49 |
|
50 |
+
return response, nodes_str
|
51 |
|
52 |
if __name__ == "__main__":
|
53 |
|
utils/__pycache__/retriever.cpython-312.pyc
CHANGED
Binary files a/utils/__pycache__/retriever.cpython-312.pyc and b/utils/__pycache__/retriever.cpython-312.pyc differ
|
|
utils/__pycache__/settings.cpython-312.pyc
CHANGED
Binary files a/utils/__pycache__/settings.cpython-312.pyc and b/utils/__pycache__/settings.cpython-312.pyc differ
|
|
utils/retriever.py
CHANGED
@@ -25,14 +25,18 @@ class QueryEngineManager:
|
|
25 |
self.reranker = None
|
26 |
self.query_engine = None
|
27 |
self._configure()
|
28 |
-
|
29 |
-
def
|
30 |
-
configure_settings()
|
31 |
self.index = load_index(path=INDEX_PATH)
|
32 |
self.nodes = list(self.index.docstore.docs.values())
|
|
|
|
|
|
|
|
|
33 |
self.reranker = LLMRerank(top_n=TOP_N_RERANKER)
|
34 |
|
35 |
def get_engine(self, bm25: bool = False, semantic: bool = False):
|
|
|
36 |
if bm25:
|
37 |
self.retriever = BM25Retriever.from_defaults(
|
38 |
nodes=self.nodes,
|
@@ -59,6 +63,6 @@ class QueryEngineManager:
|
|
59 |
|
60 |
return self.query_engine
|
61 |
|
62 |
-
def
|
63 |
engine_manager = QueryEngineManager()
|
64 |
-
return engine_manager.get_engine(bm25, semantic)
|
|
|
25 |
self.reranker = None
|
26 |
self.query_engine = None
|
27 |
self._configure()
|
28 |
+
|
29 |
+
def initialize_index(self):
|
|
|
30 |
self.index = load_index(path=INDEX_PATH)
|
31 |
self.nodes = list(self.index.docstore.docs.values())
|
32 |
+
|
33 |
+
def _configure(self):
|
34 |
+
configure_settings()
|
35 |
+
self.initialize_index()
|
36 |
self.reranker = LLMRerank(top_n=TOP_N_RERANKER)
|
37 |
|
38 |
def get_engine(self, bm25: bool = False, semantic: bool = False):
|
39 |
+
|
40 |
if bm25:
|
41 |
self.retriever = BM25Retriever.from_defaults(
|
42 |
nodes=self.nodes,
|
|
|
63 |
|
64 |
return self.query_engine
|
65 |
|
66 |
+
def get_engine(bm25: bool = False, semantic: bool = False):
|
67 |
engine_manager = QueryEngineManager()
|
68 |
+
return engine_manager.get_engine(bm25, semantic), engine_manager
|
utils/settings.py
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
from llama_index.core import Settings
|
2 |
-
|
3 |
-
from llama_index.llms.
|
|
|
4 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
5 |
from llama_index.core.node_parser import SemanticSplitterNodeParser
|
6 |
|
@@ -16,15 +17,20 @@ def load_llm():
|
|
16 |
- Do not include information from external sources not provided by the user.
|
17 |
'''
|
18 |
|
19 |
-
|
20 |
-
|
21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
# )
|
23 |
|
24 |
-
llm = OpenAI(
|
25 |
-
model = "gpt-4o-mini",
|
26 |
-
system_prompt=system_prompt
|
27 |
-
)
|
28 |
|
29 |
return llm
|
30 |
|
@@ -43,6 +49,7 @@ def configure_settings():
|
|
43 |
llm_replicate = load_llm()
|
44 |
embed_model = load_embed_model()
|
45 |
|
|
|
46 |
Settings.llm = llm_replicate
|
47 |
Settings.embed_model = embed_model
|
48 |
Settings.node_parser = SemanticSplitterNodeParser(
|
|
|
1 |
from llama_index.core import Settings
|
2 |
+
from llama_index.llms.cohere import Cohere
|
3 |
+
#from llama_index.llms.anthropic import Anthropic
|
4 |
+
#from llama_index.llms.openai import OpenAI
|
5 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
6 |
from llama_index.core.node_parser import SemanticSplitterNodeParser
|
7 |
|
|
|
17 |
- Do not include information from external sources not provided by the user.
|
18 |
'''
|
19 |
|
20 |
+
llm = Cohere(
|
21 |
+
system_prompt=system_prompt,
|
22 |
|
23 |
+
)
|
24 |
+
|
25 |
+
# llm = OpenAI(
|
26 |
+
# model = "gpt-3.5-turbo",
|
27 |
+
# system_prompt=system_prompt
|
28 |
+
# )
|
29 |
+
# llm = Anthropic(
|
30 |
+
# model="claude-3-opus-20240229",
|
31 |
+
# system_prompt=system_prompt
|
32 |
# )
|
33 |
|
|
|
|
|
|
|
|
|
34 |
|
35 |
return llm
|
36 |
|
|
|
49 |
llm_replicate = load_llm()
|
50 |
embed_model = load_embed_model()
|
51 |
|
52 |
+
#Settings.tokenizer = Anthropic().tokenizer
|
53 |
Settings.llm = llm_replicate
|
54 |
Settings.embed_model = embed_model
|
55 |
Settings.node_parser = SemanticSplitterNodeParser(
|