mirror of
https://github.com/ION606/ML-pipeline.git
synced 2026-05-14 21:06:54 +00:00
coloring fixes
This commit is contained in:
@@ -59,10 +59,10 @@ def create_vector_store(chunks):
|
||||
embeddings = HuggingFaceEmbeddings(model_name="all-MiniLM-L6-v2")
|
||||
|
||||
vector_store = Chroma.from_texts(
|
||||
chunks,
|
||||
embeddings,
|
||||
persist_directory=Config.chroma_path()
|
||||
)
|
||||
chunks,
|
||||
embeddings,
|
||||
persist_directory=Config.chroma_path()
|
||||
)
|
||||
|
||||
debugMod.log("Vector store created")
|
||||
return vector_store
|
||||
@@ -183,14 +183,14 @@ def orchestrate(query, vector_store=None, comm_outp=print, comm_inp=input):
|
||||
return [final_response, links]
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
if __name__ == "__main__":
|
||||
debugMod.moveDebugLog()
|
||||
|
||||
parser = argparse.ArgumentParser()
|
||||
parser.add_argument('--file', type=str, default="",
|
||||
help='Path to data file for analysis')
|
||||
help='Path to data file for analysis')
|
||||
parser.add_argument('--cli', type=str, default="false",
|
||||
help="whether to use the CLI for input or run the API")
|
||||
help="whether to use the CLI for input or run the API")
|
||||
args = parser.parse_args()
|
||||
|
||||
vector_store = None
|
||||
@@ -227,7 +227,7 @@ if __name__ == "__main__":
|
||||
if code_blocks:
|
||||
show_thinking('[running code...]')
|
||||
orchestrate_code(orchestrate, vector_store, chunks,
|
||||
user_env, code_blocks, query, response, links)
|
||||
user_env, code_blocks, query, response, links)
|
||||
|
||||
# clean up
|
||||
try:
|
||||
|
||||
Reference in New Issue
Block a user