Llama-2-7B
Browse files
app.py
CHANGED
@@ -35,8 +35,8 @@ def get_vector_store(target_source_chunks):
|
|
35 |
return retriver
|
36 |
|
37 |
def get_conversation_chain(retriever):
|
38 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
39 |
-
model = AutoModelForCausalLM.from_pretrained("
|
40 |
memory = ConversationBufferMemory(memory_key='chat_history', return_messages=True,)
|
41 |
chain = RetrievalQA.from_llm(
|
42 |
llm=model,
|
|
|
35 |
return retriver
|
36 |
|
37 |
def get_conversation_chain(retriever):
|
38 |
+
tokenizer = AutoTokenizer.from_pretrained("TinyPixel/Llama-2-7B-bf16-sharded")
|
39 |
+
model = AutoModelForCausalLM.from_pretrained("TinyPixel/Llama-2-7B-bf16-sharded")
|
40 |
memory = ConversationBufferMemory(memory_key='chat_history', return_messages=True,)
|
41 |
chain = RetrievalQA.from_llm(
|
42 |
llm=model,
|