umerforsure commited on
Commit
ba51d64
Β·
1 Parent(s): 894d55e

πŸ” Switched to flan-t5-large for stable CPU generation

Browse files
Files changed (2) hide show
  1. app.py +7 -4
  2. requirements.txt +1 -1
app.py CHANGED
@@ -12,13 +12,15 @@ from langchain_community.embeddings import HuggingFaceEmbeddings
12
  from langchain.text_splitter import RecursiveCharacterTextSplitter
13
  from langchain_core.documents import Document
14
 
15
- # Load Reasoning Model (CPU-friendly with better reasoning)
16
- model_id = "microsoft/phi-1_5"
 
 
17
  tokenizer = AutoTokenizer.from_pretrained(model_id)
18
- model = AutoModelForCausalLM.from_pretrained(model_id)
19
 
20
  reasoning_pipeline = pipeline(
21
- "text-generation",
22
  model=model,
23
  tokenizer=tokenizer,
24
  max_new_tokens=512,
@@ -26,6 +28,7 @@ reasoning_pipeline = pipeline(
26
  top_p=0.9
27
  )
28
 
 
29
  # Embedding Model
30
  embedding_model = HuggingFaceEmbeddings(model_name="all-MiniLM-L6-v2")
31
  vectorstore = None
 
12
  from langchain.text_splitter import RecursiveCharacterTextSplitter
13
  from langchain_core.documents import Document
14
 
15
+ # Load Reasoning Model (CPU-friendly + stronger than base)
16
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline
17
+
18
+ model_id = "google/flan-t5-large"
19
  tokenizer = AutoTokenizer.from_pretrained(model_id)
20
+ model = AutoModelForSeq2SeqLM.from_pretrained(model_id)
21
 
22
  reasoning_pipeline = pipeline(
23
+ "text2text-generation",
24
  model=model,
25
  tokenizer=tokenizer,
26
  max_new_tokens=512,
 
28
  top_p=0.9
29
  )
30
 
31
+
32
  # Embedding Model
33
  embedding_model = HuggingFaceEmbeddings(model_name="all-MiniLM-L6-v2")
34
  vectorstore = None
requirements.txt CHANGED
@@ -6,6 +6,6 @@ python-docx
6
  gradio
7
  PyPDF2
8
  langchain
9
- langchain_community
10
  sentence-transformers
 
11
  accelerate
 
6
  gradio
7
  PyPDF2
8
  langchain
 
9
  sentence-transformers
10
+ langchain_community
11
  accelerate