RakeshUtekar commited on
Commit
b009562
·
verified ·
1 Parent(s): 2df5312

Update qwen_model.py

Browse files
Files changed (1) hide show
  1. qwen_model.py +9 -27
qwen_model.py CHANGED
@@ -18,35 +18,17 @@ qwen_pipeline = pipeline(
18
  tokenizer=tokenizer
19
  )
20
 
21
- def generate_response(retrieved_texts, query, max_new_tokens=500):
22
- """
23
- Generates a response based on the retrieved texts and query using Qwen.
24
- Args:
25
- retrieved_texts (list): List of retrieved text strings (e.g., from BLIP).
26
- query (str): The user's question about the image.
27
- max_new_tokens (int): Maximum tokens to generate for the answer.
28
- Returns:
29
- str: The generated answer.
30
- """
31
- # Construct a prompt that includes the image details as context
32
  context = "\n".join(retrieved_texts)
33
  prompt = f"This is the detail about the image:\n{context}\n\nQuestion: {query}\nAnswer:"
34
-
35
- # Generate the text
36
- result = qwen_pipeline(
37
- prompt,
38
- max_new_tokens=max_new_tokens,
39
- do_sample=True, # or False if you want deterministic output
40
- temperature=0.7, # tweak as needed
41
- )
42
-
43
- # The pipeline returns a list of dicts with key "generated_text"
44
- full_generation = result[0]["generated_text"]
45
-
46
- # Optionally parse out the final answer if the model repeats the prompt
47
- if "Answer:" in full_generation:
48
- final_answer = full_generation.split("Answer:")[-1].strip()
49
  else:
50
- final_answer = full_generation
51
 
52
  return final_answer
 
 
18
  tokenizer=tokenizer
19
  )
20
 
21
+ def generate_response(retrieved_texts, query, max_new_tokens=200):
 
 
 
 
 
 
 
 
 
 
22
  context = "\n".join(retrieved_texts)
23
  prompt = f"This is the detail about the image:\n{context}\n\nQuestion: {query}\nAnswer:"
24
+
25
+ result = qwen_pipeline(prompt, max_new_tokens=max_new_tokens, ...)
26
+ generated_text = result[0]["generated_text"]
27
+
28
+ if "Answer:" in generated_text:
29
+ final_answer = generated_text.split("Answer:")[-1].strip()
 
 
 
 
 
 
 
 
 
30
  else:
31
+ final_answer = generated_text
32
 
33
  return final_answer
34
+