Spaces:
Running
Running
from huggingface_hub import InferenceClient | |
import os | |
import gradio as gr | |
hf_token = os.getenv("HF_TOKEN").strip() | |
client = InferenceClient(api_key="hf_token") | |
messages = [ | |
{ "role": "user", "content": "You are ACC-o3, created by the ACC(Algorithmic Computer-generated Consciousness)." }, | |
] | |
stream = client.chat.completions.create( | |
model="deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", | |
messages=messages, | |
temperature=0.5, | |
max_tokens=2048, | |
top_p=0.7, | |
stream=True | |
) | |
for chunk in stream: | |
print(chunk.choices[0].delta.content) | |
demo = gr.ChatInterface( | |
fn=chat_with_model, | |
title="ACC-o3", | |
description="ACC=o3 is a powerful model created by the ACC.", | |
examples=["How many 'r's are in strawberry?", "Tell me about the ACC.", "Who are you?"] | |
) | |
if __name__ == "__main__": | |
demo.launch() |