Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -27,4 +27,15 @@ speaker_embedding = torch.tensor(embeddings_dataset[7306]["xvector"]).unsqueeze(
|
|
27 |
def text_to_audio(text):
|
28 |
speech = text_to_audio_pipe(text, forward_params={"speaker_embeddings": speaker_embedding})
|
29 |
sf.write("speech.wav", speech["audio"], samplerate=speech["sampling_rate"])
|
30 |
-
return "speech.wav"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
def text_to_audio(text):
|
28 |
speech = text_to_audio_pipe(text, forward_params={"speaker_embeddings": speaker_embedding})
|
29 |
sf.write("speech.wav", speech["audio"], samplerate=speech["sampling_rate"])
|
30 |
+
return "speech.wav"
|
31 |
+
|
32 |
+
# Gradio Interface
|
33 |
+
iface = gr.Interface(
|
34 |
+
fn=lambda image: text_to_audio(image_to_text(image)),
|
35 |
+
inputs=gr.Image(type="filepath"),
|
36 |
+
outputs=gr.Audio(label="Generated Audio"),
|
37 |
+
title="Image to Audio",
|
38 |
+
description="Upload an image to generate audio description."
|
39 |
+
)
|
40 |
+
|
41 |
+
iface.launch()
|