Spaces:
Runtime error
Runtime error
| from transformers import pipeline, SpeechT5Processor, SpeechT5ForTextToSpeech, SpeechT5HifiGan | |
| import gradio as gr | |
| import torch | |
| import numpy as np | |
| from datasets import load_dataset, Audio | |
| from transformers import pipeline | |
| import librosa | |
| from openai import OpenAI | |
| # Load ASR model | |
| asr_pipe = pipeline(model="divakaivan/glaswegian-asr") | |
| # Load TTS components | |
| processor = SpeechT5Processor.from_pretrained("microsoft/speecht5_tts") | |
| tts_model = SpeechT5ForTextToSpeech.from_pretrained("divakaivan/glaswegian_tts") | |
| vocoder = SpeechT5HifiGan.from_pretrained("microsoft/speecht5_hifigan") | |
| # Load dataset for speaker embedding | |
| dataset = load_dataset("divakaivan/glaswegian_audio") | |
| dataset = dataset.cast_column("audio", Audio(sampling_rate=16000))['train'] | |
| def transcribe(audio): | |
| text = asr_pipe(audio)["text"] | |
| return text | |
| def generate_response(text, api_key): | |
| client = OpenAI(api_key=api_key) | |
| response = client.chat.completions.create( | |
| model='gpt-3.5-turbo-0125', | |
| messages=[{"role": "user", "content": text}] | |
| ) | |
| return response.choices[0].message.content | |
| def synthesize_speech(text): | |
| inputs = processor(text=text, return_tensors="pt") | |
| speaker_embeddings = create_speaker_embedding(dataset[0]["audio"]["array"]) | |
| spectrogram = tts_model.generate_speech(inputs["input_ids"], torch.tensor([speaker_embeddings])) | |
| with torch.no_grad(): | |
| speech = vocoder(spectrogram) | |
| speech = (speech.numpy() * 32767).astype(np.int16) | |
| return (16000, speech) | |
| def create_speaker_embedding(waveform): | |
| import os | |
| from speechbrain.inference.speaker import EncoderClassifier | |
| spk_model_name = "speechbrain/spkrec-xvect-voxceleb" | |
| device = "cuda" if torch.cuda.is_available() else "cpu" | |
| speaker_model = EncoderClassifier.from_hparams( | |
| source=spk_model_name, | |
| run_opts={"device": device}, | |
| savedir=os.path.join("/tmp", spk_model_name), | |
| ) | |
| with torch.no_grad(): | |
| speaker_embeddings = speaker_model.encode_batch(torch.tensor(waveform)) | |
| speaker_embeddings = torch.nn.functional.normalize(speaker_embeddings, dim=2) | |
| speaker_embeddings = speaker_embeddings.squeeze().cpu().numpy() | |
| return speaker_embeddings | |
| def voice_assistant(audio, api_key): | |
| transcribed_text = transcribe(audio) | |
| response_text = generate_response(transcribed_text, api_key) | |
| speech_audio = synthesize_speech(response_text) | |
| return speech_audio | |
| iface = gr.Interface( | |
| fn=voice_assistant, | |
| inputs=[ | |
| gr.Audio(type="filepath"), | |
| gr.Textbox(label="OpenAI API Key", type="password") | |
| ], | |
| outputs=gr.Audio(label="Response Speech", type="numpy"), | |
| title="Your Glaswegian Assistant" | |
| ) | |
| iface.launch() | |