|
6 | 6 | import queue
|
7 | 7 | import re
|
8 | 8 | import json
|
9 |
| -import torch |
10 |
| -import pyaudio |
11 |
| -import wave |
12 |
| -from typing import Optional, List, Dict, Any |
13 |
| -from gtts import gTTS |
14 |
| -from faster_whisper import WhisperModel |
15 |
| -import pygame |
16 |
| - |
17 |
| -# Audio Configuration |
18 |
| -FORMAT = pyaudio.paInt16 |
19 |
| -CHANNELS = 1 |
20 |
| -RATE = 16000 |
21 |
| -CHUNK = 512 |
22 |
| - |
23 |
| -# State Management |
24 |
| -is_speaking = False |
25 |
| -should_stop_speaking = False |
26 |
| -tts_sequence = 0 |
27 |
| -recording_data = [] |
28 |
| -buffer_data = [] |
29 |
| -is_recording = False |
30 |
| -last_speech_time = 0 |
31 |
| -running = True |
32 |
| - |
33 |
| -# Queues |
34 |
| -audio_queue = queue.Queue() |
35 |
| -tts_queue = queue.PriorityQueue() |
36 |
| -cleanup_files = [] |
37 |
| - |
38 |
| -# Initialize pygame mixer |
39 |
| -pygame.mixer.quit() |
40 |
| -pygame.mixer.init(frequency=44100, size=-16, channels=2, buffer=512) |
41 |
| - |
42 |
| -# Device selection |
43 |
| -device = "cpu" |
44 |
| -print(f"Using device: {device}") |
45 | 9 |
|
46 |
| -# Load VAD model |
47 |
| -print("Loading Silero VAD model...") |
48 |
| -vad_model, _ = torch.hub.load( |
49 |
| - repo_or_dir="snakers4/silero-vad", |
50 |
| - model="silero_vad", |
51 |
| - force_reload=False, |
52 |
| - onnx=False, |
53 |
| - verbose=False, |
54 |
| -) |
55 |
| -vad_model.to(device) |
| 10 | +import subprocess |
56 | 11 |
|
57 |
| -# Load Whisper model |
58 |
| -print("Loading Whisper model...") |
59 |
| -whisper_model = WhisperModel("base", device=device, compute_type="int8") |
| 12 | +try: |
| 13 | + import torch |
| 14 | + import pyaudio |
| 15 | + import wave |
| 16 | + from typing import Optional, List, Dict, Any |
| 17 | + from gtts import gTTS |
| 18 | + from faster_whisper import WhisperModel |
| 19 | + import pygame |
| 20 | + |
| 21 | + FORMAT = pyaudio.paInt16 |
| 22 | + CHANNELS = 1 |
| 23 | + RATE = 16000 |
| 24 | + CHUNK = 512 |
| 25 | + |
| 26 | + # State Management |
| 27 | + is_speaking = False |
| 28 | + should_stop_speaking = False |
| 29 | + tts_sequence = 0 |
| 30 | + recording_data = [] |
| 31 | + buffer_data = [] |
| 32 | + is_recording = False |
| 33 | + last_speech_time = 0 |
| 34 | + running = True |
| 35 | + |
| 36 | + # Queues |
| 37 | + audio_queue = queue.Queue() |
| 38 | + tts_queue = queue.PriorityQueue() |
| 39 | + cleanup_files = [] |
| 40 | + |
| 41 | + # Initialize pygame mixer |
| 42 | + pygame.mixer.quit() |
| 43 | + pygame.mixer.init(frequency=44100, size=-16, channels=2, buffer=512) |
| 44 | + |
| 45 | + # Device selection |
| 46 | + device = "cpu" |
| 47 | + print(f"Using device: {device}") |
| 48 | + |
| 49 | + # Load VAD model |
| 50 | + print("Loading Silero VAD model...") |
| 51 | + vad_model, _ = torch.hub.load( |
| 52 | + repo_or_dir="snakers4/silero-vad", |
| 53 | + model="silero_vad", |
| 54 | + force_reload=False, |
| 55 | + onnx=False, |
| 56 | + verbose=False, |
| 57 | + ) |
| 58 | + vad_model.to(device) |
60 | 59 |
|
61 |
| -# Conversation History Management |
62 |
| -history = [] |
63 |
| -max_history = 10 |
64 |
| -memory_file = "conversation_history.json" |
| 60 | + # Load Whisper model |
| 61 | + print("Loading Whisper model...") |
| 62 | + whisper_model = WhisperModel("base", device=device, compute_type="int8") |
65 | 63 |
|
66 |
| -# Import the get_llm_response function |
67 |
| -from npcsh.llm_funcs import get_llm_response |
68 |
| -from npcsh.npc_sysenv import NPCSH_CHAT_PROVIDER, NPCSH_CHAT_MODEL, NPCSH_API_URL |
69 |
| -import subprocess |
| 64 | + # Conversation History Management |
| 65 | + history = [] |
| 66 | + max_history = 10 |
| 67 | + memory_file = "conversation_history.json" |
70 | 68 |
|
71 |
| -print(NPCSH_CHAT_PROVIDER, NPCSH_CHAT_MODEL, NPCSH_API_URL) |
72 | 69 |
|
73 |
| -import subprocess |
| 70 | +except: |
| 71 | + print("audio dependencies not installed") |
74 | 72 |
|
75 | 73 |
|
76 | 74 | def convert_mp3_to_wav(mp3_file, wav_file):
|
|
0 commit comments