freddyaboulton HF staff commited on
Commit
891e37e
1 Parent(s): 0fcc790

first draft

Browse files
Files changed (1) hide show
  1. app.py +84 -0
app.py CHANGED
@@ -21,10 +21,94 @@ thread.start()
21
 
22
  API_URL = "http://0.0.0.0:60808/chat"
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  OUT_CHUNK = 4096
25
  OUT_RATE = 24000
26
  OUT_CHANNELS = 1
27
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  def process_audio(audio):
29
  filepath = audio
30
  print(f"filepath: {filepath}")
 
21
 
22
  API_URL = "http://0.0.0.0:60808/chat"
23
 
24
+ # recording parameters
25
+ IN_FORMAT = pyaudio.paInt16
26
+ IN_CHANNELS = 1
27
+ IN_RATE = 24000
28
+ IN_CHUNK = 1024
29
+ IN_SAMPLE_WIDTH = 2
30
+ VAD_STRIDE = 0.5
31
+
32
+ # playing parameters
33
+ OUT_FORMAT = pyaudio.paInt16
34
+ OUT_CHANNELS = 1
35
+ OUT_RATE = 24000
36
+ OUT_SAMPLE_WIDTH = 2
37
+ OUT_CHUNK = 5760
38
+
39
+
40
+
41
  OUT_CHUNK = 4096
42
  OUT_RATE = 24000
43
  OUT_CHANNELS = 1
44
 
45
+ def run_vad(ori_audio, sr):
46
+ _st = time.time()
47
+ try:
48
+ audio = np.frombuffer(ori_audio, dtype=np.int16)
49
+ audio = audio.astype(np.float32) / 32768.0
50
+ sampling_rate = 16000
51
+ if sr != sampling_rate:
52
+ audio = librosa.resample(audio, orig_sr=sr, target_sr=sampling_rate)
53
+
54
+ vad_parameters = {}
55
+ vad_parameters = VadOptions(**vad_parameters)
56
+ speech_chunks = get_speech_timestamps(audio, vad_parameters)
57
+ audio = collect_chunks(audio, speech_chunks)
58
+ duration_after_vad = audio.shape[0] / sampling_rate
59
+
60
+ if sr != sampling_rate:
61
+ # resample to original sampling rate
62
+ vad_audio = librosa.resample(audio, orig_sr=sampling_rate, target_sr=sr)
63
+ else:
64
+ vad_audio = audio
65
+ vad_audio = np.round(vad_audio * 32768.0).astype(np.int16)
66
+ vad_audio_bytes = vad_audio.tobytes()
67
+
68
+ return duration_after_vad, vad_audio_bytes, round(time.time() - _st, 4)
69
+ except Exception as e:
70
+ msg = f"[asr vad error] audio_len: {len(ori_audio)/(sr*2):.3f} s, trace: {traceback.format_exc()}"
71
+ print(msg)
72
+ return -1, ori_audio, round(time.time() - _st, 4)
73
+
74
+
75
+ def warm_up():
76
+ frames = b"\x00\x00" * 1024 * 2 # 1024 frames of 2 bytes each
77
+ dur, frames, tcost = run_vad(frames, 16000)
78
+ print(f"warm up done, time_cost: {tcost:.3f} s")
79
+
80
+ warm_up()
81
+
82
+ def determine_pause():
83
+ temp_audio = b""
84
+ vad_audio = b""
85
+
86
+ start_talking = False
87
+ last_temp_audio = None
88
+
89
+ while st.session_state.recording:
90
+ status.success("Listening...")
91
+ audio_bytes = stream.read(IN_CHUNK)
92
+ temp_audio += audio_bytes
93
+
94
+ if len(temp_audio) > IN_SAMPLE_WIDTH * IN_RATE * IN_CHANNELS * VAD_STRIDE:
95
+ dur_vad, vad_audio_bytes, time_vad = run_vad(temp_audio, IN_RATE)
96
+
97
+ print(f"duration_after_vad: {dur_vad:.3f} s, time_vad: {time_vad:.3f} s")
98
+
99
+ if dur_vad > 0.2 and not start_talking:
100
+ if last_temp_audio is not None:
101
+ st.session_state.frames.append(last_temp_audio)
102
+ start_talking = True
103
+ if start_talking:
104
+ st.session_state.frames.append(temp_audio)
105
+ if dur_vad < 0.1 and start_talking:
106
+ st.session_state.recording = False
107
+ print(f"speech end detected. excit")
108
+ last_temp_audio = temp_audio
109
+ temp_audio = b""
110
+
111
+
112
  def process_audio(audio):
113
  filepath = audio
114
  print(f"filepath: {filepath}")