private void Recognizer_Recognizing(object sender, SpeechRecognitionEventArgs e) { Dispatcher.Invoke(() => { MessageBlock.Text = $"{lastLine}\n{e.Result.Text.ToUpper()}"; }); }
// "Recognizing" events are fired every time we receive interim results during recognition (i.e. hypotheses) private void RecognizingHandler(object sender, SpeechRecognitionEventArgs e) { //If we are recognizing speech if (e.Result.Reason == ResultReason.RecognizingSpeech) { //Check who is talking requester = new Request("2", byteStream); requester.Start(); UnityEngine.Debug.LogFormat($"HYPOTHESIS: Text={e.Result.Text}"); UnityEngine.Debug.LogFormat($"Returned Message: {requester.ServerMessage}"); //Join thread. Wait for action to complete requester.Stop(); // Populate recognized string, ready for transcription. recognizedString = e.Result.Text; //If the user is detected if (requester.ServerMessage == userName) { // Dont transcribe and disable to stop Update() from changing it. recognizedText.enabled = false; UnityEngine.Debug.LogFormat($"User found or null value"); recognizedText.text = ""; } else { // Transcribe recognizedText.text = recognizedString; } } }
private void Recognized(object sender, SpeechRecognitionEventArgs e) { Console.WriteLine(); var test = GetResponse(e.Result.Properties); SpeechResult.Add(test); }
private void _speechRecognizer_Recognized(object sender, SpeechRecognitionEventArgs e) { var result = e.Result; // Checks result. if (result.Reason == ResultReason.RecognizedSpeech) { if (!String.IsNullOrWhiteSpace(result.Text)) { Console.WriteLine($"- {result.Text}"); _reconizedSpeeches.Add( new ReconizedSpeech { Result = result.Text }); } } else if (result.Reason == ResultReason.NoMatch) { Console.WriteLine($"NOMATCH: Speech could not be recognized."); } else if (result.Reason == ResultReason.Canceled) { var cancellation = CancellationDetails.FromResult(result); Console.WriteLine($"CANCELED: Reason={cancellation.Reason}"); if (cancellation.Reason == CancellationReason.Error) { Console.WriteLine($"CANCELED: ErrorCode={cancellation.ErrorCode}"); Console.WriteLine($"CANCELED: ErrorDetails={cancellation.ErrorDetails}"); Console.WriteLine($"CANCELED: Did you update the subscription info?"); } } }
// 识别过程中的中间结果 private void Recognizer_Recognizing(object sender, SpeechRecognitionEventArgs e) { if (!string.IsNullOrEmpty(e.Result.Text)) { Log("中间结果: " + e.Result.Text); } }
private static void Recognizer_Recognized(object sender, SpeechRecognitionEventArgs e) { if (GlobalFlowControl.moduleActivated == false || GlobalFlowControl.ChatbotInterrupted == true) { queryResult = null; return; } if (stopRecognizing) { return; } Debug.WriteLine("Recognized: " + e.Result.Text); if (isAskingQuestion) { if (e.Result.Text.Length > 9) { queryResult = e.Result.Text; stopRecognizing = true; } } else { if (IsContainKeyword(e.Result.Text)) { //queryResult = e.Result.Text; stopRecognizing = true; } } }
private void SpeechBotConnector_Recognizing(object sender, SpeechRecognitionEventArgs e) { if (e.Result.Reason == ResultReason.RecognizingKeyword) { Trace.TraceInformation($"Keyword Recognition: Verifying: {e.Result.Text}"); } }
/// <summary> /// Logs the final recognition result /// </summary> private void RecognizedEventHandler(SpeechRecognitionEventArgs e, RecoType rt) { TextBox log; if (rt == RecoType.Base) { log = this.baseModelLogText; this.SetCurrentText(this.baseModelCurrentText, e.Result.Text); } else { log = this.customModelLogText; this.SetCurrentText(this.customModelCurrentText, e.Result.Text); } this.WriteLine(log); this.WriteLine(log, $" --- Final result received. Reason: {e.Result.Reason.ToString()}. --- "); if (!string.IsNullOrEmpty(e.Result.Text)) { //write log this.WriteLine(log, e.Result.Text); //write result this.WriteLine(this.recognizedText, $"[{DateTime.Now.ToString("HH:mm:ss")}] Me: {e.Result.Text}"); SendMessageToBot(e.Result.Text); } }
public void Recognized(object s, SpeechRecognitionEventArgs e) { var rec = s as SpeechRecognizer; var result = e.Result; // Checks result. string newMessage = outputText.text; if (result.Reason == ResultReason.RecognizedSpeech) { buf += result.Text; } else if (result.Reason == ResultReason.NoMatch) { } else if (result.Reason == ResultReason.Canceled) { var cancellation = CancellationDetails.FromResult(result); newMessage = $"CANCELED: Reason={cancellation.Reason} ErrorDetails={cancellation.ErrorDetails}"; } lock (threadLocker) { waitingForReco = false; } }
private void _speechClient_Recognized(object sender, SpeechRecognitionEventArgs e) { Debug.WriteLine($"{e.SessionId} > Final result: {e.Result.Text}"); SpeechRecognizer r = (SpeechRecognizer)sender; SendTranscript(r.SpeechRecognitionLanguage, e.Result.Text); }
private void SpeechRecognition_Recognized(object sender, SpeechRecognitionEventArgs e) { if ((e.confidence >= SpeechRecognitionConfidenceThreshold)) { switch (e.command) { case Commands.ZOOM_IN: if (ptz != null) { ptz.ZoomLevel = 1; } break; case Commands.ZOOM_OUT: if (ptz != null) { ptz.ZoomLevel = 0; } break; case Commands.TRACK_ON: TrackingPresenter = true; break; case Commands.TRACK_OFF: TrackingPresenter = false; break; } } }
private void SpeechRecognition_Recognized(object sender, SpeechRecognitionEventArgs e) { Dictionary <string, ICommand> commands; if (EditorVisible) { commands = commands_Editor; } else if (VisualizerVisible) { commands = commands_Visualizer; } else { commands = commands_Manager; } string speechCommand = e.command; ICommand cmd; if (e.confidence >= SpeechRecognitionConfidenceThreshold) { if (!commands.TryGetValue(speechCommand, out cmd)) { ExecuteGesture(speechCommand, VisualizerVisible); //if not a UI command, assuming its a Gesture name - if Visualizer is visible, also executing keyboard shortcut for the gesture (apart from highighting it) } else if (cmd.CanExecute(null)) { cmd.Execute(null); } } }
private static async void OnSpeechRecognizedAsync(object sender, SpeechRecognitionEventArgs e) { if (e.Result.Reason == ResultReason.NoMatch) { System.Console.WriteLine("NO MATCH: Speech could not be recognized."); return; } string spokenText = e.Result.Text; System.Console.WriteLine($"RECOGNIZED: Text={spokenText}"); // Ignore everything that doesn't include the bot's name if (!spokenText.IncludesTheWords(s_twitchBotSettings.BotName)) { return; } spokenText = spokenText.RemoveText(s_twitchBotSettings.BotName); // Process the request if (spokenText.IncludesTheWords("sleep") || spokenText.IncludesTheWords("shut", "down")) { await SpeakAsync("Shutting down"); s_stopRecognition.TrySetCanceled(new CancellationToken(true)); } else if (spokenText.IncludesTheWords("Twitch")) { s_twitchBot.HandleTwitchCommand(spokenText); } }
/// <summary> /// Logs the final recognition result /// </summary> private void RecognizedEventHandler(SpeechRecognitionEventArgs e, RecoType rt) { SpeechRecognizedEventArgs eventArgs = new SpeechRecognizedEventArgs(); #if DEBUG if (rt == RecoType.Base) { eventArgs.Message = e.Result.Text; OnSpeechRecognizedEvent(eventArgs); } else { eventArgs.Message = e.Result.Text; OnSpeechRecognizedEvent(eventArgs); } eventArgs.Message = $" --- Final result received. Reason: {e.Result.Reason.ToString()}. --- "; OnSpeechRecognizedEvent(eventArgs); #endif if (!string.IsNullOrEmpty(e.Result.Text)) { //this.WriteLine(log, e.Result.Text); eventArgs.Message = e.Result.Text; OnSpeechRecognizedEvent(eventArgs); } // if access to the JSON is needed it can be obtained from Properties string json = e.Result.Properties.GetProperty(PropertyId.SpeechServiceResponse_JsonResult); }
private static void Recognizer_Recognizing(object sender, SpeechRecognitionEventArgs e) { //Console.ForegroundColor = ConsoleColor.Gray; //Console.WriteLine(); //Console.WriteLine($"SpeechRecognitionEventArgs={e.Result}"); //Console.WriteLine($"SpeechRecognitionEventArgs={e.Result.Text}"); }
private void RecognisedEventHandler(object sender, SpeechRecognitionEventArgs e) { if (!string.IsNullOrEmpty(e.Result.Text)) { onSpeech(e.Result.Text); } }
private async Task RecognizedEventHandlerAsync(SpeechRecognitionEventArgs e, RecoType rt) { TextBox log; if (rt == RecoType.Base) { log = this.baseModelLogText; this.SetCurrentText(this.baseModelCurrentText, e.Result.Text); if (e.Result.Text.Equals("What's the weather like?")) { var config = SpeechConfig.FromSubscription(this.SubscriptionKey, this.Region); config.SpeechRecognitionLanguage = this.RecognitionLanguage; SpeechSynthesizer basicRecognizer; if (this.UseMicrophone) { using (basicRecognizer = new SpeechSynthesizer(config)) { await basicRecognizer.SpeakTextAsync("cuacanya panas banget boo!"); } } } } else { log = this.customModelLogText; this.SetCurrentText(this.customModelCurrentText, e.Result.Text); } }
private void OnSpeechRecognized(object sender, SpeechRecognitionEventArgs e) { string previousResult = Interlocked.Exchange(ref _lastResult, e.Result.Text); switch (e.Result.Reason) { case ResultReason.RecognizingSpeech: Interlocked.Increment(ref _intermediateResultsReceived); Console.Out.Write($"[{e.Result.Text.Length}]\r"); if (previousResult != null && previousResult.Equals(e.Result.Text)) { Interlocked.Increment(ref _identicalResults); } break; case ResultReason.RecognizedSpeech: Interlocked.Increment(ref _finalResultsReceived); foreach (DetailedSpeechRecognitionResult result in SpeechRecognitionResultExtensions.Best(e.Result)) { string confidence = result.Confidence.ToString("F2"); string text = $"{confidence}|{result.Text}"; Trace.WriteLine(text); _transcript.WriteLine(text); Console.Out.WriteLine(text); } _transcript.WriteLine(); _transcript.Flush(); break; } }
private static void Recognizer_Recognizing(object sender, SpeechRecognitionEventArgs e) { if (GlobalFlowControl.moduleActivated == false || GlobalFlowControl.ChatbotInterrupted == true) { queryResult = null; return; } if (GlobalFlowControl.AskByPressingButton) { return; } queryResult = e.Result.Text; Debug.WriteLine("Recognizing: " + e.Result.Text); if (stopRecognizing) { return; } if (isAskingQuestion == false) { if (IsContainKeyword(queryResult)) { //queryResult = e.Result.Text; stopRecognizing = true; } } else { recognizingCount++; } }
private void RecognisingSpeechHandler(object sender, SpeechRecognitionEventArgs e) { if (e.Result.Text != null) { this.RecognisedSpeech?.Invoke(e.Result.Text); } }
private void SpeechBotConnector_Recognized(object sender, SpeechRecognitionEventArgs e) { if (e.Result.Reason == ResultReason.RecognizedSpeech) { this.RecognizedText = e.Result.Text; this.speechDuration = (int)e.Result.Duration.TotalMilliseconds; Trace.TraceInformation($"[{DateTime.Now.ToString("h:mm:ss tt", CultureInfo.CurrentCulture)}] Recognized event received. SessionId = {e.SessionId}, Speech duration = {this.speechDuration}, Recognized text = {this.RecognizedText}"); } else if (e.Result.Reason == ResultReason.RecognizedKeyword) { this.RecognizedKeyword = e.Result.Text; Trace.TraceInformation($"[{DateTime.Now.ToString("h:mm:ss tt", CultureInfo.CurrentCulture)}] Recognized event received. SessionId = {e.SessionId}"); Trace.TraceInformation($"Keyword Recognition Verified : {e.Result.Text}"); } else if (e.Result.Reason == ResultReason.NoMatch) { Trace.TraceInformation($"[{DateTime.Now.ToString("h:mm:ss tt", CultureInfo.CurrentCulture)}] Recognized event received. Speech could not be recognized. SessionId = {e.SessionId}"); Trace.TraceInformation($"No match details = {NoMatchDetails.FromResult(e.Result)}"); } else { Trace.TraceInformation($"[{DateTime.Now.ToString("h:mm:ss tt", CultureInfo.CurrentCulture)}] Recognized event received. e.Result.Reason = {e.Result.Reason}. SessionId = {e.SessionId}"); } }
//Writes the response result. private void EchoResponse(SpeechRecognitionEventArgs e) { WriteLine("Speech To Text Result:"); //handle the case when there are no results. //common situation is when there is a pause from user and audio captured has no speech in it if (e.Result.Text.Length == 0) { WriteLine("No phrase response is available."); WriteLine(); } else { WriteLine( "Text=\"{0}\"", e.Result.Text); WriteLine(); var botReplyTask = this.GetBotReplyAsync(e.Result.Text); string result = string.Empty; //send transcribed text to bot and get the response result = botReplyTask.Result; //Play audio from text to speech API var speakTask = this.PlaySpeechAudioAsync(result); } }
/// <summary> /// Logs the final recognition result. /// </summary> private void RecognizedEventHandler(SpeechRecognitionEventArgs e) { string json = e.Result.Properties.GetProperty(PropertyId.SpeechServiceResponse_JsonResult); var utt = JsonConvert.DeserializeObject <RealTimeUtt>(json); FinalResultsCumulative.Add(utt); }
private void SpeechRecognizer_Recognized(object sender, SpeechRecognitionEventArgs e) { DispatcherHelper.CheckBeginInvokeOnUI(() => { SelectedItem.Text += $" {e.Result.Text}"; }); }
/// <summary> /// Logs the final recognition result /// </summary> private void RecognizedEventHandler(SpeechRecognitionEventArgs e, RecoType rt) { TextBox log; if (rt == RecoType.Base) { log = this.baseModelLogText; this.SetCurrentText(this.baseModelCurrentText, e.Result.Text); } else { log = this.customModelLogText; this.SetCurrentText(this.customModelCurrentText, e.Result.Text); } this.WriteLine(log); this.WriteLine(log, $" --- Final result received. Reason: {e.Result.Reason.ToString()}. --- "); if (!string.IsNullOrEmpty(e.Result.Text)) { this.WriteLine(log, e.Result.Text); } // if access to the JSON is needed it can be obtained from Properties string json = e.Result.Properties.GetProperty(PropertyId.SpeechServiceResponse_JsonResult); }
private static void Recognized(object sender, SpeechRecognitionEventArgs e) { switch (e.Recognized) { case "英語": _language = Language.English; return; case "韓国語": _language = Language.Korean; return; case "スペイン語": _language = Language.Spanish; return; case "フランス語": _language = Language.French; return; case "ロシア語": _language = Language.Russian; return; } Console.WriteLine(e.Recognized); var translated = Translate(e.Recognized, _language); Console.WriteLine(translated); var mp3 = TextToSpeech(translated, _language); Play(mp3); }
// "Recognized" events are fired when the utterance end was detected by the server private void RecognizedHandler(object sender, SpeechRecognitionEventArgs e) { //This if statement creates the transcript :) if (e.Result.Reason == ResultReason.RecognizedSpeech) { UnityEngine.Debug.LogFormat($"RECOGNIZED: Text={e.Result.Text}"); //Will eliminate false speech pick-ups from transcript if (e.Result.Text.Contains(" ")) { transcript += (e.Result.Text + "\n"); } //Will accept valid answer as long as text length > 5 if (e.Result.Text.Length > 5) { questionNum++; //Goes to next question canPlay[questionNum] = true; //This (new) question's canPlay value is now true, meaning it will be played. } UnityEngine.Debug.Log(transcript); } else if (e.Result.Reason == ResultReason.NoMatch) { UnityEngine.Debug.LogFormat($"NOMATCH: Speech could not be recognized."); } }
// "Recognizing" events are fired every time we receive interim results during recognition (i.e. hypotheses) private void RecognizingHandler(object sender, SpeechRecognitionEventArgs e) { if (e.Result.Reason == ResultReason.RecognizingSpeech) { UnityEngine.Debug.LogFormat($"HYPOTHESIS: Text={e.Result.Text}"); } }
private static void Recognizer_Recognizing(object sender, SpeechRecognitionEventArgs e) { var result = e.Result; if (result.Reason == ResultReason.RecognizingSpeech) { Console.WriteLine($"We recognized: {result.Text}"); } else if (result.Reason == ResultReason.NoMatch) { Console.WriteLine($"NOMATCH: Speech could not be recognized."); } else if (result.Reason == ResultReason.Canceled) { var cancellation = CancellationDetails.FromResult(result); Console.WriteLine($"CANCELED: Reason={cancellation.Reason}"); if (cancellation.Reason == CancellationReason.Error) { Console.WriteLine($"CANCELED: ErrorCode={cancellation.ErrorCode}"); Console.WriteLine($"CANCELED: ErrorDetails={cancellation.ErrorDetails}"); Console.WriteLine($"CANCELED: Did you update the subscription info?"); } } }
private void RecognizingHandler(object sender, SpeechRecognitionEventArgs e) { lock (threadLocker) { message = e.Result.Text; } }
private void SpeechRecognition_Recognized(object sender, SpeechRecognitionEventArgs e) { Dictionary<string, ICommand> commands; if (EditorVisible) commands = commands_Editor; else if (VisualizerVisible) commands = commands_Visualizer; else commands = commands_Manager; string speechCommand = e.command; ICommand cmd; if (e.confidence >= SpeechRecognitionConfidenceThreshold) if (!commands.TryGetValue(speechCommand, out cmd)) ExecuteGesture(speechCommand, VisualizerVisible); //if not a UI command, assuming its a Gesture name - if Visualizer is visible, also executing keyboard shortcut for the gesture (apart from highighting it) else if (cmd.CanExecute(null)) cmd.Execute(null); }
void ipClient_SpeechRecognized(object sender, SpeechRecognitionEventArgs e) { RaiseSpeechRecognized(new CallInputEventArgs(e.LineNumber, e.Phrase)); }