.Make certain compatibility along with multiple platforms, including.NET 6.0,. Internet Structure 4.6.2, and.NET Specification 2.0 as well as above.Reduce dependences to stop variation conflicts and the demand for tiing redirects.Transcribing Audio Information.One of the primary functionalities of the SDK is actually audio transcription. Designers can record audio files asynchronously or in real-time. Below is actually an example of how to translate an audio documents:.using AssemblyAI.using AssemblyAI.Transcripts.var customer = brand new AssemblyAIClient(" YOUR_API_KEY").var records = wait for client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local data, similar code could be used to achieve transcription.wait for utilizing var stream = brand-new FileStream("./ nbc.mp3", FileMode.Open).var records = wait for client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK likewise sustains real-time sound transcription using Streaming Speech-to-Text. This feature is particularly helpful for treatments calling for urgent handling of audio information.utilizing AssemblyAI.Realtime.await using var scribe = brand-new RealtimeTranscriber( brand new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Ultimate: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for obtaining sound from a microphone for instance.GetAudio( async (part) => await transcriber.SendAudioAsync( chunk)).wait for transcriber.CloseAsync().Making Use Of LeMUR for LLM Functions.The SDK includes with LeMUR to make it possible for programmers to develop large foreign language style (LLM) functions on vocal records. Below is an example:.var lemurTaskParams = brand-new LemurTaskParams.Trigger="Offer a quick rundown of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Cleverness Models.Furthermore, the SDK comes with built-in help for audio intelligence designs, enabling belief review and various other innovative components.var records = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = accurate. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For additional information, go to the main AssemblyAI blog.Image source: Shutterstock.