.Make certain compatibility along with a number of platforms, including.NET 6.0,. Internet Platform 4.6.2, and.NET Criterion 2.0 as well as above.Reduce reliances to avoid version conflicts and the demand for tiing redirects.Translating Audio Information.Among the primary performances of the SDK is actually audio transcription. Developers can easily record audio data asynchronously or in real-time. Below is an instance of how to record an audio data:.using AssemblyAI.utilizing AssemblyAI.Transcripts.var client = brand-new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For nearby documents, identical code could be made use of to accomplish transcription.wait for making use of var stream = brand new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK likewise holds real-time sound transcription using Streaming Speech-to-Text. This feature is actually especially practical for uses needing urgent processing of audio records.using AssemblyAI.Realtime.wait for making use of var transcriber = new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Last: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for acquiring sound from a microphone for instance.GetAudio( async (portion) => await transcriber.SendAudioAsync( part)).wait for transcriber.CloseAsync().Utilizing LeMUR for LLM Functions.The SDK combines with LeMUR to make it possible for programmers to develop large language style (LLM) apps on vocal information. Below is actually an instance:.var lemurTaskParams = brand new LemurTaskParams.Cue="Provide a quick recap of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var reaction = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Cleverness Models.Furthermore, the SDK comes with built-in assistance for audio knowledge styles, enabling view analysis and various other sophisticated attributes.var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// BENEFICIAL, NEUTRAL, or even downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, see the formal AssemblyAI blog.Image resource: Shutterstock.