.Make certain being compatible along with multiple frameworks, including.NET 6.0,. NET Structure 4.6.2, and.NET Standard 2.0 as well as above.Reduce dependencies to avoid variation disagreements and the requirement for tiing redirects.Translating Audio Information.One of the primary functionalities of the SDK is audio transcription. Creators may record audio documents asynchronously or even in real-time. Below is an example of just how to record an audio report:.making use of AssemblyAI.utilizing AssemblyAI.Transcripts.var customer = new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For regional documents, identical code can be made use of to achieve transcription.await using var stream = brand new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK additionally holds real-time audio transcription utilizing Streaming Speech-to-Text. This component is actually specifically practical for uses requiring prompt processing of audio information.using AssemblyAI.Realtime.wait for making use of var transcriber = brand new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Final: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for getting audio coming from a microphone for example.GetAudio( async (part) => await transcriber.SendAudioAsync( chunk)).wait for transcriber.CloseAsync().Making Use Of LeMUR for LLM Apps.The SDK includes with LeMUR to permit programmers to develop large language style (LLM) apps on vocal data. Here is actually an example:.var lemurTaskParams = brand new LemurTaskParams.Cause="Provide a brief review of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Knowledge Designs.Furthermore, the SDK features built-in support for audio intelligence styles, allowing sentiment review and also various other innovative components.var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// POSITIVE, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To read more, see the formal AssemblyAI blog.Image resource: Shutterstock.