.Make sure being compatible with various frameworks, including.NET 6.0,. NET Framework 4.6.2, and.NET Standard 2.0 and above.Minimize reliances to stop version disputes as well as the demand for binding redirects.Recording Audio Info.Some of the primary functions of the SDK is audio transcription. Developers may translate audio reports asynchronously or even in real-time. Below is actually an example of just how to transcribe an audio data:.utilizing AssemblyAI.using AssemblyAI.Transcripts.var customer = brand new AssemblyAIClient(" YOUR_API_KEY").var transcript = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For neighborhood reports, comparable code may be made use of to obtain transcription.await making use of var stream = brand new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK likewise holds real-time audio transcription using Streaming Speech-to-Text. This component is especially beneficial for applications needing quick handling of audio records.using AssemblyAI.Realtime.await utilizing var transcriber = new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Last: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for getting sound from a mic for example.GetAudio( async (portion) => wait for transcriber.SendAudioAsync( part)).await transcriber.CloseAsync().Utilizing LeMUR for LLM Apps.The SDK integrates with LeMUR to make it possible for creators to develop big foreign language style (LLM) functions on vocal information. Here is an example:.var lemurTaskParams = new LemurTaskParams.Trigger="Offer a short rundown of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intelligence Versions.Furthermore, the SDK includes built-in help for audio intellect versions, making it possible for sentiment evaluation and also other state-of-the-art features.var transcript = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For additional information, go to the official AssemblyAI blog.Image source: Shutterstock.