.Guarantee being compatible along with several frameworks, including.NET 6.0,. Web Framework 4.6.2, and.NET Requirement 2.0 as well as above.Reduce dependences to prevent variation conflicts and also the need for tiing redirects.Translating Sound Information.One of the main capabilities of the SDK is audio transcription. Developers may transcribe audio reports asynchronously or in real-time. Below is an example of just how to translate an audio data:.making use of AssemblyAI.making use of AssemblyAI.Transcripts.var customer = brand new AssemblyAIClient(" YOUR_API_KEY").var records = wait for client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For regional files, comparable code could be used to achieve transcription.await using var stream = new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK additionally holds real-time audio transcription using Streaming Speech-to-Text. This component is specifically valuable for applications demanding urgent handling of audio data.making use of AssemblyAI.Realtime.wait for utilizing var transcriber = brand-new RealtimeTranscriber( brand new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Final: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for receiving audio from a mic for example.GetAudio( async (portion) => await transcriber.SendAudioAsync( portion)).wait for transcriber.CloseAsync().Using LeMUR for LLM Applications.The SDK incorporates along with LeMUR to make it possible for designers to build large foreign language model (LLM) functions on voice data. Below is an instance:.var lemurTaskParams = brand-new LemurTaskParams.Trigger="Offer a brief recap of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Cleverness Designs.Also, the SDK features integrated help for audio cleverness versions, permitting conviction study and also various other sophisticated components.var records = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = accurate. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// POSITIVE, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, visit the formal AssemblyAI blog.Image source: Shutterstock.