.Make sure being compatible with several platforms, including.NET 6.0,. Internet Platform 4.6.2, and.NET Standard 2.0 and above.Minimize dependences to stop version problems and also the necessity for binding redirects.Transcribing Sound Record.One of the major functions of the SDK is audio transcription. Developers can translate audio reports asynchronously or even in real-time. Below is actually an example of how to translate an audio report:.utilizing AssemblyAI.making use of AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For neighborhood reports, comparable code could be utilized to attain transcription.await utilizing var stream = brand-new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.flow,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK also reinforces real-time sound transcription making use of Streaming Speech-to-Text. This feature is especially useful for treatments demanding urgent handling of audio information.making use of AssemblyAI.Realtime.await using var transcriber = new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Final: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for obtaining audio from a microphone as an example.GetAudio( async (chunk) => await transcriber.SendAudioAsync( portion)).await transcriber.CloseAsync().Making Use Of LeMUR for LLM Apps.The SDK combines with LeMUR to allow developers to develop big language version (LLM) functions on voice data. Right here is actually an example:.var lemurTaskParams = new LemurTaskParams.Trigger="Give a short recap of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intellect Designs.Furthermore, the SDK features integrated support for audio intellect models, allowing belief evaluation as well as various other innovative features.var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// POSITIVE, NEUTRAL, or even downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To find out more, see the main AssemblyAI blog.Image resource: Shutterstock.