.Make sure compatibility along with numerous structures, including.NET 6.0,. Internet Structure 4.6.2, and.NET Specification 2.0 and also above.Reduce dependencies to stop model problems as well as the requirement for binding redirects.Recording Sound Files.Among the main capabilities of the SDK is actually audio transcription. Designers can record audio reports asynchronously or in real-time. Below is an example of exactly how to translate an audio data:.using AssemblyAI.using AssemblyAI.Transcripts.var client = new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For nearby reports, comparable code may be made use of to obtain transcription.wait for using var flow = brand new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.flow,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK additionally supports real-time sound transcription using Streaming Speech-to-Text. This function is specifically beneficial for uses requiring immediate processing of audio records.making use of AssemblyAI.Realtime.wait for using var scribe = brand-new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Ultimate: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for obtaining audio from a microphone for example.GetAudio( async (portion) => wait for transcriber.SendAudioAsync( part)).await transcriber.CloseAsync().Using LeMUR for LLM Apps.The SDK integrates along with LeMUR to make it possible for programmers to create big language design (LLM) applications on vocal records. Listed here is an example:.var lemurTaskParams = new LemurTaskParams.Urge="Provide a short rundown of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intellect Models.Additionally, the SDK features integrated support for audio intellect styles, allowing view study and various other sophisticated components.var transcript = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more details, explore the official AssemblyAI blog.Image source: Shutterstock.