.Guarantee compatibility along with numerous frameworks, including.NET 6.0,. Internet Structure 4.6.2, and.NET Criterion 2.0 and above.Decrease reliances to stop variation conflicts as well as the necessity for tiing redirects.Recording Sound Information.Among the main functions of the SDK is audio transcription. Designers may transcribe audio documents asynchronously or even in real-time. Below is an example of how to record an audio documents:.making use of AssemblyAI.making use of AssemblyAI.Transcripts.var client = new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local area reports, similar code could be used to accomplish transcription.await using var flow = brand-new FileStream("./ nbc.mp3", FileMode.Open).var transcript = wait for client.Transcripts.TranscribeAsync(.flow,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK additionally supports real-time audio transcription utilizing Streaming Speech-to-Text. This function is especially practical for treatments calling for urgent processing of audio information.using AssemblyAI.Realtime.await using var scribe = brand-new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Last: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for obtaining sound coming from a mic as an example.GetAudio( async (piece) => wait for transcriber.SendAudioAsync( chunk)).wait for transcriber.CloseAsync().Utilizing LeMUR for LLM Applications.The SDK includes along with LeMUR to allow creators to build large language design (LLM) apps on voice records. Listed below is actually an instance:.var lemurTaskParams = new LemurTaskParams.Trigger="Deliver a quick summary of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var feedback = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intellect Styles.Also, the SDK features integrated help for audio knowledge styles, allowing conviction analysis as well as other innovative components.var transcript = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = correct. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For additional information, see the main AssemblyAI blog.Image source: Shutterstock.