.Make sure being compatible along with numerous structures, including.NET 6.0,. Web Platform 4.6.2, and.NET Specification 2.0 and above.Minimize reliances to prevent version problems and the requirement for binding redirects.Recording Sound Info.Among the primary capabilities of the SDK is actually audio transcription. Developers can easily translate audio data asynchronously or even in real-time. Below is actually an example of how to transcribe an audio data:.making use of AssemblyAI.making use of AssemblyAI.Transcripts.var customer = brand-new AssemblyAIClient(" YOUR_API_KEY").var records = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For neighborhood reports, identical code could be made use of to accomplish transcription.wait for making use of var flow = new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.stream,.new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK also holds real-time sound transcription using Streaming Speech-to-Text. This component is specifically practical for treatments needing instant processing of audio information.using AssemblyAI.Realtime.await making use of var scribe = brand new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Last: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for acquiring sound from a mic for example.GetAudio( async (piece) => wait for transcriber.SendAudioAsync( chunk)).wait for transcriber.CloseAsync().Using LeMUR for LLM Functions.The SDK incorporates with LeMUR to make it possible for programmers to construct big foreign language design (LLM) applications on voice data. Listed below is an example:.var lemurTaskParams = brand new LemurTaskParams.Trigger="Give a short recap of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intellect Styles.Additionally, the SDK includes integrated assistance for audio intelligence styles, making it possible for belief review and various other innovative attributes.var transcript = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = accurate. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, explore the formal AssemblyAI blog.Image source: Shutterstock.