mirror of
https://github.com/jetkvm/kvm.git
synced 2025-09-16 08:38:14 +00:00
Consolidate audio frame validation functions into a single optimized implementation and add dynamic OPUS encoder parameter updates based on quality settings. Initialize validation cache at startup for consistent performance. Add latency profiler for end-to-end audio pipeline monitoring. Update test cases to use unified validation function and initialize cache. The changes improve performance by reducing function call overhead and enabling runtime optimization of audio encoding parameters based on quality presets.
232 lines
5.7 KiB
Go
232 lines
5.7 KiB
Go
package audio
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"reflect"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/jetkvm/kvm/internal/logging"
|
|
"github.com/pion/webrtc/v4/pkg/media"
|
|
"github.com/rs/zerolog"
|
|
)
|
|
|
|
// AudioRelay handles forwarding audio frames from the audio server subprocess
|
|
// to WebRTC without any CGO audio processing. This runs in the main process.
|
|
type AudioRelay struct {
|
|
// Atomic fields MUST be first for ARM32 alignment (int64 fields need 8-byte alignment)
|
|
framesRelayed int64
|
|
framesDropped int64
|
|
|
|
client *AudioOutputClient
|
|
ctx context.Context
|
|
cancel context.CancelFunc
|
|
wg sync.WaitGroup
|
|
logger *zerolog.Logger
|
|
running bool
|
|
mutex sync.RWMutex
|
|
bufferPool *AudioBufferPool // Buffer pool for memory optimization
|
|
|
|
// WebRTC integration
|
|
audioTrack AudioTrackWriter
|
|
config AudioConfig
|
|
muted bool
|
|
}
|
|
|
|
// AudioTrackWriter interface for WebRTC audio track
|
|
type AudioTrackWriter interface {
|
|
WriteSample(sample media.Sample) error
|
|
}
|
|
|
|
// NewAudioRelay creates a new audio relay for the main process
|
|
func NewAudioRelay() *AudioRelay {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
logger := logging.GetDefaultLogger().With().Str("component", "audio-relay").Logger()
|
|
|
|
return &AudioRelay{
|
|
ctx: ctx,
|
|
cancel: cancel,
|
|
logger: &logger,
|
|
bufferPool: NewAudioBufferPool(GetMaxAudioFrameSize()),
|
|
}
|
|
}
|
|
|
|
// Start begins the audio relay process
|
|
func (r *AudioRelay) Start(audioTrack AudioTrackWriter, config AudioConfig) error {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
|
|
if r.running {
|
|
return nil // Already running
|
|
}
|
|
|
|
// Create audio client to connect to subprocess
|
|
client := NewAudioOutputClient()
|
|
r.client = client
|
|
r.audioTrack = audioTrack
|
|
r.config = config
|
|
|
|
// Connect to the audio output server
|
|
if err := client.Connect(); err != nil {
|
|
return fmt.Errorf("failed to connect to audio output server: %w", err)
|
|
}
|
|
|
|
// Start relay goroutine
|
|
r.wg.Add(1)
|
|
go r.relayLoop()
|
|
|
|
r.running = true
|
|
r.logger.Info().Msg("Audio relay connected to output server")
|
|
return nil
|
|
}
|
|
|
|
// Stop stops the audio relay
|
|
func (r *AudioRelay) Stop() {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
|
|
if !r.running {
|
|
return
|
|
}
|
|
|
|
r.cancel()
|
|
r.wg.Wait()
|
|
|
|
if r.client != nil {
|
|
r.client.Disconnect()
|
|
r.client = nil
|
|
}
|
|
|
|
r.running = false
|
|
r.logger.Info().Msgf("Audio relay stopped after relaying %d frames", r.framesRelayed)
|
|
}
|
|
|
|
// SetMuted sets the mute state
|
|
func (r *AudioRelay) SetMuted(muted bool) {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
r.muted = muted
|
|
}
|
|
|
|
// IsMuted returns the current mute state (checks both relay and global mute)
|
|
func (r *AudioRelay) IsMuted() bool {
|
|
r.mutex.RLock()
|
|
defer r.mutex.RUnlock()
|
|
return r.muted || IsAudioMuted()
|
|
}
|
|
|
|
// GetStats returns relay statistics
|
|
func (r *AudioRelay) GetStats() (framesRelayed, framesDropped int64) {
|
|
r.mutex.RLock()
|
|
defer r.mutex.RUnlock()
|
|
return r.framesRelayed, r.framesDropped
|
|
}
|
|
|
|
// UpdateTrack updates the WebRTC audio track for the relay
|
|
func (r *AudioRelay) UpdateTrack(audioTrack AudioTrackWriter) {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
r.audioTrack = audioTrack
|
|
}
|
|
|
|
func (r *AudioRelay) relayLoop() {
|
|
defer r.wg.Done()
|
|
r.logger.Debug().Msg("Audio relay loop started")
|
|
|
|
var maxConsecutiveErrors = GetConfig().MaxConsecutiveErrors
|
|
consecutiveErrors := 0
|
|
|
|
for {
|
|
select {
|
|
case <-r.ctx.Done():
|
|
r.logger.Debug().Msg("audio relay loop stopping")
|
|
return
|
|
default:
|
|
frame, err := r.client.ReceiveFrame()
|
|
if err != nil {
|
|
consecutiveErrors++
|
|
r.logger.Error().Err(err).Int("consecutive_errors", consecutiveErrors).Msg("error reading frame from audio output server")
|
|
r.incrementDropped()
|
|
|
|
if consecutiveErrors >= maxConsecutiveErrors {
|
|
r.logger.Error().Int("consecutive_errors", consecutiveErrors).Int("max_errors", maxConsecutiveErrors).Msg("too many consecutive read errors, stopping audio relay")
|
|
return
|
|
}
|
|
time.Sleep(GetConfig().ShortSleepDuration)
|
|
continue
|
|
}
|
|
|
|
consecutiveErrors = 0
|
|
if err := r.forwardToWebRTC(frame); err != nil {
|
|
r.logger.Warn().Err(err).Msg("failed to forward frame to webrtc")
|
|
r.incrementDropped()
|
|
} else {
|
|
r.incrementRelayed()
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// forwardToWebRTC forwards a frame to the WebRTC audio track
|
|
func (r *AudioRelay) forwardToWebRTC(frame []byte) error {
|
|
// Use ultra-fast validation for critical audio path
|
|
if err := ValidateAudioFrame(frame); err != nil {
|
|
r.incrementDropped()
|
|
r.logger.Debug().Err(err).Msg("invalid frame data in relay")
|
|
return err
|
|
}
|
|
|
|
r.mutex.RLock()
|
|
defer r.mutex.RUnlock()
|
|
|
|
audioTrack := r.audioTrack
|
|
config := r.config
|
|
muted := r.muted
|
|
|
|
// Comprehensive nil check for audioTrack to prevent panic
|
|
if audioTrack == nil {
|
|
return nil // No audio track available
|
|
}
|
|
|
|
// Check if interface contains nil pointer using reflection
|
|
if reflect.ValueOf(audioTrack).IsNil() {
|
|
return nil // Audio track interface contains nil pointer
|
|
}
|
|
|
|
// Prepare sample data
|
|
var sampleData []byte
|
|
if muted {
|
|
// Send silence when muted - use buffer pool to avoid allocation
|
|
sampleData = r.bufferPool.Get()
|
|
sampleData = sampleData[:len(frame)] // Resize to frame length
|
|
// Clear the buffer to create silence
|
|
for i := range sampleData {
|
|
sampleData[i] = 0
|
|
}
|
|
defer r.bufferPool.Put(sampleData) // Return to pool after use
|
|
} else {
|
|
sampleData = frame
|
|
}
|
|
|
|
// Write sample to WebRTC track while holding the read lock
|
|
return audioTrack.WriteSample(media.Sample{
|
|
Data: sampleData,
|
|
Duration: config.FrameSize,
|
|
})
|
|
}
|
|
|
|
// incrementRelayed atomically increments the relayed frames counter
|
|
func (r *AudioRelay) incrementRelayed() {
|
|
r.mutex.Lock()
|
|
r.framesRelayed++
|
|
r.mutex.Unlock()
|
|
}
|
|
|
|
// incrementDropped atomically increments the dropped frames counter
|
|
func (r *AudioRelay) incrementDropped() {
|
|
r.mutex.Lock()
|
|
r.framesDropped++
|
|
r.mutex.Unlock()
|
|
}
|