mirror of
https://github.com/jetkvm/kvm.git
synced 2025-09-16 08:38:14 +00:00
This change replaces all instances of GetConfig() function calls with direct access to the Config variable throughout the audio package. The modification improves performance by eliminating function call overhead and simplifies the codebase by removing unnecessary indirection. The commit also includes minor optimizations in validation logic and connection handling, while maintaining all existing functionality. Error handling remains robust with appropriate fallbacks when config values are not available. Additional improvements include: - Enhanced connection health monitoring in UnifiedAudioClient - Optimized validation functions using cached config values - Reduced memory allocations in hot paths - Improved error recovery during quality changes
232 lines
5.7 KiB
Go
232 lines
5.7 KiB
Go
package audio
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"reflect"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/jetkvm/kvm/internal/logging"
|
|
"github.com/pion/webrtc/v4/pkg/media"
|
|
"github.com/rs/zerolog"
|
|
)
|
|
|
|
// AudioRelay handles forwarding audio frames from the audio server subprocess
|
|
// to WebRTC without any CGO audio processing. This runs in the main process.
|
|
type AudioRelay struct {
|
|
// Atomic fields MUST be first for ARM32 alignment (int64 fields need 8-byte alignment)
|
|
framesRelayed int64
|
|
framesDropped int64
|
|
|
|
client *AudioOutputClient
|
|
ctx context.Context
|
|
cancel context.CancelFunc
|
|
wg sync.WaitGroup
|
|
logger *zerolog.Logger
|
|
running bool
|
|
mutex sync.RWMutex
|
|
bufferPool *AudioBufferPool // Buffer pool for memory optimization
|
|
|
|
// WebRTC integration
|
|
audioTrack AudioTrackWriter
|
|
config AudioConfig
|
|
muted bool
|
|
}
|
|
|
|
// AudioTrackWriter interface for WebRTC audio track
|
|
type AudioTrackWriter interface {
|
|
WriteSample(sample media.Sample) error
|
|
}
|
|
|
|
// NewAudioRelay creates a new audio relay for the main process
|
|
func NewAudioRelay() *AudioRelay {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
logger := logging.GetDefaultLogger().With().Str("component", "audio-relay").Logger()
|
|
|
|
return &AudioRelay{
|
|
ctx: ctx,
|
|
cancel: cancel,
|
|
logger: &logger,
|
|
bufferPool: NewAudioBufferPool(GetMaxAudioFrameSize()),
|
|
}
|
|
}
|
|
|
|
// Start begins the audio relay process
|
|
func (r *AudioRelay) Start(audioTrack AudioTrackWriter, config AudioConfig) error {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
|
|
if r.running {
|
|
return nil // Already running
|
|
}
|
|
|
|
// Create audio client to connect to subprocess
|
|
client := NewAudioOutputClient()
|
|
r.client = client
|
|
r.audioTrack = audioTrack
|
|
r.config = config
|
|
|
|
// Connect to the audio output server
|
|
if err := client.Connect(); err != nil {
|
|
return fmt.Errorf("failed to connect to audio output server: %w", err)
|
|
}
|
|
|
|
// Start relay goroutine
|
|
r.wg.Add(1)
|
|
go r.relayLoop()
|
|
|
|
r.running = true
|
|
r.logger.Info().Msg("Audio relay connected to output server")
|
|
return nil
|
|
}
|
|
|
|
// Stop stops the audio relay
|
|
func (r *AudioRelay) Stop() {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
|
|
if !r.running {
|
|
return
|
|
}
|
|
|
|
r.cancel()
|
|
r.wg.Wait()
|
|
|
|
if r.client != nil {
|
|
r.client.Disconnect()
|
|
r.client = nil
|
|
}
|
|
|
|
r.running = false
|
|
r.logger.Info().Msgf("Audio relay stopped after relaying %d frames", r.framesRelayed)
|
|
}
|
|
|
|
// SetMuted sets the mute state
|
|
func (r *AudioRelay) SetMuted(muted bool) {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
r.muted = muted
|
|
}
|
|
|
|
// IsMuted returns the current mute state (checks both relay and global mute)
|
|
func (r *AudioRelay) IsMuted() bool {
|
|
r.mutex.RLock()
|
|
defer r.mutex.RUnlock()
|
|
return r.muted || IsAudioMuted()
|
|
}
|
|
|
|
// GetStats returns relay statistics
|
|
func (r *AudioRelay) GetStats() (framesRelayed, framesDropped int64) {
|
|
r.mutex.RLock()
|
|
defer r.mutex.RUnlock()
|
|
return r.framesRelayed, r.framesDropped
|
|
}
|
|
|
|
// UpdateTrack updates the WebRTC audio track for the relay
|
|
func (r *AudioRelay) UpdateTrack(audioTrack AudioTrackWriter) {
|
|
r.mutex.Lock()
|
|
defer r.mutex.Unlock()
|
|
r.audioTrack = audioTrack
|
|
}
|
|
|
|
func (r *AudioRelay) relayLoop() {
|
|
defer r.wg.Done()
|
|
r.logger.Debug().Msg("Audio relay loop started")
|
|
|
|
var maxConsecutiveErrors = Config.MaxConsecutiveErrors
|
|
consecutiveErrors := 0
|
|
|
|
for {
|
|
select {
|
|
case <-r.ctx.Done():
|
|
r.logger.Debug().Msg("audio relay loop stopping")
|
|
return
|
|
default:
|
|
frame, err := r.client.ReceiveFrame()
|
|
if err != nil {
|
|
consecutiveErrors++
|
|
r.logger.Error().Err(err).Int("consecutive_errors", consecutiveErrors).Msg("error reading frame from audio output server")
|
|
r.incrementDropped()
|
|
|
|
if consecutiveErrors >= maxConsecutiveErrors {
|
|
r.logger.Error().Int("consecutive_errors", consecutiveErrors).Int("max_errors", maxConsecutiveErrors).Msg("too many consecutive read errors, stopping audio relay")
|
|
return
|
|
}
|
|
time.Sleep(Config.ShortSleepDuration)
|
|
continue
|
|
}
|
|
|
|
consecutiveErrors = 0
|
|
if err := r.forwardToWebRTC(frame); err != nil {
|
|
r.logger.Warn().Err(err).Msg("failed to forward frame to webrtc")
|
|
r.incrementDropped()
|
|
} else {
|
|
r.incrementRelayed()
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// forwardToWebRTC forwards a frame to the WebRTC audio track
|
|
func (r *AudioRelay) forwardToWebRTC(frame []byte) error {
|
|
// Use ultra-fast validation for critical audio path
|
|
if err := ValidateAudioFrame(frame); err != nil {
|
|
r.incrementDropped()
|
|
r.logger.Debug().Err(err).Msg("invalid frame data in relay")
|
|
return err
|
|
}
|
|
|
|
r.mutex.RLock()
|
|
defer r.mutex.RUnlock()
|
|
|
|
audioTrack := r.audioTrack
|
|
config := r.config
|
|
muted := r.muted
|
|
|
|
// Comprehensive nil check for audioTrack to prevent panic
|
|
if audioTrack == nil {
|
|
return nil // No audio track available
|
|
}
|
|
|
|
// Check if interface contains nil pointer using reflection
|
|
if reflect.ValueOf(audioTrack).IsNil() {
|
|
return nil // Audio track interface contains nil pointer
|
|
}
|
|
|
|
// Prepare sample data
|
|
var sampleData []byte
|
|
if muted {
|
|
// Send silence when muted - use buffer pool to avoid allocation
|
|
sampleData = r.bufferPool.Get()
|
|
sampleData = sampleData[:len(frame)] // Resize to frame length
|
|
// Clear the buffer to create silence
|
|
for i := range sampleData {
|
|
sampleData[i] = 0
|
|
}
|
|
defer r.bufferPool.Put(sampleData) // Return to pool after use
|
|
} else {
|
|
sampleData = frame
|
|
}
|
|
|
|
// Write sample to WebRTC track while holding the read lock
|
|
return audioTrack.WriteSample(media.Sample{
|
|
Data: sampleData,
|
|
Duration: config.FrameSize,
|
|
})
|
|
}
|
|
|
|
// incrementRelayed atomically increments the relayed frames counter
|
|
func (r *AudioRelay) incrementRelayed() {
|
|
r.mutex.Lock()
|
|
r.framesRelayed++
|
|
r.mutex.Unlock()
|
|
}
|
|
|
|
// incrementDropped atomically increments the dropped frames counter
|
|
func (r *AudioRelay) incrementDropped() {
|
|
r.mutex.Lock()
|
|
r.framesDropped++
|
|
r.mutex.Unlock()
|
|
}
|