mirror of
https://github.com/Swiftgram/Telegram-iOS.git
synced 2025-06-16 05:55:20 +00:00
Attempt to fix the voice chat tone volume
This commit is contained in:
parent
46d46a9b4b
commit
5dec00188e
@ -218,6 +218,7 @@ private final class AudioPlayerRendererContext {
|
|||||||
let lowWaterSizeInSeconds: Int = 2
|
let lowWaterSizeInSeconds: Int = 2
|
||||||
|
|
||||||
let audioSession: MediaPlayerAudioSessionControl
|
let audioSession: MediaPlayerAudioSessionControl
|
||||||
|
let useVoiceProcessingMode: Bool
|
||||||
let controlTimebase: CMTimebase
|
let controlTimebase: CMTimebase
|
||||||
let updatedRate: () -> Void
|
let updatedRate: () -> Void
|
||||||
let audioPaused: () -> Void
|
let audioPaused: () -> Void
|
||||||
@ -250,7 +251,7 @@ private final class AudioPlayerRendererContext {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
init(controlTimebase: CMTimebase, audioSession: MediaPlayerAudioSessionControl, playAndRecord: Bool, ambient: Bool, forceAudioToSpeaker: Bool, baseRate: Double, audioLevelPipe: ValuePipe<Float>, updatedRate: @escaping () -> Void, audioPaused: @escaping () -> Void) {
|
init(controlTimebase: CMTimebase, audioSession: MediaPlayerAudioSessionControl, playAndRecord: Bool, useVoiceProcessingMode: Bool, ambient: Bool, forceAudioToSpeaker: Bool, baseRate: Double, audioLevelPipe: ValuePipe<Float>, updatedRate: @escaping () -> Void, audioPaused: @escaping () -> Void) {
|
||||||
assert(audioPlayerRendererQueue.isCurrent())
|
assert(audioPlayerRendererQueue.isCurrent())
|
||||||
|
|
||||||
self.audioSession = audioSession
|
self.audioSession = audioSession
|
||||||
@ -263,6 +264,7 @@ private final class AudioPlayerRendererContext {
|
|||||||
self.audioPaused = audioPaused
|
self.audioPaused = audioPaused
|
||||||
|
|
||||||
self.playAndRecord = playAndRecord
|
self.playAndRecord = playAndRecord
|
||||||
|
self.useVoiceProcessingMode = useVoiceProcessingMode
|
||||||
self.ambient = ambient
|
self.ambient = ambient
|
||||||
|
|
||||||
self.audioStreamDescription = audioRendererNativeStreamDescription()
|
self.audioStreamDescription = audioRendererNativeStreamDescription()
|
||||||
@ -407,7 +409,11 @@ private final class AudioPlayerRendererContext {
|
|||||||
var outputNode: AUNode = 0
|
var outputNode: AUNode = 0
|
||||||
var outputDesc = AudioComponentDescription()
|
var outputDesc = AudioComponentDescription()
|
||||||
outputDesc.componentType = kAudioUnitType_Output
|
outputDesc.componentType = kAudioUnitType_Output
|
||||||
|
if self.useVoiceProcessingMode {
|
||||||
|
outputDesc.componentSubType = kAudioUnitSubType_VoiceProcessingIO
|
||||||
|
} else {
|
||||||
outputDesc.componentSubType = kAudioUnitSubType_RemoteIO
|
outputDesc.componentSubType = kAudioUnitSubType_RemoteIO
|
||||||
|
}
|
||||||
outputDesc.componentFlags = 0
|
outputDesc.componentFlags = 0
|
||||||
outputDesc.componentFlagsMask = 0
|
outputDesc.componentFlagsMask = 0
|
||||||
outputDesc.componentManufacturer = kAudioUnitManufacturer_Apple
|
outputDesc.componentManufacturer = kAudioUnitManufacturer_Apple
|
||||||
@ -753,7 +759,7 @@ public final class MediaPlayerAudioRenderer {
|
|||||||
private let audioClock: CMClock
|
private let audioClock: CMClock
|
||||||
public let audioTimebase: CMTimebase
|
public let audioTimebase: CMTimebase
|
||||||
|
|
||||||
public init(audioSession: MediaPlayerAudioSessionControl, playAndRecord: Bool, ambient: Bool, forceAudioToSpeaker: Bool, baseRate: Double, audioLevelPipe: ValuePipe<Float>, updatedRate: @escaping () -> Void, audioPaused: @escaping () -> Void) {
|
public init(audioSession: MediaPlayerAudioSessionControl, playAndRecord: Bool, useVoiceProcessingMode: Bool = false, ambient: Bool, forceAudioToSpeaker: Bool, baseRate: Double, audioLevelPipe: ValuePipe<Float>, updatedRate: @escaping () -> Void, audioPaused: @escaping () -> Void) {
|
||||||
var audioClock: CMClock?
|
var audioClock: CMClock?
|
||||||
CMAudioClockCreate(allocator: nil, clockOut: &audioClock)
|
CMAudioClockCreate(allocator: nil, clockOut: &audioClock)
|
||||||
if audioClock == nil {
|
if audioClock == nil {
|
||||||
@ -766,7 +772,7 @@ public final class MediaPlayerAudioRenderer {
|
|||||||
self.audioTimebase = audioTimebase!
|
self.audioTimebase = audioTimebase!
|
||||||
|
|
||||||
audioPlayerRendererQueue.async {
|
audioPlayerRendererQueue.async {
|
||||||
let context = AudioPlayerRendererContext(controlTimebase: audioTimebase!, audioSession: audioSession, playAndRecord: playAndRecord, ambient: ambient, forceAudioToSpeaker: forceAudioToSpeaker, baseRate: baseRate, audioLevelPipe: audioLevelPipe, updatedRate: updatedRate, audioPaused: audioPaused)
|
let context = AudioPlayerRendererContext(controlTimebase: audioTimebase!, audioSession: audioSession, playAndRecord: playAndRecord, useVoiceProcessingMode: useVoiceProcessingMode, ambient: ambient, forceAudioToSpeaker: forceAudioToSpeaker, baseRate: baseRate, audioLevelPipe: audioLevelPipe, updatedRate: updatedRate, audioPaused: audioPaused)
|
||||||
self.contextRef = Unmanaged.passRetained(context)
|
self.contextRef = Unmanaged.passRetained(context)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -34,7 +34,7 @@ final class PresentationCallToneRenderer {
|
|||||||
|
|
||||||
self.toneRenderer = MediaPlayerAudioRenderer(audioSession: .custom({ control in
|
self.toneRenderer = MediaPlayerAudioRenderer(audioSession: .custom({ control in
|
||||||
return controlImpl?(control) ?? EmptyDisposable
|
return controlImpl?(control) ?? EmptyDisposable
|
||||||
}), playAndRecord: false, ambient: false, forceAudioToSpeaker: false, baseRate: 1.0, audioLevelPipe: self.audioLevelPipe, updatedRate: {}, audioPaused: {})
|
}), playAndRecord: false, useVoiceProcessingMode: true, ambient: false, forceAudioToSpeaker: false, baseRate: 1.0, audioLevelPipe: self.audioLevelPipe, updatedRate: {}, audioPaused: {})
|
||||||
|
|
||||||
controlImpl = { [weak self] control in
|
controlImpl = { [weak self] control in
|
||||||
queue.async {
|
queue.async {
|
||||||
|
Loading…
x
Reference in New Issue
Block a user