diff --git a/submodules/MediaPlayer/Sources/ChunkMediaPlayerV2.swift b/submodules/MediaPlayer/Sources/ChunkMediaPlayerV2.swift index 2026c7b8c7..9825eb7b4c 100644 --- a/submodules/MediaPlayer/Sources/ChunkMediaPlayerV2.swift +++ b/submodules/MediaPlayer/Sources/ChunkMediaPlayerV2.swift @@ -33,6 +33,101 @@ private final class ChunkMediaPlayerExternalSourceImpl: ChunkMediaPlayerSourceIm } public final class ChunkMediaPlayerV2: ChunkMediaPlayer { + public final class AudioContext { + fileprivate let audioSessionManager: ManagedAudioSession + private var audioSessionDisposable: Disposable? + private(set) var hasAudioSession: Bool = false + private(set) var isAmbientMode: Bool = false + private(set) var isInitialized: Bool = false + + private var updatedListeners = Bag<() -> Void>() + + public init( + audioSessionManager: ManagedAudioSession + ) { + self.audioSessionManager = audioSessionManager + } + + deinit { + self.audioSessionDisposable?.dispose() + } + + func onUpdated(_ f: @escaping () -> Void) -> Disposable { + let index = self.updatedListeners.add(f) + return ActionDisposable { [weak self] in + Queue.mainQueue().async { + guard let self else { + return + } + self.updatedListeners.remove(index) + } + } + } + + func setIsAmbient(isAmbient: Bool) { + self.hasAudioSession = false + + for f in self.updatedListeners.copyItems() { + f() + } + + self.audioSessionDisposable?.dispose() + self.audioSessionDisposable = nil + } + + func update(type: ManagedAudioSessionType?) { + if let type { + if self.audioSessionDisposable == nil { + self.isInitialized = true + + self.audioSessionDisposable = self.audioSessionManager.push(params: ManagedAudioSessionClientParams( + audioSessionType: type, + activateImmediately: false, + manualActivate: { [weak self] control in + control.setupAndActivate(synchronous: false, { state in + Queue.mainQueue().async { + guard let self else { + return + } + self.hasAudioSession = true + for f in self.updatedListeners.copyItems() { + f() + } + } + }) + }, + deactivate: { [weak self] _ in + return Signal { subscriber in + guard let self else { + subscriber.putCompletion() + return EmptyDisposable + } + + self.hasAudioSession = false + for f in self.updatedListeners.copyItems() { + f() + } + subscriber.putCompletion() + + return EmptyDisposable + } + |> runOn(.mainQueue()) + }, + headsetConnectionStatusChanged: { _ in }, + availableOutputsChanged: { _, _ in } + )) + } + } else { + if let audioSessionDisposable = self.audioSessionDisposable { + self.audioSessionDisposable = nil + audioSessionDisposable.dispose() + } + + self.hasAudioSession = false + } + } + } + public enum SourceDescription { public final class ResourceDescription { public let postbox: Postbox @@ -166,10 +261,10 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { private let dataQueue: Queue private let mediaDataReaderParams: MediaDataReaderParams - private let audioSessionManager: ManagedAudioSession private let onSeeked: (() -> Void)? private weak var playerNode: MediaPlayerNode? + private let audioContext: AudioContext private let renderSynchronizer: AVSampleBufferRenderSynchronizer private var videoRenderer: AVSampleBufferDisplayLayer private var audioRenderer: AVSampleBufferAudioRenderer? @@ -198,13 +293,20 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { } public var actionAtEnd: MediaPlayerActionAtEnd = .stop + public weak var migrateToNextPlayerOnEnd: ChunkMediaPlayerV2? { + didSet { + if self.migrateToNextPlayerOnEnd !== oldValue { + self.updateInternalState() + } + } + } private var didSeekOnce: Bool = false private var isPlaying: Bool = false private var baseRate: Double = 1.0 private var isSoundEnabled: Bool private var isMuted: Bool - private var isAmbientMode: Bool + private var initialIsAmbient: Bool private var seekId: Int = 0 private var seekTimestamp: Double = 0.0 @@ -223,12 +325,11 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { private var partsStateDisposable: Disposable? private var updateTimer: Foundation.Timer? - private var audioSessionDisposable: Disposable? - private var hasAudioSession: Bool = false + private var audioContextUpdatedDisposable: Disposable? public init( params: MediaDataReaderParams, - audioSessionManager: ManagedAudioSession, + audioContext: AudioContext, source: SourceDescription, video: Bool, playAutomatically: Bool = false, @@ -247,7 +348,7 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { self.dataQueue = ChunkMediaPlayerV2.sharedDataQueue self.mediaDataReaderParams = params - self.audioSessionManager = audioSessionManager + self.audioContext = audioContext self.onSeeked = onSeeked self.playerNode = playerNode @@ -257,7 +358,7 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { self.isSoundEnabled = enableSound self.isMuted = soundMuted - self.isAmbientMode = ambient + self.initialIsAmbient = ambient self.baseRate = baseRate self.renderSynchronizer = AVSampleBufferRenderSynchronizer() @@ -296,12 +397,19 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { } else { self.renderSynchronizer.addRenderer(self.videoRenderer) } + + self.audioContextUpdatedDisposable = self.audioContext.onUpdated({ [weak self] in + guard let self else { + return + } + self.updateInternalState() + }) } deinit { self.partsStateDisposable?.dispose() self.updateTimer?.invalidate() - self.audioSessionDisposable?.dispose() + self.audioContextUpdatedDisposable?.dispose() if #available(iOS 17.0, *) { self.videoRenderer.sampleBufferRenderer.stopRequestingMediaData() @@ -321,51 +429,19 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { } private func updateInternalState() { + var audioSessionType: ManagedAudioSessionType? if self.isSoundEnabled && self.hasSound { - if self.audioSessionDisposable == nil { - self.audioSessionDisposable = self.audioSessionManager.push(params: ManagedAudioSessionClientParams( - audioSessionType: self.isAmbientMode ? .ambient : .play(mixWithOthers: false), - activateImmediately: false, - manualActivate: { [weak self] control in - control.setupAndActivate(synchronous: false, { state in - Queue.mainQueue().async { - guard let self else { - return - } - self.hasAudioSession = true - self.updateInternalState() - } - }) - }, - deactivate: { [weak self] _ in - return Signal { subscriber in - guard let self else { - subscriber.putCompletion() - return EmptyDisposable - } - - self.hasAudioSession = false - self.updateInternalState() - subscriber.putCompletion() - - return EmptyDisposable - } - |> runOn(.mainQueue()) - }, - headsetConnectionStatusChanged: { _ in }, - availableOutputsChanged: { _, _ in } - )) + let isAmbient: Bool + if self.audioContext.isInitialized { + isAmbient = self.audioContext.isAmbientMode + } else { + isAmbient = self.initialIsAmbient } - } else { - if let audioSessionDisposable = self.audioSessionDisposable { - self.audioSessionDisposable = nil - audioSessionDisposable.dispose() - } - - self.hasAudioSession = false + audioSessionType = isAmbient ? .ambient : .play(mixWithOthers: false) } + self.audioContext.update(type: audioSessionType) - if self.isSoundEnabled && self.hasSound && self.hasAudioSession { + if self.isSoundEnabled && self.hasSound && self.audioContext.hasAudioSession { if self.audioRenderer == nil { let audioRenderer = AVSampleBufferAudioRenderer() audioRenderer.isMuted = self.isMuted @@ -799,13 +875,9 @@ public final class ChunkMediaPlayerV2: ChunkMediaPlayer { } public func continueWithOverridingAmbientMode(isAmbient: Bool) { - if self.isAmbientMode != isAmbient { - self.isAmbientMode = isAmbient - - self.hasAudioSession = false - self.updateInternalState() - self.audioSessionDisposable?.dispose() - self.audioSessionDisposable = nil + if self.audioContext.isAmbientMode != isAmbient { + self.initialIsAmbient = isAmbient + self.audioContext.setIsAmbient(isAmbient: isAmbient) let currentTimestamp: CMTime if let pendingSeekTimestamp = self.pendingSeekTimestamp { diff --git a/submodules/TelegramCallsUI/Sources/Components/LivestreamVideoViewV1.swift b/submodules/TelegramCallsUI/Sources/Components/LivestreamVideoViewV1.swift index e9d61c78a6..030ecff2b5 100644 --- a/submodules/TelegramCallsUI/Sources/Components/LivestreamVideoViewV1.swift +++ b/submodules/TelegramCallsUI/Sources/Components/LivestreamVideoViewV1.swift @@ -70,7 +70,7 @@ final class LivestreamVideoViewV1: UIView { var onSeeked: (() -> Void)? self.player = ChunkMediaPlayerV2( params: ChunkMediaPlayerV2.MediaDataReaderParams(context: context), - audioSessionManager: audioSessionManager, + audioContext: ChunkMediaPlayerV2.AudioContext(audioSessionManager: audioSessionManager), source: .externalParts(self.chunkPlayerPartsState.get()), video: true, enableSound: true, diff --git a/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryContent.swift b/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryContent.swift index b4e4c11784..725a27a253 100644 --- a/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryContent.swift +++ b/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryContent.swift @@ -6,6 +6,7 @@ import SwiftSignalKit import TelegramCore import Postbox import TelegramPresentationData +import UniversalMediaPlayer public final class StoryContentItem: Equatable { public final class ExternalState { @@ -32,6 +33,7 @@ public final class StoryContentItem: Equatable { public final class SharedState { public var replyDrafts: [StoryId: NSAttributedString] = [:] public var baseRate: Double = 1.0 + public var audioContext: ChunkMediaPlayerV2.AudioContext? public init() { } diff --git a/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemContentComponent.swift b/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemContentComponent.swift index f7adeea7cd..e50d333e51 100644 --- a/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemContentComponent.swift +++ b/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemContentComponent.swift @@ -15,6 +15,275 @@ import ButtonComponent import MultilineTextComponent import TelegramPresentationData +private protocol StoryVideoView: UIView { + var audioMode: StoryContentItem.AudioMode { get set } + var playbackCompleted: (() -> Void)? { get set } + var status: Signal { get } + + func play() + func pause() + func seek(timestamp: Double) + func setSoundMuted(soundMuted: Bool) + func continueWithOverridingAmbientMode(isAmbient: Bool) + func setBaseRate(baseRate: Double) + func update(size: CGSize, transition: ComponentTransition) +} + +private final class LegacyStoryVideoView: UIView, StoryVideoView { + private let videoNode: UniversalVideoNode + + var audioMode: StoryContentItem.AudioMode + var playbackCompleted: (() -> Void)? + + var status: Signal { + return self.videoNode.status + } + + init( + context: AccountContext, + file: FileMediaReference, + audioMode: StoryContentItem.AudioMode, + baseRate: Double, + isCaptureProtected: Bool + ) { + self.audioMode = audioMode + + var userLocation: MediaResourceUserLocation = .other + switch file { + case let .story(peer, _, _): + userLocation = .peer(peer.id) + default: + break + } + var hasSentFramesToDisplay: (() -> Void)? + self.videoNode = UniversalVideoNode( + context: context, + postbox: context.account.postbox, + audioSession: context.sharedContext.mediaManager.audioSession, + manager: context.sharedContext.mediaManager.universalVideoManager, + decoration: StoryVideoDecoration(), + content: NativeVideoContent( + id: .contextResult(0, "\(UInt64.random(in: 0 ... UInt64.max))"), + userLocation: userLocation, + fileReference: file, + imageReference: nil, + streamVideo: .story, + loopVideo: true, + enableSound: true, + soundMuted: audioMode == .off, + beginWithAmbientSound: audioMode == .ambient, + mixWithOthers: true, + useLargeThumbnail: false, + autoFetchFullSizeThumbnail: false, + tempFilePath: nil, + captureProtected: isCaptureProtected, + hintDimensions: file.media.dimensions?.cgSize, + storeAfterDownload: nil, + displayImage: false, + hasSentFramesToDisplay: { + hasSentFramesToDisplay?() + } + ), + priority: .gallery + ) + self.videoNode.isHidden = true + self.videoNode.setBaseRate(baseRate) + + super.init(frame: CGRect()) + + hasSentFramesToDisplay = { [weak self] in + guard let self else { + return + } + self.videoNode.isHidden = false + } + + self.videoNode.playbackCompleted = { [weak self] in + guard let self else { + return + } + self.playbackCompleted?() + } + + self.addSubview(self.videoNode.view) + + self.videoNode.ownsContentNodeUpdated = { [weak self] value in + guard let self else { + return + } + if value { + self.videoNode.seek(0.0) + if self.audioMode != .off { + self.videoNode.playOnceWithSound(playAndRecord: false, actionAtEnd: .stop) + } else { + self.videoNode.play() + } + } + } + self.videoNode.canAttachContent = true + } + + required init?(coder: NSCoder) { + fatalError("init(coder:) has not been implemented") + } + + func play() { + self.videoNode.play() + } + + func pause() { + self.videoNode.pause() + } + + func seek(timestamp: Double) { + self.videoNode.seek(timestamp) + } + + func setSoundMuted(soundMuted: Bool) { + self.videoNode.setSoundMuted(soundMuted: soundMuted) + } + + func continueWithOverridingAmbientMode(isAmbient: Bool) { + self.videoNode.continueWithOverridingAmbientMode(isAmbient: isAmbient) + } + + func setBaseRate(baseRate: Double) { + self.videoNode.setBaseRate(baseRate) + } + + func update(size: CGSize, transition: ComponentTransition) { + transition.setFrame(view: self.videoNode.view, frame: CGRect(origin: CGPoint(), size: size)) + self.videoNode.updateLayout(size: size, transition: transition.containedViewLayoutTransition) + } +} + +private final class ModernStoryVideoView: UIView, StoryVideoView { + private let player: ChunkMediaPlayerV2 + private let playerNode: MediaPlayerNode + + var audioMode: StoryContentItem.AudioMode + var playbackCompleted: (() -> Void)? + var isFirstPlay: Bool = true + + var status: Signal { + return self.player.status |> map(Optional.init) + } + + init( + context: AccountContext, + audioContext: ChunkMediaPlayerV2.AudioContext, + file: FileMediaReference, + audioMode: StoryContentItem.AudioMode, + baseRate: Double, + isCaptureProtected: Bool + ) { + self.audioMode = audioMode + + self.playerNode = MediaPlayerNode( + backgroundThread: false, + captureProtected: isCaptureProtected + ) + + var userLocation: MediaResourceUserLocation = .other + switch file { + case let .story(peer, _, _): + userLocation = .peer(peer.id) + default: + break + } + + self.player = ChunkMediaPlayerV2( + params: ChunkMediaPlayerV2.MediaDataReaderParams(context: context), + audioContext: audioContext, + source: .directFetch(ChunkMediaPlayerV2.SourceDescription.ResourceDescription( + postbox: context.account.postbox, + size: file.media.size ?? 0, + reference: file.resourceReference(file.media.resource), + userLocation: userLocation, + userContentType: .story, + statsCategory: statsCategoryForFileWithAttributes(file.media.attributes), + fetchAutomatically: false + )), + video: true, + playAutomatically: false, + enableSound: true, + baseRate: baseRate, + soundMuted: audioMode == .off, + ambient: audioMode == .ambient, + mixWithOthers: true, + continuePlayingWithoutSoundOnLostAudioSession: false, + isAudioVideoMessage: false, + playerNode: self.playerNode + ) + self.playerNode.isHidden = true + self.player.setBaseRate(baseRate) + + super.init(frame: CGRect()) + + self.addSubview(self.playerNode.view) + + self.playerNode.hasSentFramesToDisplay = { [weak self] in + guard let self else { + return + } + self.playerNode.isHidden = false + } + + self.player.actionAtEnd = .action({ [weak self] in + guard let self else { + return + } + self.playbackCompleted?() + }) + } + + required init?(coder: NSCoder) { + fatalError("init(coder:) has not been implemented") + } + + func play() { + if self.isFirstPlay { + self.isFirstPlay = false + + if self.audioMode != .off { + self.player.playOnceWithSound(playAndRecord: false, seek: .start) + } else { + self.player.play() + } + } else { + self.player.play() + } + } + + func pause() { + self.player.pause() + } + + func seek(timestamp: Double) { + self.player.seek(timestamp: timestamp, play: nil) + } + + func setSoundMuted(soundMuted: Bool) { + self.player.setSoundMuted(soundMuted: soundMuted) + } + + func continueWithOverridingAmbientMode(isAmbient: Bool) { + self.player.continueWithOverridingAmbientMode(isAmbient: isAmbient) + } + + func setBaseRate(baseRate: Double) { + self.player.setBaseRate(baseRate) + } + + func update(size: CGSize, transition: ComponentTransition) { + transition.containedViewLayoutTransition.updateFrame(node: self.playerNode, frame: CGRect(origin: CGPoint(), size: size)) + } + + func updateNext(nextVideoView: ModernStoryVideoView?) { + self.player.migrateToNextPlayerOnEnd = nextVideoView?.player + } +} + final class StoryItemContentComponent: Component { typealias EnvironmentType = StoryContentItem.Environment @@ -91,10 +360,11 @@ final class StoryItemContentComponent: Component { final class View: StoryContentItem.View { private let imageView: StoryItemImageView private let overlaysView: StoryItemOverlaysView - private var videoNode: UniversalVideoNode? private var loadingEffectView: StoryItemLoadingEffectView? private var loadingEffectAppearanceTimer: SwiftSignalKit.Timer? + private var videoView: StoryVideoView? + private var mediaAreasEffectView: StoryItemLoadingEffectView? private var currentMessageMedia: EngineMedia? @@ -129,6 +399,8 @@ final class StoryItemContentComponent: Component { private var fetchPriorityResourceId: String? private var currentFetchPriority: (isMain: Bool, disposable: Disposable)? + private weak var nextItemView: StoryItemContentComponent.View? + override init(frame: CGRect) { self.hierarchyTrackingLayer = HierarchyTrackingLayer() self.imageView = StoryItemImageView() @@ -186,10 +458,7 @@ final class StoryItemContentComponent: Component { } private func initializeVideoIfReady(update: Bool) { - if self.videoNode != nil { - return - } - if case .pause = self.progressMode { + if self.videoView != nil { return } @@ -197,48 +466,49 @@ final class StoryItemContentComponent: Component { return } + var useLegacyImplementation = false + if let data = component.context.currentAppConfiguration.with({ $0 }).data, let value = data["ios_video_legacystoryplayer"] as? Double { + useLegacyImplementation = value != 0.0 + } + + if case .pause = self.progressMode { + if useLegacyImplementation { + return + } + } + if case let .file(file) = currentMessageMedia, let peerReference = PeerReference(component.peer._asPeer()) { - if self.videoNode == nil { - let videoNode = UniversalVideoNode( - context: component.context, - postbox: component.context.account.postbox, - audioSession: component.context.sharedContext.mediaManager.audioSession, - manager: component.context.sharedContext.mediaManager.universalVideoManager, - decoration: StoryVideoDecoration(), - content: NativeVideoContent( - id: .contextResult(0, "\(UInt64.random(in: 0 ... UInt64.max))"), - userLocation: .peer(peerReference.id), - fileReference: .story(peer: peerReference, id: component.item.id, media: file), - imageReference: nil, - streamVideo: .story, - loopVideo: true, - enableSound: true, - soundMuted: component.audioMode == .off, - beginWithAmbientSound: component.audioMode == .ambient, - mixWithOthers: true, - useLargeThumbnail: false, - autoFetchFullSizeThumbnail: false, - tempFilePath: nil, - captureProtected: component.item.isForwardingDisabled, - hintDimensions: file.dimensions?.cgSize, - storeAfterDownload: nil, - displayImage: false, - hasSentFramesToDisplay: { [weak self] in - guard let self else { - return - } - self.videoNode?.isHidden = false - } - ), - priority: .gallery - ) - videoNode.isHidden = true - videoNode.setBaseRate(component.baseRate) + if self.videoView == nil { + let videoView: StoryVideoView + if useLegacyImplementation { + videoView = LegacyStoryVideoView( + context: component.context, + file: .story(peer: peerReference, id: component.item.id, media: file), + audioMode: component.audioMode, + baseRate: component.baseRate, + isCaptureProtected: component.item.isForwardingDisabled + ) + } else { + let audioContext: ChunkMediaPlayerV2.AudioContext + if let current = self.environment?.sharedState.audioContext { + audioContext = current + } else { + audioContext = ChunkMediaPlayerV2.AudioContext(audioSessionManager: component.context.sharedContext.mediaManager.audioSession) + self.environment?.sharedState.audioContext = audioContext + } + videoView = ModernStoryVideoView( + context: component.context, + audioContext: audioContext, + file: .story(peer: peerReference, id: component.item.id, media: file), + audioMode: component.audioMode, + baseRate: component.baseRate, + isCaptureProtected: component.item.isForwardingDisabled + ) + } + self.videoView = videoView + self.insertSubview(videoView, aboveSubview: self.imageView) - self.videoNode = videoNode - self.insertSubview(videoNode.view, aboveSubview: self.imageView) - - videoNode.playbackCompleted = { [weak self] in + videoView.playbackCompleted = { [weak self] in guard let self else { return } @@ -253,38 +523,24 @@ final class StoryItemContentComponent: Component { if shouldLoop { self.rewind() - if let videoNode = self.videoNode { + if let videoView = self.videoView { if self.contentLoaded { - videoNode.play() + videoView.play() } } } else { self.environment?.presentationProgressUpdated(1.0, false, true) } } - videoNode.ownsContentNodeUpdated = { [weak self] value in - guard let self, let component = self.component else { - return - } - if value { - self.videoNode?.seek(0.0) - if component.audioMode != .off { - self.videoNode?.playOnceWithSound(playAndRecord: false, actionAtEnd: .stop) - } else { - self.videoNode?.play() - } - } - } - videoNode.canAttachContent = true if update { self.state?.updated(transition: .immediate) } } } - if let videoNode = self.videoNode { + if let videoView = self.videoView { if self.videoProgressDisposable == nil { - self.videoProgressDisposable = (videoNode.status + self.videoProgressDisposable = (videoView.status |> deliverOnMainQueue).start(next: { [weak self] status in guard let self, let status else { return @@ -296,7 +552,17 @@ final class StoryItemContentComponent: Component { } }) } + + let canPlay = self.progressMode != .pause && self.contentLoaded && self.hierarchyTrackingLayer.isInHierarchy + + if canPlay { + videoView.play() + } else { + videoView.pause() + } } + + self.updateVideoNextItem() } override func setProgressMode(_ progressMode: StoryContentItem.ProgressMode) { @@ -310,48 +576,62 @@ final class StoryItemContentComponent: Component { } } + func setNextItemView(nextItemView: StoryItemContentComponent.View?) { + if self.nextItemView !== nextItemView { + self.nextItemView = nextItemView + self.updateVideoNextItem() + } + } + + private func updateVideoNextItem() { + if let videoView = self.videoView as? ModernStoryVideoView { + let nextVideoView = self.nextItemView?.videoView as? ModernStoryVideoView + videoView.updateNext(nextVideoView: nextVideoView) + } + } + override func rewind() { self.currentProgressTimerValue = 0.0 - if let videoNode = self.videoNode { + if let videoView = self.videoView { if self.contentLoaded { - videoNode.seek(0.0) + videoView.seek(timestamp: 0.0) } } } override func leaveAmbientMode() { - if let videoNode = self.videoNode { + if let videoView = self.videoView { self.ignoreBufferingTimestamp = CFAbsoluteTimeGetCurrent() - videoNode.setSoundMuted(soundMuted: false) - videoNode.continueWithOverridingAmbientMode(isAmbient: false) + videoView.setSoundMuted(soundMuted: false) + videoView.continueWithOverridingAmbientMode(isAmbient: false) } } override func enterAmbientMode(ambient: Bool) { - if let videoNode = self.videoNode { + if let videoView = self.videoView { self.ignoreBufferingTimestamp = CFAbsoluteTimeGetCurrent() if ambient { - videoNode.continueWithOverridingAmbientMode(isAmbient: true) + videoView.continueWithOverridingAmbientMode(isAmbient: true) } else { - videoNode.setSoundMuted(soundMuted: true) + videoView.setSoundMuted(soundMuted: true) } } } override func setBaseRate(_ baseRate: Double) { - if let videoNode = self.videoNode { - videoNode.setBaseRate(baseRate) + if let videoView = self.videoView { + videoView.setBaseRate(baseRate: baseRate) } } private func updateProgressMode(update: Bool) { - if let videoNode = self.videoNode { + if let videoView = self.videoView { let canPlay = self.progressMode != .pause && self.contentLoaded && self.hierarchyTrackingLayer.isInHierarchy if canPlay { - videoNode.play() + videoView.play() } else { - videoNode.pause() + videoView.pause() } } @@ -566,11 +846,11 @@ final class StoryItemContentComponent: Component { private var isSeeking = false func seekTo(_ timestamp: Double, apply: Bool) { - guard let videoNode = self.videoNode else { + guard let videoView = self.videoView else { return } if apply { - videoNode.seek(min(timestamp, self.effectiveDuration - 0.3)) + videoView.seek(timestamp: min(timestamp, self.effectiveDuration - 0.3)) } self.isSeeking = true self.updateVideoPlaybackProgress(timestamp) @@ -588,6 +868,10 @@ final class StoryItemContentComponent: Component { let environment = environment[StoryContentItem.Environment.self].value self.environment = environment + if let videoView = self.videoView { + videoView.audioMode = component.audioMode + } + var synchronousLoad = false if let hint = transition.userData(Hint.self) { synchronousLoad = hint.synchronousLoad @@ -632,12 +916,12 @@ final class StoryItemContentComponent: Component { self.currentMessageMedia = messageMedia reloadMedia = true - if let videoNode = self.videoNode { + if let videoView = self.videoView { self.videoProgressDisposable?.dispose() self.videoProgressDisposable = nil - self.videoNode = nil - videoNode.view.removeFromSuperview() + self.videoView = nil + videoView.removeFromSuperview() } } self.currentMessageMetadataMedia = component.item.media @@ -767,10 +1051,10 @@ final class StoryItemContentComponent: Component { } let _ = imageSize - if let videoNode = self.videoNode { + if let videoView = self.videoView { let videoSize = dimensions.aspectFilled(availableSize) - videoNode.frame = CGRect(origin: CGPoint(x: floor((availableSize.width - videoSize.width) * 0.5), y: floor((availableSize.height - videoSize.height) * 0.5)), size: videoSize) - videoNode.updateLayout(size: videoSize, transition: .immediate) + videoView.frame = CGRect(origin: CGPoint(x: floor((availableSize.width - videoSize.width) * 0.5), y: floor((availableSize.height - videoSize.height) * 0.5)), size: videoSize) + videoView.update(size: videoSize, transition: .immediate) } } } diff --git a/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemSetContainerComponent.swift b/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemSetContainerComponent.swift index 856df6d79b..27b69c1316 100644 --- a/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemSetContainerComponent.swift +++ b/submodules/TelegramUI/Components/Stories/StoryContainerScreen/Sources/StoryItemSetContainerComponent.swift @@ -1478,7 +1478,7 @@ public final class StoryItemSetContainerComponent: Component { } if itemLayout.contentScaleFraction <= 0.0001 && !self.preparingToDisplayViewList { - if index != centralIndex { + if index != centralIndex && index != centralIndex + 1 { itemVisible = false } } @@ -1870,6 +1870,19 @@ public final class StoryItemSetContainerComponent: Component { } } + for i in 0 ..< component.slice.allItems.count { + guard let visibleItem = self.visibleItems[component.slice.allItems[i].id] else { + continue + } + var nextVisibleItem: VisibleItem? + if i != component.slice.allItems.count { + nextVisibleItem = self.visibleItems[component.slice.allItems[i + 1].id] + } + if let itemView = visibleItem.view.view as? StoryItemContentComponent.View { + itemView.setNextItemView(nextItemView: nextVisibleItem?.view.view as? StoryItemContentComponent.View) + } + } + self.trulyValidIds = trulyValidIds var removeIds: [StoryId] = [] diff --git a/submodules/TelegramUniversalVideoContent/Sources/HLSVideoJSNativeContentNode.swift b/submodules/TelegramUniversalVideoContent/Sources/HLSVideoJSNativeContentNode.swift index 2115c94488..e4dca05f71 100644 --- a/submodules/TelegramUniversalVideoContent/Sources/HLSVideoJSNativeContentNode.swift +++ b/submodules/TelegramUniversalVideoContent/Sources/HLSVideoJSNativeContentNode.swift @@ -1093,7 +1093,7 @@ final class HLSVideoJSNativeContentNode: ASDisplayNode, UniversalVideoContentNod var onSeeked: (() -> Void)? self.player = ChunkMediaPlayerV2( params: ChunkMediaPlayerV2.MediaDataReaderParams(context: context), - audioSessionManager: audioSessionManager, + audioContext: ChunkMediaPlayerV2.AudioContext(audioSessionManager: audioSessionManager), source: .externalParts(self.chunkPlayerPartsState.get()), video: true, enableSound: self.enableSound, diff --git a/submodules/TelegramUniversalVideoContent/Sources/NativeVideoContent.swift b/submodules/TelegramUniversalVideoContent/Sources/NativeVideoContent.swift index e8b67bd497..afb8dadbba 100644 --- a/submodules/TelegramUniversalVideoContent/Sources/NativeVideoContent.swift +++ b/submodules/TelegramUniversalVideoContent/Sources/NativeVideoContent.swift @@ -520,7 +520,7 @@ private final class NativeVideoContentNode: ASDisplayNode, UniversalVideoContent } else { let mediaPlayer = ChunkMediaPlayerV2( params: ChunkMediaPlayerV2.MediaDataReaderParams(context: context), - audioSessionManager: audioSessionManager, + audioContext: ChunkMediaPlayerV2.AudioContext(audioSessionManager: audioSessionManager), source: .directFetch(ChunkMediaPlayerV2.SourceDescription.ResourceDescription( postbox: postbox, size: selectedFile.size ?? 0,