// Copyright (c) Microsoft Corporation. // Licensed under the MIT License. using System; using System.Threading.Tasks; using Unity.Collections; using Unity.Collections.LowLevel.Unsafe; using UnityEngine; using UnityEngine.Rendering; using UnityEngine.XR; namespace Microsoft.MixedReality.WebRTC.Unity { /// /// Custom video source capturing the Unity scene content as rendered by a given camera, /// and sending it as a video track through the selected peer connection. /// public class SceneVideoSource : CustomVideoSource { /// /// Camera used to capture the scene content, whose rendering is used as /// video content for the track. /// /// /// If the project uses Multi-Pass stereoscopic rendering, then this camera needs to /// render to a single eye to produce a single video frame. Generally this means that /// this needs to be a separate Unity camera from the one used for XR rendering, which /// is generally rendering to both eyes. /// /// If the project uses Single-Pass Instanced stereoscopic rendering, then Unity 2019.1+ /// is required to make this component work, due to the fact earlier versions of Unity /// are missing some command buffer API calls to be able to efficiently access the camera /// backbuffer in this mode. For Unity 2018.3 users who cannot upgrade, use Single-Pass /// (non-instanced) instead. /// [Header("Camera")] [Tooltip("Camera used to capture the scene content sent by the track.")] [CaptureCamera] public Camera SourceCamera; /// /// Camera event indicating the point in time during the Unity frame rendering /// when the camera rendering is to be captured. /// /// This defaults to , which is a reasonable /// default to capture the entire scene rendering, but can be customized to achieve /// other effects like capturing only a part of the scene. /// [Tooltip("Camera event when to insert the scene capture at.")] public CameraEvent CameraEvent = CameraEvent.AfterEverything; /// /// Command buffer attached to the camera to capture its rendered content from the GPU /// and transfer it to the CPU for dispatching to WebRTC. /// private CommandBuffer _commandBuffer; /// /// Read-back texture where the content of the camera backbuffer is copied before being /// transferred from GPU to CPU. The size of the texture is . /// private RenderTexture _readBackTex; /// /// Cached width, in pixels, of the readback texture and video frame produced. /// private int _readBackWidth; /// /// Cached height, in pixels, of the readback texture and video frame produced. /// private int _readBackHeight; /// /// Temporary storage for frames generated by GPU readback until consumed by WebRTC. /// private VideoFrameQueue _frameQueue = new VideoFrameQueue(3); protected override void OnEnable() { if (!SystemInfo.supportsAsyncGPUReadback) { Debug.LogError("This platform does not support async GPU readback. Cannot use the SceneVideoSender component."); enabled = false; return; } // If no camera provided, attempt to fallback to main camera if (SourceCamera == null) { var mainCameraGameObject = GameObject.FindGameObjectWithTag("MainCamera"); if (mainCameraGameObject != null) { SourceCamera = mainCameraGameObject.GetComponent(); } } if (SourceCamera == null) { throw new NullReferenceException("Empty source camera for SceneVideoSource, and could not find MainCamera as fallback."); } CreateCommandBuffer(); SourceCamera.AddCommandBuffer(CameraEvent, _commandBuffer); // Create the track source base.OnEnable(); } protected override void OnDisable() { base.OnDisable(); if (_commandBuffer != null) { // The camera sometimes goes away before this component. if (SourceCamera != null) { SourceCamera.RemoveCommandBuffer(CameraEvent, _commandBuffer); } _commandBuffer.Dispose(); _commandBuffer = null; } } /// /// Create the command buffer reading the scene content from the source camera back into CPU memory /// and delivering it via the callback to /// the underlying WebRTC track. /// private void CreateCommandBuffer() { if (_commandBuffer != null) { throw new InvalidOperationException("Command buffer already initialized."); } // By default, use the camera's render target texture size _readBackWidth = SourceCamera.scaledPixelWidth; _readBackHeight = SourceCamera.scaledPixelHeight; // Offset and scale into source render target. Vector2 srcScale = Vector2.one; Vector2 srcOffset = Vector2.zero; RenderTextureFormat srcFormat = RenderTextureFormat.ARGB32; // Handle stereoscopic rendering for VR/AR. // See https://unity3d.com/how-to/XR-graphics-development-tips for details. if (SourceCamera.stereoEnabled) { // Readback size is the size of the texture for a single eye. // The readback will occur on the left eye (chosen arbitrarily). _readBackWidth = XRSettings.eyeTextureWidth; _readBackHeight = XRSettings.eyeTextureHeight; srcFormat = XRSettings.eyeTextureDesc.colorFormat; if (XRSettings.stereoRenderingMode == XRSettings.StereoRenderingMode.MultiPass) { // Multi-pass is similar to non-stereo, nothing to do. // Ensure camera is not rendering to both eyes in multi-pass stereo, otherwise the command buffer // is executed twice (once per eye) and will produce twice as many frames, which leads to stuttering // when playing back the video stream resulting from combining those frames. if (SourceCamera.stereoTargetEye == StereoTargetEyeMask.Both) { throw new InvalidOperationException("SourceCamera has stereoscopic rendering enabled to both eyes" + " with multi-pass rendering (XRSettings.stereoRenderingMode = MultiPass). This is not supported" + " with SceneVideoSource, as this would produce one image per eye. Either set XRSettings." + "stereoRenderingMode to single-pass (instanced or not), or use multi-pass with a camera rendering" + " to a single eye (Camera.stereoTargetEye != Both)."); } } else if (XRSettings.stereoRenderingMode == XRSettings.StereoRenderingMode.SinglePass) { // Single-pass (non-instanced) stereo use "wide-buffer" packing. // Left eye corresponds to the left half of the buffer. srcScale.x = 0.5f; } else if ((XRSettings.stereoRenderingMode == XRSettings.StereoRenderingMode.SinglePassInstanced) || (XRSettings.stereoRenderingMode == XRSettings.StereoRenderingMode.SinglePassMultiview)) // same as instanced (OpenGL) { // Single-pass instanced stereo use texture array packing. // Left eye corresponds to the first array slice. #if !UNITY_2019_1_OR_NEWER // https://unity3d.com/unity/alpha/2019.1.0a13 // "Graphics: Graphics.Blit and CommandBuffer.Blit methods now support blitting to and from texture arrays." throw new NotSupportedException("Capturing scene content in single-pass instanced stereo rendering requires" + " blitting from the Texture2DArray render target of the camera, which is not supported before Unity 2019.1." + " To use this feature, either upgrade your project to Unity 2019.1+ or use single-pass non-instanced stereo" + " rendering (XRSettings.stereoRenderingMode = SinglePass)."); #endif } } _readBackTex = new RenderTexture(_readBackWidth, _readBackHeight, 0, srcFormat, RenderTextureReadWrite.Linear); _commandBuffer = new CommandBuffer(); _commandBuffer.name = "SceneVideoSource"; // Explicitly set the render target to instruct the GPU to discard previous content. // https://docs.unity3d.com/ScriptReference/Rendering.CommandBuffer.Blit.html recommends this. //< TODO - This doesn't work //_commandBuffer.SetRenderTarget(_readBackTex, RenderBufferLoadAction.DontCare, RenderBufferStoreAction.Store); // Copy camera target to readback texture _commandBuffer.BeginSample("Blit"); #if UNITY_2019_1_OR_NEWER int srcSliceIndex = 0; // left eye int dstSliceIndex = 0; _commandBuffer.Blit(BuiltinRenderTextureType.CameraTarget, /*BuiltinRenderTextureType.CurrentActive*/_readBackTex, srcScale, srcOffset, srcSliceIndex, dstSliceIndex); #else _commandBuffer.Blit(BuiltinRenderTextureType.CameraTarget, /*BuiltinRenderTextureType.CurrentActive*/_readBackTex, srcScale, srcOffset); #endif _commandBuffer.EndSample("Blit"); // Copy readback texture to RAM asynchronously, invoking the given callback once done _commandBuffer.BeginSample("Readback"); _commandBuffer.RequestAsyncReadback(_readBackTex, 0, TextureFormat.BGRA32, OnSceneFrameReady); _commandBuffer.EndSample("Readback"); } protected override void OnFrameRequested(in FrameRequest request) { // Try to dequeue a frame from the internal frame queue if (_frameQueue.TryDequeue(out Argb32VideoFrameStorage storage)) { var frame = new Argb32VideoFrame { width = storage.Width, height = storage.Height, stride = (int)storage.Width * 4 }; unsafe { fixed (void* ptr = storage.Buffer) { // Complete the request with a view over the frame buffer (no allocation) // while the buffer is pinned into memory. The native implementation will // make a copy into a native memory buffer if necessary before returning. frame.data = new IntPtr(ptr); request.CompleteRequest(frame); } } // Put the allocated buffer back in the pool for reuse _frameQueue.RecycleStorage(storage); } } /// /// Callback invoked by the command buffer when the scene frame GPU readback has completed /// and the frame is available in CPU memory. /// /// The completed and possibly failed GPU readback request. private void OnSceneFrameReady(AsyncGPUReadbackRequest request) { // Read back the data from GPU, if available if (request.hasError) { return; } NativeArray rawData = request.GetData(); Debug.Assert(rawData.Length >= _readBackWidth * _readBackHeight * 4); unsafe { byte* ptr = (byte*)NativeArrayUnsafeUtility.GetUnsafePtr(rawData); // Enqueue a frame in the internal frame queue. This will make a copy // of the frame into a pooled buffer owned by the frame queue. var frame = new Argb32VideoFrame { data = (IntPtr)ptr, stride = _readBackWidth * 4, width = (uint)_readBackWidth, height = (uint)_readBackHeight }; _frameQueue.Enqueue(frame); } } } }