mirror of
https://github.com/mollyim/webrtc.git
synced 2025-05-17 23:57:59 +01:00

In https://webrtc-review.googlesource.com/c/src/+/1560 we moved WebRTC from src/webrtc to src/ (in order to preserve an healthy git history). This CL takes care of fixing header guards, #include paths, etc... NOPRESUBMIT=true NOTREECHECKS=true NOTRY=true TBR=tommi@webrtc.org Bug: chromium:611808 Change-Id: Iea91618212bee0af16aa3f05071eab8f93706578 Reviewed-on: https://webrtc-review.googlesource.com/1561 Reviewed-by: Mirko Bonadei <mbonadei@webrtc.org> Reviewed-by: Henrik Kjellander <kjellander@webrtc.org> Commit-Queue: Mirko Bonadei <mbonadei@webrtc.org> Cr-Commit-Position: refs/heads/master@{#19846}
179 lines
6.2 KiB
Text
179 lines
6.2 KiB
Text
/*
|
|
* Copyright 2015 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "avfoundationvideocapturer.h"
|
|
|
|
#import <AVFoundation/AVFoundation.h>
|
|
|
|
#import "RTCAVFoundationVideoCapturerInternal.h"
|
|
#import "RTCDispatcher+Private.h"
|
|
#import "WebRTC/RTCLogging.h"
|
|
#import "WebRTC/RTCVideoFrameBuffer.h"
|
|
|
|
#include "avfoundationformatmapper.h"
|
|
|
|
#include "api/video/video_rotation.h"
|
|
#include "rtc_base/bind.h"
|
|
#include "rtc_base/checks.h"
|
|
#include "rtc_base/logging.h"
|
|
#include "rtc_base/thread.h"
|
|
#include "sdk/objc/Framework/Classes/Video/objc_frame_buffer.h"
|
|
|
|
namespace webrtc {
|
|
|
|
enum AVFoundationVideoCapturerMessageType : uint32_t {
|
|
kMessageTypeFrame,
|
|
};
|
|
|
|
AVFoundationVideoCapturer::AVFoundationVideoCapturer() : _capturer(nil) {
|
|
_capturer =
|
|
[[RTCAVFoundationVideoCapturerInternal alloc] initWithCapturer:this];
|
|
|
|
std::set<cricket::VideoFormat> front_camera_video_formats =
|
|
GetSupportedVideoFormatsForDevice([_capturer frontCaptureDevice]);
|
|
std::set<cricket::VideoFormat> back_camera_video_formats =
|
|
GetSupportedVideoFormatsForDevice([_capturer backCaptureDevice]);
|
|
std::vector<cricket::VideoFormat> intersection_video_formats;
|
|
if (back_camera_video_formats.empty()) {
|
|
intersection_video_formats.assign(front_camera_video_formats.begin(),
|
|
front_camera_video_formats.end());
|
|
|
|
} else if (front_camera_video_formats.empty()) {
|
|
intersection_video_formats.assign(back_camera_video_formats.begin(),
|
|
back_camera_video_formats.end());
|
|
} else {
|
|
std::set_intersection(
|
|
front_camera_video_formats.begin(), front_camera_video_formats.end(),
|
|
back_camera_video_formats.begin(), back_camera_video_formats.end(),
|
|
std::back_inserter(intersection_video_formats));
|
|
}
|
|
SetSupportedFormats(intersection_video_formats);
|
|
}
|
|
|
|
AVFoundationVideoCapturer::~AVFoundationVideoCapturer() {
|
|
_capturer = nil;
|
|
}
|
|
|
|
cricket::CaptureState AVFoundationVideoCapturer::Start(
|
|
const cricket::VideoFormat& format) {
|
|
if (!_capturer) {
|
|
LOG(LS_ERROR) << "Failed to create AVFoundation capturer.";
|
|
return cricket::CaptureState::CS_FAILED;
|
|
}
|
|
if (_capturer.isRunning) {
|
|
LOG(LS_ERROR) << "The capturer is already running.";
|
|
return cricket::CaptureState::CS_FAILED;
|
|
}
|
|
|
|
AVCaptureDevice* device = [_capturer getActiveCaptureDevice];
|
|
AVCaptureSession* session = _capturer.captureSession;
|
|
|
|
if (!SetFormatForCaptureDevice(device, session, format)) {
|
|
return cricket::CaptureState::CS_FAILED;
|
|
}
|
|
|
|
SetCaptureFormat(&format);
|
|
// This isn't super accurate because it takes a while for the AVCaptureSession
|
|
// to spin up, and this call returns async.
|
|
// TODO(tkchin): make this better.
|
|
[_capturer start];
|
|
SetCaptureState(cricket::CaptureState::CS_RUNNING);
|
|
|
|
return cricket::CaptureState::CS_STARTING;
|
|
}
|
|
|
|
void AVFoundationVideoCapturer::Stop() {
|
|
[_capturer stop];
|
|
SetCaptureFormat(NULL);
|
|
}
|
|
|
|
bool AVFoundationVideoCapturer::IsRunning() {
|
|
return _capturer.isRunning;
|
|
}
|
|
|
|
AVCaptureSession* AVFoundationVideoCapturer::GetCaptureSession() {
|
|
return _capturer.captureSession;
|
|
}
|
|
|
|
bool AVFoundationVideoCapturer::CanUseBackCamera() const {
|
|
return _capturer.canUseBackCamera;
|
|
}
|
|
|
|
void AVFoundationVideoCapturer::SetUseBackCamera(bool useBackCamera) {
|
|
_capturer.useBackCamera = useBackCamera;
|
|
}
|
|
|
|
bool AVFoundationVideoCapturer::GetUseBackCamera() const {
|
|
return _capturer.useBackCamera;
|
|
}
|
|
|
|
void AVFoundationVideoCapturer::AdaptOutputFormat(int width, int height, int fps) {
|
|
cricket::VideoFormat format(width, height, cricket::VideoFormat::FpsToInterval(fps), 0);
|
|
video_adapter()->OnOutputFormatRequest(format);
|
|
}
|
|
|
|
void AVFoundationVideoCapturer::CaptureSampleBuffer(
|
|
CMSampleBufferRef sample_buffer, VideoRotation rotation) {
|
|
if (CMSampleBufferGetNumSamples(sample_buffer) != 1 ||
|
|
!CMSampleBufferIsValid(sample_buffer) ||
|
|
!CMSampleBufferDataIsReady(sample_buffer)) {
|
|
return;
|
|
}
|
|
|
|
CVImageBufferRef image_buffer = CMSampleBufferGetImageBuffer(sample_buffer);
|
|
if (image_buffer == NULL) {
|
|
return;
|
|
}
|
|
|
|
int captured_width = CVPixelBufferGetWidth(image_buffer);
|
|
int captured_height = CVPixelBufferGetHeight(image_buffer);
|
|
|
|
int adapted_width;
|
|
int adapted_height;
|
|
int crop_width;
|
|
int crop_height;
|
|
int crop_x;
|
|
int crop_y;
|
|
int64_t translated_camera_time_us;
|
|
|
|
if (!AdaptFrame(captured_width, captured_height,
|
|
rtc::TimeNanos() / rtc::kNumNanosecsPerMicrosec,
|
|
rtc::TimeMicros(), &adapted_width, &adapted_height,
|
|
&crop_width, &crop_height, &crop_x, &crop_y,
|
|
&translated_camera_time_us)) {
|
|
return;
|
|
}
|
|
|
|
RTCCVPixelBuffer* rtcPixelBuffer = [[RTCCVPixelBuffer alloc] initWithPixelBuffer:image_buffer
|
|
adaptedWidth:adapted_width
|
|
adaptedHeight:adapted_height
|
|
cropWidth:crop_width
|
|
cropHeight:crop_height
|
|
cropX:crop_x
|
|
cropY:crop_y];
|
|
rtc::scoped_refptr<VideoFrameBuffer> buffer =
|
|
new rtc::RefCountedObject<ObjCFrameBuffer>(rtcPixelBuffer);
|
|
|
|
// Applying rotation is only supported for legacy reasons and performance is
|
|
// not critical here.
|
|
if (apply_rotation() && rotation != kVideoRotation_0) {
|
|
buffer = I420Buffer::Rotate(*buffer->ToI420(), rotation);
|
|
if (rotation == kVideoRotation_90 || rotation == kVideoRotation_270) {
|
|
std::swap(captured_width, captured_height);
|
|
}
|
|
|
|
rotation = kVideoRotation_0;
|
|
}
|
|
|
|
OnFrame(webrtc::VideoFrame(buffer, rotation, translated_camera_time_us),
|
|
captured_width, captured_height);
|
|
}
|
|
|
|
} // namespace webrtc
|