blob: 9d6a08c60c19460c3b6e0d3a520341f278ef30b4 [file] [log] [blame]
// Copyright 2016 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "media/capture/content/android/screen_capture_machine_android.h"
#include <utility>
#include "base/android/jni_android.h"
#include "base/android/scoped_java_ref.h"
#include "media/base/video_frame.h"
#include "media/capture/content/android/screen_capture_jni_headers/ScreenCapture_jni.h"
#include "media/capture/content/android/thread_safe_capture_oracle.h"
#include "media/capture/content/video_capture_oracle.h"
#include "media/capture/video_capture_types.h"
#include "third_party/libyuv/include/libyuv.h"
using base::android::AttachCurrentThread;
using base::android::JavaRef;
using base::android::ScopedJavaLocalRef;
namespace media {
ScreenCaptureMachineAndroid::ScreenCaptureMachineAndroid() {}
ScreenCaptureMachineAndroid::~ScreenCaptureMachineAndroid() {}
// static
ScopedJavaLocalRef<jobject>
ScreenCaptureMachineAndroid::createScreenCaptureMachineAndroid(
jlong nativeScreenCaptureMachineAndroid) {
return (Java_ScreenCapture_createScreenCaptureMachine(
AttachCurrentThread(), nativeScreenCaptureMachineAndroid));
}
void ScreenCaptureMachineAndroid::OnRGBAFrameAvailable(
JNIEnv* env,
const JavaRef<jobject>& obj,
const JavaRef<jobject>& buf,
jint row_stride,
jint left,
jint top,
jint width,
jint height,
jlong timestamp) {
const VideoCaptureOracle::Event event = VideoCaptureOracle::kCompositorUpdate;
const uint64_t absolute_micro =
timestamp / base::Time::kNanosecondsPerMicrosecond;
const base::TimeTicks start_time =
base::TimeTicks() + base::Microseconds(absolute_micro);
scoped_refptr<VideoFrame> frame;
ThreadSafeCaptureOracle::CaptureFrameCallback capture_frame_cb;
if (!oracle_proxy_->ObserveEventAndDecideCapture(
event, gfx::Rect(), start_time, &frame, &capture_frame_cb)) {
return;
}
DCHECK(frame->format() == PIXEL_FORMAT_I420);
scoped_refptr<VideoFrame> temp_frame = frame;
if (frame->visible_rect().width() != width ||
frame->visible_rect().height() != height) {
temp_frame = VideoFrame::CreateFrame(
PIXEL_FORMAT_I420, gfx::Size(width, height), gfx::Rect(width, height),
gfx::Size(width, height), base::TimeDelta());
}
uint8_t* const src =
reinterpret_cast<uint8_t*>(env->GetDirectBufferAddress(buf.obj()));
CHECK(src);
const int offset = top * row_stride + left * 4;
// ABGR little endian (rgba in memory) to I420.
libyuv::ABGRToI420(
src + offset, row_stride, temp_frame->visible_data(VideoFrame::kYPlane),
temp_frame->stride(VideoFrame::kYPlane),
temp_frame->visible_data(VideoFrame::kUPlane),
temp_frame->stride(VideoFrame::kUPlane),
temp_frame->visible_data(VideoFrame::kVPlane),
temp_frame->stride(VideoFrame::kVPlane),
temp_frame->visible_rect().width(), temp_frame->visible_rect().height());
if (temp_frame != frame) {
libyuv::I420Scale(
temp_frame->visible_data(VideoFrame::kYPlane),
temp_frame->stride(VideoFrame::kYPlane),
temp_frame->visible_data(VideoFrame::kUPlane),
temp_frame->stride(VideoFrame::kUPlane),
temp_frame->visible_data(VideoFrame::kVPlane),
temp_frame->stride(VideoFrame::kVPlane),
temp_frame->visible_rect().width(), temp_frame->visible_rect().height(),
frame->visible_data(VideoFrame::kYPlane),
frame->stride(VideoFrame::kYPlane),
frame->visible_data(VideoFrame::kUPlane),
frame->stride(VideoFrame::kUPlane),
frame->visible_data(VideoFrame::kVPlane),
frame->stride(VideoFrame::kVPlane), frame->visible_rect().width(),
frame->visible_rect().height(), libyuv::kFilterBilinear);
}
std::move(capture_frame_cb).Run(frame, start_time, true);
lastFrame_ = frame;
}
void ScreenCaptureMachineAndroid::OnI420FrameAvailable(
JNIEnv* env,
const JavaRef<jobject>& obj,
const JavaRef<jobject>& y_buffer,
jint y_stride,
const JavaRef<jobject>& u_buffer,
const JavaRef<jobject>& v_buffer,
jint uv_row_stride,
jint uv_pixel_stride,
jint left,
jint top,
jint width,
jint height,
jlong timestamp) {
const VideoCaptureOracle::Event event = VideoCaptureOracle::kCompositorUpdate;
const uint64_t absolute_micro =
timestamp / base::Time::kNanosecondsPerMicrosecond;
const base::TimeTicks start_time =
base::TimeTicks() + base::Microseconds(absolute_micro);
scoped_refptr<VideoFrame> frame;
ThreadSafeCaptureOracle::CaptureFrameCallback capture_frame_cb;
if (!oracle_proxy_->ObserveEventAndDecideCapture(
event, gfx::Rect(), start_time, &frame, &capture_frame_cb)) {
return;
}
DCHECK(frame->format() == PIXEL_FORMAT_I420);
scoped_refptr<VideoFrame> temp_frame = frame;
if (frame->visible_rect().width() != width ||
frame->visible_rect().height() != height) {
temp_frame = VideoFrame::CreateFrame(
PIXEL_FORMAT_I420, gfx::Size(width, height), gfx::Rect(width, height),
gfx::Size(width, height), base::TimeDelta());
}
uint8_t* const y_src =
reinterpret_cast<uint8_t*>(env->GetDirectBufferAddress(y_buffer.obj()));
CHECK(y_src);
uint8_t* u_src =
reinterpret_cast<uint8_t*>(env->GetDirectBufferAddress(u_buffer.obj()));
CHECK(u_src);
uint8_t* v_src =
reinterpret_cast<uint8_t*>(env->GetDirectBufferAddress(v_buffer.obj()));
CHECK(v_src);
const int y_offset = top * y_stride + left;
const int uv_offset = (top / 2) * uv_row_stride + left / 2;
libyuv::Android420ToI420(
y_src + y_offset, y_stride, u_src + uv_offset, uv_row_stride,
v_src + uv_offset, uv_row_stride, uv_pixel_stride,
temp_frame->visible_data(VideoFrame::kYPlane),
temp_frame->stride(VideoFrame::kYPlane),
temp_frame->visible_data(VideoFrame::kUPlane),
temp_frame->stride(VideoFrame::kUPlane),
temp_frame->visible_data(VideoFrame::kVPlane),
temp_frame->stride(VideoFrame::kVPlane),
temp_frame->visible_rect().width(), temp_frame->visible_rect().height());
if (temp_frame != frame) {
libyuv::I420Scale(
temp_frame->visible_data(VideoFrame::kYPlane),
temp_frame->stride(VideoFrame::kYPlane),
temp_frame->visible_data(VideoFrame::kUPlane),
temp_frame->stride(VideoFrame::kUPlane),
temp_frame->visible_data(VideoFrame::kVPlane),
temp_frame->stride(VideoFrame::kVPlane),
temp_frame->visible_rect().width(), temp_frame->visible_rect().height(),
frame->visible_data(VideoFrame::kYPlane),
frame->stride(VideoFrame::kYPlane),
frame->visible_data(VideoFrame::kUPlane),
frame->stride(VideoFrame::kUPlane),
frame->visible_data(VideoFrame::kVPlane),
frame->stride(VideoFrame::kVPlane), frame->visible_rect().width(),
frame->visible_rect().height(), libyuv::kFilterBilinear);
}
std::move(capture_frame_cb).Run(frame, start_time, true);
lastFrame_ = frame;
}
void ScreenCaptureMachineAndroid::OnActivityResult(JNIEnv* env,
const JavaRef<jobject>& obj,
jboolean result) {
if (!result) {
oracle_proxy_->ReportError(
media::VideoCaptureError::
kAndroidScreenCaptureTheUserDeniedScreenCapture,
FROM_HERE, "The user denied screen capture");
return;
}
if (Java_ScreenCapture_startCapture(env, obj))
oracle_proxy_->ReportStarted();
else
oracle_proxy_->ReportError(
media::VideoCaptureError::
kAndroidScreenCaptureFailedToStartScreenCapture,
FROM_HERE, "Failed to start Screen Capture");
}
void ScreenCaptureMachineAndroid::OnOrientationChange(
JNIEnv* env,
const JavaRef<jobject>& obj,
jint rotation) {
DeviceOrientation orientation = kDefault;
switch (rotation) {
case 0:
case 180:
orientation = kPortrait;
break;
case 90:
case 270:
orientation = kLandscape;
break;
default:
break;
}
gfx::Size capture_size = oracle_proxy_->GetCaptureSize();
const int width = capture_size.width();
const int height = capture_size.height();
if ((orientation == kLandscape && width < height) ||
(orientation == kPortrait && height < width)) {
capture_size.SetSize(height, width);
oracle_proxy_->UpdateCaptureSize(capture_size);
}
}
bool ScreenCaptureMachineAndroid::Start(
scoped_refptr<ThreadSafeCaptureOracle> oracle_proxy,
const VideoCaptureParams& params) {
DCHECK(oracle_proxy.get());
oracle_proxy_ = std::move(oracle_proxy);
j_capture_.Reset(
createScreenCaptureMachineAndroid(reinterpret_cast<intptr_t>(this)));
if (j_capture_.obj() == nullptr) {
DLOG(ERROR) << "Failed to createScreenCaptureAndroid";
return false;
}
DCHECK(params.requested_format.frame_size.GetArea());
jboolean ret =
Java_ScreenCapture_allocate(AttachCurrentThread(), j_capture_,
params.requested_format.frame_size.width(),
params.requested_format.frame_size.height());
if (!ret) {
DLOG(ERROR) << "Failed to init ScreenCaptureAndroid";
return false;
}
ret = Java_ScreenCapture_startPrompt(AttachCurrentThread(), j_capture_);
// NOTE: Result of user prompt will be delivered to OnActivityResult(), and
// this will report the device started/error state via the |oracle_proxy_|.
return !!ret;
}
void ScreenCaptureMachineAndroid::Stop() {
if (j_capture_.obj() != nullptr) {
Java_ScreenCapture_stopCapture(AttachCurrentThread(), j_capture_);
}
}
// ScreenCapture on Android works in a passive way and there are no captured
// frames when there is no update to the screen. When the oracle asks for a
// capture refresh, the cached captured frame is redelivered.
void ScreenCaptureMachineAndroid::MaybeCaptureForRefresh() {
if (lastFrame_.get() == nullptr)
return;
const VideoCaptureOracle::Event event = VideoCaptureOracle::kRefreshRequest;
const base::TimeTicks start_time = base::TimeTicks::Now();
scoped_refptr<VideoFrame> frame;
ThreadSafeCaptureOracle::CaptureFrameCallback capture_frame_cb;
if (!oracle_proxy_->ObserveEventAndDecideCapture(
event, gfx::Rect(), start_time, &frame, &capture_frame_cb)) {
return;
}
DCHECK(frame->format() == PIXEL_FORMAT_I420);
libyuv::I420Scale(
lastFrame_->visible_data(VideoFrame::kYPlane),
lastFrame_->stride(VideoFrame::kYPlane),
lastFrame_->visible_data(VideoFrame::kUPlane),
lastFrame_->stride(VideoFrame::kUPlane),
lastFrame_->visible_data(VideoFrame::kVPlane),
lastFrame_->stride(VideoFrame::kVPlane),
lastFrame_->visible_rect().width(), lastFrame_->visible_rect().height(),
frame->visible_data(VideoFrame::kYPlane),
frame->stride(VideoFrame::kYPlane),
frame->visible_data(VideoFrame::kUPlane),
frame->stride(VideoFrame::kUPlane),
frame->visible_data(VideoFrame::kVPlane),
frame->stride(VideoFrame::kVPlane), frame->visible_rect().width(),
frame->visible_rect().height(), libyuv::kFilterBilinear);
std::move(capture_frame_cb).Run(frame, start_time, true);
}
} // namespace media