| // Copyright (c) 2013 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/base/android/media_codec_bridge_impl.h" |
| |
| #include <algorithm> |
| #include <limits> |
| #include <memory> |
| #include <utility> |
| |
| #include "base/android/build_info.h" |
| #include "base/android/jni_android.h" |
| #include "base/android/jni_array.h" |
| #include "base/android/jni_string.h" |
| #include "base/logging.h" |
| #include "base/memory/ptr_util.h" |
| #include "base/numerics/safe_conversions.h" |
| #include "base/strings/string_util.h" |
| #include "jni/MediaCodecBridgeBuilder_jni.h" |
| #include "jni/MediaCodecBridge_jni.h" |
| #include "media/base/android/jni_hdr_metadata.h" |
| #include "media/base/android/media_codec_util.h" |
| #include "media/base/audio_codecs.h" |
| #include "media/base/bit_reader.h" |
| #include "media/base/subsample_entry.h" |
| #include "media/base/video_codecs.h" |
| |
| using base::android::AttachCurrentThread; |
| using base::android::ConvertJavaStringToUTF8; |
| using base::android::ConvertUTF8ToJavaString; |
| using base::android::JavaIntArrayToIntVector; |
| using base::android::JavaRef; |
| using base::android::ScopedJavaGlobalRef; |
| using base::android::ScopedJavaLocalRef; |
| using base::android::ToJavaByteArray; |
| |
| #define RETURN_ON_ERROR(condition) \ |
| do { \ |
| if (!(condition)) { \ |
| LOG(ERROR) << "Unable to parse AAC header: " #condition; \ |
| return false; \ |
| } \ |
| } while (0) |
| |
| namespace media { |
| namespace { |
| |
| enum { |
| kBufferFlagSyncFrame = 1, // BUFFER_FLAG_SYNC_FRAME |
| kBufferFlagEndOfStream = 4, // BUFFER_FLAG_END_OF_STREAM |
| kConfigureFlagEncode = 1, // CONFIGURE_FLAG_ENCODE |
| }; |
| |
| using CodecSpecificData = std::vector<uint8_t>; |
| |
| // Parses |extra_data| to get info to be added to a Java MediaFormat. |
| bool GetCodecSpecificDataForAudio(AudioCodec codec, |
| const uint8_t* extra_data, |
| size_t extra_data_size, |
| int64_t codec_delay_ns, |
| int64_t seek_preroll_ns, |
| CodecSpecificData* output_csd0, |
| CodecSpecificData* output_csd1, |
| CodecSpecificData* output_csd2, |
| bool* output_frame_has_adts_header) { |
| *output_frame_has_adts_header = false; |
| if (extra_data_size == 0 && codec != kCodecOpus) |
| return true; |
| |
| switch (codec) { |
| case kCodecVorbis: { |
| if (extra_data[0] != 2) { |
| LOG(ERROR) << "Invalid number of vorbis headers before the codec " |
| << "header: " << extra_data[0]; |
| return false; |
| } |
| |
| size_t header_length[2]; |
| // |total_length| keeps track of the total number of bytes before the last |
| // header. |
| size_t total_length = 1; |
| const uint8_t* current_pos = extra_data; |
| // Calculate the length of the first 2 headers. |
| for (int i = 0; i < 2; ++i) { |
| header_length[i] = 0; |
| while (total_length < extra_data_size) { |
| size_t size = *(++current_pos); |
| total_length += 1 + size; |
| if (total_length > 0x80000000) { |
| LOG(ERROR) << "Vorbis header size too large"; |
| return false; |
| } |
| header_length[i] += size; |
| if (size < 0xFF) |
| break; |
| } |
| if (total_length >= extra_data_size) { |
| LOG(ERROR) << "Invalid vorbis header size in the extra data"; |
| return false; |
| } |
| } |
| current_pos++; |
| |
| // The first header is the identification header. |
| output_csd0->assign(current_pos, current_pos + header_length[0]); |
| |
| // The last header is the codec header. |
| output_csd1->assign(extra_data + total_length, |
| extra_data + extra_data_size - total_length); |
| break; |
| } |
| case kCodecAAC: { |
| media::BitReader reader(extra_data, extra_data_size); |
| |
| // The following code is copied from aac.cc |
| // TODO(qinmin): refactor the code in aac.cc to make it more reusable. |
| uint8_t profile = 0; |
| uint8_t frequency_index = 0; |
| uint8_t channel_config = 0; |
| RETURN_ON_ERROR(reader.ReadBits(5, &profile)); |
| RETURN_ON_ERROR(reader.ReadBits(4, &frequency_index)); |
| |
| if (0xf == frequency_index) |
| RETURN_ON_ERROR(reader.SkipBits(24)); |
| RETURN_ON_ERROR(reader.ReadBits(4, &channel_config)); |
| |
| if (profile == 5 || profile == 29) { |
| // Read extension config. |
| RETURN_ON_ERROR(reader.ReadBits(4, &frequency_index)); |
| if (frequency_index == 0xf) |
| RETURN_ON_ERROR(reader.SkipBits(24)); |
| RETURN_ON_ERROR(reader.ReadBits(5, &profile)); |
| } |
| |
| if (profile < 1 || profile > 4 || frequency_index == 0xf || |
| channel_config > 7) { |
| LOG(ERROR) << "Invalid AAC header"; |
| return false; |
| } |
| |
| output_csd0->push_back(profile << 3 | frequency_index >> 1); |
| output_csd0->push_back((frequency_index & 0x01) << 7 | channel_config |
| << 3); |
| *output_frame_has_adts_header = true; |
| break; |
| } |
| case kCodecOpus: { |
| if (!extra_data || extra_data_size == 0 || codec_delay_ns < 0 || |
| seek_preroll_ns < 0) { |
| LOG(ERROR) << "Invalid Opus Header"; |
| return false; |
| } |
| |
| // csd0 - Opus Header |
| output_csd0->assign(extra_data, extra_data + extra_data_size); |
| |
| // csd1 - Codec Delay |
| const uint8_t* codec_delay_ns_ptr = |
| reinterpret_cast<const uint8_t*>(&codec_delay_ns); |
| output_csd1->assign(codec_delay_ns_ptr, |
| codec_delay_ns_ptr + sizeof(int64_t)); |
| |
| // csd2 - Seek Preroll |
| const uint8_t* seek_preroll_ns_ptr = |
| reinterpret_cast<const uint8_t*>(&seek_preroll_ns); |
| output_csd2->assign(seek_preroll_ns_ptr, |
| seek_preroll_ns_ptr + sizeof(int64_t)); |
| break; |
| } |
| default: |
| LOG(ERROR) << "Invalid header encountered for codec: " |
| << GetCodecName(codec); |
| return false; |
| } |
| return true; |
| } |
| |
| } // namespace |
| |
| // static |
| std::unique_ptr<MediaCodecBridge> MediaCodecBridgeImpl::CreateAudioDecoder( |
| const AudioDecoderConfig& config, |
| const JavaRef<jobject>& media_crypto) { |
| DVLOG(2) << __func__ << ": " << config.AsHumanReadableString() |
| << " media_crypto:" << media_crypto.obj(); |
| |
| if (!MediaCodecUtil::IsMediaCodecAvailable()) |
| return nullptr; |
| |
| const std::string mime = |
| MediaCodecUtil::CodecToAndroidMimeType(config.codec()); |
| if (mime.empty()) |
| return nullptr; |
| |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jstring> j_mime = ConvertUTF8ToJavaString(env, mime); |
| |
| const int channel_count = |
| ChannelLayoutToChannelCount(config.channel_layout()); |
| |
| // It's important that the multiplication is first in this calculation to |
| // reduce the precision loss due to integer truncation. |
| const int64_t codec_delay_ns = base::Time::kNanosecondsPerSecond * |
| config.codec_delay() / |
| config.samples_per_second(); |
| const int64_t seek_preroll_ns = config.seek_preroll().InMicroseconds() * |
| base::Time::kNanosecondsPerMicrosecond; |
| |
| CodecSpecificData csd0, csd1, csd2; |
| bool output_frame_has_adts_header; |
| if (!GetCodecSpecificDataForAudio(config.codec(), config.extra_data().data(), |
| config.extra_data().size(), codec_delay_ns, |
| seek_preroll_ns, &csd0, &csd1, &csd2, |
| &output_frame_has_adts_header)) { |
| return nullptr; |
| } |
| |
| ScopedJavaLocalRef<jbyteArray> j_csd0 = ToJavaByteArray(env, csd0); |
| ScopedJavaLocalRef<jbyteArray> j_csd1 = ToJavaByteArray(env, csd1); |
| ScopedJavaLocalRef<jbyteArray> j_csd2 = ToJavaByteArray(env, csd2); |
| |
| ScopedJavaGlobalRef<jobject> j_bridge( |
| Java_MediaCodecBridgeBuilder_createAudioDecoder( |
| env, j_mime, media_crypto, config.samples_per_second(), channel_count, |
| j_csd0, j_csd1, j_csd2, output_frame_has_adts_header)); |
| |
| if (j_bridge.is_null()) |
| return nullptr; |
| |
| return base::WrapUnique(new MediaCodecBridgeImpl(std::move(j_bridge))); |
| } |
| |
| // static |
| std::unique_ptr<MediaCodecBridge> MediaCodecBridgeImpl::CreateVideoDecoder( |
| VideoCodec codec, |
| CodecType codec_type, |
| const gfx::Size& size, |
| const JavaRef<jobject>& surface, |
| const JavaRef<jobject>& media_crypto, |
| const CodecSpecificData& csd0, |
| const CodecSpecificData& csd1, |
| const VideoColorSpace& color_space, |
| const base::Optional<HDRMetadata>& hdr_metadata, |
| bool allow_adaptive_playback) { |
| if (!MediaCodecUtil::IsMediaCodecAvailable()) |
| return nullptr; |
| |
| const std::string mime = MediaCodecUtil::CodecToAndroidMimeType(codec); |
| if (mime.empty()) |
| return nullptr; |
| |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jstring> j_mime = ConvertUTF8ToJavaString(env, mime); |
| |
| ScopedJavaLocalRef<jbyteArray> j_csd0 = |
| ToJavaByteArray(env, csd0.data(), csd0.size()); |
| ScopedJavaLocalRef<jbyteArray> j_csd1 = |
| ToJavaByteArray(env, csd1.data(), csd1.size()); |
| |
| std::unique_ptr<JniHdrMetadata> jni_hdr_metadata; |
| if (hdr_metadata.has_value()) { |
| jni_hdr_metadata.reset( |
| new JniHdrMetadata(color_space, hdr_metadata.value())); |
| } |
| ScopedJavaLocalRef<jobject> j_hdr_metadata( |
| jni_hdr_metadata ? jni_hdr_metadata->obj() : nullptr); |
| |
| ScopedJavaGlobalRef<jobject> j_bridge( |
| Java_MediaCodecBridgeBuilder_createVideoDecoder( |
| env, j_mime, static_cast<int>(codec_type), media_crypto, size.width(), |
| size.height(), surface, j_csd0, j_csd1, j_hdr_metadata, |
| allow_adaptive_playback)); |
| if (j_bridge.is_null()) |
| return nullptr; |
| |
| return base::WrapUnique(new MediaCodecBridgeImpl(std::move(j_bridge))); |
| } |
| |
| // static |
| std::unique_ptr<MediaCodecBridge> MediaCodecBridgeImpl::CreateVideoEncoder( |
| VideoCodec codec, |
| const gfx::Size& size, |
| int bit_rate, |
| int frame_rate, |
| int i_frame_interval, |
| int color_format) { |
| if (!MediaCodecUtil::IsMediaCodecAvailable()) |
| return nullptr; |
| |
| const std::string mime = MediaCodecUtil::CodecToAndroidMimeType(codec); |
| if (mime.empty()) |
| return nullptr; |
| |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jstring> j_mime = ConvertUTF8ToJavaString(env, mime); |
| ScopedJavaGlobalRef<jobject> j_bridge( |
| Java_MediaCodecBridgeBuilder_createVideoEncoder( |
| env, j_mime, size.width(), size.height(), bit_rate, frame_rate, |
| i_frame_interval, color_format)); |
| |
| if (j_bridge.is_null()) |
| return nullptr; |
| |
| return base::WrapUnique(new MediaCodecBridgeImpl(std::move(j_bridge))); |
| } |
| |
| MediaCodecBridgeImpl::MediaCodecBridgeImpl( |
| ScopedJavaGlobalRef<jobject> j_bridge) |
| : j_bridge_(std::move(j_bridge)) { |
| DCHECK(!j_bridge_.is_null()); |
| } |
| |
| MediaCodecBridgeImpl::~MediaCodecBridgeImpl() { |
| JNIEnv* env = AttachCurrentThread(); |
| if (j_bridge_.obj()) |
| Java_MediaCodecBridge_release(env, j_bridge_); |
| } |
| |
| void MediaCodecBridgeImpl::Stop() { |
| JNIEnv* env = AttachCurrentThread(); |
| Java_MediaCodecBridge_stop(env, j_bridge_); |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::Flush() { |
| JNIEnv* env = AttachCurrentThread(); |
| return static_cast<MediaCodecStatus>( |
| Java_MediaCodecBridge_flush(env, j_bridge_)); |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::GetOutputSize(gfx::Size* size) { |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jobject> result = |
| Java_MediaCodecBridge_getOutputFormat(env, j_bridge_); |
| MediaCodecStatus status = static_cast<MediaCodecStatus>( |
| Java_GetOutputFormatResult_status(env, result)); |
| if (status == MEDIA_CODEC_OK) { |
| size->SetSize(Java_GetOutputFormatResult_width(env, result), |
| Java_GetOutputFormatResult_height(env, result)); |
| } |
| return status; |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::GetOutputSamplingRate( |
| int* sampling_rate) { |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jobject> result = |
| Java_MediaCodecBridge_getOutputFormat(env, j_bridge_); |
| MediaCodecStatus status = static_cast<MediaCodecStatus>( |
| Java_GetOutputFormatResult_status(env, result)); |
| if (status == MEDIA_CODEC_OK) |
| *sampling_rate = Java_GetOutputFormatResult_sampleRate(env, result); |
| return status; |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::GetOutputChannelCount( |
| int* channel_count) { |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jobject> result = |
| Java_MediaCodecBridge_getOutputFormat(env, j_bridge_); |
| MediaCodecStatus status = static_cast<MediaCodecStatus>( |
| Java_GetOutputFormatResult_status(env, result)); |
| if (status == MEDIA_CODEC_OK) |
| *channel_count = Java_GetOutputFormatResult_channelCount(env, result); |
| return status; |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::QueueInputBuffer( |
| int index, |
| const uint8_t* data, |
| size_t data_size, |
| base::TimeDelta presentation_time) { |
| DVLOG(3) << __func__ << " " << index << ": " << data_size; |
| if (data_size > |
| base::checked_cast<size_t>(std::numeric_limits<int32_t>::max())) { |
| return MEDIA_CODEC_ERROR; |
| } |
| if (data && !FillInputBuffer(index, data, data_size)) |
| return MEDIA_CODEC_ERROR; |
| JNIEnv* env = AttachCurrentThread(); |
| return static_cast<MediaCodecStatus>(Java_MediaCodecBridge_queueInputBuffer( |
| env, j_bridge_, index, 0, data_size, presentation_time.InMicroseconds(), |
| 0)); |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::QueueSecureInputBuffer( |
| int index, |
| const uint8_t* data, |
| size_t data_size, |
| const std::string& key_id, |
| const std::string& iv, |
| const std::vector<SubsampleEntry>& subsamples, |
| const EncryptionScheme& encryption_scheme, |
| base::TimeDelta presentation_time) { |
| DVLOG(3) << __func__ << " " << index << ": " << data_size; |
| if (data_size > |
| base::checked_cast<size_t>(std::numeric_limits<int32_t>::max())) { |
| return MEDIA_CODEC_ERROR; |
| } |
| if (data && !FillInputBuffer(index, data, data_size)) |
| return MEDIA_CODEC_ERROR; |
| |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jbyteArray> j_key_id = ToJavaByteArray( |
| env, reinterpret_cast<const uint8_t*>(key_id.data()), key_id.size()); |
| ScopedJavaLocalRef<jbyteArray> j_iv = ToJavaByteArray( |
| env, reinterpret_cast<const uint8_t*>(iv.data()), iv.size()); |
| |
| // The MediaCodec.CryptoInfo documentation says to pass NULL for |clear_array| |
| // to indicate that all data is encrypted. But it doesn't specify what |
| // |cypher_array| and |subsamples_size| should be in that case. We pass |
| // one subsample here just to be on the safe side. |
| int num_subsamples = std::max(static_cast<size_t>(1), subsamples.size()); |
| |
| std::unique_ptr<jint[]> native_clear_array(new jint[num_subsamples]); |
| std::unique_ptr<jint[]> native_cypher_array(new jint[num_subsamples]); |
| |
| if (subsamples.empty()) { |
| native_clear_array[0] = 0; |
| native_cypher_array[0] = data_size; |
| } else { |
| for (size_t i = 0; i < subsamples.size(); ++i) { |
| DCHECK(subsamples[i].clear_bytes <= std::numeric_limits<uint16_t>::max()); |
| if (subsamples[i].cypher_bytes > |
| static_cast<uint32_t>(std::numeric_limits<jint>::max())) { |
| return MEDIA_CODEC_ERROR; |
| } |
| |
| native_clear_array[i] = subsamples[i].clear_bytes; |
| native_cypher_array[i] = subsamples[i].cypher_bytes; |
| } |
| } |
| |
| ScopedJavaLocalRef<jintArray> clear_array = base::android::ToJavaIntArray( |
| env, native_clear_array.get(), num_subsamples); |
| ScopedJavaLocalRef<jintArray> cypher_array = base::android::ToJavaIntArray( |
| env, native_cypher_array.get(), num_subsamples); |
| |
| return static_cast<MediaCodecStatus>( |
| Java_MediaCodecBridge_queueSecureInputBuffer( |
| env, j_bridge_, index, 0, j_iv, j_key_id, clear_array, cypher_array, |
| num_subsamples, static_cast<int>(encryption_scheme.mode()), |
| static_cast<int>(encryption_scheme.pattern().crypt_byte_block()), |
| static_cast<int>(encryption_scheme.pattern().skip_byte_block()), |
| presentation_time.InMicroseconds())); |
| } |
| |
| void MediaCodecBridgeImpl::QueueEOS(int input_buffer_index) { |
| DVLOG(3) << __func__ << ": " << input_buffer_index; |
| JNIEnv* env = AttachCurrentThread(); |
| Java_MediaCodecBridge_queueInputBuffer(env, j_bridge_, input_buffer_index, 0, |
| 0, 0, kBufferFlagEndOfStream); |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::DequeueInputBuffer( |
| base::TimeDelta timeout, |
| int* index) { |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jobject> result = Java_MediaCodecBridge_dequeueInputBuffer( |
| env, j_bridge_, timeout.InMicroseconds()); |
| *index = Java_DequeueInputResult_index(env, result); |
| MediaCodecStatus status = static_cast<MediaCodecStatus>( |
| Java_DequeueInputResult_status(env, result)); |
| DVLOG(3) << __func__ << ": status: " << status << ", index: " << *index; |
| return status; |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::DequeueOutputBuffer( |
| base::TimeDelta timeout, |
| int* index, |
| size_t* offset, |
| size_t* size, |
| base::TimeDelta* presentation_time, |
| bool* end_of_stream, |
| bool* key_frame) { |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jobject> result = |
| Java_MediaCodecBridge_dequeueOutputBuffer(env, j_bridge_, |
| timeout.InMicroseconds()); |
| *index = Java_DequeueOutputResult_index(env, result); |
| *offset = |
| base::checked_cast<size_t>(Java_DequeueOutputResult_offset(env, result)); |
| *size = base::checked_cast<size_t>( |
| Java_DequeueOutputResult_numBytes(env, result)); |
| if (presentation_time) { |
| *presentation_time = base::TimeDelta::FromMicroseconds( |
| Java_DequeueOutputResult_presentationTimeMicroseconds(env, result)); |
| } |
| int flags = Java_DequeueOutputResult_flags(env, result); |
| if (end_of_stream) |
| *end_of_stream = flags & kBufferFlagEndOfStream; |
| if (key_frame) |
| *key_frame = flags & kBufferFlagSyncFrame; |
| MediaCodecStatus status = static_cast<MediaCodecStatus>( |
| Java_DequeueOutputResult_status(env, result)); |
| DVLOG(3) << __func__ << ": status: " << status << ", index: " << *index |
| << ", offset: " << *offset << ", size: " << *size |
| << ", flags: " << flags; |
| return status; |
| } |
| |
| void MediaCodecBridgeImpl::ReleaseOutputBuffer(int index, bool render) { |
| DVLOG(3) << __func__ << ": " << index; |
| JNIEnv* env = AttachCurrentThread(); |
| Java_MediaCodecBridge_releaseOutputBuffer(env, j_bridge_, index, render); |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::GetInputBuffer(int input_buffer_index, |
| uint8_t** data, |
| size_t* capacity) { |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jobject> j_buffer( |
| Java_MediaCodecBridge_getInputBuffer(env, j_bridge_, input_buffer_index)); |
| if (j_buffer.is_null()) |
| return MEDIA_CODEC_ERROR; |
| |
| *data = static_cast<uint8_t*>(env->GetDirectBufferAddress(j_buffer.obj())); |
| *capacity = |
| base::checked_cast<size_t>(env->GetDirectBufferCapacity(j_buffer.obj())); |
| return MEDIA_CODEC_OK; |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::CopyFromOutputBuffer(int index, |
| size_t offset, |
| void* dst, |
| size_t num) { |
| const uint8_t* src_data = nullptr; |
| size_t src_capacity = 0; |
| MediaCodecStatus status = |
| GetOutputBufferAddress(index, offset, &src_data, &src_capacity); |
| if (status == MEDIA_CODEC_OK) { |
| CHECK_GE(src_capacity, num); |
| memcpy(dst, src_data, num); |
| } |
| return status; |
| } |
| |
| MediaCodecStatus MediaCodecBridgeImpl::GetOutputBufferAddress( |
| int index, |
| size_t offset, |
| const uint8_t** addr, |
| size_t* capacity) { |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jobject> j_buffer( |
| Java_MediaCodecBridge_getOutputBuffer(env, j_bridge_, index)); |
| if (j_buffer.is_null()) |
| return MEDIA_CODEC_ERROR; |
| const size_t total_capacity = env->GetDirectBufferCapacity(j_buffer.obj()); |
| CHECK_GE(total_capacity, offset); |
| *addr = reinterpret_cast<const uint8_t*>( |
| env->GetDirectBufferAddress(j_buffer.obj())) + |
| offset; |
| *capacity = total_capacity - offset; |
| return MEDIA_CODEC_OK; |
| } |
| |
| std::string MediaCodecBridgeImpl::GetName() { |
| if (base::android::BuildInfo::GetInstance()->sdk_int() < |
| base::android::SDK_VERSION_JELLY_BEAN_MR2) |
| return ""; |
| JNIEnv* env = AttachCurrentThread(); |
| ScopedJavaLocalRef<jstring> j_name = |
| Java_MediaCodecBridge_getName(env, j_bridge_); |
| return ConvertJavaStringToUTF8(env, j_name); |
| } |
| |
| bool MediaCodecBridgeImpl::SetSurface(const JavaRef<jobject>& surface) { |
| DCHECK_GE(base::android::BuildInfo::GetInstance()->sdk_int(), |
| base::android::SDK_VERSION_MARSHMALLOW); |
| JNIEnv* env = AttachCurrentThread(); |
| return Java_MediaCodecBridge_setSurface(env, j_bridge_, surface); |
| } |
| |
| void MediaCodecBridgeImpl::SetVideoBitrate(int bps, int frame_rate) { |
| JNIEnv* env = AttachCurrentThread(); |
| Java_MediaCodecBridge_setVideoBitrate(env, j_bridge_, bps, frame_rate); |
| } |
| |
| void MediaCodecBridgeImpl::RequestKeyFrameSoon() { |
| JNIEnv* env = AttachCurrentThread(); |
| Java_MediaCodecBridge_requestKeyFrameSoon(env, j_bridge_); |
| } |
| |
| bool MediaCodecBridgeImpl::FillInputBuffer(int index, |
| const uint8_t* data, |
| size_t size) { |
| uint8_t* dst = nullptr; |
| size_t capacity = 0; |
| if (GetInputBuffer(index, &dst, &capacity) != MEDIA_CODEC_OK) { |
| LOG(ERROR) << "GetInputBuffer failed"; |
| return false; |
| } |
| CHECK(dst); |
| |
| if (size > capacity) { |
| LOG(ERROR) << "Input buffer size " << size |
| << " exceeds MediaCodec input buffer capacity: " << capacity; |
| return false; |
| } |
| |
| memcpy(dst, data, size); |
| return true; |
| } |
| |
| } // namespace media |