| // Copyright 2020 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include <algorithm> |
| #include <numeric> |
| #include <vector> |
| |
| #include "base/command_line.h" |
| #include "base/files/file_util.h" |
| #include "base/json/json_writer.h" |
| #include "base/strings/stringprintf.h" |
| #include "media/base/bitstream_buffer.h" |
| #include "media/base/test_data_util.h" |
| #include "media/gpu/test/video.h" |
| #include "media/gpu/test/video_encoder/video_encoder.h" |
| #include "media/gpu/test/video_encoder/video_encoder_client.h" |
| #include "media/gpu/test/video_encoder/video_encoder_test_environment.h" |
| #include "testing/gtest/include/gtest/gtest.h" |
| |
| namespace media { |
| namespace test { |
| |
| namespace { |
| |
| // Video encoder perf tests usage message. Make sure to also update the |
| // documentation under docs/media/gpu/video_encoder_perf_test_usage.md when |
| // making changes here. |
| // TODO(dstaessens): Add video_encoder_perf_test_usage.md |
| constexpr const char* usage_msg = |
| "usage: video_encode_accelerator_perf_tests\n" |
| " [--codec=<codec>]\n" |
| " [-v=<level>] [--vmodule=<config>] [--output_folder]\n" |
| " [--gtest_help] [--help]\n" |
| " [<video path>] [<video metadata path>]\n"; |
| |
| // Video encoder performance tests help message. |
| constexpr const char* help_msg = |
| "Run the video encode accelerator performance tests on the video\n" |
| "specified by <video path>. If no <video path> is given the default\n" |
| "\"bear_320x192_40frames.yuv.webm\" video will be used.\n" |
| "\nThe <video metadata path> should specify the location of a json file\n" |
| "containing the video's metadata. By default <video path>.json will be\n" |
| "used.\n" |
| "\nThe following arguments are supported:\n" |
| " --codec codec profile to encode, \"h264 (baseline)\",\n" |
| " \"h264main, \"h264high\", \"vp8\" and \"vp9\"\n" |
| " -v enable verbose mode, e.g. -v=2.\n" |
| " --vmodule enable verbose mode for the specified module,\n" |
| " --output_folder overwrite the output folder used to store\n" |
| " performance metrics, if not specified results\n" |
| " will be stored in the current working directory.\n" |
| " --gtest_help display the gtest help and exit.\n" |
| " --help display this help and exit.\n"; |
| |
| // Default video to be used if no test video was specified. |
| constexpr base::FilePath::CharType kDefaultTestVideoPath[] = |
| FILE_PATH_LITERAL("bear_320x192_40frames.yuv.webm"); |
| |
| media::test::VideoEncoderTestEnvironment* g_env; |
| |
| // Default output folder used to store performance metrics. |
| constexpr const base::FilePath::CharType* kDefaultOutputFolder = |
| FILE_PATH_LITERAL("perf_metrics"); |
| |
| // Struct storing various time-related statistics. |
| struct PerformanceTimeStats { |
| PerformanceTimeStats() {} |
| explicit PerformanceTimeStats(const std::vector<double>& times); |
| double avg_ms_ = 0.0; |
| double percentile_25_ms_ = 0.0; |
| double percentile_50_ms_ = 0.0; |
| double percentile_75_ms_ = 0.0; |
| }; |
| |
| PerformanceTimeStats::PerformanceTimeStats(const std::vector<double>& times) { |
| if (times.empty()) |
| return; |
| |
| avg_ms_ = std::accumulate(times.begin(), times.end(), 0.0) / times.size(); |
| std::vector<double> sorted_times = times; |
| std::sort(sorted_times.begin(), sorted_times.end()); |
| percentile_25_ms_ = sorted_times[sorted_times.size() / 4]; |
| percentile_50_ms_ = sorted_times[sorted_times.size() / 2]; |
| percentile_75_ms_ = sorted_times[(sorted_times.size() * 3) / 4]; |
| } |
| |
| // TODO(dstaessens): Investigate using more appropriate metrics for encoding. |
| struct PerformanceMetrics { |
| // Write the collected performance metrics to the console. |
| void WriteToConsole() const; |
| // Write the collected performance metrics to file. |
| void WriteToFile() const; |
| |
| // Total measurement duration. |
| base::TimeDelta total_duration_; |
| // The number of bitstreams encoded. |
| size_t bitstreams_encoded_ = 0; |
| // The overall number of bitstreams encoded per second. |
| double bitstreams_per_second_ = 0.0; |
| // List of times between subsequent bitstream buffer deliveries. This is |
| // important in real-time encoding scenarios, where the delivery time should |
| // be less than the frame rate used. |
| std::vector<double> bitstream_delivery_times_; |
| // Statistics related to the time between bitstream buffer deliveries. |
| PerformanceTimeStats bitstream_delivery_stats_; |
| // List of times between queuing an encode operation and getting back the |
| // encoded bitstream buffer. |
| std::vector<double> bitstream_encode_times_; |
| // Statistics related to the encode times. |
| PerformanceTimeStats bitstream_encode_stats_; |
| }; |
| |
| // The performance evaluator can be plugged into the video encoder to collect |
| // various performance metrics. |
| class PerformanceEvaluator : public BitstreamProcessor { |
| public: |
| // Create a new performance evaluator. |
| PerformanceEvaluator() {} |
| |
| void ProcessBitstream(scoped_refptr<BitstreamRef> bitstream, |
| size_t frame_index) override; |
| bool WaitUntilDone() override { return true; } |
| |
| // Start/Stop collecting performance metrics. |
| void StartMeasuring(); |
| void StopMeasuring(); |
| |
| // Get the collected performance metrics. |
| const PerformanceMetrics& Metrics() const { return perf_metrics_; } |
| |
| private: |
| // Start/end time of the measurement period. |
| base::TimeTicks start_time_; |
| base::TimeTicks end_time_; |
| |
| // Time at which the previous bitstream was delivered. |
| base::TimeTicks prev_bitstream_delivery_time_; |
| |
| // Collection of various performance metrics. |
| PerformanceMetrics perf_metrics_; |
| }; |
| |
| void PerformanceEvaluator::ProcessBitstream( |
| scoped_refptr<BitstreamRef> bitstream, |
| size_t frame_index) { |
| base::TimeTicks now = base::TimeTicks::Now(); |
| |
| base::TimeDelta delivery_time = (now - prev_bitstream_delivery_time_); |
| perf_metrics_.bitstream_delivery_times_.push_back( |
| delivery_time.InMillisecondsF()); |
| prev_bitstream_delivery_time_ = now; |
| |
| base::TimeDelta encode_time = |
| now.since_origin() - bitstream->metadata.timestamp; |
| perf_metrics_.bitstream_encode_times_.push_back( |
| encode_time.InMillisecondsF()); |
| } |
| |
| void PerformanceEvaluator::StartMeasuring() { |
| start_time_ = base::TimeTicks::Now(); |
| prev_bitstream_delivery_time_ = start_time_; |
| } |
| |
| void PerformanceEvaluator::StopMeasuring() { |
| DCHECK_EQ(perf_metrics_.bitstream_delivery_times_.size(), |
| perf_metrics_.bitstream_encode_times_.size()); |
| |
| end_time_ = base::TimeTicks::Now(); |
| perf_metrics_.total_duration_ = end_time_ - start_time_; |
| perf_metrics_.bitstreams_encoded_ = |
| perf_metrics_.bitstream_encode_times_.size(); |
| perf_metrics_.bitstreams_per_second_ = |
| perf_metrics_.bitstreams_encoded_ / |
| perf_metrics_.total_duration_.InSecondsF(); |
| |
| // Calculate delivery and encode time metrics. |
| perf_metrics_.bitstream_delivery_stats_ = |
| PerformanceTimeStats(perf_metrics_.bitstream_delivery_times_); |
| perf_metrics_.bitstream_encode_stats_ = |
| PerformanceTimeStats(perf_metrics_.bitstream_encode_times_); |
| } |
| |
| void PerformanceMetrics::WriteToConsole() const { |
| std::cout << "Bitstreams encoded: " << bitstreams_encoded_ << std::endl; |
| std::cout << "Total duration: " << total_duration_.InMillisecondsF() |
| << "ms" << std::endl; |
| std::cout << "FPS: " << bitstreams_per_second_ |
| << std::endl; |
| std::cout << "Bitstream delivery time - average: " |
| << bitstream_delivery_stats_.avg_ms_ << "ms" << std::endl; |
| std::cout << "Bitstream delivery time - percentile 25: " |
| << bitstream_delivery_stats_.percentile_25_ms_ << "ms" << std::endl; |
| std::cout << "Bitstream delivery time - percentile 50: " |
| << bitstream_delivery_stats_.percentile_50_ms_ << "ms" << std::endl; |
| std::cout << "Bitstream delivery time - percentile 75: " |
| << bitstream_delivery_stats_.percentile_75_ms_ << "ms" << std::endl; |
| std::cout << "Bitstream encode time - average: " |
| << bitstream_encode_stats_.avg_ms_ << "ms" << std::endl; |
| std::cout << "Bitstream encode time - percentile 25: " |
| << bitstream_encode_stats_.percentile_25_ms_ << "ms" << std::endl; |
| std::cout << "Bitstream encode time - percentile 50: " |
| << bitstream_encode_stats_.percentile_50_ms_ << "ms" << std::endl; |
| std::cout << "Bitstream encode time - percentile 75: " |
| << bitstream_encode_stats_.percentile_75_ms_ << "ms" << std::endl; |
| } |
| |
| void PerformanceMetrics::WriteToFile() const { |
| base::FilePath output_folder_path = base::FilePath(g_env->OutputFolder()); |
| if (!DirectoryExists(output_folder_path)) |
| base::CreateDirectory(output_folder_path); |
| output_folder_path = base::MakeAbsoluteFilePath(output_folder_path); |
| |
| // Write performance metrics to json. |
| base::Value metrics(base::Value::Type::DICTIONARY); |
| metrics.SetKey("BitstreamsEncoded", |
| base::Value(base::checked_cast<int>(bitstreams_encoded_))); |
| metrics.SetKey("TotalDurationMs", |
| base::Value(total_duration_.InMillisecondsF())); |
| metrics.SetKey("FPS", base::Value(bitstreams_per_second_)); |
| metrics.SetKey("BitstreamDeliveryTimeAverage", |
| base::Value(bitstream_delivery_stats_.avg_ms_)); |
| metrics.SetKey("BitstreamDeliveryTimePercentile25", |
| base::Value(bitstream_delivery_stats_.percentile_25_ms_)); |
| metrics.SetKey("BitstreamDeliveryTimePercentile50", |
| base::Value(bitstream_delivery_stats_.percentile_50_ms_)); |
| metrics.SetKey("BitstreamDeliveryTimePercentile75", |
| base::Value(bitstream_delivery_stats_.percentile_75_ms_)); |
| metrics.SetKey("BitstreamEncodeTimeAverage", |
| base::Value(bitstream_encode_stats_.avg_ms_)); |
| metrics.SetKey("BitstreamEncodeTimePercentile25", |
| base::Value(bitstream_encode_stats_.percentile_25_ms_)); |
| metrics.SetKey("BitstreamEncodeTimePercentile50", |
| base::Value(bitstream_encode_stats_.percentile_50_ms_)); |
| metrics.SetKey("BitstreamEncodeTimePercentile75", |
| base::Value(bitstream_encode_stats_.percentile_75_ms_)); |
| |
| // Write bitstream delivery times to json. |
| base::Value delivery_times(base::Value::Type::LIST); |
| for (double bitstream_delivery_time : bitstream_delivery_times_) { |
| delivery_times.Append(bitstream_delivery_time); |
| } |
| metrics.SetKey("BitstreamDeliveryTimes", std::move(delivery_times)); |
| |
| // Write bitstream encodes times to json. |
| base::Value encode_times(base::Value::Type::LIST); |
| for (double bitstream_encode_time : bitstream_encode_times_) { |
| encode_times.Append(bitstream_encode_time); |
| } |
| metrics.SetKey("BitstreamEncodeTimes", std::move(encode_times)); |
| |
| // Write json to file. |
| std::string metrics_str; |
| ASSERT_TRUE(base::JSONWriter::WriteWithOptions( |
| metrics, base::JSONWriter::OPTIONS_PRETTY_PRINT, &metrics_str)); |
| base::FilePath metrics_file_path = output_folder_path.Append( |
| g_env->GetTestOutputFilePath().AddExtension(FILE_PATH_LITERAL(".json"))); |
| // Make sure that the directory into which json is saved is created. |
| LOG_ASSERT(base::CreateDirectory(metrics_file_path.DirName())); |
| base::File metrics_output_file( |
| base::FilePath(metrics_file_path), |
| base::File::FLAG_CREATE_ALWAYS | base::File::FLAG_WRITE); |
| int bytes_written = metrics_output_file.WriteAtCurrentPos( |
| metrics_str.data(), metrics_str.length()); |
| ASSERT_EQ(bytes_written, static_cast<int>(metrics_str.length())); |
| VLOG(0) << "Wrote performance metrics to: " << metrics_file_path; |
| } |
| |
| // Video encode test class. Performs setup and teardown for each single test. |
| class VideoEncoderTest : public ::testing::Test { |
| public: |
| // Create a new video encoder instance. |
| std::unique_ptr<VideoEncoder> CreateVideoEncoder(const Video* video, |
| VideoCodecProfile profile) { |
| LOG_ASSERT(video); |
| |
| std::vector<std::unique_ptr<BitstreamProcessor>> bitstream_processors; |
| auto performance_evaluator = std::make_unique<PerformanceEvaluator>(); |
| performance_evaluator_ = performance_evaluator.get(); |
| bitstream_processors.push_back(std::move(performance_evaluator)); |
| |
| VideoEncoderClientConfig config; |
| config.framerate = video->FrameRate(); |
| config.output_profile = profile; |
| auto video_encoder = |
| VideoEncoder::Create(config, std::move(bitstream_processors)); |
| LOG_ASSERT(video_encoder); |
| LOG_ASSERT(video_encoder->Initialize(video)); |
| |
| return video_encoder; |
| } |
| |
| PerformanceEvaluator* performance_evaluator_; |
| }; |
| |
| } // namespace |
| |
| // Encode video from start to end while measuring uncapped performance. This |
| // test will encode a video as fast as possible, and gives an idea about the |
| // maximum output of the encoder. |
| TEST_F(VideoEncoderTest, MeasureUncappedPerformance) { |
| auto encoder = CreateVideoEncoder(g_env->Video(), g_env->Profile()); |
| |
| performance_evaluator_->StartMeasuring(); |
| encoder->Encode(); |
| EXPECT_TRUE(encoder->WaitForFlushDone()); |
| performance_evaluator_->StopMeasuring(); |
| |
| auto metrics = performance_evaluator_->Metrics(); |
| metrics.WriteToConsole(); |
| metrics.WriteToFile(); |
| |
| EXPECT_EQ(encoder->GetFlushDoneCount(), 1u); |
| EXPECT_EQ(encoder->GetFrameReleasedCount(), g_env->Video()->NumFrames()); |
| } |
| |
| } // namespace test |
| } // namespace media |
| |
| int main(int argc, char** argv) { |
| // Set the default test data path. |
| media::test::Video::SetTestDataPath(media::GetTestDataPath()); |
| |
| // Print the help message if requested. This needs to be done before |
| // initializing gtest, to overwrite the default gtest help message. |
| base::CommandLine::Init(argc, argv); |
| const base::CommandLine* cmd_line = base::CommandLine::ForCurrentProcess(); |
| LOG_ASSERT(cmd_line); |
| if (cmd_line->HasSwitch("help")) { |
| std::cout << media::test::usage_msg << "\n" << media::test::help_msg; |
| return 0; |
| } |
| |
| // Check if a video was specified on the command line. |
| base::CommandLine::StringVector args = cmd_line->GetArgs(); |
| base::FilePath video_path = |
| (args.size() >= 1) ? base::FilePath(args[0]) |
| : base::FilePath(media::test::kDefaultTestVideoPath); |
| base::FilePath video_metadata_path = |
| (args.size() >= 2) ? base::FilePath(args[1]) : base::FilePath(); |
| std::string codec = "h264"; |
| |
| // Parse command line arguments. |
| base::FilePath::StringType output_folder = media::test::kDefaultOutputFolder; |
| base::CommandLine::SwitchMap switches = cmd_line->GetSwitches(); |
| for (base::CommandLine::SwitchMap::const_iterator it = switches.begin(); |
| it != switches.end(); ++it) { |
| if (it->first.find("gtest_") == 0 || // Handled by GoogleTest |
| it->first == "v" || it->first == "vmodule") { // Handled by Chrome |
| continue; |
| } |
| |
| if (it->first == "output_folder") { |
| output_folder = it->second; |
| } else if (it->first == "codec") { |
| codec = it->second; |
| } else { |
| std::cout << "unknown option: --" << it->first << "\n" |
| << media::test::usage_msg; |
| return EXIT_FAILURE; |
| } |
| } |
| |
| testing::InitGoogleTest(&argc, argv); |
| |
| // Set up our test environment. |
| media::test::VideoEncoderTestEnvironment* test_environment = |
| media::test::VideoEncoderTestEnvironment::Create( |
| video_path, video_metadata_path, false, base::FilePath(output_folder), |
| codec, false /* output_bitstream */); |
| if (!test_environment) |
| return EXIT_FAILURE; |
| |
| media::test::g_env = static_cast<media::test::VideoEncoderTestEnvironment*>( |
| testing::AddGlobalTestEnvironment(test_environment)); |
| |
| return RUN_ALL_TESTS(); |
| } |