1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762
|
// Copyright 2019 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifdef UNSAFE_BUFFERS_BUILD
// TODO(crbug.com/40285824): Remove this and spanify to fix the errors.
#pragma allow_unsafe_buffers
#endif
#include "media/gpu/v4l2/v4l2_video_decoder_backend_stateless.h"
#include <fcntl.h>
#include <linux/media.h>
#include <sys/ioctl.h>
#include <memory>
#include "base/containers/contains.h"
#include "base/functional/bind.h"
#include "base/functional/callback_helpers.h"
#include "base/metrics/histogram_macros.h"
#include "base/not_fatal_until.h"
#include "base/numerics/safe_conversions.h"
#include "base/posix/eintr_wrapper.h"
#include "base/task/sequenced_task_runner.h"
#include "base/trace_event/trace_event.h"
#include "media/base/decoder_status.h"
#include "media/base/video_codecs.h"
#include "media/base/video_frame.h"
#include "media/gpu/accelerated_video_decoder.h"
#include "media/gpu/chromeos/dmabuf_video_frame_pool.h"
#include "media/gpu/chromeos/video_frame_resource.h"
#include "media/gpu/macros.h"
#include "media/gpu/v4l2/v4l2_device.h"
#include "media/gpu/v4l2/v4l2_video_decoder_delegate_av1.h"
#include "media/gpu/v4l2/v4l2_video_decoder_delegate_h264.h"
#include "media/gpu/v4l2/v4l2_video_decoder_delegate_h265.h"
#include "media/gpu/v4l2/v4l2_video_decoder_delegate_vp8.h"
#include "media/gpu/v4l2/v4l2_video_decoder_delegate_vp9.h"
namespace media {
namespace {
// Size of the timestamp cache, needs to be large enough for frame-reordering.
constexpr size_t kTimestampCacheSize = 128;
// Number of requests to allocate for submitting input buffers, if requests
// are used.
} // namespace
struct V4L2StatelessVideoDecoderBackend::OutputRequest {
static OutputRequest Surface(scoped_refptr<V4L2DecodeSurface> s,
base::TimeDelta t) {
return OutputRequest(std::move(s), t);
}
static OutputRequest FlushFence() { return OutputRequest(kFlushFence); }
static OutputRequest ChangeResolutionFence() {
return OutputRequest(kChangeResolutionFence);
}
OutputRequest(const OutputRequest&) = delete;
OutputRequest& operator=(const OutputRequest&) = delete;
bool IsReady() const {
return (type != OutputRequestType::kSurface) || surface->decoded();
}
// Allow move, but not copy.
OutputRequest(OutputRequest&&) = default;
enum OutputRequestType {
// The surface to be outputted.
kSurface,
// The fence to indicate the flush request.
kFlushFence,
// The fence to indicate resolution change request.
kChangeResolutionFence,
};
// The type of the request.
const OutputRequestType type;
// The surface to be outputted.
scoped_refptr<V4L2DecodeSurface> surface;
// The timestamp of the output frame. Because a surface might be outputted
// multiple times with different timestamp, we need to store timestamp out of
// surface.
base::TimeDelta timestamp;
private:
OutputRequest(scoped_refptr<V4L2DecodeSurface> s, base::TimeDelta t)
: type(kSurface), surface(std::move(s)), timestamp(t) {}
explicit OutputRequest(OutputRequestType t) : type(t) {}
};
V4L2StatelessVideoDecoderBackend::DecodeRequest::DecodeRequest(
scoped_refptr<DecoderBuffer> buf,
VideoDecoder::DecodeCB cb,
int32_t id)
: buffer(std::move(buf)), decode_cb(std::move(cb)), bitstream_id(id) {}
V4L2StatelessVideoDecoderBackend::DecodeRequest::DecodeRequest(
DecodeRequest&&) = default;
V4L2StatelessVideoDecoderBackend::DecodeRequest&
V4L2StatelessVideoDecoderBackend::DecodeRequest::operator=(DecodeRequest&&) =
default;
V4L2StatelessVideoDecoderBackend::DecodeRequest::~DecodeRequest() = default;
V4L2StatelessVideoDecoderBackend::V4L2StatelessVideoDecoderBackend(
Client* const client,
scoped_refptr<V4L2Device> device,
VideoCodecProfile profile,
const VideoColorSpace& color_space,
scoped_refptr<base::SequencedTaskRunner> task_runner,
CdmContext* cdm_context)
: V4L2VideoDecoderBackend(client, std::move(device)),
profile_(profile),
color_space_(color_space),
bitstream_id_to_timestamp_(kTimestampCacheSize),
task_runner_(task_runner),
cdm_context_(cdm_context) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
weak_this_ = weak_this_factory_.GetWeakPtr();
}
V4L2StatelessVideoDecoderBackend::~V4L2StatelessVideoDecoderBackend() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
LOG_IF(WARNING, !surfaces_at_device_.empty())
<< "There is/are " << surfaces_at_device_.size()
<< " pending CAPTURE queue buffers pending dequeuing. This might be "
<< "fine or a problem depending on the destruction semantics (of the "
<< "client code).";
if (!output_request_queue_.empty() || flush_cb_ || current_decode_request_ ||
!decode_request_queue_.empty()) {
VLOGF(1) << "Should not destroy backend during pending decode!";
}
}
bool V4L2StatelessVideoDecoderBackend::Initialize() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
if (!IsSupportedProfile(profile_)) {
VLOGF(1) << "Unsupported profile " << GetProfileName(profile_);
return false;
}
if (!CreateDecoder())
return false;
CHECK(input_queue_->SupportsRequests());
requests_queue_ = device_->GetRequestsQueue();
return !!requests_queue_;
}
// static
void V4L2StatelessVideoDecoderBackend::ReuseOutputBufferThunk(
scoped_refptr<base::SequencedTaskRunner> task_runner,
std::optional<base::WeakPtr<V4L2StatelessVideoDecoderBackend>> weak_this,
V4L2ReadableBufferRef buffer) {
DVLOGF(3);
DCHECK(weak_this);
if (task_runner->RunsTasksInCurrentSequence()) {
if (*weak_this) {
(*weak_this)->ReuseOutputBuffer(std::move(buffer));
}
} else {
task_runner->PostTask(
FROM_HERE,
base::BindOnce(&V4L2StatelessVideoDecoderBackend::ReuseOutputBuffer,
*weak_this, std::move(buffer)));
}
}
void V4L2StatelessVideoDecoderBackend::ReuseOutputBuffer(
V4L2ReadableBufferRef buffer) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3) << "Reuse output surface #" << buffer->BufferId();
// Resume decoding in case of ran out of surface.
if (pause_reason_ == PauseReason::kRanOutOfSurfaces) {
pause_reason_ = PauseReason::kNone;
task_runner_->PostTask(
FROM_HERE,
base::BindOnce(&V4L2StatelessVideoDecoderBackend::DoDecodeWork,
weak_this_));
}
}
void V4L2StatelessVideoDecoderBackend::OnOutputBufferDequeued(
V4L2ReadableBufferRef dequeued_buffer) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
// Mark the output buffer decoded, and try to output surface.
DCHECK(!surfaces_at_device_.empty());
auto surface = std::move(surfaces_at_device_.front());
DCHECK_EQ(static_cast<size_t>(surface->output_record()),
dequeued_buffer->BufferId());
surfaces_at_device_.pop();
surface->SetDecoded();
auto reuse_buffer_cb =
base::BindOnce(&V4L2StatelessVideoDecoderBackend::ReuseOutputBufferThunk,
task_runner_, weak_this_, std::move(dequeued_buffer));
if (output_queue_->GetMemoryType() == V4L2_MEMORY_MMAP) {
// Keep a reference to the V4L2 buffer until the frame is reused, because
// the frame is backed up by the V4L2 buffer's memory.
surface->frame()->AddDestructionObserver(std::move(reuse_buffer_cb));
} else {
// Keep a reference to the V4L2 buffer until the buffer is reused. The
// reason for this is that the we currently use V4L2 buffer IDs to generate
// timestamps to
// reference frames, therefore we cannot reuse the same V4L2 buffer ID for
// another decode operation until all references to that frame are gone.
surface->SetReleaseCallback(std::move(reuse_buffer_cb));
}
PumpOutputSurfaces();
}
scoped_refptr<V4L2DecodeSurface>
V4L2StatelessVideoDecoderBackend::CreateSecureSurface(uint64_t secure_handle) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(4);
// Request V4L2 input and output buffers.
auto input_buf = input_queue_->GetFreeBuffer();
auto output_buf = output_queue_->GetFreeBuffer();
if (!input_buf || !output_buf) {
DVLOGF(3) << "There is no free V4L2 buffer.";
return nullptr;
}
DmabufVideoFramePool* pool = client_->GetVideoFramePool();
scoped_refptr<FrameResource> frame;
if (!pool) {
// Get FrameResource from the V4L2 buffer because now we allocate from V4L2
// driver via MMAP. The FrameResource received from V4L2 buffer will remain
// until deallocating V4L2Queue. But we need to know when the buffer is not
// used by the client. So we wrap the frame here.
DCHECK_EQ(output_queue_->GetMemoryType(), V4L2_MEMORY_MMAP);
scoped_refptr<FrameResource> origin_frame = output_buf->GetFrameResource();
if (!origin_frame) {
LOG(ERROR) << "There is no available FrameResource from the V4L2 buffer.";
return nullptr;
}
frame = origin_frame->CreateWrappingFrame();
} else {
// This is used in cases when the video decoder format does not need
// conversion before being sent to Chrome's Media pipeline. On ChromeOS,
// currently only RK3399 (scarlet) supports this.
DCHECK_EQ(output_queue_->GetMemoryType(), V4L2_MEMORY_DMABUF);
frame = pool->GetFrame();
if (!frame) {
// We allocate the same number of output buffer slot in V4L2 device and
// the output FrameResource. If there is free output buffer slot but no
// free FrameResource, it means the FrameResource is not released at
// client side. Post DoDecodeWork when the pool has available frames.
DVLOGF(3) << "There is no available FrameResource.";
pool->NotifyWhenFrameAvailable(base::BindOnce(
base::IgnoreResult(&base::SequencedTaskRunner::PostTask),
task_runner_, FROM_HERE,
base::BindOnce(&V4L2StatelessVideoDecoderBackend::DoDecodeWork,
weak_this_)));
return nullptr;
}
}
scoped_refptr<V4L2DecodeSurface> dec_surface;
CHECK(input_queue_->SupportsRequests());
std::optional<V4L2RequestRef> request_ref = requests_queue_->GetFreeRequest();
if (!request_ref) {
DVLOGF(1) << "Could not get free request.";
return nullptr;
}
return base::MakeRefCounted<V4L2RequestDecodeSurface>(
std::move(*input_buf), std::move(*output_buf), std::move(frame),
secure_handle, std::move(*request_ref));
}
scoped_refptr<V4L2DecodeSurface>
V4L2StatelessVideoDecoderBackend::CreateSurface() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(4);
return CreateSecureSurface(0);
}
bool V4L2StatelessVideoDecoderBackend::SubmitSlice(
V4L2DecodeSurface* dec_surface,
const uint8_t* data,
size_t size) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3);
size_t plane_size = dec_surface->input_buffer().GetPlaneSize(0);
size_t bytes_used = dec_surface->input_buffer().GetPlaneBytesUsed(0);
if (size > plane_size - bytes_used) {
LOG(ERROR) << "The size of submitted slice(" << size
<< ") is larger than the remaining buffer size("
<< plane_size - bytes_used << "). Plane size is " << plane_size;
client_->OnBackendError();
return false;
}
// Secure playback will submit a nullptr for |data|, the target data already
// will exist in the secure buffer.
if (data) {
void* mapping = dec_surface->input_buffer().GetPlaneMapping(0);
memcpy(reinterpret_cast<uint8_t*>(mapping) + bytes_used, data, size);
}
dec_surface->input_buffer().SetPlaneBytesUsed(0, bytes_used + size);
return true;
}
void V4L2StatelessVideoDecoderBackend::DecodeSurface(
scoped_refptr<V4L2DecodeSurface> dec_surface) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3);
DCHECK(current_decode_request_);
const auto timestamp = current_decode_request_->buffer->timestamp();
buffer_tracers_[timestamp] = std::make_unique<ScopedDecodeTrace>(
"V4L2VideoDecoderBackendStateless", *(current_decode_request_->buffer));
enqueuing_timestamps_[timestamp.InMilliseconds()] = base::TimeTicks::Now();
if (!dec_surface->Submit()) {
LOG(ERROR) << "Error while submitting frame for decoding!";
client_->OnBackendError();
return;
}
surfaces_at_device_.push(std::move(dec_surface));
}
void V4L2StatelessVideoDecoderBackend::SurfaceReady(
scoped_refptr<V4L2DecodeSurface> dec_surface,
int32_t bitstream_id,
const gfx::Rect& visible_rect,
const VideoColorSpace& color_space) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3);
// Find the timestamp associated with |bitstream_id|. It's possible that a
// surface is output multiple times for different |bitstream_id|s (e.g. VP9
// show_existing_frame feature). This means we need to output the same frame
// again with a different timestamp.
// On some rare occasions it's also possible that a single DecoderBuffer
// produces multiple surfaces with the same |bitstream_id|, so we shouldn't
// remove the timestamp from the cache.
const auto it = bitstream_id_to_timestamp_.Peek(bitstream_id);
CHECK(it != bitstream_id_to_timestamp_.end());
base::TimeDelta timestamp = it->second;
dec_surface->SetVisibleRect(visible_rect);
dec_surface->SetColorSpace(color_space);
output_request_queue_.push(
OutputRequest::Surface(std::move(dec_surface), timestamp));
PumpOutputSurfaces();
}
void V4L2StatelessVideoDecoderBackend::ResumeDecoding() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DoDecodeWork();
}
void V4L2StatelessVideoDecoderBackend::EnqueueDecodeTask(
scoped_refptr<DecoderBuffer> buffer,
VideoDecoder::DecodeCB decode_cb) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
const int32_t bitstream_id =
bitstream_id_generator_.GenerateNextId().GetUnsafeValue();
if (!buffer->end_of_stream())
bitstream_id_to_timestamp_.Put(bitstream_id, buffer->timestamp());
decode_request_queue_.push(
DecodeRequest(std::move(buffer), std::move(decode_cb), bitstream_id));
// If we are already decoding, then we don't need to pump again.
if (!current_decode_request_)
DoDecodeWork();
}
void V4L2StatelessVideoDecoderBackend::DoDecodeWork() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
if (!client_->IsDecoding())
return;
if (!PumpDecodeTask()) {
LOG(ERROR) << "Failed to do decode work.";
client_->OnBackendError();
}
}
bool V4L2StatelessVideoDecoderBackend::PumpDecodeTask() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3) << " Number of Decode requests: " << decode_request_queue_.size();
pause_reason_ = PauseReason::kNone;
while (true) {
switch (decoder_->Decode()) {
case AcceleratedVideoDecoder::kConfigChange:
if (decoder_->GetBitDepth() != 8u && decoder_->GetBitDepth() != 10u) {
VLOGF(2) << "Unsupported bit depth: "
<< base::strict_cast<int>(decoder_->GetBitDepth());
return false;
}
if (profile_ != decoder_->GetProfile()) {
DVLOGF(3) << "Profile is changed: " << profile_ << " -> "
<< decoder_->GetProfile();
if (!IsSupportedProfile(decoder_->GetProfile())) {
VLOGF(2) << "Unsupported profile: " << decoder_->GetProfile();
return false;
}
profile_ = decoder_->GetProfile();
}
if (pic_size_ == decoder_->GetPicSize()) {
// There is no need to do anything in V4L2 API when only a profile is
// changed.
DVLOGF(3) << "Only profile is changed. No need to do anything.";
continue;
}
DVLOGF(3) << "Need to change resolution. Pause decoding.";
client_->InitiateFlush();
output_request_queue_.push(OutputRequest::ChangeResolutionFence());
PumpOutputSurfaces();
return true;
case AcceleratedVideoDecoder::kRanOutOfStreamData:
// Current decode request is finished processing.
if (current_decode_request_) {
std::move(current_decode_request_->decode_cb)
.Run(DecoderStatus::Codes::kOk);
current_decode_request_ = std::nullopt;
}
// Process next decode request.
if (decode_request_queue_.empty())
return true;
current_decode_request_ = std::move(decode_request_queue_.front());
decode_request_queue_.pop();
if (current_decode_request_->buffer->end_of_stream()) {
if (!decoder_->Flush()) {
VLOGF(1) << "Failed flushing the decoder.";
return false;
}
// Put the decoder in an idle state, ready to resume.
decoder_->Reset();
client_->InitiateFlush();
DCHECK(!flush_cb_);
flush_cb_ = std::move(current_decode_request_->decode_cb);
output_request_queue_.push(OutputRequest::FlushFence());
PumpOutputSurfaces();
current_decode_request_ = std::nullopt;
return true;
}
decoder_->SetStream(current_decode_request_->bitstream_id,
*current_decode_request_->buffer);
break;
case AcceleratedVideoDecoder::kRanOutOfSurfaces:
DVLOGF(3) << "Ran out of surfaces. Resume when buffer is returned.";
pause_reason_ = PauseReason::kRanOutOfSurfaces;
return true;
case AcceleratedVideoDecoder::kDecodeError:
DVLOGF(3) << "Error decoding stream";
return false;
case AcceleratedVideoDecoder::kTryAgain:
// In this case we are waiting for an async operation relating to secure
// content. When that is complete, ResumeDecoding will be invoked and we
// will start decoding again; or a reset will occur and that will resume
// decoding.
return true;
}
}
}
void V4L2StatelessVideoDecoderBackend::PumpOutputSurfaces() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3) << "Number of display surfaces: " << output_request_queue_.size();
bool resume_decode = false;
while (!output_request_queue_.empty()) {
if (!output_request_queue_.front().IsReady()) {
DVLOGF(3) << "The first surface is not ready yet.";
// It is possible that that V4L2 buffers for this output surface are not
// even queued yet. Make sure that decoder_->Decode() is called to
// continue that work and prevent the decoding thread from starving.
resume_decode = true;
break;
}
OutputRequest request = std::move(output_request_queue_.front());
output_request_queue_.pop();
switch (request.type) {
case OutputRequest::kFlushFence:
DCHECK(output_request_queue_.empty());
DVLOGF(2) << "Flush finished.";
std::move(flush_cb_).Run(DecoderStatus::Codes::kOk);
resume_decode = true;
client_->CompleteFlush();
break;
case OutputRequest::kChangeResolutionFence:
DCHECK(output_request_queue_.empty());
ChangeResolution();
break;
case OutputRequest::kSurface:
scoped_refptr<V4L2DecodeSurface> surface = std::move(request.surface);
DCHECK(surface->frame());
client_->OutputFrame(surface->frame(), surface->visible_rect(),
surface->color_space(), request.timestamp);
{
const auto timestamp = surface->frame()->timestamp();
const auto flat_timestamp = timestamp.InMilliseconds();
// TODO(b/190615065) |flat_timestamp| might be repeated with H.264
// bitstreams, investigate why, and change the if() to DCHECK().
if (base::Contains(enqueuing_timestamps_, flat_timestamp)) {
const auto decoding_begin = enqueuing_timestamps_[flat_timestamp];
const auto decoding_end = base::TimeTicks::Now();
UMA_HISTOGRAM_TIMES("Media.PlatformVideoDecoding.Decode",
decoding_end - decoding_begin);
enqueuing_timestamps_.erase(flat_timestamp);
}
auto iter = buffer_tracers_.find(timestamp);
if (iter != buffer_tracers_.end()) {
iter->second->EndTrace(DecoderStatus::Codes::kOk);
buffer_tracers_.erase(iter);
}
}
break;
}
}
if (resume_decode) {
task_runner_->PostTask(
FROM_HERE,
base::BindOnce(&V4L2StatelessVideoDecoderBackend::DoDecodeWork,
weak_this_));
}
}
void V4L2StatelessVideoDecoderBackend::ChangeResolution() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
// We change resolution after outputting all pending surfaces, there should
// be no V4L2DecodeSurface left.
DCHECK(surfaces_at_device_.empty());
DCHECK(output_request_queue_.empty());
const size_t num_codec_reference_frames = decoder_->GetNumReferenceFrames();
// Verify |num_codec_reference_frames| has a reasonable value. Anecdotally 16
// is the largest amount of reference frames seen, on an ITU-T H.264 test
// vector (CAPCM*1_Sand_E.h264).
CHECK_LE(num_codec_reference_frames, 32u);
const gfx::Rect visible_rect = decoder_->GetVisibleRect();
const gfx::Size pic_size = decoder_->GetPicSize();
const uint8_t bit_depth = decoder_->GetBitDepth();
// Set output format with the new resolution.
DCHECK(!pic_size.IsEmpty());
DVLOGF(3) << "Change resolution to " << pic_size.ToString();
client_->ChangeResolution(pic_size, visible_rect, num_codec_reference_frames,
bit_depth);
}
bool V4L2StatelessVideoDecoderBackend::ApplyResolution(
const gfx::Size& pic_size,
const gfx::Rect& visible_rect) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DCHECK_EQ(input_queue_->QueuedBuffersCount(), 0u);
auto ret = input_queue_->GetFormat().first;
if (!ret) {
VPLOGF(1) << "Failed getting OUTPUT format";
return false;
}
struct v4l2_format format = std::move(*ret);
format.fmt.pix_mp.width = pic_size.width();
format.fmt.pix_mp.height = pic_size.height();
if (device_->Ioctl(VIDIOC_S_FMT, &format) != 0) {
RecordVidiocIoctlErrorUMA(VidiocIoctlRequests::kVidiocSFmt);
VPLOGF(1) << "Failed setting OUTPUT format";
return false;
}
return true;
}
void V4L2StatelessVideoDecoderBackend::OnChangeResolutionDone(
CroStatus status) {
if (status == CroStatus::Codes::kResetRequired)
return;
if (status != CroStatus::Codes::kOk) {
LOG(ERROR) << "Backend failure when changing resolution ("
<< static_cast<int>(status.code()) << ").";
client_->OnBackendError();
return;
}
pic_size_ = decoder_->GetPicSize();
client_->CompleteFlush();
task_runner_->PostTask(
FROM_HERE, base::BindOnce(&V4L2StatelessVideoDecoderBackend::DoDecodeWork,
weak_this_));
}
void V4L2StatelessVideoDecoderBackend::OnStreamStopped(bool stop_input_queue) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3);
// The V4L2 stream has been stopped stopped, so all surfaces on the device
// have been returned to the client.
surfaces_at_device_ = {};
}
void V4L2StatelessVideoDecoderBackend::ClearPendingRequests(
DecoderStatus status) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3);
if (decoder_) {
// If we reset during resolution change, re-create AVD. Then the new AVD
// will trigger resolution change again after reset.
if (pic_size_ != decoder_->GetPicSize()) {
CreateDecoder();
} else {
decoder_->Reset();
}
}
// Clear output_request_queue_.
while (!output_request_queue_.empty())
output_request_queue_.pop();
if (flush_cb_)
std::move(flush_cb_).Run(status);
// Clear current_decode_request_ and decode_request_queue_.
if (current_decode_request_) {
std::move(current_decode_request_->decode_cb).Run(status);
current_decode_request_ = std::nullopt;
}
while (!decode_request_queue_.empty()) {
auto request = std::move(decode_request_queue_.front());
decode_request_queue_.pop();
std::move(request.decode_cb).Run(status);
}
}
bool V4L2StatelessVideoDecoderBackend::StopInputQueueOnResChange() const {
return true;
}
size_t V4L2StatelessVideoDecoderBackend::GetNumOUTPUTQueueBuffers(
bool secure_mode) const {
// Some H.264 test vectors (CAPCM*1_Sand_E.h264) need 16 reference frames; add
// one to calculate the number of OUTPUT buffers, to account for the frame
// being decoded.
// For secure mode, we are very memory constrained so only allocate 8 buffers.
// TODO(b/249325255): reduce this number to e.g. 8 or even less when it does
// not artificially limit the size of the CAPTURE (decoded video frames)
// queue.
constexpr size_t kNumInputBuffers = 16 + 1;
constexpr size_t kNumInputBuffersSecureMode = 8;
return secure_mode ? kNumInputBuffersSecureMode : kNumInputBuffers;
}
bool V4L2StatelessVideoDecoderBackend::IsSupportedProfile(
VideoCodecProfile profile) {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DCHECK(device_);
if (supported_profiles_.empty()) {
const std::vector<uint32_t> kSupportedInputFourccs = {
V4L2_PIX_FMT_H264_SLICE,
#if BUILDFLAG(ENABLE_HEVC_PARSER_AND_HW_DECODER)
V4L2_PIX_FMT_HEVC_SLICE,
#endif
V4L2_PIX_FMT_VP8_FRAME,
V4L2_PIX_FMT_VP9_FRAME,
V4L2_PIX_FMT_AV1_FRAME,
};
auto device = base::MakeRefCounted<V4L2Device>();
VideoDecodeAccelerator::SupportedProfiles profiles =
device->GetSupportedDecodeProfiles(kSupportedInputFourccs);
for (const auto& entry : profiles)
supported_profiles_.push_back(entry.profile);
}
return base::Contains(supported_profiles_, profile);
}
bool V4L2StatelessVideoDecoderBackend::CreateDecoder() {
DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
DVLOGF(3);
pic_size_ = gfx::Size();
CHECK(input_queue_->SupportsRequests());
if (profile_ >= H264PROFILE_MIN && profile_ <= H264PROFILE_MAX) {
decoder_ = std::make_unique<H264Decoder>(
std::make_unique<V4L2VideoDecoderDelegateH264>(this, device_.get(),
cdm_context_),
profile_, color_space_);
#if BUILDFLAG(ENABLE_HEVC_PARSER_AND_HW_DECODER)
} else if (profile_ >= HEVCPROFILE_MIN && profile_ <= HEVCPROFILE_MAX) {
decoder_ = std::make_unique<H265Decoder>(
std::make_unique<V4L2VideoDecoderDelegateH265>(this, device_.get()),
profile_, color_space_);
#endif
} else if (profile_ >= VP8PROFILE_MIN && profile_ <= VP8PROFILE_MAX) {
decoder_ = std::make_unique<VP8Decoder>(
std::make_unique<V4L2VideoDecoderDelegateVP8>(this, device_.get()),
color_space_);
} else if (profile_ >= VP9PROFILE_MIN && profile_ <= VP9PROFILE_MAX) {
decoder_ = std::make_unique<VP9Decoder>(
std::make_unique<V4L2VideoDecoderDelegateVP9>(this, device_.get()),
profile_, color_space_);
#if BUILDFLAG(IS_CHROMEOS)
} else if (profile_ >= AV1PROFILE_MIN && profile_ <= AV1PROFILE_MAX) {
decoder_ = std::make_unique<AV1Decoder>(
std::make_unique<V4L2VideoDecoderDelegateAV1>(this, device_.get()),
profile_, color_space_);
#endif
} else {
VLOGF(1) << "Unsupported profile " << GetProfileName(profile_);
return false;
}
return true;
}
} // namespace media
|