Index: media/gpu/v4l2_jpeg_decode_accelerator.cc |
diff --git a/media/gpu/v4l2_jpeg_decode_accelerator.cc b/media/gpu/v4l2_jpeg_decode_accelerator.cc |
index e8f4cc6ac7a088a185fef86c540f2a1c75657cab..b3109191e150e665911354e8fd7965368a789eb9 100644 |
--- a/media/gpu/v4l2_jpeg_decode_accelerator.cc |
+++ b/media/gpu/v4l2_jpeg_decode_accelerator.cc |
@@ -113,8 +113,8 @@ V4L2JpegDecodeAccelerator::BufferRecord::BufferRecord() |
V4L2JpegDecodeAccelerator::BufferRecord::~BufferRecord() {} |
V4L2JpegDecodeAccelerator::JobRecord::JobRecord( |
- const media::BitstreamBuffer& bitstream_buffer, |
- scoped_refptr<media::VideoFrame> video_frame) |
+ const BitstreamBuffer& bitstream_buffer, |
+ scoped_refptr<VideoFrame> video_frame) |
: bitstream_buffer_id(bitstream_buffer.id()), |
shm(bitstream_buffer, true), |
out_frame(video_frame) {} |
@@ -226,8 +226,8 @@ bool V4L2JpegDecodeAccelerator::Initialize(Client* client) { |
} |
void V4L2JpegDecodeAccelerator::Decode( |
- const media::BitstreamBuffer& bitstream_buffer, |
- const scoped_refptr<media::VideoFrame>& video_frame) { |
+ const BitstreamBuffer& bitstream_buffer, |
+ const scoped_refptr<VideoFrame>& video_frame) { |
DVLOG(1) << "Decode(): input_id=" << bitstream_buffer.id() |
<< ", size=" << bitstream_buffer.size(); |
DCHECK(io_task_runner_->BelongsToCurrentThread()); |
@@ -240,7 +240,7 @@ void V4L2JpegDecodeAccelerator::Decode( |
return; |
} |
- if (video_frame->format() != media::PIXEL_FORMAT_I420) { |
+ if (video_frame->format() != PIXEL_FORMAT_I420) { |
PostNotifyError(bitstream_buffer.id(), UNSUPPORTED_JPEG); |
return; |
} |
@@ -388,8 +388,8 @@ bool V4L2JpegDecodeAccelerator::CreateOutputBuffers() { |
DCHECK(!running_jobs_.empty()); |
linked_ptr<JobRecord> job_record = running_jobs_.front(); |
- size_t frame_size = media::VideoFrame::AllocationSize( |
- media::PIXEL_FORMAT_I420, job_record->out_frame->coded_size()); |
+ size_t frame_size = VideoFrame::AllocationSize( |
+ PIXEL_FORMAT_I420, job_record->out_frame->coded_size()); |
struct v4l2_format format; |
memset(&format, 0, sizeof(format)); |
format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; |
@@ -413,7 +413,7 @@ bool V4L2JpegDecodeAccelerator::CreateOutputBuffers() { |
DCHECK(output_buffer_map_.empty()); |
output_buffer_map_.resize(reqbufs.count); |
- media::VideoPixelFormat output_format = |
+ VideoPixelFormat output_format = |
V4L2Device::V4L2PixFmtToVideoPixelFormat(output_buffer_pixelformat_); |
for (size_t i = 0; i < output_buffer_map_.size(); ++i) { |
@@ -427,7 +427,7 @@ bool V4L2JpegDecodeAccelerator::CreateOutputBuffers() { |
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QUERYBUF, &buffer); |
DCHECK_GE(buffer.length, |
- media::VideoFrame::AllocationSize( |
+ VideoFrame::AllocationSize( |
output_format, |
gfx::Size(format.fmt.pix.width, format.fmt.pix.height))); |
@@ -616,16 +616,16 @@ void V4L2JpegDecodeAccelerator::EnqueueOutput() { |
static bool CopyOutputImage(const uint32_t src_pixelformat, |
const void* src_addr, |
const gfx::Size& src_coded_size, |
- const scoped_refptr<media::VideoFrame>& dst_frame) { |
- media::VideoPixelFormat format = |
+ const scoped_refptr<VideoFrame>& dst_frame) { |
+ VideoPixelFormat format = |
V4L2Device::V4L2PixFmtToVideoPixelFormat(src_pixelformat); |
- size_t src_size = media::VideoFrame::AllocationSize(format, src_coded_size); |
- uint8_t* dst_y = dst_frame->data(media::VideoFrame::kYPlane); |
- uint8_t* dst_u = dst_frame->data(media::VideoFrame::kUPlane); |
- uint8_t* dst_v = dst_frame->data(media::VideoFrame::kVPlane); |
- size_t dst_y_stride = dst_frame->stride(media::VideoFrame::kYPlane); |
- size_t dst_u_stride = dst_frame->stride(media::VideoFrame::kUPlane); |
- size_t dst_v_stride = dst_frame->stride(media::VideoFrame::kVPlane); |
+ size_t src_size = VideoFrame::AllocationSize(format, src_coded_size); |
+ uint8_t* dst_y = dst_frame->data(VideoFrame::kYPlane); |
+ uint8_t* dst_u = dst_frame->data(VideoFrame::kUPlane); |
+ uint8_t* dst_v = dst_frame->data(VideoFrame::kVPlane); |
+ size_t dst_y_stride = dst_frame->stride(VideoFrame::kYPlane); |
+ size_t dst_u_stride = dst_frame->stride(VideoFrame::kUPlane); |
+ size_t dst_v_stride = dst_frame->stride(VideoFrame::kVPlane); |
// If the source format is I420, ConvertToI420 will simply copy the frame. |
if (libyuv::ConvertToI420(static_cast<uint8_t*>(const_cast<void*>(src_addr)), |
@@ -746,7 +746,7 @@ static bool AddHuffmanTable(const void* input_ptr, |
uint8_t marker1, marker2; |
READ_U8_OR_RETURN_FALSE(reader, &marker1); |
READ_U8_OR_RETURN_FALSE(reader, &marker2); |
- if (marker1 != media::JPEG_MARKER_PREFIX || marker2 != media::JPEG_SOI) { |
+ if (marker1 != JPEG_MARKER_PREFIX || marker2 != JPEG_SOI) { |
DLOG(ERROR) << __func__ << ": The input is not a Jpeg"; |
return false; |
} |
@@ -758,13 +758,13 @@ static bool AddHuffmanTable(const void* input_ptr, |
while (!has_marker_sos && !has_marker_dht) { |
const char* start_addr = reader.ptr(); |
READ_U8_OR_RETURN_FALSE(reader, &marker1); |
- if (marker1 != media::JPEG_MARKER_PREFIX) { |
+ if (marker1 != JPEG_MARKER_PREFIX) { |
DLOG(ERROR) << __func__ << ": marker1 != 0xFF"; |
return false; |
} |
do { |
READ_U8_OR_RETURN_FALSE(reader, &marker2); |
- } while (marker2 == media::JPEG_MARKER_PREFIX); // skip fill bytes |
+ } while (marker2 == JPEG_MARKER_PREFIX); // skip fill bytes |
uint16_t size; |
READ_U16_OR_RETURN_FALSE(reader, &size); |
@@ -777,11 +777,11 @@ static bool AddHuffmanTable(const void* input_ptr, |
size -= sizeof(size); |
switch (marker2) { |
- case media::JPEG_DHT: { |
+ case JPEG_DHT: { |
has_marker_dht = true; |
break; |
} |
- case media::JPEG_SOS: { |
+ case JPEG_SOS: { |
if (!has_marker_dht) { |
memcpy(static_cast<uint8_t*>(output_ptr) + current_offset, |
kDefaultDhtSeg, sizeof(kDefaultDhtSeg)); |