| OLD | NEW |
| 1 // Copyright 2011 The Chromium Authors. All rights reserved. | 1 // Copyright 2011 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "config.h" | 5 #include "config.h" |
| 6 | 6 |
| 7 #include "cc/video_layer_impl.h" | 7 #include "cc/video_layer_impl.h" |
| 8 | 8 |
| 9 #include "NotImplemented.h" | 9 #include "NotImplemented.h" |
| 10 #include "cc/io_surface_draw_quad.h" | 10 #include "cc/io_surface_draw_quad.h" |
| (...skipping 104 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 115 if (!m_provider) { | 115 if (!m_provider) { |
| 116 m_frame = 0; | 116 m_frame = 0; |
| 117 return; | 117 return; |
| 118 } | 118 } |
| 119 | 119 |
| 120 m_frame = m_provider->getCurrentFrame(); | 120 m_frame = m_provider->getCurrentFrame(); |
| 121 | 121 |
| 122 if (!m_frame) | 122 if (!m_frame) |
| 123 return; | 123 return; |
| 124 | 124 |
| 125 // If these fail, we'll have to add draw logic that handles offset bitmap/ |
| 126 // texture UVs. For now, just expect (0, 0) offset, since all our decoders |
| 127 // so far don't offset. |
| 128 DCHECK_EQ(m_frame->visibleRect().x, 0); |
| 129 DCHECK_EQ(m_frame->visibleRect().y, 0); |
| 130 |
| 125 m_format = convertVFCFormatToGC3DFormat(*m_frame); | 131 m_format = convertVFCFormatToGC3DFormat(*m_frame); |
| 126 | 132 |
| 127 if (m_format == GL_INVALID_VALUE) { | 133 if (m_format == GL_INVALID_VALUE) { |
| 128 m_provider->putCurrentFrame(m_frame); | 134 m_provider->putCurrentFrame(m_frame); |
| 129 m_frame = 0; | 135 m_frame = 0; |
| 130 return; | 136 return; |
| 131 } | 137 } |
| 132 | 138 |
| 133 if (m_frame->planes() > WebKit::WebVideoFrame::maxPlanes) { | 139 if (m_frame->planes() > WebKit::WebVideoFrame::maxPlanes) { |
| 134 m_provider->putCurrentFrame(m_frame); | 140 m_provider->putCurrentFrame(m_frame); |
| (...skipping 24 matching lines...) Expand all Loading... |
| 159 if (!m_frame) | 165 if (!m_frame) |
| 160 return; | 166 return; |
| 161 | 167 |
| 162 SharedQuadState* sharedQuadState = quadSink.useSharedQuadState(createSharedQ
uadState()); | 168 SharedQuadState* sharedQuadState = quadSink.useSharedQuadState(createSharedQ
uadState()); |
| 163 appendDebugBorderQuad(quadSink, sharedQuadState, appendQuadsData); | 169 appendDebugBorderQuad(quadSink, sharedQuadState, appendQuadsData); |
| 164 | 170 |
| 165 // FIXME: When we pass quads out of process, we need to double-buffer, or | 171 // FIXME: When we pass quads out of process, we need to double-buffer, or |
| 166 // otherwise synchonize use of all textures in the quad. | 172 // otherwise synchonize use of all textures in the quad. |
| 167 | 173 |
| 168 IntRect quadRect(IntPoint(), contentBounds()); | 174 IntRect quadRect(IntPoint(), contentBounds()); |
| 175 WebKit::WebRect visibleRect = m_frame->visibleRect(); |
| 176 WebKit::WebSize textureSize = m_frame->textureSize(); |
| 177 |
| 178 // pixels for macroblocked formats. |
| 179 const float texWidthScale = |
| 180 static_cast<float>(visibleRect.width) / textureSize.width; |
| 181 const float texHeightScale = |
| 182 static_cast<float>(visibleRect.height) / textureSize.height; |
| 169 | 183 |
| 170 switch (m_format) { | 184 switch (m_format) { |
| 171 case GL_LUMINANCE: { | 185 case GL_LUMINANCE: { |
| 172 // YUV software decoder. | 186 // YUV software decoder. |
| 173 const FramePlane& yPlane = m_framePlanes[WebKit::WebVideoFrame::yPlane]; | 187 const FramePlane& yPlane = m_framePlanes[WebKit::WebVideoFrame::yPlane]; |
| 174 const FramePlane& uPlane = m_framePlanes[WebKit::WebVideoFrame::uPlane]; | 188 const FramePlane& uPlane = m_framePlanes[WebKit::WebVideoFrame::uPlane]; |
| 175 const FramePlane& vPlane = m_framePlanes[WebKit::WebVideoFrame::vPlane]; | 189 const FramePlane& vPlane = m_framePlanes[WebKit::WebVideoFrame::vPlane]; |
| 176 scoped_ptr<YUVVideoDrawQuad> yuvVideoQuad = YUVVideoDrawQuad::create(sha
redQuadState, quadRect, yPlane, uPlane, vPlane); | 190 FloatSize texScale(texWidthScale, texHeightScale); |
| 191 scoped_ptr<YUVVideoDrawQuad> yuvVideoQuad = YUVVideoDrawQuad::create( |
| 192 sharedQuadState, quadRect, texScale, yPlane, uPlane, vPlane); |
| 177 quadSink.append(yuvVideoQuad.PassAs<DrawQuad>(), appendQuadsData); | 193 quadSink.append(yuvVideoQuad.PassAs<DrawQuad>(), appendQuadsData); |
| 178 break; | 194 break; |
| 179 } | 195 } |
| 180 case GL_RGBA: { | 196 case GL_RGBA: { |
| 181 // RGBA software decoder. | 197 // RGBA software decoder. |
| 182 const FramePlane& plane = m_framePlanes[WebKit::WebVideoFrame::rgbPlane]
; | 198 const FramePlane& plane = m_framePlanes[WebKit::WebVideoFrame::rgbPlane]
; |
| 183 float widthScaleFactor = static_cast<float>(plane.visibleSize.width()) /
plane.size.width(); | |
| 184 | |
| 185 bool premultipliedAlpha = true; | 199 bool premultipliedAlpha = true; |
| 186 FloatRect uvRect(0, 0, widthScaleFactor, 1); | 200 FloatRect uvRect(0, 0, texWidthScale, texHeightScale); |
| 187 bool flipped = false; | 201 bool flipped = false; |
| 188 scoped_ptr<TextureDrawQuad> textureQuad = TextureDrawQuad::create(shared
QuadState, quadRect, plane.resourceId, premultipliedAlpha, uvRect, flipped); | 202 scoped_ptr<TextureDrawQuad> textureQuad = TextureDrawQuad::create(shared
QuadState, quadRect, plane.resourceId, premultipliedAlpha, uvRect, flipped); |
| 189 quadSink.append(textureQuad.PassAs<DrawQuad>(), appendQuadsData); | 203 quadSink.append(textureQuad.PassAs<DrawQuad>(), appendQuadsData); |
| 190 break; | 204 break; |
| 191 } | 205 } |
| 192 case GL_TEXTURE_2D: { | 206 case GL_TEXTURE_2D: { |
| 193 // NativeTexture hardware decoder. | 207 // NativeTexture hardware decoder. |
| 194 bool premultipliedAlpha = true; | 208 bool premultipliedAlpha = true; |
| 195 FloatRect uvRect(0, 0, 1, 1); | 209 FloatRect uvRect(0, 0, texWidthScale, texHeightScale); |
| 196 bool flipped = false; | 210 bool flipped = false; |
| 197 scoped_ptr<TextureDrawQuad> textureQuad = TextureDrawQuad::create(shared
QuadState, quadRect, m_externalTextureResource, premultipliedAlpha, uvRect, flip
ped); | 211 scoped_ptr<TextureDrawQuad> textureQuad = TextureDrawQuad::create(shared
QuadState, quadRect, m_externalTextureResource, premultipliedAlpha, uvRect, flip
ped); |
| 198 quadSink.append(textureQuad.PassAs<DrawQuad>(), appendQuadsData); | 212 quadSink.append(textureQuad.PassAs<DrawQuad>(), appendQuadsData); |
| 199 break; | 213 break; |
| 200 } | 214 } |
| 201 case GL_TEXTURE_RECTANGLE_ARB: { | 215 case GL_TEXTURE_RECTANGLE_ARB: { |
| 202 IntSize textureSize(m_frame->width(), m_frame->height()); | 216 IntSize visibleSize(visibleRect.width, visibleRect.height); |
| 203 scoped_ptr<IOSurfaceDrawQuad> ioSurfaceQuad = IOSurfaceDrawQuad::create(
sharedQuadState, quadRect, textureSize, m_frame->textureId(), IOSurfaceDrawQuad:
:Unflipped); | 217 scoped_ptr<IOSurfaceDrawQuad> ioSurfaceQuad = IOSurfaceDrawQuad::create(
sharedQuadState, quadRect, visibleSize, m_frame->textureId(), IOSurfaceDrawQuad:
:Unflipped); |
| 204 quadSink.append(ioSurfaceQuad.PassAs<DrawQuad>(), appendQuadsData); | 218 quadSink.append(ioSurfaceQuad.PassAs<DrawQuad>(), appendQuadsData); |
| 205 break; | 219 break; |
| 206 } | 220 } |
| 207 case GL_TEXTURE_EXTERNAL_OES: { | 221 case GL_TEXTURE_EXTERNAL_OES: { |
| 208 // StreamTexture hardware decoder. | 222 // StreamTexture hardware decoder. |
| 209 scoped_ptr<StreamVideoDrawQuad> streamVideoQuad = StreamVideoDrawQuad::c
reate(sharedQuadState, quadRect, m_frame->textureId(), m_streamTextureMatrix); | 223 WebKit::WebTransformationMatrix transform(m_streamTextureMatrix); |
| 224 transform.scaleNonUniform(texWidthScale, texHeightScale); |
| 225 scoped_ptr<StreamVideoDrawQuad> streamVideoQuad = |
| 226 StreamVideoDrawQuad::create(sharedQuadState, quadRect, |
| 227 m_frame->textureId(), |
| 228 m_streamTextureMatrix); |
| 210 quadSink.append(streamVideoQuad.PassAs<DrawQuad>(), appendQuadsData); | 229 quadSink.append(streamVideoQuad.PassAs<DrawQuad>(), appendQuadsData); |
| 211 break; | 230 break; |
| 212 } | 231 } |
| 213 default: | 232 default: |
| 214 CRASH(); // Someone updated convertVFCFormatToGC3DFormat above but updat
e this! | 233 CRASH(); // Someone updated convertVFCFormatToGC3DFormat above but updat
e this! |
| 215 } | 234 } |
| 216 } | 235 } |
| 217 | 236 |
| 218 void VideoLayerImpl::didDraw(ResourceProvider* resourceProvider) | 237 void VideoLayerImpl::didDraw(ResourceProvider* resourceProvider) |
| 219 { | 238 { |
| (...skipping 12 matching lines...) Expand all Loading... |
| 232 resourceProvider->deleteResource(m_externalTextureResource); | 251 resourceProvider->deleteResource(m_externalTextureResource); |
| 233 m_externalTextureResource = 0; | 252 m_externalTextureResource = 0; |
| 234 } | 253 } |
| 235 | 254 |
| 236 m_provider->putCurrentFrame(m_frame); | 255 m_provider->putCurrentFrame(m_frame); |
| 237 m_frame = 0; | 256 m_frame = 0; |
| 238 | 257 |
| 239 m_providerLock.Release(); | 258 m_providerLock.Release(); |
| 240 } | 259 } |
| 241 | 260 |
| 242 static int videoFrameDimension(int originalDimension, unsigned plane, int format
) | 261 bool VideoLayerImpl::FramePlane::allocateData( |
| 243 { | 262 ResourceProvider* resourceProvider) |
| 244 if (format == WebKit::WebVideoFrame::FormatYV12 && plane != WebKit::WebVideo
Frame::yPlane) | |
| 245 return originalDimension / 2; | |
| 246 return originalDimension; | |
| 247 } | |
| 248 | |
| 249 static bool hasPaddingBytes(const WebKit::WebVideoFrame& frame, unsigned plane) | |
| 250 { | |
| 251 return frame.stride(plane) > videoFrameDimension(frame.width(), plane, frame
.format()); | |
| 252 } | |
| 253 | |
| 254 IntSize VideoLayerImpl::computeVisibleSize(const WebKit::WebVideoFrame& frame, u
nsigned plane) | |
| 255 { | |
| 256 int visibleWidth = videoFrameDimension(frame.width(), plane, frame.format())
; | |
| 257 int originalWidth = visibleWidth; | |
| 258 int visibleHeight = videoFrameDimension(frame.height(), plane, frame.format(
)); | |
| 259 | |
| 260 // When there are dead pixels at the edge of the texture, decrease | |
| 261 // the frame width by 1 to prevent the rightmost pixels from | |
| 262 // interpolating with the dead pixels. | |
| 263 if (hasPaddingBytes(frame, plane)) | |
| 264 --visibleWidth; | |
| 265 | |
| 266 // In YV12, every 2x2 square of Y values corresponds to one U and | |
| 267 // one V value. If we decrease the width of the UV plane, we must decrease t
he | |
| 268 // width of the Y texture by 2 for proper alignment. This must happen | |
| 269 // always, even if Y's texture does not have padding bytes. | |
| 270 if (plane == WebKit::WebVideoFrame::yPlane && frame.format() == WebKit::WebV
ideoFrame::FormatYV12) { | |
| 271 if (hasPaddingBytes(frame, WebKit::WebVideoFrame::uPlane)) | |
| 272 visibleWidth = originalWidth - 2; | |
| 273 } | |
| 274 | |
| 275 return IntSize(visibleWidth, visibleHeight); | |
| 276 } | |
| 277 | |
| 278 bool VideoLayerImpl::FramePlane::allocateData(ResourceProvider* resourceProvider
) | |
| 279 { | 263 { |
| 280 if (resourceId) | 264 if (resourceId) |
| 281 return true; | 265 return true; |
| 282 | 266 |
| 283 resourceId = resourceProvider->createResource(Renderer::ImplPool, size, form
at, ResourceProvider::TextureUsageAny); | 267 resourceId = resourceProvider->createResource(Renderer::ImplPool, size, form
at, ResourceProvider::TextureUsageAny); |
| 284 return resourceId; | 268 return resourceId; |
| 285 } | 269 } |
| 286 | 270 |
| 287 void VideoLayerImpl::FramePlane::freeData(ResourceProvider* resourceProvider) | 271 void VideoLayerImpl::FramePlane::freeData(ResourceProvider* resourceProvider) |
| 288 { | 272 { |
| 289 if (!resourceId) | 273 if (!resourceId) |
| 290 return; | 274 return; |
| 291 | 275 |
| 292 resourceProvider->deleteResource(resourceId); | 276 resourceProvider->deleteResource(resourceId); |
| 293 resourceId = 0; | 277 resourceId = 0; |
| 294 } | 278 } |
| 295 | 279 |
| 296 bool VideoLayerImpl::allocatePlaneData(ResourceProvider* resourceProvider) | 280 bool VideoLayerImpl::allocatePlaneData(ResourceProvider* resourceProvider) |
| 297 { | 281 { |
| 282 WebKit::WebSize textureSize = m_frame->textureSize(); |
| 298 int maxTextureSize = resourceProvider->maxTextureSize(); | 283 int maxTextureSize = resourceProvider->maxTextureSize(); |
| 299 for (unsigned planeIndex = 0; planeIndex < m_frame->planes(); ++planeIndex)
{ | 284 for (unsigned planeIdx = 0; planeIdx < m_frame->planes(); ++planeIdx) { |
| 300 VideoLayerImpl::FramePlane& plane = m_framePlanes[planeIndex]; | 285 VideoLayerImpl::FramePlane& plane = m_framePlanes[planeIdx]; |
| 301 | 286 |
| 302 IntSize requiredTextureSize(m_frame->stride(planeIndex), videoFrameDimen
sion(m_frame->height(), planeIndex, m_frame->format())); | 287 IntSize requiredTextureSize(textureSize.width, textureSize.height); |
| 303 // FIXME: Remove the test against maxTextureSize when tiled layers are i
mplemented. | 288 // FIXME: Remove the test against maxTextureSize when tiled layers are |
| 304 if (requiredTextureSize.isZero() || requiredTextureSize.width() > maxTex
tureSize || requiredTextureSize.height() > maxTextureSize) | 289 // implemented. |
| 290 if (requiredTextureSize.isZero() || |
| 291 requiredTextureSize.width() > maxTextureSize || |
| 292 requiredTextureSize.height() > maxTextureSize) |
| 305 return false; | 293 return false; |
| 306 | 294 |
| 307 if (plane.size != requiredTextureSize || plane.format != m_format) { | 295 if (plane.size != requiredTextureSize || plane.format != m_format) { |
| 308 plane.freeData(resourceProvider); | 296 plane.freeData(resourceProvider); |
| 309 plane.size = requiredTextureSize; | 297 plane.size = requiredTextureSize; |
| 310 plane.format = m_format; | 298 plane.format = m_format; |
| 311 } | 299 } |
| 312 | 300 |
| 313 if (!plane.resourceId) { | 301 if (!plane.allocateData(resourceProvider)) |
| 314 if (!plane.allocateData(resourceProvider)) | 302 return false; |
| 315 return false; | |
| 316 plane.visibleSize = computeVisibleSize(*m_frame, planeIndex); | |
| 317 } | |
| 318 } | 303 } |
| 319 return true; | 304 return true; |
| 320 } | 305 } |
| 321 | 306 |
| 322 bool VideoLayerImpl::copyPlaneData(ResourceProvider* resourceProvider) | 307 bool VideoLayerImpl::copyPlaneData(ResourceProvider* resourceProvider) |
| 323 { | 308 { |
| 324 size_t softwarePlaneCount = m_frame->planes(); | 309 WebKit::WebSize textureSize = m_frame->textureSize(); |
| 325 if (!softwarePlaneCount) | 310 for (unsigned planeIdx = 0; planeIdx < m_frame->planes(); ++planeIdx) { |
| 326 return true; | 311 VideoLayerImpl::FramePlane& plane = m_framePlanes[planeIdx]; |
| 312 const uint8_t* planePixels = static_cast<const uint8_t*>(m_frame->data(p
laneIdx)); |
| 327 | 313 |
| 328 for (size_t softwarePlaneIndex = 0; softwarePlaneIndex < softwarePlaneCount;
++softwarePlaneIndex) { | 314 // Only non-FormatNativeTexture planes should need upload. |
| 329 VideoLayerImpl::FramePlane& plane = m_framePlanes[softwarePlaneIndex]; | 315 DCHECK_EQ(plane.format, GL_LUMINANCE); |
| 330 const uint8_t* softwarePlanePixels = static_cast<const uint8_t*>(m_frame
->data(softwarePlaneIndex)); | 316 |
| 331 IntRect planeRect(IntPoint(), plane.size); | 317 IntRect planeRect(0, 0, plane.size.width(), plane.size.height()); |
| 332 resourceProvider->upload(plane.resourceId, softwarePlanePixels, planeRec
t, planeRect, IntSize()); | 318 IntRect visibleRect(0, 0, textureSize.width, textureSize.height); |
| 319 resourceProvider->upload(plane.resourceId, planePixels, planeRect, |
| 320 visibleRect, IntSize()); |
| 333 } | 321 } |
| 334 return true; | 322 return true; |
| 335 } | 323 } |
| 336 | 324 |
| 337 void VideoLayerImpl::freePlaneData(ResourceProvider* resourceProvider) | 325 void VideoLayerImpl::freePlaneData(ResourceProvider* resourceProvider) |
| 338 { | 326 { |
| 339 for (unsigned i = 0; i < WebKit::WebVideoFrame::maxPlanes; ++i) | 327 for (unsigned i = 0; i < WebKit::WebVideoFrame::maxPlanes; ++i) |
| 340 m_framePlanes[i].freeData(resourceProvider); | 328 m_framePlanes[i].freeData(resourceProvider); |
| 341 } | 329 } |
| 342 | 330 |
| (...skipping 35 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 378 str->append("video layer\n"); | 366 str->append("video layer\n"); |
| 379 LayerImpl::dumpLayerProperties(str, indent); | 367 LayerImpl::dumpLayerProperties(str, indent); |
| 380 } | 368 } |
| 381 | 369 |
| 382 const char* VideoLayerImpl::layerTypeAsString() const | 370 const char* VideoLayerImpl::layerTypeAsString() const |
| 383 { | 371 { |
| 384 return "VideoLayer"; | 372 return "VideoLayer"; |
| 385 } | 373 } |
| 386 | 374 |
| 387 } | 375 } |
| OLD | NEW |