/* * Copyright 2012 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "SkBitmap.h" #include "SkBitmapCache.h" #include "SkCanvas.h" #include "SkData.h" #include "SkImageEncoder.h" #include "SkImageGenerator.h" #include "SkImagePriv.h" #include "SkImageShader.h" #include "SkImage_Base.h" #include "SkNextID.h" #include "SkPixelRef.h" #include "SkPixelSerializer.h" #include "SkReadPixelsRec.h" #include "SkString.h" #include "SkSurface.h" #if SK_SUPPORT_GPU #include "GrTexture.h" #include "GrContext.h" #include "SkImage_Gpu.h" #endif SkImage::SkImage(int width, int height, uint32_t uniqueID) : fWidth(width) , fHeight(height) , fUniqueID(kNeedNewImageUniqueID == uniqueID ? SkNextID::ImageID() : uniqueID) { SkASSERT(width > 0); SkASSERT(height > 0); } const void* SkImage::peekPixels(SkImageInfo* info, size_t* rowBytes) const { SkImageInfo infoStorage; size_t rowBytesStorage; if (nullptr == info) { info = &infoStorage; } if (nullptr == rowBytes) { rowBytes = &rowBytesStorage; } return as_IB(this)->onPeekPixels(info, rowBytes); } bool SkImage::readPixels(const SkImageInfo& dstInfo, void* dstPixels, size_t dstRowBytes, int srcX, int srcY, CachingHint chint) const { SkReadPixelsRec rec(dstInfo, dstPixels, dstRowBytes, srcX, srcY); if (!rec.trim(this->width(), this->height())) { return false; } return as_IB(this)->onReadPixels(rec.fInfo, rec.fPixels, rec.fRowBytes, rec.fX, rec.fY, chint); } bool SkImage::scalePixels(const SkPixmap& dst, SkFilterQuality quality, CachingHint chint) const { if (this->width() == dst.width() && this->height() == dst.height()) { return this->readPixels(dst, 0, 0, chint); } // Idea: If/when SkImageGenerator supports a native-scaling API (where the generator itself // can scale more efficiently) we should take advantage of it here. // SkBitmap bm; if (as_IB(this)->getROPixels(&bm, chint)) { bm.lockPixels(); SkPixmap pmap; // Note: By calling the pixmap scaler, we never cache the final result, so the chint // is (currently) only being applied to the getROPixels. If we get a request to // also attempt to cache the final (scaled) result, we would add that logic here. // return bm.peekPixels(&pmap) && pmap.scalePixels(dst, quality); } return false; } void SkImage::preroll(GrContext* ctx) const { // For now, and to maintain parity w/ previous pixelref behavior, we just force the image // to produce a cached raster-bitmap form, so that drawing to a raster canvas should be fast. // SkBitmap bm; if (as_IB(this)->getROPixels(&bm)) { bm.lockPixels(); bm.unlockPixels(); } } /////////////////////////////////////////////////////////////////////////////////////////////////// SkShader* SkImage::newShader(SkShader::TileMode tileX, SkShader::TileMode tileY, const SkMatrix* localMatrix) const { return SkImageShader::Create(this, tileX, tileY, localMatrix); } SkData* SkImage::encode(SkImageEncoder::Type type, int quality) const { SkBitmap bm; if (as_IB(this)->getROPixels(&bm)) { return SkImageEncoder::EncodeData(bm, type, quality); } return nullptr; } SkData* SkImage::encode(SkPixelSerializer* serializer) const { SkAutoTUnref<SkPixelSerializer> defaultSerializer; SkPixelSerializer* effectiveSerializer = serializer; if (!effectiveSerializer) { defaultSerializer.reset(SkImageEncoder::CreatePixelSerializer()); SkASSERT(defaultSerializer.get()); effectiveSerializer = defaultSerializer.get(); } SkAutoTUnref<SkData> encoded(this->refEncoded()); if (encoded && effectiveSerializer->useEncodedData(encoded->data(), encoded->size())) { return encoded.detach(); } SkBitmap bm; SkAutoPixmapUnlock apu; if (as_IB(this)->getROPixels(&bm) && bm.requestLock(&apu)) { return effectiveSerializer->encode(apu.pixmap()); } return nullptr; } SkData* SkImage::refEncoded() const { GrContext* ctx = nullptr; // should we allow the caller to pass in a ctx? return as_IB(this)->onRefEncoded(ctx); } SkImage* SkImage::NewFromEncoded(SkData* encoded, const SkIRect* subset) { if (nullptr == encoded || 0 == encoded->size()) { return nullptr; } SkImageGenerator* generator = SkImageGenerator::NewFromEncoded(encoded); return generator ? SkImage::NewFromGenerator(generator, subset) : nullptr; } const char* SkImage::toString(SkString* str) const { str->appendf("image: (id:%d (%d, %d) %s)", this->uniqueID(), this->width(), this->height(), this->isOpaque() ? "opaque" : ""); return str->c_str(); } SkImage* SkImage::newSubset(const SkIRect& subset) const { if (subset.isEmpty()) { return nullptr; } const SkIRect bounds = SkIRect::MakeWH(this->width(), this->height()); if (!bounds.contains(subset)) { return nullptr; } // optimization : return self if the subset == our bounds if (bounds == subset) { return SkRef(const_cast<SkImage*>(this)); } return as_IB(this)->onNewSubset(subset); } #if SK_SUPPORT_GPU GrTexture* SkImage::getTexture() const { return as_IB(this)->peekTexture(); } bool SkImage::isTextureBacked() const { return SkToBool(as_IB(this)->getTexture()); } GrBackendObject SkImage::getTextureHandle(bool flushPendingGrContextIO) const { GrTexture* texture = as_IB(this)->getTexture(); if (texture) { GrContext* context = texture->getContext(); if (context) { if (flushPendingGrContextIO) { context->prepareSurfaceForExternalIO(texture); } } return texture->getTextureHandle(); } return 0; } #else GrTexture* SkImage::getTexture() const { return nullptr; } bool SkImage::isTextureBacked() const { return false; } GrBackendObject SkImage::getTextureHandle(bool) const { return 0; } #endif /////////////////////////////////////////////////////////////////////////////// static bool raster_canvas_supports(const SkImageInfo& info) { switch (info.colorType()) { case kN32_SkColorType: return kUnpremul_SkAlphaType != info.alphaType(); case kRGB_565_SkColorType: return true; case kAlpha_8_SkColorType: return true; default: break; } return false; } SkImage_Base::SkImage_Base(int width, int height, uint32_t uniqueID) : INHERITED(width, height, uniqueID) , fAddedToCache(false) {} SkImage_Base::~SkImage_Base() { if (fAddedToCache.load()) { SkNotifyBitmapGenIDIsStale(this->uniqueID()); } } bool SkImage_Base::onReadPixels(const SkImageInfo& dstInfo, void* dstPixels, size_t dstRowBytes, int srcX, int srcY, CachingHint) const { if (!raster_canvas_supports(dstInfo)) { return false; } SkBitmap bm; bm.installPixels(dstInfo, dstPixels, dstRowBytes); SkCanvas canvas(bm); SkPaint paint; paint.setXfermodeMode(SkXfermode::kSrc_Mode); canvas.drawImage(this, -SkIntToScalar(srcX), -SkIntToScalar(srcY), &paint); return true; } /////////////////////////////////////////////////////////////////////////////////////////////////// bool SkImage::peekPixels(SkPixmap* pmap) const { SkImageInfo info; size_t rowBytes; const void* pixels = this->peekPixels(&info, &rowBytes); if (pixels) { if (pmap) { pmap->reset(info, pixels, rowBytes); } return true; } return false; } bool SkImage::readPixels(const SkPixmap& pmap, int srcX, int srcY, CachingHint chint) const { return this->readPixels(pmap.info(), pmap.writable_addr(), pmap.rowBytes(), srcX, srcY, chint); } #if SK_SUPPORT_GPU #include "GrTextureToYUVPlanes.h" #endif #include "SkRGBAToYUV.h" bool SkImage::readYUV8Planes(const SkISize sizes[3], void* const planes[3], const size_t rowBytes[3], SkYUVColorSpace colorSpace) const { #if SK_SUPPORT_GPU if (GrTexture* texture = as_IB(this)->peekTexture()) { if (GrTextureToYUVPlanes(texture, sizes, planes, rowBytes, colorSpace)) { return true; } } #endif return SkRGBAToYUV(this, sizes, planes, rowBytes, colorSpace); } /////////////////////////////////////////////////////////////////////////////////////////////////// SkImage* SkImage::NewFromBitmap(const SkBitmap& bm) { SkPixelRef* pr = bm.pixelRef(); if (nullptr == pr) { return nullptr; } #if SK_SUPPORT_GPU if (GrTexture* tex = pr->getTexture()) { SkAutoTUnref<GrTexture> unrefCopy; if (!bm.isImmutable()) { tex = GrDeepCopyTexture(tex, SkBudgeted::kNo); if (nullptr == tex) { return nullptr; } unrefCopy.reset(tex); } const SkImageInfo info = bm.info(); return new SkImage_Gpu(info.width(), info.height(), bm.getGenerationID(), info.alphaType(), tex, SkBudgeted::kNo); } #endif // This will check for immutable (share or copy) return SkNewImageFromRasterBitmap(bm); } bool SkImage::asLegacyBitmap(SkBitmap* bitmap, LegacyBitmapMode mode) const { return as_IB(this)->onAsLegacyBitmap(bitmap, mode); } bool SkImage_Base::onAsLegacyBitmap(SkBitmap* bitmap, LegacyBitmapMode mode) const { // As the base-class, all we can do is make a copy (regardless of mode). // Subclasses that want to be more optimal should override. SkImageInfo info = SkImageInfo::MakeN32(this->width(), this->height(), this->isOpaque() ? kOpaque_SkAlphaType : kPremul_SkAlphaType); if (!bitmap->tryAllocPixels(info)) { return false; } if (!this->readPixels(bitmap->info(), bitmap->getPixels(), bitmap->rowBytes(), 0, 0)) { bitmap->reset(); return false; } if (kRO_LegacyBitmapMode == mode) { bitmap->setImmutable(); } return true; } SkImage* SkImage::NewFromPicture(const SkPicture* picture, const SkISize& dimensions, const SkMatrix* matrix, const SkPaint* paint) { if (!picture) { return nullptr; } return NewFromGenerator(SkImageGenerator::NewFromPicture(dimensions, picture, matrix, paint)); } bool SkImage::isLazyGenerated() const { return as_IB(this)->onIsLazyGenerated(); } ////////////////////////////////////////////////////////////////////////////////////// #if !SK_SUPPORT_GPU SkImage* SkImage::NewFromTexture(GrContext*, const GrBackendTextureDesc&, SkAlphaType, TextureReleaseProc, ReleaseContext) { return nullptr; } SkImage* SkImage::NewFromAdoptedTexture(GrContext*, const GrBackendTextureDesc&, SkAlphaType) { return nullptr; } SkImage* SkImage::NewFromTextureCopy(GrContext*, const GrBackendTextureDesc&, SkAlphaType) { return nullptr; } SkImage* SkImage::newTextureImage(GrContext*) const { return nullptr; } #endif