/*
* Copyright 2016 Google Inc.
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "GrVkResourceProvider.h"
#include "GrContextPriv.h"
#include "GrSamplerState.h"
#include "GrVkCommandBuffer.h"
#include "GrVkCommandPool.h"
#include "GrVkCopyPipeline.h"
#include "GrVkGpu.h"
#include "GrVkPipeline.h"
#include "GrVkRenderTarget.h"
#include "GrVkUniformBuffer.h"
#include "GrVkUtil.h"
#include "SkTaskGroup.h"
#ifdef SK_TRACE_VK_RESOURCES
std::atomic<uint32_t> GrVkResource::fKeyCounter{0};
#endif
GrVkResourceProvider::GrVkResourceProvider(GrVkGpu* gpu)
: fGpu(gpu)
, fPipelineCache(VK_NULL_HANDLE) {
fPipelineStateCache = new PipelineStateCache(gpu);
}
GrVkResourceProvider::~GrVkResourceProvider() {
SkASSERT(0 == fRenderPassArray.count());
SkASSERT(0 == fExternalRenderPasses.count());
SkASSERT(VK_NULL_HANDLE == fPipelineCache);
delete fPipelineStateCache;
}
VkPipelineCache GrVkResourceProvider::pipelineCache() {
if (fPipelineCache == VK_NULL_HANDLE) {
VkPipelineCacheCreateInfo createInfo;
memset(&createInfo, 0, sizeof(VkPipelineCacheCreateInfo));
createInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_CACHE_CREATE_INFO;
createInfo.pNext = nullptr;
createInfo.flags = 0;
auto persistentCache = fGpu->getContext()->priv().getPersistentCache();
sk_sp<SkData> cached;
if (persistentCache) {
uint32_t key = GrVkGpu::kPipelineCache_PersistentCacheKeyType;
sk_sp<SkData> keyData = SkData::MakeWithoutCopy(&key, sizeof(uint32_t));
cached = persistentCache->load(*keyData);
}
bool usedCached = false;
if (cached) {
uint32_t* cacheHeader = (uint32_t*)cached->data();
if (cacheHeader[1] == VK_PIPELINE_CACHE_HEADER_VERSION_ONE) {
// For version one of the header, the total header size is 16 bytes plus
// VK_UUID_SIZE bytes. See Section 9.6 (Pipeline Cache) in the vulkan spec to see
// the breakdown of these bytes.
SkASSERT(cacheHeader[0] == 16 + VK_UUID_SIZE);
const VkPhysicalDeviceProperties& devProps = fGpu->physicalDeviceProperties();
const uint8_t* supportedPipelineCacheUUID = devProps.pipelineCacheUUID;
if (cacheHeader[2] == devProps.vendorID && cacheHeader[3] == devProps.deviceID &&
!memcmp(&cacheHeader[4], supportedPipelineCacheUUID, VK_UUID_SIZE)) {
createInfo.initialDataSize = cached->size();
createInfo.pInitialData = cached->data();
usedCached = true;
}
}
}
if (!usedCached) {
createInfo.initialDataSize = 0;
createInfo.pInitialData = nullptr;
}
VkResult result = GR_VK_CALL(fGpu->vkInterface(),
CreatePipelineCache(fGpu->device(), &createInfo, nullptr,
&fPipelineCache));
SkASSERT(VK_SUCCESS == result);
if (VK_SUCCESS != result) {
fPipelineCache = VK_NULL_HANDLE;
}
}
return fPipelineCache;
}
void GrVkResourceProvider::init() {
// Init uniform descriptor objects
GrVkDescriptorSetManager* dsm = GrVkDescriptorSetManager::CreateUniformManager(fGpu);
fDescriptorSetManagers.emplace_back(dsm);
SkASSERT(1 == fDescriptorSetManagers.count());
fUniformDSHandle = GrVkDescriptorSetManager::Handle(0);
}
GrVkPipeline* GrVkResourceProvider::createPipeline(int numColorSamples,
const GrPrimitiveProcessor& primProc,
const GrPipeline& pipeline,
const GrStencilSettings& stencil,
VkPipelineShaderStageCreateInfo* shaderStageInfo,
int shaderStageCount,
GrPrimitiveType primitiveType,
VkRenderPass compatibleRenderPass,
VkPipelineLayout layout) {
return GrVkPipeline::Create(fGpu, numColorSamples, primProc, pipeline, stencil, shaderStageInfo,
shaderStageCount, primitiveType, compatibleRenderPass, layout,
this->pipelineCache());
}
GrVkCopyPipeline* GrVkResourceProvider::findOrCreateCopyPipeline(
const GrVkRenderTarget* dst,
VkPipelineShaderStageCreateInfo* shaderStageInfo,
VkPipelineLayout pipelineLayout) {
// Find or Create a compatible pipeline
GrVkCopyPipeline* pipeline = nullptr;
for (int i = 0; i < fCopyPipelines.count() && !pipeline; ++i) {
if (fCopyPipelines[i]->isCompatible(*dst->simpleRenderPass())) {
pipeline = fCopyPipelines[i];
}
}
if (!pipeline) {
pipeline = GrVkCopyPipeline::Create(fGpu, shaderStageInfo,
pipelineLayout,
dst->numColorSamples(),
*dst->simpleRenderPass(),
this->pipelineCache());
if (!pipeline) {
return nullptr;
}
fCopyPipelines.push_back(pipeline);
}
SkASSERT(pipeline);
pipeline->ref();
return pipeline;
}
// To create framebuffers, we first need to create a simple RenderPass that is
// only used for framebuffer creation. When we actually render we will create
// RenderPasses as needed that are compatible with the framebuffer.
const GrVkRenderPass*
GrVkResourceProvider::findCompatibleRenderPass(const GrVkRenderTarget& target,
CompatibleRPHandle* compatibleHandle) {
for (int i = 0; i < fRenderPassArray.count(); ++i) {
if (fRenderPassArray[i].isCompatible(target)) {
const GrVkRenderPass* renderPass = fRenderPassArray[i].getCompatibleRenderPass();
renderPass->ref();
if (compatibleHandle) {
*compatibleHandle = CompatibleRPHandle(i);
}
return renderPass;
}
}
const GrVkRenderPass* renderPass =
fRenderPassArray.emplace_back(fGpu, target).getCompatibleRenderPass();
renderPass->ref();
if (compatibleHandle) {
*compatibleHandle = CompatibleRPHandle(fRenderPassArray.count() - 1);
}
return renderPass;
}
const GrVkRenderPass*
GrVkResourceProvider::findCompatibleRenderPass(const CompatibleRPHandle& compatibleHandle) {
SkASSERT(compatibleHandle.isValid() && compatibleHandle.toIndex() < fRenderPassArray.count());
int index = compatibleHandle.toIndex();
const GrVkRenderPass* renderPass = fRenderPassArray[index].getCompatibleRenderPass();
renderPass->ref();
return renderPass;
}
const GrVkRenderPass* GrVkResourceProvider::findCompatibleExternalRenderPass(
VkRenderPass renderPass, uint32_t colorAttachmentIndex) {
for (int i = 0; i < fExternalRenderPasses.count(); ++i) {
if (fExternalRenderPasses[i]->isCompatibleExternalRP(renderPass)) {
fExternalRenderPasses[i]->ref();
#ifdef SK_DEBUG
uint32_t cachedColorIndex;
SkASSERT(fExternalRenderPasses[i]->colorAttachmentIndex(&cachedColorIndex));
SkASSERT(cachedColorIndex == colorAttachmentIndex);
#endif
return fExternalRenderPasses[i];
}
}
const GrVkRenderPass* newRenderPass = new GrVkRenderPass(renderPass, colorAttachmentIndex);
fExternalRenderPasses.push_back(newRenderPass);
newRenderPass->ref();
return newRenderPass;
}
const GrVkRenderPass* GrVkResourceProvider::findRenderPass(
const GrVkRenderTarget& target,
const GrVkRenderPass::LoadStoreOps& colorOps,
const GrVkRenderPass::LoadStoreOps& stencilOps,
CompatibleRPHandle* compatibleHandle) {
GrVkResourceProvider::CompatibleRPHandle tempRPHandle;
GrVkResourceProvider::CompatibleRPHandle* pRPHandle = compatibleHandle ? compatibleHandle
: &tempRPHandle;
*pRPHandle = target.compatibleRenderPassHandle();
// This will get us the handle to (and possible create) the compatible set for the specific
// GrVkRenderPass we are looking for.
this->findCompatibleRenderPass(target, compatibleHandle);
return this->findRenderPass(*pRPHandle, colorOps, stencilOps);
}
const GrVkRenderPass*
GrVkResourceProvider::findRenderPass(const CompatibleRPHandle& compatibleHandle,
const GrVkRenderPass::LoadStoreOps& colorOps,
const GrVkRenderPass::LoadStoreOps& stencilOps) {
SkASSERT(compatibleHandle.isValid() && compatibleHandle.toIndex() < fRenderPassArray.count());
CompatibleRenderPassSet& compatibleSet = fRenderPassArray[compatibleHandle.toIndex()];
const GrVkRenderPass* renderPass = compatibleSet.getRenderPass(fGpu,
colorOps,
stencilOps);
renderPass->ref();
return renderPass;
}
GrVkDescriptorPool* GrVkResourceProvider::findOrCreateCompatibleDescriptorPool(
VkDescriptorType type, uint32_t count) {
return new GrVkDescriptorPool(fGpu, type, count);
}
GrVkSampler* GrVkResourceProvider::findOrCreateCompatibleSampler(
const GrSamplerState& params, const GrVkYcbcrConversionInfo& ycbcrInfo) {
GrVkSampler* sampler = fSamplers.find(GrVkSampler::GenerateKey(params, ycbcrInfo));
if (!sampler) {
sampler = GrVkSampler::Create(fGpu, params, ycbcrInfo);
if (!sampler) {
return nullptr;
}
fSamplers.add(sampler);
}
SkASSERT(sampler);
sampler->ref();
return sampler;
}
GrVkSamplerYcbcrConversion* GrVkResourceProvider::findOrCreateCompatibleSamplerYcbcrConversion(
const GrVkYcbcrConversionInfo& ycbcrInfo) {
GrVkSamplerYcbcrConversion* ycbcrConversion =
fYcbcrConversions.find(GrVkSamplerYcbcrConversion::GenerateKey(ycbcrInfo));
if (!ycbcrConversion) {
ycbcrConversion = GrVkSamplerYcbcrConversion::Create(fGpu, ycbcrInfo);
if (!ycbcrConversion) {
return nullptr;
}
fYcbcrConversions.add(ycbcrConversion);
}
SkASSERT(ycbcrConversion);
ycbcrConversion->ref();
return ycbcrConversion;
}
GrVkPipelineState* GrVkResourceProvider::findOrCreateCompatiblePipelineState(
GrRenderTarget* renderTarget, GrSurfaceOrigin origin,
const GrPipeline& pipeline, const GrPrimitiveProcessor& proc,
const GrTextureProxy* const primProcProxies[], GrPrimitiveType primitiveType,
VkRenderPass compatibleRenderPass) {
return fPipelineStateCache->refPipelineState(renderTarget, origin, proc, primProcProxies,
pipeline, primitiveType, compatibleRenderPass);
}
void GrVkResourceProvider::getSamplerDescriptorSetHandle(VkDescriptorType type,
const GrVkUniformHandler& uniformHandler,
GrVkDescriptorSetManager::Handle* handle) {
SkASSERT(handle);
SkASSERT(VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER == type ||
VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER == type);
for (int i = 0; i < fDescriptorSetManagers.count(); ++i) {
if (fDescriptorSetManagers[i]->isCompatible(type, &uniformHandler)) {
*handle = GrVkDescriptorSetManager::Handle(i);
return;
}
}
GrVkDescriptorSetManager* dsm = GrVkDescriptorSetManager::CreateSamplerManager(fGpu, type,
uniformHandler);
fDescriptorSetManagers.emplace_back(dsm);
*handle = GrVkDescriptorSetManager::Handle(fDescriptorSetManagers.count() - 1);
}
void GrVkResourceProvider::getSamplerDescriptorSetHandle(VkDescriptorType type,
const SkTArray<uint32_t>& visibilities,
GrVkDescriptorSetManager::Handle* handle) {
SkASSERT(handle);
SkASSERT(VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER == type ||
VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER == type);
for (int i = 0; i < fDescriptorSetManagers.count(); ++i) {
if (fDescriptorSetManagers[i]->isCompatible(type, visibilities)) {
*handle = GrVkDescriptorSetManager::Handle(i);
return;
}
}
GrVkDescriptorSetManager* dsm = GrVkDescriptorSetManager::CreateSamplerManager(fGpu, type,
visibilities);
fDescriptorSetManagers.emplace_back(dsm);
*handle = GrVkDescriptorSetManager::Handle(fDescriptorSetManagers.count() - 1);
}
VkDescriptorSetLayout GrVkResourceProvider::getUniformDSLayout() const {
SkASSERT(fUniformDSHandle.isValid());
return fDescriptorSetManagers[fUniformDSHandle.toIndex()]->layout();
}
VkDescriptorSetLayout GrVkResourceProvider::getSamplerDSLayout(
const GrVkDescriptorSetManager::Handle& handle) const {
SkASSERT(handle.isValid());
return fDescriptorSetManagers[handle.toIndex()]->layout();
}
const GrVkDescriptorSet* GrVkResourceProvider::getUniformDescriptorSet() {
SkASSERT(fUniformDSHandle.isValid());
return fDescriptorSetManagers[fUniformDSHandle.toIndex()]->getDescriptorSet(fGpu,
fUniformDSHandle);
}
const GrVkDescriptorSet* GrVkResourceProvider::getSamplerDescriptorSet(
const GrVkDescriptorSetManager::Handle& handle) {
SkASSERT(handle.isValid());
return fDescriptorSetManagers[handle.toIndex()]->getDescriptorSet(fGpu, handle);
}
void GrVkResourceProvider::recycleDescriptorSet(const GrVkDescriptorSet* descSet,
const GrVkDescriptorSetManager::Handle& handle) {
SkASSERT(descSet);
SkASSERT(handle.isValid());
int managerIdx = handle.toIndex();
SkASSERT(managerIdx < fDescriptorSetManagers.count());
fDescriptorSetManagers[managerIdx]->recycleDescriptorSet(descSet);
}
GrVkCommandPool* GrVkResourceProvider::findOrCreateCommandPool() {
std::unique_lock<std::recursive_mutex> lock(fBackgroundMutex);
GrVkCommandPool* result;
if (fAvailableCommandPools.count()) {
result = fAvailableCommandPools.back();
fAvailableCommandPools.pop_back();
} else {
result = GrVkCommandPool::Create(fGpu);
}
SkASSERT(result->unique());
SkDEBUGCODE(
for (const GrVkCommandPool* pool : fActiveCommandPools) {
SkASSERT(pool != result);
}
for (const GrVkCommandPool* pool : fAvailableCommandPools) {
SkASSERT(pool != result);
}
)
fActiveCommandPools.push_back(result);
result->ref();
return result;
}
void GrVkResourceProvider::checkCommandBuffers() {
for (int i = fActiveCommandPools.count() - 1; i >= 0; --i) {
GrVkCommandPool* pool = fActiveCommandPools[i];
if (!pool->isOpen()) {
GrVkPrimaryCommandBuffer* buffer = pool->getPrimaryCommandBuffer();
if (buffer->finished(fGpu)) {
fActiveCommandPools.removeShuffle(i);
this->backgroundReset(pool);
}
}
}
}
void GrVkResourceProvider::addFinishedProcToActiveCommandBuffers(
GrGpuFinishedProc finishedProc, GrGpuFinishedContext finishedContext) {
sk_sp<GrRefCntedCallback> procRef(new GrRefCntedCallback(finishedProc, finishedContext));
for (int i = 0; i < fActiveCommandPools.count(); ++i) {
GrVkCommandPool* pool = fActiveCommandPools[i];
if (!pool->isOpen()) {
GrVkPrimaryCommandBuffer* buffer = pool->getPrimaryCommandBuffer();
buffer->addFinishedProc(procRef);
}
}
}
const GrVkResource* GrVkResourceProvider::findOrCreateStandardUniformBufferResource() {
const GrVkResource* resource = nullptr;
int count = fAvailableUniformBufferResources.count();
if (count > 0) {
resource = fAvailableUniformBufferResources[count - 1];
fAvailableUniformBufferResources.removeShuffle(count - 1);
} else {
resource = GrVkUniformBuffer::CreateResource(fGpu, GrVkUniformBuffer::kStandardSize);
}
return resource;
}
void GrVkResourceProvider::recycleStandardUniformBufferResource(const GrVkResource* resource) {
fAvailableUniformBufferResources.push_back(resource);
}
void GrVkResourceProvider::destroyResources(bool deviceLost) {
SkTaskGroup* taskGroup = fGpu->getContext()->priv().getTaskGroup();
if (taskGroup) {
taskGroup->wait();
}
// Release all copy pipelines
for (int i = 0; i < fCopyPipelines.count(); ++i) {
fCopyPipelines[i]->unref(fGpu);
}
// loop over all render pass sets to make sure we destroy all the internal VkRenderPasses
for (int i = 0; i < fRenderPassArray.count(); ++i) {
fRenderPassArray[i].releaseResources(fGpu);
}
fRenderPassArray.reset();
for (int i = 0; i < fExternalRenderPasses.count(); ++i) {
fExternalRenderPasses[i]->unref(fGpu);
}
fExternalRenderPasses.reset();
// Iterate through all store GrVkSamplers and unref them before resetting the hash.
SkTDynamicHash<GrVkSampler, GrVkSampler::Key>::Iter iter(&fSamplers);
for (; !iter.done(); ++iter) {
(*iter).unref(fGpu);
}
fSamplers.reset();
fPipelineStateCache->release();
GR_VK_CALL(fGpu->vkInterface(), DestroyPipelineCache(fGpu->device(), fPipelineCache, nullptr));
fPipelineCache = VK_NULL_HANDLE;
for (GrVkCommandPool* pool : fActiveCommandPools) {
SkASSERT(pool->unique());
pool->unref(fGpu);
}
fActiveCommandPools.reset();
for (GrVkCommandPool* pool : fAvailableCommandPools) {
SkASSERT(pool->unique());
pool->unref(fGpu);
}
fAvailableCommandPools.reset();
// We must release/destroy all command buffers and pipeline states before releasing the
// GrVkDescriptorSetManagers
for (int i = 0; i < fDescriptorSetManagers.count(); ++i) {
fDescriptorSetManagers[i]->release(fGpu);
}
fDescriptorSetManagers.reset();
// release our uniform buffers
for (int i = 0; i < fAvailableUniformBufferResources.count(); ++i) {
SkASSERT(fAvailableUniformBufferResources[i]->unique());
fAvailableUniformBufferResources[i]->unref(fGpu);
}
fAvailableUniformBufferResources.reset();
}
void GrVkResourceProvider::abandonResources() {
SkTaskGroup* taskGroup = fGpu->getContext()->priv().getTaskGroup();
if (taskGroup) {
taskGroup->wait();
}
// Abandon all command pools
for (int i = 0; i < fActiveCommandPools.count(); ++i) {
SkASSERT(fActiveCommandPools[i]->unique());
fActiveCommandPools[i]->unrefAndAbandon();
}
fActiveCommandPools.reset();
for (int i = 0; i < fAvailableCommandPools.count(); ++i) {
SkASSERT(fAvailableCommandPools[i]->unique());
fAvailableCommandPools[i]->unrefAndAbandon();
}
fAvailableCommandPools.reset();
// Abandon all copy pipelines
for (int i = 0; i < fCopyPipelines.count(); ++i) {
fCopyPipelines[i]->unrefAndAbandon();
}
// loop over all render pass sets to make sure we destroy all the internal VkRenderPasses
for (int i = 0; i < fRenderPassArray.count(); ++i) {
fRenderPassArray[i].abandonResources();
}
fRenderPassArray.reset();
for (int i = 0; i < fExternalRenderPasses.count(); ++i) {
fExternalRenderPasses[i]->unrefAndAbandon();
}
fExternalRenderPasses.reset();
// Iterate through all store GrVkSamplers and unrefAndAbandon them before resetting the hash.
SkTDynamicHash<GrVkSampler, GrVkSampler::Key>::Iter iter(&fSamplers);
for (; !iter.done(); ++iter) {
(*iter).unrefAndAbandon();
}
fSamplers.reset();
fPipelineStateCache->abandon();
fPipelineCache = VK_NULL_HANDLE;
// We must abandon all command buffers and pipeline states before abandoning the
// GrVkDescriptorSetManagers
for (int i = 0; i < fDescriptorSetManagers.count(); ++i) {
fDescriptorSetManagers[i]->abandon();
}
fDescriptorSetManagers.reset();
// release our uniform buffers
for (int i = 0; i < fAvailableUniformBufferResources.count(); ++i) {
SkASSERT(fAvailableUniformBufferResources[i]->unique());
fAvailableUniformBufferResources[i]->unrefAndAbandon();
}
fAvailableUniformBufferResources.reset();
}
void GrVkResourceProvider::backgroundReset(GrVkCommandPool* pool) {
SkASSERT(pool->unique());
pool->releaseResources(fGpu);
SkTaskGroup* taskGroup = fGpu->getContext()->priv().getTaskGroup();
if (taskGroup) {
taskGroup->add([this, pool]() {
this->reset(pool);
});
} else {
this->reset(pool);
}
}
void GrVkResourceProvider::reset(GrVkCommandPool* pool) {
SkASSERT(pool->unique());
pool->reset(fGpu);
std::unique_lock<std::recursive_mutex> providerLock(fBackgroundMutex);
fAvailableCommandPools.push_back(pool);
}
void GrVkResourceProvider::storePipelineCacheData() {
size_t dataSize = 0;
VkResult result = GR_VK_CALL(fGpu->vkInterface(), GetPipelineCacheData(fGpu->device(),
this->pipelineCache(),
&dataSize, nullptr));
SkASSERT(result == VK_SUCCESS);
std::unique_ptr<uint8_t[]> data(new uint8_t[dataSize]);
result = GR_VK_CALL(fGpu->vkInterface(), GetPipelineCacheData(fGpu->device(),
this->pipelineCache(),
&dataSize,
(void*)data.get()));
SkASSERT(result == VK_SUCCESS);
uint32_t key = GrVkGpu::kPipelineCache_PersistentCacheKeyType;
sk_sp<SkData> keyData = SkData::MakeWithoutCopy(&key, sizeof(uint32_t));
fGpu->getContext()->priv().getPersistentCache()->store(
*keyData, *SkData::MakeWithoutCopy(data.get(), dataSize));
}
////////////////////////////////////////////////////////////////////////////////
GrVkResourceProvider::CompatibleRenderPassSet::CompatibleRenderPassSet(
const GrVkGpu* gpu,
const GrVkRenderTarget& target)
: fLastReturnedIndex(0) {
fRenderPasses.emplace_back(new GrVkRenderPass());
fRenderPasses[0]->initSimple(gpu, target);
}
bool GrVkResourceProvider::CompatibleRenderPassSet::isCompatible(
const GrVkRenderTarget& target) const {
// The first GrVkRenderpass should always exists since we create the basic load store
// render pass on create
SkASSERT(fRenderPasses[0]);
return fRenderPasses[0]->isCompatible(target);
}
GrVkRenderPass* GrVkResourceProvider::CompatibleRenderPassSet::getRenderPass(
const GrVkGpu* gpu,
const GrVkRenderPass::LoadStoreOps& colorOps,
const GrVkRenderPass::LoadStoreOps& stencilOps) {
for (int i = 0; i < fRenderPasses.count(); ++i) {
int idx = (i + fLastReturnedIndex) % fRenderPasses.count();
if (fRenderPasses[idx]->equalLoadStoreOps(colorOps, stencilOps)) {
fLastReturnedIndex = idx;
return fRenderPasses[idx];
}
}
GrVkRenderPass* renderPass = fRenderPasses.emplace_back(new GrVkRenderPass());
renderPass->init(gpu, *this->getCompatibleRenderPass(), colorOps, stencilOps);
fLastReturnedIndex = fRenderPasses.count() - 1;
return renderPass;
}
void GrVkResourceProvider::CompatibleRenderPassSet::releaseResources(GrVkGpu* gpu) {
for (int i = 0; i < fRenderPasses.count(); ++i) {
if (fRenderPasses[i]) {
fRenderPasses[i]->unref(gpu);
fRenderPasses[i] = nullptr;
}
}
}
void GrVkResourceProvider::CompatibleRenderPassSet::abandonResources() {
for (int i = 0; i < fRenderPasses.count(); ++i) {
if (fRenderPasses[i]) {
fRenderPasses[i]->unrefAndAbandon();
fRenderPasses[i] = nullptr;
}
}
}