/* * Copyright (C) 2010 The Android Open Source Project * Copyright (c) 2011-2014 The Linux Foundation. All rights reserved. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include <atomic> #include <limits.h> #include <unistd.h> #include <fcntl.h> #include <cutils/properties.h> #include <sys/mman.h> #include "gr.h" #include "gpu.h" #include "memalloc.h" #include "alloc_controller.h" #include <qdMetaData.h> #include <linux/msm_ion.h> using namespace gralloc; static uint64_t next_backing_store_id() { static std::atomic<uint64_t> next_id(1); return next_id++; } gpu_context_t::gpu_context_t(const private_module_t* module, IAllocController* alloc_ctrl ) : mAllocCtrl(alloc_ctrl) { // Zero out the alloc_device_t memset(static_cast<alloc_device_t*>(this), 0, sizeof(alloc_device_t)); // Initialize the procs common.tag = HARDWARE_DEVICE_TAG; common.version = 0; common.module = const_cast<hw_module_t*>(&module->base.common); common.close = gralloc_close; alloc = gralloc_alloc; free = gralloc_free; } int gpu_context_t::gralloc_alloc_buffer(unsigned int size, int usage, buffer_handle_t* pHandle, int bufferType, int format, int width, int height) { int err = 0; int flags = 0; size = roundUpToPageSize(size); alloc_data data; data.offset = 0; data.fd = -1; data.base = 0; if(format == HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED) data.align = 8192; else data.align = getpagesize(); if (usage & GRALLOC_USAGE_PROTECTED) { if (usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY) { /* The alignment here reflects qsee mmu V7L/V8L requirement */ data.align = SZ_2M; } else { data.align = SECURE_ALIGN; } size = ALIGN(size, data.align); } data.size = size; data.pHandle = (uintptr_t) pHandle; err = mAllocCtrl->allocate(data, usage); if (!err) { /* allocate memory for enhancement data */ alloc_data eData; eData.fd = -1; eData.base = 0; eData.offset = 0; eData.size = ROUND_UP_PAGESIZE(sizeof(MetaData_t)); eData.pHandle = data.pHandle; eData.align = getpagesize(); int eDataUsage = 0; int eDataErr = mAllocCtrl->allocate(eData, eDataUsage); ALOGE_IF(eDataErr, "gralloc failed for eDataErr=%s", strerror(-eDataErr)); if (usage & GRALLOC_USAGE_PRIVATE_EXTERNAL_ONLY) { flags |= private_handle_t::PRIV_FLAGS_EXTERNAL_ONLY; } if (usage & GRALLOC_USAGE_PRIVATE_INTERNAL_ONLY) { flags |= private_handle_t::PRIV_FLAGS_INTERNAL_ONLY; } if (usage & GRALLOC_USAGE_HW_VIDEO_ENCODER ) { flags |= private_handle_t::PRIV_FLAGS_VIDEO_ENCODER; } if (usage & GRALLOC_USAGE_HW_CAMERA_WRITE) { flags |= private_handle_t::PRIV_FLAGS_CAMERA_WRITE; } if (usage & GRALLOC_USAGE_HW_CAMERA_READ) { flags |= private_handle_t::PRIV_FLAGS_CAMERA_READ; } if (usage & GRALLOC_USAGE_HW_COMPOSER) { flags |= private_handle_t::PRIV_FLAGS_HW_COMPOSER; } if (usage & GRALLOC_USAGE_HW_TEXTURE) { flags |= private_handle_t::PRIV_FLAGS_HW_TEXTURE; } if(usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY) { flags |= private_handle_t::PRIV_FLAGS_SECURE_DISPLAY; } if(isMacroTileEnabled(format, usage)) { flags |= private_handle_t::PRIV_FLAGS_TILE_RENDERED; } if (isUBwcEnabled(format, usage)) { flags |= private_handle_t::PRIV_FLAGS_UBWC_ALIGNED; } if(usage & (GRALLOC_USAGE_SW_READ_MASK | GRALLOC_USAGE_SW_WRITE_MASK)) { flags |= private_handle_t::PRIV_FLAGS_CPU_RENDERED; } if (usage & (GRALLOC_USAGE_HW_VIDEO_ENCODER | GRALLOC_USAGE_HW_CAMERA_WRITE | GRALLOC_USAGE_HW_RENDER | GRALLOC_USAGE_HW_FB)) { flags |= private_handle_t::PRIV_FLAGS_NON_CPU_WRITER; } if(usage & GRALLOC_USAGE_HW_COMPOSER) { flags |= private_handle_t::PRIV_FLAGS_DISP_CONSUMER; } if(false == data.uncached) { flags |= private_handle_t::PRIV_FLAGS_CACHED; } flags |= data.allocType; uint64_t eBaseAddr = (uint64_t)(eData.base) + eData.offset; private_handle_t *hnd = new private_handle_t(data.fd, size, flags, bufferType, format, width, height, eData.fd, eData.offset, eBaseAddr); hnd->offset = data.offset; hnd->base = (uint64_t)(data.base) + data.offset; hnd->gpuaddr = 0; ColorSpace_t colorSpace = ITU_R_601; setMetaData(hnd, UPDATE_COLOR_SPACE, (void*) &colorSpace); *pHandle = hnd; } ALOGE_IF(err, "gralloc failed err=%s", strerror(-err)); return err; } void gpu_context_t::getGrallocInformationFromFormat(int inputFormat, int *bufferType) { *bufferType = BUFFER_TYPE_VIDEO; if (isUncompressedRgbFormat(inputFormat) == TRUE) { // RGB formats *bufferType = BUFFER_TYPE_UI; } } int gpu_context_t::gralloc_alloc_framebuffer_locked(int usage, buffer_handle_t* pHandle) { private_module_t* m = reinterpret_cast<private_module_t*>(common.module); // This allocation will only happen when gralloc is in fb mode if (m->framebuffer == NULL) { ALOGE("%s: Invalid framebuffer", __FUNCTION__); return -EINVAL; } const unsigned int bufferMask = m->bufferMask; const uint32_t numBuffers = m->numBuffers; unsigned int bufferSize = m->finfo.line_length * m->info.yres; //adreno needs FB size to be page aligned bufferSize = roundUpToPageSize(bufferSize); if (numBuffers == 1) { // If we have only one buffer, we never use page-flipping. Instead, // we return a regular buffer which will be memcpy'ed to the main // screen when post is called. int newUsage = (usage & ~GRALLOC_USAGE_HW_FB) | GRALLOC_USAGE_HW_2D; return gralloc_alloc_buffer(bufferSize, newUsage, pHandle, BUFFER_TYPE_UI, m->fbFormat, m->info.xres, m->info.yres); } if (bufferMask >= ((1LU<<numBuffers)-1)) { // We ran out of buffers. return -ENOMEM; } // create a "fake" handle for it uint64_t vaddr = uint64_t(m->framebuffer->base); // As GPU needs ION FD, the private handle is created // using ION fd and ION flags are set private_handle_t* hnd = new private_handle_t( dup(m->framebuffer->fd), bufferSize, private_handle_t::PRIV_FLAGS_USES_ION | private_handle_t::PRIV_FLAGS_FRAMEBUFFER, BUFFER_TYPE_UI, m->fbFormat, m->info.xres, m->info.yres); // find a free slot for (uint32_t i=0 ; i<numBuffers ; i++) { if ((bufferMask & (1LU<<i)) == 0) { m->bufferMask |= (uint32_t)(1LU<<i); break; } vaddr += bufferSize; } hnd->base = vaddr; hnd->offset = (unsigned int)(vaddr - m->framebuffer->base); *pHandle = hnd; return 0; } int gpu_context_t::gralloc_alloc_framebuffer(int usage, buffer_handle_t* pHandle) { private_module_t* m = reinterpret_cast<private_module_t*>(common.module); pthread_mutex_lock(&m->lock); int err = gralloc_alloc_framebuffer_locked(usage, pHandle); pthread_mutex_unlock(&m->lock); return err; } int gpu_context_t::alloc_impl(int w, int h, int format, int usage, buffer_handle_t* pHandle, int* pStride, unsigned int bufferSize) { if (!pHandle || !pStride) return -EINVAL; unsigned int size; int alignedw, alignedh; int grallocFormat = format; int bufferType; //If input format is HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED then based on //the usage bits, gralloc assigns a format. if(format == HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED || format == HAL_PIXEL_FORMAT_YCbCr_420_888) { if (usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC) grallocFormat = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC; else if(usage & GRALLOC_USAGE_HW_VIDEO_ENCODER) grallocFormat = HAL_PIXEL_FORMAT_NV12_ENCODEABLE; //NV12 else if((usage & GRALLOC_USAGE_HW_CAMERA_MASK) == GRALLOC_USAGE_HW_CAMERA_ZSL) grallocFormat = HAL_PIXEL_FORMAT_NV21_ZSL; //NV21 ZSL else if(usage & GRALLOC_USAGE_HW_CAMERA_READ) grallocFormat = HAL_PIXEL_FORMAT_YCrCb_420_SP; //NV21 else if(usage & GRALLOC_USAGE_HW_CAMERA_WRITE) { if (format == HAL_PIXEL_FORMAT_YCbCr_420_888) { grallocFormat = HAL_PIXEL_FORMAT_NV21_ZSL; //NV21 } else { grallocFormat = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS; //NV12 preview } } else if(usage & GRALLOC_USAGE_HW_COMPOSER) //XXX: If we still haven't set a format, default to RGBA8888 grallocFormat = HAL_PIXEL_FORMAT_RGBA_8888; else if(format == HAL_PIXEL_FORMAT_YCbCr_420_888) { //If no other usage flags are detected, default the //flexible YUV format to NV21_ZSL grallocFormat = HAL_PIXEL_FORMAT_NV21_ZSL; } } bool useFbMem = false; char property[PROPERTY_VALUE_MAX]; char isUBWC[PROPERTY_VALUE_MAX]; if (usage & GRALLOC_USAGE_HW_FB) { if ((property_get("debug.gralloc.map_fb_memory", property, NULL) > 0) && (!strncmp(property, "1", PROPERTY_VALUE_MAX ) || (!strncasecmp(property,"true", PROPERTY_VALUE_MAX )))) { useFbMem = true; } else { usage &= ~GRALLOC_USAGE_PRIVATE_ALLOC_UBWC; if (property_get("debug.gralloc.enable_fb_ubwc", isUBWC, NULL) > 0){ if ((!strncmp(isUBWC, "1", PROPERTY_VALUE_MAX)) || (!strncasecmp(isUBWC, "true", PROPERTY_VALUE_MAX))) { // Allocate UBWC aligned framebuffer usage |= GRALLOC_USAGE_PRIVATE_ALLOC_UBWC; } } } } getGrallocInformationFromFormat(grallocFormat, &bufferType); size = getBufferSizeAndDimensions(w, h, grallocFormat, usage, alignedw, alignedh); if ((unsigned int)size <= 0) return -EINVAL; size = (bufferSize >= size)? bufferSize : size; int err = 0; if(useFbMem) { err = gralloc_alloc_framebuffer(usage, pHandle); } else { err = gralloc_alloc_buffer(size, usage, pHandle, bufferType, grallocFormat, alignedw, alignedh); } if (err < 0) { return err; } auto hnd = (private_handle_t*) *pHandle; hnd->backing_store = next_backing_store_id(); hnd->original_width = w; hnd->original_format = format; *pStride = alignedw; return 0; } int gpu_context_t::free_impl(private_handle_t const* hnd) { private_module_t* m = reinterpret_cast<private_module_t*>(common.module); if (hnd->flags & private_handle_t::PRIV_FLAGS_FRAMEBUFFER) { const unsigned int bufferSize = m->finfo.line_length * m->info.yres; unsigned int index = (unsigned int) ((hnd->base - m->framebuffer->base) / bufferSize); m->bufferMask &= (uint32_t)~(1LU<<index); } else { terminateBuffer(&m->base, const_cast<private_handle_t*>(hnd)); IMemAlloc* memalloc = mAllocCtrl->getAllocator(hnd->flags); int err = memalloc->free_buffer((void*)hnd->base, hnd->size, hnd->offset, hnd->fd); if(err) return err; // free the metadata space unsigned int size = ROUND_UP_PAGESIZE(sizeof(MetaData_t)); err = memalloc->free_buffer((void*)hnd->base_metadata, size, hnd->offset_metadata, hnd->fd_metadata); if (err) return err; } delete hnd; return 0; } int gpu_context_t::gralloc_alloc(alloc_device_t* dev, int w, int h, int format, int usage, buffer_handle_t* pHandle, int* pStride) { if (!dev) { return -EINVAL; } gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, 0); } int gpu_context_t::gralloc_alloc_size(alloc_device_t* dev, int w, int h, int format, int usage, buffer_handle_t* pHandle, int* pStride, int bufferSize) { if (!dev) { return -EINVAL; } gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, bufferSize); } int gpu_context_t::gralloc_free(alloc_device_t* dev, buffer_handle_t handle) { if (private_handle_t::validate(handle) < 0) return -EINVAL; private_handle_t const* hnd = reinterpret_cast<private_handle_t const*>(handle); gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev); return gpu->free_impl(hnd); } /*****************************************************************************/ int gpu_context_t::gralloc_close(struct hw_device_t *dev) { gpu_context_t* ctx = reinterpret_cast<gpu_context_t*>(dev); if (ctx) { /* TODO: keep a list of all buffer_handle_t created, and free them * all here. */ delete ctx; } return 0; }