| /* |
| * Copyright 2018 Google Inc. |
| * |
| * Use of this source code is governed by a BSD-style license that can be |
| * found in the LICENSE file. |
| */ |
| |
| #include "include/private/GrTypesPriv.h" |
| #include "src/gpu/GrGpuResourcePriv.h" |
| #include "src/gpu/GrStagingBufferManager.h" |
| #include "src/gpu/mtl/GrMtlBuffer.h" |
| #include "src/gpu/mtl/GrMtlCommandBuffer.h" |
| #include "src/gpu/mtl/GrMtlGpu.h" |
| |
| #if !__has_feature(objc_arc) |
| #error This file must be compiled with Arc. Use -fobjc-arc flag |
| #endif |
| |
| #ifdef SK_DEBUG |
| #define VALIDATE() this->validate() |
| #else |
| #define VALIDATE() do {} while(false) |
| #endif |
| |
| GR_NORETAIN_BEGIN |
| |
| #ifdef SK_ENABLE_MTL_DEBUG_INFO |
| NSString* kBufferTypeNames[kGrGpuBufferTypeCount] = { |
| @"Vertex", |
| @"Index", |
| @"Indirect", |
| @"Xfer CPU to GPU", |
| @"Xfer GPU to CPU", |
| @"Uniform", |
| }; |
| #endif |
| |
| sk_sp<GrMtlBuffer> GrMtlBuffer::Make(GrMtlGpu* gpu, size_t size, GrGpuBufferType intendedType, |
| GrAccessPattern accessPattern, const void* data) { |
| sk_sp<GrMtlBuffer> buffer(new GrMtlBuffer(gpu, size, intendedType, accessPattern)); |
| if (data && !buffer->onUpdateData(data, size)) { |
| return nullptr; |
| } |
| return buffer; |
| } |
| |
| GrMtlBuffer::GrMtlBuffer(GrMtlGpu* gpu, size_t size, GrGpuBufferType intendedType, |
| GrAccessPattern accessPattern) |
| : INHERITED(gpu, size, intendedType, accessPattern) |
| , fIsDynamic(accessPattern != kStatic_GrAccessPattern) { |
| NSUInteger options = 0; |
| if (@available(macOS 10.11, iOS 9.0, *)) { |
| if (fIsDynamic) { |
| #ifdef SK_BUILD_FOR_MAC |
| if (gpu->mtlCaps().isMac()) { |
| options |= MTLResourceStorageModeManaged; |
| } else { |
| options |= MTLResourceStorageModeShared; |
| } |
| #else |
| options |= MTLResourceStorageModeShared; |
| #endif |
| } else { |
| options |= MTLResourceStorageModePrivate; |
| } |
| } |
| |
| size = SkAlignTo(size, gpu->mtlCaps().getMinBufferAlignment()); |
| fMtlBuffer = size == 0 ? nil : |
| [gpu->device() newBufferWithLength: size |
| options: options]; |
| #ifdef SK_ENABLE_MTL_DEBUG_INFO |
| fMtlBuffer.label = kBufferTypeNames[(int)intendedType]; |
| #endif |
| this->registerWithCache(SkBudgeted::kYes); |
| VALIDATE(); |
| } |
| |
| GrMtlBuffer::~GrMtlBuffer() { |
| SkASSERT(!fMtlBuffer); |
| SkASSERT(!fMapPtr); |
| } |
| |
| bool GrMtlBuffer::onUpdateData(const void* src, size_t sizeInBytes) { |
| if (this->wasDestroyed()) { |
| return false; |
| } |
| |
| if (sizeInBytes > this->size()) { |
| return false; |
| } |
| |
| if (fIsDynamic) { |
| this->internalMap(sizeInBytes); |
| if (!fMapPtr) { |
| return false; |
| } |
| memcpy(fMapPtr, src, sizeInBytes); |
| this->internalUnmap(sizeInBytes); |
| } else { |
| // copy data to gpu buffer |
| GrStagingBufferManager::Slice slice; |
| slice = this->mtlGpu()->stagingBufferManager()->allocateStagingBufferSlice( |
| sizeInBytes, this->mtlGpu()->mtlCaps().getMinBufferAlignment()); |
| if (!slice.fBuffer) { |
| return false; |
| } |
| memcpy(slice.fOffsetMapPtr, src, sizeInBytes); |
| |
| GrMtlCommandBuffer* cmdBuffer = this->mtlGpu()->commandBuffer(); |
| id<MTLBlitCommandEncoder> GR_NORETAIN blitCmdEncoder = cmdBuffer->getBlitCommandEncoder(); |
| if (!blitCmdEncoder) { |
| return false; |
| } |
| GrMtlBuffer* mtlBuffer = static_cast<GrMtlBuffer*>(slice.fBuffer); |
| id<MTLBuffer> transferBuffer = mtlBuffer->mtlBuffer(); |
| [blitCmdEncoder copyFromBuffer: transferBuffer |
| sourceOffset: slice.fOffset |
| toBuffer: fMtlBuffer |
| destinationOffset: 0 |
| size: sizeInBytes]; |
| } |
| |
| return true; |
| } |
| |
| inline GrMtlGpu* GrMtlBuffer::mtlGpu() const { |
| SkASSERT(!this->wasDestroyed()); |
| return static_cast<GrMtlGpu*>(this->getGpu()); |
| } |
| |
| void GrMtlBuffer::onAbandon() { |
| fMtlBuffer = nil; |
| fMapPtr = nullptr; |
| VALIDATE(); |
| INHERITED::onAbandon(); |
| } |
| |
| void GrMtlBuffer::onRelease() { |
| if (!this->wasDestroyed()) { |
| VALIDATE(); |
| fMtlBuffer = nil; |
| fMapPtr = nullptr; |
| VALIDATE(); |
| } |
| INHERITED::onRelease(); |
| } |
| |
| void GrMtlBuffer::internalMap(size_t sizeInBytes) { |
| if (fIsDynamic) { |
| VALIDATE(); |
| SkASSERT(sizeInBytes <= this->size()); |
| SkASSERT(!this->isMapped()); |
| fMapPtr = static_cast<char*>(fMtlBuffer.contents); |
| VALIDATE(); |
| } |
| } |
| |
| void GrMtlBuffer::internalUnmap(size_t sizeInBytes) { |
| SkASSERT(fMtlBuffer); |
| if (fIsDynamic) { |
| VALIDATE(); |
| SkASSERT(sizeInBytes <= this->size()); |
| SkASSERT(this->isMapped()); |
| #ifdef SK_BUILD_FOR_MAC |
| if (this->mtlGpu()->mtlCaps().isMac()) { |
| [fMtlBuffer didModifyRange: NSMakeRange(0, sizeInBytes)]; |
| } |
| #endif |
| fMapPtr = nullptr; |
| } |
| } |
| |
| void GrMtlBuffer::onMap() { |
| if (!this->wasDestroyed()) { |
| this->internalMap(this->size()); |
| } |
| } |
| |
| void GrMtlBuffer::onUnmap() { |
| if (!this->wasDestroyed()) { |
| this->internalUnmap(this->size()); |
| } |
| } |
| |
| #ifdef SK_DEBUG |
| void GrMtlBuffer::validate() const { |
| SkASSERT(fMtlBuffer == nil || |
| this->intendedType() == GrGpuBufferType::kVertex || |
| this->intendedType() == GrGpuBufferType::kIndex || |
| this->intendedType() == GrGpuBufferType::kXferCpuToGpu || |
| this->intendedType() == GrGpuBufferType::kXferGpuToCpu || |
| this->intendedType() == GrGpuBufferType::kDrawIndirect || |
| this->intendedType() == GrGpuBufferType::kUniform); |
| SkASSERT((fMapPtr && fMtlBuffer) || !fMapPtr); |
| } |
| #endif |
| |
| GR_NORETAIN_END |