mirror of
https://github.com/intel/compute-runtime.git
synced 2026-01-03 23:03:02 +08:00
Clean unused code
remove parts with device queue Signed-off-by: Katarzyna Cencelewska <katarzyna.cencelewska@intel.com>
This commit is contained in:
committed by
Compute-Runtime-Automation
parent
df4bec03e8
commit
25ce4d0c04
@@ -10,9 +10,6 @@
|
||||
|
||||
#include <cstdint>
|
||||
|
||||
// It's max SSH size per kernel (MAX_BINDING_TABLE_INDEX * 64)
|
||||
const uint32_t SAMPLER_OBJECT_ID_SHIFT = 253 * 64;
|
||||
|
||||
// Sampler Patch Token Enums
|
||||
enum SAMPLER_PATCH_ENUM {
|
||||
CLK_DEFAULT_SAMPLER = 0x00,
|
||||
|
||||
@@ -22,7 +22,6 @@
|
||||
#include "shared/source/helpers/per_thread_data.h"
|
||||
#include "shared/source/helpers/ptr_math.h"
|
||||
#include "shared/source/helpers/surface_format_info.h"
|
||||
#include "shared/source/kernel/kernel_arg_descriptor_extended_device_side_enqueue.h"
|
||||
#include "shared/source/kernel/kernel_arg_descriptor_extended_vme.h"
|
||||
#include "shared/source/memory_manager/memory_manager.h"
|
||||
#include "shared/source/memory_manager/unified_memory_manager.h"
|
||||
@@ -1581,12 +1580,6 @@ cl_int Kernel::setArgImageWithMipLevel(uint32_t argIndex,
|
||||
patch<uint32_t, uint64_t>(imageDesc.image_array_size, crossThreadData, argAsImg.metadataPayload.arraySize);
|
||||
patch<uint32_t, cl_channel_type>(imageFormat.image_channel_data_type, crossThreadData, argAsImg.metadataPayload.channelDataType);
|
||||
patch<uint32_t, cl_channel_order>(imageFormat.image_channel_order, crossThreadData, argAsImg.metadataPayload.channelOrder);
|
||||
if (arg.getExtendedTypeInfo().hasDeviceSideEnqueueExtendedDescriptor) {
|
||||
const auto &explicitArgsExtendedDescriptors = kernelInfo.kernelDescriptor.payloadMappings.explicitArgsExtendedDescriptors;
|
||||
UNRECOVERABLE_IF(argIndex >= explicitArgsExtendedDescriptors.size());
|
||||
auto deviceSideEnqueueDescriptor = static_cast<ArgDescriptorDeviceSideEnqueue *>(explicitArgsExtendedDescriptors[argIndex].get());
|
||||
patch<uint32_t, uint32_t>(argAsImg.bindful, crossThreadData, deviceSideEnqueueDescriptor->objectId);
|
||||
}
|
||||
|
||||
auto pixelSize = pImage->getSurfaceFormatInfo().surfaceFormat.ImageElementSizeInBytes;
|
||||
patch<uint64_t, uint64_t>(graphicsAllocation->getGpuAddress(), crossThreadData, argAsImg.metadataPayload.flatBaseOffset);
|
||||
@@ -1670,12 +1663,6 @@ cl_int Kernel::setArgSampler(uint32_t argIndex,
|
||||
patch<uint32_t, uint32_t>(pSampler->getSnapWaValue(), crossThreadData, argAsSmp.metadataPayload.samplerSnapWa);
|
||||
patch<uint32_t, uint32_t>(getAddrModeEnum(pSampler->addressingMode), crossThreadData, argAsSmp.metadataPayload.samplerAddressingMode);
|
||||
patch<uint32_t, uint32_t>(getNormCoordsEnum(pSampler->normalizedCoordinates), crossThreadData, argAsSmp.metadataPayload.samplerNormalizedCoords);
|
||||
if (arg.getExtendedTypeInfo().hasDeviceSideEnqueueExtendedDescriptor) {
|
||||
const auto &explicitArgsExtendedDescriptors = kernelInfo.kernelDescriptor.payloadMappings.explicitArgsExtendedDescriptors;
|
||||
UNRECOVERABLE_IF(argIndex >= explicitArgsExtendedDescriptors.size());
|
||||
auto deviceSideEnqueueDescriptor = static_cast<ArgDescriptorDeviceSideEnqueue *>(explicitArgsExtendedDescriptors[argIndex].get());
|
||||
patch<uint32_t, uint32_t>(SAMPLER_OBJECT_ID_SHIFT + argAsSmp.bindful, crossThreadData, deviceSideEnqueueDescriptor->objectId);
|
||||
}
|
||||
|
||||
retVal = CL_SUCCESS;
|
||||
}
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (C) 2018-2021 Intel Corporation
|
||||
* Copyright (C) 2018-2022 Intel Corporation
|
||||
*
|
||||
* SPDX-License-Identifier: MIT
|
||||
*
|
||||
@@ -47,7 +47,6 @@ void KernelImageArgTest::SetUp() {
|
||||
pKernelInfo->addArgImage(3, 0);
|
||||
|
||||
pKernelInfo->addArgImage(4, 0x20);
|
||||
pKernelInfo->addExtendedDeviceSideEnqueueDescriptor(4, 0);
|
||||
|
||||
pKernelInfo->kernelDescriptor.kernelAttributes.bufferAddressingMode = ApiSpecificConfig::getBindlessConfiguration() ? KernelDescriptor::AddressingMode::BindlessAndStateless : KernelDescriptor::AddressingMode::BindfulAndStateless;
|
||||
pKernelInfo->kernelDescriptor.kernelAttributes.imageAddressingMode = ApiSpecificConfig::getBindlessConfiguration() ? KernelDescriptor::AddressingMode::Bindless : KernelDescriptor::AddressingMode::Bindful;
|
||||
|
||||
@@ -92,7 +92,7 @@ class CloneKernelTest : public MultiRootDeviceWithSubDevicesFixture {
|
||||
char surfaceStateHeap[128];
|
||||
};
|
||||
|
||||
TEST_F(CloneKernelTest, GivenKernelWithPrivateSurfaceWhenCloningKernelThenClonedKernelProgramItsOwnPrivateSurfaceAddress) {
|
||||
TEST_F(CloneKernelTest, givenKernelWithPrivateSurfaceWhenCloningKernelThenClonedKernelProgramItsOwnPrivateSurfaceAddress) {
|
||||
for (auto &rootDeviceIndex : this->context->getRootDeviceIndices()) {
|
||||
auto pSourcePrivateSurface = pSourceKernel[rootDeviceIndex]->privateSurface;
|
||||
auto pClonedPrivateSurface = pClonedKernel[rootDeviceIndex]->privateSurface;
|
||||
@@ -119,7 +119,7 @@ TEST_F(CloneKernelTest, GivenKernelWithPrivateSurfaceWhenCloningKernelThenCloned
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenUnsetArgWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenUnsetArgWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
pKernelInfo->addArgBuffer(0);
|
||||
for (auto &rootDeviceIndex : this->context->getRootDeviceIndices()) {
|
||||
EXPECT_EQ(1u, pSourceKernel[rootDeviceIndex]->getKernelArguments().size());
|
||||
@@ -147,7 +147,7 @@ TEST_F(CloneKernelTest, GivenUnsetArgWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgLocalWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgLocalWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
const size_t slmSize = 0x800;
|
||||
pKernelInfo->addArgLocal(0, 0, 1);
|
||||
|
||||
@@ -184,7 +184,7 @@ TEST_F(CloneKernelTest, GivenArgLocalWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgBufferWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgBufferWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
pKernelInfo->addArgBuffer(0, 0x20, sizeof(void *));
|
||||
|
||||
auto buffer = clUniquePtr(Buffer::create(context.get(), 0, MemoryConstants::pageSize, nullptr, retVal));
|
||||
@@ -226,7 +226,7 @@ TEST_F(CloneKernelTest, GivenArgBufferWhenCloningKernelThenKernelInfoIsCorrect)
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgPipeWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgPipeWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
pKernelInfo->addArgPipe(0, 0x20, sizeof(void *));
|
||||
auto pipe = clUniquePtr(Pipe::create(context.get(), 0, 1, 20, nullptr, retVal));
|
||||
EXPECT_EQ(CL_SUCCESS, retVal);
|
||||
@@ -265,13 +265,12 @@ TEST_F(CloneKernelTest, GivenArgPipeWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
EXPECT_EQ(pipe->getGraphicsAllocation(rootDeviceIndex)->getGpuAddressToPatch(), reinterpret_cast<uint64_t>(*pKernelArg));
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgImageWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgImageWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
pKernelInfo->addArgImage(0, 0x20);
|
||||
auto &metaPayload = pKernelInfo->argAsImg(0).metadataPayload;
|
||||
metaPayload.imgWidth = 0x4;
|
||||
metaPayload.imgHeight = 0x8;
|
||||
metaPayload.imgDepth = 0xc;
|
||||
pKernelInfo->addExtendedDeviceSideEnqueueDescriptor(0, 0);
|
||||
|
||||
auto image = std::unique_ptr<Image>(Image2dHelper<>::create(context.get()));
|
||||
ASSERT_NE(nullptr, image);
|
||||
@@ -310,7 +309,6 @@ TEST_F(CloneKernelTest, GivenArgImageWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
|
||||
auto crossThreadData = reinterpret_cast<uint32_t *>(pClonedKernel[rootDeviceIndex]->getCrossThreadData());
|
||||
auto &clonedArg = pClonedKernel[rootDeviceIndex]->getKernelInfo().getArgDescriptorAt(0).as<ArgDescImage>();
|
||||
EXPECT_EQ(clonedArg.bindful, *crossThreadData);
|
||||
|
||||
auto pImgWidth = ptrOffset(crossThreadData, clonedArg.metadataPayload.imgWidth);
|
||||
EXPECT_EQ(imageWidth, *pImgWidth);
|
||||
@@ -322,7 +320,7 @@ TEST_F(CloneKernelTest, GivenArgImageWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
EXPECT_EQ(imageDepth, *pImgDepth);
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgAcceleratorWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgAcceleratorWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
pKernelInfo->addArgAccelerator(0, undefined<SurfaceStateHeapOffset>, 0x4, 0x14, 0x1c, 0xc);
|
||||
|
||||
cl_motion_estimation_desc_intel desc = {
|
||||
@@ -384,14 +382,13 @@ TEST_F(CloneKernelTest, GivenArgAcceleratorWhenCloningKernelThenKernelInfoIsCorr
|
||||
EXPECT_EQ(CL_SUCCESS, retVal);
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgSamplerWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgSamplerWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
auto sampler = clUniquePtr<Sampler>(new MockSampler(context.get(),
|
||||
true,
|
||||
(cl_addressing_mode)CL_ADDRESS_MIRRORED_REPEAT,
|
||||
(cl_filter_mode)CL_FILTER_NEAREST));
|
||||
|
||||
pKernelInfo->addArgSampler(0, 0x20, 0x8, 0x10, 0x4);
|
||||
pKernelInfo->addExtendedDeviceSideEnqueueDescriptor(0, 0);
|
||||
|
||||
cl_sampler samplerObj = sampler.get();
|
||||
auto rootDeviceIndex = *context->getRootDeviceIndices().begin();
|
||||
@@ -423,7 +420,6 @@ TEST_F(CloneKernelTest, GivenArgSamplerWhenCloningKernelThenKernelInfoIsCorrect)
|
||||
|
||||
auto crossThreadData = reinterpret_cast<uint32_t *>(pClonedKernel[rootDeviceIndex]->getCrossThreadData());
|
||||
const auto &clonedArg = pClonedKernel[rootDeviceIndex]->getKernelInfo().getArgDescriptorAt(0).as<ArgDescSampler>();
|
||||
EXPECT_EQ(SAMPLER_OBJECT_ID_SHIFT + clonedArg.bindful, *crossThreadData);
|
||||
|
||||
auto pSnapWa = ptrOffset(crossThreadData, clonedArg.metadataPayload.samplerSnapWa);
|
||||
EXPECT_EQ(sampler->getSnapWaValue(), *pSnapWa);
|
||||
@@ -436,7 +432,7 @@ TEST_F(CloneKernelTest, GivenArgSamplerWhenCloningKernelThenKernelInfoIsCorrect)
|
||||
EXPECT_EQ(3, sampler->getRefInternalCount());
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgSvmWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgSvmWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
char *svmPtr = new char[256];
|
||||
|
||||
pKernelInfo->addArgBuffer(0, 0x20, sizeof(void *));
|
||||
@@ -474,7 +470,7 @@ TEST_F(CloneKernelTest, GivenArgSvmWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
delete[] svmPtr;
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgSvmAllocWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgSvmAllocWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
pKernelInfo->addArgBuffer(0, 0x20, sizeof(void *));
|
||||
|
||||
char memory[100] = {};
|
||||
@@ -517,7 +513,7 @@ TEST_F(CloneKernelTest, GivenArgSvmAllocWhenCloningKernelThenKernelInfoIsCorrect
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenArgImmediateWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenArgImmediateWhenCloningKernelThenKernelInfoIsCorrect) {
|
||||
pKernelInfo->addArgImmediate(0, sizeof(void *), 0x20);
|
||||
|
||||
using TypeParam = unsigned long;
|
||||
@@ -554,7 +550,7 @@ TEST_F(CloneKernelTest, GivenArgImmediateWhenCloningKernelThenKernelInfoIsCorrec
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenExecInfoWhenCloningKernelThenSvmAllocationIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenExecInfoWhenCloningKernelThenSvmAllocationIsCorrect) {
|
||||
REQUIRE_SVM_OR_SKIP(device1);
|
||||
void *ptrSVM = context->getSVMAllocsManager()->createSVMAlloc(256, {}, context->getRootDeviceIndices(), context->getDeviceBitfields());
|
||||
ASSERT_NE(nullptr, ptrSVM);
|
||||
@@ -582,7 +578,7 @@ TEST_F(CloneKernelTest, GivenExecInfoWhenCloningKernelThenSvmAllocationIsCorrect
|
||||
context->getSVMAllocsManager()->freeSVMAlloc(ptrSVM);
|
||||
}
|
||||
|
||||
TEST_F(CloneKernelTest, GivenUnifiedMemoryExecInfoWhenCloningKernelThenUnifiedMemoryAllocationIsCorrect) {
|
||||
TEST_F(CloneKernelTest, givenUnifiedMemoryExecInfoWhenCloningKernelThenUnifiedMemoryAllocationIsCorrect) {
|
||||
REQUIRE_SVM_OR_SKIP(device1);
|
||||
void *ptrSVM = context->getSVMAllocsManager()->createSVMAlloc(256, {}, context->getRootDeviceIndices(), context->getDeviceBitfields());
|
||||
ASSERT_NE(nullptr, ptrSVM);
|
||||
|
||||
@@ -28,7 +28,6 @@ TEST_F(KernelImageArgTest, GivenKernelWithImageArgsWhenCheckingDifferentScenario
|
||||
size_t imageWidth = image->getImageDesc().image_width;
|
||||
size_t imageHeight = image->getImageDesc().image_height;
|
||||
size_t imageDepth = image->getImageDesc().image_depth;
|
||||
uint32_t objectId = pKernelInfo->argAsImg(4).bindful;
|
||||
|
||||
cl_mem memObj = image.get();
|
||||
|
||||
@@ -49,8 +48,6 @@ TEST_F(KernelImageArgTest, GivenKernelWithImageArgsWhenCheckingDifferentScenario
|
||||
|
||||
auto imgDepthOffset = ptrOffset(crossThreadData, 0x30);
|
||||
EXPECT_EQ(imageDepth, *imgDepthOffset);
|
||||
|
||||
EXPECT_EQ(objectId, *crossThreadData);
|
||||
}
|
||||
|
||||
TEST_F(KernelImageArgTest, givenKernelWithFlatImageTokensWhenArgIsSetThenPatchAllParams) {
|
||||
|
||||
@@ -11,7 +11,7 @@
|
||||
|
||||
#include "patch_g7.h"
|
||||
|
||||
TEST_F(KernelDataTest, GIVENpatchTokenAllocateStatelessEventPoolSurfaceWHENdecodeTokensTHENtokenLocatedInPatchInfo) {
|
||||
TEST_F(KernelDataTest, givenPatchTokenAllocateStatelessEventPoolSurfaceWhenDecodeTokensThenTokenLocatedInPatchInfo) {
|
||||
iOpenCL::SPatchAllocateStatelessEventPoolSurface allocateStatelessEventPoolSurface;
|
||||
allocateStatelessEventPoolSurface.Token = PATCH_TOKEN_ALLOCATE_STATELESS_EVENT_POOL_SURFACE;
|
||||
allocateStatelessEventPoolSurface.Size = sizeof(SPatchAllocateStatelessEventPoolSurface);
|
||||
@@ -31,70 +31,7 @@ TEST_F(KernelDataTest, GIVENpatchTokenAllocateStatelessEventPoolSurfaceWHENdecod
|
||||
EXPECT_EQ_VAL(allocateStatelessEventPoolSurface.SurfaceStateHeapOffset, eventPoolArg.bindful);
|
||||
}
|
||||
|
||||
TEST_F(KernelDataTest, GIVENpatchTokenAllocateStatelessDefaultDeviceQueueSurfaceWHENdecodeTokensTHENtokenLocatedInPatchInfo) {
|
||||
iOpenCL::SPatchAllocateStatelessDefaultDeviceQueueSurface allocateStatelessDefaultDeviceQueueSurface;
|
||||
allocateStatelessDefaultDeviceQueueSurface.Token = PATCH_TOKEN_ALLOCATE_STATELESS_DEFAULT_DEVICE_QUEUE_SURFACE;
|
||||
allocateStatelessDefaultDeviceQueueSurface.Size = sizeof(SPatchAllocateStatelessDefaultDeviceQueueSurface);
|
||||
|
||||
allocateStatelessDefaultDeviceQueueSurface.DataParamSize = 7;
|
||||
allocateStatelessDefaultDeviceQueueSurface.DataParamOffset = 0xABC;
|
||||
allocateStatelessDefaultDeviceQueueSurface.SurfaceStateHeapOffset = 0xDEF;
|
||||
|
||||
pPatchList = &allocateStatelessDefaultDeviceQueueSurface;
|
||||
patchListSize = allocateStatelessDefaultDeviceQueueSurface.Size;
|
||||
|
||||
buildAndDecode();
|
||||
|
||||
const auto &defaultQueueSurfaceAddress = pKernelInfo->kernelDescriptor.payloadMappings.implicitArgs.deviceSideEnqueueDefaultQueueSurfaceAddress;
|
||||
EXPECT_EQ(allocateStatelessDefaultDeviceQueueSurface.DataParamOffset, defaultQueueSurfaceAddress.stateless);
|
||||
EXPECT_EQ(allocateStatelessDefaultDeviceQueueSurface.DataParamSize, defaultQueueSurfaceAddress.pointerSize);
|
||||
EXPECT_EQ(allocateStatelessDefaultDeviceQueueSurface.SurfaceStateHeapOffset, defaultQueueSurfaceAddress.bindful);
|
||||
}
|
||||
|
||||
TEST_F(KernelDataTest, GIVENpatchTokenStatelessDeviceQueueKernelArgumentWHENdecodeTokensTHENapropriateKernelArgInfoFilled) {
|
||||
iOpenCL::SPatchStatelessDeviceQueueKernelArgument deviceQueueKernelArgument;
|
||||
deviceQueueKernelArgument.Token = PATCH_TOKEN_STATELESS_DEVICE_QUEUE_KERNEL_ARGUMENT;
|
||||
deviceQueueKernelArgument.Size = sizeof(SPatchStatelessDeviceQueueKernelArgument);
|
||||
|
||||
deviceQueueKernelArgument.ArgumentNumber = 3;
|
||||
deviceQueueKernelArgument.DataParamSize = 7;
|
||||
deviceQueueKernelArgument.DataParamOffset = 0xABC;
|
||||
deviceQueueKernelArgument.SurfaceStateHeapOffset = 0xDEF;
|
||||
|
||||
pPatchList = &deviceQueueKernelArgument;
|
||||
patchListSize = deviceQueueKernelArgument.Size;
|
||||
|
||||
buildAndDecode();
|
||||
|
||||
ASSERT_GE(pKernelInfo->getExplicitArgs().size(), size_t(4u));
|
||||
EXPECT_TRUE(pKernelInfo->getArgDescriptorAt(3).getExtendedTypeInfo().isDeviceQueue);
|
||||
const auto &argAsPtr = pKernelInfo->getArgDescriptorAt(3).as<ArgDescPointer>();
|
||||
EXPECT_EQ(deviceQueueKernelArgument.DataParamOffset, argAsPtr.stateless);
|
||||
EXPECT_EQ(deviceQueueKernelArgument.DataParamSize, argAsPtr.pointerSize);
|
||||
EXPECT_EQ(deviceQueueKernelArgument.SurfaceStateHeapOffset, argAsPtr.bindful);
|
||||
}
|
||||
|
||||
TEST_F(KernelDataTest, GIVENdataParameterParentEventWHENdecodeTokensTHENoffsetLocatedInWorkloadInfo) {
|
||||
const uint32_t offsetSimdSize = 0xABC;
|
||||
|
||||
SPatchDataParameterBuffer dataParameterToken;
|
||||
dataParameterToken.Token = PATCH_TOKEN_DATA_PARAMETER_BUFFER;
|
||||
dataParameterToken.Size = sizeof(SPatchDataParameterBuffer);
|
||||
dataParameterToken.Type = DATA_PARAMETER_PARENT_EVENT;
|
||||
dataParameterToken.ArgumentNumber = 0;
|
||||
dataParameterToken.Offset = offsetSimdSize;
|
||||
dataParameterToken.DataSize = sizeof(uint32_t);
|
||||
dataParameterToken.SourceOffset = 0;
|
||||
|
||||
pPatchList = &dataParameterToken;
|
||||
patchListSize = dataParameterToken.Size;
|
||||
|
||||
buildAndDecode();
|
||||
|
||||
EXPECT_EQ(pKernelInfo->kernelDescriptor.payloadMappings.implicitArgs.deviceSideEnqueueParentEvent, offsetSimdSize);
|
||||
}
|
||||
|
||||
TEST_F(KernelDataTest, GIVENdataParameterPreferredWorkgroupMultipleTokenWHENbinaryIsdecodedTHENcorrectOffsetIsAssigned) {
|
||||
TEST_F(KernelDataTest, givenDataParameterPreferredWorkgroupMultipleTokenWhenBinaryIsdecodedThenCorrectOffsetIsAssigned) {
|
||||
const uint32_t offset = 0x100;
|
||||
|
||||
SPatchDataParameterBuffer dataParameterToken;
|
||||
@@ -113,27 +50,3 @@ TEST_F(KernelDataTest, GIVENdataParameterPreferredWorkgroupMultipleTokenWHENbina
|
||||
|
||||
EXPECT_EQ(pKernelInfo->kernelDescriptor.payloadMappings.implicitArgs.preferredWkgMultiple, offset);
|
||||
}
|
||||
|
||||
TEST_F(KernelDataTest, GIVENdataParameterObjectIdWHENdecodeTokensTHENoffsetLocatedInKernelArgInfo) {
|
||||
const uint32_t offsetObjectId = 0xABC;
|
||||
const uint32_t argNum = 7;
|
||||
|
||||
SPatchDataParameterBuffer dataParameterToken;
|
||||
dataParameterToken.Token = PATCH_TOKEN_DATA_PARAMETER_BUFFER;
|
||||
dataParameterToken.Size = sizeof(SPatchDataParameterBuffer);
|
||||
dataParameterToken.Type = DATA_PARAMETER_OBJECT_ID;
|
||||
dataParameterToken.ArgumentNumber = argNum;
|
||||
dataParameterToken.Offset = offsetObjectId;
|
||||
dataParameterToken.DataSize = sizeof(uint32_t);
|
||||
dataParameterToken.SourceOffset = 0;
|
||||
|
||||
pPatchList = &dataParameterToken;
|
||||
patchListSize = dataParameterToken.Size;
|
||||
|
||||
buildAndDecode();
|
||||
|
||||
ASSERT_GE(pKernelInfo->getExplicitArgs().size(), size_t(argNum + 1));
|
||||
EXPECT_TRUE(pKernelInfo->getArgDescriptorAt(argNum).getExtendedTypeInfo().hasDeviceSideEnqueueExtendedDescriptor);
|
||||
auto deviceSideEnqueueDesc = reinterpret_cast<NEO::ArgDescriptorDeviceSideEnqueue *>(pKernelInfo->kernelDescriptor.payloadMappings.explicitArgsExtendedDescriptors[argNum].get());
|
||||
EXPECT_EQ(offsetObjectId, deviceSideEnqueueDesc->objectId);
|
||||
}
|
||||
|
||||
@@ -43,7 +43,6 @@ class SamplerSetArgFixture : public ClDeviceFixture {
|
||||
|
||||
// setup kernel arg offsets
|
||||
pKernelInfo->addArgSampler(0, 0x40, 0x8, 0x10, 0x4);
|
||||
pKernelInfo->addExtendedDeviceSideEnqueueDescriptor(0, 0x0);
|
||||
|
||||
pKernelInfo->addArgSampler(1, 0x40);
|
||||
|
||||
@@ -311,10 +310,8 @@ HWTEST_F(SamplerSetArgTest, GivenFilteringNearestAndAddressingClampWhenSettingKe
|
||||
auto snapWaCrossThreadData = ptrOffset(crossThreadData, 0x4);
|
||||
|
||||
unsigned int snapWaValue = 0xffffffff;
|
||||
unsigned int objectId = SAMPLER_OBJECT_ID_SHIFT + pKernelInfo->argAsSmp(0).bindful;
|
||||
|
||||
EXPECT_EQ(snapWaValue, *snapWaCrossThreadData);
|
||||
EXPECT_EQ(objectId, *crossThreadData);
|
||||
}
|
||||
|
||||
HWTEST_F(SamplerSetArgTest, GivenKernelWithoutObjIdOffsetWhenSettingArgThenObjIdNotPatched) {
|
||||
|
||||
Reference in New Issue
Block a user