2022-03-29 15:39:27 +00:00
|
|
|
/*
|
2024-01-18 18:39:41 +00:00
|
|
|
* Copyright (C) 2021-2024 Intel Corporation
|
2022-03-29 15:39:27 +00:00
|
|
|
*
|
|
|
|
|
* SPDX-License-Identifier: MIT
|
|
|
|
|
*
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#include "level_zero/tools/source/debug/debug_session.h"
|
|
|
|
|
|
2024-02-06 22:01:29 +00:00
|
|
|
#include "shared/source/gmm_helper/gmm_helper.h"
|
|
|
|
|
#include "shared/source/helpers/aligned_memory.h"
|
2024-02-09 22:02:00 +00:00
|
|
|
#include "shared/source/helpers/basic_math.h"
|
2024-02-06 10:47:36 +00:00
|
|
|
#include "shared/source/helpers/hw_info.h"
|
2023-12-05 21:03:49 +00:00
|
|
|
#include "shared/source/helpers/sleep.h"
|
2024-02-06 22:01:29 +00:00
|
|
|
#include "shared/source/memory_manager/memory_manager.h"
|
2023-10-19 20:35:35 +00:00
|
|
|
#include "shared/source/os_interface/linux/drm_allocation.h"
|
|
|
|
|
#include "shared/source/os_interface/linux/drm_neo.h"
|
2023-12-05 21:03:49 +00:00
|
|
|
#include "shared/source/os_interface/linux/sys_calls.h"
|
2023-10-19 20:35:35 +00:00
|
|
|
|
2022-03-29 15:39:27 +00:00
|
|
|
#include "level_zero/core/source/device/device.h"
|
2024-02-06 10:47:36 +00:00
|
|
|
#include "level_zero/core/source/gfx_core_helpers/l0_gfx_core_helper.h"
|
|
|
|
|
#include "level_zero/include/zet_intel_gpu_debug.h"
|
2023-10-23 23:35:40 +00:00
|
|
|
#include "level_zero/tools/source/debug/linux/debug_session.h"
|
2023-10-19 20:35:35 +00:00
|
|
|
#include "level_zero/tools/source/debug/linux/debug_session_factory.h"
|
2024-02-06 10:47:36 +00:00
|
|
|
#include "level_zero/tools/source/debug/linux/drm_helper.h"
|
2022-03-29 15:39:27 +00:00
|
|
|
|
|
|
|
|
namespace L0 {
|
2023-11-23 17:09:04 +00:00
|
|
|
DebugSessionLinuxAllocatorFn debugSessionLinuxFactory[DEBUG_SESSION_LINUX_TYPE_MAX] = {};
|
2022-03-29 15:39:27 +00:00
|
|
|
|
2022-09-23 14:56:32 +00:00
|
|
|
DebugSession *DebugSession::create(const zet_debug_config_t &config, Device *device, ze_result_t &result, bool isRootAttach) {
|
2023-10-19 20:35:35 +00:00
|
|
|
|
|
|
|
|
if (!device->getOsInterface().isDebugAttachAvailable()) {
|
|
|
|
|
result = ZE_RESULT_ERROR_UNSUPPORTED_FEATURE;
|
|
|
|
|
return nullptr;
|
|
|
|
|
}
|
|
|
|
|
auto drm = device->getOsInterface().getDriverModel()->as<NEO::Drm>();
|
|
|
|
|
auto drmVersion = NEO::Drm::getDrmVersion(drm->getFileDescriptor());
|
|
|
|
|
|
|
|
|
|
DebugSessionLinuxAllocatorFn allocator = nullptr;
|
|
|
|
|
if ("xe" == drmVersion) {
|
2023-11-23 17:09:04 +00:00
|
|
|
allocator = debugSessionLinuxFactory[DEBUG_SESSION_LINUX_TYPE_XE];
|
2023-10-19 20:35:35 +00:00
|
|
|
} else {
|
2023-11-23 17:09:04 +00:00
|
|
|
allocator = debugSessionLinuxFactory[DEBUG_SESSION_LINUX_TYPE_I915];
|
2023-10-19 20:35:35 +00:00
|
|
|
}
|
2023-12-05 21:03:49 +00:00
|
|
|
UNRECOVERABLE_IF(!allocator)
|
2023-10-19 20:35:35 +00:00
|
|
|
|
|
|
|
|
auto debugSession = allocator(config, device, result, isRootAttach);
|
|
|
|
|
if (result != ZE_RESULT_SUCCESS) {
|
|
|
|
|
return nullptr;
|
|
|
|
|
}
|
|
|
|
|
debugSession->setAttachMode(isRootAttach);
|
|
|
|
|
result = debugSession->initialize();
|
|
|
|
|
if (result != ZE_RESULT_SUCCESS) {
|
|
|
|
|
debugSession->closeConnection();
|
|
|
|
|
delete debugSession;
|
|
|
|
|
debugSession = nullptr;
|
|
|
|
|
} else {
|
|
|
|
|
debugSession->startAsyncThread();
|
|
|
|
|
}
|
|
|
|
|
return debugSession;
|
2022-03-29 15:39:27 +00:00
|
|
|
}
|
2023-10-23 23:35:40 +00:00
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::translateDebuggerOpenErrno(int error) {
|
|
|
|
|
|
|
|
|
|
ze_result_t result = ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
switch (error) {
|
|
|
|
|
case ENODEV:
|
|
|
|
|
result = ZE_RESULT_ERROR_UNSUPPORTED_FEATURE;
|
|
|
|
|
break;
|
|
|
|
|
case EBUSY:
|
|
|
|
|
result = ZE_RESULT_ERROR_NOT_AVAILABLE;
|
|
|
|
|
break;
|
|
|
|
|
case EACCES:
|
|
|
|
|
result = ZE_RESULT_ERROR_INSUFFICIENT_PERMISSIONS;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
return result;
|
|
|
|
|
}
|
|
|
|
|
|
2023-12-05 21:03:49 +00:00
|
|
|
bool DebugSessionLinux::closeFd() {
|
|
|
|
|
if (fd == 0) {
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
auto ret = NEO::SysCalls::close(fd);
|
|
|
|
|
|
|
|
|
|
if (ret != 0) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("Debug connection close() on fd: %d failed: retCode: %d\n", fd, ret);
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
fd = 0;
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void *DebugSessionLinux::readInternalEventsThreadFunction(void *arg) {
|
|
|
|
|
DebugSessionLinux *self = reinterpret_cast<DebugSessionLinux *>(arg);
|
|
|
|
|
PRINT_DEBUGGER_INFO_LOG("Debugger internal event thread started\n", "");
|
|
|
|
|
self->internalThreadHasStarted = true;
|
|
|
|
|
|
|
|
|
|
while (self->internalEventThread.threadActive) {
|
|
|
|
|
self->readInternalEventsAsync();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
PRINT_DEBUGGER_INFO_LOG("Debugger internal event thread closing\n", "");
|
|
|
|
|
|
|
|
|
|
return nullptr;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
std::unique_ptr<uint64_t[]> DebugSessionLinux::getInternalEvent() {
|
|
|
|
|
std::unique_ptr<uint64_t[]> eventMemory;
|
|
|
|
|
|
|
|
|
|
{
|
|
|
|
|
std::unique_lock<std::mutex> lock(internalEventThreadMutex);
|
|
|
|
|
|
|
|
|
|
if (internalEventQueue.empty()) {
|
|
|
|
|
NEO::waitOnCondition(internalEventCondition, lock, std::chrono::milliseconds(100));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!internalEventQueue.empty()) {
|
|
|
|
|
eventMemory = std::move(internalEventQueue.front());
|
|
|
|
|
internalEventQueue.pop();
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
return eventMemory;
|
|
|
|
|
}
|
|
|
|
|
|
2024-01-18 18:39:41 +00:00
|
|
|
void DebugSessionLinux::closeAsyncThread() {
|
|
|
|
|
asyncThread.close();
|
|
|
|
|
internalEventThread.close();
|
|
|
|
|
}
|
|
|
|
|
|
2024-02-06 10:47:36 +00:00
|
|
|
bool DebugSessionLinux::checkForceExceptionBit(uint64_t memoryHandle, EuThread::ThreadId threadId, uint32_t *cr0, const SIP::regset_desc *regDesc) {
|
|
|
|
|
|
|
|
|
|
auto gpuVa = getContextStateSaveAreaGpuVa(memoryHandle);
|
|
|
|
|
auto threadSlotOffset = calculateThreadSlotOffset(threadId);
|
|
|
|
|
auto startRegOffset = threadSlotOffset + calculateRegisterOffsetInThreadSlot(regDesc, 0);
|
|
|
|
|
|
|
|
|
|
[[maybe_unused]] int ret = readGpuMemory(memoryHandle, reinterpret_cast<char *>(cr0), 1 * regDesc->bytes, gpuVa + startRegOffset);
|
|
|
|
|
DEBUG_BREAK_IF(ret != ZE_RESULT_SUCCESS);
|
|
|
|
|
|
|
|
|
|
const uint32_t cr0ForcedExcpetionBitmask = 0x04000000;
|
|
|
|
|
if (cr0[1] & cr0ForcedExcpetionBitmask) {
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void DebugSessionLinux::checkStoppedThreadsAndGenerateEvents(const std::vector<EuThread::ThreadId> &threads, uint64_t memoryHandle, uint32_t deviceIndex) {
|
|
|
|
|
|
|
|
|
|
std::vector<EuThread::ThreadId> threadsWithAttention;
|
|
|
|
|
std::vector<EuThread::ThreadId> stoppedThreadsToReport;
|
|
|
|
|
NEO::sleep(std::chrono::microseconds(1));
|
|
|
|
|
|
|
|
|
|
if (threads.size() > 1) {
|
|
|
|
|
auto hwInfo = connectedDevice->getHwInfo();
|
|
|
|
|
auto &l0GfxCoreHelper = connectedDevice->getL0GfxCoreHelper();
|
|
|
|
|
|
|
|
|
|
std::unique_ptr<uint8_t[]> bitmask;
|
|
|
|
|
size_t bitmaskSize;
|
|
|
|
|
[[maybe_unused]] auto attReadResult = threadControl(threads, deviceIndex, ThreadControlCmd::stopped, bitmask, bitmaskSize);
|
|
|
|
|
|
|
|
|
|
// error querying STOPPED threads - no threads available ( for example: threads have completed )
|
|
|
|
|
if (attReadResult != 0) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("checkStoppedThreadsAndGenerateEvents ATTENTION read failed: %d errno = %d \n", (int)attReadResult, DrmHelper::getErrno(connectedDevice));
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
threadsWithAttention = l0GfxCoreHelper.getThreadsFromAttentionBitmask(hwInfo, deviceIndex, bitmask.get(), bitmaskSize);
|
|
|
|
|
|
|
|
|
|
if (threadsWithAttention.size() == 0) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const auto &threadsToCheck = threadsWithAttention.size() > 0 ? threadsWithAttention : threads;
|
|
|
|
|
stoppedThreadsToReport.reserve(threadsToCheck.size());
|
|
|
|
|
|
|
|
|
|
const auto regSize = std::max(getRegisterSize(ZET_DEBUG_REGSET_TYPE_CR_INTEL_GPU), 64u);
|
|
|
|
|
auto cr0 = std::make_unique<uint32_t[]>(regSize / sizeof(uint32_t));
|
|
|
|
|
auto regDesc = typeToRegsetDesc(ZET_DEBUG_REGSET_TYPE_CR_INTEL_GPU);
|
|
|
|
|
|
|
|
|
|
for (auto &threadId : threadsToCheck) {
|
|
|
|
|
SIP::sr_ident srMagic = {{0}};
|
|
|
|
|
srMagic.count = 0;
|
|
|
|
|
|
|
|
|
|
if (readSystemRoutineIdent(allThreads[threadId].get(), memoryHandle, srMagic)) {
|
|
|
|
|
bool wasStopped = allThreads[threadId]->isStopped();
|
|
|
|
|
bool checkIfStopped = true;
|
|
|
|
|
|
|
|
|
|
if (srMagic.count % 2 == 1) {
|
|
|
|
|
memset(cr0.get(), 0, regSize);
|
|
|
|
|
checkIfStopped = !checkForceExceptionBit(memoryHandle, threadId, cr0.get(), regDesc);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (checkIfStopped && allThreads[threadId]->verifyStopped(srMagic.count)) {
|
|
|
|
|
allThreads[threadId]->stopThread(memoryHandle);
|
|
|
|
|
if (!wasStopped) {
|
|
|
|
|
stoppedThreadsToReport.push_back(threadId);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
} else {
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
generateEventsForStoppedThreads(stoppedThreadsToReport);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::resumeImp(const std::vector<EuThread::ThreadId> &threads, uint32_t deviceIndex) {
|
|
|
|
|
std::unique_ptr<uint8_t[]> bitmask;
|
|
|
|
|
size_t bitmaskSize;
|
|
|
|
|
|
|
|
|
|
auto result = threadControl(threads, deviceIndex, ThreadControlCmd::resume, bitmask, bitmaskSize);
|
|
|
|
|
|
|
|
|
|
return result == 0 ? ZE_RESULT_SUCCESS : ZE_RESULT_ERROR_NOT_AVAILABLE;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::interruptImp(uint32_t deviceIndex) {
|
|
|
|
|
std::unique_ptr<uint8_t[]> bitmask;
|
|
|
|
|
size_t bitmaskSize;
|
|
|
|
|
|
|
|
|
|
auto result = threadControl({}, deviceIndex, ThreadControlCmd::interruptAll, bitmask, bitmaskSize);
|
|
|
|
|
|
|
|
|
|
return result == 0 ? ZE_RESULT_SUCCESS : ZE_RESULT_ERROR_NOT_AVAILABLE;
|
|
|
|
|
}
|
|
|
|
|
|
2024-02-06 22:01:29 +00:00
|
|
|
ze_result_t DebugSessionLinux::readGpuMemory(uint64_t vmHandle, char *output, size_t size, uint64_t gpuVa) {
|
|
|
|
|
|
|
|
|
|
int vmDebugFd = openVmFd(vmHandle, true);
|
|
|
|
|
if (vmDebugFd < 0) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("VM_OPEN failed = %d\n", vmDebugFd);
|
|
|
|
|
return ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
int64_t retVal = 0;
|
|
|
|
|
auto gmmHelper = connectedDevice->getNEODevice()->getGmmHelper();
|
|
|
|
|
gpuVa = gmmHelper->decanonize(gpuVa);
|
|
|
|
|
if (flushVmCache(vmDebugFd) != 0) {
|
|
|
|
|
return ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
if (NEO::debugManager.flags.EnableDebuggerMmapMemoryAccess.get()) {
|
|
|
|
|
uint64_t alignedMem = alignDown(gpuVa, MemoryConstants::pageSize);
|
|
|
|
|
uint64_t alignedDiff = gpuVa - alignedMem;
|
|
|
|
|
uint64_t alignedSize = size + alignedDiff;
|
|
|
|
|
|
|
|
|
|
void *mappedPtr = ioctlHandler->mmap(NULL, alignedSize, PROT_READ, MAP_SHARED, vmDebugFd, alignedMem);
|
|
|
|
|
|
|
|
|
|
if (mappedPtr == MAP_FAILED) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("Reading memory failed, errno = %d\n", errno);
|
|
|
|
|
retVal = -1;
|
|
|
|
|
} else {
|
|
|
|
|
char *realSourceVA = static_cast<char *>(mappedPtr) + alignedDiff;
|
|
|
|
|
retVal = memcpy_s(output, size, static_cast<void *>(realSourceVA), size);
|
|
|
|
|
ioctlHandler->munmap(mappedPtr, alignedSize);
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
size_t pendingSize = size;
|
|
|
|
|
uint8_t retry = 0;
|
|
|
|
|
const uint8_t maxRetries = 3;
|
|
|
|
|
size_t retrySize = size;
|
|
|
|
|
do {
|
|
|
|
|
PRINT_DEBUGGER_MEM_ACCESS_LOG("Reading (pread) memory from gpu va = %#" PRIx64 ", size = %zu\n", gpuVa, pendingSize);
|
|
|
|
|
retVal = ioctlHandler->pread(vmDebugFd, output, pendingSize, gpuVa);
|
|
|
|
|
output += retVal;
|
|
|
|
|
gpuVa += retVal;
|
|
|
|
|
pendingSize -= retVal;
|
|
|
|
|
if (retVal == 0) {
|
|
|
|
|
if (pendingSize < retrySize) {
|
|
|
|
|
retry = 0;
|
|
|
|
|
}
|
|
|
|
|
retry++;
|
|
|
|
|
retrySize = pendingSize;
|
|
|
|
|
}
|
|
|
|
|
} while (((retVal == 0) && (retry < maxRetries)) || ((retVal > 0) && (pendingSize > 0)));
|
|
|
|
|
|
|
|
|
|
if (retVal < 0) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("Reading memory failed, errno = %d\n", errno);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
retVal = pendingSize;
|
|
|
|
|
}
|
|
|
|
|
if (flushVmCache(vmDebugFd) != 0) {
|
|
|
|
|
return ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
NEO::SysCalls::close(vmDebugFd);
|
|
|
|
|
|
|
|
|
|
return (retVal == 0) ? ZE_RESULT_SUCCESS : ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::writeGpuMemory(uint64_t vmHandle, const char *input, size_t size, uint64_t gpuVa) {
|
|
|
|
|
|
|
|
|
|
int vmDebugFd = openVmFd(vmHandle, false);
|
|
|
|
|
if (vmDebugFd < 0) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("VM_OPEN failed = %d\n", vmDebugFd);
|
|
|
|
|
return ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
int64_t retVal = 0;
|
|
|
|
|
auto gmmHelper = connectedDevice->getNEODevice()->getGmmHelper();
|
|
|
|
|
gpuVa = gmmHelper->decanonize(gpuVa);
|
|
|
|
|
if (flushVmCache(vmDebugFd) != 0) {
|
|
|
|
|
return ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
if (NEO::debugManager.flags.EnableDebuggerMmapMemoryAccess.get()) {
|
|
|
|
|
uint64_t alignedMem = alignDown(gpuVa, MemoryConstants::pageSize);
|
|
|
|
|
uint64_t alignedDiff = gpuVa - alignedMem;
|
|
|
|
|
uint64_t alignedSize = size + alignedDiff;
|
|
|
|
|
|
|
|
|
|
void *mappedPtr = ioctlHandler->mmap(NULL, alignedSize, PROT_WRITE, MAP_SHARED, vmDebugFd, alignedMem);
|
|
|
|
|
|
|
|
|
|
if (mappedPtr == MAP_FAILED) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("Writing memory failed, errno = %d\n", errno);
|
|
|
|
|
retVal = -1;
|
|
|
|
|
} else {
|
|
|
|
|
char *realDestVA = static_cast<char *>(mappedPtr) + alignedDiff;
|
|
|
|
|
retVal = memcpy_s(static_cast<void *>(realDestVA), size, input, size);
|
|
|
|
|
ioctlHandler->munmap(mappedPtr, alignedSize);
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
size_t pendingSize = size;
|
|
|
|
|
uint8_t retry = 0;
|
|
|
|
|
const uint8_t maxRetries = 3;
|
|
|
|
|
size_t retrySize = size;
|
|
|
|
|
do {
|
|
|
|
|
PRINT_DEBUGGER_MEM_ACCESS_LOG("Writing (pwrite) memory to gpu va = %#" PRIx64 ", size = %zu\n", gpuVa, pendingSize);
|
|
|
|
|
retVal = ioctlHandler->pwrite(vmDebugFd, input, pendingSize, gpuVa);
|
|
|
|
|
input += retVal;
|
|
|
|
|
gpuVa += retVal;
|
|
|
|
|
pendingSize -= retVal;
|
|
|
|
|
if (retVal == 0) {
|
|
|
|
|
if (pendingSize < retrySize) {
|
|
|
|
|
retry = 0;
|
|
|
|
|
}
|
|
|
|
|
retry++;
|
|
|
|
|
retrySize = pendingSize;
|
|
|
|
|
}
|
|
|
|
|
} while (((retVal == 0) && (retry < maxRetries)) || ((retVal > 0) && (pendingSize > 0)));
|
|
|
|
|
|
|
|
|
|
if (retVal < 0) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("Writing memory failed, errno = %d\n", errno);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
retVal = pendingSize;
|
|
|
|
|
}
|
|
|
|
|
if (flushVmCache(vmDebugFd) != 0) {
|
|
|
|
|
return ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
NEO::SysCalls::close(vmDebugFd);
|
|
|
|
|
|
|
|
|
|
return (retVal == 0) ? ZE_RESULT_SUCCESS : ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
|
2024-02-06 16:53:04 +00:00
|
|
|
int DebugSessionLinux::threadControl(const std::vector<EuThread::ThreadId> &threads, uint32_t tile, ThreadControlCmd threadCmd, std::unique_ptr<uint8_t[]> &bitmaskOut, size_t &bitmaskSizeOut) {
|
|
|
|
|
|
|
|
|
|
auto hwInfo = connectedDevice->getHwInfo();
|
|
|
|
|
auto classInstance = DrmHelper::getEngineInstance(connectedDevice, tile, hwInfo.capabilityTable.defaultEngineType);
|
|
|
|
|
UNRECOVERABLE_IF(!classInstance);
|
|
|
|
|
|
|
|
|
|
auto &l0GfxCoreHelper = connectedDevice->getL0GfxCoreHelper();
|
|
|
|
|
|
|
|
|
|
bitmaskSizeOut = 0;
|
|
|
|
|
std::unique_ptr<uint8_t[]> bitmask;
|
|
|
|
|
size_t bitmaskSize = 0;
|
|
|
|
|
|
|
|
|
|
if (threadCmd == ThreadControlCmd::interrupt ||
|
|
|
|
|
threadCmd == ThreadControlCmd::resume ||
|
|
|
|
|
threadCmd == ThreadControlCmd::stopped) {
|
|
|
|
|
l0GfxCoreHelper.getAttentionBitmaskForSingleThreads(threads, hwInfo, bitmask, bitmaskSize);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
uint64_t seqnoRet = 0;
|
|
|
|
|
uint64_t bitmaskSizeRet = 0;
|
|
|
|
|
auto euControlRetVal = euControlIoctl(threadCmd, classInstance, bitmask, bitmaskSize, seqnoRet, bitmaskSizeRet);
|
|
|
|
|
if (euControlRetVal != 0) {
|
|
|
|
|
PRINT_DEBUGGER_ERROR_LOG("euControl IOCTL failed: retCode: %d errno = %d threadCmd = %d\n", euControlRetVal, errno, threadCmd);
|
|
|
|
|
} else {
|
|
|
|
|
PRINT_DEBUGGER_INFO_LOG("euControl IOCTL: seqno = %llu threadCmd = %u\n", seqnoRet, threadCmd);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (threadCmd == ThreadControlCmd::interrupt ||
|
|
|
|
|
threadCmd == ThreadControlCmd::interruptAll) {
|
|
|
|
|
if (euControlRetVal == 0) {
|
|
|
|
|
euControlInterruptSeqno[tile] = seqnoRet;
|
|
|
|
|
} else {
|
|
|
|
|
euControlInterruptSeqno[tile] = invalidHandle;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (threadCmd == ThreadControlCmd::stopped) {
|
|
|
|
|
bitmaskOut = std::move(bitmask);
|
|
|
|
|
bitmaskSizeOut = bitmaskSizeRet;
|
|
|
|
|
}
|
|
|
|
|
return euControlRetVal;
|
|
|
|
|
}
|
|
|
|
|
|
2024-02-09 22:02:00 +00:00
|
|
|
ze_result_t DebugSessionLinux::readElfSpace(const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer,
|
|
|
|
|
const char *&elfData, const uint64_t offset) {
|
|
|
|
|
|
|
|
|
|
int retVal = -1;
|
|
|
|
|
elfData += offset;
|
|
|
|
|
retVal = memcpy_s(buffer, size, elfData, size);
|
|
|
|
|
return (retVal == 0) ? ZE_RESULT_SUCCESS : ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::readMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer) {
|
|
|
|
|
ze_result_t status = validateThreadAndDescForMemoryAccess(thread, desc);
|
|
|
|
|
if (status != ZE_RESULT_SUCCESS) {
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (desc->type == ZET_DEBUG_MEMORY_SPACE_TYPE_DEFAULT) {
|
|
|
|
|
status = readDefaultMemory(thread, desc, size, buffer);
|
|
|
|
|
} else {
|
|
|
|
|
auto threadId = convertToThreadId(thread);
|
|
|
|
|
status = slmMemoryAccess<void *, false>(threadId, desc, size, buffer);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::readDefaultMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer) {
|
|
|
|
|
ze_result_t status = ZE_RESULT_SUCCESS;
|
|
|
|
|
|
|
|
|
|
bool isa = tryReadIsa(connectedDevice->getNEODevice()->getDeviceBitfield(), desc, size, buffer, status);
|
|
|
|
|
if (isa) {
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
bool elf = tryReadElf(desc, size, buffer, status);
|
|
|
|
|
if (elf) {
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (DebugSession::isThreadAll(thread)) {
|
|
|
|
|
return accessDefaultMemForThreadAll(desc, size, const_cast<void *>(buffer), false);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
auto threadId = convertToThreadId(thread);
|
|
|
|
|
auto vmHandle = allThreads[threadId]->getMemoryHandle();
|
|
|
|
|
if (vmHandle == invalidHandle) {
|
|
|
|
|
return ZE_RESULT_ERROR_NOT_AVAILABLE;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return readGpuMemory(vmHandle, static_cast<char *>(buffer), size, desc->address);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::writeMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, const void *buffer) {
|
|
|
|
|
ze_result_t status = validateThreadAndDescForMemoryAccess(thread, desc);
|
|
|
|
|
if (status != ZE_RESULT_SUCCESS) {
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (desc->type == ZET_DEBUG_MEMORY_SPACE_TYPE_DEFAULT) {
|
|
|
|
|
status = writeDefaultMemory(thread, desc, size, buffer);
|
|
|
|
|
} else {
|
|
|
|
|
auto threadId = convertToThreadId(thread);
|
|
|
|
|
status = slmMemoryAccess<const void *, true>(threadId, desc, size, buffer);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::writeDefaultMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, const void *buffer) {
|
|
|
|
|
ze_result_t status = ZE_RESULT_SUCCESS;
|
|
|
|
|
|
|
|
|
|
auto deviceBitfield = connectedDevice->getNEODevice()->getDeviceBitfield();
|
|
|
|
|
|
|
|
|
|
bool isa = tryWriteIsa(deviceBitfield, desc, size, buffer, status);
|
|
|
|
|
if (isa) {
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (DebugSession::isThreadAll(thread)) {
|
|
|
|
|
return accessDefaultMemForThreadAll(desc, size, const_cast<void *>(buffer), true);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
auto threadId = convertToThreadId(thread);
|
|
|
|
|
auto threadVmHandle = allThreads[threadId]->getMemoryHandle();
|
|
|
|
|
if (threadVmHandle == invalidHandle) {
|
|
|
|
|
return ZE_RESULT_ERROR_NOT_AVAILABLE;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return writeGpuMemory(threadVmHandle, static_cast<const char *>(buffer), size, desc->address);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
bool DebugSessionLinux::tryWriteIsa(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, const void *buffer, ze_result_t &status) {
|
|
|
|
|
return tryAccessIsa(deviceBitfield, desc, size, const_cast<void *>(buffer), true, status);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
bool DebugSessionLinux::tryReadIsa(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, ze_result_t &status) {
|
|
|
|
|
return tryAccessIsa(deviceBitfield, desc, size, buffer, false, status);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
bool DebugSessionLinux::tryReadElf(const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, ze_result_t &status) {
|
|
|
|
|
const char *elfData = nullptr;
|
|
|
|
|
uint64_t offset = 0;
|
|
|
|
|
|
|
|
|
|
std::lock_guard<std::mutex> memLock(asyncThreadMutex);
|
|
|
|
|
|
|
|
|
|
status = getElfOffset(desc, size, elfData, offset);
|
|
|
|
|
if (status == ZE_RESULT_ERROR_INVALID_ARGUMENT) {
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (elfData) {
|
|
|
|
|
status = readElfSpace(desc, size, buffer, elfData, offset);
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::accessDefaultMemForThreadAll(const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, bool write) {
|
|
|
|
|
auto status = ZE_RESULT_ERROR_UNINITIALIZED;
|
|
|
|
|
std::vector<uint64_t> allVms;
|
|
|
|
|
|
|
|
|
|
allVms = getAllMemoryHandles();
|
|
|
|
|
|
|
|
|
|
if (allVms.size() > 0) {
|
|
|
|
|
for (auto vmHandle : allVms) {
|
|
|
|
|
if (write) {
|
|
|
|
|
status = writeGpuMemory(vmHandle, static_cast<char *>(buffer), size, desc->address);
|
|
|
|
|
} else {
|
|
|
|
|
status = readGpuMemory(vmHandle, static_cast<char *>(buffer), size, desc->address);
|
|
|
|
|
}
|
|
|
|
|
if (status == ZE_RESULT_SUCCESS) {
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
status = ZE_RESULT_ERROR_NOT_AVAILABLE;
|
|
|
|
|
}
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
bool DebugSessionLinux::tryAccessIsa(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, bool write, ze_result_t &status) {
|
|
|
|
|
status = ZE_RESULT_ERROR_NOT_AVAILABLE;
|
|
|
|
|
uint64_t vmHandle[NEO::EngineLimits::maxHandleCount] = {invalidHandle};
|
|
|
|
|
uint32_t deviceIndex = Math::getMinLsbSet(static_cast<uint32_t>(deviceBitfield.to_ulong()));
|
|
|
|
|
|
|
|
|
|
bool isaAccess = false;
|
|
|
|
|
|
|
|
|
|
auto checkIfAnyFailed = [](const auto &result) { return result != ZE_RESULT_SUCCESS; };
|
|
|
|
|
|
|
|
|
|
{
|
|
|
|
|
std::lock_guard<std::mutex> memLock(asyncThreadMutex);
|
|
|
|
|
|
|
|
|
|
if (deviceBitfield.count() == 1) {
|
|
|
|
|
status = getISAVMHandle(deviceIndex, desc, size, vmHandle[deviceIndex]);
|
|
|
|
|
if (status == ZE_RESULT_SUCCESS) {
|
|
|
|
|
isaAccess = true;
|
|
|
|
|
}
|
|
|
|
|
if (status == ZE_RESULT_ERROR_INVALID_ARGUMENT) {
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
isaAccess = getIsaInfoForAllInstances(deviceBitfield, desc, size, vmHandle, status);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (isaAccess && status == ZE_RESULT_SUCCESS) {
|
|
|
|
|
|
|
|
|
|
if (write) {
|
|
|
|
|
if (deviceBitfield.count() == 1) {
|
|
|
|
|
if (vmHandle[deviceIndex] != invalidHandle) {
|
|
|
|
|
status = writeGpuMemory(vmHandle[deviceIndex], static_cast<char *>(buffer), size, desc->address);
|
|
|
|
|
} else {
|
|
|
|
|
status = ZE_RESULT_ERROR_UNINITIALIZED;
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
std::vector<ze_result_t> results(NEO::EngineLimits::maxHandleCount);
|
|
|
|
|
|
|
|
|
|
for (uint32_t i = 0; i < NEO::EngineLimits::maxHandleCount; i++) {
|
|
|
|
|
results[i] = ZE_RESULT_SUCCESS;
|
|
|
|
|
|
|
|
|
|
if (deviceBitfield.test(i) && vmHandle[i] != invalidHandle) {
|
|
|
|
|
results[i] = writeGpuMemory(vmHandle[i], static_cast<char *>(buffer), size, desc->address);
|
|
|
|
|
|
|
|
|
|
if (results[i] != ZE_RESULT_SUCCESS) {
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
const bool anyFailed = std::any_of(results.begin(), results.end(), checkIfAnyFailed);
|
|
|
|
|
|
|
|
|
|
if (anyFailed) {
|
|
|
|
|
status = ZE_RESULT_ERROR_UNKNOWN;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
|
|
|
|
|
if (deviceBitfield.count() > 1) {
|
|
|
|
|
for (uint32_t i = 0; i < NEO::EngineLimits::maxHandleCount; i++) {
|
|
|
|
|
if (vmHandle[i] != invalidHandle) {
|
|
|
|
|
deviceIndex = i;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (vmHandle[deviceIndex] != invalidHandle) {
|
|
|
|
|
status = readGpuMemory(vmHandle[deviceIndex], static_cast<char *>(buffer), size, desc->address);
|
|
|
|
|
} else {
|
|
|
|
|
status = ZE_RESULT_ERROR_UNINITIALIZED;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
return isaAccess;
|
|
|
|
|
}
|
|
|
|
|
|
2024-02-13 21:49:56 +00:00
|
|
|
std::vector<uint64_t> DebugSessionLinux::getAllMemoryHandles() {
|
|
|
|
|
std::vector<uint64_t> allVms;
|
|
|
|
|
std::unique_lock<std::mutex> memLock(asyncThreadMutex);
|
|
|
|
|
|
|
|
|
|
auto &vmIds = getClientConnection(clientHandle)->vmIds;
|
|
|
|
|
allVms.resize(vmIds.size());
|
|
|
|
|
std::copy(vmIds.begin(), vmIds.end(), allVms.begin());
|
|
|
|
|
return allVms;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ze_result_t DebugSessionLinux::getElfOffset(const zet_debug_memory_space_desc_t *desc, size_t size, const char *&elfData, uint64_t &offset) {
|
|
|
|
|
auto clientConnection = getClientConnection(clientHandle);
|
|
|
|
|
auto &elfMap = clientConnection->elfMap;
|
|
|
|
|
auto accessVA = desc->address;
|
|
|
|
|
ze_result_t status = ZE_RESULT_ERROR_UNINITIALIZED;
|
|
|
|
|
elfData = nullptr;
|
|
|
|
|
|
|
|
|
|
if (elfMap.size() > 0) {
|
|
|
|
|
uint64_t baseVa;
|
|
|
|
|
uint64_t ceilVa;
|
|
|
|
|
for (auto &elf : elfMap) {
|
|
|
|
|
baseVa = elf.first;
|
|
|
|
|
ceilVa = elf.first + clientConnection->getElfSize(elf.second);
|
|
|
|
|
if (accessVA >= baseVa && accessVA < ceilVa) {
|
|
|
|
|
if (accessVA + size > ceilVa) {
|
|
|
|
|
status = ZE_RESULT_ERROR_INVALID_ARGUMENT;
|
|
|
|
|
} else {
|
|
|
|
|
DEBUG_BREAK_IF(clientConnection->getElfData(elf.second) == nullptr);
|
|
|
|
|
elfData = clientConnection->getElfData(elf.second);
|
|
|
|
|
offset = accessVA - baseVa;
|
|
|
|
|
status = ZE_RESULT_SUCCESS;
|
|
|
|
|
}
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return status;
|
|
|
|
|
}
|
|
|
|
|
|
2022-03-29 15:39:27 +00:00
|
|
|
} // namespace L0
|