/* * Copyright (C) 2022-2023 Intel Corporation * * SPDX-License-Identifier: MIT * */ #pragma once #include "shared/source/helpers/basic_math.h" #include "shared/source/helpers/topology_map.h" #include "shared/source/os_interface/linux/drm_debug.h" #include "shared/source/os_interface/linux/i915_prelim.h" #include "shared/source/os_interface/linux/sys_calls.h" #include "level_zero/core/source/device/device.h" #include "level_zero/core/source/device/device_imp.h" #include "level_zero/tools/source/debug/debug_session.h" #include "level_zero/tools/source/debug/debug_session_imp.h" #include #include #include #include namespace NEO { class Thread; struct EngineClassInstance; } // namespace NEO namespace L0 { struct TileDebugSessionLinux; struct DebugSessionLinux : DebugSessionImp { friend struct TileDebugSessionLinux; ~DebugSessionLinux() override; DebugSessionLinux(const zet_debug_config_t &config, Device *device, int debugFd, void *params); ze_result_t initialize() override; bool closeConnection() override; ze_result_t readMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer) override; ze_result_t writeMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, const void *buffer) override; ze_result_t acknowledgeEvent(const zet_debug_event_t *event) override; struct IoctlHandler { MOCKABLE_VIRTUAL ~IoctlHandler() = default; MOCKABLE_VIRTUAL int ioctl(int fd, unsigned long request, void *arg) { int ret = 0; int error = 0; bool shouldRetryIoctl = false; do { shouldRetryIoctl = false; ret = NEO::SysCalls::ioctl(fd, request, arg); error = errno; if (ret == -1) { shouldRetryIoctl = (error == EINTR || error == EAGAIN || error == EBUSY); if (request == PRELIM_I915_DEBUG_IOCTL_EU_CONTROL) { shouldRetryIoctl = (error == EINTR || error == EAGAIN); } } } while (shouldRetryIoctl); return ret; } MOCKABLE_VIRTUAL int poll(pollfd *pollFd, unsigned long int numberOfFds, int timeout) { return NEO::SysCalls::poll(pollFd, numberOfFds, timeout); } MOCKABLE_VIRTUAL int64_t pread(int fd, void *buf, size_t count, off_t offset) { return NEO::SysCalls::pread(fd, buf, count, offset); } MOCKABLE_VIRTUAL int64_t pwrite(int fd, const void *buf, size_t count, off_t offset) { return NEO::SysCalls::pwrite(fd, buf, count, offset); } MOCKABLE_VIRTUAL void *mmap(void *addr, size_t size, int prot, int flags, int fd, off_t off) { return NEO::SysCalls::mmap(addr, size, prot, flags, fd, off); } MOCKABLE_VIRTUAL int munmap(void *addr, size_t size) { return NEO::SysCalls::munmap(addr, size); } }; static constexpr size_t maxEventSize = 4096; using ContextHandle = uint64_t; struct ContextParams { ContextHandle handle = 0; uint64_t vm = UINT64_MAX; std::vector engines; }; struct UuidData { uint64_t handle = 0; uint64_t classHandle = 0; NEO::DrmResourceClass classIndex = NEO::DrmResourceClass::MaxSize; std::unique_ptr data; size_t dataSize = 0; uint64_t ptr = 0; }; struct BindInfo { uint64_t gpuVa = 0; uint64_t size = 0; }; struct IsaAllocation { BindInfo bindInfo; uint64_t elfUuidHandle; uint64_t vmHandle; bool tileInstanced = false; bool perKernelModule = true; NEO::DeviceBitfield deviceBitfield; uint64_t moduleBegin; uint64_t moduleEnd; std::unordered_set cookies; int vmBindCounter = 0; bool moduleLoadEventAck = false; std::vector ackEvents; }; struct Module { std::unordered_set loadAddresses[NEO::EngineLimits::maxHandleCount]; uint64_t moduleUuidHandle; uint64_t elfUuidHandle; uint32_t segmentCount; NEO::DeviceBitfield deviceBitfield; int segmentVmBindCounter[NEO::EngineLimits::maxHandleCount]; std::vector ackEvents[NEO::EngineLimits::maxHandleCount]; bool moduleLoadEventAcked[NEO::EngineLimits::maxHandleCount]; }; static bool apiEventCompare(const zet_debug_event_t &event1, const zet_debug_event_t &event2) { return memcmp(&event1, &event2, sizeof(zet_debug_event_t)) == 0; }; struct ClientConnection { prelim_drm_i915_debug_event_client client = {}; std::unordered_map contextsCreated; std::unordered_map> classHandleToIndex; std::unordered_map uuidMap; std::unordered_set vmIds; std::unordered_map vmToModuleDebugAreaBindInfo; std::unordered_map vmToContextStateSaveAreaBindInfo; std::unordered_map vmToStateBaseAreaBindInfo; std::unordered_map vmToTile; std::unordered_map> isaMap[NEO::EngineLimits::maxHandleCount]; std::unordered_map elfMap; std::unordered_map lrcToContextHandle; uint64_t moduleDebugAreaGpuVa = 0; uint64_t contextStateSaveAreaGpuVa = 0; uint64_t stateBaseAreaGpuVa = 0; std::unordered_map uuidToModule; }; static ze_result_t translateDebuggerOpenErrno(int error); constexpr static uint64_t invalidClientHandle = std::numeric_limits::max(); constexpr static uint64_t invalidHandle = std::numeric_limits::max(); protected: enum class ThreadControlCmd { Interrupt, Resume, Stopped, InterruptAll }; MOCKABLE_VIRTUAL void handleEvent(prelim_drm_i915_debug_event *event); bool checkAllEventsCollected(); ze_result_t readEventImp(prelim_drm_i915_debug_event *drmDebugEvent); ze_result_t resumeImp(const std::vector &threads, uint32_t deviceIndex) override; ze_result_t interruptImp(uint32_t deviceIndex) override; void checkStoppedThreadsAndGenerateEvents(const std::vector &threads, uint64_t memoryHandle, uint32_t deviceIndex) override; void enqueueApiEvent(zet_debug_event_t &debugEvent) override { pushApiEvent(debugEvent); } void pushApiEvent(zet_debug_event_t &debugEvent) { return pushApiEvent(debugEvent, invalidHandle); } void pushApiEvent(zet_debug_event_t &debugEvent, uint64_t moduleUuidHandle) { std::unique_lock lock(asyncThreadMutex); if (moduleUuidHandle != invalidHandle && (debugEvent.flags & ZET_DEBUG_EVENT_FLAG_NEED_ACK)) { eventsToAck.push_back( std::pair(debugEvent, moduleUuidHandle)); } apiEvents.push(debugEvent); apiEventCondition.notify_all(); } MOCKABLE_VIRTUAL void createTileSessionsIfEnabled(); MOCKABLE_VIRTUAL TileDebugSessionLinux *createTileSession(const zet_debug_config_t &config, Device *device, DebugSessionImp *rootDebugSession); static void *asyncThreadFunction(void *arg); static void *readInternalEventsThreadFunction(void *arg); void startAsyncThread() override; void closeAsyncThread(); MOCKABLE_VIRTUAL void startInternalEventsThread() { internalEventThread.thread = NEO::Thread::create(readInternalEventsThreadFunction, reinterpret_cast(this)); } void closeInternalEventsThread() { internalEventThread.close(); } bool closeFd(); virtual std::vector getAllMemoryHandles() { std::vector allVms; std::unique_lock memLock(asyncThreadMutex); auto &vmIds = clientHandleToConnection[clientHandle]->vmIds; allVms.resize(vmIds.size()); std::copy(vmIds.begin(), vmIds.end(), allVms.begin()); return allVms; } void handleEventsAsync(); void readInternalEventsAsync(); MOCKABLE_VIRTUAL std::unique_ptr getInternalEvent(); MOCKABLE_VIRTUAL float getThreadStartLimitTime() { return 0.5; } bool handleVmBindEvent(prelim_drm_i915_debug_event_vm_bind *vmBind); void handleContextParamEvent(prelim_drm_i915_debug_event_context_param *contextParam); void handleAttentionEvent(prelim_drm_i915_debug_event_eu_attention *attention); void handleEnginesEvent(prelim_drm_i915_debug_event_engines *engines); virtual bool ackIsaEvents(uint32_t deviceIndex, uint64_t isaVa); virtual bool ackModuleEvents(uint32_t deviceIndex, uint64_t moduleUuidHandle); MOCKABLE_VIRTUAL void processPendingVmBindEvents(); void attachTile() override { UNRECOVERABLE_IF(true); } void detachTile() override { UNRECOVERABLE_IF(true); } void cleanRootSessionAfterDetach(uint32_t deviceIndex) override; void extractUuidData(uint64_t client, const UuidData &uuidData); uint64_t extractVaFromUuidString(std::string &uuid); bool readModuleDebugArea() override; ze_result_t readSbaBuffer(EuThread::ThreadId, NEO::SbaTrackedAddresses &sbaBuffer) override; void readStateSaveAreaHeader() override; ze_result_t readGpuMemory(uint64_t vmHandle, char *output, size_t size, uint64_t gpuVa) override; ze_result_t writeGpuMemory(uint64_t vmHandle, const char *input, size_t size, uint64_t gpuVa) override; ze_result_t getISAVMHandle(uint32_t deviceIndex, const zet_debug_memory_space_desc_t *desc, size_t size, uint64_t &vmHandle); bool getIsaInfoForAllInstances(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, uint64_t vmHandles[], ze_result_t &status); ze_result_t getElfOffset(const zet_debug_memory_space_desc_t *desc, size_t size, const char *&elfData, uint64_t &offset); ze_result_t readElfSpace(const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, const char *&elfData, const uint64_t offset); virtual bool tryReadElf(const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, ze_result_t &status); bool tryWriteIsa(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, const void *buffer, ze_result_t &status); bool tryReadIsa(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, ze_result_t &status); virtual bool tryAccessIsa(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, bool write, ze_result_t &status); ze_result_t accessDefaultMemForThreadAll(const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, bool write); ze_result_t readDefaultMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer); ze_result_t writeDefaultMemory(ze_device_thread_t thread, const zet_debug_memory_space_desc_t *desc, size_t size, const void *buffer); MOCKABLE_VIRTUAL int threadControl(const std::vector &threads, uint32_t tile, ThreadControlCmd threadCmd, std::unique_ptr &bitmask, size_t &bitmaskSize); uint64_t getContextStateSaveAreaGpuVa(uint64_t memoryHandle) override; virtual uint64_t getSbaBufferGpuVa(uint64_t memoryHandle); void printContextVms(); bool isTileWithinDeviceBitfield(uint32_t tileIndex) { return connectedDevice->getNEODevice()->getDeviceBitfield().test(tileIndex); } bool checkAllOtherTileIsaAllocationsPresent(uint32_t tileIndex, uint64_t isaVa) { bool allInstancesPresent = true; for (uint32_t i = 0; i < NEO::EngineLimits::maxHandleCount; i++) { if (i != tileIndex && connectedDevice->getNEODevice()->getDeviceBitfield().test(i)) { if (clientHandleToConnection[clientHandle]->isaMap[i].find(isaVa) == clientHandleToConnection[clientHandle]->isaMap[i].end()) { allInstancesPresent = false; break; } } } return allInstancesPresent; } bool checkAllOtherTileIsaAllocationsRemoved(uint32_t tileIndex, uint64_t isaVa) { bool allInstancesRemoved = true; for (uint32_t i = 0; i < NEO::EngineLimits::maxHandleCount; i++) { if (i != tileIndex && connectedDevice->getNEODevice()->getDeviceBitfield().test(i)) { if (clientHandleToConnection[clientHandle]->isaMap[i].find(isaVa) != clientHandleToConnection[clientHandle]->isaMap[i].end()) { allInstancesRemoved = false; break; } } } return allInstancesRemoved; } bool checkAllOtherTileModuleSegmentsPresent(uint32_t tileIndex, const Module &module) { bool allInstancesPresent = true; for (uint32_t i = 0; i < NEO::EngineLimits::maxHandleCount; i++) { if (i != tileIndex && connectedDevice->getNEODevice()->getDeviceBitfield().test(i)) { if (module.loadAddresses[i].size() != module.segmentCount) { allInstancesPresent = false; break; } } } return allInstancesPresent; } bool checkAllOtherTileModuleSegmentsRemoved(uint32_t tileIndex, const Module &module) { bool allInstancesRemoved = true; for (uint32_t i = 0; i < NEO::EngineLimits::maxHandleCount; i++) { if (i != tileIndex && connectedDevice->getNEODevice()->getDeviceBitfield().test(i)) { if (module.loadAddresses[i].size() != 0) { allInstancesRemoved = false; break; } } } return allInstancesRemoved; } ThreadHelper internalEventThread; std::mutex internalEventThreadMutex; std::condition_variable internalEventCondition; std::queue> internalEventQueue; std::vector> eventsToAck; // debug event, uuid handle to module std::vector> pendingVmBindEvents; int fd = 0; uint32_t i915DebuggerVersion = 0; virtual int ioctl(unsigned long request, void *arg); std::unique_ptr ioctlHandler; std::atomic detached{false}; uint64_t clientHandle = invalidClientHandle; uint64_t clientHandleClosed = invalidClientHandle; std::unordered_map uuidL0CommandQueueHandleToDevice; uint64_t euControlInterruptSeqno[NEO::EngineLimits::maxHandleCount]; std::unordered_map> clientHandleToConnection; std::atomic internalThreadHasStarted{false}; bool blockOnFenceMode = false; // false - blocking VM_BIND on CPU - autoack events until last blocking event // true - blocking on fence - do not auto-ack events }; struct TileDebugSessionLinux : DebugSessionLinux { TileDebugSessionLinux(zet_debug_config_t config, Device *device, DebugSessionImp *rootDebugSession) : DebugSessionLinux(config, device, 0, nullptr), rootDebugSession(reinterpret_cast(rootDebugSession)) { tileIndex = Math::log2(static_cast(connectedDevice->getNEODevice()->getDeviceBitfield().to_ulong())); } ~TileDebugSessionLinux() override = default; bool closeConnection() override { return true; } ze_result_t initialize() override { createEuThreads(); return ZE_RESULT_SUCCESS; } bool insertModule(zet_debug_event_info_module_t module); bool removeModule(zet_debug_event_info_module_t module); bool processEntry(); bool processExit(); void attachTile() override; void detachTile() override; bool isAttached = false; protected: void startAsyncThread() override { UNRECOVERABLE_IF(true); }; void cleanRootSessionAfterDetach(uint32_t deviceIndex) override { UNRECOVERABLE_IF(true); }; bool readModuleDebugArea() override { return true; }; uint64_t getContextStateSaveAreaGpuVa(uint64_t memoryHandle) override { return rootDebugSession->getContextStateSaveAreaGpuVa(memoryHandle); }; void readStateSaveAreaHeader() override; uint64_t getSbaBufferGpuVa(uint64_t memoryHandle) override { return rootDebugSession->getSbaBufferGpuVa(memoryHandle); } int ioctl(unsigned long request, void *arg) override { return rootDebugSession->ioctl(request, arg); } std::vector getAllMemoryHandles() override { return rootDebugSession->getAllMemoryHandles(); } bool tryReadElf(const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, ze_result_t &status) override { return rootDebugSession->tryReadElf(desc, size, buffer, status); } bool tryAccessIsa(NEO::DeviceBitfield deviceBitfield, const zet_debug_memory_space_desc_t *desc, size_t size, void *buffer, bool write, ze_result_t &status) override { return rootDebugSession->tryAccessIsa(deviceBitfield, desc, size, buffer, write, status); } bool ackIsaEvents(uint32_t deviceIndex, uint64_t isaVa) override { return rootDebugSession->ackIsaEvents(this->tileIndex, isaVa); } bool ackModuleEvents(uint32_t deviceIndex, uint64_t moduleUuidHandle) override { return rootDebugSession->ackModuleEvents(this->tileIndex, moduleUuidHandle); } ze_result_t readGpuMemory(uint64_t vmHandle, char *output, size_t size, uint64_t gpuVa) override { return rootDebugSession->readGpuMemory(vmHandle, output, size, gpuVa); } ze_result_t writeGpuMemory(uint64_t vmHandle, const char *input, size_t size, uint64_t gpuVa) override { return rootDebugSession->writeGpuMemory(vmHandle, input, size, gpuVa); } ze_result_t resumeImp(const std::vector &threads, uint32_t deviceIndex) override { return rootDebugSession->resumeImp(threads, this->tileIndex); } ze_result_t interruptImp(uint32_t deviceIndex) override { return rootDebugSession->interruptImp(this->tileIndex); } DebugSessionLinux *rootDebugSession = nullptr; uint32_t tileIndex = std::numeric_limits::max(); bool processEntryState = false; std::unordered_map modules; }; } // namespace L0