<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629) in Graphics</title>
    <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745556#M151092</link>
    <description>&lt;P&gt;I am doing a similar use case for the B70 but have found a different issue that limits on dual and single card with the ASrocks Thaichi creator which supports PCIE 5 /16&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;BLOCKQUOTE&gt;&lt;P class=""&gt;Arc Pro B70 (BMG-G31) advertises LnkCap Gen 1 x1 on dual-card configuration — card-level ceiling, not platform&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;HR /&gt;&lt;H2&gt;Summary&lt;/H2&gt;&lt;P class=""&gt;I have two Intel Arc Pro B70 GPUs in a dual-card AI inference workstation. Both cards train at PCIe Gen 1 x1 (2.5 GT/s × 1) regardless of platform configuration. After extensive platform-side diagnostics — including a full motherboard BIOS update from AGESA 1.2.0.3e to AGESA 1.3.0.0a — the LnkCap register on both cards continues to advertise a maximum of Gen 1 x1, meaning the ceiling is at the card / silicon / firmware level rather than the motherboard or riser. Platform-side variables have been exhaustively ruled out. This is blocking production deployment because tensor parallelism (required for Intel LLM-Scaler vLLM multi-card serving) would be crippled at this link speed.&lt;/P&gt;&lt;P class=""&gt;Requesting engineering engagement to determine whether this is a known launch-firmware issue on BMG-G31, whether a newer firmware exists beyond the currently-installed BMG__31.1058, or whether this represents a hardware defect requiring RMA.&lt;/P&gt;&lt;HR /&gt;&lt;H2&gt;Hardware&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;GPUs under test:&lt;/STRONG&gt; 2x Intel Arc Pro B70 (32 GB GDDR6, BMG-G31 silicon)&lt;UL class=""&gt;&lt;LI&gt;Card 1: PCI 0000:03:00.0, MEI device /dev/mei0, firmware BMG__31.1058&lt;/LI&gt;&lt;LI&gt;Card 2: PCI 0000:08:00.0, MEI device /dev/mei1, firmware BMG__31.1058&lt;/LI&gt;&lt;LI&gt;Device ID: 8086:e223&lt;/LI&gt;&lt;LI&gt;Subsystem ID: 8086:1701&lt;/LI&gt;&lt;LI&gt;Both cards stock Intel reference design, purchased Q1 2026&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Motherboard:&lt;/STRONG&gt; ASRock X870 Taichi Creator&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;BIOS pre-test:&lt;/STRONG&gt; 3.33 (AGESA 1.2.0.3e)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;BIOS post-update:&lt;/STRONG&gt; 4.10 (AGESA 1.3.0.0a, released 2026-02-10)&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;CPU:&lt;/STRONG&gt; AMD Ryzen 9 9900X (Zen 5, 12C/24T)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Memory:&lt;/STRONG&gt; 30 GB DDR5&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Power:&lt;/STRONG&gt; Adequate PSU for 2× 250W B70s + CPU (not a power-limit issue)&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;Operating System / Software Stack&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;Ubuntu 24.04.4 LTS&lt;/LI&gt;&lt;LI&gt;Kernel: 6.17.0-20-generic&lt;/LI&gt;&lt;LI&gt;Kernel driver: xe (both cards claimed correctly, lspci -k confirms)&lt;/LI&gt;&lt;LI&gt;Intel compute-runtime: 26.09.37435.1 (latest from intel/compute-runtime GitHub)&lt;/LI&gt;&lt;LI&gt;Intel Graphics Compiler (IGC): v2.30.1 build 20950&lt;/LI&gt;&lt;LI&gt;Intel oneAPI DPC++ Compiler: 2025.3.3 (2025.3.3.20260319)&lt;/LI&gt;&lt;LI&gt;GuC/HuC firmware: Latest HEAD from linux-firmware.git&lt;UL class=""&gt;&lt;LI&gt;/lib/firmware/xe/bmg_guc_70.bin.zst&lt;/LI&gt;&lt;LI&gt;/lib/firmware/xe/bmg_huc.bin.zst&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;SYCL runtime confirms both cards as SYCL devices:&lt;UL class=""&gt;&lt;LI&gt;[level_zero:gpu][0] Intel(R) Graphics [0xe223]&lt;/LI&gt;&lt;LI&gt;[level_zero:gpu][1] Intel(R) Graphics [0xe223]&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;igsc version 0.9.3 installed and successfully enumerates both cards&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;The Specific Symptom&lt;/H2&gt;&lt;P class=""&gt;lspci -vv output for both cards (identical behavior):&lt;/P&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;PRE&gt;&lt;SPAN&gt;03:00.0 VGA compatible controller [0300]: Intel Corporation Device [8086:e223]&lt;/SPAN&gt;&lt;SPAN&gt;    ...&lt;/SPAN&gt;&lt;SPAN&gt;    LnkCap: Port #0, Speed 2.5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s &amp;lt;64ns, L1 &amp;lt;1us&lt;/SPAN&gt;&lt;SPAN&gt;    LnkSta: Speed 2.5GT/s, Width x1&lt;/SPAN&gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;PRE&gt;&lt;SPAN&gt;08:00.0 VGA compatible controller [0300]: Intel Corporation Device [8086:e223]&lt;/SPAN&gt;&lt;SPAN&gt;    ...&lt;/SPAN&gt;&lt;SPAN&gt;    LnkCap: Port #0, Speed 2.5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s &amp;lt;64ns, L1 &amp;lt;1us&lt;/SPAN&gt;&lt;SPAN&gt;    LnkSta: Speed 2.5GT/s, Width x1&lt;/SPAN&gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P class=""&gt;&lt;STRONG&gt;Critical observation:&lt;/STRONG&gt; LnkCap (the advertised maximum capability) is Speed 2.5GT/s, Width x1. This is not a link-down negotiation — the cards are telling the system they cannot go faster. At the LnkCap level, the ceiling is set by the downstream endpoint (the B70 card itself), not the upstream root complex.&lt;/P&gt;&lt;P class=""&gt;&lt;STRONG&gt;Upstream path is healthy.&lt;/STRONG&gt; dmesg reports:&lt;/P&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;PRE&gt;&lt;SPAN&gt;pci 0000:01:00.0: 126.024 Gb/s available PCIe bandwidth, limited by 16.0 GT/s PCIe x8 link&lt;/SPAN&gt;&lt;SPAN&gt;pci 0000:06:00.0: 126.024 Gb/s available PCIe bandwidth, limited by 16.0 GT/s PCIe x8 link&lt;/SPAN&gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P class=""&gt;The CPU-to-slot path provides full PCIe 4.0 x8 bandwidth (≈16 GB/s per card). Motherboard and slot training to Gen 4 x8 is fine. The B70 cards are the link-training constraint.&lt;/P&gt;&lt;H2&gt;Variables Ruled Out (Exhaustive Platform-Side Testing)&lt;/H2&gt;&lt;OL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;Riser cable signal integrity&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Tested with Conbull PCIe 5.0 riser (brand A) → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;Replaced with different-brand PCIe 5.0 riser → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Tested both cards direct in motherboard PCIE1 and PCIE2 slots, no risers → still Gen 1 x1&lt;/STRONG&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;BIOS auto-negotiation&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Forced PCIe Gen 4 in BIOS (pre-update 3.33) → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;Default/Auto PCIe negotiation (post-update 4.10) → Gen 1 x1&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Motherboard BIOS / AGESA&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Updated from 3.33 (AGESA 1.2.0.3e) to 4.10 (AGESA 1.3.0.0a) → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;This is a major AGESA revision boundary. No change in B70 link state.&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Idle power saving / ASPM&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Link state monitored via /sys/class/drm/card*/device and lspci -vv at 0.5-second intervals during active LLM inference&lt;/LI&gt;&lt;LI&gt;Link state remains Gen 1 x1 across all observed activity levels&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Driver / compute-runtime versions&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Multiple combinations tested; currently on latest upstream (26.09.37435.1 + IGC v2.30.1)&lt;/LI&gt;&lt;LI&gt;Battlemage GuC/HuC firmware loaded from current linux-firmware.git&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Slot bifurcation&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;BIOS 4.10 exposes explicit PCIe Gen 5 x16 and x8/x8 bifurcation options (new vs 3.33)&lt;/LI&gt;&lt;LI&gt;Neither affects B70 link training — both cards report the same LnkCap regardless&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;H2&gt;Variables NOT Yet Ruled Out&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;B70 on-card firmware.&lt;/STRONG&gt; Currently BMG__31.1058 on both cards. No newer firmware has been published on Intel's Linux support channels. Intel's Linux FW support article (000096950) states: &lt;EM&gt;"The Linux driver package does not update FW."&lt;/EM&gt; Community workaround of extracting firmware from the Windows driver package is available but unsupported and risks voiding warranty. Not attempted.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Silicon-level defect on both cards.&lt;/STRONG&gt; Low probability (identical behavior on two cards from likely-different production batches suggests pattern, not random defect) but cannot rule out entirely.&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;Performance Impact&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;Steady-state inference:&lt;/STRONG&gt; unaffected at this link speed because weights stay in VRAM&lt;UL class=""&gt;&lt;LI&gt;Current benchmark on Qwen3.5-9B Q8_0 via llama.cpp SYCL: 47 tok/s generation&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Model load latency:&lt;/STRONG&gt; Extended (~60 s for a ~9 GB model at Gen 1 x1, vs. &amp;lt;1 s at Gen 5 x8)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Multi-GPU tensor parallelism:&lt;/STRONG&gt; Effectively blocked. oneCCL/Level Zero cross-card transfers at 250 MB/s per card make tensor parallelism worse than single-card independent serving.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Business impact:&lt;/STRONG&gt; Cannot migrate to Intel LLM-Scaler vLLM container (26.18.8.2, released 2026-04-22) for multi-user concurrent serving, which is the documented production path for dual-B70 deployments at the advertised 140+ tok/s throughput (reference: Hal9000AIML 2×B70 setup, also Intel Project Battlematrix validation).&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;Requested Engineering Response&lt;/H2&gt;&lt;OL class=""&gt;&lt;LI&gt;Is BMG__31.1058 the current production firmware for B70?&lt;/LI&gt;&lt;LI&gt;Is there a known link-training issue on early BMG-G31 cards that would cause the endpoint to advertise LnkCap Gen 1 x1? A launch-batch firmware regression would be consistent with two cards from the same SKU exhibiting identical behavior.&lt;/LI&gt;&lt;LI&gt;If a newer firmware exists, what is the supported path for updating on Linux? Can Intel provide the firmware binary (.bin file) for direct application with igsc fw update, as currently needed by workstation Linux deployments? The "install Windows to update" path is impractical for production AI inference servers.&lt;/LI&gt;&lt;LI&gt;If firmware is current and behavior is expected at launch, what is the expected firmware release window for a link-training fix?&lt;/LI&gt;&lt;LI&gt;If neither of the above applies, please initiate RMA evaluation for both cards.&lt;/LI&gt;&lt;/OL&gt;</description>
    <pubDate>Thu, 23 Apr 2026 23:49:18 GMT</pubDate>
    <dc:creator>gadget</dc:creator>
    <dc:date>2026-04-23T23:49:18Z</dc:date>
    <item>
      <title>Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744546#M150906</link>
      <description>&lt;P&gt;Full disclosure, I've been working with Codex and Claude on this generation/upscaling pipeline project. This current issue and following message about it has been drafted with assistance from Claude.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Subject: Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629) — DPC access violation during sustained PyTorch XPU compute&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;## Project Background&lt;/P&gt;&lt;P&gt;I'm running a local AI image generation and upscaling pipeline on Windows 10, powered by an Intel Arc Pro B70. The stack includes:&lt;/P&gt;&lt;P&gt;- **ComfyUI** (embedded Python 3.13.11) for image generation via Flux.1/Flux.2 models&lt;BR /&gt;- **SUPIR** (SDXL-based image restoration/upscaling) running as a subprocess bridge&lt;BR /&gt;- **PyTorch 2.9.1+xpu / torchvision 0.24.1+xpu / torchaudio 2.9.1+xpu** (oneDNN + Level Zero backend)&lt;BR /&gt;- Mixed precision: VAE in bf16, UNet in fp16&lt;BR /&gt;- Tiled VAE encode/decode to manage VRAM, chunked SDPA (512-token blocks) to avoid OOM on large sequences&lt;/P&gt;&lt;P&gt;The B70's 32 GB VRAM is a key reason I chose this card — it handles models that won't fit on consumer GPUs.&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## The Problem&lt;/P&gt;&lt;P&gt;After several minutes of sustained high GPU utilization during SUPIR inference (tiled VAE → EDM denoising loop), the system hard crashes with **BSOD 0xD1 (DRIVER_IRQL_NOT_LESS_OR_EQUAL)**. This has occurred **6+ times within a 24-hour window**, all on the same driver build. No Python exception precedes the crash — the process simply stops producing output, then the OS reboots.&lt;/P&gt;&lt;P&gt;Driver **32.0.101.8629** (released 2026-04-02) is confirmed installed and reported as **up to date** by Intel Pro Graphics Software. This is not a "update your driver" situation.&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## System Configuration&lt;/P&gt;&lt;P&gt;| Component | Detail |&lt;BR /&gt;|---|---|&lt;BR /&gt;| **GPU** | Intel Arc Pro B70 (32 GB VRAM) |&lt;BR /&gt;| **CPU** | AMD Ryzen 9 7900X 12-Core, 4.70 GHz |&lt;BR /&gt;| **Motherboard** | ASUS ROG STRIX B650E-F |&lt;BR /&gt;| **RAM** | G.Skill Flare X5 32 GB (2×16 GB) DDR5-6000 |&lt;BR /&gt;| **Storage** | Crucial P3 2 TB PCIe Gen3 NVMe M.2 |&lt;BR /&gt;| **PSU** | Seasonic Prime 750W Platinum |&lt;BR /&gt;| **OS** | Windows 10 Pro 22H2 (build 19045) |&lt;BR /&gt;| **GPU Driver** | 32.0.101.8629 (DriverDate 2026-04-01) |&lt;BR /&gt;| **PyTorch** | 2.9.1+xpu |&lt;BR /&gt;| **Level Zero runtime** | 1.14.37111 (from oneDNN log) |&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## Dump Analysis — Root Cause&lt;/P&gt;&lt;P&gt;I captured a full kernel dump and analyzed it with `cdb.exe` (WinDbg/Microsoft Debugging Tools) with Microsoft symbols. The results are unambiguous:&lt;/P&gt;&lt;P&gt;```&lt;BR /&gt;BUGCHECK_CODE: D1 (DRIVER_IRQL_NOT_LESS_OR_EQUAL)&lt;BR /&gt;BUGCHECK_P1: ffffe301ecde483c ← invalid kernel pointer (read)&lt;BR /&gt;BUGCHECK_P2: b ← IRQL 11 (DISPATCH_LEVEL+)&lt;BR /&gt;BUGCHECK_P3: 0 ← read operation&lt;BR /&gt;BUGCHECK_P4: fffff807a0ce42c9 ← faulting RIP&lt;/P&gt;&lt;P&gt;FAULTING_MODULE: igdkmdnd64.sys&lt;BR /&gt;IMAGE_VERSION: 32.0.101.8629&lt;BR /&gt;SYMBOL_NAME: igdkmdnd64+0x3a4c29&lt;BR /&gt;FAILURE_BUCKET_ID: AV_igdkmdnd64!unknown_function&lt;BR /&gt;FAILURE_ID_HASH: {f72986a3-e8f9-3600-9d7c-dd8a40f557df}&lt;/P&gt;&lt;P&gt;Faulting instruction:&lt;BR /&gt;fffff807`a0ce42c9 4181 3c81 fd100011 cmp dword ptr [r9+rax*4], 110010FDh&lt;BR /&gt;← reads [r9+rax*4] at an invalid/freed kernel address&lt;BR /&gt;```&lt;/P&gt;&lt;P&gt;**Call stack (bottom → top):**&lt;BR /&gt;```&lt;BR /&gt;nt!KiIdleLoop&lt;BR /&gt;nt!KiRetireDpcList&lt;BR /&gt;nt!KiExecuteAllDpcs&lt;BR /&gt;dxgkrnl!DpiFdoDpcForIsr ← graphics ISR completion DPC&lt;BR /&gt;igdkmdnd64+0x39be8&lt;BR /&gt;igdkmdnd64+0x17cc3&lt;BR /&gt;igdkmdnd64+0x45bca0&lt;BR /&gt;igdkmdnd64+0x45bbca&lt;BR /&gt;igdkmdnd64+0x45a85a&lt;BR /&gt;igdkmdnd64+0x45c731&lt;BR /&gt;dxgkrnl!DpSynchronizeExecution&lt;BR /&gt;nt!KeSynchronizeExecution&lt;BR /&gt;igdkmdnd64+0x45c854&lt;BR /&gt;igdkmdnd64+0x3b2dd9&lt;BR /&gt;igdkmdnd64+0x395566&lt;BR /&gt;igdkmdnd64+0x3969ed&lt;BR /&gt;igdkmdnd64+0x3a07e6&lt;BR /&gt;igdkmdnd64+0x3a3bcd&lt;BR /&gt;igdkmdnd64+0x3a4c29 ← ACCESS VIOLATION&lt;BR /&gt;nt!KiPageFault&lt;BR /&gt;nt!KiBugCheckDispatch&lt;BR /&gt;nt!KeBugCheckEx&lt;BR /&gt;```&lt;/P&gt;&lt;P&gt;The crash is inside the **graphics ISR completion DPC** — the code path exercised by sustained Level Zero command-list submission. The KMD is dereferencing what appears to be a stale or freed object pointer (`r9`) during DPC synchronization after prolonged compute submission.&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## Reproduction Pattern&lt;/P&gt;&lt;P&gt;1. Load SDXL-class model to XPU in bf16/fp16 mixed precision (~9 GB resident)&lt;BR /&gt;2. Run tiled VAE encode over ~956×1399 image at tile_size=1024 → ~546 kernel submissions, ~10 s **(succeeds)**&lt;BR /&gt;3. Run tiled VAE decode of latent (352×240) at decoder_tile_size=128 → ~738 kernel submissions, ~10 s **(succeeds)**&lt;BR /&gt;4. Re-encode stage-1 output → ~546 submissions, ~3 s **(succeeds)**&lt;BR /&gt;5. Enter 35-step EDM denoise loop (full SDXL UNet forward per step, chunked attention, many oneDNN matmul/conv primitives); GPU sustains ~100% utilization&lt;BR /&gt;6. After several minutes (typically between step 3 and step 20 by elapsed-time correlation), system dies — no Python exception, bridge process stops, OS reboots&lt;/P&gt;&lt;P&gt;**Crash timeline — all same bugcheck 0xD1, same driver 32.0.101.8629:**&lt;BR /&gt;```&lt;BR /&gt;2026-04-15 01:45:22 (MEMORY.DMP attached from this crash)&lt;BR /&gt;2026-04-15 00:51:05&lt;BR /&gt;2026-04-15 00:04:50&lt;BR /&gt;2026-04-14 22:40:59&lt;BR /&gt;2026-04-14 19:40:06&lt;BR /&gt;2026-04-14 17:35:32&lt;BR /&gt;```&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## What Has Been Ruled Out&lt;/P&gt;&lt;P&gt;| Hypothesis | Test | Result |&lt;BR /&gt;|---|---|---|&lt;BR /&gt;| TDR timeout | Set `TdrDelay=60`, `TdrDdiDelay=60`, rebooted | Still crashed — not a timeout |&lt;BR /&gt;| Python/user-space bug | Moved UNet denoise entirely to CPU (fp32) | Still BSODed — crash is in KMD regardless |&lt;BR /&gt;| Out of memory | Monitor VRAM during run | &amp;lt;12 GB used; no OOM exceptions |&lt;BR /&gt;| Stale driver | Checked Intel Pro Graphics Software | 32.0.101.8629 = latest available |&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## Associated User-Space Errors (Possibly Related)&lt;/P&gt;&lt;P&gt;Prior runs logged oneDNN errors before the crash path matured:&lt;BR /&gt;```&lt;BR /&gt;oneDNN error: CL_INVALID_BINARY at src/gpu/intel/ocl/engine.cpp:269&lt;BR /&gt;RuntimeError: could not create a primitive&lt;BR /&gt;```&lt;/P&gt;&lt;P&gt;This suggests the SPIR-V JIT / Level Zero kernel binary pipeline can produce a binary the runtime rejects. We've applied env-var cache suppression as a local mitigation (`NEO_CACHE_PERSISTENT=0`, `SYCL_CACHE_PERSISTENT=0`, `ZE_ENABLE_LOADER_CACHE=0`, `MKLDNN_PRIMITIVE_CACHE_CAPACITY=0`, `ONEDNN_PRIMITIVE_CACHE_CAPACITY=0`) but BSODs continue.&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## Attachments&lt;/P&gt;&lt;P&gt;- **`MEMORY.DMP`** — full kernel dump, 2026-04-15 01:45:10 (~2.4 GB, since this exceeds upload limit, I cannot attach it but it is available)&lt;BR /&gt;- **`analyze_20260415.txt`** — complete `!analyze -v` + `lmkv` output from `cdb.exe`&lt;BR /&gt;- Bridge process log with `[supir-phase]` breadcrumbs available on request — gives the exact Level Zero submission call site at the moment of crash on the next repro&lt;/P&gt;&lt;P&gt;---&lt;/P&gt;&lt;P&gt;## Ask&lt;/P&gt;&lt;P&gt;1. Please match `igdkmdnd64+0x3a4c29` against Intel private symbols to identify the freed/stale object being dereferenced&lt;BR /&gt;2. If a fix exists in a beta/Xe2 insider branch or a newer Pro B-series driver build, please share a build number to test against&lt;BR /&gt;3. If a known workaround exists (Level Zero command-list batching knob, KMD feature flag, submission frequency limit), please advise&lt;/P&gt;&lt;P&gt;Happy to run any additional diagnostic captures or test against a driver drop. This is a reproducible, kernel-confirmed defect on the latest shipping driver.&lt;/P&gt;</description>
      <pubDate>Thu, 16 Apr 2026 03:11:03 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744546#M150906</guid>
      <dc:creator>RyanR3</dc:creator>
      <dc:date>2026-04-16T03:11:03Z</dc:date>
    </item>
    <item>
      <title>Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744670#M150932</link>
      <description>&lt;P&gt;Hi &amp;nbsp;RyanR3,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thank you for reaching out.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;I appreciate your transparency about working with Codex and Claude on your generation/upscaling pipeline project.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;To help us better understand and troubleshoot the issue you're experiencing, would it be possible for you to share a sample project or some code snippets? Having something we can reproduce on our end would significantly help us identify what's causing the problem and work toward a solution.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Additionally, are you working as a developer or is this more of a personal/research project?&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;This information will help me determine the cause and provide appropriate troubleshooting steps.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;We're looking forward to your response.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Best regards&lt;/P&gt;&lt;P&gt;Jonzyl B.&lt;/P&gt;&lt;P&gt;Intel Customer Support Technician&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Thu, 16 Apr 2026 22:14:50 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744670#M150932</guid>
      <dc:creator>Jonzyl_Intel</dc:creator>
      <dc:date>2026-04-16T22:14:50Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744689#M150934</link>
      <description>&lt;P&gt;Thank you for the quick response. I hope you don't mind but I used Claude to try to get the exact information on what you're needing. This is a personal/enthusiast project and my code experience is &lt;STRONG&gt;long&lt;/STRONG&gt; out of date (think Web 1.0 &lt;span class="lia-unicode-emoji" title=":face_with_tears_of_joy:"&gt;😂&lt;/span&gt;) so any issues with code structure are purely on Codex/Claude. I've used those models to&amp;nbsp;build a local AI image generation and upscaling pipeline for home use because I didn't like the limitations of a lot of stuff out there, like Comfy's visual spaghetti mess, and wanted something that was more intuitive for my use while making use of Comfy's mature backend.&lt;BR /&gt;&lt;BR /&gt;I had been working with a 4060ti 16GB previously running Forge UI with SDXL and Flux.1. I decided to upgrade hardware with the launch of the B70, and expand into working with Flux.2, SUPIR, hopefully Wan video, and possibly other LLMs with the extra freedom the 32GB VRAM in B70 offers.&lt;BR /&gt;&lt;BR /&gt;Here's what Claude helped me put together for you. If it left anything out, please let me know and I'll try to provide it:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;1. Why Code Snippets May Not Be the Full Picture&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;I want to be transparent: the crash is in igdkmdnd64.sys kernel-mode code, confirmed by full dump analysis with cdb.exe and Microsoft symbols. The Python application is not throwing an exception — the OS crashes inside a graphics ISR completion DPC. That said, I can provide both a minimal standalone reproduction script and the actual modified source files so your team can understand the submission pattern that triggers it.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;2. Minimal Standalone Reproduction Script&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;This does not require SUPIR or any AI models. It mimics the same sustained Level Zero compute pattern (chunked attention + convolution) that reliably triggers the crash on my system. Expected runtime before crash: 3–10 minutes at ~100% GPU utilization.&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;"""
Minimal XPU stress repro for igdkmdnd64.sys BSOD (0xD1).
Tested on: Intel Arc Pro B70, driver 32.0.101.8629, PyTorch 2.9.1+xpu, Windows 10 22H2.
Expected: system crashes with DRIVER_IRQL_NOT_LESS_OR_EQUAL after several minutes.
"""
 
import torch
import time
import gc
 
assert hasattr(torch, "xpu") and torch.xpu.is_available(), "XPU not available"
device = torch.device("xpu")
print(f"Device : {torch.xpu.get_device_name(0)}")
print(f"PyTorch: {torch.__version__}")
print("Starting sustained XPU compute -- expect BSOD within ~3-10 minutes on affected driver.\n")
 
start = time.time()
 
for step in range(500):
    # --- Chunked multi-head attention (mirrors SDXL UNet CrossAttention on XPU) ---
    # seq_len=84480 is the actual token count for a 352x240 latent; use 8192 here
    # for a faster repro that still exercises the same kernel submission path.
    B, H, N, D = 1, 8, 8192, 160          # batch, heads, seq_len, head_dim
    q = torch.randn(B, H, N, D, dtype=torch.float16, device=device)
    k = torch.randn(B, H, N, D, dtype=torch.float16, device=device)
    v = torch.randn(B, H, N, D, dtype=torch.float16, device=device)
 
    CHUNK = 512
    scale = D ** -0.5
    chunks = []
    for i in range(0, N, CHUNK):
        q_c  = q[:, :, i:i + CHUNK]
        attn = torch.einsum("bhid,bhjd-&amp;gt;bhij", q_c, k) * scale
        attn = attn.softmax(dim=-1)
        chunks.append(torch.einsum("bhij,bhjd-&amp;gt;bhid", attn, v))
        del attn, q_c
    out = torch.cat(chunks, dim=2)
    del q, k, v, chunks, out
 
    # --- Conv2d block (mirrors UNet ResBlock) ---
    x    = torch.randn(1, 512, 64, 64, dtype=torch.float16, device=device)
    conv = torch.nn.Conv2d(512, 512, 3, padding=1).half().to(device)
    y    = conv(x)
    del x, y, conv
 
    elapsed = time.time() - start
    print(f"Step {step+1:&amp;gt;4} | {elapsed:6.1f}s elapsed", flush=True)
 
    if step % 20 == 19:
        torch.xpu.empty_cache()
        gc.collect()
 
print("\nDone -- no crash this run.")&lt;/LI-CODE&gt;&lt;H4&gt;&amp;nbsp;To Run&lt;/H4&gt;&lt;LI-CODE lang="markup"&gt;pip install torch==2.9.1+xpu torchvision==0.24.1+xpu torchaudio==2.9.1+xpu \
    --index-url https://download.pytorch.org/whl/xpu
python xpu_stress_repro.py&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;3. Actual Application Code — Key Modified Files&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;These are the three files our pipeline modifies from the &lt;STRONG&gt;stock SUPIR repository&lt;/STRONG&gt; that are relevant to this crash:&lt;/P&gt;&lt;H4&gt;3.1&amp;nbsp; tilevae.py — XPU VRAM Detection + oneDNN Primitive Fallback&lt;/H4&gt;&lt;LI-CODE lang="markup"&gt;# XPU-aware VRAM detection (added; stock code was CUDA-only)
def _get_gpu_total_memory_mb():
    if torch.cuda.is_available():
        return torch.cuda.get_device_properties(devices.device).total_memory // 2**20
    if hasattr(torch, "xpu") and torch.xpu.is_available():
        return torch.xpu.get_device_properties(devices.device).total_memory // 2**20
return 0
 
# Decoder tile size fix (CRITICAL -- stock code passed pixel-space tile_size to decoder,
# which operates in latent space at 1/8 scale, causing tiling to be skipped entirely
# and triggering one massive untiled forward pass -&amp;gt; earlier BSOD path)
def get_recommend_decoder_tile_size():
    total_memory = _get_gpu_total_memory_mb()
    if   total_memory &amp;gt; 30*1000: return 256   # B70 with 32 GB lands here
    elif total_memory &amp;gt; 16*1000: return 192
    elif total_memory &amp;gt; 12*1000: return 128
    elif total_memory &amp;gt; 8*1000:  return 96
    else:                        return 64
 
# oneDNN primitive failure fallback (added in attn_forward)
try:
    h_ = _compute_attention(q, k, v)
except RuntimeError as err:
    msg = str(err)
    if ("could not create a primitive" not in msg) and ("CL_INVALID_BINARY" not in msg):
        raise
    # Intel XPU oneDNN intermittently fails primitive creation for bmm.
    # Fall back to CPU for this block.
    print("[Tiled VAE] attn_forward fallback to CPU after XPU primitive creation failure")
    h_ = _compute_attention(q.float().cpu(), k.float().cpu(), v.float().cpu()
         ).to(device=v.device, dtype=v.dtype)&lt;/LI-CODE&gt;&lt;H4&gt;3.2&amp;nbsp; attention.py — Chunked SDPA for XPU (No Flash Attention Backend)&lt;STRONG&gt;&lt;BR /&gt;&lt;/STRONG&gt;&lt;/H4&gt;&lt;LI-CODE lang="markup"&gt;# XPU has no flash/efficient SDPA backend. Naive path for large sequences
# (e.g. 84,480 tokens from a 352x240 latent) tries to allocate a 49.85 GB
# attention matrix and OOMs. We chunk along the query dimension instead.
 
_N      = q.shape[2]
_is_xpu = q.device.type == "xpu"
 
if _is_xpu and mask is None and _N &amp;gt; 2048:
    _CHUNK = 512          # ~3.2 GB per chunk on B70; well within free VRAM
    scale  = q.shape[-1] ** -0.5
    out_chunks = []
    for _i in range(0, _N, _CHUNK):
        q_c    = q[:, :, _i:_i + _CHUNK]
        attn_c = torch.einsum("b h i d, b h j d -&amp;gt; b h i j", q_c, k) * scale
        attn_c = attn_c.softmax(dim=-1)
        out_c  = torch.einsum("b h i j, b h j d -&amp;gt; b h i d", attn_c, v)
        out_chunks.append(out_c)
        del attn_c, q_c
    out = torch.cat(out_chunks, dim=2)
    del out_chunks
else:
    # CUDA path unchanged; nullcontext used on XPU (sdp_kernel is CUDA-only)
    with sdp_kernel(**BACKEND_MAP[self.backend]):
        out = F.scaled_dot_product_attention(q, k, v, attn_mask=mask)&lt;/LI-CODE&gt;&lt;H4&gt;3.3&amp;nbsp; supir_worker.py — Staged Model Load + Decoder Tile Size Fix&lt;STRONG&gt;&lt;BR /&gt;&lt;/STRONG&gt;&lt;/H4&gt;&lt;LI-CODE lang="markup"&gt;# Decoder tile size fix applied at load time:
# pixel-space tile_size (e.g. 1024) -&amp;gt; latent-space (//8), clamped [64, 256]
_decoder_ts = max(64, min(normalized_tile_size // 8, 256))
model.init_tile_vae(
    encoder_tile_size=normalized_tile_size,   # 1024 -&amp;gt; encoder stays pixel-space
    decoder_tile_size=_decoder_ts,            # 1024 -&amp;gt; 128 for decoder
)
 
# VAE in bf16, UNet/denoiser in fp16 (matches runtime YAML config).
# Blanket .half() previously put the VAE in fp16 while autocast wrapped
# it in bf16 -&amp;gt; dtype mismatch -&amp;gt; unnecessary oneDNN kernel recompilation.
_bf16_names = {"first_stage_model"}
for name, child in model.named_children():
    target_dtype = torch.bfloat16 if name in _bf16_names else torch.float16
    child.to(dtype=target_dtype)
 
# Staged device transfer: move submodules individually with XPU cache flushes
# rather than a single model.to(device) for ~12 GB.
for name, child in model.named_children():
    mb = sum(p.numel() * p.element_size() for p in child.parameters()) / 1024**2
    child.to(device)
    if mb &amp;gt; 100:
        torch.xpu.empty_cache()
        time.sleep(0.5)&lt;/LI-CODE&gt;&lt;H4&gt;&lt;BR /&gt;&lt;STRONG&gt;4. Cache Suppression Environment Variables&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;Applied as a mitigation — does not prevent the crash, but reduces poisoned JIT cache risk:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;NEO_CACHE_PERSISTENT=0
SYCL_CACHE_PERSISTENT=0
ZE_ENABLE_LOADER_CACHE=0
MKLDNN_PRIMITIVE_CACHE_CAPACITY=0
ONEDNN_PRIMITIVE_CACHE_CAPACITY=0&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;5. Attachments Available&lt;/STRONG&gt;&lt;/H4&gt;&lt;TABLE width="0"&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;&lt;STRONG&gt;File&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;&lt;STRONG&gt;Description&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;MEMORY.DMP (~2.4 GB)&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Full kernel dump from confirmed crash, 2026-04-15 01:45:10&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;analyze_20260415.txt&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Complete !analyze -v output pinpointing igdkmdnd64+0x3a4c29&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;Bridge log with [supir-phase] breadcrumbs&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Available on next repro — correlates exact Level Zero submission with fault moment&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&lt;BR /&gt;Please advise on your preferred method to transfer the dump file, if needed, as it will exceed forum upload limits.&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;6. Original Bug Report (INTEL_BUG_REPORT.md)&lt;/STRONG&gt;&lt;/H4&gt;&lt;H4&gt;&lt;STRONG&gt;Title&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys 32.0.101.8629 — DPC AV during sustained PyTorch 2.9.1+xpu (oneDNN/Level Zero) compute&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Severity&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;High — reproducible system crash on current latest driver. Kernel-mode fault, not recoverable by application restart.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Summary&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;On Windows 10 (19045) with an Intel Arc Pro B70, running sustained PyTorch 2.9.1+xpu compute (SUPIR image upscaling: tiled VAE encode/decode + 35-step EDM denoise loop), the system reliably blue-screens with bugcheck 0xD1 (DRIVER_IRQL_NOT_LESS_OR_EQUAL) after several minutes of high GPU utilization. Full kernel dump captured; !analyze -v pinpoints igdkmdnd64.sys reading an invalid kernel pointer inside a DPC off the graphics ISR.&lt;/P&gt;&lt;P&gt;This reproduces on driver 32.0.101.8629 (released 2026-04-02), reported by Intel Pro Graphics Software as the latest available for this GPU.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Hardware / OS&lt;/STRONG&gt;&lt;/H4&gt;&lt;TABLE width="0"&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;&lt;STRONG&gt;Component&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;&lt;STRONG&gt;Detail&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;GPU&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Intel(R) Arc(TM) Pro B70 Graphics&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;PCI ID&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;VEN_8086 DEV_E223 SUBSYS_17018086 REV_00&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;GPU Driver&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;32.0.101.8629 (DriverDate 2026-04-01)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;VRAM&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;32 GB&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;CPU&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;AMD Ryzen 9 7900X 12-Core, 4.70 GHz&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;Motherboard&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;ASUS ROG STRIX B650E-F&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;RAM&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;G.Skill Flare X5 32 GB (2x16 GB) DDR5-6000&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;Storage&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Crucial P3 2 TB PCIe Gen3 NVMe M.2&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;PSU&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Seasonic Prime 750W Platinum&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;OS&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Windows 10 Pro 22H2 64-bit, 10.0.19045&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;BIOS&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;AMI SMBIOS 3842 (2026-03-10)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;H4&gt;&lt;STRONG&gt;&lt;BR /&gt;Software Stack&lt;/STRONG&gt;&lt;/H4&gt;&lt;TABLE width="0"&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;&lt;STRONG&gt;Item&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;&lt;STRONG&gt;Value&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;Python&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;3.13.11 (Comfy embedded)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;PyTorch&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;2.9.1+xpu&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;torchvision&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;0.24.1+xpu&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;torchaudio&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;2.9.1+xpu&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;Level Zero runtime&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;1.14.37111 (from oneDNN log)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;Framework&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;SUPIR (SDXL-based image restoration)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;Attention path&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;F.scaled_dot_product_attention, chunked by 512-token query blocks on XPU&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;VAE dtype&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;bf16 (ae_dtype)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="233"&gt;&lt;P&gt;UNet dtype&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;fp16 (diffusion_dtype)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;H4&gt;&lt;STRONG&gt;&lt;BR /&gt;Bugcheck Details (from !analyze -v)&lt;/STRONG&gt;&lt;/H4&gt;&lt;LI-CODE lang="markup"&gt;BUGCHECK_CODE:        D1  (DRIVER_IRQL_NOT_LESS_OR_EQUAL)
BUGCHECK_P1:          ffffe301ecde483c   (read address -- invalid kernel pointer)
BUGCHECK_P2:          b                  (IRQL 11 -- DISPATCH_LEVEL+)
BUGCHECK_P3:          0                  (read)
BUGCHECK_P4:          fffff807a0ce42c9   (faulting RIP)
 
PROCESS_NAME:         System (DPC context)
FAULTING_THREAD:      ffffe301e9ecb640
READ_ADDRESS:         ffffe301ecde483c
 
IP_IN_PAGED_CODE:     igdkmdnd64+0x3a4c29
  fffff807`a0ce42c9   4181 3c81 fd100011   cmp dword ptr [r9+rax*4], 110010FDh
 
SYMBOL_NAME:          igdkmdnd64+0x3a4c29
MODULE_NAME:          igdkmdnd64
IMAGE_NAME:           igdkmdnd64.sys
IMAGE_VERSION:        32.0.101.8629
FAILURE_BUCKET_ID:    AV_igdkmdnd64!unknown_function
FAILURE_ID_HASH:      {f72986a3-e8f9-3600-9d7c-dd8a40f557df}&lt;/LI-CODE&gt;&lt;H4&gt;&lt;STRONG&gt;Call Stack (bottom to top)&lt;/STRONG&gt;&lt;/H4&gt;&lt;LI-CODE lang="markup"&gt;nt!KiIdleLoop+0x9e
nt!KiRetireDpcList+0x1f4
nt!KiExecuteAllDpcs+0x30e
dxgkrnl!DpiFdoDpcForIsr+0x66            &amp;lt;- graphics ISR completion DPC
igdkmdnd64+0x39be8
igdkmdnd64+0x17cc3
igdkmdnd64+0x45bca0
igdkmdnd64+0x45bbca
igdkmdnd64+0x45a85a
igdkmdnd64+0x45c731
dxgkrnl!DpSynchronizeExecution+0xac
nt!KeSynchronizeExecution+0x48
igdkmdnd64+0x45c854
igdkmdnd64+0x3b2dd9
igdkmdnd64+0x395566
igdkmdnd64+0x3969ed
igdkmdnd64+0x3a07e6
igdkmdnd64+0x3a3bcd
igdkmdnd64+0x3a4c29                      &amp;lt;- ACCESS VIOLATION here
nt!KiPageFault+0x478
nt!KiBugCheckDispatch+0x69
nt!KeBugCheckEx&lt;/LI-CODE&gt;&lt;H4&gt;&lt;BR /&gt;Reproduction Pattern&lt;/H4&gt;&lt;OL&gt;&lt;LI&gt;Load SDXL-class model to XPU in bf16/fp16 mixed precision (~9 GB resident).&lt;/LI&gt;&lt;LI&gt;Run tiled VAE encode over a ~956x1399 image at tile_size=1024 -- ~546 kernel submissions, ~8-12s (succeeds).&lt;/LI&gt;&lt;LI&gt;Run tiled VAE decode of latent (352x240) at decoder_tile_size=128 -- ~738 kernel submissions, ~10-12s (succeeds).&lt;/LI&gt;&lt;LI&gt;Re-encode stage-1 image -- ~546 submissions, ~3s (succeeds).&lt;/LI&gt;&lt;LI&gt;Enter 35-step EDM denoise loop; each step runs a full SDXL UNet forward (chunked attention, many oneDNN matmul/conv primitives). GPU stays at ~100% utilization.&lt;/LI&gt;&lt;LI&gt;After several minutes of denoising (usually between step 3 and step 20), the system dies with the bugcheck above. No Python exception precedes it.&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&lt;STRONG&gt;Crash timeline — all same bugcheck 0xD1, same driver 32.0.101.8629:&lt;/STRONG&gt;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;2026-04-15 01:45:22   (MEMORY.DMP attached from this crash)
2026-04-15 00:51:05
2026-04-15 00:04:50
2026-04-14 22:40:59
2026-04-14 19:40:06
2026-04-14 17:35:32&lt;/LI-CODE&gt;&lt;H4&gt;&lt;STRONG&gt;&lt;BR /&gt;What Has Been Ruled Out&lt;/STRONG&gt;&lt;/H4&gt;&lt;TABLE width="0"&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="187"&gt;&lt;P&gt;&lt;STRONG&gt;Hypothesis&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;&lt;STRONG&gt;Test&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;&lt;STRONG&gt;Result&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="187"&gt;&lt;P&gt;Not a TDR timeout&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;Set TdrDelay=60 and TdrDdiDelay=60, rebooted&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;Still crashed — not a timeout&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="187"&gt;&lt;P&gt;Not a Python/user-space bug&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;Moved UNet denoise entirely to CPU (fp32)&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;Still BSODed — crash is in KMD regardless&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="187"&gt;&lt;P&gt;Out of memory&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;Monitored VRAM during run&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;Less than 12 GB used; no OOM exceptions&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="187"&gt;&lt;P&gt;Stale driver&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;Checked Intel Pro Graphics Software&lt;/P&gt;&lt;/TD&gt;&lt;TD width="219"&gt;&lt;P&gt;32.0.101.8629 = latest available&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;H4&gt;&lt;BR /&gt;&lt;STRONG&gt;Associated User-Space Errors (Possibly Related)&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;Prior runs logged oneDNN errors before the crash path matured:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;oneDNN error: CL_INVALID_BINARY at src/gpu/intel/ocl/engine.cpp:269
RuntimeError: could not create a primitive&lt;/LI-CODE&gt;&lt;P&gt;This suggests the SPIR-V JIT / Level Zero kernel binary pipeline can produce a binary the runtime rejects. Cache suppression env vars have been applied as a local mitigation but BSODs continue.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;&amp;nbsp;&lt;/STRONG&gt;&lt;/H4&gt;</description>
      <pubDate>Fri, 17 Apr 2026 00:21:24 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744689#M150934</guid>
      <dc:creator>RyanR3</dc:creator>
      <dc:date>2026-04-17T00:21:24Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744775#M150950</link>
      <description>&lt;P&gt;Some additional findings I've come up with I thought I'd pass along in case it's helpful.&amp;nbsp;&lt;SPAN&gt;In further personal testing, I have found that using SUPIR to upscale a 730x730 image 2x to 1472x1472 will &lt;STRONG&gt;not&lt;/STRONG&gt; crash the system but, attempting to upscale that same 730x730 4x will BSOD, as will attempting to upscale a 1472x1472 image only 2x.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 17 Apr 2026 14:41:31 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744775#M150950</guid>
      <dc:creator>RyanR3</dc:creator>
      <dc:date>2026-04-17T14:41:31Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744989#M150988</link>
      <description>&lt;P&gt;Further testing this weekend has identified what would be considered "high risk" upscaling runs leading to BSOD condition.&lt;BR /&gt;&lt;BR /&gt;Report compiled by Claude over a couple test runs:&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;High-Risk Workload Profile&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;&lt;EM&gt;Master Forge / SUPIR on Intel Arc Pro B70&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;Conditions under which the Intel Arc Pro B70 KMD (igdkmdnd64.sys) has been observed to BSOD with DPC AV (0xD1) in this application.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Hardware / Driver Context&lt;/STRONG&gt;&lt;/H4&gt;&lt;UL&gt;&lt;LI&gt;&lt;STRONG&gt;GPU:&lt;/STRONG&gt; Intel Arc Pro B70 (Battlemage), 32 GB&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Driver:&lt;/STRONG&gt; igdkmdnd64.sys — faulting module across all observed crashes&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Stack:&lt;/STRONG&gt; PyTorch 2.9.1+xpu, Level Zero / oneDNN, Python 3.12, Windows 11&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Fault signature:&lt;/STRONG&gt; PAGE_FAULT_IN_NONPAGED_AREA (0xD1) at igdkmdnd64+0x3a4c29 inside a DPC&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Workload Class&lt;/STRONG&gt;&lt;/H4&gt;&lt;UL&gt;&lt;LI&gt;&lt;STRONG&gt;SDXL-family UNet (SUPIR fork)&lt;/STRONG&gt; with CFG-batched chunked SDPA&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Dtype:&lt;/STRONG&gt; bf16 for both UNet and latent tensors (previously fp16, but bf16-switch did not resolve the crash)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Attention path:&lt;/STRONG&gt; naive (non-flash / non-xformers) chunked scaled-dot-product implemented in fp32 with stabilized softmax — XPU has no flash/efficient SDPA backend&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;VAE:&lt;/STRONG&gt; tiled decode/encode, tile_size ≤ 768&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Submission-Rate Risk Factors (the smoking gun)&lt;/STRONG&gt;&lt;/H4&gt;&lt;UL&gt;&lt;LI&gt;&lt;STRONG&gt;Attention sequence length N &amp;gt; ~130,000 tokens&lt;/STRONG&gt; (latent 368×368 or larger)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Per-attention-call chunk count &amp;gt; ~256&lt;/STRONG&gt; when running the chunked-SDPA fallback&lt;/LI&gt;&lt;LI&gt;Observed crash: ~530 chunks × CFG-batch × ~70 attention blocks × 35 steps ≈ ~1.3 M submissions across a single image&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Sustained Level Zero submissions over several minutes&lt;/STRONG&gt; — faults typically surface 1–5 minutes into a heavy run&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Observed latency anomaly:&lt;/STRONG&gt; denoise step 0 took &lt;STRONG&gt;113.6 s&lt;/STRONG&gt; on a 368×368 latent vs &lt;STRONG&gt;4.2 s&lt;/STRONG&gt; on a 168×168 latent in the same session (27× slowdown = GPU saturation signal preceding the BSOD)&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Output-Size Risk Thresholds (empirical)&lt;/STRONG&gt;&lt;/H4&gt;&lt;UL&gt;&lt;LI&gt;&lt;STRONG&gt;High-risk output (warn):&lt;/STRONG&gt; output max_dim ≥ 1792 OR output pixels ≥ 4 Mpix&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Crash-zone output (block):&lt;/STRONG&gt; output max_dim &amp;gt; 2048 OR output pixels &amp;gt; 5 Mpix&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Known-crashing inputs:&lt;/STRONG&gt;&lt;/LI&gt;&lt;LI&gt;655×655 → 1344×1344 at tile_size=1024, upscale=2, steps=35 (§52 of our logs — 1.8 Mpix output, below the old “high-risk” gate)&lt;/LI&gt;&lt;LI&gt;1472×1472 → 2944×2944 at tile_size=768, upscale=2, steps=35 (§55 — 8.66 Mpix output, crash on first image of batch)&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Tiling / Memory Pressure&lt;/STRONG&gt;&lt;/H4&gt;&lt;UL&gt;&lt;LI&gt;tile_size=1024 triggers the fault even on sub-2-Mpix outputs&lt;/LI&gt;&lt;LI&gt;tile_size=768 mitigates only the VAE path — UNet attention still saturates the KMD on large latents&lt;/LI&gt;&lt;LI&gt;Occurs well below VRAM limits (~9.3 GB resident, ~30% of 32 GB) — &lt;STRONG&gt;not&lt;/STRONG&gt; an OOM, it’s a submission-rate / queue-depth issue&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Temporal / Batch Patterns That Elevate Risk&lt;/STRONG&gt;&lt;/H4&gt;&lt;UL&gt;&lt;LI&gt;&lt;STRONG&gt;Second consecutive run&lt;/STRONG&gt; within ~5 s of a previous completion (resolved in-app with an 8 s inter-run cool-down, but the vulnerability is driver-side)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Back-to-back high-risk images&lt;/STRONG&gt; in a batch with no inter-image settle (resolved in-app with a 6 s per-image cool-down after any job with output ≥ 1792 px / 4 Mpix)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Model reload + immediate heavy attention call&lt;/STRONG&gt; — the second run after a reload has been observed to fault more often than the first&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;What Does NOT Trigger It (ruled out)&lt;/STRONG&gt;&lt;/H4&gt;&lt;UL&gt;&lt;LI&gt;Not an fp16 overflow (reproduced on bf16)&lt;/LI&gt;&lt;LI&gt;Not an OOM (VRAM stays ~30%)&lt;/LI&gt;&lt;LI&gt;Not specific to 4× upscale (2× is sufficient when the output is large enough)&lt;/LI&gt;&lt;LI&gt;Not specific to a single image content (reproduced across unrelated source images)&lt;/LI&gt;&lt;LI&gt;Not ComfyUI / Flux generation — SUPIR is the observed reproducer, but the common factor is &lt;STRONG&gt;sustained chunked-SDPA submissions on XPU&lt;/STRONG&gt;, so any SDXL/SDXL-derivative workload without a flash-attention backend on B70 is a theoretical candidate&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Minimal Repro Recipe&lt;/STRONG&gt;&lt;/H4&gt;&lt;OL&gt;&lt;LI&gt;SUPIR (or any SDXL CrossAttention fork) with naive chunked-SDPA fallback on XPU&lt;/LI&gt;&lt;LI&gt;Input latent ≥ 168×168 (N ≥ 28,224 tokens) is sufficient for §52’s crash at tile_size=1024&lt;/LI&gt;&lt;LI&gt;bf16 UNet + bf16 latent&lt;/LI&gt;&lt;LI&gt;35 denoise steps, CFG-batched (cond + uncond), num_samples=1&lt;/LI&gt;&lt;LI&gt;Run twice back-to-back; second run reliably crashes within ~60 s on the affected machine&lt;/LI&gt;&lt;/OL&gt;</description>
      <pubDate>Sun, 19 Apr 2026 23:58:30 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1744989#M150988</guid>
      <dc:creator>RyanR3</dc:creator>
      <dc:date>2026-04-19T23:58:30Z</dc:date>
    </item>
    <item>
      <title>Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745090#M151007</link>
      <description>&lt;P&gt;Hi&amp;nbsp;RyanR3,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thank you for providing this information.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I will do further research on this matter and post the response on this thread once it is available.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If you have questions, please let us know. Thank you.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Best regards&lt;/P&gt;&lt;P&gt;Jonzyl B.&lt;/P&gt;&lt;P&gt;Intel Customer Support Technician&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Mon, 20 Apr 2026 19:21:04 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745090#M151007</guid>
      <dc:creator>Jonzyl_Intel</dc:creator>
      <dc:date>2026-04-20T19:21:04Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745308#M151055</link>
      <description>&lt;P&gt;Wanted to update with additional information that may potentially be useful.&lt;BR /&gt;&lt;BR /&gt;My reported information, so far, has revolved around image upscaling. While waiting on word on that current issue, I decided to try some other AI related tasks. After some trial and error with OOM and memory management, I've gotten successful image generation in Flux.1 Dev FP8 (I was only using NF4 on my 4060ti) and Flux.2 Dev FP4. Moving on from that I decided to try some LoRA training starting with Flux.2 since it is the more demanding of the two.&lt;BR /&gt;&lt;BR /&gt;I'm now running into the same BSOD issue under that as I was with upscaling, and the underlying issue seems to again be BSOD after sustained XPU load.&lt;BR /&gt;&lt;BR /&gt;I understand that the B70 is brand new tech and there's always a bit of a "penalty" that early adopters pay with running into new issues, and I know you all are working hard on this issue (and others) but, the B70 is billed as a pro level workhorse and if I'm being honest, I feel like it's letting me down with what should be fairly basic, mainstream AI stuff, and I haven't even tried video generation yet. All of which should be fairly easy with 32GB VRAM to work with. For perspective, I could pretty easily run 100 batch, 45 step, 2MP Flux.1 image generation runs on my 4060ti 16GB for hours on end with zero issue, while trying to get what I'd consider performance commensurate with a 32GB card seems to just be a continual fight because it seems like the B70 can't handle sustained workload.&lt;BR /&gt;&lt;BR /&gt;Regardless, just wanted to provide some customer feedback on my B70 experience so far before getting on to the latest information which I had Codex help compile into a new report after the failed LoRA training:&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Driver Stability Update: Escalation from User-Mode AV to BSOD (0xD1)&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;Prepared April 22, 2026 | Driver branch in use: 32.0.101.x | Faulting kernel module: igdkmdnd64.sys&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We have new telemetry that strengthens the case for a driver/runtime stability defect under sustained XPU compute load.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;H4&gt;&lt;STRONG&gt;New Incident Summary (April 21-22, 2026)&lt;/STRONG&gt;&lt;/H4&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;A) Repeated user-mode access violations&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;- python.exe crashes with Exception code 0xC0000005&lt;/P&gt;&lt;P&gt;- Faulting module: torch\lib\c10.dll&lt;/P&gt;&lt;P&gt;- Repeated fault signature:&lt;/P&gt;&lt;P&gt;&amp;nbsp; - c10.dll fault offset: 0x000000000008f514&lt;/P&gt;&lt;P&gt;&amp;nbsp; - Example event times:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; - April 21, 2026 11:13:34 PM (Application Error 1000)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; - April 22, 2026 12:21:45 AM (Application Error 1000)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;B) Follow-on kernel bugcheck&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;- System later BSODs with 0x000000D1 (DRIVER_IRQL_NOT_LESS_OR_EQUAL)&lt;/P&gt;&lt;P&gt;- WER SystemErrorReporting confirms dump creation:&lt;/P&gt;&lt;P&gt;&amp;nbsp; - April 22, 2026 12:50:13 AM&lt;/P&gt;&lt;P&gt;&amp;nbsp; - Bugcheck params:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; - P1=0xffff89820710a83c&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; - P2=0x000000000000000b&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; - P3=0x0000000000000000&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; - P4=0xfffff8018d2f42c9&lt;/P&gt;&lt;P&gt;- Prior matching D1 event also present:&lt;/P&gt;&lt;P&gt;&amp;nbsp; - April 20, 2026 5:42:32 PM&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;H4&gt;&lt;STRONG&gt;Observed Failure Pattern&lt;/STRONG&gt;&lt;/H4&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Under sustained XPU workload, failure now appears as a two-stage progression:&lt;/P&gt;&lt;P&gt;1) User-mode AV in PyTorch runtime (c10.dll, 0xC0000005)&lt;/P&gt;&lt;P&gt;2) System instability escalation to kernel bugcheck 0xD1 (igdkmdnd64 path)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This is materially different from normal application exceptions and strongly indicates a low-level runtime/driver fault path.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;H4&gt;&lt;STRONG&gt;Correlated Runtime Telemetry&lt;/STRONG&gt;&lt;/H4&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;- Long-running compute phase proceeds normally for several minutes.&lt;/P&gt;&lt;P&gt;- Logs then end abruptly during active compute (no graceful Python exception path).&lt;/P&gt;&lt;P&gt;- In the latest run, workload entered active iteration and continued until abrupt termination; system subsequently recorded unexpected shutdown and bugcheck events.&lt;/P&gt;&lt;P&gt;- This behavior is consistent with runtime/device state corruption rather than deterministic script failure.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;H4&gt;&lt;STRONG&gt;Why This Points to Driver/Runtime (Not App Logic)&lt;/STRONG&gt;&lt;/H4&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;- Crash signatures are stable across runs and dates.&lt;/P&gt;&lt;P&gt;- User-mode AV occurs inside c10.dll (native runtime), not Python-level tracebacks.&lt;/P&gt;&lt;P&gt;- Kernel crash class is consistent and recurring (0xD1).&lt;/P&gt;&lt;P&gt;- Same machine has now produced multiple D1 incidents over separate sessions.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;H4&gt;&lt;STRONG&gt;Artifacts Available&lt;/STRONG&gt;&lt;/H4&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;- MEMORY.DMP&lt;/P&gt;&lt;P&gt;- 042226-272031-01.dmp&lt;/P&gt;&lt;P&gt;- WER System bugcheck report ID:&lt;/P&gt;&lt;P&gt;&amp;nbsp; - 0d951bab-690c-4693-ac8f-4de6fb0a6d5b&lt;/P&gt;&lt;P&gt;- App crash report ID (python.exe / c10.dll):&lt;/P&gt;&lt;P&gt;&amp;nbsp; - 412a16dc-2f53-4fb9-9164-90eecd4203c2&lt;/P&gt;&lt;P&gt;- Additional earlier bugcheck report ID:&lt;/P&gt;&lt;P&gt;&amp;nbsp; - 49b93ebc-1f05-4362-b468-ba73da016f26&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;H4&gt;&lt;STRONG&gt;Request to Intel&lt;/STRONG&gt;&lt;/H4&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please treat this as an escalation of the existing issue:&lt;/P&gt;&lt;P&gt;- Investigate the repeated user-mode AV in c10.dll as a potential precursor to KMD failure.&lt;/P&gt;&lt;P&gt;- Correlate with repeated 0xD1 incidents in igdkmdnd64.sys under sustained XPU compute.&lt;/P&gt;</description>
      <pubDate>Wed, 22 Apr 2026 06:14:20 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745308#M151055</guid>
      <dc:creator>RyanR3</dc:creator>
      <dc:date>2026-04-22T06:14:20Z</dc:date>
    </item>
    <item>
      <title>Re: Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745556#M151092</link>
      <description>&lt;P&gt;I am doing a similar use case for the B70 but have found a different issue that limits on dual and single card with the ASrocks Thaichi creator which supports PCIE 5 /16&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;BLOCKQUOTE&gt;&lt;P class=""&gt;Arc Pro B70 (BMG-G31) advertises LnkCap Gen 1 x1 on dual-card configuration — card-level ceiling, not platform&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;HR /&gt;&lt;H2&gt;Summary&lt;/H2&gt;&lt;P class=""&gt;I have two Intel Arc Pro B70 GPUs in a dual-card AI inference workstation. Both cards train at PCIe Gen 1 x1 (2.5 GT/s × 1) regardless of platform configuration. After extensive platform-side diagnostics — including a full motherboard BIOS update from AGESA 1.2.0.3e to AGESA 1.3.0.0a — the LnkCap register on both cards continues to advertise a maximum of Gen 1 x1, meaning the ceiling is at the card / silicon / firmware level rather than the motherboard or riser. Platform-side variables have been exhaustively ruled out. This is blocking production deployment because tensor parallelism (required for Intel LLM-Scaler vLLM multi-card serving) would be crippled at this link speed.&lt;/P&gt;&lt;P class=""&gt;Requesting engineering engagement to determine whether this is a known launch-firmware issue on BMG-G31, whether a newer firmware exists beyond the currently-installed BMG__31.1058, or whether this represents a hardware defect requiring RMA.&lt;/P&gt;&lt;HR /&gt;&lt;H2&gt;Hardware&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;GPUs under test:&lt;/STRONG&gt; 2x Intel Arc Pro B70 (32 GB GDDR6, BMG-G31 silicon)&lt;UL class=""&gt;&lt;LI&gt;Card 1: PCI 0000:03:00.0, MEI device /dev/mei0, firmware BMG__31.1058&lt;/LI&gt;&lt;LI&gt;Card 2: PCI 0000:08:00.0, MEI device /dev/mei1, firmware BMG__31.1058&lt;/LI&gt;&lt;LI&gt;Device ID: 8086:e223&lt;/LI&gt;&lt;LI&gt;Subsystem ID: 8086:1701&lt;/LI&gt;&lt;LI&gt;Both cards stock Intel reference design, purchased Q1 2026&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Motherboard:&lt;/STRONG&gt; ASRock X870 Taichi Creator&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;BIOS pre-test:&lt;/STRONG&gt; 3.33 (AGESA 1.2.0.3e)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;BIOS post-update:&lt;/STRONG&gt; 4.10 (AGESA 1.3.0.0a, released 2026-02-10)&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;CPU:&lt;/STRONG&gt; AMD Ryzen 9 9900X (Zen 5, 12C/24T)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Memory:&lt;/STRONG&gt; 30 GB DDR5&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Power:&lt;/STRONG&gt; Adequate PSU for 2× 250W B70s + CPU (not a power-limit issue)&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;Operating System / Software Stack&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;Ubuntu 24.04.4 LTS&lt;/LI&gt;&lt;LI&gt;Kernel: 6.17.0-20-generic&lt;/LI&gt;&lt;LI&gt;Kernel driver: xe (both cards claimed correctly, lspci -k confirms)&lt;/LI&gt;&lt;LI&gt;Intel compute-runtime: 26.09.37435.1 (latest from intel/compute-runtime GitHub)&lt;/LI&gt;&lt;LI&gt;Intel Graphics Compiler (IGC): v2.30.1 build 20950&lt;/LI&gt;&lt;LI&gt;Intel oneAPI DPC++ Compiler: 2025.3.3 (2025.3.3.20260319)&lt;/LI&gt;&lt;LI&gt;GuC/HuC firmware: Latest HEAD from linux-firmware.git&lt;UL class=""&gt;&lt;LI&gt;/lib/firmware/xe/bmg_guc_70.bin.zst&lt;/LI&gt;&lt;LI&gt;/lib/firmware/xe/bmg_huc.bin.zst&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;SYCL runtime confirms both cards as SYCL devices:&lt;UL class=""&gt;&lt;LI&gt;[level_zero:gpu][0] Intel(R) Graphics [0xe223]&lt;/LI&gt;&lt;LI&gt;[level_zero:gpu][1] Intel(R) Graphics [0xe223]&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;igsc version 0.9.3 installed and successfully enumerates both cards&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;The Specific Symptom&lt;/H2&gt;&lt;P class=""&gt;lspci -vv output for both cards (identical behavior):&lt;/P&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;PRE&gt;&lt;SPAN&gt;03:00.0 VGA compatible controller [0300]: Intel Corporation Device [8086:e223]&lt;/SPAN&gt;&lt;SPAN&gt;    ...&lt;/SPAN&gt;&lt;SPAN&gt;    LnkCap: Port #0, Speed 2.5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s &amp;lt;64ns, L1 &amp;lt;1us&lt;/SPAN&gt;&lt;SPAN&gt;    LnkSta: Speed 2.5GT/s, Width x1&lt;/SPAN&gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;PRE&gt;&lt;SPAN&gt;08:00.0 VGA compatible controller [0300]: Intel Corporation Device [8086:e223]&lt;/SPAN&gt;&lt;SPAN&gt;    ...&lt;/SPAN&gt;&lt;SPAN&gt;    LnkCap: Port #0, Speed 2.5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s &amp;lt;64ns, L1 &amp;lt;1us&lt;/SPAN&gt;&lt;SPAN&gt;    LnkSta: Speed 2.5GT/s, Width x1&lt;/SPAN&gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P class=""&gt;&lt;STRONG&gt;Critical observation:&lt;/STRONG&gt; LnkCap (the advertised maximum capability) is Speed 2.5GT/s, Width x1. This is not a link-down negotiation — the cards are telling the system they cannot go faster. At the LnkCap level, the ceiling is set by the downstream endpoint (the B70 card itself), not the upstream root complex.&lt;/P&gt;&lt;P class=""&gt;&lt;STRONG&gt;Upstream path is healthy.&lt;/STRONG&gt; dmesg reports:&lt;/P&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;PRE&gt;&lt;SPAN&gt;pci 0000:01:00.0: 126.024 Gb/s available PCIe bandwidth, limited by 16.0 GT/s PCIe x8 link&lt;/SPAN&gt;&lt;SPAN&gt;pci 0000:06:00.0: 126.024 Gb/s available PCIe bandwidth, limited by 16.0 GT/s PCIe x8 link&lt;/SPAN&gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P class=""&gt;The CPU-to-slot path provides full PCIe 4.0 x8 bandwidth (≈16 GB/s per card). Motherboard and slot training to Gen 4 x8 is fine. The B70 cards are the link-training constraint.&lt;/P&gt;&lt;H2&gt;Variables Ruled Out (Exhaustive Platform-Side Testing)&lt;/H2&gt;&lt;OL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;Riser cable signal integrity&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Tested with Conbull PCIe 5.0 riser (brand A) → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;Replaced with different-brand PCIe 5.0 riser → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Tested both cards direct in motherboard PCIE1 and PCIE2 slots, no risers → still Gen 1 x1&lt;/STRONG&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;BIOS auto-negotiation&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Forced PCIe Gen 4 in BIOS (pre-update 3.33) → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;Default/Auto PCIe negotiation (post-update 4.10) → Gen 1 x1&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Motherboard BIOS / AGESA&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Updated from 3.33 (AGESA 1.2.0.3e) to 4.10 (AGESA 1.3.0.0a) → Gen 1 x1&lt;/LI&gt;&lt;LI&gt;This is a major AGESA revision boundary. No change in B70 link state.&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Idle power saving / ASPM&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Link state monitored via /sys/class/drm/card*/device and lspci -vv at 0.5-second intervals during active LLM inference&lt;/LI&gt;&lt;LI&gt;Link state remains Gen 1 x1 across all observed activity levels&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Driver / compute-runtime versions&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;Multiple combinations tested; currently on latest upstream (26.09.37435.1 + IGC v2.30.1)&lt;/LI&gt;&lt;LI&gt;Battlemage GuC/HuC firmware loaded from current linux-firmware.git&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Slot bifurcation&lt;/STRONG&gt;&lt;UL class=""&gt;&lt;LI&gt;BIOS 4.10 exposes explicit PCIe Gen 5 x16 and x8/x8 bifurcation options (new vs 3.33)&lt;/LI&gt;&lt;LI&gt;Neither affects B70 link training — both cards report the same LnkCap regardless&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;H2&gt;Variables NOT Yet Ruled Out&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;B70 on-card firmware.&lt;/STRONG&gt; Currently BMG__31.1058 on both cards. No newer firmware has been published on Intel's Linux support channels. Intel's Linux FW support article (000096950) states: &lt;EM&gt;"The Linux driver package does not update FW."&lt;/EM&gt; Community workaround of extracting firmware from the Windows driver package is available but unsupported and risks voiding warranty. Not attempted.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Silicon-level defect on both cards.&lt;/STRONG&gt; Low probability (identical behavior on two cards from likely-different production batches suggests pattern, not random defect) but cannot rule out entirely.&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;Performance Impact&lt;/H2&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;Steady-state inference:&lt;/STRONG&gt; unaffected at this link speed because weights stay in VRAM&lt;UL class=""&gt;&lt;LI&gt;Current benchmark on Qwen3.5-9B Q8_0 via llama.cpp SYCL: 47 tok/s generation&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Model load latency:&lt;/STRONG&gt; Extended (~60 s for a ~9 GB model at Gen 1 x1, vs. &amp;lt;1 s at Gen 5 x8)&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Multi-GPU tensor parallelism:&lt;/STRONG&gt; Effectively blocked. oneCCL/Level Zero cross-card transfers at 250 MB/s per card make tensor parallelism worse than single-card independent serving.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Business impact:&lt;/STRONG&gt; Cannot migrate to Intel LLM-Scaler vLLM container (26.18.8.2, released 2026-04-22) for multi-user concurrent serving, which is the documented production path for dual-B70 deployments at the advertised 140+ tok/s throughput (reference: Hal9000AIML 2×B70 setup, also Intel Project Battlematrix validation).&lt;/LI&gt;&lt;/UL&gt;&lt;H2&gt;Requested Engineering Response&lt;/H2&gt;&lt;OL class=""&gt;&lt;LI&gt;Is BMG__31.1058 the current production firmware for B70?&lt;/LI&gt;&lt;LI&gt;Is there a known link-training issue on early BMG-G31 cards that would cause the endpoint to advertise LnkCap Gen 1 x1? A launch-batch firmware regression would be consistent with two cards from the same SKU exhibiting identical behavior.&lt;/LI&gt;&lt;LI&gt;If a newer firmware exists, what is the supported path for updating on Linux? Can Intel provide the firmware binary (.bin file) for direct application with igsc fw update, as currently needed by workstation Linux deployments? The "install Windows to update" path is impractical for production AI inference servers.&lt;/LI&gt;&lt;LI&gt;If firmware is current and behavior is expected at launch, what is the expected firmware release window for a link-training fix?&lt;/LI&gt;&lt;LI&gt;If neither of the above applies, please initiate RMA evaluation for both cards.&lt;/LI&gt;&lt;/OL&gt;</description>
      <pubDate>Thu, 23 Apr 2026 23:49:18 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745556#M151092</guid>
      <dc:creator>gadget</dc:creator>
      <dc:date>2026-04-23T23:49:18Z</dc:date>
    </item>
    <item>
      <title>Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745981#M151174</link>
      <description>&lt;P&gt;Hi&amp;nbsp;RyanR3,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thanks for patiently waiting! I wanted to give you a quick update on your case.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;We're actively looking into the issue you reported, and our team is working hard to get to the bottom of it. To help us investigate further, we'll need a dump file from when the issue occurs. &lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Steps to Create a DMP File:&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;1) Enable Memory Dump Collection:&lt;/P&gt;&lt;UL&gt;&lt;UL&gt;&lt;LI&gt;Right-click "This PC" and select "Properties"&lt;/LI&gt;&lt;LI&gt;Click "Advanced system settings"&lt;/LI&gt;&lt;LI&gt;Under "Startup and Recovery," click "Settings"&lt;/LI&gt;&lt;LI&gt;In the "Write debugging information" dropdown, select "Complete memory dump"&lt;/LI&gt;&lt;LI&gt;Click "OK" to save&lt;/LI&gt;&lt;/UL&gt;&lt;/UL&gt;&lt;P&gt;2) Saving Dump Files:&lt;/P&gt;&lt;UL&gt;&lt;UL&gt;&lt;LI&gt;The system will automatically create a dump file (usually saved in C:\Windows\MEMORY.DMP)&lt;/LI&gt;&lt;LI&gt;After the system restarts, locate this file and send it to us&lt;/LI&gt;&lt;/UL&gt;&lt;/UL&gt;&lt;P&gt;3) Alternative Method (if the above doesn't work):&lt;/P&gt;&lt;UL&gt;&lt;UL&gt;&lt;LI&gt;Download and install WinDbg from Microsoft&lt;/LI&gt;&lt;LI&gt;Use Task Manager to create a dump when the issue happens&lt;/LI&gt;&lt;LI&gt;Go to Task Manager &amp;gt; Details tab &amp;gt; Right-click the problematic process &amp;gt; "Create dump file"&lt;/LI&gt;&lt;/UL&gt;&lt;/UL&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Let me know once you've got that dump file ready, and we'll dive right into analyzing it!&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Once you have the dump file ready, please share it using your preferred file sharing service (Google Drive, OneDrive, Dropbox, etc.) and send us the download link. These files can be quite large, so using a cloud service will make it much easier for both of us!&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thanks for your patience, and feel free to reach out if you need any help with these steps.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Best regards&lt;/P&gt;&lt;P&gt;Jonzyl B.&lt;/P&gt;&lt;P&gt;Intel Customer Support Technician&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Mon, 27 Apr 2026 23:16:39 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745981#M151174</guid>
      <dc:creator>Jonzyl_Intel</dc:creator>
      <dc:date>2026-04-27T23:16:39Z</dc:date>
    </item>
    <item>
      <title>Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745983#M151175</link>
      <description>&lt;P&gt;Hi gadget,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;To address your concern about that - I'd recommend opening up a separate case for it. This way, we can make sure it gets the proper focus it deserves without getting mixed up with what we're currently working on. It'll help us stay organized and ensure both issues get the attention they need to be resolved properly.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Best regards&lt;/P&gt;&lt;P&gt;Jonzyl B.&lt;/P&gt;&lt;P&gt;Intel Customer Support Technician&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Mon, 27 Apr 2026 23:17:12 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745983#M151175</guid>
      <dc:creator>Jonzyl_Intel</dc:creator>
      <dc:date>2026-04-27T23:17:12Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745989#M151177</link>
      <description>&lt;P&gt;Hi Jonzyl,&lt;BR /&gt;I have the dump file from April 15, when I opened this ticket, uploaded to my Google Drive account.&lt;BR /&gt;&lt;BR /&gt;Where can I send the link without posting it publicly? I did not see an option to DM you directly in your profile.&lt;/P&gt;</description>
      <pubDate>Tue, 28 Apr 2026 00:33:06 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1745989#M151177</guid>
      <dc:creator>RyanR3</dc:creator>
      <dc:date>2026-04-28T00:33:06Z</dc:date>
    </item>
    <item>
      <title>Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746094#M151203</link>
      <description>&lt;P&gt;Hi RyanR3,&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Thank you for your response. I've sent an email to your active email address you.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;Please check both your Inbox and Spam folder for my email, and once received, kindly send the file to us.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Best regards&lt;/P&gt;&lt;P&gt;Jonzyl B.&lt;/P&gt;&lt;P&gt;Intel Customer Support Technician&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;BR /&gt;</description>
      <pubDate>Tue, 28 Apr 2026 17:33:04 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746094#M151203</guid>
      <dc:creator>Jonzyl_Intel</dc:creator>
      <dc:date>2026-04-28T17:33:04Z</dc:date>
    </item>
    <item>
      <title>Re: Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746131#M151216</link>
      <description>&lt;H5&gt;I wanted to add additional information that since my April 22 post, I've updated the Intel driver to the April 16 released 32.0.101.8724 version. After that update, I attempted a LORA training run that did not result in a BSOD but did result in an error. Per Claude and Codex it may be related to the initial issue or the new driver released.&lt;/H5&gt;&lt;P&gt;Here's the report on it I had Claude compile. I will upload the mentioned files in a .zip to my Google Drive and send the link privately as before:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H3&gt;&lt;STRONG&gt;Intel Arc Pro B70 — Userspace stoull failure pinpointed to sycl::device::ext_oneapi_supports_cl_extension in sycl8.dll 2025.3.0.0&lt;/STRONG&gt;&lt;/H3&gt;&lt;P&gt;&lt;EM&gt;Date: 2026-04-28. Posted as a follow-up to the existing BSOD thread ("Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys 32.0.101.8629") because the new finding below was captured on the same hardware, after applying the 32.0.101.8724 driver update, and may be related to either the original kernel-mode defect Intel is already working on or to a separate userspace defect introduced or exposed by the 8724 driver / oneAPI runtime stack on this GPU.&lt;/EM&gt;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Headline Result&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;&lt;EM&gt;&lt;STRONG&gt;A user-mode process dump captured on 2026-04-28 at the moment of the first C++ exception throw, combined with WinDbg/cdb stack reconstruction, identifies the failing API as:&lt;/STRONG&gt;&lt;/EM&gt;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;sycl::_V1::device::ext_oneapi_supports_cl_extension
  in sycl8.dll  version 2025.3.0.0  (Intel oneAPI DPC++ Library, build 2026-01-12)&lt;/LI-CODE&gt;&lt;P&gt;The throw originates inside sycl8.dll, propagates as sycl::_V1::exception (observed 12 times in rapid succession by the dump tool), and is eventually re-raised by torch-xpu as a Python RuntimeError carrying the message "invalid stoull argument". The C++ stack establishes that the upstream trigger is torch's XPU device-properties query (c10::xpu::get_raw_device → at::xpu::getDeviceProperties), called as part of an ordinary tensor.contiguous() / convolution forward path on Arc Pro B70.&lt;/P&gt;&lt;P&gt;The std::stoull failure inside sycl8.dll is consistent with the SYCL extension-query path parsing a driver-returned extension-version string and encountering an empty or non-numeric token. We did not observe a clean PI / Level Zero error code in the throw chain (no PI_ERROR_* or ze_result_t string in the dump's exception path); the behavior matches an internal SYCL parse failure on metadata returned through the extension-support API.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Context: Existing BSOD Thread on This Hardware&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;For continuity with the in-progress investigation, the relevant facts from the original BSOD report on this hardware:&lt;/P&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Bugcheck&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;&lt;STRONG&gt;0xD1 DRIVER_IRQL_NOT_LESS_OR_EQUAL&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Faulting module&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;igdkmdnd64.sys (Intel Graphics Kernel-Mode Driver)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Driver image version in dump&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;32.0.101.8629&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Fault site (per WinDbg)&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;igdkmdnd64+0x3a4c29&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Failure bucket&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;AV_igdkmdnd64!unknown_function&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Failure hash&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;{f72986a3-e8f9-3600-9d7c-dd8a40f557df}&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Stack shape&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;KiPageFault followed by multiple igdkmdnd64 frames and dxgkrnl DPC/ISR synchronization frames (invalid kernel-address access in DPC context, not application exception)&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Trigger conditions&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;Sustained Intel XPU compute from PyTorch / oneDNN / Level Zero workloads. Increasing TDR delay did not prevent the BSOD.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;Mitigation attempted&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;Updated graphics driver to 32.0.101.8724 (released 2026-04-16). This is the driver under test for the new finding below.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The original kernel-mode MEMORY.DMP that produced the data above is already provided to Intel through the existing thread channel. The new evidence in this post is a USER-MODE process dump from a different failure mode that began appearing after the 8724 driver update (see below).&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Driver and Software State at Capture&lt;/STRONG&gt;&lt;/H4&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="175"&gt;&lt;P&gt;&lt;STRONG&gt;Graphics driver&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;&lt;STRONG&gt;32.0.101.8724 (Intel graphics driver, released 2026-04-16). Installed prior to the 2026-04-28 capture.&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="175"&gt;&lt;P&gt;&lt;STRONG&gt;Prior graphics driver&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;32.0.101.8629 (released 2026-04-02). Replaced by 8724 before this capture.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="175"&gt;&lt;P&gt;&lt;STRONG&gt;Level Zero driver version&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;1.15.37669 (reported by oneDNN verbose info banner during this capture).&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="175"&gt;&lt;P&gt;&lt;STRONG&gt;GPU&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Intel Arc Pro B70 Graphics, binary_kernels:enabled&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="175"&gt;&lt;P&gt;&lt;STRONG&gt;PyTorch XPU stack&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;torch 2.11.0+xpu / torchvision 0.26.0+xpu / torchaudio 2.11.0+xpu. Embedded Python 3.13.11 inside the ComfyUI portable distribution. ComfyUI 0.16.4.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="175"&gt;&lt;P&gt;&lt;STRONG&gt;Workload&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="391"&gt;&lt;P&gt;Flux-style transformer LoRA training, batch size 1, BF16 mixed precision, AdamW, gradient checkpointing on, block-swap CPU&amp;lt;-&amp;gt;XPU, SDPA + split_attn, LoRA rank/alpha clamped to 16. Throw reproduces in the cache-latents phase (VAE encode) before training begins.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Capture Methodology (this report)&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;A full user-mode process dump was captured of the failing python.exe training subprocess at the moment of the first C++ exception throw, with the following procedure:&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;Master Forge launched in PowerShell with MF_ONEDNN_VERBOSE=1 set on the parent shell, propagating ONEDNN_VERBOSE=1 to the training subprocess. Confirmed by master_forge.log banner showing "MF_ONEDNN_VERBOSE='1' (oneDNN verbose tracing requested)" and by subsequent oneDNN verbose output in the training subprocess stdout.&lt;/LI&gt;&lt;LI&gt;ProcDump (Sysinternals v11.1) launched in a separate elevated PowerShell, attached to the cache-latents subprocess by PID. Filter: -ma -e 1 -f sycl -f invalid_argument (full dump on first-chance C++ exception whose mangled type name contains "sycl" or "invalid_argument").&lt;/LI&gt;&lt;LI&gt;Training run started against a 30-image, 1024-bucket image-caption dataset that the Arc Pro B70 / sycl8 had not previously processed at those bucket sizes. The cache-latents phase (VAE encode, no training yet) reproduced the throw within ~5 seconds of dispatch.&lt;/LI&gt;&lt;LI&gt;ProcDump observed 12 sycl::_V1::exception throws in rapid succession, then std::invalid_argument throws, all at the same wall-clock second. Dump triggered on the first sycl::exception and written to disk at 4.16 GB.&lt;/LI&gt;&lt;LI&gt;Dump analyzed with cdb (Microsoft Console Debugger), Microsoft public symbol server enabled. Microsoft public symbols resolved KERNELBASE and VCRUNTIME140 frames cleanly; sycl8.dll, c10_xpu.dll, torch_xpu.dll frames showed nearest-exported-symbol approximations because Intel and PyTorch private PDBs are not on the public symbol server.&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;ProcDump exception-monitor record at the moment of capture (verbatim):&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;[time]Exception: E06D7363.?AVexception@_V1@sycl@@   (12 occurrences)
[time]Exception: E06D7363.?AVinvalid_argument@std@@   (4 occurrences)
[time]Exception: E06D7363.msc
[time]Process Exit: PID NNNNN, Exit Code 0x00000001&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;C++ Stack at the First sycl::exception Throw&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;Frames marked "+0xNNNN" against an exported symbol such as getBorderColor are nearest-exported-symbol approximations because sycl8.dll private PDBs are not on Microsoft's public symbol server. The module name, the publicly-named ext_oneapi_supports_cl_extension caller, and the torch-xpu / c10-xpu caller frames are unambiguous and constitute the actionable triage signal.&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;00 KERNELBASE!RaiseException+0x69                              ; OS exception raise
01 VCRUNTIME140!_CxxThrowException+0x97                        ; MSVC C++ throw
02 sycl8!sycl::_V1::detail::getBorderColor+0xa402              ; THROW INSIDE sycl8.dll
03 sycl8!sycl::_V1::detail::getBorderColor+0xa87b              ; sycl8 anonymous internals
04 sycl8!sycl::_V1::detail::getBorderColor+0xae7d
05 sycl8!sycl::_V1::detail::getBorderColor+0xfc89
06 sycl8!sycl::_V1::device::ext_oneapi_supports_cl_extension+0x77   ; THE CULPRIT API
07 c10_xpu!c10::xpu::get_raw_device+0xa85                      ; torch-xpu device facade
08 torch_xpu!at::xpu::getDeviceProperties+0x99                 ; device-properties query
09 torch_xpu!at::native::xpu::copysign_kernel+0x65df2          ; (anon, near copysign)
0a torch_xpu!at::native::xpu::copy_kernel+0xbe1ed              ; copy_kernel internals
0b torch_xpu!at::native::xpu::copy_kernel+0xb0a4e
0c torch_xpu!at::native::xpu::copy_kernel+0xa9ab5
0d torch_xpu!at::native::xpu::copy_kernel+0x114b
0e torch_xpu!at::native::xpu::copy_kernel+0xa86
0f torch_xpu!at::native::xpu::copy_kernel+0xab
10 torch_xpu!at::native::structured_max_pool2d_with_indices_out_xpu::impl+0x184e
11 torch_xpu!at::native::structured_max_pool2d_with_indices_out_xpu::impl+0xa40
12 torch_cpu!at::native::copy_ignoring_overlaps+0xaad          ; cpu-side copy bridge
13 torch_cpu!at::native::copy_+0x10d
14 torch_cpu!at::_ops::copy_::call+0x188
15 torch_cpu!at::native::clone+0x21e
16 torch_cpu!at::compositeexplicitautograd::view_copy_symint_outf+0x37ce
17 torch_cpu!at::compositeexplicitautograd::bucketize_outf+0x594c7
18 torch_cpu!at::_ops::clone::call+0x17e
19 torch_cpu!at::native::contiguous+0xf5                        ; .contiguous() inner
1a torch_cpu!at::compositeimplicitautograd::where+0x213e
1b torch_cpu!at::compositeimplicitautograd::broadcast_to_symint+0x49c87
1c torch_cpu!at::_ops::contiguous::call+0x17e                   ; Tensor.contiguous() entry
1d torch_cpu!at::TensorBase::__dispatch_contiguous+0x29
1e torch_cpu!at::TensorBase::contiguous+0xb3
1f torch_cpu!at::Tensor::contiguous+0x1c
20 torch_xpu!at::native::structured_mm_out_xpu::impl+0x6b6b     ; matmul on XPU
21 torch_xpu!at::native::structured_mm_out_xpu::impl+0xbc7f
22 torch_xpu!at::native::structured_mm_out_xpu::impl+0x951f
23 torch_cpu!at::_ops::convolution_overrideable::call+0x52c     ; convolution forward
24 torch_cpu!at::native::_convolution+0x2173
25 torch_cpu!at::compositeexplicitautograd::view_copy_symint_outf+0x1c9e
26 torch_cpu!at::compositeexplicitautograd::bucketize_outf+0x56d47
27 torch_cpu!at::_ops::_convolution::call+0x3cf
28 torch_cpu!at::native::convolution+0x1f4
...   (continuing up through autograd dispatch and into the Python C API)&lt;/LI-CODE&gt;&lt;H4&gt;&lt;STRONG&gt;Modules Implicated (versions for triage)&lt;/STRONG&gt;&lt;/H4&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;sycl8.dll&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;&lt;STRONG&gt;Path: &amp;lt;embedded_python&amp;gt;\Library\bin\sycl8.dll. Version 2025.3.0.0. Build timestamp Mon Jan 12 17:05:33 2026. Intel oneAPI DPC++ Library. This is the module that throws.&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;ze_loader.dll&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;Path: C:\Windows\System32\ze_loader.dll. Version 1.28.2.0. Build timestamp Fri Feb 20 14:42:08 2026. oneAPI Level Zero Loader for Windows. Loaded but no PI / Level Zero error code observed in the throw chain itself.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;torch_xpu.dll&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;Path: &amp;lt;embedded_python&amp;gt;\Lib\site-packages\torch\lib\torch_xpu.dll. PyTorch 2.11.0+xpu. Build timestamp Sat Mar 21 01:22:23 2026. Caller of c10::xpu::get_raw_device.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;c10_xpu.dll&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;Path: &amp;lt;embedded_python&amp;gt;\Lib\site-packages\torch\lib\c10_xpu.dll. PyTorch 2.11.0+xpu. Build timestamp Fri Mar 20 23:39:19 2026. Provides c10::xpu::get_raw_device, which calls into sycl::device::ext_oneapi_supports_cl_extension.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;torch_cpu.dll&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;PyTorch 2.11.0+xpu. Build timestamp Sat Mar 21 00:02:28 2026.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="170"&gt;&lt;P&gt;&lt;STRONG&gt;torch_python.dll&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="395"&gt;&lt;P&gt;PyTorch 2.11.0+xpu. Build timestamp Sat Mar 21 01:24:17 2026.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Python-Level Symptom vs C++ Root Cause (this capture)&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;The Python-level traceback for this capture surfaces the failure inside F.group_norm (torch.nn.functional.group_norm → torch.group_norm → C++ engine) with the message "invalid stoull argument". The dump shows that the actual C++ throw is upstream of group_norm: it occurs during the tensor.contiguous() / clone / copy chain that runs inside the convolution forward pass earlier in the VAE encoder block. That chain calls into torch-xpu's at::xpu::getDeviceProperties and c10::xpu::get_raw_device, which in turn calls sycl::device::ext_oneapi_supports_cl_extension on sycl8.dll. The extension-support query throws sycl::_V1::exception, the exception propagates up the C++ stack, eventually a downstream torch wrapper catches a related std::invalid_argument and re-raises as the Python RuntimeError. The Python op at the top of the stack at re-raise (group_norm) is incidental — it is simply the op that happened to be dispatching when the cached exception state surfaced.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Possible Relationship to the Existing BSOD Investigation&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;I cannot establish a causal link between the kernel-mode 0xD1 BSOD on 32.0.101.8629 and the userspace sycl::exception throw on 32.0.101.8724 from the evidence alone. They are different layers (kernel-mode driver vs userspace oneAPI runtime) and different failure mechanisms (IRQL violation in DPC context vs std::stoull parse failure on extension metadata). They do, however, share enough context to warrant Intel's attention as a single investigation:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Same hardware (Arc Pro B70).&lt;/LI&gt;&lt;LI&gt;Same workload class (PyTorch XPU training, sustained Level Zero submissions).&lt;/LI&gt;&lt;LI&gt;The 8724 driver update is the only major environmental change between the BSOD reproductions on 8629 and the userspace stoull reproductions on 8724.&lt;/LI&gt;&lt;LI&gt;The userspace stoull crash occurs early enough in training that sustained Level Zero compute is never reached, which is why I cannot currently confirm or deny whether 8724 fixes the original 0xD1 KMD defect. The userspace defect is gating the BSOD retest.&lt;/LI&gt;&lt;LI&gt;If the 8724 driver returns malformed extension-version metadata that trips sycl8's std::stoull parser, that same metadata change might also be a symptom of, or co-located with, the kernel-mode change that is the subject of the existing thread.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Practical implication: even if the two issues are owned by different Intel teams (KMD vs oneAPI DPC++), resolving the userspace stoull is a prerequisite for confirming whether the KMD fix on 8724 actually landed.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Hypothesis on the std::stoull Parse Failure&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;Without sycl8.dll private PDBs we cannot point at the exact std::stoull call inside getBorderColor offsets. Based on the public API name (ext_oneapi_supports_cl_extension), the most likely scenario is:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;ext_oneapi_supports_cl_extension takes a cl_intel_* or cl_khr_* extension name and returns whether the device supports it, potentially with a minimum version.&lt;/LI&gt;&lt;LI&gt;Internally, sycl8 likely queries the device's reported extension list and version metadata via Level Zero / OpenCL (clGetDeviceInfo CL_DEVICE_EXTENSIONS_WITH_VERSION or equivalent Level Zero property).&lt;/LI&gt;&lt;LI&gt;The returned metadata contains version tokens that sycl8 parses with std::stoull (or std::stoi/strtoul wrapped into invalid_argument).&lt;/LI&gt;&lt;LI&gt;On Intel Arc Pro B70 with the current Level Zero driver build, one or more extension entries returns an empty token (or a non-numeric token, e.g. an empty version field after a delimiter). std::stoull on an empty input throws std::invalid_argument. SYCL's extension subsystem catches and re-throws as sycl::exception.&lt;/LI&gt;&lt;LI&gt;12 throws in succession suggest the query iterates over a list of extensions, all of which trip the same parse, OR a single query iterates internally and counts each parse failure as one throw.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Either way, the fix surface is in two places: (a) sycl8 should treat empty / non-numeric extension version tokens as benign (fail-soft, treat as version 0 or unsupported, not throw); (b) the Level Zero driver for Arc Pro B70 should not return malformed extension version tokens. Intel can confirm which side owns the fix.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Mitigations Already Applied (for completeness)&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;None of the following resolve the bug, but they eliminate confounds and give the report a clean reproducer:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Inherited environment scrub before training subprocess launch: PYTORCH_*, TORCH_*, CUDA_*, IPEX_*, INTEL_*, SYCL_*, ZE_*, L0_*, LEVEL_ZERO_*, ONEAPI_*, NEO_*, ONEDNN_*, MKL_DEBUG, MKLDNN_* prefixes stripped on child startup.&lt;/LI&gt;&lt;LI&gt;Allocator-config keys actively pop'd from child env (PYTORCH_ALLOC_CONF, PYTORCH_XPU_ALLOC_CONF, PYTORCH_CUDA_ALLOC_CONF) — earlier failure mode on torch 2.9.1+xpu was an unrelated allocator-parser stoull that has not recurred on torch 2.11.0+xpu.&lt;/LI&gt;&lt;LI&gt;Single heavy-GPU-job coordination so generation, training, SUPIR, and captioning never contend for the XPU.&lt;/LI&gt;&lt;LI&gt;Verified system-RAM release at every workflow boundary (post-completion ComfyUI /free + gc.collect + poll-until-target).&lt;/LI&gt;&lt;LI&gt;DiT compatibility check now reads only the safetensors JSON header (no full mmap), avoiding pagefile-commit failures on Windows.&lt;/LI&gt;&lt;LI&gt;Compatibility shim around the Flux patchify rearrange to avoid non-contiguous tensor.reshape on XPU (view + permute + .contiguous() + view chain).&lt;/LI&gt;&lt;LI&gt;Pre-loop XPU cooldown sleep before first training step.&lt;/LI&gt;&lt;LI&gt;DataLoader workers forced to 0 on Windows+XPU.&lt;/LI&gt;&lt;LI&gt;LoRA rank clamped to 16 for 32 GB Arc Pro B70 headroom.&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Sanitized Repro Shape&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;The cache-latents phase reproduces the throw without any training. Command shape (paths sanitized):&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;&amp;lt;embedded_python&amp;gt;\python.exe -u src/musubi_tuner/flux_1_dev_cache_latents.py \
  --dataset_config &amp;lt;workspace&amp;gt;\configs\&amp;lt;lora_name&amp;gt;.toml \
  --vae &amp;lt;comfy_models&amp;gt;\vae\ae.safetensors \
  --model_version dev \
  --vae_dtype bfloat16 \
  --skip_existing \
  --device xpu&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Dataset shape: a small image-caption dataset (30 images, 1024-bucket, two aspect-ratio sub-buckets), batch size 1. Specific image content and captions are not relevant — the throw fires during the encoder forward pass, before any image data is consumed beyond the first tensor allocation.&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;What I Can and Cannot Claim&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;&lt;STRONG&gt;Can claim:&lt;/STRONG&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;The throw originates inside sycl8.dll 2025.3.0.0 at sycl::_V1::device::ext_oneapi_supports_cl_extension. Confirmed by WinDbg/cdb stack walk of a captured user-mode dump.&lt;/LI&gt;&lt;LI&gt;The upstream caller is c10::xpu::get_raw_device → at::xpu::getDeviceProperties from torch 2.11.0+xpu.&lt;/LI&gt;&lt;LI&gt;The exception is sycl::_V1::exception (12 in succession at the same wall-clock second), then std::invalid_argument(s), then surfaced to Python as RuntimeError("invalid stoull argument").&lt;/LI&gt;&lt;LI&gt;The std::stoull failure is internal to sycl8 — no PI / Level Zero error code appears in the throw chain.&lt;/LI&gt;&lt;LI&gt;Subsequent runs against the same input shape work because the compiled kernel binary is cached in the persistent SYCL kernel cache; fresh shapes that have never been compiled re-trigger the throw.&lt;/LI&gt;&lt;LI&gt;Driver 32.0.101.8629 produced the original BSOD (kernel-mode dump available); driver 32.0.101.8724 has been installed and has not produced a recurrence, but training cannot reach sustained-compute to confirm.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;STRONG&gt;Cannot claim:&lt;/STRONG&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;The exact line inside sycl8.dll where std::stoull is called or what string it parses. Nearest-exported-symbol approximations (the getBorderColor +0xNNNN pattern in the stack) are caused by sycl8 private PDBs not being on the Microsoft public symbol server.&lt;/LI&gt;&lt;LI&gt;Whether the malformed extension-version metadata originates in sycl8's parser being too strict or in the Level Zero driver returning bad metadata. Intel guidance on which component owns the fix is requested.&lt;/LI&gt;&lt;LI&gt;Whether 32.0.101.8724 fixed the original igdkmdnd64.sys 0xD1 defect.&lt;/LI&gt;&lt;/UL&gt;&lt;H4&gt;&lt;STRONG&gt;Materials Available via Private Channel&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;All items below are from the 2026-04-28 capture session. They are available on request through Intel's private support upload path. Not attached to the public forum thread.&lt;/P&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD width="177"&gt;&lt;P&gt;&lt;STRONG&gt;User-mode crash dump&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="389"&gt;&lt;P&gt;&lt;STRONG&gt;python.exe_260428_184651.dmp, 4.16 GB. Captured 2026-04-28 by Sysinternals ProcDump v11.1 on the cache-latents subprocess. Triggered on first sycl::_V1::exception throw (filter: sycl OR invalid_argument). Full process dump (-ma).&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="177"&gt;&lt;P&gt;&lt;STRONG&gt;WinDbg/cdb analysis output&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="389"&gt;&lt;P&gt;analyze_v2.txt, 223 KB. Output of cdb script run against the above dump with Microsoft public symbols. Contains .lastevent / .exr -1 / .ecxr / kbn 200 / ~* kbn 50 / lmDvm for sycl8 / ze_loader / *xpu / c10* / torch* modules.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="177"&gt;&lt;P&gt;&lt;STRONG&gt;oneDNN verbose trace&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="389"&gt;&lt;P&gt;Captured during the same training session that produced the dump (ONEDNN_VERBOSE=1 set on the training subprocess). Confirms 100 percent of oneDNN dispatches on this workload were GPU matmul jit:gemm:any (885+ exec events, all successful). Establishes that oneDNN is not the suspect.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="177"&gt;&lt;P&gt;&lt;STRONG&gt;master_forge.log section&lt;/STRONG&gt;&lt;/P&gt;&lt;/TD&gt;&lt;TD width="389"&gt;&lt;P&gt;Log excerpt of the 2026-04-28 session covering app startup, ComfyUI initialization, LoRA-server boot, training launch, cache-latents subprocess execution, and the failure traceback.&lt;/P&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;TR&gt;&lt;TD width="177"&gt;&amp;nbsp;&lt;/TD&gt;&lt;TD width="389"&gt;&amp;nbsp;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;H4&gt;&lt;STRONG&gt;Questions for Intel Engineering&lt;/STRONG&gt;&lt;/H4&gt;&lt;OL&gt;&lt;LI&gt;Which Intel team owns sycl8.dll (Intel oneAPI DPC++ Library) version 2025.3.0.0? Specifically, the team responsible for sycl::_V1::device::ext_oneapi_supports_cl_extension and the extension-version-string parsing path. Should this be triaged in the existing thread, or split into a sibling thread that can be cross-referenced?&lt;/LI&gt;&lt;LI&gt;Is the std::stoull failure in ext_oneapi_supports_cl_extension a known issue on Arc-class GPUs with Level Zero driver 1.15.37669 / graphics driver 32.0.101.8724? If so, is a fix available in a newer sycl8.dll release?&lt;/LI&gt;&lt;LI&gt;If the parse failure is caused by malformed extension-version metadata returned by the Level Zero driver for Arc Pro B70, would the appropriate fix be: (a) sycl8 fail-soft on empty version tokens, or (b) the L0 driver returning well-formed tokens, or (c) both?&lt;/LI&gt;&lt;LI&gt;Does graphics driver 32.0.101.8724 (released 2026-04-16) carry a fix for the igdkmdnd64.sys 0xD1 bucket cited earlier in this thread (failure bucket AV_igdkmdnd64!unknown_function, hash {f72986a3-e8f9-3600-9d7c-dd8a40f557df})? If so, the original MEMORY.DMP already provided in this thread should be a useful confirmation; if not, please advise on a target driver release. The userspace stoull crash described above is currently gating any sustained-compute retest on 8724.&lt;/LI&gt;&lt;LI&gt;Are debug builds of sycl8.dll (with private PDBs) available through Intel's NDA / support channels so we could pinpoint the exact std::stoull caller line and the offending extension name? If yes we can run the reproducer once more with the debug build loaded.&lt;/LI&gt;&lt;/OL&gt;&lt;H4&gt;&lt;STRONG&gt;Requested Next Step&lt;/STRONG&gt;&lt;/H4&gt;&lt;P&gt;Continue the existing kernel-mode triage on the 8629 BSOD bucket and, in parallel, route the userspace sycl8 finding above to the appropriate oneAPI DPC++ team. Please advise whether to keep both items in this thread or split the userspace finding into a sibling thread cross-referenced from here. We are available to run additional diagnostics on request, including ONEDNN_VERBOSE=2, Level Zero / SYCL trace, ETW / GPUView, a debug-build sycl8.dll repro, or a live cdb attach with Intel-provided private symbols.&lt;/P&gt;</description>
      <pubDate>Wed, 29 Apr 2026 00:02:46 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746131#M151216</guid>
      <dc:creator>RyanR3</dc:creator>
      <dc:date>2026-04-29T00:02:46Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746266#M151255</link>
      <description>&lt;P&gt;Hi &lt;a href="https://community.intel.com/t5/user/viewprofilepage/user-id/477489"&gt;@Jonzyl_Intel&lt;/a&gt;&amp;nbsp;and&amp;nbsp;&lt;a href="https://community.intel.com/t5/user/viewprofilepage/user-id/481129"&gt;@gadget&lt;/a&gt;&amp;nbsp;:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I can confirm the issue has happened on two out of four cards for me so far. In my case the affected cards are irreversibly dowgrading to gen4 affecting P2P as can be easily seen in synthetic measurements using ze_peak. No clear trigger, no fault of my own, user cannot intentionally cause the behavior or revert it. Boards returned. Needs a hotfix right now. Please provide the link here to the new issue so I can add data to it (I will try to fish for it now).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;DG&lt;/P&gt;</description>
      <pubDate>Wed, 29 Apr 2026 21:30:32 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746266#M151255</guid>
      <dc:creator>DCGo</dc:creator>
      <dc:date>2026-04-29T21:30:32Z</dc:date>
    </item>
    <item>
      <title>Re: Re:Arc Pro B70 BSOD 0xD1 in igdkmdnd64.sys (32.0.101.8629)</title>
      <link>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746276#M151259</link>
      <description>&lt;P&gt;Reported&amp;nbsp;&lt;A href="https://community.intel.com/t5/Graphics/Arc-Pro-B70-cards-permanently-downgraded-from-PCIe-Gen5-x16-to/m-p/1746275/highlight/true#M151258" target="_blank" rel="noopener"&gt;here&lt;/A&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 29 Apr 2026 22:41:23 GMT</pubDate>
      <guid>https://community.intel.com/t5/Graphics/Arc-Pro-B70-BSOD-0xD1-in-igdkmdnd64-sys-32-0-101-8629/m-p/1746276#M151259</guid>
      <dc:creator>DCGo</dc:creator>
      <dc:date>2026-04-29T22:41:23Z</dc:date>
    </item>
  </channel>
</rss>

