I want a Arc 24GB VRAM clamshell (just like the 4060 Ti 16GB one is) version for AI inference and later, if there's going to be a Arc B700 series, a B750/B770 32GB VRAM clamshell version as well, because the 24GB VRAM are kinda not enough for fast inference to fit bigger LLMs.
Tbh, I think my (and maybe others') demands have changed since the release of the GeForce 5090 which has 32GB VRAM: To fit a decent 30B-32B LLM quant (e.g. Qwen2.5-32B-Instruct-Q6_K.gguf) in 1 GPU, because my other PCI(e) slots are used (maybe I could reshuffle stuff but then it wouldn't be optimal) or because I don't want to bother with 2 GPUs, one would need a Arc B770 / B780 32GB VRAM GPU clamshell version. So it would be nice if Intel priorized the 256bit clamshell version over the 192bit one.
Another way to make the Arc Battlemage lots-of-vram-for-ai clamshell maybe cheaper: Use GDDR7 (NV already show that despite GDDR7, 5070' price is still 549) and reduce the memory bus width from 256 to 192 bit to make the chip smaller and cheaper and use 3GB GDDR7 chips:
Before: 8 chips * 2GB per chip * 2 [clamshell] = 32GB VRAM (@ ~608 GB/s)
After: 6 chips * 3GB per chip * 2 [clamshell] = 36GB VRAM (@ ~672 GB/s)