ASRock's M_2 VGA module adds legacy D-Sub connector to any modern system.
ASRock Launches Legacy M.2 Graphics Card : Read more
ASRock Launches Legacy M.2 Graphics Card : Read more
It's not really intended for a "real" desktop computer using a typical x86 cpu. It looks like it could be used to add removable graphics for embedded systems that use ARM... or much more likely it's probably meant to add a GUI to a headless server.Ummmm, interesting but who actually use it? I mean, it's gotta be worse than Intel integrated graphics even on their non-tigerlake processors
This would be good for WorkStation/Server boxes . . .
According to StarTech themselves, their USB to whatever video adapters still require a GPU to drive the display output.But in such a case, why not use a USB graphics adapter? Just quickly searching on Amazon, I found a StarTech USB 3.0 - dual HDMI adapter that supports 4K @ 60 Hz, for $111.41. And I found an Ablewe USB 3.0 - single HDMI adapter that supports 1080P @ 60 Hz for $15.99.
For a headless server, a USB graphics adapter is something a technician could carry in their toolbox and use as-needed. Or maybe the server room would have a single monitor with a USB graphics adapter, and the technician would cart the monitor+adapter over to whichever server needed servicing at that time. That seems easier and cheaper than installing an M.2 card in every single server.
And for permanent installations, a USB graphics adapter is just fine as well, if all you need is 2D graphics.
Maybe for some people, M.2 PCIe 2D VGA makes sense. But for me personally, if I had a computer that only needed occasional weak graphics, I'd just use a USB adapter.
But in such a case, why not use a USB graphics adapter? Just quickly searching on Amazon, I found a StarTech USB 3.0 - dual HDMI adapter that supports 4K @ 60 Hz, for $111.41. And I found an Ablewe USB 3.0 - single HDMI adapter that supports 1080P @ 60 Hz for $15.99.
For a headless server, a USB graphics adapter is something a technician could carry in their toolbox and use as-needed. Or maybe the server room would have a single monitor with a USB graphics adapter, and the technician would cart the monitor+adapter over to whichever server needed servicing at that time. That seems easier and cheaper than installing an M.2 card in every single server.
And for permanent installations, a USB graphics adapter is just fine as well, if all you need is 2D graphics.
Maybe for some people, M.2 PCIe 2D VGA makes sense. But for me personally, if I had a computer that only needed occasional weak graphics, I'd just use a USB adapter.
I think IOT is the target audience.This would be good for WorkStation/Server boxes where you don't often access them or need the pcie slots for other things like folding cards, mining cards, etc.
I don't think it's IOT because there's no need for such devices to have a user-serviceable PCIe stack. And if the device doesn't have a need for a display, the software isn't going to be built to drive one. Any output otherwise is likely through some UART based debug port.I think IOT is the target audience.
The Second option would be raspberry pi's class boards for stuff like your washing machine or vacuum cleaner or maybe even cars.
If they manage to make this connector some kind of standard for random items that need minimal computing power, those things can save half of the cost while not having a gpu on board and sometimes you just need to connect to get a service message.
imagine micro cpu, 1 core 1 pcie slot, 1-5W of computing.
pull wifi card out, push gpu -> debug until no more errors on screen, put wifi back.