There are a few reasons, some technical, some commercial, and also economical really
- GPU manufacturers would always rather you just bought a new GPU rather than upgrading it yourself constantly.
- The GPUs can often be limited to the bus size, so making a card that only utilises a portion of this just to leave headroom for expansions would add a lot more development cost.
- You would need sockets and buses on the GPU for the expansion which would massively increase development and production cost and also increase the size of the GPU much bigger than it already is.
- That and by the time that you would need to add more GPU memory, you are probably better off buying newer GPU technology anyway. You are likely already maximising the usage of the GPU with the VRAM it has.
- Power requirements stipulated by manufacturers becomes more convoluted as each person could be using up different amounts of power in different ways. Not only power requirements, but cooling requirement also - all of a sudden, maintaining the cards life becomes much more difficult.
- Theoretically you are increasing the amount of things that could go wrong, warranties could suffer, compatibility could suffer, and maintenance could also be worse. There are enough issues with DDR RAM compatibility (such as mixing modules) so imagine adding these problems to GPUs too.
Also you'd need to be able to do all of this, without needing to taking the card apart - the costs and inefficiencies just add up.
So ultimately, it massively saves cost for the manufacturer, helps standardise practice and usage far more, and ultimately likely helps you in the long run!
I believe there are a select few out there that have a kind of expandable memory, but they're not really aimed at general users.