- Joined
- Jan 14, 2019
- Messages
- 14,039 (6.35/day)
- Location
- Midlands, UK
Processor | Various Intel and AMD CPUs |
---|---|
Motherboard | Micro-ATX and mini-ITX |
Cooling | Yes |
Memory | Overclocking is overrated |
Video Card(s) | Various Nvidia and AMD GPUs |
Storage | A lot |
Display(s) | Monitors and TVs |
Case | The smaller the better |
Audio Device(s) | Speakers and headphones |
Power Supply | 300 to 750 W, bronze to gold |
Mouse | Wireless |
Keyboard | Mechanic |
VR HMD | Not yet |
Software | Linux gaming master race |
The question is simple.
We're seeing an increasing number of game remakes and franchise iterations and a decreasing number of original creations - from the big AAA studios, at least.
The latest controversial releases are Portal RTX the The Witcher 3 Next-Gen Update (what is "next" gen, anyway? If it was released last month, then isn't it current gen already?).
These remakes beg the question: are computer graphics really that important? Or is it an agenda that Intel/AMD/Nvidia are trying to force on us to buy their hardware? Have big game studios reached their creative limits? The discussion is open. Fire away!
We're seeing an increasing number of game remakes and franchise iterations and a decreasing number of original creations - from the big AAA studios, at least.
The latest controversial releases are Portal RTX the The Witcher 3 Next-Gen Update (what is "next" gen, anyway? If it was released last month, then isn't it current gen already?).
These remakes beg the question: are computer graphics really that important? Or is it an agenda that Intel/AMD/Nvidia are trying to force on us to buy their hardware? Have big game studios reached their creative limits? The discussion is open. Fire away!