The GPU in your computer is about 10 to 100 times more powerful than the CPU, depending on workload. For real-time graphics rendering and machine learning, you are enjoying that power, and doing those workloads on a CPU is not viable. Why aren’t we exploiting that power for other workloads? What prevents a GPU from being a more general purpose computer?| Raph Levien’s blog
I promised in my last post to show an example of the importance of knowing how much precision a float has at a particular value. Here goes. As a general rule this type of data should never be store…| Random ASCII - tech blog of Bruce Dawson
Behind the scenes - how I'm making redblobgames.com - interactive explanations of game development algorithms and math| simblob.blogspot.com