Do everything in Java...
via Digitalmars-d
digitalmars-d at puremagic.com
Sat Dec 6 04:04:46 PST 2014
On Saturday, 6 December 2014 at 09:24:57 UTC, John Colvin wrote:
> Big simulations still benefit from dedicated clusters. Good
> performance often requires uniformly extremely low latencies
> between nodes, as well as the very fastest in distributed
> storage (read *and* write).
The question is not performance between nodes if you can
partition the dataset (which I made a requirement), but how much
you pay in total for getting the job done. So you can have
inefficiency and still profit by renting CPU time because the
total cost of ownership from having a local under-utilized server
farm can be quite high.
But if the simulation requires a NUMA-like architecture… then you
don't have a dataset that you can partition and solve in a
map-reduce style.
> P.S. GPUs are not a panacea for all hpc problems. For example,
> rdma is only a recent thing for GPUs across different nodes. In
> general there is a communication bandwidth and latency issue:
> the more power you pack in each compute unit (GPU or CPU or
> whatever), the more bandwidth you need connecting them.
HPC is a special case and different architectures will suit
different problems, so you have to tailor the hardware
architecture to the problems you want to solve, but then we are
not talking $10.000 servers… If you need RDMA, then you are
basically in NUMA land, which is not really suitable for a
generic cloud solution in the first place?
More information about the Digitalmars-d
mailing list