Unofficial wish list status.(Jul 2008)
Me Here
p9e883002 at sneakemail.com
Fri Jul 4 09:04:12 PDT 2008
superdan wrote:
> Oskar Linde Wrote:
>
> > superdan wrote:
> > > Me Here Wrote:
> > >
> > >> Walter Bright wrote:
> > > >
> > >>> Yes, but the onus will be on you (the programmer) to prevent data races
> > and >>> do proper synchronization.
> > >> In the scenario described, the main thread initialises the array of
> > data. Then, >> non-overlapping slices of that are tioned out to N worker
> > threads. Only one >> thread ever modifies any given segment. When the
> > worker threads are complete, >> the 'results' are left in the original
> > array available in its entirety only to >> the main thread.
> > > >
> > >>> You have to be very wary of cache effects when
> > >>> writing data in one thread and expecting to see it in another.
> > >> Are you saying that there is some combination of OS and/or hardware L1/L2
> > >> caching that would allow one thread to read a memory location
> > (previously) >> modified by another thread, and see 'old data'?
> > > >
> > >> Cos if you are, its a deeply serious bug that if its not already very
> > well >> documented by the OS writer or hardware manufacturers, then here's
> > your chance >> to get slashdotted (and diggited and redited etc. all
> > concurrently) as the >> discoveerer of a fatel processor flaw.
> > >
> > > google for "relaxed memory consistency model" or "memory barriers". geez.
> >
> > I presume the discussion regards symmetric multiprocessing (SMP).
> >
> > Cache coherency is a very important element of any SMP design. It
> > basically means that caches should be fully transparent, i.e. the
> > behavior should not change by the addition or removal of caches.
>
> you are perfectly correct... as of ten years ago. you are right in that cache
> coherency protocols ensure the memory model is respected regardless of adding
> or eliminating caches. (i should know coz i implemented a couple for a
> simulator.) the problem is that the memory model has been aggressively
> changed recently towards providing less and less implied ordering and
> requiring programs to write explicit synchronization directives.
>
> > So the above scenario should never occur. If thread A writes something
> > prior to thread B reading it, B should never get the old value.
>
> yeah the problem is it's hard to define what "prior" means.
>
> > "Memory barriers" have nothing to do with cache consistency. A memory
> > barrier only prevents a single CPU thread from reordering load/store
> > instructions across that specific barrier.
>
> memory barriers strengthen the relaxed memory model that was pushed
> aggressively by the need for faster caches.
Since in the scenario I describe, Each thread or cpu is dealing with a single
section of memory. And each section of memory is being dealt with by a single
thread or cpu, the is effectively no shared state whilst the threads run, Hence
no possibility of cache inconsistancy due to pipeline reordering. Ie.
main thread populates a[ 0 .. 1000 ];
for thread 1 .. 10
spawn( thread, \a[ ((thread-1 ) *100 ) .. ((thread-1 + 100) * 100 ] );
main thread waits for all threads to terminate;
main thread does something with a[];
In any case, cache consistancy issues due to pipeline reordering do not survive
context switches, so the issue is a non-issue for the purposes of the
discussion at hand. Ie. threading
Pipelines cover single digit or low double digit runs of non-branching
instructsion at most. A context switch consists of hundreds if not thousands of
instructions on all but the most highly tuned of real-time kernels. This is a
very localised issue, for the compiler writer, not the application programmer
to worry about.
I know Walter *is* a compiler writer, but this is a complete red-herring in the
context of this discussion.
b.
--
More information about the Digitalmars-d
mailing list