Ah, I see. There are several ways this could be done. With the "substitute the cas funcall" line I thought you were going for an intermediate solution that would help the LLVM backend but not the native codegen. I was thinking you would leave the out-of-line primop definition for, e.g., casMutVar#, but fix the ccall to "cas" within that primop, so that you don't need a C function call sequence. But it sounds like you are going whole hog and going right for inline primops! Great.Actually, there are some places where I am ignorant of what optimizations the backend(s) can do (and I haven't been able to learn the answer from the commentary yet). For example, I assume calls to C are never inlinable, but are "out of line" primops inlinable?You alluded to the double call over head -- first for out-of-line casMutVar# and then to the C function "cas". Does that mean "no" they are not inlinable? (There is one sentence in the commentary that makes it sound like "no": This also changes to code generator to push the continuation of any follow on code onto the stack.)One thing that I now understand looking at Tibbe's patches, is that going to inline primops does NOT mean forgoing FFI calls necessarily. That patch still uses emitForeignCall within emitPopCntCall. Is that what you were planning to do for the atomic primops?The alternative, which seemed laborious, is to take code like this:cas(StgVolatilePtr p, StgWord o, StgWord n){#if i386_HOST_ARCH || x86_64_HOST_ARCH__asm__ __volatile__ ("lock\ncmpxchg %3,%1":"=a"(o), "=m" (*(volatile unsigned int *)p):"0" (o), "r" (n));return o;#elif powerpc_HOST_ARCH....and embed its logic within the codegen for the inline primops.-----------------------------------------------------------------------------Anyway, to answer your question about which primops I'd like to see:
- CAS on MutVars, MutableArray#, and MutableByteArray#
- fetch and add on MutableByteArray#
- barriers / memory fences
- Drafts of .cmm for these can be found here. Note that *only* casMutVar# is currently shipped with GHC.
These are the ones I'm using currently. But there's no reason that we shouldn't aim for a fairly "complete set". For example, why not have fetch-and-sub and the other "atomicrmw" variants? Relating these to the LLVM atomics and memory orderings, they become:
- CAS variants = LLVM cmpxchg with SequentiallyConsistent ordering
- fetch-and-X variants = LLVM atomicrmw with SequentiallyConsistent
- store_load_barrier = LLVM fenceInst with SequentiallyConsistent
- write_barrier and load_load_barrier = I *think* these are both covered by a FenceInst with AcquireRelease ordering...
Someone else double checking these would be good, since I'm not yet familiar with LLVM and am just going off the documentation you linked.Btw, I'm not sure why SMP.h uses "lock; addl $0,0(%%esp)" instead of the mfence instruction for store_load_barrier on x86, but I believe they should be the same.-Ryan[1] I note that the LLVM documentation says "store-store fences are generally not exposed to IR because they are extremely difficult to use correctly."On Sat, Jul 20, 2013 at 3:19 AM, Carter Schonwald <carter.schonwald@gmail.com> wrote:
Ryan, you misunderstand (or maybe i'm not understanding quite). It is 330 am after all! (I might be better at explaining tomorrow afternoon)the idea is to provide CMM/haskell level primops, not to "pattern match on the ccall". I leave the updating of any cmm code to use such intrinsics as distinct task to be done subsequently :)If you look at the example patches for pop count that David Terei referred me to, https://github.com/ghc/ghc/commit/2d0438f329ac153f9e59155f405d27fac0c43d65 (for the native code gen) and https://github.com/ghc/ghc/commit/2906db6c3a3f1000bd7347c7d8e45e65eb2806cb for the llvm code gen, the pattern is pretty clear, adding new "first class" primiopsPoint being, dont' worry about that right now, (its 3am after all)What I want from you is a clear description of the CMM / Haskell level PrimOps you want for making your life easier in supporting great parallelism in GHC, in terms of those LLVM operations and their semantics that I've referred you to.what the final names of these will be can be bike shedded some other time, doesn't matter currently. For now, please read my ticket and the llvm links when you have the bandwidth, and layout what you'd want primop wise!thanks-CarterOn Sat, Jul 20, 2013 at 2:47 AM, Ryan Newton <rrnewton@gmail.com> wrote:
Sorry, "rewrite" was too overloaded a term to use here. I was just referring to the proposal to "substitute the cas funcall with the right llvm operation".That is, the approach would pattern match for the CMM code "ccall cas" or "foreign "C" cas" (I'm afraid I don't know the difference between those) and replace it with the equivalent LLVM op, right?I think the assumption there is that the native codegen would still have to suffer the funcall overhead and use the C versions. I don't know exactly what the changes would look like to make barriers/CAS all proper inline primops, because it would have to reproduce in the code generator all the platform-specific #ifdef'd C code that is currently in SMP.h. Which I guess is doable, but probably only for someone who knows the native GHC codegen properly...On Sat, Jul 20, 2013 at 2:30 AM, Carter Schonwald <carter.schonwald@gmail.com> wrote:
Ryan, could you explain what you want more precisely? Specifically what you want in terms of exposed primops using the terminology / vocabulary in http://llvm.org/docs/LangRef.html#ordering and http://llvm.org/docs/Atomics.html ?I'll first do the work for just the LLVM backend, and I"ll likely need some active guidance / monitoring for the native codegen analogues(also asked this on ticket for documentation purposes)On Sat, Jul 20, 2013 at 2:18 AM, Ryan Newton <rrnewton@gmail.com> wrote:
Hi Carter,Yes, SMP.h is where I've copy pasted the duplicate functionality from (since I can't presently rely on linking the symbols).Your proposal for the LLVM backend sounds *great*. But it also is going to provide additional constraints for getting "atomic-primops" right.The goal of atomic-primops is to be a stable Haskell-level interface into the relevant CAS and fetch-and-add stuff. The reason this is important is that one has to be very careful to defeat the GHC optimizer in all the relevant places and make pointer equality a reliable property. I would like to get atomic-primops to work reliably in 7.4, 7.6 [and 7.8] and have more "native" support in future GHC releases, where maybe the foreign primops would become unecessary. (They are a pain and have already exposed one blocking cabal bug, fixed in upcoming 1.17.)A couple additional suggestions for the proposal in ticket #7883:
- we should use more unique symbols than "cas", especially for this rewriting trick. How about "ghc_cas" or something?
- it would be great to get at least fetch-and-add in addition to CAS and barriers
- if we reliably provide this set of special symbols, libraries like atomic-primops may use them in the .cmm and benefit from the CMM->LLVM substitutions
- if we include all the primops I need in GHC proper the previous bullet will stop applying ;-)
Cheers,-RyanP.S. Just as a bit of motivation, here are some recent performance numbers. We often wonder about how close our "pure values in a box" approach comes to efficient lock-free structures. Well here are some numbers about using a proper unboxed counter in the Haskell heap, vs using an IORef Int and atomicModifyIORef': Up to 100X performance difference on some platforms for microbenchmarks that hammer a counter:And here are the performance and scaling advantages of using ChaseLev (based on atomic-primops), over a traditional pure-in-a-box structure (IORef Data.Seq). The following are timings of ChaseLev/traditional respectively on a 32 core westmere:fib(42) 1 threads: 21sfib(42) 2 threads: 10.1sfib(42) 4 threads: 5.2s (100%prod)fib(42) 8 threads: 2.7s - 3.2s (100%prod)fib(42) 16 threads: 1.28sfib(42) 24 threads: 1.85sfib(42) 32 threads: 4.8s (high variance)(hive) fib(42) 1 threads: 41.8s (95% prod)(hive) fib(42) 2 threads: 25.2s (66% prod)(hive) fib(42) 4 threads: 14.6s (27% prod, 135GB alloc)(hive) fib(42) 8 threads: 17.1s (26% prod)(hive) fib(42) 16 threads: 16.3s (13% prod)(hive) fib(42) 24 threads: 21.2s (30% prod)(hive) fib(42) 32 threads: 29.3s (33% prod)And that is WITH the inefficiency of doing a "ccall" on every single atomic operation.Notes on parfib performance are here:On Fri, Jul 19, 2013 at 5:05 PM, Carter Schonwald <carter.schonwald@gmail.com> wrote:
ryan, the relevant machinery on the C side is here, see ./includes/stg/SMP.h : https://github.com/ghc/ghc/blob/7cc8a3cc5c2970009b83844ff9cc4e27913b8559/includes/stg/SMP.h(unless i'm missing something)On Fri, Jul 19, 2013 at 4:53 PM, Carter Schonwald <carter.schonwald@gmail.com> wrote:
Ryan,if you look at line 270, you'll see the CAS is a C call https://github.com/ghc/ghc/blob/95e6865ecf06b2bd80fa737e4fa4a24beaae25c5/rts/PrimOps.cmm#L270What Simon is alluding to is some work I started (but need to finish)http://ghc.haskell.org/trac/ghc/ticket/7883 is the relevant ticket, and I'll need to sort out doing the same on the native code gen too
there ARE no write barrier primops, they're baked into the CAS machinery in ghc's rtsOn Fri, Jul 19, 2013 at 1:02 PM, Ryan Newton <rrnewton@gmail.com> wrote:
Yes, I'd absolutely rather not suffer C call overhead for these functions (or the CAS functions). But isn't that how it's done currently for the casMutVar# primop?To avoid the overhead, is it necessary to make each primop in-line rather than out-of-line, or just to get rid of the "ccall"?Another reason it would be good to package these with GHC is that I'm having trouble building robust libraries of foreign primops that work under all "ways" (e.g. GHCI). For example, this bug:If I write .cmm code that depends on RTS functionality like stg_MUT_VAR_CLEAN_info, then it seems to work fine when in compiled mode (with/without threading, profiling), but I get link errors from GHCI where these symbols aren't defined.I've got a draft of the relevant primops here:Which includes:
- variants of CAS for MutableArray# and MutableByteArray#
- fetch-and-add for MutableByteArray#
Also, there are some tweaks to support the new "ticketed" interface for safer CAS:I started adding some of these primops to GHC proper (still as out-of-line), but not all of them. I had gone with the foreign primop route instead...-RyanP.S. Where is the write barrier primop? I don't see it listed in prelude/primops.txt...On Fri, Jul 19, 2013 at 11:41 AM, Carter Schonwald <carter.schonwald@gmail.com> wrote:
I guess I should find the time to finish the CAS primop work I volunteered to do then. Ill look into in a few days.
On Friday, July 19, 2013, Simon Marlow wrote:_______________________________________________On 18/07/13 14:17, Ryan Newton wrote:
The "atomic-primops" library depends on symbols such as
store_load_barrier and "cas", which are defined in SMP.h. Thus the
result is that if the program is linked WITHOUT "-threaded", the user
gets a linker error about undefined symbols.
The specific place it's used is in the 'foreign "C"' bits of this .cmm code:
https://github.com/rrnewton/haskell-lockfree-queue/blob/87e63b21b2a6c375e93c30b98c28c1d04f88781c/AtomicPrimops/cbits/primops.cmm
I'm trying to explore hacks that will enable me to pull in those
functions during compile time, without duplicating a whole bunch of code
from the RTS. But it's a fragile business.
It seems to me that some of these routines have general utility. In
future versions of GHC, could we consider linking in those routines
irrespective of "-threaded"?
We should make the non-THREADED versions EXTERN_INLINE too, so that there will be (empty) functions to call in rts/Inlines.c. Want to submit a patch?
A better solution would be to make them into primops. You don't really want to be calling out to a C function to implement a memory barrier. We have this for write_barrier(), but none of the others so far. Of couse that's a larger change.
Cheers,
Simon
ghc-devs mailing list
ghc-devs@haskell.org
http://www.haskell.org/mailman/listinfo/ghc-devs