summaryrefslogtreecommitdiff
path: root/debian/patches/lock-memory-clobber.patch
blob: 3d16074ed07c2701eda6b75c806923a0cd082598 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
Index: libpthread/sysdeps/i386/bits/memory.h
===================================================================
RCS file: /cvsroot/hurd/hurd/libpthread/sysdeps/i386/bits/memory.h,v
retrieving revision 1.3
diff -u -p -r1.3 memory.h
--- libpthread/sysdeps/i386/bits/memory.h	1 Jul 2008 11:43:17 -0000	1.3
+++ libpthread/sysdeps/i386/bits/memory.h	2 Aug 2008 20:38:31 -0000
@@ -28,7 +28,7 @@ __memory_barrier (void)
 
   /* Any lock'ed instruction will do.  We just do a simple
      increment.  */
-  __asm__ __volatile ("lock; incl %0" : "=m" (i) : "m" (i));
+  __asm__ __volatile ("lock; incl %0" : "=m" (i) : "m" (i) : "memory");
 }
 
 /* Prevent read reordering across this function.  */
Index: libthreads/i386/cthreads.h
===================================================================
RCS file: /cvsroot/hurd/hurd/libthreads/i386/cthreads.h,v
retrieving revision 1.3
diff -u -p -r1.3 cthreads.h
--- libthreads/i386/cthreads.h	3 Mar 2007 23:57:37 -0000	1.3
+++ libthreads/i386/cthreads.h	2 Aug 2008 20:38:32 -0000
@@ -98,14 +98,14 @@ typedef volatile int spin_lock_t;
 	({  register int _u__ ; \
 	    __asm__ volatile("xorl %0, %0; \n\
 			  xchgl %0, %1" \
-			: "=&r" (_u__), "=m" (*(p)) ); \
+			: "=&r" (_u__), "=m" (*(p)) :: "memory" ); \
 	    0; })
 
 #define	spin_try_lock(p)\
 	(!({  boolean_t _r__; \
 	    __asm__ volatile("movl $1, %0; \n\
 			  xchgl %0, %1" \
-			: "=&r" (_r__), "=m" (*(p)) ); \
+			: "=&r" (_r__), "=m" (*(p)) :: "memory" ); \
 	    _r__; }))
 
 #define	cthread_sp() \