From 24a435f52cb62b3a148358f78dbb5867a97b18c4 Mon Sep 17 00:00:00 2001 From: Khem Raj Date: Fri, 15 Aug 2008 05:24:04 +0000 Subject: __asm__ and __volatile__ conversion for arm. --- libpthread/nptl/sysdeps/unix/sysv/linux/arm/lowlevellock.h | 4 ++-- libpthread/nptl/sysdeps/unix/sysv/linux/arm/unwind-resume.c | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) (limited to 'libpthread/nptl/sysdeps/unix/sysv/linux') diff --git a/libpthread/nptl/sysdeps/unix/sysv/linux/arm/lowlevellock.h b/libpthread/nptl/sysdeps/unix/sysv/linux/arm/lowlevellock.h index 26783b761..df5833c46 100644 --- a/libpthread/nptl/sysdeps/unix/sysv/linux/arm/lowlevellock.h +++ b/libpthread/nptl/sysdeps/unix/sysv/linux/arm/lowlevellock.h @@ -107,7 +107,7 @@ __lll_mutex_trylock (int *futex) else if (old > 1) flag = atomic_exchange_acq (futex, old); #else - asm volatile ( + __asm__ __volatile__ ( "\tswp %[old], %[flag], [%[futex]] @ try to take the lock\n" "\tcmp %[old], #1 @ check old lock value\n" "\tmovlo %[flag], #0 @ if we got it, return 0\n" @@ -134,7 +134,7 @@ __lll_mutex_cond_trylock (int *futex) else if (old > 1) flag = atomic_exchange_acq (futex, old); #else - asm volatile ( + __asm__ __volatile__ ( "\tswp %[old], %[flag], [%[futex]] @ try to take the lock\n" "\tcmp %[old], #1 @ check old lock value\n" "\tmovlo %[flag], #0 @ if we got it, return 0\n" diff --git a/libpthread/nptl/sysdeps/unix/sysv/linux/arm/unwind-resume.c b/libpthread/nptl/sysdeps/unix/sysv/linux/arm/unwind-resume.c index 344f80535..8188249ef 100644 --- a/libpthread/nptl/sysdeps/unix/sysv/linux/arm/unwind-resume.c +++ b/libpthread/nptl/sysdeps/unix/sysv/linux/arm/unwind-resume.c @@ -53,7 +53,7 @@ init (void) ARM unwinder relies on register state at entrance. So we write this in assembly. */ -asm ( +__asm__ ( " .globl _Unwind_Resume\n" " .type _Unwind_Resume, %function\n" "_Unwind_Resume:\n" -- cgit v1.2.3