From 360ef36e5033b52b84d0c3b20d673bb184bfac97 Mon Sep 17 00:00:00 2001 From: Doug Thompson Date: Thu, 26 Jul 2007 10:41:15 -0700 Subject: [PATCH] --- yaml --- r: 62996 b: refs/heads/master c: 39c29657fcf6060d71e04f1e52e5bb4b2999644f h: refs/heads/master v: v3 --- [refs] | 2 +- trunk/include/asm-mips/edac.h | 35 +++++++++++++++++++++++++++++++++++ 2 files changed, 36 insertions(+), 1 deletion(-) create mode 100644 trunk/include/asm-mips/edac.h diff --git a/[refs] b/[refs] index 2ac7e34cb4d6..89ccd1a0024f 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: d4c1465b7de9686c4c5aa533b15c09ab014aab3a +refs/heads/master: 39c29657fcf6060d71e04f1e52e5bb4b2999644f diff --git a/trunk/include/asm-mips/edac.h b/trunk/include/asm-mips/edac.h new file mode 100644 index 000000000000..83719eee2d13 --- /dev/null +++ b/trunk/include/asm-mips/edac.h @@ -0,0 +1,35 @@ +#ifndef ASM_EDAC_H +#define ASM_EDAC_H + +/* ECC atomic, DMA, SMP and interrupt safe scrub function */ + +static inline void atomic_scrub(void *va, u32 size) +{ + unsigned long *virt_addr = va; + unsigned long temp; + u32 i; + + for (i = 0; i < size / sizeof(unsigned long); i++, virt_addr++) { + + /* + * Very carefully read and write to memory atomically + * so we are interrupt, DMA and SMP safe. + * + * Intel: asm("lock; addl $0, %0"::"m"(*virt_addr)); + */ + + __asm__ __volatile__ ( + " .set mips3 \n" + "1: ll %0, %1 # atomic_add \n" + " ll %0, %1 # atomic_add \n" + " addu %0, $0 \n" + " sc %0, %1 \n" + " beqz %0, 1b \n" + " .set mips0 \n" + : "=&r" (temp), "=m" (*virt_addr) + : "m" (*virt_addr)); + + } +} + +#endif