Blame view

kernel/linux-rt-4.4.41/arch/avr32/include/asm/cmpxchg.h 2.8 KB
5113f6f70   김현기   kernel add
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
  /*
   * Atomic operations that C can't guarantee us.  Useful for
   * resource counting etc.
   *
   * But use these as seldom as possible since they are slower than
   * regular operations.
   *
   * Copyright (C) 2004-2006 Atmel Corporation
   *
   * This program is free software; you can redistribute it and/or modify
   * it under the terms of the GNU General Public License version 2 as
   * published by the Free Software Foundation.
   */
  #ifndef __ASM_AVR32_CMPXCHG_H
  #define __ASM_AVR32_CMPXCHG_H
  
  #define xchg(ptr,x) \
  	((__typeof__(*(ptr)))__xchg((unsigned long)(x),(ptr),sizeof(*(ptr))))
  
  extern void __xchg_called_with_bad_pointer(void);
  
  static inline unsigned long xchg_u32(u32 val, volatile u32 *m)
  {
  	u32 ret;
  
  	asm volatile("xchg %[ret], %[m], %[val]"
  			: [ret] "=&r"(ret), "=m"(*m)
  			: "m"(*m), [m] "r"(m), [val] "r"(val)
  			: "memory");
  	return ret;
  }
  
  static inline unsigned long __xchg(unsigned long x,
  				       volatile void *ptr,
  				       int size)
  {
  	switch(size) {
  	case 4:
  		return xchg_u32(x, ptr);
  	default:
  		__xchg_called_with_bad_pointer();
  		return x;
  	}
  }
  
  static inline unsigned long __cmpxchg_u32(volatile int *m, unsigned long old,
  					  unsigned long new)
  {
  	__u32 ret;
  
  	asm volatile(
  		"1:	ssrf	5
  "
  		"	ld.w	%[ret], %[m]
  "
  		"	cp.w	%[ret], %[old]
  "
  		"	brne	2f
  "
  		"	stcond	%[m], %[new]
  "
  		"	brne	1b
  "
  		"2:
  "
  		: [ret] "=&r"(ret), [m] "=m"(*m)
  		: "m"(m), [old] "ir"(old), [new] "r"(new)
  		: "memory", "cc");
  	return ret;
  }
  
  extern unsigned long __cmpxchg_u64_unsupported_on_32bit_kernels(
          volatile int * m, unsigned long old, unsigned long new);
  #define __cmpxchg_u64 __cmpxchg_u64_unsupported_on_32bit_kernels
  
  /* This function doesn't exist, so you'll get a linker error
     if something tries to do an invalid cmpxchg().  */
  extern void __cmpxchg_called_with_bad_pointer(void);
  
  static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old,
  				      unsigned long new, int size)
  {
  	switch (size) {
  	case 4:
  		return __cmpxchg_u32(ptr, old, new);
  	case 8:
  		return __cmpxchg_u64(ptr, old, new);
  	}
  
  	__cmpxchg_called_with_bad_pointer();
  	return old;
  }
  
  #define cmpxchg(ptr, old, new)					\
  	((typeof(*(ptr)))__cmpxchg((ptr), (unsigned long)(old),	\
  				   (unsigned long)(new),	\
  				   sizeof(*(ptr))))
  
  #include <asm-generic/cmpxchg-local.h>
  
  static inline unsigned long __cmpxchg_local(volatile void *ptr,
  				      unsigned long old,
  				      unsigned long new, int size)
  {
  	switch (size) {
  	case 4:
  		return __cmpxchg_u32(ptr, old, new);
  	default:
  		return __cmpxchg_local_generic(ptr, old, new, size);
  	}
  
  	return old;
  }
  
  #define cmpxchg_local(ptr, old, new)					\
  	((typeof(*(ptr)))__cmpxchg_local((ptr), (unsigned long)(old),	\
  				   (unsigned long)(new),		\
  				   sizeof(*(ptr))))
  
  #define cmpxchg64_local(ptr, o, n) __cmpxchg64_local_generic((ptr), (o), (n))
  
  #endif /* __ASM_AVR32_CMPXCHG_H */