File: atomic.h

package info (click to toggle)
xbmc-pvr-addons 13.0%2Bgit20140512%2Bg91cc731%2Bdfsg1-3
  • links: PTS, VCS
  • area: main
  • in suites: stretch
  • size: 12,740 kB
  • ctags: 10,440
  • sloc: cpp: 59,563; ansic: 17,810; xml: 3,635; makefile: 400; sh: 42
file content (216 lines) | stat: -rw-r--r-- 5,836 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
/*
 *  Copyright (C) 2004-2008, Eric Lund, Jon Gettler
 *  http://www.mvpmc.org/
 *
 *  This library is free software; you can redistribute it and/or
 *  modify it under the terms of the GNU Lesser General Public
 *  License as published by the Free Software Foundation; either
 *  version 2.1 of the License, or (at your option) any later version.
 *
 *  This library is distributed in the hope that it will be useful,
 *  but WITHOUT ANY WARRANTY; without even the implied warranty of
 *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 *  Lesser General Public License for more details.
 *
 *  You should have received a copy of the GNU Lesser General Public
 *  License along with this library; if not, write to the Free Software
 *  Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
 */

#ifndef __MVP_ATOMIC_H
#define __MVP_ATOMIC_H

#ifdef __APPLE__
#pragma GCC optimization_level 0
#endif

#if defined(_MSC_VER)
#include <windows.h>
#define inline __inline
#endif

#if defined(__APPLE__)
#include <libkern/OSAtomic.h>

typedef	volatile int32_t mvp_atomic_t;

#define __mvp_atomic_increment(x)	OSAtomicIncrement32(x)
#define __mvp_atomic_decrement(x)	OSAtomicDecrement32(x)
#else
typedef	volatile unsigned int mvp_atomic_t;

/**
 * Atomically incremente a reference count variable.
 * \param valp address of atomic variable
 * \return incremented reference count
 */
static inline unsigned
__mvp_atomic_increment(mvp_atomic_t *valp)
{
	mvp_atomic_t __val;
#if defined __i486__ || defined __i586__ || defined __i686__
	__asm__ __volatile__(
		"lock xaddl %0, (%1);"
		"     inc   %0;"
		: "=r" (__val)
		: "r" (valp), "0" (0x1)
		: "cc", "memory"
		);
#elif defined __i386__ || defined __x86_64__
	__asm__ volatile (".byte 0xf0, 0x0f, 0xc1, 0x02" /*lock; xaddl %eax, (%edx) */
		      : "=a" (__val)
		      : "0" (1), "m" (*valp), "d" (valp)
		      : "memory");
	/* on the x86 __val is the pre-increment value, so normalize it. */
	++__val;
#elif defined __powerpc__ || defined __ppc__
	asm volatile ("1:	lwarx   %0,0,%1\n"
		      "	addic.   %0,%0,1\n"
		      "	dcbt    %0,%1\n"
		      "	stwcx.  %0,0,%1\n"
		      "	bne-    1b\n"
		      "	isync\n"
		      : "=&r" (__val)
		      : "r" (valp)
		      : "cc", "memory");
#elif defined _MSC_VER
	__val = InterlockedIncrement(valp);
#elif defined ANDROID
	__val = __sync_add_and_fetch(valp, 1);
#elif defined __arm__ && !defined __thumb__
	int tmp1, tmp2;
	int inc = 1;
	__asm__ __volatile__ (
		"\n"
		"0:"
		"ldr     %0, [%3]\n"
		"add     %1, %0, %4\n"
		"swp     %2, %1, [%3]\n"
		"cmp     %0, %2\n"
		"swpne   %0, %2, [%3]\n"
		"bne     0b\n"
		: "=&r"(tmp1), "=&r"(__val), "=&r"(tmp2)
		: "r" (valp), "r"(inc)
		: "cc", "memory");
#elif defined __GNUC__
	/*
	 * Don't know how to atomic increment for a generic architecture
	 * so try to use GCC builtin
	 */
	__val = __sync_add_and_fetch(valp,1);
#else
#warning unknown architecture, atomic increment is not...
	__val = ++(*valp);
#endif
	return __val;
}

/**
 * Atomically decrement a reference count variable.
 * \param valp address of atomic variable
 * \return decremented reference count
 */
static inline unsigned
__mvp_atomic_decrement(mvp_atomic_t *valp)
{
	mvp_atomic_t __val;
#if defined __i486__ || defined __i586__ || defined __i686__
	__asm__ __volatile__(
		"lock xaddl %0, (%1);"
		"     dec   %0;"
		: "=r" (__val)
		: "r" (valp), "0" (-0x1)
		: "cc", "memory"
		);
#elif defined __i386__ || defined __x86_64__
	__asm__ volatile (".byte 0xf0, 0x0f, 0xc1, 0x02" /*lock; xaddl %eax, (%edx) */
		      : "=a" (__val)
		      : "0" (-1), "m" (*valp), "d" (valp)
		      : "memory");
	/* __val is the pre-decrement value, so normalize it */
	--__val;
#elif defined __powerpc__ || defined __ppc__
	asm volatile ("1:	lwarx   %0,0,%1\n"
		      "	addic.   %0,%0,-1\n"
		      "	dcbt    %0,%1\n"
		      "	stwcx.  %0,0,%1\n"
		      "	bne-    1b\n"
		      "	isync\n"
		      : "=&r" (__val)
		      : "r" (valp)
		      : "cc", "memory");
#elif defined ANDROID
	__val = __sync_sub_and_fetch(valp, 1);
#elif defined __arm__ && !defined __thumb__
	int tmp1, tmp2;
	int inc = -1;
	__asm__ __volatile__ (
		"\n"
		"0:"
		"ldr     %0, [%3]\n"
		"add     %1, %0, %4\n"
		"swp     %2, %1, [%3]\n"
		"cmp     %0, %2\n"
		"swpne   %0, %2, [%3]\n"
		"bne     0b\n"
		: "=&r"(tmp1), "=&r"(__val), "=&r"(tmp2)
		: "r" (valp), "r"(inc)
		: "cc", "memory");
#elif defined __sparcv9__
	mvp_atomic_t __newval, __oldval = (*valp);
	do
		{
			__newval = __oldval - 1;
			__asm__ ("cas	[%4], %2, %0"
				 : "=r" (__oldval), "=m" (*valp)
				 : "r" (__oldval), "m" (*valp), "r"((valp)), "0" (__newval));
		}
	while (__newval != __oldval);
	/*  The value for __val is in '__oldval' */
	__val = __oldval;
#elif defined _MSC_VER
	__val = InterlockedDecrement(valp);
#elif defined __GNUC__
	/*
	 * Don't know how to atomic decrement for a generic architecture
	 * so use GCC builtin
	 */
	__val = __sync_sub_and_fetch(valp,1);
#else
#warning unknown architecture, atomic deccrement is not...
	__val = --(*valp);
#endif
	return __val;
}
#endif

#define mvp_atomic_inc __mvp_atomic_inc
static inline int mvp_atomic_inc(mvp_atomic_t *a) {
	return __mvp_atomic_increment(a);
};

#define mvp_atomic_dec __mvp_atomic_dec
static inline int mvp_atomic_dec(mvp_atomic_t *a) {
	return __mvp_atomic_decrement(a);
};

#define mvp_atomic_dec_and_test __mvp_atomic_dec_and_test
static inline int mvp_atomic_dec_and_test(mvp_atomic_t *a) {
	return (__mvp_atomic_decrement(a) == 0);
};

#define mvp_atomic_set __mvp_atomic_set
static inline void mvp_atomic_set(mvp_atomic_t *a, unsigned val) {
	*a = val;
};

#define mvp_atomic_val __mvp_atomic_val
static inline int mvp_atomic_val(mvp_atomic_t *a) {
	return *a;
};

#ifdef __APPLE__
#pragma GCC optimization_level reset
#endif

#endif  /* __MVP_ATOMIC_H */