Code Review
/
kvmfornfv.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
review
|
tree
raw
|
inline
| side by side
These changes are the raw update to linux-4.4.6-rt14. Kernel sources
[kvmfornfv.git]
/
kernel
/
arch
/
blackfin
/
mach-bf561
/
atomic.S
diff --git
a/kernel/arch/blackfin/mach-bf561/atomic.S
b/kernel/arch/blackfin/mach-bf561/atomic.S
index
2a08df8
..
26fccb5
100644
(file)
--- a/
kernel/arch/blackfin/mach-bf561/atomic.S
+++ b/
kernel/arch/blackfin/mach-bf561/atomic.S
@@
-587,10
+587,10
@@
ENDPROC(___raw_write_unlock_asm)
* r0 = ptr
* r1 = value
*
* r0 = ptr
* r1 = value
*
- * A
dd
a signed value to a 32bit word and return the new value atomically.
+ * A
DD
a signed value to a 32bit word and return the new value atomically.
* Clobbers: r3:0, p1:0
*/
* Clobbers: r3:0, p1:0
*/
-ENTRY(___raw_atomic_
update
_asm)
+ENTRY(___raw_atomic_
add
_asm)
p1 = r0;
r3 = r1;
[--sp] = rets;
p1 = r0;
r3 = r1;
[--sp] = rets;
@@
-603,19
+603,19
@@
ENTRY(___raw_atomic_update_asm)
r0 = r3;
rets = [sp++];
rts;
r0 = r3;
rets = [sp++];
rts;
-ENDPROC(___raw_atomic_
update
_asm)
+ENDPROC(___raw_atomic_
add
_asm)
/*
* r0 = ptr
* r1 = mask
*
/*
* r0 = ptr
* r1 = mask
*
- *
Clear
the mask bits from a 32bit word and return the old 32bit value
+ *
AND
the mask bits from a 32bit word and return the old 32bit value
* atomically.
* Clobbers: r3:0, p1:0
*/
* atomically.
* Clobbers: r3:0, p1:0
*/
-ENTRY(___raw_atomic_
clear
_asm)
+ENTRY(___raw_atomic_
and
_asm)
p1 = r0;
p1 = r0;
- r3 =
~
r1;
+ r3 = r1;
[--sp] = rets;
call _get_core_lock;
r2 = [p1];
[--sp] = rets;
call _get_core_lock;
r2 = [p1];
@@
-627,17
+627,17
@@
ENTRY(___raw_atomic_clear_asm)
r0 = r3;
rets = [sp++];
rts;
r0 = r3;
rets = [sp++];
rts;
-ENDPROC(___raw_atomic_
clear
_asm)
+ENDPROC(___raw_atomic_
and
_asm)
/*
* r0 = ptr
* r1 = mask
*
/*
* r0 = ptr
* r1 = mask
*
- *
Set
the mask bits into a 32bit word and return the old 32bit value
+ *
OR
the mask bits into a 32bit word and return the old 32bit value
* atomically.
* Clobbers: r3:0, p1:0
*/
* atomically.
* Clobbers: r3:0, p1:0
*/
-ENTRY(___raw_atomic_
set
_asm)
+ENTRY(___raw_atomic_
or
_asm)
p1 = r0;
r3 = r1;
[--sp] = rets;
p1 = r0;
r3 = r1;
[--sp] = rets;
@@
-651,7
+651,7
@@
ENTRY(___raw_atomic_set_asm)
r0 = r3;
rets = [sp++];
rts;
r0 = r3;
rets = [sp++];
rts;
-ENDPROC(___raw_atomic_
set
_asm)
+ENDPROC(___raw_atomic_
or
_asm)
/*
* r0 = ptr
/*
* r0 = ptr
@@
-787,7
+787,7
@@
ENTRY(___raw_bit_set_asm)
r2 = r1;
r1 = 1;
r1 <<= r2;
r2 = r1;
r1 = 1;
r1 <<= r2;
- jump ___raw_atomic_
set
_asm
+ jump ___raw_atomic_
or
_asm
ENDPROC(___raw_bit_set_asm)
/*
ENDPROC(___raw_bit_set_asm)
/*
@@
-798,10
+798,10
@@
ENDPROC(___raw_bit_set_asm)
* Clobbers: r3:0, p1:0
*/
ENTRY(___raw_bit_clear_asm)
* Clobbers: r3:0, p1:0
*/
ENTRY(___raw_bit_clear_asm)
- r2 =
r
1;
- r
1 =
1;
- r1
<<=
r2;
- jump ___raw_atomic_
clear
_asm
+ r2 = 1;
+ r
2 <<= r
1;
+ r1
= ~
r2;
+ jump ___raw_atomic_
and
_asm
ENDPROC(___raw_bit_clear_asm)
/*
ENDPROC(___raw_bit_clear_asm)
/*