comparison src/os_cpu/solaris_x86/vm/solaris_x86_64.il @ 0:a61af66fc99e jdk7-b24

Initial load
author duke
date Sat, 01 Dec 2007 00:00:00 +0000
parents
children 7b920868b475
comparison
equal deleted inserted replaced
-1:000000000000 0:a61af66fc99e
1 //
2 // Copyright 2004-2007 Sun Microsystems, Inc. All Rights Reserved.
3 // DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 //
5 // This code is free software; you can redistribute it and/or modify it
6 // under the terms of the GNU General Public License version 2 only, as
7 // published by the Free Software Foundation.
8 //
9 // This code is distributed in the hope that it will be useful, but WITHOUT
10 // ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 // FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 // version 2 for more details (a copy is included in the LICENSE file that
13 // accompanied this code).
14 //
15 // You should have received a copy of the GNU General Public License version
16 // 2 along with this work; if not, write to the Free Software Foundation,
17 // Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 //
19 // Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
20 // CA 95054 USA or visit www.sun.com if you need additional information or
21 // have any questions.
22 //
23 //
24
25 // The argument size of each inline directive is ignored by the compiler
26 // and is set to the number of arguments as documentation.
27
28 // Get the raw thread ID from %gs:0
29 .inline _raw_thread_id,0
30 movq %fs:0, %rax
31 .end
32
33 // Get the frame pointer from previous frame.
34 .inline _get_previous_fp,0
35 movq %rbp, %rax
36 movq %rax, %rax
37 .end
38
39 // Support for jint Atomic::add(jint add_value, volatile jint* dest)
40 // An additional bool (os::is_MP()) is passed as the last argument.
41 .inline _Atomic_add,3
42 movl %edi, %eax // save add_value for return
43 testl %edx, %edx // MP test
44 je 1f
45 lock
46 1: xaddl %edi, (%rsi)
47 addl %edi, %eax
48 .end
49
50 // Support for jlong Atomic::add(jlong add_value, volatile jlong* dest)
51 // An additional bool (os::is_MP()) is passed as the last argument.
52 .inline _Atomic_add_long,3
53 movq %rdi, %rax // save add_value for return
54 testq %rdx, %rdx // MP test
55 je 1f
56 lock
57 1: xaddq %rdi, (%rsi)
58 addq %rdi, %rax
59 .end
60
61 // Support for jint Atomic::xchg(jint exchange_value, volatile jint* dest).
62 .inline _Atomic_xchg,2
63 xchgl (%rsi), %edi
64 movl %edi, %eax
65 .end
66
67 // Support for jlong Atomic::xchg(jlong exchange_value, volatile jlong* dest).
68 .inline _Atomic_xchg_long,2
69 xchgq (%rsi), %rdi
70 movq %rdi, %rax
71 .end
72
73 // Support for jint Atomic::cmpxchg(jint exchange_value,
74 // volatile jint *dest,
75 // jint compare_value)
76 // An additional bool (os::is_MP()) is passed as the last argument.
77 .inline _Atomic_cmpxchg,4
78 movl %edx, %eax // compare_value
79 testl %ecx, %ecx // MP test
80 je 1f
81 lock
82 1: cmpxchgl %edi, (%rsi)
83 .end
84
85 // Support for jlong Atomic::cmpxchg(jlong exchange_value,
86 // volatile jlong* dest,
87 // jlong compare_value)
88 // An additional bool (os::is_MP()) is passed as the last argument.
89 .inline _Atomic_cmpxchg_long,6
90 movq %rdx, %rax // compare_value
91 testq %rcx, %rcx // MP test
92 je 1f
93 lock
94 1: cmpxchgq %rdi, (%rsi)
95 .end
96
97 // Support for OrderAccess::acquire()
98 .inline _OrderAccess_acquire,0
99 movl 0(%rsp), %eax
100 .end
101
102 // Support for OrderAccess::fence()
103 .inline _OrderAccess_fence,0
104 lock
105 addl $0, (%rsp)
106 .end
107
108 // Support for u2 Bytes::swap_u2(u2 x)
109 .inline _raw_swap_u2,1
110 movw %di, %ax
111 rorw $8, %ax
112 .end
113
114 // Support for u4 Bytes::swap_u4(u4 x)
115 .inline _raw_swap_u4,1
116 movl %edi, %eax
117 bswapl %eax
118 .end
119
120 // Support for u8 Bytes::swap_u8(u8 x)
121 .inline _raw_swap_u8,1
122 movq %rdi, %rax
123 bswapq %rax
124 .end
125
126 // Support for void Prefetch::read
127 .inline _Prefetch_read,2
128 prefetcht0 (%rdi, %rsi, 1)
129 .end
130
131 // Support for void Prefetch::write
132 // We use prefetcht0 because em64t doesn't support prefetchw.
133 // prefetchw is a 3dnow instruction.
134 .inline _Prefetch_write,2
135 prefetcht0 (%rdi, %rsi, 1)
136 .end