342
|
1 /*
|
|
2 * Copyright 2001-2007 Sun Microsystems, Inc. All Rights Reserved.
|
|
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
4 *
|
|
5 * This code is free software; you can redistribute it and/or modify it
|
|
6 * under the terms of the GNU General Public License version 2 only, as
|
|
7 * published by the Free Software Foundation.
|
|
8 *
|
|
9 * This code is distributed in the hope that it will be useful, but WITHOUT
|
|
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
12 * version 2 for more details (a copy is included in the LICENSE file that
|
|
13 * accompanied this code).
|
|
14 *
|
|
15 * You should have received a copy of the GNU General Public License version
|
|
16 * 2 along with this work; if not, write to the Free Software Foundation,
|
|
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
18 *
|
|
19 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
|
|
20 * CA 95054 USA or visit www.sun.com if you need additional information or
|
|
21 * have any questions.
|
|
22 *
|
|
23 */
|
|
24
|
|
25 inline HeapWord* G1BlockOffsetTable::block_start(const void* addr) {
|
|
26 if (addr >= _bottom && addr < _end) {
|
|
27 return block_start_unsafe(addr);
|
|
28 } else {
|
|
29 return NULL;
|
|
30 }
|
|
31 }
|
|
32
|
|
33 inline HeapWord*
|
|
34 G1BlockOffsetTable::block_start_const(const void* addr) const {
|
|
35 if (addr >= _bottom && addr < _end) {
|
|
36 return block_start_unsafe_const(addr);
|
|
37 } else {
|
|
38 return NULL;
|
|
39 }
|
|
40 }
|
|
41
|
|
42 inline size_t G1BlockOffsetSharedArray::index_for(const void* p) const {
|
|
43 char* pc = (char*)p;
|
|
44 assert(pc >= (char*)_reserved.start() &&
|
|
45 pc < (char*)_reserved.end(),
|
|
46 "p not in range.");
|
|
47 size_t delta = pointer_delta(pc, _reserved.start(), sizeof(char));
|
|
48 size_t result = delta >> LogN;
|
|
49 assert(result < _vs.committed_size(), "bad index from address");
|
|
50 return result;
|
|
51 }
|
|
52
|
|
53 inline HeapWord*
|
|
54 G1BlockOffsetSharedArray::address_for_index(size_t index) const {
|
|
55 assert(index < _vs.committed_size(), "bad index");
|
|
56 HeapWord* result = _reserved.start() + (index << LogN_words);
|
|
57 assert(result >= _reserved.start() && result < _reserved.end(),
|
|
58 "bad address from index");
|
|
59 return result;
|
|
60 }
|
|
61
|
|
62 inline HeapWord*
|
|
63 G1BlockOffsetArray::block_at_or_preceding(const void* addr,
|
|
64 bool has_max_index,
|
|
65 size_t max_index) const {
|
|
66 assert(_array->offset_array(0) == 0, "objects can't cross covered areas");
|
|
67 size_t index = _array->index_for(addr);
|
|
68 // We must make sure that the offset table entry we use is valid. If
|
|
69 // "addr" is past the end, start at the last known one and go forward.
|
|
70 if (has_max_index) {
|
|
71 index = MIN2(index, max_index);
|
|
72 }
|
|
73 HeapWord* q = _array->address_for_index(index);
|
|
74
|
|
75 uint offset = _array->offset_array(index); // Extend u_char to uint.
|
|
76 while (offset >= N_words) {
|
|
77 // The excess of the offset from N_words indicates a power of Base
|
|
78 // to go back by.
|
|
79 size_t n_cards_back = BlockOffsetArray::entry_to_cards_back(offset);
|
|
80 q -= (N_words * n_cards_back);
|
|
81 assert(q >= _sp->bottom(), "Went below bottom!");
|
|
82 index -= n_cards_back;
|
|
83 offset = _array->offset_array(index);
|
|
84 }
|
|
85 assert(offset < N_words, "offset too large");
|
|
86 q -= offset;
|
|
87 return q;
|
|
88 }
|
|
89
|
|
90 inline HeapWord*
|
|
91 G1BlockOffsetArray::
|
|
92 forward_to_block_containing_addr_const(HeapWord* q, HeapWord* n,
|
|
93 const void* addr) const {
|
|
94 if (csp() != NULL) {
|
|
95 if (addr >= csp()->top()) return csp()->top();
|
|
96 while (n <= addr) {
|
|
97 q = n;
|
|
98 oop obj = oop(q);
|
|
99 if (obj->klass() == NULL) return q;
|
|
100 n += obj->size();
|
|
101 }
|
|
102 } else {
|
|
103 while (n <= addr) {
|
|
104 q = n;
|
|
105 oop obj = oop(q);
|
|
106 if (obj->klass() == NULL) return q;
|
|
107 n += _sp->block_size(q);
|
|
108 }
|
|
109 }
|
|
110 assert(q <= n, "wrong order for q and addr");
|
|
111 assert(addr < n, "wrong order for addr and n");
|
|
112 return q;
|
|
113 }
|
|
114
|
|
115 inline HeapWord*
|
|
116 G1BlockOffsetArray::forward_to_block_containing_addr(HeapWord* q,
|
|
117 const void* addr) {
|
|
118 if (oop(q)->klass() == NULL) return q;
|
|
119 HeapWord* n = q + _sp->block_size(q);
|
|
120 // In the normal case, where the query "addr" is a card boundary, and the
|
|
121 // offset table chunks are the same size as cards, the block starting at
|
|
122 // "q" will contain addr, so the test below will fail, and we'll fall
|
|
123 // through quickly.
|
|
124 if (n <= addr) {
|
|
125 q = forward_to_block_containing_addr_slow(q, n, addr);
|
|
126 }
|
|
127 assert(q <= addr, "wrong order for current and arg");
|
|
128 return q;
|
|
129 }
|
|
130
|
|
131 //////////////////////////////////////////////////////////////////////////
|
|
132 // BlockOffsetArrayNonContigSpace inlines
|
|
133 //////////////////////////////////////////////////////////////////////////
|
|
134 inline void G1BlockOffsetArray::freed(HeapWord* blk_start, HeapWord* blk_end) {
|
|
135 // Verify that the BOT shows [blk_start, blk_end) to be one block.
|
|
136 verify_single_block(blk_start, blk_end);
|
|
137 // adjust _unallocated_block upward or downward
|
|
138 // as appropriate
|
|
139 if (BlockOffsetArrayUseUnallocatedBlock) {
|
|
140 assert(_unallocated_block <= _end,
|
|
141 "Inconsistent value for _unallocated_block");
|
|
142 if (blk_end >= _unallocated_block && blk_start <= _unallocated_block) {
|
|
143 // CMS-specific note: a block abutting _unallocated_block to
|
|
144 // its left is being freed, a new block is being added or
|
|
145 // we are resetting following a compaction
|
|
146 _unallocated_block = blk_start;
|
|
147 }
|
|
148 }
|
|
149 }
|
|
150
|
|
151 inline void G1BlockOffsetArray::freed(HeapWord* blk, size_t size) {
|
|
152 freed(blk, blk + size);
|
|
153 }
|