Mercurial > hg > graal-jvmci-8
diff src/share/vm/memory/threadLocalAllocBuffer.inline.hpp @ 0:a61af66fc99e jdk7-b24
Initial load
author | duke |
---|---|
date | Sat, 01 Dec 2007 00:00:00 +0000 |
parents | |
children | 2dd52dea6d28 |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/src/share/vm/memory/threadLocalAllocBuffer.inline.hpp Sat Dec 01 00:00:00 2007 +0000 @@ -0,0 +1,90 @@ +/* + * Copyright 1999-2007 Sun Microsystems, Inc. All Rights Reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, + * CA 95054 USA or visit www.sun.com if you need additional information or + * have any questions. + * + */ + +inline HeapWord* ThreadLocalAllocBuffer::allocate(size_t size) { + invariants(); + HeapWord* obj = top(); + if (pointer_delta(end(), obj) >= size) { + // successful thread-local allocation + + DEBUG_ONLY(Copy::fill_to_words(obj, size, badHeapWordVal)); + // This addition is safe because we know that top is + // at least size below end, so the add can't wrap. + set_top(obj + size); + + invariants(); + return obj; + } + return NULL; +} + +inline size_t ThreadLocalAllocBuffer::compute_size(size_t obj_size) { + const size_t aligned_obj_size = align_object_size(obj_size); + + // Compute the size for the new TLAB. + // The "last" tlab may be smaller to reduce fragmentation. + // unsafe_max_tlab_alloc is just a hint. + const size_t available_size = Universe::heap()->unsafe_max_tlab_alloc(myThread()) / + HeapWordSize; + size_t new_tlab_size = MIN2(available_size, desired_size() + aligned_obj_size); + + // Make sure there's enough room for object and filler int[]. + const size_t obj_plus_filler_size = aligned_obj_size + alignment_reserve(); + if (new_tlab_size < obj_plus_filler_size) { + // If there isn't enough room for the allocation, return failure. + if (PrintTLAB && Verbose) { + gclog_or_tty->print_cr("ThreadLocalAllocBuffer::compute_size(" SIZE_FORMAT ")" + " returns failure", + obj_size); + } + return 0; + } + if (PrintTLAB && Verbose) { + gclog_or_tty->print_cr("ThreadLocalAllocBuffer::compute_size(" SIZE_FORMAT ")" + " returns " SIZE_FORMAT, + obj_size, new_tlab_size); + } + return new_tlab_size; +} + + +void ThreadLocalAllocBuffer::record_slow_allocation(size_t obj_size) { + // Raise size required to bypass TLAB next time. Why? Else there's + // a risk that a thread that repeatedly allocates objects of one + // size will get stuck on this slow path. + + set_refill_waste_limit(refill_waste_limit() + refill_waste_limit_increment()); + + _slow_allocations++; + + if (PrintTLAB && Verbose) { + Thread* thrd = myThread(); + gclog_or_tty->print("TLAB: %s thread: "INTPTR_FORMAT" [id: %2d]" + " obj: "SIZE_FORMAT + " free: "SIZE_FORMAT + " waste: "SIZE_FORMAT"\n", + "slow", thrd, thrd->osthread()->thread_id(), + obj_size, free(), refill_waste_limit()); + } +}