Mercurial > hg > truffle
diff src/share/vm/utilities/copy.cpp @ 0:a61af66fc99e jdk7-b24
Initial load
author | duke |
---|---|
date | Sat, 01 Dec 2007 00:00:00 +0000 |
parents | |
children | c18cbe5936b8 |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/src/share/vm/utilities/copy.cpp Sat Dec 01 00:00:00 2007 +0000 @@ -0,0 +1,92 @@ +/* + * Copyright 2006-2007 Sun Microsystems, Inc. All Rights Reserved. + * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. + * + * This code is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 only, as + * published by the Free Software Foundation. + * + * This code is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License + * version 2 for more details (a copy is included in the LICENSE file that + * accompanied this code). + * + * You should have received a copy of the GNU General Public License version + * 2 along with this work; if not, write to the Free Software Foundation, + * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. + * + * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, + * CA 95054 USA or visit www.sun.com if you need additional information or + * have any questions. + * + */ + +# include "incls/_precompiled.incl" +# include "incls/_copy.cpp.incl" + + +// Copy bytes; larger units are filled atomically if everything is aligned. +void Copy::conjoint_memory_atomic(void* from, void* to, size_t size) { + address src = (address) from; + address dst = (address) to; + uintptr_t bits = (uintptr_t) src | (uintptr_t) dst | (uintptr_t) size; + + // (Note: We could improve performance by ignoring the low bits of size, + // and putting a short cleanup loop after each bulk copy loop. + // There are plenty of other ways to make this faster also, + // and it's a slippery slope. For now, let's keep this code simple + // since the simplicity helps clarify the atomicity semantics of + // this operation. There are also CPU-specific assembly versions + // which may or may not want to include such optimizations.) + + if (bits % sizeof(jlong) == 0) { + Copy::conjoint_jlongs_atomic((jlong*) src, (jlong*) dst, size / sizeof(jlong)); + } else if (bits % sizeof(jint) == 0) { + Copy::conjoint_jints_atomic((jint*) src, (jint*) dst, size / sizeof(jint)); + } else if (bits % sizeof(jshort) == 0) { + Copy::conjoint_jshorts_atomic((jshort*) src, (jshort*) dst, size / sizeof(jshort)); + } else { + // Not aligned, so no need to be atomic. + Copy::conjoint_bytes((void*) src, (void*) dst, size); + } +} + + +// Fill bytes; larger units are filled atomically if everything is aligned. +void Copy::fill_to_memory_atomic(void* to, size_t size, jubyte value) { + address dst = (address) to; + uintptr_t bits = (uintptr_t) to | (uintptr_t) size; + if (bits % sizeof(jlong) == 0) { + jlong fill = (julong)( (jubyte)value ); // zero-extend + if (fill != 0) { + fill += fill << 8; + fill += fill << 16; + fill += fill << 32; + } + //Copy::fill_to_jlongs_atomic((jlong*) dst, size / sizeof(jlong)); + for (uintptr_t off = 0; off < size; off += sizeof(jlong)) { + *(jlong*)(dst + off) = fill; + } + } else if (bits % sizeof(jint) == 0) { + jint fill = (juint)( (jubyte)value ); // zero-extend + if (fill != 0) { + fill += fill << 8; + fill += fill << 16; + } + //Copy::fill_to_jints_atomic((jint*) dst, size / sizeof(jint)); + for (uintptr_t off = 0; off < size; off += sizeof(jint)) { + *(jint*)(dst + off) = fill; + } + } else if (bits % sizeof(jshort) == 0) { + jshort fill = (jushort)( (jubyte)value ); // zero-extend + fill += fill << 8; + //Copy::fill_to_jshorts_atomic((jshort*) dst, size / sizeof(jshort)); + for (uintptr_t off = 0; off < size; off += sizeof(jshort)) { + *(jshort*)(dst + off) = fill; + } + } else { + // Not aligned, so no need to be atomic. + Copy::fill_to_bytes(dst, size, value); + } +}