1.1 --- a/src/cpu/x86/vm/stubGenerator_x86_32.cpp Tue Apr 06 15:18:10 2010 -0700 1.2 +++ b/src/cpu/x86/vm/stubGenerator_x86_32.cpp Wed Apr 07 09:37:47 2010 -0700 1.3 @@ -812,7 +812,7 @@ 1.4 Label L_copy_64_bytes_loop, L_copy_64_bytes, L_copy_8_bytes, L_exit; 1.5 // Copy 64-byte chunks 1.6 __ jmpb(L_copy_64_bytes); 1.7 - __ align(16); 1.8 + __ align(OptoLoopAlignment); 1.9 __ BIND(L_copy_64_bytes_loop); 1.10 1.11 if(UseUnalignedLoadStores) { 1.12 @@ -874,7 +874,7 @@ 1.13 Label L_copy_64_bytes_loop, L_copy_64_bytes, L_copy_8_bytes, L_exit; 1.14 // Copy 64-byte chunks 1.15 __ jmpb(L_copy_64_bytes); 1.16 - __ align(16); 1.17 + __ align(OptoLoopAlignment); 1.18 __ BIND(L_copy_64_bytes_loop); 1.19 __ movq(mmx0, Address(from, 0)); 1.20 __ movq(mmx1, Address(from, 8)); 1.21 @@ -1144,7 +1144,7 @@ 1.22 __ movl(Address(to, count, sf, 0), rdx); 1.23 __ jmpb(L_copy_8_bytes); 1.24 1.25 - __ align(16); 1.26 + __ align(OptoLoopAlignment); 1.27 // Move 8 bytes 1.28 __ BIND(L_copy_8_bytes_loop); 1.29 if (UseXMMForArrayCopy) { 1.30 @@ -1235,7 +1235,7 @@ 1.31 } 1.32 } else { 1.33 __ jmpb(L_copy_8_bytes); 1.34 - __ align(16); 1.35 + __ align(OptoLoopAlignment); 1.36 __ BIND(L_copy_8_bytes_loop); 1.37 __ fild_d(Address(from, 0)); 1.38 __ fistp_d(Address(from, to_from, Address::times_1)); 1.39 @@ -1282,7 +1282,7 @@ 1.40 1.41 __ jmpb(L_copy_8_bytes); 1.42 1.43 - __ align(16); 1.44 + __ align(OptoLoopAlignment); 1.45 __ BIND(L_copy_8_bytes_loop); 1.46 if (VM_Version::supports_mmx()) { 1.47 if (UseXMMForArrayCopy) { 1.48 @@ -1454,7 +1454,7 @@ 1.49 // Loop control: 1.50 // for (count = -count; count != 0; count++) 1.51 // Base pointers src, dst are biased by 8*count,to last element. 1.52 - __ align(16); 1.53 + __ align(OptoLoopAlignment); 1.54 1.55 __ BIND(L_store_element); 1.56 __ movptr(to_element_addr, elem); // store the oop