2371 lines
102 KiB
ArmAsm
2371 lines
102 KiB
ArmAsm
/*
|
|
* Copyright (C) 2012 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include "asm_support_x86.S"
|
|
#include "interpreter/cfi_asm_support.h"
|
|
|
|
#include "arch/quick_alloc_entrypoints.S"
|
|
|
|
// For x86, the CFA is esp+4, the address above the pushed return address on the stack.
|
|
|
|
/*
|
|
* Macro that sets up the callee save frame to conform with
|
|
* Runtime::CreateCalleeSaveMethod(kSaveRefsOnly)
|
|
* and preserves the value of temp_reg at entry.
|
|
*/
|
|
MACRO1(SETUP_SAVE_REFS_ONLY_FRAME_PRESERVE_TEMP_REG, temp_reg)
|
|
PUSH edi // Save callee saves (ebx is saved/restored by the upcall)
|
|
PUSH esi
|
|
PUSH ebp
|
|
PUSH RAW_VAR(temp_reg) // Save temp_reg
|
|
subl MACRO_LITERAL(8), %esp // Grow stack by 2 words.
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
|
|
LOAD_RUNTIME_INSTANCE \temp_reg
|
|
// Push save all callee-save method.
|
|
pushl RUNTIME_SAVE_REFS_ONLY_METHOD_OFFSET(REG_VAR(temp_reg))
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
// Store esp as the top quick frame.
|
|
movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET
|
|
// Restore temp_reg.
|
|
movl 12(%esp), REG_VAR(temp_reg)
|
|
CFI_RESTORE(RAW_VAR(temp_reg))
|
|
|
|
// Ugly compile-time check, but we only have the preprocessor.
|
|
// Last +4: implicit return address pushed on stack when caller made call.
|
|
#if (FRAME_SIZE_SAVE_REFS_ONLY != 3*4 + 16 + 4)
|
|
#error "FRAME_SIZE_SAVE_REFS_ONLY(X86) size not as expected."
|
|
#endif
|
|
END_MACRO
|
|
|
|
/*
|
|
* Macro that sets up the callee save frame to conform with
|
|
* Runtime::CreateCalleeSaveMethod(kSaveRefsAndArgs)
|
|
*/
|
|
MACRO1(SETUP_SAVE_REFS_AND_ARGS_FRAME, temp_reg)
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME_REGISTERS_ONLY
|
|
|
|
LOAD_RUNTIME_INSTANCE \temp_reg
|
|
// Push save all callee-save method.
|
|
pushl RUNTIME_SAVE_REFS_AND_ARGS_METHOD_OFFSET(REG_VAR(temp_reg))
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
// Store esp as the stop quick frame.
|
|
movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET
|
|
END_MACRO
|
|
|
|
/*
|
|
* Macro that sets up the callee save frame to conform with
|
|
* Runtime::CreateCalleeSaveMethod(kSaveRefsAndArgs) where the method is passed in EAX.
|
|
*/
|
|
MACRO0(SETUP_SAVE_REFS_AND_ARGS_FRAME_WITH_METHOD_IN_EAX)
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME_REGISTERS_ONLY
|
|
|
|
pushl %eax // Store the ArtMethod reference at the bottom of the stack.
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
// Store esp as the stop quick frame.
|
|
movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET
|
|
END_MACRO
|
|
|
|
// Restore register and jump to routine
|
|
// Inputs: EDI contains pointer to code.
|
|
// Notes: Need to pop EAX too (restores Method*)
|
|
MACRO0(RESTORE_SAVE_REFS_AND_ARGS_FRAME_AND_JUMP)
|
|
POP eax // Restore Method*
|
|
|
|
// Restore FPRs.
|
|
movsd 0(%esp), %xmm0
|
|
movsd 8(%esp), %xmm1
|
|
movsd 16(%esp), %xmm2
|
|
movsd 24(%esp), %xmm3
|
|
|
|
addl MACRO_LITERAL(32), %esp // Remove FPRs.
|
|
CFI_ADJUST_CFA_OFFSET(-32)
|
|
|
|
POP ecx // Restore args except eax
|
|
POP edx
|
|
POP ebx
|
|
POP ebp // Restore callee saves
|
|
POP esi
|
|
xchgl 0(%esp),%edi // restore EDI and place code pointer as only value on stack
|
|
ret
|
|
END_MACRO
|
|
|
|
/*
|
|
* Macro that sets up the callee save frame to conform with
|
|
* Runtime::CreateCalleeSaveMethod(kSaveEverything)
|
|
* when EDI and ESI are already saved.
|
|
*/
|
|
MACRO2(SETUP_SAVE_EVERYTHING_FRAME_EDI_ESI_SAVED, temp_reg, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET)
|
|
// Save core registers from highest to lowest to agree with core spills bitmap.
|
|
// EDI and ESI, or at least placeholders for them, are already on the stack.
|
|
PUSH ebp
|
|
PUSH ebx
|
|
PUSH edx
|
|
PUSH ecx
|
|
PUSH eax
|
|
// Create space for FPR registers and stack alignment padding.
|
|
subl MACRO_LITERAL(12 + 8 * 8), %esp
|
|
CFI_ADJUST_CFA_OFFSET(12 + 8 * 8)
|
|
// Save FPRs.
|
|
movsd %xmm0, 12(%esp)
|
|
movsd %xmm1, 20(%esp)
|
|
movsd %xmm2, 28(%esp)
|
|
movsd %xmm3, 36(%esp)
|
|
movsd %xmm4, 44(%esp)
|
|
movsd %xmm5, 52(%esp)
|
|
movsd %xmm6, 60(%esp)
|
|
movsd %xmm7, 68(%esp)
|
|
|
|
LOAD_RUNTIME_INSTANCE \temp_reg
|
|
// Push save everything callee-save method.
|
|
pushl \runtime_method_offset(REG_VAR(temp_reg))
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
// Store esp as the stop quick frame.
|
|
movl %esp, %fs:THREAD_TOP_QUICK_FRAME_OFFSET
|
|
|
|
// Ugly compile-time check, but we only have the preprocessor.
|
|
// Last +4: implicit return address pushed on stack when caller made call.
|
|
#if (FRAME_SIZE_SAVE_EVERYTHING != 7*4 + 8*8 + 12 + 4 + 4)
|
|
#error "FRAME_SIZE_SAVE_EVERYTHING(X86) size not as expected."
|
|
#endif
|
|
END_MACRO
|
|
|
|
/*
|
|
* Macro that sets up the callee save frame to conform with
|
|
* Runtime::CreateCalleeSaveMethod(kSaveEverything)
|
|
* when EDI is already saved.
|
|
*/
|
|
MACRO2(SETUP_SAVE_EVERYTHING_FRAME_EDI_SAVED, temp_reg, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET)
|
|
// Save core registers from highest to lowest to agree with core spills bitmap.
|
|
// EDI, or at least a placeholder for it, is already on the stack.
|
|
PUSH esi
|
|
SETUP_SAVE_EVERYTHING_FRAME_EDI_ESI_SAVED RAW_VAR(temp_reg), \runtime_method_offset
|
|
END_MACRO
|
|
|
|
/*
|
|
* Macro that sets up the callee save frame to conform with
|
|
* Runtime::CreateCalleeSaveMethod(kSaveEverything)
|
|
*/
|
|
MACRO2(SETUP_SAVE_EVERYTHING_FRAME, temp_reg, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET)
|
|
PUSH edi
|
|
SETUP_SAVE_EVERYTHING_FRAME_EDI_SAVED RAW_VAR(temp_reg), \runtime_method_offset
|
|
END_MACRO
|
|
|
|
MACRO0(RESTORE_SAVE_EVERYTHING_FRAME_FRPS)
|
|
// Restore FPRs. Method and padding is still on the stack.
|
|
movsd 16(%esp), %xmm0
|
|
movsd 24(%esp), %xmm1
|
|
movsd 32(%esp), %xmm2
|
|
movsd 40(%esp), %xmm3
|
|
movsd 48(%esp), %xmm4
|
|
movsd 56(%esp), %xmm5
|
|
movsd 64(%esp), %xmm6
|
|
movsd 72(%esp), %xmm7
|
|
END_MACRO
|
|
|
|
MACRO0(RESTORE_SAVE_EVERYTHING_FRAME_GPRS_EXCEPT_EAX)
|
|
// Restore core registers (except eax).
|
|
POP ecx
|
|
POP edx
|
|
POP ebx
|
|
POP ebp
|
|
POP esi
|
|
POP edi
|
|
END_MACRO
|
|
|
|
MACRO0(RESTORE_SAVE_EVERYTHING_FRAME)
|
|
RESTORE_SAVE_EVERYTHING_FRAME_FRPS
|
|
|
|
// Remove save everything callee save method, stack alignment padding and FPRs.
|
|
addl MACRO_LITERAL(16 + 8 * 8), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-(16 + 8 * 8))
|
|
|
|
POP eax
|
|
RESTORE_SAVE_EVERYTHING_FRAME_GPRS_EXCEPT_EAX
|
|
END_MACRO
|
|
|
|
MACRO0(RESTORE_SAVE_EVERYTHING_FRAME_KEEP_EAX)
|
|
RESTORE_SAVE_EVERYTHING_FRAME_FRPS
|
|
|
|
// Remove save everything callee save method, stack alignment padding and FPRs, skip EAX.
|
|
addl MACRO_LITERAL(16 + 8 * 8 + 4), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-(16 + 8 * 8 + 4))
|
|
|
|
RESTORE_SAVE_EVERYTHING_FRAME_GPRS_EXCEPT_EAX
|
|
END_MACRO
|
|
|
|
MACRO2(NO_ARG_RUNTIME_EXCEPTION, c_name, cxx_name)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx // save all registers as basis for long jump context
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(12), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call CALLVAR(cxx_name) // cxx_name(Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO2(NO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING, c_name, cxx_name)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx // save all registers as basis for long jump context
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(12), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call CALLVAR(cxx_name) // cxx_name(Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO2(ONE_ARG_RUNTIME_EXCEPTION, c_name, cxx_name)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx // save all registers as basis for long jump context
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(8), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO2(TWO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING, c_name, cxx_name)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx // save all registers as basis for long jump context
|
|
// Outgoing argument set up
|
|
PUSH eax // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, arg2, Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
/*
|
|
* Called by managed code to create and deliver a NullPointerException.
|
|
*/
|
|
NO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_null_pointer_exception, artThrowNullPointerExceptionFromCode
|
|
|
|
/*
|
|
* Call installed by a signal handler to create and deliver a NullPointerException.
|
|
*/
|
|
DEFINE_FUNCTION_CUSTOM_CFA art_quick_throw_null_pointer_exception_from_signal, 2 * __SIZEOF_POINTER__
|
|
// Fault address and return address were saved by the fault handler.
|
|
// Save all registers as basis for long jump context; EDI will replace fault address later.
|
|
SETUP_SAVE_EVERYTHING_FRAME_EDI_SAVED ebx
|
|
// Retrieve fault address and save EDI.
|
|
movl (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__)(%esp), %eax
|
|
movl %edi, (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__)(%esp)
|
|
CFI_REL_OFFSET(%edi, (FRAME_SIZE_SAVE_EVERYTHING - 2 * __SIZEOF_POINTER__))
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(8), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax // pass arg1
|
|
call SYMBOL(artThrowNullPointerExceptionFromSignal) // (addr, self)
|
|
UNREACHABLE
|
|
END_FUNCTION art_quick_throw_null_pointer_exception_from_signal
|
|
|
|
/*
|
|
* Called by managed code to create and deliver an ArithmeticException.
|
|
*/
|
|
NO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_div_zero, artThrowDivZeroFromCode
|
|
|
|
/*
|
|
* Called by managed code to create and deliver a StackOverflowError.
|
|
*/
|
|
NO_ARG_RUNTIME_EXCEPTION art_quick_throw_stack_overflow, artThrowStackOverflowFromCode
|
|
|
|
/*
|
|
* Called by managed code, saves callee saves and then calls artThrowException
|
|
* that will place a mock Method* at the bottom of the stack. Arg1 holds the exception.
|
|
*/
|
|
ONE_ARG_RUNTIME_EXCEPTION art_quick_deliver_exception, artDeliverExceptionFromCode
|
|
|
|
/*
|
|
* Called by managed code to create and deliver an ArrayIndexOutOfBoundsException. Arg1 holds
|
|
* index, arg2 holds limit.
|
|
*/
|
|
TWO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_array_bounds, artThrowArrayBoundsFromCode
|
|
|
|
/*
|
|
* Called by managed code to create and deliver a StringIndexOutOfBoundsException
|
|
* as if thrown from a call to String.charAt(). Arg1 holds index, arg2 holds limit.
|
|
*/
|
|
TWO_ARG_RUNTIME_EXCEPTION_SAVE_EVERYTHING art_quick_throw_string_bounds, artThrowStringBoundsFromCode
|
|
|
|
/*
|
|
* All generated callsites for interface invokes and invocation slow paths will load arguments
|
|
* as usual - except instead of loading arg0/r0 with the target Method*, arg0/r0 will contain
|
|
* the method_idx. This wrapper will save arg1-arg3 and call the appropriate C helper.
|
|
* NOTE: "this" is first visible argument of the target, and so can be found in arg1/r1.
|
|
*
|
|
* The helper will attempt to locate the target and return a 64-bit result in r0/r1 consisting
|
|
* of the target Method* in r0 and method->code_ in r1.
|
|
*
|
|
* If unsuccessful, the helper will return null/null and there will be a pending exception in the
|
|
* thread and we branch to another stub to deliver it.
|
|
*
|
|
* On success this wrapper will restore arguments and *jump* to the target, leaving the lr
|
|
* pointing back to the original caller.
|
|
*/
|
|
MACRO1(INVOKE_TRAMPOLINE_BODY, cxx_name)
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME ebx
|
|
movl %esp, %edx // remember SP
|
|
|
|
// Outgoing argument set up
|
|
PUSH edx // pass SP
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, arg2, Thread*, SP)
|
|
movl %edx, %edi // save code pointer in EDI
|
|
addl MACRO_LITERAL(20), %esp // Pop arguments skip eax
|
|
CFI_ADJUST_CFA_OFFSET(-20)
|
|
|
|
// Restore FPRs.
|
|
movsd 0(%esp), %xmm0
|
|
movsd 8(%esp), %xmm1
|
|
movsd 16(%esp), %xmm2
|
|
movsd 24(%esp), %xmm3
|
|
|
|
// Remove space for FPR args.
|
|
addl MACRO_LITERAL(4 * 8), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-4 * 8)
|
|
|
|
POP ecx // Restore args except eax
|
|
POP edx
|
|
POP ebx
|
|
POP ebp // Restore callee saves
|
|
POP esi
|
|
// Swap EDI callee save with code pointer.
|
|
xchgl %edi, (%esp)
|
|
testl %eax, %eax // Branch forward if exception pending.
|
|
jz 1f
|
|
// Tail call to intended method.
|
|
ret
|
|
1:
|
|
addl MACRO_LITERAL(4), %esp // Pop code pointer off stack
|
|
CFI_ADJUST_CFA_OFFSET(-4)
|
|
DELIVER_PENDING_EXCEPTION
|
|
END_MACRO
|
|
MACRO2(INVOKE_TRAMPOLINE, c_name, cxx_name)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
INVOKE_TRAMPOLINE_BODY RAW_VAR(cxx_name)
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
INVOKE_TRAMPOLINE art_quick_invoke_interface_trampoline_with_access_check, artInvokeInterfaceTrampolineWithAccessCheck
|
|
|
|
INVOKE_TRAMPOLINE art_quick_invoke_static_trampoline_with_access_check, artInvokeStaticTrampolineWithAccessCheck
|
|
INVOKE_TRAMPOLINE art_quick_invoke_direct_trampoline_with_access_check, artInvokeDirectTrampolineWithAccessCheck
|
|
INVOKE_TRAMPOLINE art_quick_invoke_super_trampoline_with_access_check, artInvokeSuperTrampolineWithAccessCheck
|
|
INVOKE_TRAMPOLINE art_quick_invoke_virtual_trampoline_with_access_check, artInvokeVirtualTrampolineWithAccessCheck
|
|
|
|
/*
|
|
* Helper for quick invocation stub to set up XMM registers.
|
|
* Increments shorty and arg_array and clobbers temp_char.
|
|
* Branches to finished if it encounters the end of the shorty.
|
|
*/
|
|
MACRO5(LOOP_OVER_SHORTY_LOADING_XMMS, xmm_reg, shorty, arg_array, temp_char, finished)
|
|
1: // LOOP
|
|
movb (REG_VAR(shorty)), REG_VAR(temp_char) // temp_char := *shorty
|
|
addl MACRO_LITERAL(1), REG_VAR(shorty) // shorty++
|
|
cmpb MACRO_LITERAL(0), REG_VAR(temp_char) // if (temp_char == '\0')
|
|
je VAR(finished) // goto finished
|
|
cmpb MACRO_LITERAL(68), REG_VAR(temp_char) // if (temp_char == 'D')
|
|
je 2f // goto FOUND_DOUBLE
|
|
cmpb MACRO_LITERAL(70), REG_VAR(temp_char) // if (temp_char == 'F')
|
|
je 3f // goto FOUND_FLOAT
|
|
addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++
|
|
// Handle extra space in arg array taken by a long.
|
|
cmpb MACRO_LITERAL(74), REG_VAR(temp_char) // if (temp_char != 'J')
|
|
jne 1b // goto LOOP
|
|
addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++
|
|
jmp 1b // goto LOOP
|
|
2: // FOUND_DOUBLE
|
|
movsd (REG_VAR(arg_array)), REG_VAR(xmm_reg)
|
|
addl MACRO_LITERAL(8), REG_VAR(arg_array) // arg_array+=2
|
|
jmp 4f
|
|
3: // FOUND_FLOAT
|
|
movss (REG_VAR(arg_array)), REG_VAR(xmm_reg)
|
|
addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++
|
|
4:
|
|
END_MACRO
|
|
|
|
/*
|
|
* Helper for quick invocation stub to set up GPR registers.
|
|
* Increments shorty and arg_array, and returns the current short character in
|
|
* temp_char. Branches to finished if it encounters the end of the shorty.
|
|
*/
|
|
MACRO4(SKIP_OVER_FLOATS, shorty, arg_array, temp_char, finished)
|
|
1: // LOOP:
|
|
movb (REG_VAR(shorty)), REG_VAR(temp_char) // temp_char := *shorty
|
|
addl MACRO_LITERAL(1), REG_VAR(shorty) // shorty++
|
|
cmpb MACRO_LITERAL(0), REG_VAR(temp_char) // if (temp_char == '\0')
|
|
je VAR(finished) // goto finished
|
|
cmpb MACRO_LITERAL(70), REG_VAR(temp_char) // if (temp_char == 'F')
|
|
je 3f // goto SKIP_FLOAT
|
|
cmpb MACRO_LITERAL(68), REG_VAR(temp_char) // if (temp_char == 'D')
|
|
je 4f // goto SKIP_DOUBLE
|
|
jmp 5f // goto end
|
|
3: // SKIP_FLOAT
|
|
addl MACRO_LITERAL(4), REG_VAR(arg_array) // arg_array++
|
|
jmp 1b // goto LOOP
|
|
4: // SKIP_DOUBLE
|
|
addl MACRO_LITERAL(8), REG_VAR(arg_array) // arg_array+=2
|
|
jmp 1b // goto LOOP
|
|
5:
|
|
END_MACRO
|
|
|
|
/*
|
|
* Quick invocation stub (non-static).
|
|
* On entry:
|
|
* [sp] = return address
|
|
* [sp + 4] = method pointer
|
|
* [sp + 8] = argument array or null for no argument methods
|
|
* [sp + 12] = size of argument array in bytes
|
|
* [sp + 16] = (managed) thread pointer
|
|
* [sp + 20] = JValue* result
|
|
* [sp + 24] = shorty
|
|
*/
|
|
DEFINE_FUNCTION art_quick_invoke_stub
|
|
// Save the non-volatiles.
|
|
PUSH ebp // save ebp
|
|
PUSH ebx // save ebx
|
|
PUSH esi // save esi
|
|
PUSH edi // save edi
|
|
// Set up argument XMM registers.
|
|
mov 24+16(%esp), %esi // ESI := shorty + 1 ; ie skip return arg character.
|
|
addl LITERAL(1), %esi
|
|
mov 8+16(%esp), %edi // EDI := arg_array + 4 ; ie skip this pointer.
|
|
addl LITERAL(4), %edi
|
|
// Clobbers ESI, EDI, EAX.
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm0, esi, edi, al, .Lxmm_setup_finished
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm1, esi, edi, al, .Lxmm_setup_finished
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm2, esi, edi, al, .Lxmm_setup_finished
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm3, esi, edi, al, .Lxmm_setup_finished
|
|
.balign 16
|
|
.Lxmm_setup_finished:
|
|
mov %esp, %ebp // copy value of stack pointer into base pointer
|
|
CFI_DEF_CFA_REGISTER(ebp)
|
|
mov 28(%ebp), %ebx // get arg array size
|
|
// reserve space for return addr, method*, ebx, ebp, esi, and edi in frame
|
|
addl LITERAL(36), %ebx
|
|
// align frame size to 16 bytes
|
|
andl LITERAL(0xFFFFFFF0), %ebx
|
|
subl LITERAL(20), %ebx // remove space for return address, ebx, ebp, esi and edi
|
|
subl %ebx, %esp // reserve stack space for argument array
|
|
|
|
movl LITERAL(0), (%esp) // store null for method*
|
|
|
|
// Copy arg array into stack.
|
|
movl 28(%ebp), %ecx // ECX = size of args
|
|
movl 24(%ebp), %esi // ESI = argument array
|
|
leal 4(%esp), %edi // EDI = just after Method* in stack arguments
|
|
rep movsb // while (ecx--) { *edi++ = *esi++ }
|
|
|
|
mov 40(%ebp), %esi // ESI := shorty + 1 ; ie skip return arg character.
|
|
addl LITERAL(1), %esi
|
|
mov 24(%ebp), %edi // EDI := arg_array
|
|
mov 0(%edi), %ecx // ECX := this pointer
|
|
addl LITERAL(4), %edi // EDI := arg_array + 4 ; ie skip this pointer.
|
|
|
|
// Enumerate the possible cases for loading GPRS.
|
|
// edx (and maybe ebx):
|
|
SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished
|
|
cmpb LITERAL(74), %al // if (al == 'J') goto FOUND_LONG
|
|
je .LfirstLong
|
|
// Must be an integer value.
|
|
movl (%edi), %edx
|
|
addl LITERAL(4), %edi // arg_array++
|
|
|
|
// Now check ebx
|
|
SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished
|
|
// Must be first word of a long, or an integer. First word of long doesn't
|
|
// go into EBX, but can be loaded there anyways, as it is harmless.
|
|
movl (%edi), %ebx
|
|
jmp .Lgpr_setup_finished
|
|
.LfirstLong:
|
|
movl (%edi), %edx
|
|
movl 4(%edi), %ebx
|
|
// Nothing left to load.
|
|
.Lgpr_setup_finished:
|
|
mov 20(%ebp), %eax // move method pointer into eax
|
|
call *ART_METHOD_QUICK_CODE_OFFSET_32(%eax) // call the method
|
|
mov %ebp, %esp // restore stack pointer
|
|
CFI_DEF_CFA_REGISTER(esp)
|
|
POP edi // pop edi
|
|
POP esi // pop esi
|
|
POP ebx // pop ebx
|
|
POP ebp // pop ebp
|
|
mov 20(%esp), %ecx // get result pointer
|
|
mov %eax, (%ecx) // store the result assuming its a long, int or Object*
|
|
mov %edx, 4(%ecx) // store the other half of the result
|
|
mov 24(%esp), %edx // get the shorty
|
|
cmpb LITERAL(68), (%edx) // test if result type char == 'D'
|
|
je .Lreturn_double_quick
|
|
cmpb LITERAL(70), (%edx) // test if result type char == 'F'
|
|
je .Lreturn_float_quick
|
|
ret
|
|
.Lreturn_double_quick:
|
|
movsd %xmm0, (%ecx) // store the floating point result
|
|
ret
|
|
.Lreturn_float_quick:
|
|
movss %xmm0, (%ecx) // store the floating point result
|
|
ret
|
|
END_FUNCTION art_quick_invoke_stub
|
|
|
|
/*
|
|
* Quick invocation stub (static).
|
|
* On entry:
|
|
* [sp] = return address
|
|
* [sp + 4] = method pointer
|
|
* [sp + 8] = argument array or null for no argument methods
|
|
* [sp + 12] = size of argument array in bytes
|
|
* [sp + 16] = (managed) thread pointer
|
|
* [sp + 20] = JValue* result
|
|
* [sp + 24] = shorty
|
|
*/
|
|
DEFINE_FUNCTION art_quick_invoke_static_stub
|
|
// Save the non-volatiles.
|
|
PUSH ebp // save ebp
|
|
PUSH ebx // save ebx
|
|
PUSH esi // save esi
|
|
PUSH edi // save edi
|
|
// Set up argument XMM registers.
|
|
mov 24+16(%esp), %esi // ESI := shorty + 1 ; ie skip return arg character.
|
|
addl LITERAL(1), %esi
|
|
mov 8+16(%esp), %edi // EDI := arg_array
|
|
// Clobbers ESI, EDI, EAX.
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm0, esi, edi, al, .Lxmm_setup_finished2
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm1, esi, edi, al, .Lxmm_setup_finished2
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm2, esi, edi, al, .Lxmm_setup_finished2
|
|
LOOP_OVER_SHORTY_LOADING_XMMS xmm3, esi, edi, al, .Lxmm_setup_finished2
|
|
.balign 16
|
|
.Lxmm_setup_finished2:
|
|
mov %esp, %ebp // copy value of stack pointer into base pointer
|
|
CFI_DEF_CFA_REGISTER(ebp)
|
|
mov 28(%ebp), %ebx // get arg array size
|
|
// reserve space for return addr, method*, ebx, ebp, esi, and edi in frame
|
|
addl LITERAL(36), %ebx
|
|
// align frame size to 16 bytes
|
|
andl LITERAL(0xFFFFFFF0), %ebx
|
|
subl LITERAL(20), %ebx // remove space for return address, ebx, ebp, esi and edi
|
|
subl %ebx, %esp // reserve stack space for argument array
|
|
|
|
movl LITERAL(0), (%esp) // store null for method*
|
|
|
|
// Copy arg array into stack.
|
|
movl 28(%ebp), %ecx // ECX = size of args
|
|
movl 24(%ebp), %esi // ESI = argument array
|
|
leal 4(%esp), %edi // EDI = just after Method* in stack arguments
|
|
rep movsb // while (ecx--) { *edi++ = *esi++ }
|
|
|
|
mov 40(%ebp), %esi // ESI := shorty + 1 ; ie skip return arg character.
|
|
addl LITERAL(1), %esi
|
|
mov 24(%ebp), %edi // EDI := arg_array
|
|
|
|
// Enumerate the possible cases for loading GPRS.
|
|
// ecx (and maybe edx)
|
|
SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2
|
|
cmpb LITERAL(74), %al // if (al == 'J') goto FOUND_LONG
|
|
je .LfirstLong2
|
|
// Must be an integer value. Load into ECX.
|
|
movl (%edi), %ecx
|
|
addl LITERAL(4), %edi // arg_array++
|
|
|
|
// Now check edx (and maybe ebx).
|
|
SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2
|
|
cmpb LITERAL(74), %al // if (al == 'J') goto FOUND_LONG
|
|
je .LSecondLong2
|
|
// Must be an integer. Load into EDX.
|
|
movl (%edi), %edx
|
|
addl LITERAL(4), %edi // arg_array++
|
|
|
|
// Is there anything for ebx?
|
|
SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2
|
|
// Must be first word of a long, or an integer. First word of long doesn't
|
|
// go into EBX, but can be loaded there anyways, as it is harmless.
|
|
movl (%edi), %ebx
|
|
jmp .Lgpr_setup_finished2
|
|
.LSecondLong2:
|
|
// EDX:EBX is long. That is all.
|
|
movl (%edi), %edx
|
|
movl 4(%edi), %ebx
|
|
jmp .Lgpr_setup_finished2
|
|
.LfirstLong2:
|
|
// ECX:EDX is a long
|
|
movl (%edi), %ecx
|
|
movl 4(%edi), %edx
|
|
addl LITERAL(8), %edi // arg_array += 2
|
|
|
|
// Anything for EBX?
|
|
SKIP_OVER_FLOATS esi, edi, al, .Lgpr_setup_finished2
|
|
// Must be first word of a long, or an integer. First word of long doesn't
|
|
// go into EBX, but can be loaded there anyways, as it is harmless.
|
|
movl (%edi), %ebx
|
|
jmp .Lgpr_setup_finished2
|
|
// Nothing left to load.
|
|
.Lgpr_setup_finished2:
|
|
mov 20(%ebp), %eax // move method pointer into eax
|
|
call *ART_METHOD_QUICK_CODE_OFFSET_32(%eax) // call the method
|
|
mov %ebp, %esp // restore stack pointer
|
|
CFI_DEF_CFA_REGISTER(esp)
|
|
POP edi // pop edi
|
|
POP esi // pop esi
|
|
POP ebx // pop ebx
|
|
POP ebp // pop ebp
|
|
mov 20(%esp), %ecx // get result pointer
|
|
mov %eax, (%ecx) // store the result assuming its a long, int or Object*
|
|
mov %edx, 4(%ecx) // store the other half of the result
|
|
mov 24(%esp), %edx // get the shorty
|
|
cmpb LITERAL(68), (%edx) // test if result type char == 'D'
|
|
je .Lreturn_double_quick2
|
|
cmpb LITERAL(70), (%edx) // test if result type char == 'F'
|
|
je .Lreturn_float_quick2
|
|
ret
|
|
.Lreturn_double_quick2:
|
|
movsd %xmm0, (%ecx) // store the floating point result
|
|
ret
|
|
.Lreturn_float_quick2:
|
|
movss %xmm0, (%ecx) // store the floating point result
|
|
ret
|
|
END_FUNCTION art_quick_invoke_static_stub
|
|
|
|
MACRO3(ONE_ARG_DOWNCALL, c_name, cxx_name, return_macro)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(8), %esp // push padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
CALL_MACRO(return_macro) // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO3(TWO_ARG_DOWNCALL, c_name, cxx_name, return_macro)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
PUSH eax // push padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, arg2, Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
CALL_MACRO(return_macro) // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO3(THREE_ARG_DOWNCALL, c_name, cxx_name, return_macro)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH edx // pass arg3
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, arg2, arg3, Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
CALL_MACRO(return_macro) // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO3(FOUR_ARG_DOWNCALL, c_name, cxx_name, return_macro)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_REFS_ONLY_FRAME_PRESERVE_TEMP_REG ebx // save ref containing registers for GC
|
|
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(12), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ebx // pass arg4
|
|
PUSH edx // pass arg3
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, arg2, arg3, arg4, Thread*)
|
|
addl MACRO_LITERAL(32), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-32)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
CALL_MACRO(return_macro) // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO3(ONE_ARG_REF_DOWNCALL, c_name, cxx_name, return_macro)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(8), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
CALL_MACRO(return_macro) // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO3(TWO_ARG_REF_DOWNCALL, c_name, cxx_name, return_macro)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
PUSH eax // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, arg2, referrer, Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
CALL_MACRO(return_macro) // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO3(THREE_ARG_REF_DOWNCALL, c_name, cxx_name, return_macro)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH edx // pass arg3
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, arg2, arg3, Thread*)
|
|
addl LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-32)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
CALL_MACRO(return_macro) // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
/*
|
|
* Macro for resolution and initialization of indexed DEX file
|
|
* constants such as classes and strings.
|
|
*/
|
|
MACRO3(ONE_ARG_SAVE_EVERYTHING_DOWNCALL, c_name, cxx_name, runtime_method_offset = RUNTIME_SAVE_EVERYTHING_METHOD_OFFSET)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx, \runtime_method_offset // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(8), %esp // push padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax // pass the index of the constant as arg1
|
|
call CALLVAR(cxx_name) // cxx_name(arg1, Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
testl %eax, %eax // If result is null, deliver the OOME.
|
|
jz 1f
|
|
CFI_REMEMBER_STATE
|
|
RESTORE_SAVE_EVERYTHING_FRAME_KEEP_EAX // restore frame up to return address
|
|
ret // return
|
|
CFI_RESTORE_STATE_AND_DEF_CFA esp, FRAME_SIZE_SAVE_EVERYTHING
|
|
1:
|
|
DELIVER_PENDING_EXCEPTION_FRAME_READY
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
MACRO2(ONE_ARG_SAVE_EVERYTHING_DOWNCALL_FOR_CLINIT, c_name, cxx_name)
|
|
ONE_ARG_SAVE_EVERYTHING_DOWNCALL \c_name, \cxx_name, RUNTIME_SAVE_EVERYTHING_FOR_CLINIT_METHOD_OFFSET
|
|
END_MACRO
|
|
|
|
MACRO0(RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER)
|
|
testl %eax, %eax // eax == 0 ?
|
|
jz 1f // if eax == 0 goto 1
|
|
ret // return
|
|
1: // deliver exception on current thread
|
|
DELIVER_PENDING_EXCEPTION
|
|
END_MACRO
|
|
|
|
MACRO0(RETURN_IF_EAX_ZERO)
|
|
testl %eax, %eax // eax == 0 ?
|
|
jnz 1f // if eax != 0 goto 1
|
|
ret // return
|
|
1: // deliver exception on current thread
|
|
DELIVER_PENDING_EXCEPTION
|
|
END_MACRO
|
|
|
|
// Generate the allocation entrypoints for each allocator.
|
|
GENERATE_ALLOC_ENTRYPOINTS_FOR_NON_TLAB_ALLOCATORS
|
|
|
|
// Comment out allocators that have x86 specific asm.
|
|
// Region TLAB:
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_RESOLVED(_region_tlab, RegionTLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_INITIALIZED(_region_tlab, RegionTLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_WITH_ACCESS_CHECK(_region_tlab, RegionTLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_OBJECT(_region_tlab, RegionTLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED(_region_tlab, RegionTLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED8(_region_tlab, RegionTLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED16(_region_tlab, RegionTLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED32(_region_tlab, RegionTLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED64(_region_tlab, RegionTLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_BYTES(_region_tlab, RegionTLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_CHARS(_region_tlab, RegionTLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_STRING(_region_tlab, RegionTLAB)
|
|
// Normal TLAB:
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_RESOLVED(_tlab, TLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_INITIALIZED(_tlab, TLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_WITH_ACCESS_CHECK(_tlab, TLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_OBJECT(_tlab, TLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED(_tlab, TLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED8(_tlab, TLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED16(_tlab, TLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED32(_tlab, TLAB)
|
|
// GENERATE_ALLOC_ENTRYPOINTS_ALLOC_ARRAY_RESOLVED64(_tlab, TLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_BYTES(_tlab, TLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_CHARS(_tlab, TLAB)
|
|
GENERATE_ALLOC_ENTRYPOINTS_ALLOC_STRING_FROM_STRING(_tlab, TLAB)
|
|
|
|
// A hand-written override for GENERATE_ALLOC_ENTRYPOINTS_ALLOC_OBJECT_RESOLVED(_rosalloc, RosAlloc).
|
|
MACRO2(ART_QUICK_ALLOC_OBJECT_ROSALLOC, c_name, cxx_name)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
// Fast path rosalloc allocation.
|
|
// eax: type/return value
|
|
// ecx, ebx, edx: free
|
|
movl %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread
|
|
// Check if the thread local allocation
|
|
// stack has room
|
|
movl THREAD_LOCAL_ALLOC_STACK_TOP_OFFSET(%ebx), %ecx
|
|
cmpl THREAD_LOCAL_ALLOC_STACK_END_OFFSET(%ebx), %ecx
|
|
jae .Lslow_path\c_name
|
|
|
|
movl MIRROR_CLASS_OBJECT_SIZE_ALLOC_FAST_PATH_OFFSET(%eax), %ecx // Load the object size (ecx)
|
|
// Check if the size is for a thread
|
|
// local allocation. Also does the
|
|
// finalizable and initialization check.
|
|
cmpl LITERAL(ROSALLOC_MAX_THREAD_LOCAL_BRACKET_SIZE), %ecx
|
|
ja .Lslow_path\c_name
|
|
shrl LITERAL(ROSALLOC_BRACKET_QUANTUM_SIZE_SHIFT), %ecx // Calculate the rosalloc bracket index
|
|
// from object size.
|
|
// Load thread local rosalloc run (ebx)
|
|
// Subtract __SIZEOF_POINTER__ to subtract
|
|
// one from edi as there is no 0 byte run
|
|
// and the size is already aligned.
|
|
movl (THREAD_ROSALLOC_RUNS_OFFSET - __SIZEOF_POINTER__)(%ebx, %ecx, __SIZEOF_POINTER__), %ebx
|
|
// Load free_list head (edi),
|
|
// this will be the return value.
|
|
movl (ROSALLOC_RUN_FREE_LIST_OFFSET + ROSALLOC_RUN_FREE_LIST_HEAD_OFFSET)(%ebx), %ecx
|
|
jecxz .Lslow_path\c_name
|
|
// Point of no slow path. Won't go to
|
|
// the slow path from here on.
|
|
// Load the next pointer of the head
|
|
// and update head of free list with
|
|
// next pointer
|
|
movl ROSALLOC_SLOT_NEXT_OFFSET(%ecx), %edx
|
|
movl %edx, (ROSALLOC_RUN_FREE_LIST_OFFSET + ROSALLOC_RUN_FREE_LIST_HEAD_OFFSET)(%ebx)
|
|
// Decrement size of free list by 1
|
|
decl (ROSALLOC_RUN_FREE_LIST_OFFSET + ROSALLOC_RUN_FREE_LIST_SIZE_OFFSET)(%ebx)
|
|
// Store the class pointer in the
|
|
// header. This also overwrites the
|
|
// next pointer. The offsets are
|
|
// asserted to match.
|
|
#if ROSALLOC_SLOT_NEXT_OFFSET != MIRROR_OBJECT_CLASS_OFFSET
|
|
#error "Class pointer needs to overwrite next pointer."
|
|
#endif
|
|
POISON_HEAP_REF eax
|
|
movl %eax, MIRROR_OBJECT_CLASS_OFFSET(%ecx)
|
|
movl %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread
|
|
// Push the new object onto the thread
|
|
// local allocation stack and
|
|
// increment the thread local
|
|
// allocation stack top.
|
|
movl THREAD_LOCAL_ALLOC_STACK_TOP_OFFSET(%ebx), %eax
|
|
movl %ecx, (%eax)
|
|
addl LITERAL(COMPRESSED_REFERENCE_SIZE), %eax
|
|
movl %eax, THREAD_LOCAL_ALLOC_STACK_TOP_OFFSET(%ebx)
|
|
// No fence needed for x86.
|
|
movl %ecx, %eax // Move object to return register
|
|
ret
|
|
.Lslow_path\c_name:
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
subl LITERAL(8), %esp // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax
|
|
call SYMBOL(artAllocObjectFromCodeResolvedRosAlloc) // cxx_name(arg0, Thread*)
|
|
addl LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
ART_QUICK_ALLOC_OBJECT_ROSALLOC art_quick_alloc_object_resolved_rosalloc, artAllocObjectFromCodeResolvedRosAlloc
|
|
ART_QUICK_ALLOC_OBJECT_ROSALLOC art_quick_alloc_object_initialized_rosalloc, artAllocObjectFromCodeInitializedRosAlloc
|
|
|
|
// The common fast path code for art_quick_alloc_object_resolved/initialized_tlab
|
|
// and art_quick_alloc_object_resolved/initialized_region_tlab.
|
|
//
|
|
// EAX: type/return_value
|
|
MACRO1(ALLOC_OBJECT_RESOLVED_TLAB_FAST_PATH, slowPathLabel)
|
|
movl %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread
|
|
movl THREAD_LOCAL_END_OFFSET(%ebx), %edi // Load thread_local_end.
|
|
subl THREAD_LOCAL_POS_OFFSET(%ebx), %edi // Compute the remaining buffer size.
|
|
movl MIRROR_CLASS_OBJECT_SIZE_ALLOC_FAST_PATH_OFFSET(%eax), %ecx // Load the object size.
|
|
cmpl %edi, %ecx // Check if it fits.
|
|
ja VAR(slowPathLabel)
|
|
movl THREAD_LOCAL_POS_OFFSET(%ebx), %edx // Load thread_local_pos
|
|
// as allocated object.
|
|
addl %edx, %ecx // Add the object size.
|
|
movl %ecx, THREAD_LOCAL_POS_OFFSET(%ebx) // Update thread_local_pos.
|
|
incl THREAD_LOCAL_OBJECTS_OFFSET(%ebx) // Increase thread_local_objects.
|
|
// Store the class pointer in the header.
|
|
// No fence needed for x86.
|
|
POISON_HEAP_REF eax
|
|
movl %eax, MIRROR_OBJECT_CLASS_OFFSET(%edx)
|
|
movl %edx, %eax
|
|
POP edi
|
|
ret // Fast path succeeded.
|
|
END_MACRO
|
|
|
|
// The common slow path code for art_quick_alloc_object_resolved/initialized_tlab
|
|
// and art_quick_alloc_object_resolved/initialized_region_tlab.
|
|
MACRO1(ALLOC_OBJECT_RESOLVED_TLAB_SLOW_PATH, cxx_name)
|
|
POP edi
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
subl LITERAL(8), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax
|
|
call CALLVAR(cxx_name) // cxx_name(arg0, Thread*)
|
|
addl LITERAL(16), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception
|
|
END_MACRO
|
|
|
|
MACRO2(ART_QUICK_ALLOC_OBJECT_TLAB, c_name, cxx_name)
|
|
DEFINE_FUNCTION VAR(c_name)
|
|
// Fast path tlab allocation.
|
|
// EAX: type
|
|
// EBX, ECX, EDX: free.
|
|
PUSH edi
|
|
ALLOC_OBJECT_RESOLVED_TLAB_FAST_PATH .Lslow_path\c_name
|
|
.Lslow_path\c_name:
|
|
ALLOC_OBJECT_RESOLVED_TLAB_SLOW_PATH RAW_VAR(cxx_name)
|
|
END_FUNCTION VAR(c_name)
|
|
END_MACRO
|
|
|
|
ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_resolved_tlab, artAllocObjectFromCodeResolvedTLAB
|
|
ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_initialized_tlab, artAllocObjectFromCodeInitializedTLAB
|
|
ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_resolved_region_tlab, artAllocObjectFromCodeResolvedRegionTLAB
|
|
ART_QUICK_ALLOC_OBJECT_TLAB art_quick_alloc_object_initialized_region_tlab, artAllocObjectFromCodeInitializedRegionTLAB
|
|
|
|
// The fast path code for art_quick_alloc_array_region_tlab.
|
|
// Inputs: EAX: the class, ECX: int32_t component_count, EDX: total_size
|
|
// Free temp: EBX
|
|
// Output: EAX: return value.
|
|
MACRO1(ALLOC_ARRAY_TLAB_FAST_PATH_RESOLVED_WITH_SIZE, slowPathLabel)
|
|
mov %fs:THREAD_SELF_OFFSET, %ebx // ebx = thread
|
|
// Mask out the unaligned part to make sure we are 8 byte aligned.
|
|
andl LITERAL(OBJECT_ALIGNMENT_MASK_TOGGLED), %edx
|
|
movl THREAD_LOCAL_END_OFFSET(%ebx), %edi
|
|
subl THREAD_LOCAL_POS_OFFSET(%ebx), %edi
|
|
cmpl %edi, %edx // Check if it fits.
|
|
ja RAW_VAR(slowPathLabel)
|
|
movl THREAD_LOCAL_POS_OFFSET(%ebx), %edi
|
|
addl %edi, %edx // Add the object size.
|
|
movl %edx, THREAD_LOCAL_POS_OFFSET(%ebx) // Update thread_local_pos_
|
|
addl LITERAL(1), THREAD_LOCAL_OBJECTS_OFFSET(%ebx) // Increase thread_local_objects.
|
|
// Store the class pointer in the
|
|
// header.
|
|
// No fence needed for x86.
|
|
POISON_HEAP_REF eax
|
|
movl %eax, MIRROR_OBJECT_CLASS_OFFSET(%edi)
|
|
movl %ecx, MIRROR_ARRAY_LENGTH_OFFSET(%edi)
|
|
movl %edi, %eax
|
|
POP edi
|
|
ret // Fast path succeeded.
|
|
END_MACRO
|
|
|
|
MACRO1(COMPUTE_ARRAY_SIZE_UNKNOWN, slow_path)
|
|
// Possibly a large object, go slow.
|
|
// Also does negative array size check.
|
|
cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_WIDE_ARRAY_DATA_OFFSET) / 8), %ecx
|
|
ja RAW_VAR(slow_path)
|
|
PUSH ecx
|
|
movl %ecx, %edx
|
|
movl MIRROR_CLASS_COMPONENT_TYPE_OFFSET(%eax), %ecx // Load component type.
|
|
UNPOISON_HEAP_REF ecx
|
|
movl MIRROR_CLASS_OBJECT_PRIMITIVE_TYPE_OFFSET(%ecx), %ecx // Load primitive type.
|
|
shr MACRO_LITERAL(PRIMITIVE_TYPE_SIZE_SHIFT_SHIFT), %ecx // Get component size shift.
|
|
sall %cl, %edx // Calculate array count shifted.
|
|
// Add array header + alignment rounding.
|
|
add MACRO_LITERAL(MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK), %edx
|
|
// Add 4 extra bytes if we are doing a long array.
|
|
add MACRO_LITERAL(1), %ecx
|
|
and MACRO_LITERAL(4), %ecx
|
|
#if MIRROR_WIDE_ARRAY_DATA_OFFSET != MIRROR_INT_ARRAY_DATA_OFFSET + 4
|
|
#error Long array data offset must be 4 greater than int array data offset.
|
|
#endif
|
|
addl %ecx, %edx
|
|
POP ecx
|
|
END_MACRO
|
|
|
|
MACRO1(COMPUTE_ARRAY_SIZE_8, slow_path)
|
|
// EAX: mirror::Class* klass, ECX: int32_t component_count
|
|
// Possibly a large object, go slow.
|
|
// Also does negative array size check.
|
|
cmpl LITERAL(MIN_LARGE_OBJECT_THRESHOLD - MIRROR_INT_ARRAY_DATA_OFFSET), %ecx
|
|
ja RAW_VAR(slow_path)
|
|
// Add array header + alignment rounding.
|
|
leal (MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK)(%ecx), %edx
|
|
END_MACRO
|
|
|
|
MACRO1(COMPUTE_ARRAY_SIZE_16, slow_path)
|
|
// EAX: mirror::Class* klass, ECX: int32_t component_count
|
|
// Possibly a large object, go slow.
|
|
// Also does negative array size check.
|
|
cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_INT_ARRAY_DATA_OFFSET) / 2), %ecx
|
|
ja RAW_VAR(slow_path)
|
|
// Add array header + alignment rounding.
|
|
leal ((MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK) / 2)(%ecx), %edx
|
|
sall MACRO_LITERAL(1), %edx
|
|
END_MACRO
|
|
|
|
MACRO1(COMPUTE_ARRAY_SIZE_32, slow_path)
|
|
// EAX: mirror::Class* klass, ECX: int32_t component_count
|
|
// Possibly a large object, go slow.
|
|
// Also does negative array size check.
|
|
cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_INT_ARRAY_DATA_OFFSET) / 4), %ecx
|
|
ja RAW_VAR(slow_path)
|
|
// Add array header + alignment rounding.
|
|
leal ((MIRROR_INT_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK) / 4)(%ecx), %edx
|
|
sall MACRO_LITERAL(2), %edx
|
|
END_MACRO
|
|
|
|
MACRO1(COMPUTE_ARRAY_SIZE_64, slow_path)
|
|
// EAX: mirror::Class* klass, ECX: int32_t component_count
|
|
// Possibly a large object, go slow.
|
|
// Also does negative array size check.
|
|
cmpl LITERAL((MIN_LARGE_OBJECT_THRESHOLD - MIRROR_WIDE_ARRAY_DATA_OFFSET) / 8), %ecx
|
|
ja RAW_VAR(slow_path)
|
|
// Add array header + alignment rounding.
|
|
leal ((MIRROR_WIDE_ARRAY_DATA_OFFSET + OBJECT_ALIGNMENT_MASK) / 8)(%ecx), %edx
|
|
sall MACRO_LITERAL(3), %edx
|
|
END_MACRO
|
|
|
|
MACRO3(GENERATE_ALLOC_ARRAY_TLAB, c_entrypoint, cxx_name, size_setup)
|
|
DEFINE_FUNCTION VAR(c_entrypoint)
|
|
// EAX: mirror::Class* klass, ECX: int32_t component_count
|
|
PUSH edi
|
|
CALL_MACRO(size_setup) .Lslow_path\c_entrypoint
|
|
ALLOC_ARRAY_TLAB_FAST_PATH_RESOLVED_WITH_SIZE .Lslow_path\c_entrypoint
|
|
.Lslow_path\c_entrypoint:
|
|
POP edi
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
PUSH eax // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx
|
|
PUSH eax
|
|
call CALLVAR(cxx_name) // cxx_name(arg0, arg1, Thread*)
|
|
addl LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception
|
|
END_FUNCTION VAR(c_entrypoint)
|
|
END_MACRO
|
|
|
|
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_UNKNOWN
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved8_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_8
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved16_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_16
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved32_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_32
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved64_region_tlab, artAllocArrayFromCodeResolvedRegionTLAB, COMPUTE_ARRAY_SIZE_64
|
|
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_UNKNOWN
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved8_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_8
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved16_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_16
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved32_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_32
|
|
GENERATE_ALLOC_ARRAY_TLAB art_quick_alloc_array_resolved64_tlab, artAllocArrayFromCodeResolvedTLAB, COMPUTE_ARRAY_SIZE_64
|
|
|
|
ONE_ARG_SAVE_EVERYTHING_DOWNCALL_FOR_CLINIT art_quick_initialize_static_storage, artInitializeStaticStorageFromCode
|
|
ONE_ARG_SAVE_EVERYTHING_DOWNCALL_FOR_CLINIT art_quick_resolve_type, artResolveTypeFromCode
|
|
ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_type_and_verify_access, artResolveTypeAndVerifyAccessFromCode
|
|
ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_method_handle, artResolveMethodHandleFromCode
|
|
ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_method_type, artResolveMethodTypeFromCode
|
|
ONE_ARG_SAVE_EVERYTHING_DOWNCALL art_quick_resolve_string, artResolveStringFromCode
|
|
|
|
TWO_ARG_REF_DOWNCALL art_quick_handle_fill_data, artHandleFillArrayDataFromCode, RETURN_IF_EAX_ZERO
|
|
|
|
/*
|
|
* Entry from managed code that tries to lock the object in a fast path and
|
|
* calls `artLockObjectFromCode()` for the difficult cases, may block for GC.
|
|
* EAX holds the possibly null object to lock.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_lock_object
|
|
testl %eax, %eax
|
|
jz SYMBOL(art_quick_lock_object_no_inline)
|
|
movl %eax, %ecx // Move obj to a different register.
|
|
LOCK_OBJECT_FAST_PATH ecx, edx, /*saved_eax*/ none, .Llock_object_slow
|
|
.Llock_object_slow:
|
|
movl %ecx, %eax // Move obj back to EAX.
|
|
jmp SYMBOL(art_quick_lock_object_no_inline)
|
|
END_FUNCTION art_quick_lock_object
|
|
|
|
/*
|
|
* Entry from managed code that calls `artLockObjectFromCode()`, may block for GC.
|
|
* EAX holds the possibly null object to lock.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_lock_object_no_inline
|
|
// This is also the slow path for art_quick_lock_object.
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
INCREASE_FRAME 8 // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH_ARG eax // pass object
|
|
call SYMBOL(artLockObjectFromCode) // artLockObjectFromCode(object, Thread*)
|
|
DECREASE_FRAME 16 // pop arguments
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
RETURN_IF_EAX_ZERO
|
|
END_FUNCTION art_quick_lock_object_no_inline
|
|
|
|
/*
|
|
* Entry from managed code that tries to unlock the object in a fast path and calls
|
|
* `artUnlockObjectFromCode()` for the difficult cases and delivers exception on failure.
|
|
* EAX holds the possibly null object to unlock.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_unlock_object
|
|
testl %eax, %eax
|
|
jz SYMBOL(art_quick_unlock_object_no_inline)
|
|
movl %eax, %ecx // Move obj to a different register.
|
|
UNLOCK_OBJECT_FAST_PATH ecx, edx, /*saved_eax*/ none, .Lunlock_object_slow
|
|
.Lunlock_object_slow:
|
|
movl %ecx, %eax // Move obj back to EAX.
|
|
jmp SYMBOL(art_quick_unlock_object_no_inline)
|
|
END_FUNCTION art_quick_unlock_object
|
|
|
|
/*
|
|
* Entry from managed code that calls `artUnlockObjectFromCode()`
|
|
* and delivers exception on failure.
|
|
* EAX holds the possibly null object to unlock.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_unlock_object_no_inline
|
|
// This is also the slow path for art_quick_unlock_object.
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
INCREASE_FRAME 8 // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH_ARG eax // pass object
|
|
call SYMBOL(artUnlockObjectFromCode) // artUnlockObjectFromCode(object, Thread*)
|
|
DECREASE_FRAME 16 // pop arguments
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
RETURN_IF_EAX_ZERO
|
|
END_FUNCTION art_quick_unlock_object_no_inline
|
|
|
|
DEFINE_FUNCTION art_quick_instance_of
|
|
PUSH eax // alignment padding
|
|
PUSH ecx // pass arg2 - obj->klass
|
|
PUSH eax // pass arg1 - checked class
|
|
call SYMBOL(artInstanceOfFromCode) // (Object* obj, Class* ref_klass)
|
|
addl LITERAL(12), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
ret
|
|
END_FUNCTION art_quick_instance_of
|
|
|
|
DEFINE_FUNCTION art_quick_check_instance_of
|
|
// Type check using the bit string passes null as the target class. In that case just throw.
|
|
testl %ecx, %ecx
|
|
jz .Lthrow_class_cast_exception_for_bitstring_check
|
|
|
|
PUSH eax // alignment padding
|
|
PUSH ecx // pass arg2 - checked class
|
|
PUSH eax // pass arg1 - obj
|
|
call SYMBOL(artInstanceOfFromCode) // (Object* obj, Class* ref_klass)
|
|
testl %eax, %eax
|
|
jz .Lthrow_class_cast_exception // jump forward if not assignable
|
|
addl LITERAL(12), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
ret
|
|
CFI_ADJUST_CFA_OFFSET(12) // Reset unwind info so following code unwinds.
|
|
|
|
.Lthrow_class_cast_exception:
|
|
POP eax // pop arguments
|
|
POP ecx
|
|
addl LITERAL(4), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-4)
|
|
|
|
.Lthrow_class_cast_exception_for_bitstring_check:
|
|
SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx // save all registers as basis for long jump context
|
|
// Outgoing argument set up
|
|
PUSH eax // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
call SYMBOL(artThrowClassCastExceptionForObject) // (Object* src, Class* dest, Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION art_quick_check_instance_of
|
|
|
|
// Restore reg's value if reg is not the same as exclude_reg, otherwise just adjust stack.
|
|
MACRO2(POP_REG_NE, reg, exclude_reg)
|
|
.ifc RAW_VAR(reg), RAW_VAR(exclude_reg)
|
|
addl MACRO_LITERAL(4), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-4)
|
|
.else
|
|
POP RAW_VAR(reg)
|
|
.endif
|
|
END_MACRO
|
|
|
|
/*
|
|
* Macro to insert read barrier, only used in art_quick_aput_obj.
|
|
* obj_reg and dest_reg are registers, offset is a defined literal such as
|
|
* MIRROR_OBJECT_CLASS_OFFSET.
|
|
* pop_eax is a boolean flag, indicating if eax is popped after the call.
|
|
* TODO: When read barrier has a fast path, add heap unpoisoning support for the fast path.
|
|
*/
|
|
MACRO4(READ_BARRIER, obj_reg, offset, dest_reg, pop_eax)
|
|
#ifdef USE_READ_BARRIER
|
|
PUSH eax // save registers used in art_quick_aput_obj
|
|
PUSH ebx
|
|
PUSH edx
|
|
PUSH ecx
|
|
// Outgoing argument set up
|
|
pushl MACRO_LITERAL((RAW_VAR(offset))) // pass offset, double parentheses are necessary
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH RAW_VAR(obj_reg) // pass obj_reg
|
|
PUSH eax // pass ref, just pass eax for now since parameter ref is unused
|
|
call SYMBOL(artReadBarrierSlow) // artReadBarrierSlow(ref, obj_reg, offset)
|
|
// No need to unpoison return value in eax, artReadBarrierSlow() would do the unpoisoning.
|
|
.ifnc RAW_VAR(dest_reg), eax
|
|
movl %eax, REG_VAR(dest_reg) // save loaded ref in dest_reg
|
|
.endif
|
|
addl MACRO_LITERAL(12), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
POP_REG_NE ecx, RAW_VAR(dest_reg) // Restore args except dest_reg
|
|
POP_REG_NE edx, RAW_VAR(dest_reg)
|
|
POP_REG_NE ebx, RAW_VAR(dest_reg)
|
|
.ifc RAW_VAR(pop_eax), true
|
|
POP_REG_NE eax, RAW_VAR(dest_reg)
|
|
.endif
|
|
#else
|
|
movl RAW_VAR(offset)(REG_VAR(obj_reg)), REG_VAR(dest_reg)
|
|
UNPOISON_HEAP_REF RAW_VAR(dest_reg)
|
|
#endif // USE_READ_BARRIER
|
|
END_MACRO
|
|
|
|
DEFINE_FUNCTION art_quick_aput_obj
|
|
test %edx, %edx // store of null
|
|
jz .Ldo_aput_null
|
|
READ_BARRIER eax, MIRROR_OBJECT_CLASS_OFFSET, ebx, true
|
|
READ_BARRIER ebx, MIRROR_CLASS_COMPONENT_TYPE_OFFSET, ebx, true
|
|
// value's type == array's component type - trivial assignability
|
|
#if defined(USE_READ_BARRIER)
|
|
READ_BARRIER edx, MIRROR_OBJECT_CLASS_OFFSET, eax, false
|
|
cmpl %eax, %ebx
|
|
POP eax // restore eax from the push in the beginning of READ_BARRIER macro
|
|
// This asymmetric push/pop saves a push of eax and maintains stack alignment.
|
|
#elif defined(USE_HEAP_POISONING)
|
|
PUSH eax // save eax
|
|
movl MIRROR_OBJECT_CLASS_OFFSET(%edx), %eax
|
|
UNPOISON_HEAP_REF eax
|
|
cmpl %eax, %ebx
|
|
POP eax // restore eax
|
|
#else
|
|
cmpl MIRROR_OBJECT_CLASS_OFFSET(%edx), %ebx
|
|
#endif
|
|
jne .Lcheck_assignability
|
|
.Ldo_aput:
|
|
POISON_HEAP_REF edx
|
|
movl %edx, MIRROR_OBJECT_ARRAY_DATA_OFFSET(%eax, %ecx, 4)
|
|
movl %fs:THREAD_CARD_TABLE_OFFSET, %edx
|
|
shrl LITERAL(CARD_TABLE_CARD_SHIFT), %eax
|
|
movb %dl, (%edx, %eax)
|
|
ret
|
|
.Ldo_aput_null:
|
|
movl %edx, MIRROR_OBJECT_ARRAY_DATA_OFFSET(%eax, %ecx, 4)
|
|
ret
|
|
.Lcheck_assignability:
|
|
PUSH eax // save arguments
|
|
PUSH ecx
|
|
PUSH edx
|
|
#if defined(USE_READ_BARRIER)
|
|
subl LITERAL(4), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
READ_BARRIER edx, MIRROR_OBJECT_CLASS_OFFSET, eax, true
|
|
subl LITERAL(4), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax // pass arg2 - type of the value to be stored
|
|
#elif defined(USE_HEAP_POISONING)
|
|
subl LITERAL(8), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
movl MIRROR_OBJECT_CLASS_OFFSET(%edx), %eax
|
|
UNPOISON_HEAP_REF eax
|
|
PUSH eax // pass arg2 - type of the value to be stored
|
|
#else
|
|
subl LITERAL(8), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl MIRROR_OBJECT_CLASS_OFFSET(%edx) // pass arg2 - type of the value to be stored
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
#endif
|
|
PUSH ebx // pass arg1 - component type of the array
|
|
call SYMBOL(artIsAssignableFromCode) // (Class* a, Class* b)
|
|
addl LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
testl %eax, %eax
|
|
jz .Lthrow_array_store_exception
|
|
POP edx
|
|
POP ecx
|
|
POP eax
|
|
POISON_HEAP_REF edx
|
|
movl %edx, MIRROR_OBJECT_ARRAY_DATA_OFFSET(%eax, %ecx, 4) // do the aput
|
|
movl %fs:THREAD_CARD_TABLE_OFFSET, %edx
|
|
shrl LITERAL(CARD_TABLE_CARD_SHIFT), %eax
|
|
movb %dl, (%edx, %eax)
|
|
ret
|
|
CFI_ADJUST_CFA_OFFSET(12) // 3 POP after the jz for unwinding.
|
|
.Lthrow_array_store_exception:
|
|
POP edx
|
|
POP ecx
|
|
POP eax
|
|
SETUP_SAVE_ALL_CALLEE_SAVES_FRAME ebx // save all registers as basis for long jump context
|
|
// Outgoing argument set up
|
|
PUSH eax // alignment padding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH edx // pass arg2 - value
|
|
PUSH eax // pass arg1 - array
|
|
call SYMBOL(artThrowArrayStoreException) // (array, value, Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION art_quick_aput_obj
|
|
|
|
DEFINE_FUNCTION art_quick_memcpy
|
|
PUSH edx // pass arg3
|
|
PUSH ecx // pass arg2
|
|
PUSH eax // pass arg1
|
|
// PLT call requires EBX initialized to the $_GLOBAL_OFFSET_TABLE_.
|
|
SETUP_PC_REL_BASE_0 ebx
|
|
1:
|
|
addl $_GLOBAL_OFFSET_TABLE_ + (1b - 0b), %ebx
|
|
call PLT_SYMBOL(memcpy) // (void*, const void*, size_t)
|
|
addl LITERAL(12), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
ret
|
|
END_FUNCTION art_quick_memcpy
|
|
|
|
DEFINE_FUNCTION art_quick_test_suspend
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx, RUNTIME_SAVE_EVERYTHING_FOR_SUSPEND_CHECK_METHOD_OFFSET // save everything for GC
|
|
// Outgoing argument set up
|
|
subl MACRO_LITERAL(12), %esp // push padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call SYMBOL(artTestSuspendFromCode) // (Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_EVERYTHING_FRAME // restore frame up to return address
|
|
ret // return
|
|
END_FUNCTION art_quick_test_suspend
|
|
|
|
DEFINE_FUNCTION art_quick_d2l
|
|
subl LITERAL(12), %esp // alignment padding, room for argument
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
movsd %xmm0, 0(%esp) // arg a
|
|
call SYMBOL(art_d2l) // (jdouble a)
|
|
addl LITERAL(12), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
ret
|
|
END_FUNCTION art_quick_d2l
|
|
|
|
DEFINE_FUNCTION art_quick_f2l
|
|
subl LITERAL(12), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
movss %xmm0, 0(%esp) // arg a
|
|
call SYMBOL(art_f2l) // (jfloat a)
|
|
addl LITERAL(12), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
ret
|
|
END_FUNCTION art_quick_f2l
|
|
|
|
DEFINE_FUNCTION art_quick_ldiv
|
|
subl LITERAL(12), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
PUSH ebx // pass arg4 b.hi
|
|
PUSH edx // pass arg3 b.lo
|
|
PUSH ecx // pass arg2 a.hi
|
|
PUSH eax // pass arg1 a.lo
|
|
call SYMBOL(artLdiv) // (jlong a, jlong b)
|
|
addl LITERAL(28), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-28)
|
|
ret
|
|
END_FUNCTION art_quick_ldiv
|
|
|
|
DEFINE_FUNCTION art_quick_lmod
|
|
subl LITERAL(12), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
PUSH ebx // pass arg4 b.hi
|
|
PUSH edx // pass arg3 b.lo
|
|
PUSH ecx // pass arg2 a.hi
|
|
PUSH eax // pass arg1 a.lo
|
|
call SYMBOL(artLmod) // (jlong a, jlong b)
|
|
addl LITERAL(28), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-28)
|
|
ret
|
|
END_FUNCTION art_quick_lmod
|
|
|
|
DEFINE_FUNCTION art_quick_lmul
|
|
imul %eax, %ebx // ebx = a.lo(eax) * b.hi(ebx)
|
|
imul %edx, %ecx // ecx = b.lo(edx) * a.hi(ecx)
|
|
mul %edx // edx:eax = a.lo(eax) * b.lo(edx)
|
|
add %ebx, %ecx
|
|
add %ecx, %edx // edx += (a.lo * b.hi) + (b.lo * a.hi)
|
|
ret
|
|
END_FUNCTION art_quick_lmul
|
|
|
|
DEFINE_FUNCTION art_quick_lshl
|
|
// ecx:eax << edx
|
|
xchg %edx, %ecx
|
|
shld %cl,%eax,%edx
|
|
shl %cl,%eax
|
|
test LITERAL(32), %cl
|
|
jz 1f
|
|
mov %eax, %edx
|
|
xor %eax, %eax
|
|
1:
|
|
ret
|
|
END_FUNCTION art_quick_lshl
|
|
|
|
DEFINE_FUNCTION art_quick_lshr
|
|
// ecx:eax >> edx
|
|
xchg %edx, %ecx
|
|
shrd %cl,%edx,%eax
|
|
sar %cl,%edx
|
|
test LITERAL(32),%cl
|
|
jz 1f
|
|
mov %edx, %eax
|
|
sar LITERAL(31), %edx
|
|
1:
|
|
ret
|
|
END_FUNCTION art_quick_lshr
|
|
|
|
DEFINE_FUNCTION art_quick_lushr
|
|
// ecx:eax >>> edx
|
|
xchg %edx, %ecx
|
|
shrd %cl,%edx,%eax
|
|
shr %cl,%edx
|
|
test LITERAL(32),%cl
|
|
jz 1f
|
|
mov %edx, %eax
|
|
xor %edx, %edx
|
|
1:
|
|
ret
|
|
END_FUNCTION art_quick_lushr
|
|
|
|
// Note: Functions `art{Get,Set}<Kind>{Static,Instance}FromCompiledCode` are
|
|
// defined with a macro in runtime/entrypoints/quick/quick_field_entrypoints.cc.
|
|
|
|
ONE_ARG_REF_DOWNCALL art_quick_get_boolean_static, artGetBooleanStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
ONE_ARG_REF_DOWNCALL art_quick_get_byte_static, artGetByteStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
ONE_ARG_REF_DOWNCALL art_quick_get_char_static, artGetCharStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
ONE_ARG_REF_DOWNCALL art_quick_get_short_static, artGetShortStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
ONE_ARG_REF_DOWNCALL art_quick_get32_static, artGet32StaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
ONE_ARG_REF_DOWNCALL art_quick_get64_static, artGet64StaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
ONE_ARG_REF_DOWNCALL art_quick_get_obj_static, artGetObjStaticFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
|
|
TWO_ARG_REF_DOWNCALL art_quick_get_boolean_instance, artGetBooleanInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
TWO_ARG_REF_DOWNCALL art_quick_get_byte_instance, artGetByteInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
TWO_ARG_REF_DOWNCALL art_quick_get_char_instance, artGetCharInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
TWO_ARG_REF_DOWNCALL art_quick_get_short_instance, artGetShortInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
TWO_ARG_REF_DOWNCALL art_quick_get32_instance, artGet32InstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
TWO_ARG_REF_DOWNCALL art_quick_get64_instance, artGet64InstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
TWO_ARG_REF_DOWNCALL art_quick_get_obj_instance, artGetObjInstanceFromCompiledCode, RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
|
|
TWO_ARG_REF_DOWNCALL art_quick_set8_static, artSet8StaticFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
TWO_ARG_REF_DOWNCALL art_quick_set16_static, artSet16StaticFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
TWO_ARG_REF_DOWNCALL art_quick_set32_static, artSet32StaticFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
TWO_ARG_REF_DOWNCALL art_quick_set_obj_static, artSetObjStaticFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
|
|
THREE_ARG_REF_DOWNCALL art_quick_set64_static, artSet64StaticFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
THREE_ARG_REF_DOWNCALL art_quick_set8_instance, artSet8InstanceFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
THREE_ARG_REF_DOWNCALL art_quick_set16_instance, artSet16InstanceFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
THREE_ARG_REF_DOWNCALL art_quick_set32_instance, artSet32InstanceFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
THREE_ARG_REF_DOWNCALL art_quick_set_obj_instance, artSetObjInstanceFromCompiledCode, RETURN_IF_EAX_ZERO
|
|
|
|
// Call artSet64InstanceFromCode with 4 word size arguments.
|
|
DEFINE_FUNCTION art_quick_set64_instance
|
|
movd %ebx, %xmm0
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
movd %xmm0, %ebx
|
|
// Outgoing argument set up
|
|
subl LITERAL(12), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ebx // pass high half of new_val
|
|
PUSH edx // pass low half of new_val
|
|
PUSH ecx // pass object
|
|
PUSH eax // pass field_idx
|
|
call SYMBOL(artSet64InstanceFromCompiledCode) // (field_idx, Object*, new_val, Thread*)
|
|
addl LITERAL(32), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-32)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
RETURN_IF_EAX_ZERO // return or deliver exception
|
|
END_FUNCTION art_quick_set64_instance
|
|
|
|
DEFINE_FUNCTION art_quick_proxy_invoke_handler
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME_WITH_METHOD_IN_EAX
|
|
PUSH esp // pass SP
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // pass receiver
|
|
PUSH eax // pass proxy method
|
|
call SYMBOL(artQuickProxyInvokeHandler) // (proxy method, receiver, Thread*, SP)
|
|
movd %eax, %xmm0 // place return value also into floating point return value
|
|
movd %edx, %xmm1
|
|
punpckldq %xmm1, %xmm0
|
|
addl LITERAL(16 + FRAME_SIZE_SAVE_REFS_AND_ARGS - FRAME_SIZE_SAVE_REFS_ONLY), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-(16 + FRAME_SIZE_SAVE_REFS_AND_ARGS - FRAME_SIZE_SAVE_REFS_ONLY))
|
|
RESTORE_SAVE_REFS_ONLY_FRAME
|
|
RETURN_OR_DELIVER_PENDING_EXCEPTION // return or deliver exception
|
|
END_FUNCTION art_quick_proxy_invoke_handler
|
|
|
|
/*
|
|
* Called to resolve an imt conflict.
|
|
* eax is the conflict ArtMethod.
|
|
* xmm7 is a hidden argument that holds the target interface method.
|
|
*
|
|
* Note that this stub writes to eax.
|
|
* Because of lack of free registers, it also saves and restores esi.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_imt_conflict_trampoline
|
|
PUSH ESI
|
|
movd %xmm7, %esi // Get target method index stored in xmm7, remember it in ESI.
|
|
movl ART_METHOD_JNI_OFFSET_32(%eax), %eax // Load ImtConflictTable.
|
|
.Limt_table_iterate:
|
|
cmpl %esi, 0(%eax)
|
|
jne .Limt_table_next_entry
|
|
// We successfully hit an entry in the table. Load the target method
|
|
// and jump to it.
|
|
movl __SIZEOF_POINTER__(%eax), %eax
|
|
CFI_REMEMBER_STATE
|
|
POP ESI
|
|
jmp *ART_METHOD_QUICK_CODE_OFFSET_32(%eax)
|
|
CFI_RESTORE_STATE_AND_DEF_CFA esp, 8
|
|
.Limt_table_next_entry:
|
|
// If the entry is null, the interface method is not in the ImtConflictTable.
|
|
cmpl LITERAL(0), 0(%eax)
|
|
jz .Lconflict_trampoline
|
|
// Iterate over the entries of the ImtConflictTable.
|
|
addl LITERAL(2 * __SIZEOF_POINTER__), %eax
|
|
jmp .Limt_table_iterate
|
|
.Lconflict_trampoline:
|
|
// Call the runtime stub to populate the ImtConflictTable and jump to the
|
|
// resolved method.
|
|
// Pass the interface method in first argument.
|
|
movl %esi, %eax
|
|
POP ESI
|
|
INVOKE_TRAMPOLINE_BODY artInvokeInterfaceTrampoline
|
|
END_FUNCTION art_quick_imt_conflict_trampoline
|
|
|
|
DEFINE_FUNCTION art_quick_resolution_trampoline
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME ebx
|
|
movl %esp, %edi
|
|
PUSH EDI // pass SP. do not just PUSH ESP; that messes up unwinding
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // pass receiver
|
|
PUSH eax // pass method
|
|
call SYMBOL(artQuickResolutionTrampoline) // (Method* called, receiver, Thread*, SP)
|
|
movl %eax, %edi // remember code pointer in EDI
|
|
addl LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
test %eax, %eax // if code pointer is null goto deliver pending exception
|
|
jz 1f
|
|
RESTORE_SAVE_REFS_AND_ARGS_FRAME_AND_JUMP
|
|
1:
|
|
RESTORE_SAVE_REFS_AND_ARGS_FRAME
|
|
DELIVER_PENDING_EXCEPTION
|
|
END_FUNCTION art_quick_resolution_trampoline
|
|
|
|
DEFINE_FUNCTION art_quick_generic_jni_trampoline
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME_WITH_METHOD_IN_EAX
|
|
movl %esp, %ebp // save SP at callee-save frame
|
|
CFI_DEF_CFA_REGISTER(ebp)
|
|
subl LITERAL(5120), %esp
|
|
// prepare for artQuickGenericJniTrampoline call
|
|
// (Thread*, managed_sp, reserved_area)
|
|
// (esp) 4(esp) 8(esp) <= C calling convention
|
|
// fs:... ebp esp <= where they are
|
|
|
|
movl %esp, %eax
|
|
subl LITERAL(4), %esp // Padding for 16B alignment.
|
|
pushl %eax // Pass reserved area.
|
|
pushl %ebp // Pass managed frame SP.
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current().
|
|
call SYMBOL(artQuickGenericJniTrampoline) // (Thread*, sp)
|
|
|
|
// The C call will have registered the complete save-frame on success.
|
|
// The result of the call is:
|
|
// eax: pointer to native code, 0 on error.
|
|
// The bottom of the reserved area contains values for arg registers,
|
|
// hidden arg register and SP for out args for the call.
|
|
|
|
// Check for error (class init check or locking for synchronized native method can throw).
|
|
test %eax, %eax
|
|
jz .Lexception_in_native
|
|
|
|
// On x86 there are no registers passed, so no native call args to pop here.
|
|
|
|
// Save code pointer in EDX.
|
|
movl %eax, %edx
|
|
// Load hidden arg (EAX) for @CriticalNative.
|
|
movl 16(%esp), %eax
|
|
// Load SP for out args, releasing unneeded reserved area.
|
|
movl 20(%esp), %esp
|
|
|
|
// Native call.
|
|
call *%edx
|
|
|
|
// result sign extension is handled in C code
|
|
// prepare for artQuickGenericJniEndTrampoline call
|
|
// (Thread*, result, result_f)
|
|
// (esp) 4(esp) 12(esp) <= C calling convention
|
|
// fs:... eax:edx fp0 <= where they are
|
|
|
|
subl LITERAL(20), %esp // Padding & pass float result.
|
|
fstpl (%esp)
|
|
pushl %edx // Pass int result.
|
|
pushl %eax
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current().
|
|
call SYMBOL(artQuickGenericJniEndTrampoline)
|
|
|
|
// Pending exceptions possible.
|
|
mov %fs:THREAD_EXCEPTION_OFFSET, %ebx
|
|
testl %ebx, %ebx
|
|
jnz .Lexception_in_native
|
|
|
|
// Tear down the alloca.
|
|
movl %ebp, %esp
|
|
CFI_REMEMBER_STATE
|
|
CFI_DEF_CFA_REGISTER(esp)
|
|
|
|
// Tear down the callee-save frame.
|
|
// Remove space for FPR args and EAX
|
|
addl LITERAL(4 + 4 * 8), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-(4 + 4 * 8))
|
|
|
|
POP ecx
|
|
addl LITERAL(4), %esp // Avoid edx, as it may be part of the result.
|
|
CFI_ADJUST_CFA_OFFSET(-4)
|
|
POP ebx
|
|
POP ebp // Restore callee saves
|
|
POP esi
|
|
POP edi
|
|
// Quick expects the return value to be in xmm0.
|
|
movd %eax, %xmm0
|
|
movd %edx, %xmm1
|
|
punpckldq %xmm1, %xmm0
|
|
ret
|
|
|
|
// Undo the unwinding information from above since it doesn't apply below.
|
|
CFI_RESTORE_STATE_AND_DEF_CFA ebp, 64
|
|
.Lexception_in_native:
|
|
pushl %fs:THREAD_TOP_QUICK_FRAME_OFFSET
|
|
addl LITERAL(-1), (%esp) // Remove the GenericJNI tag.
|
|
movl (%esp), %esp
|
|
call art_deliver_pending_exception
|
|
END_FUNCTION art_quick_generic_jni_trampoline
|
|
|
|
DEFINE_FUNCTION art_deliver_pending_exception
|
|
// This will create a new save-all frame, required by the runtime.
|
|
DELIVER_PENDING_EXCEPTION
|
|
END_FUNCTION art_deliver_pending_exception
|
|
|
|
DEFINE_FUNCTION art_quick_to_interpreter_bridge
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME ebx // save frame
|
|
mov %esp, %edx // remember SP
|
|
PUSH eax // alignment padding
|
|
PUSH edx // pass SP
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax // pass method
|
|
call SYMBOL(artQuickToInterpreterBridge) // (method, Thread*, SP)
|
|
addl LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
|
|
// Return eax:edx in xmm0 also.
|
|
movd %eax, %xmm0
|
|
movd %edx, %xmm1
|
|
punpckldq %xmm1, %xmm0
|
|
|
|
addl LITERAL(48), %esp // Remove FPRs and EAX, ECX, EDX, EBX.
|
|
CFI_ADJUST_CFA_OFFSET(-48)
|
|
|
|
POP ebp // Restore callee saves
|
|
POP esi
|
|
POP edi
|
|
|
|
RETURN_OR_DELIVER_PENDING_EXCEPTION // return or deliver exception
|
|
END_FUNCTION art_quick_to_interpreter_bridge
|
|
|
|
/*
|
|
* Called by managed code, saves callee saves and then calls artInvokeObsoleteMethod
|
|
*/
|
|
ONE_ARG_RUNTIME_EXCEPTION art_invoke_obsolete_method_stub, artInvokeObsoleteMethod
|
|
|
|
/*
|
|
* Routine that intercepts method calls and returns.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_instrumentation_entry
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME edx
|
|
PUSH eax // Save eax which will be clobbered by the callee-save method.
|
|
subl LITERAL(16), %esp // Align stack (12 bytes) and reserve space for the SP argument
|
|
CFI_ADJUST_CFA_OFFSET(16) // (4 bytes). We lack the scratch registers to calculate the SP
|
|
// right now, so we will just fill it in later.
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current().
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH ecx // Pass receiver.
|
|
PUSH eax // Pass Method*.
|
|
leal 32(%esp), %eax // Put original SP into eax
|
|
movl %eax, 12(%esp) // set SP
|
|
call SYMBOL(artInstrumentationMethodEntryFromCode) // (Method*, Object*, Thread*, SP)
|
|
|
|
addl LITERAL(28), %esp // Pop arguments upto saved Method*.
|
|
CFI_ADJUST_CFA_OFFSET(-28)
|
|
|
|
testl %eax, %eax
|
|
jz 1f // Test for null return (indicating exception) and handle it.
|
|
|
|
movl 60(%esp), %edi // Restore edi.
|
|
movl %eax, 60(%esp) // Place code* over edi, just under return pc.
|
|
SETUP_PC_REL_BASE ebx, .Linstrumentation_entry_pc_rel_base
|
|
leal SYMBOL(art_quick_instrumentation_exit) - .Linstrumentation_entry_pc_rel_base(%ebx), %ebx
|
|
// Place instrumentation exit as return pc. ebx holds the GOT computed on entry.
|
|
movl %ebx, 64(%esp)
|
|
movl 0(%esp), %eax // Restore eax.
|
|
// Restore FPRs (extra 4 bytes of offset due to EAX push at top).
|
|
movsd 8(%esp), %xmm0
|
|
movsd 16(%esp), %xmm1
|
|
movsd 24(%esp), %xmm2
|
|
movsd 32(%esp), %xmm3
|
|
|
|
// Restore GPRs.
|
|
movl 40(%esp), %ecx // Restore ecx.
|
|
movl 44(%esp), %edx // Restore edx.
|
|
movl 48(%esp), %ebx // Restore ebx.
|
|
movl 52(%esp), %ebp // Restore ebp.
|
|
movl 56(%esp), %esi // Restore esi.
|
|
addl LITERAL(60), %esp // Wind stack back upto code*.
|
|
CFI_ADJUST_CFA_OFFSET(-60)
|
|
ret // Call method (and pop).
|
|
1:
|
|
// Make caller handle exception
|
|
addl LITERAL(4), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-4)
|
|
RESTORE_SAVE_REFS_AND_ARGS_FRAME
|
|
DELIVER_PENDING_EXCEPTION
|
|
END_FUNCTION art_quick_instrumentation_entry
|
|
|
|
DEFINE_FUNCTION_CUSTOM_CFA art_quick_instrumentation_exit, 0
|
|
pushl LITERAL(0) // Push a fake return PC as there will be none on the stack.
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx
|
|
|
|
movl %esp, %ecx // Remember SP
|
|
subl LITERAL(8), %esp // Align stack.
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
PUSH edx // Save gpr return value. edx and eax need to be together,
|
|
// which isn't the case in kSaveEverything frame.
|
|
PUSH eax
|
|
leal 32(%esp), %eax // Get pointer to fpr_result, in kSaveEverything frame
|
|
movl %esp, %edx // Get pointer to gpr_result
|
|
PUSH eax // Pass fpr_result
|
|
PUSH edx // Pass gpr_result
|
|
PUSH ecx // Pass SP
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current.
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
|
|
call SYMBOL(artInstrumentationMethodExitFromCode) // (Thread*, SP, gpr_result*, fpr_result*)
|
|
// Return result could have been changed if it's a reference.
|
|
movl 16(%esp), %ecx
|
|
movl %ecx, (80+32)(%esp)
|
|
addl LITERAL(32), %esp // Pop arguments and grp_result.
|
|
CFI_ADJUST_CFA_OFFSET(-32)
|
|
|
|
testl %eax, %eax // Check if we returned error.
|
|
jz .Ldo_deliver_instrumentation_exception
|
|
testl %edx, %edx
|
|
jnz .Ldeoptimize
|
|
// Normal return.
|
|
movl %eax, FRAME_SIZE_SAVE_EVERYTHING-4(%esp) // Set return pc.
|
|
RESTORE_SAVE_EVERYTHING_FRAME
|
|
ret
|
|
.Ldeoptimize:
|
|
mov %edx, (FRAME_SIZE_SAVE_EVERYTHING-4)(%esp) // Set return pc.
|
|
RESTORE_SAVE_EVERYTHING_FRAME
|
|
jmp SYMBOL(art_quick_deoptimize)
|
|
.Ldo_deliver_instrumentation_exception:
|
|
DELIVER_PENDING_EXCEPTION_FRAME_READY
|
|
END_FUNCTION art_quick_instrumentation_exit
|
|
|
|
/*
|
|
* Instrumentation has requested that we deoptimize into the interpreter. The deoptimization
|
|
* will long jump to the upcall with a special exception of -1.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_deoptimize
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx
|
|
subl LITERAL(12), %esp // Align stack.
|
|
CFI_ADJUST_CFA_OFFSET(12)
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current().
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call SYMBOL(artDeoptimize) // (Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION art_quick_deoptimize
|
|
|
|
/*
|
|
* Compiled code has requested that we deoptimize into the interpreter. The deoptimization
|
|
* will long jump to the interpreter bridge.
|
|
*/
|
|
DEFINE_FUNCTION art_quick_deoptimize_from_compiled_code
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx
|
|
subl LITERAL(8), %esp // Align stack.
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current().
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH eax
|
|
call SYMBOL(artDeoptimizeFromCompiledCode) // (DeoptimizationKind, Thread*)
|
|
UNREACHABLE
|
|
END_FUNCTION art_quick_deoptimize_from_compiled_code
|
|
|
|
/*
|
|
* String's compareTo.
|
|
*
|
|
* On entry:
|
|
* eax: this string object (known non-null)
|
|
* ecx: comp string object (known non-null)
|
|
*/
|
|
DEFINE_FUNCTION art_quick_string_compareto
|
|
PUSH esi // push callee save reg
|
|
PUSH edi // push callee save reg
|
|
mov MIRROR_STRING_COUNT_OFFSET(%eax), %edx
|
|
mov MIRROR_STRING_COUNT_OFFSET(%ecx), %ebx
|
|
lea MIRROR_STRING_VALUE_OFFSET(%eax), %esi
|
|
lea MIRROR_STRING_VALUE_OFFSET(%ecx), %edi
|
|
#if (STRING_COMPRESSION_FEATURE)
|
|
/* Differ cases */
|
|
shrl LITERAL(1), %edx
|
|
jnc .Lstring_compareto_this_is_compressed
|
|
shrl LITERAL(1), %ebx
|
|
jnc .Lstring_compareto_that_is_compressed
|
|
jmp .Lstring_compareto_both_not_compressed
|
|
.Lstring_compareto_this_is_compressed:
|
|
shrl LITERAL(1), %ebx
|
|
jnc .Lstring_compareto_both_compressed
|
|
/* If (this->IsCompressed() && that->IsCompressed() == false) */
|
|
mov %edx, %eax
|
|
subl %ebx, %eax
|
|
mov %edx, %ecx
|
|
cmovg %ebx, %ecx
|
|
/* Going into loop to compare each character */
|
|
jecxz .Lstring_compareto_keep_length // check loop counter (if 0, don't compare)
|
|
.Lstring_compareto_loop_comparison_this_compressed:
|
|
movzbl (%esi), %edx // move *(this_cur_char) byte to long
|
|
movzwl (%edi), %ebx // move *(that_cur_char) word to long
|
|
addl LITERAL(1), %esi // ++this_cur_char (8-bit)
|
|
addl LITERAL(2), %edi // ++that_cur_char (16-bit)
|
|
subl %ebx, %edx
|
|
loope .Lstring_compareto_loop_comparison_this_compressed
|
|
cmovne %edx, %eax // return eax = *(this_cur_char) - *(that_cur_char)
|
|
jmp .Lstring_compareto_return
|
|
.Lstring_compareto_that_is_compressed:
|
|
mov %edx, %eax
|
|
subl %ebx, %eax
|
|
mov %edx, %ecx
|
|
cmovg %ebx, %ecx
|
|
/* If (this->IsCompressed() == false && that->IsCompressed()) */
|
|
jecxz .Lstring_compareto_keep_length // check loop counter, if 0, don't compare
|
|
.Lstring_compareto_loop_comparison_that_compressed:
|
|
movzwl (%esi), %edx // move *(this_cur_char) word to long
|
|
movzbl (%edi), %ebx // move *(that_cur_char) byte to long
|
|
addl LITERAL(2), %esi // ++this_cur_char (16-bit)
|
|
addl LITERAL(1), %edi // ++that_cur_char (8-bit)
|
|
subl %ebx, %edx
|
|
loope .Lstring_compareto_loop_comparison_that_compressed
|
|
cmovne %edx, %eax
|
|
jmp .Lstring_compareto_return // return eax = *(this_cur_char) - *(that_cur_char)
|
|
.Lstring_compareto_both_compressed:
|
|
/* Calculate min length and count diff */
|
|
mov %edx, %ecx
|
|
mov %edx, %eax
|
|
subl %ebx, %eax
|
|
cmovg %ebx, %ecx
|
|
jecxz .Lstring_compareto_keep_length
|
|
repe cmpsb
|
|
je .Lstring_compareto_keep_length
|
|
movzbl -1(%esi), %eax // get last compared char from this string (8-bit)
|
|
movzbl -1(%edi), %ecx // get last compared char from comp string (8-bit)
|
|
jmp .Lstring_compareto_count_difference
|
|
#endif // STRING_COMPRESSION_FEATURE
|
|
.Lstring_compareto_both_not_compressed:
|
|
/* Calculate min length and count diff */
|
|
mov %edx, %ecx
|
|
mov %edx, %eax
|
|
subl %ebx, %eax
|
|
cmovg %ebx, %ecx
|
|
/*
|
|
* At this point we have:
|
|
* eax: value to return if first part of strings are equal
|
|
* ecx: minimum among the lengths of the two strings
|
|
* esi: pointer to this string data
|
|
* edi: pointer to comp string data
|
|
*/
|
|
jecxz .Lstring_compareto_keep_length
|
|
repe cmpsw // find nonmatching chars in [%esi] and [%edi], up to length %ecx
|
|
je .Lstring_compareto_keep_length
|
|
movzwl -2(%esi), %eax // get last compared char from this string (16-bit)
|
|
movzwl -2(%edi), %ecx // get last compared char from comp string (16-bit)
|
|
.Lstring_compareto_count_difference:
|
|
subl %ecx, %eax
|
|
.Lstring_compareto_keep_length:
|
|
.Lstring_compareto_return:
|
|
POP edi // pop callee save reg
|
|
POP esi // pop callee save reg
|
|
ret
|
|
END_FUNCTION art_quick_string_compareto
|
|
|
|
DEFINE_FUNCTION art_quick_string_builder_append
|
|
SETUP_SAVE_REFS_ONLY_FRAME ebx // save ref containing registers for GC
|
|
// Outgoing argument set up
|
|
leal FRAME_SIZE_SAVE_REFS_ONLY + __SIZEOF_POINTER__(%esp), %edi // prepare args
|
|
push %eax // push padding
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
push %edi // pass args
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
push %eax // pass format
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call SYMBOL(artStringBuilderAppend) // (uint32_t, const unit32_t*, Thread*)
|
|
addl MACRO_LITERAL(16), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_REFS_ONLY_FRAME // restore frame up to return address
|
|
RETURN_IF_RESULT_IS_NON_ZERO_OR_DELIVER // return or deliver exception
|
|
END_FUNCTION art_quick_string_builder_append
|
|
|
|
// Create a function `name` calling the ReadBarrier::Mark routine,
|
|
// getting its argument and returning its result through register
|
|
// `reg`, saving and restoring all caller-save registers.
|
|
//
|
|
// If `reg` is different from `eax`, the generated function follows a
|
|
// non-standard runtime calling convention:
|
|
// - register `reg` is used to pass the (sole) argument of this function
|
|
// (instead of EAX);
|
|
// - register `reg` is used to return the result of this function
|
|
// (instead of EAX);
|
|
// - EAX is treated like a normal (non-argument) caller-save register;
|
|
// - everything else is the same as in the standard runtime calling
|
|
// convention (e.g. standard callee-save registers are preserved).
|
|
MACRO2(READ_BARRIER_MARK_REG, name, reg)
|
|
DEFINE_FUNCTION VAR(name)
|
|
// Null check so that we can load the lock word.
|
|
test REG_VAR(reg), REG_VAR(reg)
|
|
jz .Lret_rb_\name
|
|
.Lnot_null_\name:
|
|
// Check the mark bit, if it is 1 return.
|
|
testl LITERAL(LOCK_WORD_MARK_BIT_MASK_SHIFTED), MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(reg))
|
|
jz .Lslow_rb_\name
|
|
ret
|
|
.Lslow_rb_\name:
|
|
PUSH eax
|
|
mov MIRROR_OBJECT_LOCK_WORD_OFFSET(REG_VAR(reg)), %eax
|
|
add LITERAL(LOCK_WORD_STATE_FORWARDING_ADDRESS_OVERFLOW), %eax
|
|
// Jump if overflow, the only case where it overflows should be the forwarding address one.
|
|
// Taken ~25% of the time.
|
|
jnae .Lret_forwarding_address\name
|
|
|
|
// Save all potentially live caller-save core registers.
|
|
mov 0(%esp), %eax
|
|
PUSH ecx
|
|
PUSH edx
|
|
PUSH ebx
|
|
// 8-byte align the stack to improve (8-byte) XMM register saving and restoring.
|
|
// and create space for caller-save floating-point registers.
|
|
subl MACRO_LITERAL(4 + 8 * 8), %esp
|
|
CFI_ADJUST_CFA_OFFSET(4 + 8 * 8)
|
|
// Save all potentially live caller-save floating-point registers.
|
|
movsd %xmm0, 0(%esp)
|
|
movsd %xmm1, 8(%esp)
|
|
movsd %xmm2, 16(%esp)
|
|
movsd %xmm3, 24(%esp)
|
|
movsd %xmm4, 32(%esp)
|
|
movsd %xmm5, 40(%esp)
|
|
movsd %xmm6, 48(%esp)
|
|
movsd %xmm7, 56(%esp)
|
|
|
|
subl LITERAL(4), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
PUSH RAW_VAR(reg) // pass arg1 - obj from `reg`
|
|
call SYMBOL(artReadBarrierMark) // artReadBarrierMark(obj)
|
|
.ifnc RAW_VAR(reg), eax
|
|
movl %eax, REG_VAR(reg) // return result into `reg`
|
|
.endif
|
|
addl LITERAL(8), %esp // pop argument and remove padding
|
|
CFI_ADJUST_CFA_OFFSET(-8)
|
|
|
|
// Restore floating-point registers.
|
|
movsd 0(%esp), %xmm0
|
|
movsd 8(%esp), %xmm1
|
|
movsd 16(%esp), %xmm2
|
|
movsd 24(%esp), %xmm3
|
|
movsd 32(%esp), %xmm4
|
|
movsd 40(%esp), %xmm5
|
|
movsd 48(%esp), %xmm6
|
|
movsd 56(%esp), %xmm7
|
|
// Remove floating-point registers and padding.
|
|
addl MACRO_LITERAL(8 * 8 + 4), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-(8 * 8 + 4))
|
|
// Restore core regs, except `reg`, as it is used to return the
|
|
// result of this function (simply remove it from the stack instead).
|
|
POP_REG_NE ebx, RAW_VAR(reg)
|
|
POP_REG_NE edx, RAW_VAR(reg)
|
|
POP_REG_NE ecx, RAW_VAR(reg)
|
|
POP_REG_NE eax, RAW_VAR(reg)
|
|
.Lret_rb_\name:
|
|
ret
|
|
.Lret_forwarding_address\name:
|
|
// The overflow cleared the top bits.
|
|
sall LITERAL(LOCK_WORD_STATE_FORWARDING_ADDRESS_SHIFT), %eax
|
|
mov %eax, REG_VAR(reg)
|
|
POP_REG_NE eax, RAW_VAR(reg)
|
|
ret
|
|
END_FUNCTION VAR(name)
|
|
END_MACRO
|
|
|
|
READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg00, eax
|
|
READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg01, ecx
|
|
READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg02, edx
|
|
READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg03, ebx
|
|
READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg05, ebp
|
|
// Note: There is no art_quick_read_barrier_mark_reg04, as register 4 (ESP)
|
|
// cannot be used to pass arguments.
|
|
READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg06, esi
|
|
READ_BARRIER_MARK_REG art_quick_read_barrier_mark_reg07, edi
|
|
|
|
DEFINE_FUNCTION art_quick_read_barrier_slow
|
|
PUSH edx // pass arg3 - offset
|
|
PUSH ecx // pass arg2 - obj
|
|
PUSH eax // pass arg1 - ref
|
|
call SYMBOL(artReadBarrierSlow) // artReadBarrierSlow(ref, obj, offset)
|
|
addl LITERAL(12), %esp // pop arguments
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
ret
|
|
END_FUNCTION art_quick_read_barrier_slow
|
|
|
|
DEFINE_FUNCTION art_quick_read_barrier_for_root_slow
|
|
subl LITERAL(8), %esp // alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
PUSH eax // pass arg1 - root
|
|
call SYMBOL(artReadBarrierForRootSlow) // artReadBarrierForRootSlow(root)
|
|
addl LITERAL(12), %esp // pop argument and remove padding
|
|
CFI_ADJUST_CFA_OFFSET(-12)
|
|
ret
|
|
END_FUNCTION art_quick_read_barrier_for_root_slow
|
|
|
|
/*
|
|
* On stack replacement stub.
|
|
* On entry:
|
|
* [sp] = return address
|
|
* [sp + 4] = stack to copy
|
|
* [sp + 8] = size of stack
|
|
* [sp + 12] = pc to call
|
|
* [sp + 16] = JValue* result
|
|
* [sp + 20] = shorty
|
|
* [sp + 24] = thread
|
|
*/
|
|
DEFINE_FUNCTION art_quick_osr_stub
|
|
// Save native callee saves.
|
|
PUSH ebp
|
|
PUSH ebx
|
|
PUSH esi
|
|
PUSH edi
|
|
SAVE_SIZE=20 // 4 registers and the return address
|
|
mov 4+16(%esp), %esi // ESI = argument array
|
|
mov 8+16(%esp), %ecx // ECX = size of args
|
|
mov 12+16(%esp), %ebx // EBX = pc to call
|
|
mov %esp, %ebp // Save stack pointer
|
|
CFI_DEF_CFA(ebp, SAVE_SIZE) // CFA = ebp + SAVE_SIZE
|
|
CFI_REMEMBER_STATE
|
|
andl LITERAL(0xFFFFFFF0), %esp // Align stack
|
|
pushl %ebp // Save old stack pointer
|
|
subl LITERAL(12), %esp // Align stack
|
|
movl LITERAL(0), (%esp) // Store null for ArtMethod* slot
|
|
// ebp isn't properly spilled in the osr method, so we need use DWARF expression.
|
|
// NB: the CFI must be before the call since this is the address gdb will lookup.
|
|
// NB: gdb expects that cfa_expression returns the CFA value (not address to it).
|
|
CFI_ESCAPE( /* cfa = [sp + 12] + SAVE_SIZE */ \
|
|
0x0f, 6, /* DW_CFA_def_cfa_expression(len) */ \
|
|
0x92, 4, 12, /* DW_OP_bregx(reg,offset) */ \
|
|
0x06, /* DW_OP_deref */ \
|
|
0x23, SAVE_SIZE) /* DW_OP_plus_uconst(val) */
|
|
call .Losr_entry
|
|
mov 12(%esp), %esp // Restore stack pointer.
|
|
CFI_DEF_CFA(esp, SAVE_SIZE) // CFA = esp + SAVE_SIZE
|
|
|
|
// Restore callee saves.
|
|
POP edi
|
|
POP esi
|
|
POP ebx
|
|
POP ebp
|
|
mov 16(%esp), %ecx // Get JValue result
|
|
mov %eax, (%ecx) // Store the result.
|
|
mov %edx, 4(%ecx) // Store the other half of the result.
|
|
ret
|
|
.Losr_entry:
|
|
CFI_RESTORE_STATE_AND_DEF_CFA ebp, SAVE_SIZE // CFA = ebp + SAVE_SIZE
|
|
subl LITERAL(4), %ecx // Given stack size contains pushed frame pointer, substract it.
|
|
subl %ecx, %esp
|
|
mov %esp, %edi // EDI = beginning of stack
|
|
rep movsb // while (ecx--) { *edi++ = *esi++ }
|
|
jmp *%ebx
|
|
END_FUNCTION art_quick_osr_stub
|
|
|
|
DEFINE_FUNCTION art_quick_invoke_polymorphic
|
|
// On entry: EAX := unused, ECX := receiver
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME ebx // Save frame.
|
|
mov %esp, %edx // Remember SP
|
|
sub LITERAL(4), %esp // Alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
push %edx // Push SP
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
pushl %fs:THREAD_SELF_OFFSET // Push Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
push %ecx // Push receiver (method handle)
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call SYMBOL(artInvokePolymorphic) // invoke with (receiver, thread, SP)
|
|
addl LITERAL(16), %esp // Pop arguments.
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
mov %eax, 4(%esp) // Result is in EAX:EDX. Copy to saved FP state.
|
|
mov %edx, 8(%esp)
|
|
mov %edx, 40(%esp) // Copy EDX to saved context
|
|
RESTORE_SAVE_REFS_AND_ARGS_FRAME
|
|
RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
END_FUNCTION art_quick_invoke_polymorphic
|
|
|
|
DEFINE_FUNCTION art_quick_invoke_custom
|
|
SETUP_SAVE_REFS_AND_ARGS_FRAME ebx // Save frame.
|
|
// EAX := call_site_index
|
|
mov %esp, %ecx // Remember SP.
|
|
subl LITERAL(4), %esp // Alignment padding.
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
push %ecx // pass SP
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
push %eax // pass call_site_index
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call SYMBOL(artInvokeCustom) // artInvokeCustom(call_site_index, Thread*, SP)
|
|
addl LITERAL(16), %esp // Pop arguments.
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
mov %eax, 4(%esp) // Result is in EAX:EDX. Copy to saved FP state.
|
|
mov %edx, 8(%esp)
|
|
mov %edx, 40(%esp) // Copy EDX to saved context
|
|
RESTORE_SAVE_REFS_AND_ARGS_FRAME
|
|
RETURN_OR_DELIVER_PENDING_EXCEPTION
|
|
END_FUNCTION art_quick_invoke_custom
|
|
|
|
// Wrap ExecuteSwitchImpl in assembly method which specifies DEX PC for unwinding.
|
|
// Argument 0: ESP+4: The context pointer for ExecuteSwitchImpl.
|
|
// Argument 1: ESP+8: Pointer to the templated ExecuteSwitchImpl to call.
|
|
// Argument 2: ESP+12: The value of DEX PC (memory address of the methods bytecode).
|
|
DEFINE_FUNCTION ExecuteSwitchImplAsm
|
|
PUSH ebx // Spill EBX; Increments ESP, so arg0 is at ESP+8 now.
|
|
mov 12(%esp), %eax // EAX = C++ templated interpreter function
|
|
mov 16(%esp), %ebx // EBX = DEX PC (callee save register)
|
|
mov 8(%esp), %ecx // ECX = Context argument for the function
|
|
CFI_DEFINE_DEX_PC_WITH_OFFSET(0 /* EAX */, 3 /* EBX */, 0)
|
|
|
|
sub LITERAL(4), %esp // Alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
push %ecx // Push argument
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call *%eax // Call the wrapped function
|
|
addl LITERAL(8), %esp
|
|
CFI_ADJUST_CFA_OFFSET(-8)
|
|
|
|
POP ebx // Restore EBX
|
|
ret
|
|
END_FUNCTION ExecuteSwitchImplAsm
|
|
|
|
// On entry: eax is the class, ebp is the inline cache.
|
|
DEFINE_FUNCTION art_quick_update_inline_cache
|
|
#if (INLINE_CACHE_SIZE != 5)
|
|
#error "INLINE_CACHE_SIZE not as expected."
|
|
#endif
|
|
// Don't update the cache if we are marking.
|
|
cmpl LITERAL(0), %fs:THREAD_IS_GC_MARKING_OFFSET
|
|
jnz .Lret
|
|
PUSH ecx
|
|
movl %eax, %ecx // eax will be used for cmpxchg
|
|
.Lentry1:
|
|
movl INLINE_CACHE_CLASSES_OFFSET(%ebp), %eax
|
|
cmpl %ecx, %eax
|
|
je .Ldone
|
|
cmpl LITERAL(0), %eax
|
|
jne .Lentry2
|
|
lock cmpxchg %ecx, INLINE_CACHE_CLASSES_OFFSET(%ebp)
|
|
jz .Ldone
|
|
jmp .Lentry1
|
|
.Lentry2:
|
|
movl (INLINE_CACHE_CLASSES_OFFSET+4)(%ebp), %eax
|
|
cmpl %ecx, %eax
|
|
je .Ldone
|
|
cmpl LITERAL(0), %eax
|
|
jne .Lentry3
|
|
lock cmpxchg %ecx, (INLINE_CACHE_CLASSES_OFFSET+4)(%ebp)
|
|
jz .Ldone
|
|
jmp .Lentry2
|
|
.Lentry3:
|
|
movl (INLINE_CACHE_CLASSES_OFFSET+8)(%ebp), %eax
|
|
cmpl %ecx, %eax
|
|
je .Ldone
|
|
cmpl LITERAL(0), %eax
|
|
jne .Lentry4
|
|
lock cmpxchg %ecx, (INLINE_CACHE_CLASSES_OFFSET+8)(%ebp)
|
|
jz .Ldone
|
|
jmp .Lentry3
|
|
.Lentry4:
|
|
movl (INLINE_CACHE_CLASSES_OFFSET+12)(%ebp), %eax
|
|
cmpl %ecx, %eax
|
|
je .Ldone
|
|
cmpl LITERAL(0), %eax
|
|
jne .Lentry5
|
|
lock cmpxchg %ecx, (INLINE_CACHE_CLASSES_OFFSET+12)(%ebp)
|
|
jz .Ldone
|
|
jmp .Lentry4
|
|
.Lentry5:
|
|
// Unconditionally store, the cache is megamorphic.
|
|
movl %ecx, (INLINE_CACHE_CLASSES_OFFSET+16)(%ebp)
|
|
.Ldone:
|
|
// Restore registers
|
|
movl %ecx, %eax
|
|
POP ecx
|
|
.Lret:
|
|
ret
|
|
END_FUNCTION art_quick_update_inline_cache
|
|
|
|
// TODO: implement these!
|
|
UNIMPLEMENTED art_quick_memcmp16
|
|
|
|
// On entry, the method is at the bottom of the stack.
|
|
DEFINE_FUNCTION art_quick_compile_optimized
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx
|
|
mov FRAME_SIZE_SAVE_EVERYTHING(%esp), %eax // Fetch ArtMethod
|
|
sub LITERAL(8), %esp // Alignment padding
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
pushl %fs:THREAD_SELF_OFFSET // pass Thread::Current()
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
pushl %eax
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call SYMBOL(artCompileOptimized) // (ArtMethod*, Thread*)
|
|
addl LITERAL(16), %esp // Pop arguments.
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
RESTORE_SAVE_EVERYTHING_FRAME
|
|
ret
|
|
END_FUNCTION art_quick_compile_optimized
|
|
|
|
DEFINE_FUNCTION art_quick_method_entry_hook
|
|
SETUP_SAVE_EVERYTHING_FRAME edx
|
|
mov FRAME_SIZE_SAVE_EVERYTHING(%esp), %eax // Fetch ArtMethod
|
|
subl LITERAL(8), %esp
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current().
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
pushl %eax // Pass Method*.
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
|
|
call SYMBOL(artMethodEntryHook) // (Method*, Thread*)
|
|
|
|
addl LITERAL(16), %esp // Pop arguments.
|
|
CFI_ADJUST_CFA_OFFSET(-16)
|
|
|
|
RESTORE_SAVE_EVERYTHING_FRAME
|
|
ret
|
|
END_FUNCTION art_quick_method_entry_hook
|
|
|
|
DEFINE_FUNCTION art_quick_method_exit_hook
|
|
SETUP_SAVE_EVERYTHING_FRAME ebx
|
|
|
|
mov FRAME_SIZE_SAVE_EVERYTHING(%esp), %ebx // Remember ArtMethod*
|
|
subl LITERAL(8), %esp // Align stack.
|
|
CFI_ADJUST_CFA_OFFSET(8)
|
|
PUSH_ARG edx // Save gpr return value. edx and eax need to be together
|
|
// which isn't the case in kSaveEverything frame.
|
|
PUSH_ARG eax
|
|
movl %esp, %edx // Get pointer to gpr_result
|
|
leal 32(%esp), %eax // Get pointer to fpr_result, in kSaveEverything frame
|
|
PUSH_ARG eax // Pass fpr_result
|
|
PUSH_ARG edx // Pass gpr_result
|
|
PUSH_ARG ebx // Pass ArtMethod*
|
|
pushl %fs:THREAD_SELF_OFFSET // Pass Thread::Current.
|
|
CFI_ADJUST_CFA_OFFSET(4)
|
|
call SYMBOL(artMethodExitHook) // (Thread*, ArtMethod*, gpr_result*, fpr_result*)
|
|
|
|
// Return result could have been changed if it's a reference.
|
|
movl 16(%esp), %ecx
|
|
movl %ecx, (80+32)(%esp)
|
|
addl LITERAL(32), %esp // Pop arguments and grp_result.
|
|
CFI_ADJUST_CFA_OFFSET(-32)
|
|
|
|
cmpl LITERAL(1), %eax // Check if we returned error.
|
|
CFI_REMEMBER_STATE
|
|
je .Ldo_deliver_instrumentation_exception_exit
|
|
|
|
// Normal return.
|
|
RESTORE_SAVE_EVERYTHING_FRAME
|
|
ret
|
|
.Ldo_deliver_instrumentation_exception_exit:
|
|
CFI_RESTORE_STATE_AND_DEF_CFA esp, FRAME_SIZE_SAVE_EVERYTHING
|
|
DELIVER_PENDING_EXCEPTION_FRAME_READY
|
|
END_FUNCTION art_quick_method_exit_hook
|
|
|
|
|
|
|