OSDN Git Service

am 9db54b1e: Add dx option to always generate const-string/jumbo. do not merge.
[android-x86/dalvik.git] / vm / interp / Jit.cpp
1 /*
2  * Copyright (C) 2008 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 #ifdef WITH_JIT
17
18 /*
19  * Target independent portion of Android's Jit
20  */
21
22 #include "Dalvik.h"
23 #include "Jit.h"
24
25 #include "libdex/DexOpcodes.h"
26 #include <unistd.h>
27 #include <pthread.h>
28 #include <sys/time.h>
29 #include <signal.h>
30 #include "compiler/Compiler.h"
31 #include "compiler/CompilerUtility.h"
32 #include "compiler/CompilerIR.h"
33 #include <errno.h>
34
35 #if defined(WITH_SELF_VERIFICATION)
36 /* Allocate space for per-thread ShadowSpace data structures */
37 void* dvmSelfVerificationShadowSpaceAlloc(Thread* self)
38 {
39     self->shadowSpace = (ShadowSpace*) calloc(1, sizeof(ShadowSpace));
40     if (self->shadowSpace == NULL)
41         return NULL;
42
43     self->shadowSpace->registerSpaceSize = REG_SPACE;
44     self->shadowSpace->registerSpace =
45         (int*) calloc(self->shadowSpace->registerSpaceSize, sizeof(int));
46
47     return self->shadowSpace->registerSpace;
48 }
49
50 /* Free per-thread ShadowSpace data structures */
51 void dvmSelfVerificationShadowSpaceFree(Thread* self)
52 {
53     free(self->shadowSpace->registerSpace);
54     free(self->shadowSpace);
55 }
56
57 /*
58  * Save out PC, FP, thread state, and registers to shadow space.
59  * Return a pointer to the shadow space for JIT to use.
60  *
61  * The set of saved state from the Thread structure is:
62  *     pc  (Dalvik PC)
63  *     fp  (Dalvik FP)
64  *     retval
65  *     method
66  *     methodClassDex
67  *     interpStackEnd
68  */
69 void* dvmSelfVerificationSaveState(const u2* pc, u4* fp,
70                                    Thread* self, int targetTrace)
71 {
72     ShadowSpace *shadowSpace = self->shadowSpace;
73     unsigned preBytes = self->interpSave.method->outsSize*4 +
74         sizeof(StackSaveArea);
75     unsigned postBytes = self->interpSave.method->registersSize*4;
76
77     //ALOGD("### selfVerificationSaveState(%d) pc: %#x fp: %#x",
78     //    self->threadId, (int)pc, (int)fp);
79
80     if (shadowSpace->selfVerificationState != kSVSIdle) {
81         ALOGD("~~~ Save: INCORRECT PREVIOUS STATE(%d): %d",
82             self->threadId, shadowSpace->selfVerificationState);
83         ALOGD("********** SHADOW STATE DUMP **********");
84         ALOGD("PC: %#x FP: %#x", (int)pc, (int)fp);
85     }
86     shadowSpace->selfVerificationState = kSVSStart;
87
88     // Dynamically grow shadow register space if necessary
89     if (preBytes + postBytes > shadowSpace->registerSpaceSize * sizeof(u4)) {
90         free(shadowSpace->registerSpace);
91         shadowSpace->registerSpaceSize = (preBytes + postBytes) / sizeof(u4);
92         shadowSpace->registerSpace =
93             (int*) calloc(shadowSpace->registerSpaceSize, sizeof(u4));
94     }
95
96     // Remember original state
97     shadowSpace->startPC = pc;
98     shadowSpace->fp = fp;
99     shadowSpace->retval = self->interpSave.retval;
100     shadowSpace->interpStackEnd = self->interpStackEnd;
101
102     /*
103      * Store the original method here in case the trace ends with a
104      * return/invoke, the last method.
105      */
106     shadowSpace->method = self->interpSave.method;
107     shadowSpace->methodClassDex = self->interpSave.methodClassDex;
108
109     shadowSpace->shadowFP = shadowSpace->registerSpace +
110                             shadowSpace->registerSpaceSize - postBytes/4;
111
112     self->interpSave.curFrame = (u4*)shadowSpace->shadowFP;
113     self->interpStackEnd = (u1*)shadowSpace->registerSpace;
114
115     // Create a copy of the stack
116     memcpy(((char*)shadowSpace->shadowFP)-preBytes, ((char*)fp)-preBytes,
117         preBytes+postBytes);
118
119     // Setup the shadowed heap space
120     shadowSpace->heapSpaceTail = shadowSpace->heapSpace;
121
122     // Reset trace length
123     shadowSpace->traceLength = 0;
124
125     return shadowSpace;
126 }
127
128 /*
129  * Save ending PC, FP and compiled code exit point to shadow space.
130  * Return a pointer to the shadow space for JIT to restore state.
131  */
132 void* dvmSelfVerificationRestoreState(const u2* pc, u4* fp,
133                                       SelfVerificationState exitState,
134                                       Thread* self)
135 {
136     ShadowSpace *shadowSpace = self->shadowSpace;
137     shadowSpace->endPC = pc;
138     shadowSpace->endShadowFP = fp;
139     shadowSpace->jitExitState = exitState;
140
141     //ALOGD("### selfVerificationRestoreState(%d) pc: %#x fp: %#x endPC: %#x",
142     //    self->threadId, (int)shadowSpace->startPC, (int)shadowSpace->fp,
143     //    (int)pc);
144
145     if (shadowSpace->selfVerificationState != kSVSStart) {
146         ALOGD("~~~ Restore: INCORRECT PREVIOUS STATE(%d): %d",
147             self->threadId, shadowSpace->selfVerificationState);
148         ALOGD("********** SHADOW STATE DUMP **********");
149         ALOGD("Dalvik PC: %#x endPC: %#x", (int)shadowSpace->startPC,
150             (int)shadowSpace->endPC);
151         ALOGD("Interp FP: %#x", (int)shadowSpace->fp);
152         ALOGD("Shadow FP: %#x endFP: %#x", (int)shadowSpace->shadowFP,
153             (int)shadowSpace->endShadowFP);
154     }
155
156     // Special case when punting after a single instruction
157     if (exitState == kSVSPunt && pc == shadowSpace->startPC) {
158         shadowSpace->selfVerificationState = kSVSIdle;
159     } else {
160         shadowSpace->selfVerificationState = exitState;
161     }
162
163     /* Restore state before returning */
164     self->interpSave.pc = shadowSpace->startPC;
165     self->interpSave.curFrame = shadowSpace->fp;
166     self->interpSave.method = shadowSpace->method;
167     self->interpSave.methodClassDex = shadowSpace->methodClassDex;
168     self->interpSave.retval = shadowSpace->retval;
169     self->interpStackEnd = shadowSpace->interpStackEnd;
170
171     return shadowSpace;
172 }
173
174 /* Print contents of virtual registers */
175 static void selfVerificationPrintRegisters(int* addr, int* addrRef,
176                                            int numWords)
177 {
178     int i;
179     for (i = 0; i < numWords; i++) {
180         ALOGD("(v%d) 0x%8x%s", i, addr[i], addr[i] != addrRef[i] ? " X" : "");
181     }
182 }
183
184 /* Print values maintained in shadowSpace */
185 static void selfVerificationDumpState(const u2* pc, Thread* self)
186 {
187     ShadowSpace* shadowSpace = self->shadowSpace;
188     StackSaveArea* stackSave = SAVEAREA_FROM_FP(self->interpSave.curFrame);
189     int frameBytes = (int) shadowSpace->registerSpace +
190                      shadowSpace->registerSpaceSize*4 -
191                      (int) shadowSpace->shadowFP;
192     int localRegs = 0;
193     int frameBytes2 = 0;
194     if ((uintptr_t)self->interpSave.curFrame < (uintptr_t)shadowSpace->fp) {
195         localRegs = (stackSave->method->registersSize -
196                      stackSave->method->insSize)*4;
197         frameBytes2 = (int) shadowSpace->fp -
198                       (int)self->interpSave.curFrame - localRegs;
199     }
200     ALOGD("********** SHADOW STATE DUMP **********");
201     ALOGD("CurrentPC: %#x, Offset: 0x%04x", (int)pc,
202         (int)(pc - stackSave->method->insns));
203     ALOGD("Class: %s", shadowSpace->method->clazz->descriptor);
204     ALOGD("Method: %s", shadowSpace->method->name);
205     ALOGD("Dalvik PC: %#x endPC: %#x", (int)shadowSpace->startPC,
206         (int)shadowSpace->endPC);
207     ALOGD("Interp FP: %#x endFP: %#x", (int)shadowSpace->fp,
208         (int)self->interpSave.curFrame);
209     ALOGD("Shadow FP: %#x endFP: %#x", (int)shadowSpace->shadowFP,
210         (int)shadowSpace->endShadowFP);
211     ALOGD("Frame1 Bytes: %d Frame2 Local: %d Bytes: %d", frameBytes,
212         localRegs, frameBytes2);
213     ALOGD("Trace length: %d State: %d", shadowSpace->traceLength,
214         shadowSpace->selfVerificationState);
215 }
216
217 /* Print decoded instructions in the current trace */
218 static void selfVerificationDumpTrace(const u2* pc, Thread* self)
219 {
220     ShadowSpace* shadowSpace = self->shadowSpace;
221     StackSaveArea* stackSave = SAVEAREA_FROM_FP(self->interpSave.curFrame);
222     int i, addr, offset;
223     DecodedInstruction *decInsn;
224
225     ALOGD("********** SHADOW TRACE DUMP **********");
226     for (i = 0; i < shadowSpace->traceLength; i++) {
227         addr = shadowSpace->trace[i].addr;
228         offset =  (int)((u2*)addr - stackSave->method->insns);
229         decInsn = &(shadowSpace->trace[i].decInsn);
230         /* Not properly decoding instruction, some registers may be garbage */
231         ALOGD("%#x: (0x%04x) %s",
232             addr, offset, dexGetOpcodeName(decInsn->opcode));
233     }
234 }
235
236 /* Code is forced into this spin loop when a divergence is detected */
237 static void selfVerificationSpinLoop(ShadowSpace *shadowSpace)
238 {
239     const u2 *startPC = shadowSpace->startPC;
240     JitTraceDescription* desc = dvmCopyTraceDescriptor(startPC, NULL);
241     if (desc) {
242         dvmCompilerWorkEnqueue(startPC, kWorkOrderTraceDebug, desc);
243         /*
244          * This function effectively terminates the VM right here, so not
245          * freeing the desc pointer when the enqueuing fails is acceptable.
246          */
247     }
248     gDvmJit.selfVerificationSpin = true;
249     while(gDvmJit.selfVerificationSpin) sleep(10);
250 }
251
252 /*
253  * If here, we're re-interpreting an instruction that was included
254  * in a trace that was just executed.  This routine is called for
255  * each instruction in the original trace, and compares state
256  * when it reaches the end point.
257  *
258  * TUNING: the interpretation mechanism now supports a counted
259  * single-step mechanism.  If we were to associate an instruction
260  * count with each trace exit, we could just single-step the right
261  * number of cycles and then compare.  This would improve detection
262  * of control divergences, as well as (slightly) simplify this code.
263  */
264 void dvmCheckSelfVerification(const u2* pc, Thread* self)
265 {
266     ShadowSpace *shadowSpace = self->shadowSpace;
267     SelfVerificationState state = shadowSpace->selfVerificationState;
268
269     DecodedInstruction decInsn;
270     dexDecodeInstruction(pc, &decInsn);
271
272     //ALOGD("### DbgIntp(%d): PC: %#x endPC: %#x state: %d len: %d %s",
273     //    self->threadId, (int)pc, (int)shadowSpace->endPC, state,
274     //    shadowSpace->traceLength, dexGetOpcodeName(decInsn.opcode));
275
276     if (state == kSVSIdle || state == kSVSStart) {
277         ALOGD("~~~ DbgIntrp: INCORRECT PREVIOUS STATE(%d): %d",
278             self->threadId, state);
279         selfVerificationDumpState(pc, self);
280         selfVerificationDumpTrace(pc, self);
281     }
282
283     /*
284      * Generalize the self verification state to kSVSDebugInterp unless the
285      * entry reason is kSVSBackwardBranch or kSVSSingleStep.
286      */
287     if (state != kSVSBackwardBranch && state != kSVSSingleStep) {
288         shadowSpace->selfVerificationState = kSVSDebugInterp;
289     }
290
291     /*
292      * Check if the current pc matches the endPC. Only check for non-zero
293      * trace length when backward branches are involved.
294      */
295     if (pc == shadowSpace->endPC &&
296         (state == kSVSDebugInterp || state == kSVSSingleStep ||
297          (state == kSVSBackwardBranch && shadowSpace->traceLength != 0))) {
298
299         shadowSpace->selfVerificationState = kSVSIdle;
300
301         /* Check register space */
302         int frameBytes = (int) shadowSpace->registerSpace +
303                          shadowSpace->registerSpaceSize*4 -
304                          (int) shadowSpace->shadowFP;
305         if (memcmp(shadowSpace->fp, shadowSpace->shadowFP, frameBytes)) {
306             if (state == kSVSBackwardBranch) {
307                 /* State mismatch on backward branch - try one more iteration */
308                 shadowSpace->selfVerificationState = kSVSDebugInterp;
309                 goto log_and_continue;
310             }
311             ALOGD("~~~ DbgIntp(%d): REGISTERS DIVERGENCE!", self->threadId);
312             selfVerificationDumpState(pc, self);
313             selfVerificationDumpTrace(pc, self);
314             ALOGD("*** Interp Registers: addr: %#x bytes: %d",
315                 (int)shadowSpace->fp, frameBytes);
316             selfVerificationPrintRegisters((int*)shadowSpace->fp,
317                                            (int*)shadowSpace->shadowFP,
318                                            frameBytes/4);
319             ALOGD("*** Shadow Registers: addr: %#x bytes: %d",
320                 (int)shadowSpace->shadowFP, frameBytes);
321             selfVerificationPrintRegisters((int*)shadowSpace->shadowFP,
322                                            (int*)shadowSpace->fp,
323                                            frameBytes/4);
324             selfVerificationSpinLoop(shadowSpace);
325         }
326         /* Check new frame if it exists (invokes only) */
327         if ((uintptr_t)self->interpSave.curFrame < (uintptr_t)shadowSpace->fp) {
328             StackSaveArea* stackSave =
329                 SAVEAREA_FROM_FP(self->interpSave.curFrame);
330             int localRegs = (stackSave->method->registersSize -
331                              stackSave->method->insSize)*4;
332             int frameBytes2 = (int) shadowSpace->fp -
333                               (int) self->interpSave.curFrame - localRegs;
334             if (memcmp(((char*)self->interpSave.curFrame)+localRegs,
335                 ((char*)shadowSpace->endShadowFP)+localRegs, frameBytes2)) {
336                 if (state == kSVSBackwardBranch) {
337                     /*
338                      * State mismatch on backward branch - try one more
339                      * iteration.
340                      */
341                     shadowSpace->selfVerificationState = kSVSDebugInterp;
342                     goto log_and_continue;
343                 }
344                 ALOGD("~~~ DbgIntp(%d): REGISTERS (FRAME2) DIVERGENCE!",
345                     self->threadId);
346                 selfVerificationDumpState(pc, self);
347                 selfVerificationDumpTrace(pc, self);
348                 ALOGD("*** Interp Registers: addr: %#x l: %d bytes: %d",
349                     (int)self->interpSave.curFrame, localRegs, frameBytes2);
350                 selfVerificationPrintRegisters((int*)self->interpSave.curFrame,
351                                                (int*)shadowSpace->endShadowFP,
352                                                (frameBytes2+localRegs)/4);
353                 ALOGD("*** Shadow Registers: addr: %#x l: %d bytes: %d",
354                     (int)shadowSpace->endShadowFP, localRegs, frameBytes2);
355                 selfVerificationPrintRegisters((int*)shadowSpace->endShadowFP,
356                                                (int*)self->interpSave.curFrame,
357                                                (frameBytes2+localRegs)/4);
358                 selfVerificationSpinLoop(shadowSpace);
359             }
360         }
361
362         /* Check memory space */
363         bool memDiff = false;
364         ShadowHeap* heapSpacePtr;
365         for (heapSpacePtr = shadowSpace->heapSpace;
366              heapSpacePtr != shadowSpace->heapSpaceTail; heapSpacePtr++) {
367             int memData = *((unsigned int*) heapSpacePtr->addr);
368             if (heapSpacePtr->data != memData) {
369                 if (state == kSVSBackwardBranch) {
370                     /*
371                      * State mismatch on backward branch - try one more
372                      * iteration.
373                      */
374                     shadowSpace->selfVerificationState = kSVSDebugInterp;
375                     goto log_and_continue;
376                 }
377                 ALOGD("~~~ DbgIntp(%d): MEMORY DIVERGENCE!", self->threadId);
378                 ALOGD("Addr: %#x Intrp Data: %#x Jit Data: %#x",
379                     heapSpacePtr->addr, memData, heapSpacePtr->data);
380                 selfVerificationDumpState(pc, self);
381                 selfVerificationDumpTrace(pc, self);
382                 memDiff = true;
383             }
384         }
385         if (memDiff) selfVerificationSpinLoop(shadowSpace);
386
387
388         /*
389          * Success.  If this shadowed trace included a single-stepped
390          * instruction, we need to stay in the interpreter for one
391          * more interpretation before resuming.
392          */
393         if (state == kSVSSingleStep) {
394             assert(self->jitResumeNPC != NULL);
395             assert(self->singleStepCount == 0);
396             self->singleStepCount = 1;
397             dvmEnableSubMode(self, kSubModeCountedStep);
398         }
399
400         /*
401          * Switch off shadow replay mode.  The next shadowed trace
402          * execution will turn it back on.
403          */
404         dvmDisableSubMode(self, kSubModeJitSV);
405
406         self->jitState = kJitDone;
407         return;
408     }
409 log_and_continue:
410     /* If end not been reached, make sure max length not exceeded */
411     if (shadowSpace->traceLength >= JIT_MAX_TRACE_LEN) {
412         ALOGD("~~~ DbgIntp(%d): CONTROL DIVERGENCE!", self->threadId);
413         ALOGD("startPC: %#x endPC: %#x currPC: %#x",
414             (int)shadowSpace->startPC, (int)shadowSpace->endPC, (int)pc);
415         selfVerificationDumpState(pc, self);
416         selfVerificationDumpTrace(pc, self);
417         selfVerificationSpinLoop(shadowSpace);
418         return;
419     }
420     /* Log the instruction address and decoded instruction for debug */
421     shadowSpace->trace[shadowSpace->traceLength].addr = (int)pc;
422     shadowSpace->trace[shadowSpace->traceLength].decInsn = decInsn;
423     shadowSpace->traceLength++;
424 }
425 #endif
426
427 /*
428  * If one of our fixed tables or the translation buffer fills up,
429  * call this routine to avoid wasting cycles on future translation requests.
430  */
431 void dvmJitStopTranslationRequests()
432 {
433     /*
434      * Note 1: This won't necessarily stop all translation requests, and
435      * operates on a delayed mechanism.  Running threads look to the copy
436      * of this value in their private thread structures and won't see
437      * this change until it is refreshed (which happens on interpreter
438      * entry).
439      * Note 2: This is a one-shot memory leak on this table. Because this is a
440      * permanent off switch for Jit profiling, it is a one-time leak of 1K
441      * bytes, and no further attempt will be made to re-allocate it.  Can't
442      * free it because some thread may be holding a reference.
443      */
444     gDvmJit.pProfTable = NULL;
445     dvmJitUpdateThreadStateAll();
446 }
447
448 #if defined(WITH_JIT_TUNING)
449 /* Convenience function to increment counter from assembly code */
450 void dvmBumpNoChain(int from)
451 {
452     gDvmJit.noChainExit[from]++;
453 }
454
455 /* Convenience function to increment counter from assembly code */
456 void dvmBumpNormal()
457 {
458     gDvmJit.normalExit++;
459 }
460
461 /* Convenience function to increment counter from assembly code */
462 void dvmBumpPunt(int from)
463 {
464     gDvmJit.puntExit++;
465 }
466 #endif
467
468 /* Dumps debugging & tuning stats to the log */
469 void dvmJitStats()
470 {
471     int i;
472     int hit;
473     int not_hit;
474     int chains;
475     int stubs;
476     if (gDvmJit.pJitEntryTable) {
477         for (i=0, stubs=chains=hit=not_hit=0;
478              i < (int) gDvmJit.jitTableSize;
479              i++) {
480             if (gDvmJit.pJitEntryTable[i].dPC != 0) {
481                 hit++;
482                 if (gDvmJit.pJitEntryTable[i].codeAddress ==
483                       dvmCompilerGetInterpretTemplate())
484                     stubs++;
485             } else
486                 not_hit++;
487             if (gDvmJit.pJitEntryTable[i].u.info.chain != gDvmJit.jitTableSize)
488                 chains++;
489         }
490         ALOGD("JIT: table size is %d, entries used is %d",
491              gDvmJit.jitTableSize,  gDvmJit.jitTableEntriesUsed);
492         ALOGD("JIT: %d traces, %d slots, %d chains, %d thresh, %s",
493              hit, not_hit + hit, chains, gDvmJit.threshold,
494              gDvmJit.blockingMode ? "Blocking" : "Non-blocking");
495
496 #if defined(WITH_JIT_TUNING)
497         ALOGD("JIT: Code cache patches: %d", gDvmJit.codeCachePatches);
498
499         ALOGD("JIT: Lookups: %d hits, %d misses; %d normal, %d punt",
500              gDvmJit.addrLookupsFound, gDvmJit.addrLookupsNotFound,
501              gDvmJit.normalExit, gDvmJit.puntExit);
502
503         ALOGD("JIT: ICHits: %d", gDvmICHitCount);
504
505         ALOGD("JIT: noChainExit: %d IC miss, %d interp callsite, "
506              "%d switch overflow",
507              gDvmJit.noChainExit[kInlineCacheMiss],
508              gDvmJit.noChainExit[kCallsiteInterpreted],
509              gDvmJit.noChainExit[kSwitchOverflow]);
510
511         ALOGD("JIT: ICPatch: %d init, %d rejected, %d lock-free, %d queued, "
512              "%d dropped",
513              gDvmJit.icPatchInit, gDvmJit.icPatchRejected,
514              gDvmJit.icPatchLockFree, gDvmJit.icPatchQueued,
515              gDvmJit.icPatchDropped);
516
517         ALOGD("JIT: Invoke: %d mono, %d poly, %d native, %d return",
518              gDvmJit.invokeMonomorphic, gDvmJit.invokePolymorphic,
519              gDvmJit.invokeNative, gDvmJit.returnOp);
520         ALOGD("JIT: Inline: %d mgetter, %d msetter, %d pgetter, %d psetter",
521              gDvmJit.invokeMonoGetterInlined, gDvmJit.invokeMonoSetterInlined,
522              gDvmJit.invokePolyGetterInlined, gDvmJit.invokePolySetterInlined);
523         ALOGD("JIT: Total compilation time: %llu ms", gDvmJit.jitTime / 1000);
524         ALOGD("JIT: Avg unit compilation time: %llu us",
525              gDvmJit.numCompilations == 0 ? 0 :
526              gDvmJit.jitTime / gDvmJit.numCompilations);
527         ALOGD("JIT: Potential GC blocked by compiler: max %llu us / "
528              "avg %llu us (%d)",
529              gDvmJit.maxCompilerThreadBlockGCTime,
530              gDvmJit.numCompilerThreadBlockGC == 0 ?
531                  0 : gDvmJit.compilerThreadBlockGCTime /
532                      gDvmJit.numCompilerThreadBlockGC,
533              gDvmJit.numCompilerThreadBlockGC);
534 #endif
535
536         ALOGD("JIT: %d Translation chains, %d interp stubs",
537              gDvmJit.translationChains, stubs);
538         if (gDvmJit.profileMode == kTraceProfilingContinuous) {
539             dvmCompilerSortAndPrintTraceProfiles();
540         }
541     }
542 }
543
544
545 /* End current trace now & don't include current instruction */
546 void dvmJitEndTraceSelect(Thread* self, const u2* dPC)
547 {
548     if (self->jitState == kJitTSelect) {
549         self->jitState = kJitTSelectEnd;
550     }
551     if (self->jitState == kJitTSelectEnd) {
552         // Clean up and finish now.
553         dvmCheckJit(dPC, self);
554     }
555 }
556
557 /*
558  * Find an entry in the JitTable, creating if necessary.
559  * Returns null if table is full.
560  */
561 static JitEntry *lookupAndAdd(const u2* dPC, bool callerLocked,
562                               bool isMethodEntry)
563 {
564     u4 chainEndMarker = gDvmJit.jitTableSize;
565     u4 idx = dvmJitHash(dPC);
566
567     /*
568      * Walk the bucket chain to find an exact match for our PC and trace/method
569      * type
570      */
571     while ((gDvmJit.pJitEntryTable[idx].u.info.chain != chainEndMarker) &&
572            ((gDvmJit.pJitEntryTable[idx].dPC != dPC) ||
573             (gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry !=
574              isMethodEntry))) {
575         idx = gDvmJit.pJitEntryTable[idx].u.info.chain;
576     }
577
578     if (gDvmJit.pJitEntryTable[idx].dPC != dPC ||
579         gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry != isMethodEntry) {
580         /*
581          * No match.  Aquire jitTableLock and find the last
582          * slot in the chain. Possibly continue the chain walk in case
583          * some other thread allocated the slot we were looking
584          * at previuosly (perhaps even the dPC we're trying to enter).
585          */
586         if (!callerLocked)
587             dvmLockMutex(&gDvmJit.tableLock);
588         /*
589          * At this point, if .dPC is NULL, then the slot we're
590          * looking at is the target slot from the primary hash
591          * (the simple, and common case).  Otherwise we're going
592          * to have to find a free slot and chain it.
593          */
594         ANDROID_MEMBAR_FULL(); /* Make sure we reload [].dPC after lock */
595         if (gDvmJit.pJitEntryTable[idx].dPC != NULL) {
596             u4 prev;
597             while (gDvmJit.pJitEntryTable[idx].u.info.chain != chainEndMarker) {
598                 if (gDvmJit.pJitEntryTable[idx].dPC == dPC &&
599                     gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry ==
600                         isMethodEntry) {
601                     /* Another thread got there first for this dPC */
602                     if (!callerLocked)
603                         dvmUnlockMutex(&gDvmJit.tableLock);
604                     return &gDvmJit.pJitEntryTable[idx];
605                 }
606                 idx = gDvmJit.pJitEntryTable[idx].u.info.chain;
607             }
608             /* Here, idx should be pointing to the last cell of an
609              * active chain whose last member contains a valid dPC */
610             assert(gDvmJit.pJitEntryTable[idx].dPC != NULL);
611             /* Linear walk to find a free cell and add it to the end */
612             prev = idx;
613             while (true) {
614                 idx++;
615                 if (idx == chainEndMarker)
616                     idx = 0;  /* Wraparound */
617                 if ((gDvmJit.pJitEntryTable[idx].dPC == NULL) ||
618                     (idx == prev))
619                     break;
620             }
621             if (idx != prev) {
622                 JitEntryInfoUnion oldValue;
623                 JitEntryInfoUnion newValue;
624                 /*
625                  * Although we hold the lock so that noone else will
626                  * be trying to update a chain field, the other fields
627                  * packed into the word may be in use by other threads.
628                  */
629                 do {
630                     oldValue = gDvmJit.pJitEntryTable[prev].u;
631                     newValue = oldValue;
632                     newValue.info.chain = idx;
633                 } while (android_atomic_release_cas(oldValue.infoWord,
634                         newValue.infoWord,
635                         &gDvmJit.pJitEntryTable[prev].u.infoWord) != 0);
636             }
637         }
638         if (gDvmJit.pJitEntryTable[idx].dPC == NULL) {
639             gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry = isMethodEntry;
640             /*
641              * Initialize codeAddress and allocate the slot.  Must
642              * happen in this order (since dPC is set, the entry is live.
643              */
644             android_atomic_release_store((int32_t)dPC,
645                  (volatile int32_t *)(void *)&gDvmJit.pJitEntryTable[idx].dPC);
646             /* for simulator mode, we need to initialized codeAddress to null */
647             gDvmJit.pJitEntryTable[idx].codeAddress = NULL;
648             gDvmJit.pJitEntryTable[idx].dPC = dPC;
649             gDvmJit.jitTableEntriesUsed++;
650         } else {
651             /* Table is full */
652             idx = chainEndMarker;
653         }
654         if (!callerLocked)
655             dvmUnlockMutex(&gDvmJit.tableLock);
656     }
657     return (idx == chainEndMarker) ? NULL : &gDvmJit.pJitEntryTable[idx];
658 }
659
660 /* Dump a trace description */
661 void dvmJitDumpTraceDesc(JitTraceDescription *trace)
662 {
663     int i;
664     bool done = false;
665     const u2* dpc;
666     const u2* dpcBase;
667     int curFrag = 0;
668     ALOGD("===========================================");
669     ALOGD("Trace dump %#x, Method %s off %#x",(int)trace,
670          trace->method->name,trace->trace[curFrag].info.frag.startOffset);
671     dpcBase = trace->method->insns;
672     while (!done) {
673         DecodedInstruction decInsn;
674         if (trace->trace[curFrag].isCode) {
675             ALOGD("Frag[%d]- Insts: %d, start: %#x, hint: %#x, end: %d",
676                  curFrag, trace->trace[curFrag].info.frag.numInsts,
677                  trace->trace[curFrag].info.frag.startOffset,
678                  trace->trace[curFrag].info.frag.hint,
679                  trace->trace[curFrag].info.frag.runEnd);
680             dpc = dpcBase + trace->trace[curFrag].info.frag.startOffset;
681             for (i=0; i<trace->trace[curFrag].info.frag.numInsts; i++) {
682                 dexDecodeInstruction(dpc, &decInsn);
683                 ALOGD("    0x%04x - %s %#x",(dpc-dpcBase),
684                      dexGetOpcodeName(decInsn.opcode),(int)dpc);
685                 dpc += dexGetWidthFromOpcode(decInsn.opcode);
686             }
687             if (trace->trace[curFrag].info.frag.runEnd) {
688                 done = true;
689             }
690         } else {
691             ALOGD("Frag[%d]- META info: 0x%08x", curFrag,
692                  (int)trace->trace[curFrag].info.meta);
693         }
694         curFrag++;
695     }
696     ALOGD("-------------------------------------------");
697 }
698
699 /*
700  * Append the class ptr of "this" and the current method ptr to the current
701  * trace. That is, the trace runs will contain the following components:
702  *  + trace run that ends with an invoke (existing entry)
703  *  + thisClass (new)
704  *  + calleeMethod (new)
705  */
706 static void insertClassMethodInfo(Thread* self,
707                                   const ClassObject* thisClass,
708                                   const Method* calleeMethod,
709                                   const DecodedInstruction* insn)
710 {
711     int currTraceRun = ++self->currTraceRun;
712     self->trace[currTraceRun].info.meta = thisClass ?
713                                     (void *) thisClass->descriptor : NULL;
714     self->trace[currTraceRun].isCode = false;
715
716     currTraceRun = ++self->currTraceRun;
717     self->trace[currTraceRun].info.meta = thisClass ?
718                                     (void *) thisClass->classLoader : NULL;
719     self->trace[currTraceRun].isCode = false;
720
721     currTraceRun = ++self->currTraceRun;
722     self->trace[currTraceRun].info.meta = (void *) calleeMethod;
723     self->trace[currTraceRun].isCode = false;
724 }
725
726 /*
727  * Check if the next instruction following the invoke is a move-result and if
728  * so add it to the trace. That is, this will add the trace run that includes
729  * the move-result to the trace list.
730  *
731  *  + trace run that ends with an invoke (existing entry)
732  *  + thisClass (existing entry)
733  *  + calleeMethod (existing entry)
734  *  + move result (new)
735  *
736  * lastPC, len, offset are all from the preceding invoke instruction
737  */
738 static void insertMoveResult(const u2 *lastPC, int len, int offset,
739                              Thread *self)
740 {
741     DecodedInstruction nextDecInsn;
742     const u2 *moveResultPC = lastPC + len;
743
744     dexDecodeInstruction(moveResultPC, &nextDecInsn);
745     if ((nextDecInsn.opcode != OP_MOVE_RESULT) &&
746         (nextDecInsn.opcode != OP_MOVE_RESULT_WIDE) &&
747         (nextDecInsn.opcode != OP_MOVE_RESULT_OBJECT))
748         return;
749
750     /* We need to start a new trace run */
751     int currTraceRun = ++self->currTraceRun;
752     self->currRunHead = moveResultPC;
753     self->trace[currTraceRun].info.frag.startOffset = offset + len;
754     self->trace[currTraceRun].info.frag.numInsts = 1;
755     self->trace[currTraceRun].info.frag.runEnd = false;
756     self->trace[currTraceRun].info.frag.hint = kJitHintNone;
757     self->trace[currTraceRun].isCode = true;
758     self->totalTraceLen++;
759
760     self->currRunLen = dexGetWidthFromInstruction(moveResultPC);
761 }
762
763 /*
764  * Adds to the current trace request one instruction at a time, just
765  * before that instruction is interpreted.  This is the primary trace
766  * selection function.  NOTE: return instruction are handled a little
767  * differently.  In general, instructions are "proposed" to be added
768  * to the current trace prior to interpretation.  If the interpreter
769  * then successfully completes the instruction, is will be considered
770  * part of the request.  This allows us to examine machine state prior
771  * to interpretation, and also abort the trace request if the instruction
772  * throws or does something unexpected.  However, return instructions
773  * will cause an immediate end to the translation request - which will
774  * be passed to the compiler before the return completes.  This is done
775  * in response to special handling of returns by the interpreter (and
776  * because returns cannot throw in a way that causes problems for the
777  * translated code.
778  */
779 void dvmCheckJit(const u2* pc, Thread* self)
780 {
781     const ClassObject *thisClass = self->callsiteClass;
782     const Method* curMethod = self->methodToCall;
783     int flags, len;
784     int allDone = false;
785     /* Stay in break/single-stop mode for the next instruction */
786     bool stayOneMoreInst = false;
787
788     /* Prepare to handle last PC and stage the current PC & method*/
789     const u2 *lastPC = self->lastPC;
790
791     self->lastPC = pc;
792
793     switch (self->jitState) {
794         int offset;
795         DecodedInstruction decInsn;
796         case kJitTSelect:
797             /* First instruction - just remember the PC and exit */
798             if (lastPC == NULL) break;
799             /* Grow the trace around the last PC if jitState is kJitTSelect */
800             dexDecodeInstruction(lastPC, &decInsn);
801 #if TRACE_OPCODE_FILTER
802             /* Only add JIT support opcode to trace. End the trace if
803              * this opcode is not supported.
804              */
805             if (!dvmIsOpcodeSupportedByJit(decInsn.opcode)) {
806                 self->jitState = kJitTSelectEnd;
807                 break;
808             }
809 #endif
810             /*
811              * Treat {PACKED,SPARSE}_SWITCH as trace-ending instructions due
812              * to the amount of space it takes to generate the chaining
813              * cells.
814              */
815             if (self->totalTraceLen != 0 &&
816                 (decInsn.opcode == OP_PACKED_SWITCH ||
817                  decInsn.opcode == OP_SPARSE_SWITCH)) {
818                 self->jitState = kJitTSelectEnd;
819                 break;
820             }
821
822 #if defined(SHOW_TRACE)
823             ALOGD("TraceGen: adding %s. lpc:%#x, pc:%#x",
824                  dexGetOpcodeName(decInsn.opcode), (int)lastPC, (int)pc);
825 #endif
826             flags = dexGetFlagsFromOpcode(decInsn.opcode);
827             len = dexGetWidthFromInstruction(lastPC);
828             offset = lastPC - self->traceMethod->insns;
829             assert((unsigned) offset <
830                    dvmGetMethodInsnsSize(self->traceMethod));
831             if (lastPC != self->currRunHead + self->currRunLen) {
832                 int currTraceRun;
833                 /* We need to start a new trace run */
834                 currTraceRun = ++self->currTraceRun;
835                 self->currRunLen = 0;
836                 self->currRunHead = (u2*)lastPC;
837                 self->trace[currTraceRun].info.frag.startOffset = offset;
838                 self->trace[currTraceRun].info.frag.numInsts = 0;
839                 self->trace[currTraceRun].info.frag.runEnd = false;
840                 self->trace[currTraceRun].info.frag.hint = kJitHintNone;
841                 self->trace[currTraceRun].isCode = true;
842             }
843             self->trace[self->currTraceRun].info.frag.numInsts++;
844             self->totalTraceLen++;
845             self->currRunLen += len;
846
847             /*
848              * If the last instruction is an invoke, we will try to sneak in
849              * the move-result* (if existent) into a separate trace run.
850              */
851             {
852               int needReservedRun = (flags & kInstrInvoke) ? 1 : 0;
853
854               /* Will probably never hit this with the current trace builder */
855               if (self->currTraceRun ==
856                    (MAX_JIT_RUN_LEN - 1 - needReservedRun)) {
857                 self->jitState = kJitTSelectEnd;
858               }
859             }
860
861             if (!dexIsGoto(flags) &&
862                   ((flags & (kInstrCanBranch |
863                              kInstrCanSwitch |
864                              kInstrCanReturn |
865                              kInstrInvoke)) != 0)) {
866                     self->jitState = kJitTSelectEnd;
867 #if defined(SHOW_TRACE)
868                 ALOGD("TraceGen: ending on %s, basic block end",
869                      dexGetOpcodeName(decInsn.opcode));
870 #endif
871
872                 /*
873                  * If the current invoke is a {virtual,interface}, get the
874                  * current class/method pair into the trace as well.
875                  * If the next instruction is a variant of move-result, insert
876                  * it to the trace too.
877                  */
878                 if (flags & kInstrInvoke) {
879                     insertClassMethodInfo(self, thisClass, curMethod,
880                                           &decInsn);
881                     insertMoveResult(lastPC, len, offset, self);
882                 }
883             }
884             /* Break on throw or self-loop */
885             if ((decInsn.opcode == OP_THROW) || (lastPC == pc)){
886                 self->jitState = kJitTSelectEnd;
887             }
888             if (self->totalTraceLen >= JIT_MAX_TRACE_LEN) {
889                 self->jitState = kJitTSelectEnd;
890             }
891             if ((flags & kInstrCanReturn) != kInstrCanReturn) {
892                 break;
893             }
894             else {
895                 /*
896                  * Last instruction is a return - stay in the dbg interpreter
897                  * for one more instruction if it is a non-void return, since
898                  * we don't want to start a trace with move-result as the first
899                  * instruction (which is already included in the trace
900                  * containing the invoke.
901                  */
902                 if (decInsn.opcode != OP_RETURN_VOID) {
903                     stayOneMoreInst = true;
904                 }
905             }
906             /* NOTE: intentional fallthrough for returns */
907         case kJitTSelectEnd:
908             {
909                 /* Empty trace - set to bail to interpreter */
910                 if (self->totalTraceLen == 0) {
911                     dvmJitSetCodeAddr(self->currTraceHead,
912                                       dvmCompilerGetInterpretTemplate(),
913                                       dvmCompilerGetInterpretTemplateSet(),
914                                       false /* Not method entry */, 0);
915                     self->jitState = kJitDone;
916                     allDone = true;
917                     break;
918                 }
919
920                 int lastTraceDesc = self->currTraceRun;
921
922                 /* Extend a new empty desc if the last slot is meta info */
923                 if (!self->trace[lastTraceDesc].isCode) {
924                     lastTraceDesc = ++self->currTraceRun;
925                     self->trace[lastTraceDesc].info.frag.startOffset = 0;
926                     self->trace[lastTraceDesc].info.frag.numInsts = 0;
927                     self->trace[lastTraceDesc].info.frag.hint = kJitHintNone;
928                     self->trace[lastTraceDesc].isCode = true;
929                 }
930
931                 /* Mark the end of the trace runs */
932                 self->trace[lastTraceDesc].info.frag.runEnd = true;
933
934                 JitTraceDescription* desc =
935                    (JitTraceDescription*)malloc(sizeof(JitTraceDescription) +
936                      sizeof(JitTraceRun) * (self->currTraceRun+1));
937
938                 if (desc == NULL) {
939                     ALOGE("Out of memory in trace selection");
940                     dvmJitStopTranslationRequests();
941                     self->jitState = kJitDone;
942                     allDone = true;
943                     break;
944                 }
945
946                 desc->method = self->traceMethod;
947                 memcpy((char*)&(desc->trace[0]),
948                     (char*)&(self->trace[0]),
949                     sizeof(JitTraceRun) * (self->currTraceRun+1));
950 #if defined(SHOW_TRACE)
951                 ALOGD("TraceGen:  trace done, adding to queue");
952                 dvmJitDumpTraceDesc(desc);
953 #endif
954                 if (dvmCompilerWorkEnqueue(
955                        self->currTraceHead,kWorkOrderTrace,desc)) {
956                     /* Work order successfully enqueued */
957                     if (gDvmJit.blockingMode) {
958                         dvmCompilerDrainQueue();
959                     }
960                 } else {
961                     /*
962                      * Make sure the descriptor for the abandoned work order is
963                      * freed.
964                      */
965                     free(desc);
966                 }
967                 self->jitState = kJitDone;
968                 allDone = true;
969             }
970             break;
971         case kJitDone:
972             allDone = true;
973             break;
974         case kJitNot:
975             allDone = true;
976             break;
977         default:
978             ALOGE("Unexpected JIT state: %d", self->jitState);
979             dvmAbort();
980             break;
981     }
982
983     /*
984      * If we're done with trace selection, switch off the control flags.
985      */
986      if (allDone) {
987          dvmDisableSubMode(self, kSubModeJitTraceBuild);
988          if (stayOneMoreInst) {
989              // Clear jitResumeNPC explicitly since we know we don't need it
990              // here.
991              self->jitResumeNPC = NULL;
992              // Keep going in single-step mode for at least one more inst
993              if (self->singleStepCount == 0)
994                  self->singleStepCount = 1;
995              dvmEnableSubMode(self, kSubModeCountedStep);
996          }
997      }
998      return;
999 }
1000
1001 JitEntry *dvmJitFindEntry(const u2* pc, bool isMethodEntry)
1002 {
1003     int idx = dvmJitHash(pc);
1004
1005     /* Expect a high hit rate on 1st shot */
1006     if ((gDvmJit.pJitEntryTable[idx].dPC == pc) &&
1007         (gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry == isMethodEntry))
1008         return &gDvmJit.pJitEntryTable[idx];
1009     else {
1010         int chainEndMarker = gDvmJit.jitTableSize;
1011         while (gDvmJit.pJitEntryTable[idx].u.info.chain != chainEndMarker) {
1012             idx = gDvmJit.pJitEntryTable[idx].u.info.chain;
1013             if ((gDvmJit.pJitEntryTable[idx].dPC == pc) &&
1014                 (gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry ==
1015                 isMethodEntry))
1016                 return &gDvmJit.pJitEntryTable[idx];
1017         }
1018     }
1019     return NULL;
1020 }
1021
1022 /*
1023  * Walk through the JIT profile table and find the corresponding JIT code, in
1024  * the specified format (ie trace vs method). This routine needs to be fast.
1025  */
1026 void* getCodeAddrCommon(const u2* dPC, bool methodEntry)
1027 {
1028     int idx = dvmJitHash(dPC);
1029     const u2* pc = gDvmJit.pJitEntryTable[idx].dPC;
1030     if (pc != NULL) {
1031         bool hideTranslation = dvmJitHideTranslation();
1032         if (pc == dPC &&
1033             gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry == methodEntry) {
1034             int offset = (gDvmJit.profileMode >= kTraceProfilingContinuous) ?
1035                  0 : gDvmJit.pJitEntryTable[idx].u.info.profileOffset;
1036             intptr_t codeAddress =
1037                 (intptr_t)gDvmJit.pJitEntryTable[idx].codeAddress;
1038 #if defined(WITH_JIT_TUNING)
1039             gDvmJit.addrLookupsFound++;
1040 #endif
1041             return hideTranslation || !codeAddress ?  NULL :
1042                   (void *)(codeAddress + offset);
1043         } else {
1044             int chainEndMarker = gDvmJit.jitTableSize;
1045             while (gDvmJit.pJitEntryTable[idx].u.info.chain != chainEndMarker) {
1046                 idx = gDvmJit.pJitEntryTable[idx].u.info.chain;
1047                 if (gDvmJit.pJitEntryTable[idx].dPC == dPC &&
1048                     gDvmJit.pJitEntryTable[idx].u.info.isMethodEntry ==
1049                         methodEntry) {
1050                     int offset = (gDvmJit.profileMode >=
1051                         kTraceProfilingContinuous) ? 0 :
1052                         gDvmJit.pJitEntryTable[idx].u.info.profileOffset;
1053                     intptr_t codeAddress =
1054                         (intptr_t)gDvmJit.pJitEntryTable[idx].codeAddress;
1055 #if defined(WITH_JIT_TUNING)
1056                     gDvmJit.addrLookupsFound++;
1057 #endif
1058                     return hideTranslation || !codeAddress ? NULL :
1059                         (void *)(codeAddress + offset);
1060                 }
1061             }
1062         }
1063     }
1064 #if defined(WITH_JIT_TUNING)
1065     gDvmJit.addrLookupsNotFound++;
1066 #endif
1067     return NULL;
1068 }
1069
1070 /*
1071  * If a translated code address, in trace format, exists for the davik byte code
1072  * pointer return it.
1073  */
1074 void* dvmJitGetTraceAddr(const u2* dPC)
1075 {
1076     return getCodeAddrCommon(dPC, false /* method entry */);
1077 }
1078
1079 /*
1080  * If a translated code address, in whole-method format, exists for the davik
1081  * byte code pointer return it.
1082  */
1083 void* dvmJitGetMethodAddr(const u2* dPC)
1084 {
1085     return getCodeAddrCommon(dPC, true /* method entry */);
1086 }
1087
1088 /*
1089  * Similar to dvmJitGetTraceAddr, but returns null if the calling
1090  * thread is in a single-step mode.
1091  */
1092 void* dvmJitGetTraceAddrThread(const u2* dPC, Thread* self)
1093 {
1094     return (self->interpBreak.ctl.breakFlags != 0) ? NULL :
1095             getCodeAddrCommon(dPC, false /* method entry */);
1096 }
1097
1098 /*
1099  * Similar to dvmJitGetMethodAddr, but returns null if the calling
1100  * thread is in a single-step mode.
1101  */
1102 void* dvmJitGetMethodAddrThread(const u2* dPC, Thread* self)
1103 {
1104     return (self->interpBreak.ctl.breakFlags != 0) ? NULL :
1105             getCodeAddrCommon(dPC, true /* method entry */);
1106 }
1107
1108 /*
1109  * Register the translated code pointer into the JitTable.
1110  * NOTE: Once a codeAddress field transitions from initial state to
1111  * JIT'd code, it must not be altered without first halting all
1112  * threads.  We defer the setting of the profile prefix size until
1113  * after the new code address is set to ensure that the prefix offset
1114  * is never applied to the initial interpret-only translation.  All
1115  * translations with non-zero profile prefixes will still be correct
1116  * if entered as if the profile offset is 0, but the interpret-only
1117  * template cannot handle a non-zero prefix.
1118  * NOTE: JitTable must not be in danger of reset while this
1119  * code is executing. see Issue 4271784 for details.
1120  */
1121 void dvmJitSetCodeAddr(const u2* dPC, void *nPC, JitInstructionSetType set,
1122                        bool isMethodEntry, int profilePrefixSize)
1123 {
1124     JitEntryInfoUnion oldValue;
1125     JitEntryInfoUnion newValue;
1126     /*
1127      * Get the JitTable slot for this dPC (or create one if JitTable
1128      * has been reset between the time the trace was requested and
1129      * now.
1130      */
1131     JitEntry *jitEntry = isMethodEntry ?
1132         lookupAndAdd(dPC, false /* caller holds tableLock */, isMethodEntry) :
1133                      dvmJitFindEntry(dPC, isMethodEntry);
1134     assert(jitEntry);
1135     /* Note: order of update is important */
1136     do {
1137         oldValue = jitEntry->u;
1138         newValue = oldValue;
1139         newValue.info.isMethodEntry = isMethodEntry;
1140         newValue.info.instructionSet = set;
1141         newValue.info.profileOffset = profilePrefixSize;
1142     } while (android_atomic_release_cas(
1143              oldValue.infoWord, newValue.infoWord,
1144              &jitEntry->u.infoWord) != 0);
1145     jitEntry->codeAddress = nPC;
1146 }
1147
1148 /*
1149  * Determine if valid trace-bulding request is active.  If so, set
1150  * the proper flags in interpBreak and return.  Trace selection will
1151  * then begin normally via dvmCheckBefore.
1152  */
1153 void dvmJitCheckTraceRequest(Thread* self)
1154 {
1155     int i;
1156     /*
1157      * A note on trace "hotness" filtering:
1158      *
1159      * Our first level trigger is intentionally loose - we need it to
1160      * fire easily not just to identify potential traces to compile, but
1161      * also to allow re-entry into the code cache.
1162      *
1163      * The 2nd level filter (done here) exists to be selective about
1164      * what we actually compile.  It works by requiring the same
1165      * trace head "key" (defined as filterKey below) to appear twice in
1166      * a relatively short period of time.   The difficulty is defining the
1167      * shape of the filterKey.  Unfortunately, there is no "one size fits
1168      * all" approach.
1169      *
1170      * For spiky execution profiles dominated by a smallish
1171      * number of very hot loops, we would want the second-level filter
1172      * to be very selective.  A good selective filter is requiring an
1173      * exact match of the Dalvik PC.  In other words, defining filterKey as:
1174      *     intptr_t filterKey = (intptr_t)self->interpSave.pc
1175      *
1176      * However, for flat execution profiles we do best when aggressively
1177      * translating.  A heuristically decent proxy for this is to use
1178      * the value of the method pointer containing the trace as the filterKey.
1179      * Intuitively, this is saying that once any trace in a method appears hot,
1180      * immediately translate any other trace from that same method that
1181      * survives the first-level filter.  Here, filterKey would be defined as:
1182      *     intptr_t filterKey = (intptr_t)self->interpSave.method
1183      *
1184      * The problem is that we can't easily detect whether we're dealing
1185      * with a spiky or flat profile.  If we go with the "pc" match approach,
1186      * flat profiles perform poorly.  If we go with the loose "method" match,
1187      * we end up generating a lot of useless translations.  Probably the
1188      * best approach in the future will be to retain profile information
1189      * across runs of each application in order to determine it's profile,
1190      * and then choose once we have enough history.
1191      *
1192      * However, for now we've decided to chose a compromise filter scheme that
1193      * includes elements of both.  The high order bits of the filter key
1194      * are drawn from the enclosing method, and are combined with a slice
1195      * of the low-order bits of the Dalvik pc of the trace head.  The
1196      * looseness of the filter can be adjusted by changing with width of
1197      * the Dalvik pc slice (JIT_TRACE_THRESH_FILTER_PC_BITS).  The wider
1198      * the slice, the tighter the filter.
1199      *
1200      * Note: the fixed shifts in the function below reflect assumed word
1201      * alignment for method pointers, and half-word alignment of the Dalvik pc.
1202      * for method pointers and half-word alignment for dalvik pc.
1203      */
1204     u4 methodKey = (u4)self->interpSave.method <<
1205                    (JIT_TRACE_THRESH_FILTER_PC_BITS - 2);
1206     u4 pcKey = ((u4)self->interpSave.pc >> 1) &
1207                ((1 << JIT_TRACE_THRESH_FILTER_PC_BITS) - 1);
1208     intptr_t filterKey = (intptr_t)(methodKey | pcKey);
1209
1210     // Shouldn't be here if already building a trace.
1211     assert((self->interpBreak.ctl.subMode & kSubModeJitTraceBuild)==0);
1212
1213     /* Check if the JIT request can be handled now */
1214     if ((gDvmJit.pJitEntryTable != NULL) &&
1215         ((self->interpBreak.ctl.breakFlags & kInterpSingleStep) == 0)){
1216         /* Bypass the filter for hot trace requests or during stress mode */
1217         if (self->jitState == kJitTSelectRequest &&
1218             gDvmJit.threshold > 6) {
1219             /* Two-level filtering scheme */
1220             for (i=0; i< JIT_TRACE_THRESH_FILTER_SIZE; i++) {
1221                 if (filterKey == self->threshFilter[i]) {
1222                     self->threshFilter[i] = 0; // Reset filter entry
1223                     break;
1224                 }
1225             }
1226             if (i == JIT_TRACE_THRESH_FILTER_SIZE) {
1227                 /*
1228                  * Use random replacement policy - otherwise we could miss a
1229                  * large loop that contains more traces than the size of our
1230                  * filter array.
1231                  */
1232                 i = rand() % JIT_TRACE_THRESH_FILTER_SIZE;
1233                 self->threshFilter[i] = filterKey;
1234                 self->jitState = kJitDone;
1235             }
1236         }
1237
1238         /* If the compiler is backlogged, cancel any JIT actions */
1239         if (gDvmJit.compilerQueueLength >= gDvmJit.compilerHighWater) {
1240             self->jitState = kJitDone;
1241         }
1242
1243         /*
1244          * Check for additional reasons that might force the trace select
1245          * request to be dropped
1246          */
1247         if (self->jitState == kJitTSelectRequest ||
1248             self->jitState == kJitTSelectRequestHot) {
1249             if (dvmJitFindEntry(self->interpSave.pc, false)) {
1250                 /* In progress - nothing do do */
1251                self->jitState = kJitDone;
1252             } else {
1253                 JitEntry *slot = lookupAndAdd(self->interpSave.pc,
1254                                               false /* lock */,
1255                                               false /* method entry */);
1256                 if (slot == NULL) {
1257                     /*
1258                      * Table is full.  This should have been
1259                      * detected by the compiler thread and the table
1260                      * resized before we run into it here.  Assume bad things
1261                      * are afoot and disable profiling.
1262                      */
1263                     self->jitState = kJitDone;
1264                     ALOGD("JIT: JitTable full, disabling profiling");
1265                     dvmJitStopTranslationRequests();
1266                 }
1267             }
1268         }
1269
1270         switch (self->jitState) {
1271             case kJitTSelectRequest:
1272             case kJitTSelectRequestHot:
1273                 self->jitState = kJitTSelect;
1274                 self->traceMethod = self->interpSave.method;
1275                 self->currTraceHead = self->interpSave.pc;
1276                 self->currTraceRun = 0;
1277                 self->totalTraceLen = 0;
1278                 self->currRunHead = self->interpSave.pc;
1279                 self->currRunLen = 0;
1280                 self->trace[0].info.frag.startOffset =
1281                      self->interpSave.pc - self->interpSave.method->insns;
1282                 self->trace[0].info.frag.numInsts = 0;
1283                 self->trace[0].info.frag.runEnd = false;
1284                 self->trace[0].info.frag.hint = kJitHintNone;
1285                 self->trace[0].isCode = true;
1286                 self->lastPC = 0;
1287                 /* Turn on trace selection mode */
1288                 dvmEnableSubMode(self, kSubModeJitTraceBuild);
1289 #if defined(SHOW_TRACE)
1290                 ALOGD("Starting trace for %s at %#x",
1291                      self->interpSave.method->name, (int)self->interpSave.pc);
1292 #endif
1293                 break;
1294             case kJitDone:
1295                 break;
1296             default:
1297                 ALOGE("Unexpected JIT state: %d", self->jitState);
1298                 dvmAbort();
1299         }
1300     } else {
1301         /* Cannot build trace this time */
1302         self->jitState = kJitDone;
1303     }
1304 }
1305
1306 /*
1307  * Resizes the JitTable.  Must be a power of 2, and returns true on failure.
1308  * Stops all threads, and thus is a heavyweight operation. May only be called
1309  * by the compiler thread.
1310  */
1311 bool dvmJitResizeJitTable( unsigned int size )
1312 {
1313     JitEntry *pNewTable;
1314     JitEntry *pOldTable;
1315     JitEntry tempEntry;
1316     unsigned int oldSize;
1317     unsigned int i;
1318
1319     assert(gDvmJit.pJitEntryTable != NULL);
1320     assert(size && !(size & (size - 1)));   /* Is power of 2? */
1321
1322     ALOGI("Jit: resizing JitTable from %d to %d", gDvmJit.jitTableSize, size);
1323
1324     if (size <= gDvmJit.jitTableSize) {
1325         return true;
1326     }
1327
1328     /* Make sure requested size is compatible with chain field width */
1329     tempEntry.u.info.chain = size;
1330     if (tempEntry.u.info.chain != size) {
1331         ALOGD("Jit: JitTable request of %d too big", size);
1332         return true;
1333     }
1334
1335     pNewTable = (JitEntry*)calloc(size, sizeof(*pNewTable));
1336     if (pNewTable == NULL) {
1337         return true;
1338     }
1339     for (i=0; i< size; i++) {
1340         pNewTable[i].u.info.chain = size;  /* Initialize chain termination */
1341     }
1342
1343     /* Stop all other interpreting/jit'ng threads */
1344     dvmSuspendAllThreads(SUSPEND_FOR_TBL_RESIZE);
1345
1346     pOldTable = gDvmJit.pJitEntryTable;
1347     oldSize = gDvmJit.jitTableSize;
1348
1349     dvmLockMutex(&gDvmJit.tableLock);
1350     gDvmJit.pJitEntryTable = pNewTable;
1351     gDvmJit.jitTableSize = size;
1352     gDvmJit.jitTableMask = size - 1;
1353     gDvmJit.jitTableEntriesUsed = 0;
1354
1355     for (i=0; i < oldSize; i++) {
1356         if (pOldTable[i].dPC) {
1357             JitEntry *p;
1358             u2 chain;
1359             p = lookupAndAdd(pOldTable[i].dPC, true /* holds tableLock*/,
1360                              pOldTable[i].u.info.isMethodEntry);
1361             p->codeAddress = pOldTable[i].codeAddress;
1362             /* We need to preserve the new chain field, but copy the rest */
1363             chain = p->u.info.chain;
1364             p->u = pOldTable[i].u;
1365             p->u.info.chain = chain;
1366         }
1367     }
1368
1369     dvmUnlockMutex(&gDvmJit.tableLock);
1370
1371     free(pOldTable);
1372
1373     /* Restart the world */
1374     dvmResumeAllThreads(SUSPEND_FOR_TBL_RESIZE);
1375
1376     return false;
1377 }
1378
1379 /*
1380  * Reset the JitTable to the initial clean state.
1381  */
1382 void dvmJitResetTable()
1383 {
1384     JitEntry *jitEntry = gDvmJit.pJitEntryTable;
1385     unsigned int size = gDvmJit.jitTableSize;
1386     unsigned int i;
1387
1388     dvmLockMutex(&gDvmJit.tableLock);
1389
1390     /* Note: If need to preserve any existing counts. Do so here. */
1391     if (gDvmJit.pJitTraceProfCounters) {
1392         for (i=0; i < JIT_PROF_BLOCK_BUCKETS; i++) {
1393             if (gDvmJit.pJitTraceProfCounters->buckets[i])
1394                 memset((void *) gDvmJit.pJitTraceProfCounters->buckets[i],
1395                        0, sizeof(JitTraceCounter_t) * JIT_PROF_BLOCK_ENTRIES);
1396         }
1397         gDvmJit.pJitTraceProfCounters->next = 0;
1398     }
1399
1400     memset((void *) jitEntry, 0, sizeof(JitEntry) * size);
1401     for (i=0; i< size; i++) {
1402         jitEntry[i].u.info.chain = size;  /* Initialize chain termination */
1403     }
1404     gDvmJit.jitTableEntriesUsed = 0;
1405     dvmUnlockMutex(&gDvmJit.tableLock);
1406 }
1407
1408 /*
1409  * Return the address of the next trace profile counter.  This address
1410  * will be embedded in the generated code for the trace, and thus cannot
1411  * change while the trace exists.
1412  */
1413 JitTraceCounter_t *dvmJitNextTraceCounter()
1414 {
1415     int idx = gDvmJit.pJitTraceProfCounters->next / JIT_PROF_BLOCK_ENTRIES;
1416     int elem = gDvmJit.pJitTraceProfCounters->next % JIT_PROF_BLOCK_ENTRIES;
1417     JitTraceCounter_t *res;
1418     /* Lazily allocate blocks of counters */
1419     if (!gDvmJit.pJitTraceProfCounters->buckets[idx]) {
1420         JitTraceCounter_t *p =
1421               (JitTraceCounter_t*) calloc(JIT_PROF_BLOCK_ENTRIES, sizeof(*p));
1422         if (!p) {
1423             ALOGE("Failed to allocate block of trace profile counters");
1424             dvmAbort();
1425         }
1426         gDvmJit.pJitTraceProfCounters->buckets[idx] = p;
1427     }
1428     res = &gDvmJit.pJitTraceProfCounters->buckets[idx][elem];
1429     gDvmJit.pJitTraceProfCounters->next++;
1430     return res;
1431 }
1432
1433 /*
1434  * Float/double conversion requires clamping to min and max of integer form.  If
1435  * target doesn't support this normally, use these.
1436  */
1437 s8 dvmJitd2l(double d)
1438 {
1439     static const double kMaxLong = (double)(s8)0x7fffffffffffffffULL;
1440     static const double kMinLong = (double)(s8)0x8000000000000000ULL;
1441     if (d >= kMaxLong)
1442         return (s8)0x7fffffffffffffffULL;
1443     else if (d <= kMinLong)
1444         return (s8)0x8000000000000000ULL;
1445     else if (d != d) // NaN case
1446         return 0;
1447     else
1448         return (s8)d;
1449 }
1450
1451 s8 dvmJitf2l(float f)
1452 {
1453     static const float kMaxLong = (float)(s8)0x7fffffffffffffffULL;
1454     static const float kMinLong = (float)(s8)0x8000000000000000ULL;
1455     if (f >= kMaxLong)
1456         return (s8)0x7fffffffffffffffULL;
1457     else if (f <= kMinLong)
1458         return (s8)0x8000000000000000ULL;
1459     else if (f != f) // NaN case
1460         return 0;
1461     else
1462         return (s8)f;
1463 }
1464
1465 /* Should only be called by the compiler thread */
1466 void dvmJitChangeProfileMode(TraceProfilingModes newState)
1467 {
1468     if (gDvmJit.profileMode != newState) {
1469         gDvmJit.profileMode = newState;
1470         dvmJitUnchainAll();
1471     }
1472 }
1473
1474 void dvmJitTraceProfilingOn()
1475 {
1476     if (gDvmJit.profileMode == kTraceProfilingPeriodicOff)
1477         dvmCompilerForceWorkEnqueue(NULL, kWorkOrderProfileMode,
1478                                     (void*) kTraceProfilingPeriodicOn);
1479     else if (gDvmJit.profileMode == kTraceProfilingDisabled)
1480         dvmCompilerForceWorkEnqueue(NULL, kWorkOrderProfileMode,
1481                                     (void*) kTraceProfilingContinuous);
1482 }
1483
1484 void dvmJitTraceProfilingOff()
1485 {
1486     if (gDvmJit.profileMode == kTraceProfilingPeriodicOn)
1487         dvmCompilerForceWorkEnqueue(NULL, kWorkOrderProfileMode,
1488                                     (void*) kTraceProfilingPeriodicOff);
1489     else if (gDvmJit.profileMode == kTraceProfilingContinuous)
1490         dvmCompilerForceWorkEnqueue(NULL, kWorkOrderProfileMode,
1491                                     (void*) kTraceProfilingDisabled);
1492 }
1493
1494 /*
1495  * Update JIT-specific info in Thread structure for a single thread
1496  */
1497 void dvmJitUpdateThreadStateSingle(Thread* thread)
1498 {
1499     thread->pJitProfTable = gDvmJit.pProfTable;
1500     thread->jitThreshold = gDvmJit.threshold;
1501 }
1502
1503 /*
1504  * Walk through the thread list and refresh all local copies of
1505  * JIT global state (which was placed there for fast access).
1506  */
1507 void dvmJitUpdateThreadStateAll()
1508 {
1509     Thread* self = dvmThreadSelf();
1510     Thread* thread;
1511
1512     dvmLockThreadList(self);
1513     for (thread = gDvm.threadList; thread != NULL; thread = thread->next) {
1514         dvmJitUpdateThreadStateSingle(thread);
1515     }
1516     dvmUnlockThreadList();
1517
1518 }
1519 #endif /* WITH_JIT */