blob: f025c0a2d403d1dddf06cc876e72ab6deedc4bfb [file] [log] [blame]
xueliang.zhongf7caf682017-03-01 16:07:02 +00001/*
2 * Copyright (C) 2017 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17#include "arch/arm/instruction_set_features_arm.h"
18#include "code_generator_utils.h"
19#include "common_arm.h"
20#include "mirror/array-inl.h"
21#include "scheduler_arm.h"
22
23namespace art {
24namespace arm {
25
26using helpers::Int32ConstantFrom;
27using helpers::Uint64ConstantFrom;
28
29void SchedulingLatencyVisitorARM::HandleBinaryOperationLantencies(HBinaryOperation* instr) {
30 switch (instr->GetResultType()) {
31 case Primitive::kPrimLong:
32 // HAdd and HSub long operations translate to ADDS+ADC or SUBS+SBC pairs,
33 // so a bubble (kArmNopLatency) is added to represent the internal carry flag
34 // dependency inside these pairs.
35 last_visited_internal_latency_ = kArmIntegerOpLatency + kArmNopLatency;
36 last_visited_latency_ = kArmIntegerOpLatency;
37 break;
38 case Primitive::kPrimFloat:
39 case Primitive::kPrimDouble:
40 last_visited_latency_ = kArmFloatingPointOpLatency;
41 break;
42 default:
43 last_visited_latency_ = kArmIntegerOpLatency;
44 break;
45 }
46}
47
48void SchedulingLatencyVisitorARM::VisitAdd(HAdd* instr) {
49 HandleBinaryOperationLantencies(instr);
50}
51
52void SchedulingLatencyVisitorARM::VisitSub(HSub* instr) {
53 HandleBinaryOperationLantencies(instr);
54}
55
56void SchedulingLatencyVisitorARM::VisitMul(HMul* instr) {
57 switch (instr->GetResultType()) {
58 case Primitive::kPrimLong:
59 last_visited_internal_latency_ = 3 * kArmMulIntegerLatency;
60 last_visited_latency_ = kArmIntegerOpLatency;
61 break;
62 case Primitive::kPrimFloat:
63 case Primitive::kPrimDouble:
64 last_visited_latency_ = kArmMulFloatingPointLatency;
65 break;
66 default:
67 last_visited_latency_ = kArmMulIntegerLatency;
68 break;
69 }
70}
71
72void SchedulingLatencyVisitorARM::HandleBitwiseOperationLantencies(HBinaryOperation* instr) {
73 switch (instr->GetResultType()) {
74 case Primitive::kPrimLong:
75 last_visited_internal_latency_ = kArmIntegerOpLatency;
76 last_visited_latency_ = kArmIntegerOpLatency;
77 break;
78 case Primitive::kPrimFloat:
79 case Primitive::kPrimDouble:
80 last_visited_latency_ = kArmFloatingPointOpLatency;
81 break;
82 default:
83 last_visited_latency_ = kArmIntegerOpLatency;
84 break;
85 }
86}
87
88void SchedulingLatencyVisitorARM::VisitAnd(HAnd* instr) {
89 HandleBitwiseOperationLantencies(instr);
90}
91
92void SchedulingLatencyVisitorARM::VisitOr(HOr* instr) {
93 HandleBitwiseOperationLantencies(instr);
94}
95
96void SchedulingLatencyVisitorARM::VisitXor(HXor* instr) {
97 HandleBitwiseOperationLantencies(instr);
98}
99
100void SchedulingLatencyVisitorARM::VisitRor(HRor* instr) {
101 switch (instr->GetResultType()) {
102 case Primitive::kPrimInt:
103 last_visited_latency_ = kArmIntegerOpLatency;
104 break;
105 case Primitive::kPrimLong: {
106 // HandleLongRotate
107 HInstruction* rhs = instr->GetRight();
108 if (rhs->IsConstant()) {
109 uint64_t rot = Uint64ConstantFrom(rhs->AsConstant()) & kMaxLongShiftDistance;
110 if (rot != 0u) {
111 last_visited_internal_latency_ = 3 * kArmIntegerOpLatency;
112 last_visited_latency_ = kArmIntegerOpLatency;
113 } else {
114 last_visited_internal_latency_ = kArmIntegerOpLatency;
115 last_visited_latency_ = kArmIntegerOpLatency;
116 }
117 } else {
118 last_visited_internal_latency_ = 9 * kArmIntegerOpLatency + kArmBranchLatency;
119 last_visited_latency_ = kArmBranchLatency;
120 }
121 break;
122 }
123 default:
124 LOG(FATAL) << "Unexpected operation type " << instr->GetResultType();
125 UNREACHABLE();
126 }
127}
128
129void SchedulingLatencyVisitorARM::HandleShiftLatencies(HBinaryOperation* instr) {
130 Primitive::Type type = instr->GetResultType();
131 HInstruction* rhs = instr->GetRight();
132 switch (type) {
133 case Primitive::kPrimInt:
134 if (!rhs->IsConstant()) {
135 last_visited_internal_latency_ = kArmIntegerOpLatency;
136 }
137 last_visited_latency_ = kArmIntegerOpLatency;
138 break;
139 case Primitive::kPrimLong:
140 if (!rhs->IsConstant()) {
141 last_visited_internal_latency_ = 8 * kArmIntegerOpLatency;
142 } else {
143 uint32_t shift_value = Int32ConstantFrom(rhs->AsConstant()) & kMaxLongShiftDistance;
144 if (shift_value == 1 || shift_value >= 32) {
145 last_visited_internal_latency_ = kArmIntegerOpLatency;
146 } else {
147 last_visited_internal_latency_ = 2 * kArmIntegerOpLatency;
148 }
149 }
150 last_visited_latency_ = kArmIntegerOpLatency;
151 break;
152 default:
153 LOG(FATAL) << "Unexpected operation type " << type;
154 UNREACHABLE();
155 }
156}
157
158void SchedulingLatencyVisitorARM::VisitShl(HShl* instr) {
159 HandleShiftLatencies(instr);
160}
161
162void SchedulingLatencyVisitorARM::VisitShr(HShr* instr) {
163 HandleShiftLatencies(instr);
164}
165
166void SchedulingLatencyVisitorARM::VisitUShr(HUShr* instr) {
167 HandleShiftLatencies(instr);
168}
169
xueliang.zhongbf9e21a2017-06-15 11:01:11 +0100170void SchedulingLatencyVisitorARM::HandleGenerateConditionWithZero(IfCondition condition) {
171 switch (condition) {
172 case kCondEQ:
173 case kCondBE:
174 case kCondNE:
175 case kCondA:
176 last_visited_internal_latency_ += kArmIntegerOpLatency;
177 last_visited_latency_ = kArmIntegerOpLatency;
xueliang.zhongf7caf682017-03-01 16:07:02 +0000178 break;
xueliang.zhongbf9e21a2017-06-15 11:01:11 +0100179 case kCondGE:
180 // Mvn
181 last_visited_internal_latency_ += kArmIntegerOpLatency;
182 FALLTHROUGH_INTENDED;
183 case kCondLT:
184 // Lsr
185 last_visited_latency_ = kArmIntegerOpLatency;
186 break;
187 case kCondAE:
188 // Trivially true.
189 // Mov
190 last_visited_latency_ = kArmIntegerOpLatency;
191 break;
192 case kCondB:
193 // Trivially false.
194 // Mov
195 last_visited_latency_ = kArmIntegerOpLatency;
xueliang.zhongf7caf682017-03-01 16:07:02 +0000196 break;
197 default:
xueliang.zhongbf9e21a2017-06-15 11:01:11 +0100198 LOG(FATAL) << "Unexpected condition " << condition;
199 UNREACHABLE();
xueliang.zhongf7caf682017-03-01 16:07:02 +0000200 }
xueliang.zhongbf9e21a2017-06-15 11:01:11 +0100201}
202
203void SchedulingLatencyVisitorARM::HandleGenerateLongTestConstant(HCondition* condition) {
204 DCHECK_EQ(condition->GetLeft()->GetType(), Primitive::kPrimLong);
205
206 IfCondition cond = condition->GetCondition();
207
208 HInstruction* right = condition->InputAt(1);
209
210 int64_t value = Uint64ConstantFrom(right);
211
212 // Comparisons against 0 are common enough, so codegen has special handling for them.
213 if (value == 0) {
214 switch (cond) {
215 case kCondNE:
216 case kCondA:
217 case kCondEQ:
218 case kCondBE:
219 // Orrs
220 last_visited_internal_latency_ += kArmIntegerOpLatency;
221 return;
222 case kCondLT:
223 case kCondGE:
224 // Cmp
225 last_visited_internal_latency_ += kArmIntegerOpLatency;
226 return;
227 case kCondB:
228 case kCondAE:
229 // Cmp
230 last_visited_internal_latency_ += kArmIntegerOpLatency;
231 return;
232 default:
233 break;
234 }
235 }
236
237 switch (cond) {
238 case kCondEQ:
239 case kCondNE:
240 case kCondB:
241 case kCondBE:
242 case kCondA:
243 case kCondAE: {
244 // Cmp, IT, Cmp
245 last_visited_internal_latency_ += 3 * kArmIntegerOpLatency;
246 break;
247 }
248 case kCondLE:
249 case kCondGT:
250 // Trivially true or false.
251 if (value == std::numeric_limits<int64_t>::max()) {
252 // Cmp
253 last_visited_internal_latency_ += kArmIntegerOpLatency;
254 break;
255 }
256 FALLTHROUGH_INTENDED;
257 case kCondGE:
258 case kCondLT: {
259 // Cmp, Sbcs
260 last_visited_internal_latency_ += 2 * kArmIntegerOpLatency;
261 break;
262 }
263 default:
264 LOG(FATAL) << "Unreachable";
265 UNREACHABLE();
266 }
267}
268
269void SchedulingLatencyVisitorARM::HandleGenerateLongTest(HCondition* condition) {
270 DCHECK_EQ(condition->GetLeft()->GetType(), Primitive::kPrimLong);
271
272 IfCondition cond = condition->GetCondition();
273
274 switch (cond) {
275 case kCondEQ:
276 case kCondNE:
277 case kCondB:
278 case kCondBE:
279 case kCondA:
280 case kCondAE: {
281 // Cmp, IT, Cmp
282 last_visited_internal_latency_ += 3 * kArmIntegerOpLatency;
283 break;
284 }
285 case kCondLE:
286 case kCondGT:
287 case kCondGE:
288 case kCondLT: {
289 // Cmp, Sbcs
290 last_visited_internal_latency_ += 2 * kArmIntegerOpLatency;
291 break;
292 }
293 default:
294 LOG(FATAL) << "Unreachable";
295 UNREACHABLE();
296 }
297}
298
299// The GenerateTest series of function all counted as internal latency.
300void SchedulingLatencyVisitorARM::HandleGenerateTest(HCondition* condition) {
301 const Primitive::Type type = condition->GetLeft()->GetType();
302
303 if (type == Primitive::kPrimLong) {
304 condition->InputAt(1)->IsConstant()
305 ? HandleGenerateLongTestConstant(condition)
306 : HandleGenerateLongTest(condition);
307 } else if (Primitive::IsFloatingPointType(type)) {
308 // GenerateVcmp + Vmrs
309 last_visited_internal_latency_ += 2 * kArmFloatingPointOpLatency;
310 } else {
311 // Cmp
312 last_visited_internal_latency_ += kArmIntegerOpLatency;
313 }
314}
315
316bool SchedulingLatencyVisitorARM::CanGenerateTest(HCondition* condition) {
317 if (condition->GetLeft()->GetType() == Primitive::kPrimLong) {
318 HInstruction* right = condition->InputAt(1);
319
320 if (right->IsConstant()) {
321 IfCondition c = condition->GetCondition();
322 const uint64_t value = Uint64ConstantFrom(right);
323
324 if (c < kCondLT || c > kCondGE) {
325 if (value != 0) {
326 return false;
327 }
328 } else if (c == kCondLE || c == kCondGT) {
329 if (value < std::numeric_limits<int64_t>::max() &&
330 !codegen_->GetAssembler()->ShifterOperandCanHold(SBC, High32Bits(value + 1), kCcSet)) {
331 return false;
332 }
333 } else if (!codegen_->GetAssembler()->ShifterOperandCanHold(SBC, High32Bits(value), kCcSet)) {
334 return false;
335 }
336 }
337 }
338
339 return true;
340}
341
342void SchedulingLatencyVisitorARM::HandleGenerateConditionGeneric(HCondition* cond) {
343 HandleGenerateTest(cond);
344
345 // Unlike codegen pass, we cannot check 'out' register IsLow() here,
346 // because scheduling is before liveness(location builder) and register allocator,
347 // so we can only choose to follow one path of codegen by assuming otu.IsLow() is true.
348 last_visited_internal_latency_ += 2 * kArmIntegerOpLatency;
xueliang.zhongf7caf682017-03-01 16:07:02 +0000349 last_visited_latency_ = kArmIntegerOpLatency;
350}
351
xueliang.zhongbf9e21a2017-06-15 11:01:11 +0100352void SchedulingLatencyVisitorARM::HandleGenerateEqualLong(HCondition* cond) {
353 DCHECK_EQ(cond->GetLeft()->GetType(), Primitive::kPrimLong);
354
355 IfCondition condition = cond->GetCondition();
356
357 last_visited_internal_latency_ += 2 * kArmIntegerOpLatency;
358
359 if (condition == kCondNE) {
360 // Orrs, IT, Mov
361 last_visited_internal_latency_ += 3 * kArmIntegerOpLatency;
362 } else {
363 last_visited_internal_latency_ += kArmIntegerOpLatency;
364 HandleGenerateConditionWithZero(condition);
365 }
366}
367
368void SchedulingLatencyVisitorARM::HandleGenerateLongComparesAndJumps() {
369 last_visited_internal_latency_ += 4 * kArmIntegerOpLatency;
370 last_visited_internal_latency_ += kArmBranchLatency;
371}
372
373void SchedulingLatencyVisitorARM::HandleGenerateConditionLong(HCondition* cond) {
374 DCHECK_EQ(cond->GetLeft()->GetType(), Primitive::kPrimLong);
375
376 IfCondition condition = cond->GetCondition();
377 HInstruction* right = cond->InputAt(1);
378
379 if (right->IsConstant()) {
380 // Comparisons against 0 are common enough, so codegen has special handling for them.
381 if (Uint64ConstantFrom(right) == 0) {
382 switch (condition) {
383 case kCondNE:
384 case kCondA:
385 case kCondEQ:
386 case kCondBE:
387 // Orr
388 last_visited_internal_latency_ += kArmIntegerOpLatency;
389 HandleGenerateConditionWithZero(condition);
390 return;
391 case kCondLT:
392 case kCondGE:
393 FALLTHROUGH_INTENDED;
394 case kCondAE:
395 case kCondB:
396 HandleGenerateConditionWithZero(condition);
397 return;
398 case kCondLE:
399 case kCondGT:
400 default:
401 break;
402 }
403 }
404 }
405
406 if ((condition == kCondEQ || condition == kCondNE) &&
407 !CanGenerateTest(cond)) {
408 HandleGenerateEqualLong(cond);
409 return;
410 }
411
412 if (CanGenerateTest(cond)) {
413 HandleGenerateConditionGeneric(cond);
414 return;
415 }
416
417 HandleGenerateLongComparesAndJumps();
418
419 last_visited_internal_latency_ += kArmIntegerOpLatency;
420 last_visited_latency_ = kArmBranchLatency;;
421}
422
423void SchedulingLatencyVisitorARM::HandleGenerateConditionIntegralOrNonPrimitive(HCondition* cond) {
424 const Primitive::Type type = cond->GetLeft()->GetType();
425
426 DCHECK(Primitive::IsIntegralType(type) || type == Primitive::kPrimNot) << type;
427
428 if (type == Primitive::kPrimLong) {
429 HandleGenerateConditionLong(cond);
430 return;
431 }
432
433 IfCondition condition = cond->GetCondition();
434 HInstruction* right = cond->InputAt(1);
435 int64_t value;
436
437 if (right->IsConstant()) {
438 value = Uint64ConstantFrom(right);
439
440 // Comparisons against 0 are common enough, so codegen has special handling for them.
441 if (value == 0) {
442 switch (condition) {
443 case kCondNE:
444 case kCondA:
445 case kCondEQ:
446 case kCondBE:
447 case kCondLT:
448 case kCondGE:
449 case kCondAE:
450 case kCondB:
451 HandleGenerateConditionWithZero(condition);
452 return;
453 case kCondLE:
454 case kCondGT:
455 default:
456 break;
457 }
458 }
459 }
460
461 if (condition == kCondEQ || condition == kCondNE) {
462 if (condition == kCondNE) {
463 // CMP, IT, MOV.ne
464 last_visited_internal_latency_ += 2 * kArmIntegerOpLatency;
465 last_visited_latency_ = kArmIntegerOpLatency;
466 } else {
467 last_visited_internal_latency_ += kArmIntegerOpLatency;
468 HandleGenerateConditionWithZero(condition);
469 }
470 return;
471 }
472
473 HandleGenerateConditionGeneric(cond);
474}
475
476void SchedulingLatencyVisitorARM::HandleCondition(HCondition* cond) {
477 if (cond->IsEmittedAtUseSite()) {
478 last_visited_latency_ = 0;
479 return;
480 }
481
482 const Primitive::Type type = cond->GetLeft()->GetType();
483
484 if (Primitive::IsFloatingPointType(type)) {
485 HandleGenerateConditionGeneric(cond);
486 return;
487 }
488
489 DCHECK(Primitive::IsIntegralType(type) || type == Primitive::kPrimNot) << type;
490
491 const IfCondition condition = cond->GetCondition();
492
493 if (type == Primitive::kPrimBoolean &&
494 cond->GetRight()->GetType() == Primitive::kPrimBoolean &&
495 (condition == kCondEQ || condition == kCondNE)) {
496 if (condition == kCondEQ) {
497 last_visited_internal_latency_ = kArmIntegerOpLatency;
498 }
499 last_visited_latency_ = kArmIntegerOpLatency;
500 return;
501 }
502
503 HandleGenerateConditionIntegralOrNonPrimitive(cond);
504}
505
506void SchedulingLatencyVisitorARM::VisitCondition(HCondition* instr) {
507 HandleCondition(instr);
508}
509
xueliang.zhongf7caf682017-03-01 16:07:02 +0000510void SchedulingLatencyVisitorARM::VisitCompare(HCompare* instr) {
511 Primitive::Type type = instr->InputAt(0)->GetType();
512 switch (type) {
513 case Primitive::kPrimBoolean:
514 case Primitive::kPrimByte:
515 case Primitive::kPrimShort:
516 case Primitive::kPrimChar:
517 case Primitive::kPrimInt:
518 last_visited_internal_latency_ = 2 * kArmIntegerOpLatency;
519 break;
520 case Primitive::kPrimLong:
521 last_visited_internal_latency_ = 2 * kArmIntegerOpLatency + 3 * kArmBranchLatency;
522 break;
523 case Primitive::kPrimFloat:
524 case Primitive::kPrimDouble:
525 last_visited_internal_latency_ = kArmIntegerOpLatency + 2 * kArmFloatingPointOpLatency;
526 break;
527 default:
528 last_visited_internal_latency_ = 2 * kArmIntegerOpLatency;
529 break;
530 }
531 last_visited_latency_ = kArmIntegerOpLatency;
532}
533
534void SchedulingLatencyVisitorARM::VisitBitwiseNegatedRight(HBitwiseNegatedRight* instruction) {
535 if (instruction->GetResultType() == Primitive::kPrimInt) {
536 last_visited_latency_ = kArmIntegerOpLatency;
537 } else {
538 last_visited_internal_latency_ = kArmIntegerOpLatency;
539 last_visited_latency_ = kArmIntegerOpLatency;
540 }
541}
542
543void SchedulingLatencyVisitorARM::HandleGenerateDataProcInstruction(bool internal_latency) {
544 if (internal_latency) {
545 last_visited_internal_latency_ += kArmIntegerOpLatency;
546 } else {
547 last_visited_latency_ = kArmDataProcWithShifterOpLatency;
548 }
549}
550
551void SchedulingLatencyVisitorARM::HandleGenerateDataProc(HDataProcWithShifterOp* instruction) {
552 const HInstruction::InstructionKind kind = instruction->GetInstrKind();
553 if (kind == HInstruction::kAdd) {
554 last_visited_internal_latency_ = kArmIntegerOpLatency;
555 last_visited_latency_ = kArmIntegerOpLatency;
556 } else if (kind == HInstruction::kSub) {
557 last_visited_internal_latency_ = kArmIntegerOpLatency;
558 last_visited_latency_ = kArmIntegerOpLatency;
559 } else {
560 HandleGenerateDataProcInstruction(/* internal_latency */ true);
561 HandleGenerateDataProcInstruction();
562 }
563}
564
565void SchedulingLatencyVisitorARM::HandleGenerateLongDataProc(HDataProcWithShifterOp* instruction) {
566 DCHECK_EQ(instruction->GetType(), Primitive::kPrimLong);
567 DCHECK(HDataProcWithShifterOp::IsShiftOp(instruction->GetOpKind()));
568
569 const uint32_t shift_value = instruction->GetShiftAmount();
570 const HInstruction::InstructionKind kind = instruction->GetInstrKind();
571
572 if (shift_value >= 32) {
573 // Different shift types actually generate similar code here,
574 // no need to differentiate shift types like the codegen pass does,
575 // which also avoids handling shift types from different ARM backends.
576 HandleGenerateDataProc(instruction);
577 } else {
578 DCHECK_GT(shift_value, 1U);
579 DCHECK_LT(shift_value, 32U);
580
581 if (kind == HInstruction::kOr || kind == HInstruction::kXor) {
582 HandleGenerateDataProcInstruction(/* internal_latency */ true);
583 HandleGenerateDataProcInstruction(/* internal_latency */ true);
584 HandleGenerateDataProcInstruction();
585 } else {
586 last_visited_internal_latency_ += 2 * kArmIntegerOpLatency;
587 HandleGenerateDataProc(instruction);
588 }
589 }
590}
591
592void SchedulingLatencyVisitorARM::VisitDataProcWithShifterOp(HDataProcWithShifterOp* instruction) {
593 const HDataProcWithShifterOp::OpKind op_kind = instruction->GetOpKind();
594
595 if (instruction->GetType() == Primitive::kPrimInt) {
xueliang.zhongf7caf682017-03-01 16:07:02 +0000596 HandleGenerateDataProcInstruction();
597 } else {
598 DCHECK_EQ(instruction->GetType(), Primitive::kPrimLong);
599 if (HDataProcWithShifterOp::IsExtensionOp(op_kind)) {
600 HandleGenerateDataProc(instruction);
601 } else {
602 HandleGenerateLongDataProc(instruction);
603 }
604 }
605}
606
607void SchedulingLatencyVisitorARM::VisitIntermediateAddress(HIntermediateAddress* ATTRIBUTE_UNUSED) {
608 // Although the code generated is a simple `add` instruction, we found through empirical results
609 // that spacing it from its use in memory accesses was beneficial.
610 last_visited_internal_latency_ = kArmNopLatency;
611 last_visited_latency_ = kArmIntegerOpLatency;
612}
613
Artem Serovf0fc4c62017-05-03 15:07:15 +0100614void SchedulingLatencyVisitorARM::VisitIntermediateAddressIndex(
615 HIntermediateAddressIndex* ATTRIBUTE_UNUSED) {
616 UNIMPLEMENTED(FATAL) << "IntermediateAddressIndex is not implemented for ARM";
617}
618
xueliang.zhongf7caf682017-03-01 16:07:02 +0000619void SchedulingLatencyVisitorARM::VisitMultiplyAccumulate(HMultiplyAccumulate* ATTRIBUTE_UNUSED) {
620 last_visited_latency_ = kArmMulIntegerLatency;
621}
622
623void SchedulingLatencyVisitorARM::VisitArrayGet(HArrayGet* instruction) {
624 Primitive::Type type = instruction->GetType();
625 const bool maybe_compressed_char_at =
626 mirror::kUseStringCompression && instruction->IsStringCharAt();
627 HInstruction* array_instr = instruction->GetArray();
628 bool has_intermediate_address = array_instr->IsIntermediateAddress();
629 HInstruction* index = instruction->InputAt(1);
630
631 switch (type) {
632 case Primitive::kPrimBoolean:
633 case Primitive::kPrimByte:
634 case Primitive::kPrimShort:
635 case Primitive::kPrimChar:
636 case Primitive::kPrimInt: {
637 if (maybe_compressed_char_at) {
638 last_visited_internal_latency_ += kArmMemoryLoadLatency;
639 }
640 if (index->IsConstant()) {
641 if (maybe_compressed_char_at) {
642 last_visited_internal_latency_ +=
643 kArmIntegerOpLatency + kArmBranchLatency + kArmMemoryLoadLatency;
644 last_visited_latency_ = kArmBranchLatency;
645 } else {
646 last_visited_latency_ += kArmMemoryLoadLatency;
647 }
648 } else {
649 if (has_intermediate_address) {
650 } else {
651 last_visited_internal_latency_ += kArmIntegerOpLatency;
652 }
653 if (maybe_compressed_char_at) {
654 last_visited_internal_latency_ +=
655 kArmIntegerOpLatency + kArmBranchLatency + kArmMemoryLoadLatency;
656 last_visited_latency_ = kArmBranchLatency;
657 } else {
658 last_visited_latency_ += kArmMemoryLoadLatency;
659 }
660 }
661 break;
662 }
663
664 case Primitive::kPrimNot: {
665 if (kEmitCompilerReadBarrier && kUseBakerReadBarrier) {
666 last_visited_latency_ = kArmLoadWithBakerReadBarrierLatency;
667 } else {
668 if (index->IsConstant()) {
669 last_visited_latency_ = kArmMemoryLoadLatency;
670 } else {
671 if (has_intermediate_address) {
672 } else {
673 last_visited_internal_latency_ += kArmIntegerOpLatency;
674 }
675 last_visited_internal_latency_ = kArmMemoryLoadLatency;
676 }
677 }
678 break;
679 }
680
681 case Primitive::kPrimLong: {
682 if (index->IsConstant()) {
683 last_visited_latency_ = kArmMemoryLoadLatency;
684 } else {
685 last_visited_internal_latency_ += kArmIntegerOpLatency;
686 last_visited_latency_ = kArmMemoryLoadLatency;
687 }
688 break;
689 }
690
691 case Primitive::kPrimFloat: {
692 if (index->IsConstant()) {
693 last_visited_latency_ = kArmMemoryLoadLatency;
694 } else {
695 last_visited_internal_latency_ += kArmIntegerOpLatency;
696 last_visited_latency_ = kArmMemoryLoadLatency;
697 }
698 break;
699 }
700
701 case Primitive::kPrimDouble: {
702 if (index->IsConstant()) {
703 last_visited_latency_ = kArmMemoryLoadLatency;
704 } else {
705 last_visited_internal_latency_ += kArmIntegerOpLatency;
706 last_visited_latency_ = kArmMemoryLoadLatency;
707 }
708 break;
709 }
710
711 default:
712 LOG(FATAL) << "Unreachable type " << type;
713 UNREACHABLE();
714 }
715}
716
717void SchedulingLatencyVisitorARM::VisitArrayLength(HArrayLength* instruction) {
718 last_visited_latency_ = kArmMemoryLoadLatency;
719 if (mirror::kUseStringCompression && instruction->IsStringLength()) {
720 last_visited_internal_latency_ = kArmMemoryLoadLatency;
721 last_visited_latency_ = kArmIntegerOpLatency;
722 }
723}
724
725void SchedulingLatencyVisitorARM::VisitArraySet(HArraySet* instruction) {
726 HInstruction* index = instruction->InputAt(1);
727 Primitive::Type value_type = instruction->GetComponentType();
728 HInstruction* array_instr = instruction->GetArray();
729 bool has_intermediate_address = array_instr->IsIntermediateAddress();
730
731 switch (value_type) {
732 case Primitive::kPrimBoolean:
733 case Primitive::kPrimByte:
734 case Primitive::kPrimShort:
735 case Primitive::kPrimChar:
736 case Primitive::kPrimInt: {
737 if (index->IsConstant()) {
738 last_visited_latency_ = kArmMemoryStoreLatency;
739 } else {
740 if (has_intermediate_address) {
741 } else {
742 last_visited_internal_latency_ = kArmIntegerOpLatency;
743 }
744 last_visited_latency_ = kArmMemoryStoreLatency;
745 }
746 break;
747 }
748
749 case Primitive::kPrimNot: {
750 if (instruction->InputAt(2)->IsNullConstant()) {
751 if (index->IsConstant()) {
752 last_visited_latency_ = kArmMemoryStoreLatency;
753 } else {
754 last_visited_internal_latency_ = kArmIntegerOpLatency;
755 last_visited_latency_ = kArmMemoryStoreLatency;
756 }
757 } else {
758 // Following the exact instructions of runtime type checks is too complicated,
759 // just giving it a simple slow latency.
760 last_visited_latency_ = kArmRuntimeTypeCheckLatency;
761 }
762 break;
763 }
764
765 case Primitive::kPrimLong: {
766 if (index->IsConstant()) {
767 last_visited_latency_ = kArmMemoryLoadLatency;
768 } else {
769 last_visited_internal_latency_ = kArmIntegerOpLatency;
770 last_visited_latency_ = kArmMemoryLoadLatency;
771 }
772 break;
773 }
774
775 case Primitive::kPrimFloat: {
776 if (index->IsConstant()) {
777 last_visited_latency_ = kArmMemoryLoadLatency;
778 } else {
779 last_visited_internal_latency_ = kArmIntegerOpLatency;
780 last_visited_latency_ = kArmMemoryLoadLatency;
781 }
782 break;
783 }
784
785 case Primitive::kPrimDouble: {
786 if (index->IsConstant()) {
787 last_visited_latency_ = kArmMemoryLoadLatency;
788 } else {
789 last_visited_internal_latency_ = kArmIntegerOpLatency;
790 last_visited_latency_ = kArmMemoryLoadLatency;
791 }
792 break;
793 }
794
795 default:
796 LOG(FATAL) << "Unreachable type " << value_type;
797 UNREACHABLE();
798 }
799}
800
801void SchedulingLatencyVisitorARM::VisitBoundsCheck(HBoundsCheck* ATTRIBUTE_UNUSED) {
802 last_visited_internal_latency_ = kArmIntegerOpLatency;
803 // Users do not use any data results.
804 last_visited_latency_ = 0;
805}
806
807void SchedulingLatencyVisitorARM::HandleDivRemConstantIntegralLatencies(int32_t imm) {
808 if (imm == 0) {
809 last_visited_internal_latency_ = 0;
810 last_visited_latency_ = 0;
811 } else if (imm == 1 || imm == -1) {
812 last_visited_latency_ = kArmIntegerOpLatency;
813 } else if (IsPowerOfTwo(AbsOrMin(imm))) {
814 last_visited_internal_latency_ = 3 * kArmIntegerOpLatency;
815 last_visited_latency_ = kArmIntegerOpLatency;
816 } else {
817 last_visited_internal_latency_ = kArmMulIntegerLatency + 2 * kArmIntegerOpLatency;
818 last_visited_latency_ = kArmIntegerOpLatency;
819 }
820}
821
822void SchedulingLatencyVisitorARM::VisitDiv(HDiv* instruction) {
823 Primitive::Type type = instruction->GetResultType();
824 switch (type) {
825 case Primitive::kPrimInt: {
826 HInstruction* rhs = instruction->GetRight();
827 if (rhs->IsConstant()) {
828 int32_t imm = Int32ConstantFrom(rhs->AsConstant());
829 HandleDivRemConstantIntegralLatencies(imm);
830 } else {
831 last_visited_latency_ = kArmDivIntegerLatency;
832 }
833 break;
834 }
835 case Primitive::kPrimFloat:
836 last_visited_latency_ = kArmDivFloatLatency;
837 break;
838 case Primitive::kPrimDouble:
839 last_visited_latency_ = kArmDivDoubleLatency;
840 break;
841 default:
842 last_visited_internal_latency_ = kArmCallInternalLatency;
843 last_visited_latency_ = kArmCallLatency;
844 break;
845 }
846}
847
848void SchedulingLatencyVisitorARM::VisitInstanceFieldGet(HInstanceFieldGet* instruction) {
849 HandleFieldGetLatencies(instruction, instruction->GetFieldInfo());
850}
851
852void SchedulingLatencyVisitorARM::VisitInstanceFieldSet(HInstanceFieldSet* instruction) {
853 HandleFieldSetLatencies(instruction, instruction->GetFieldInfo());
854}
855
856void SchedulingLatencyVisitorARM::VisitInstanceOf(HInstanceOf* ATTRIBUTE_UNUSED) {
857 last_visited_internal_latency_ = kArmCallInternalLatency;
858 last_visited_latency_ = kArmIntegerOpLatency;
859}
860
861void SchedulingLatencyVisitorARM::VisitInvoke(HInvoke* ATTRIBUTE_UNUSED) {
862 last_visited_internal_latency_ = kArmCallInternalLatency;
863 last_visited_latency_ = kArmCallLatency;
864}
865
866void SchedulingLatencyVisitorARM::VisitLoadString(HLoadString* ATTRIBUTE_UNUSED) {
867 last_visited_internal_latency_ = kArmLoadStringInternalLatency;
868 last_visited_latency_ = kArmMemoryLoadLatency;
869}
870
871void SchedulingLatencyVisitorARM::VisitNewArray(HNewArray* ATTRIBUTE_UNUSED) {
872 last_visited_internal_latency_ = kArmIntegerOpLatency + kArmCallInternalLatency;
873 last_visited_latency_ = kArmCallLatency;
874}
875
876void SchedulingLatencyVisitorARM::VisitNewInstance(HNewInstance* instruction) {
877 if (instruction->IsStringAlloc()) {
878 last_visited_internal_latency_ = 2 * kArmMemoryLoadLatency + kArmCallInternalLatency;
879 } else {
880 last_visited_internal_latency_ = kArmCallInternalLatency;
881 }
882 last_visited_latency_ = kArmCallLatency;
883}
884
885void SchedulingLatencyVisitorARM::VisitRem(HRem* instruction) {
886 Primitive::Type type = instruction->GetResultType();
887 switch (type) {
888 case Primitive::kPrimInt: {
889 HInstruction* rhs = instruction->GetRight();
890 if (rhs->IsConstant()) {
891 int32_t imm = Int32ConstantFrom(rhs->AsConstant());
892 HandleDivRemConstantIntegralLatencies(imm);
893 } else {
894 last_visited_internal_latency_ = kArmDivIntegerLatency;
895 last_visited_latency_ = kArmMulIntegerLatency;
896 }
897 break;
898 }
899 default:
900 last_visited_internal_latency_ = kArmCallInternalLatency;
901 last_visited_latency_ = kArmCallLatency;
902 break;
903 }
904}
905
906void SchedulingLatencyVisitorARM::HandleFieldGetLatencies(HInstruction* instruction,
907 const FieldInfo& field_info) {
908 DCHECK(instruction->IsInstanceFieldGet() || instruction->IsStaticFieldGet());
909 DCHECK(codegen_ != nullptr);
910 bool is_volatile = field_info.IsVolatile();
911 Primitive::Type field_type = field_info.GetFieldType();
912 bool atomic_ldrd_strd = codegen_->GetInstructionSetFeatures().HasAtomicLdrdAndStrd();
913
914 switch (field_type) {
915 case Primitive::kPrimBoolean:
916 case Primitive::kPrimByte:
917 case Primitive::kPrimShort:
918 case Primitive::kPrimChar:
919 case Primitive::kPrimInt:
920 last_visited_latency_ = kArmMemoryLoadLatency;
921 break;
922
923 case Primitive::kPrimNot:
924 if (kEmitCompilerReadBarrier && kUseBakerReadBarrier) {
925 last_visited_internal_latency_ = kArmMemoryLoadLatency + kArmIntegerOpLatency;
926 last_visited_latency_ = kArmMemoryLoadLatency;
927 } else {
928 last_visited_latency_ = kArmMemoryLoadLatency;
929 }
930 break;
931
932 case Primitive::kPrimLong:
933 if (is_volatile && !atomic_ldrd_strd) {
934 last_visited_internal_latency_ = kArmMemoryLoadLatency + kArmIntegerOpLatency;
935 last_visited_latency_ = kArmMemoryLoadLatency;
936 } else {
937 last_visited_latency_ = kArmMemoryLoadLatency;
938 }
939 break;
940
941 case Primitive::kPrimFloat:
942 last_visited_latency_ = kArmMemoryLoadLatency;
943 break;
944
945 case Primitive::kPrimDouble:
946 if (is_volatile && !atomic_ldrd_strd) {
947 last_visited_internal_latency_ =
948 kArmMemoryLoadLatency + kArmIntegerOpLatency + kArmMemoryLoadLatency;
949 last_visited_latency_ = kArmIntegerOpLatency;
950 } else {
951 last_visited_latency_ = kArmMemoryLoadLatency;
952 }
953 break;
954
955 default:
956 last_visited_latency_ = kArmMemoryLoadLatency;
957 break;
958 }
959
960 if (is_volatile) {
961 last_visited_internal_latency_ += kArmMemoryBarrierLatency;
962 }
963}
964
965void SchedulingLatencyVisitorARM::HandleFieldSetLatencies(HInstruction* instruction,
966 const FieldInfo& field_info) {
967 DCHECK(instruction->IsInstanceFieldSet() || instruction->IsStaticFieldSet());
968 DCHECK(codegen_ != nullptr);
969 bool is_volatile = field_info.IsVolatile();
970 Primitive::Type field_type = field_info.GetFieldType();
971 bool needs_write_barrier =
972 CodeGenerator::StoreNeedsWriteBarrier(field_type, instruction->InputAt(1));
973 bool atomic_ldrd_strd = codegen_->GetInstructionSetFeatures().HasAtomicLdrdAndStrd();
974
975 switch (field_type) {
976 case Primitive::kPrimBoolean:
977 case Primitive::kPrimByte:
978 case Primitive::kPrimShort:
979 case Primitive::kPrimChar:
980 if (is_volatile) {
981 last_visited_internal_latency_ = kArmMemoryBarrierLatency + kArmMemoryStoreLatency;
982 last_visited_latency_ = kArmMemoryBarrierLatency;
983 } else {
984 last_visited_latency_ = kArmMemoryStoreLatency;
985 }
986 break;
987
988 case Primitive::kPrimInt:
989 case Primitive::kPrimNot:
990 if (kPoisonHeapReferences && needs_write_barrier) {
991 last_visited_internal_latency_ += kArmIntegerOpLatency * 2;
992 }
993 last_visited_latency_ = kArmMemoryStoreLatency;
994 break;
995
996 case Primitive::kPrimLong:
997 if (is_volatile && !atomic_ldrd_strd) {
998 last_visited_internal_latency_ =
999 kArmIntegerOpLatency + kArmMemoryLoadLatency + kArmMemoryStoreLatency;
1000 last_visited_latency_ = kArmIntegerOpLatency;
1001 } else {
1002 last_visited_latency_ = kArmMemoryStoreLatency;
1003 }
1004 break;
1005
1006 case Primitive::kPrimFloat:
1007 last_visited_latency_ = kArmMemoryStoreLatency;
1008 break;
1009
1010 case Primitive::kPrimDouble:
1011 if (is_volatile && !atomic_ldrd_strd) {
1012 last_visited_internal_latency_ = kArmIntegerOpLatency +
1013 kArmIntegerOpLatency + kArmMemoryLoadLatency + kArmMemoryStoreLatency;
1014 last_visited_latency_ = kArmIntegerOpLatency;
1015 } else {
1016 last_visited_latency_ = kArmMemoryStoreLatency;
1017 }
1018 break;
1019
1020 default:
1021 last_visited_latency_ = kArmMemoryStoreLatency;
1022 break;
1023 }
1024}
1025
1026void SchedulingLatencyVisitorARM::VisitStaticFieldGet(HStaticFieldGet* instruction) {
1027 HandleFieldGetLatencies(instruction, instruction->GetFieldInfo());
1028}
1029
1030void SchedulingLatencyVisitorARM::VisitStaticFieldSet(HStaticFieldSet* instruction) {
1031 HandleFieldSetLatencies(instruction, instruction->GetFieldInfo());
1032}
1033
1034void SchedulingLatencyVisitorARM::VisitSuspendCheck(HSuspendCheck* instruction) {
1035 HBasicBlock* block = instruction->GetBlock();
1036 DCHECK((block->GetLoopInformation() != nullptr) ||
1037 (block->IsEntryBlock() && instruction->GetNext()->IsGoto()));
1038 // Users do not use any data results.
1039 last_visited_latency_ = 0;
1040}
1041
1042void SchedulingLatencyVisitorARM::VisitTypeConversion(HTypeConversion* instr) {
1043 Primitive::Type result_type = instr->GetResultType();
1044 Primitive::Type input_type = instr->GetInputType();
1045
1046 switch (result_type) {
1047 case Primitive::kPrimByte:
1048 case Primitive::kPrimChar:
1049 case Primitive::kPrimShort:
1050 last_visited_latency_ = kArmIntegerOpLatency; // SBFX or UBFX
1051 break;
1052
1053 case Primitive::kPrimInt:
1054 switch (input_type) {
1055 case Primitive::kPrimLong:
1056 last_visited_latency_ = kArmIntegerOpLatency; // MOV
1057 break;
1058 case Primitive::kPrimFloat:
1059 case Primitive::kPrimDouble:
1060 last_visited_internal_latency_ = kArmTypeConversionFloatingPointIntegerLatency;
1061 last_visited_latency_ = kArmFloatingPointOpLatency;
1062 break;
1063 default:
1064 last_visited_latency_ = kArmIntegerOpLatency;
1065 break;
1066 }
1067 break;
1068
1069 case Primitive::kPrimLong:
1070 switch (input_type) {
1071 case Primitive::kPrimBoolean:
1072 case Primitive::kPrimByte:
1073 case Primitive::kPrimChar:
1074 case Primitive::kPrimShort:
1075 case Primitive::kPrimInt:
1076 // MOV and extension
1077 last_visited_internal_latency_ = kArmIntegerOpLatency;
1078 last_visited_latency_ = kArmIntegerOpLatency;
1079 break;
1080 case Primitive::kPrimFloat:
1081 case Primitive::kPrimDouble:
1082 // invokes runtime
1083 last_visited_internal_latency_ = kArmCallInternalLatency;
1084 break;
1085 default:
1086 last_visited_internal_latency_ = kArmIntegerOpLatency;
1087 last_visited_latency_ = kArmIntegerOpLatency;
1088 break;
1089 }
1090 break;
1091
1092 case Primitive::kPrimFloat:
1093 switch (input_type) {
1094 case Primitive::kPrimBoolean:
1095 case Primitive::kPrimByte:
1096 case Primitive::kPrimChar:
1097 case Primitive::kPrimShort:
1098 case Primitive::kPrimInt:
1099 last_visited_internal_latency_ = kArmTypeConversionFloatingPointIntegerLatency;
1100 last_visited_latency_ = kArmFloatingPointOpLatency;
1101 break;
1102 case Primitive::kPrimLong:
1103 // invokes runtime
1104 last_visited_internal_latency_ = kArmCallInternalLatency;
1105 break;
1106 case Primitive::kPrimDouble:
1107 last_visited_latency_ = kArmFloatingPointOpLatency;
1108 break;
1109 default:
1110 last_visited_latency_ = kArmFloatingPointOpLatency;
1111 break;
1112 }
1113 break;
1114
1115 case Primitive::kPrimDouble:
1116 switch (input_type) {
1117 case Primitive::kPrimBoolean:
1118 case Primitive::kPrimByte:
1119 case Primitive::kPrimChar:
1120 case Primitive::kPrimShort:
1121 case Primitive::kPrimInt:
1122 last_visited_internal_latency_ = kArmTypeConversionFloatingPointIntegerLatency;
1123 last_visited_latency_ = kArmFloatingPointOpLatency;
1124 break;
1125 case Primitive::kPrimLong:
1126 last_visited_internal_latency_ = 5 * kArmFloatingPointOpLatency;
1127 last_visited_latency_ = kArmFloatingPointOpLatency;
1128 break;
1129 case Primitive::kPrimFloat:
1130 last_visited_latency_ = kArmFloatingPointOpLatency;
1131 break;
1132 default:
1133 last_visited_latency_ = kArmFloatingPointOpLatency;
1134 break;
1135 }
1136 break;
1137
1138 default:
1139 last_visited_latency_ = kArmTypeConversionFloatingPointIntegerLatency;
1140 break;
1141 }
1142}
1143
xueliang.zhongf7caf682017-03-01 16:07:02 +00001144} // namespace arm
1145} // namespace art