Mathias Agopian | 61fd2ab | 2016-08-16 19:42:42 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (C) 2016 The Android Open Source Project |
| 3 | * |
| 4 | * Licensed under the Apache License, Version 2.0 (the "License"); |
| 5 | * you may not use this file except in compliance with the License. |
| 6 | * You may obtain a copy of the License at |
| 7 | * |
| 8 | * http://www.apache.org/licenses/LICENSE-2.0 |
| 9 | * |
| 10 | * Unless required by applicable law or agreed to in writing, software |
| 11 | * distributed under the License is distributed on an "AS IS" BASIS, |
| 12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 13 | * See the License for the specific language governing permissions and |
| 14 | * limitations under the License. |
| 15 | */ |
| 16 | |
| 17 | #include "Profiler.h" |
| 18 | |
| 19 | #include <stdlib.h> |
| 20 | #include <string.h> |
| 21 | #include <unistd.h> |
| 22 | |
| 23 | #include <algorithm> |
| 24 | #include <iostream> |
| 25 | |
| 26 | #if defined(__linux__) |
| 27 | |
| 28 | #include <sys/syscall.h> |
| 29 | |
| 30 | #ifdef __ARM_ARCH |
| 31 | enum ARMv8PmuPerfTypes{ |
| 32 | // Common micro-architecture events |
| 33 | ARMV8_PMUV3_PERFCTR_L1_ICACHE_REFILL = 0x01, |
| 34 | ARMV8_PMUV3_PERFCTR_L1_ICACHE_ACCESS = 0x14, |
| 35 | ARMV8_PMUV3_PERFCTR_L2_CACHE_ACCESS = 0x16, |
| 36 | ARMV8_PMUV3_PERFCTR_L2_CACHE_REFILL = 0x17, |
| 37 | ARMV8_PMUV3_PERFCTR_L2_CACHE_WB = 0x18, |
| 38 | }; |
| 39 | #endif |
| 40 | |
| 41 | static int perf_event_open(struct perf_event_attr* hw_event, pid_t pid, |
| 42 | int cpu, int group_fd, unsigned long flags) { |
| 43 | return syscall(__NR_perf_event_open, hw_event, pid, cpu, group_fd, flags); |
| 44 | } |
| 45 | |
| 46 | #endif // __linux__ |
| 47 | |
| 48 | namespace utils { |
| 49 | |
| 50 | Profiler& Profiler::get() noexcept { |
| 51 | static Profiler sProfiler; |
| 52 | return sProfiler; |
| 53 | } |
| 54 | |
| 55 | Profiler::Profiler() noexcept { |
| 56 | std::uninitialized_fill(mCountersFd.begin(), mCountersFd.end(), -1); |
| 57 | Profiler::resetEvents(EV_CPU_CYCLES | EV_L1D_RATES | EV_BPU_RATES); |
| 58 | } |
| 59 | |
| 60 | Profiler::~Profiler() noexcept { |
| 61 | for (int fd : mCountersFd) { |
| 62 | if (fd >= 0) { |
| 63 | close(fd); |
| 64 | } |
| 65 | } |
| 66 | } |
| 67 | |
| 68 | uint32_t Profiler::resetEvents(uint32_t eventMask) noexcept { |
| 69 | // close all counters |
| 70 | for (int& fd : mCountersFd) { |
| 71 | if (fd >= 0) { |
| 72 | close(fd); |
| 73 | fd = -1; |
| 74 | } |
| 75 | } |
| 76 | mEnabledEvents = 0; |
| 77 | |
| 78 | #if defined(__linux__) |
| 79 | |
| 80 | struct perf_event_attr pe; |
| 81 | memset(&pe, 0, sizeof(struct perf_event_attr)); |
| 82 | pe.type = PERF_TYPE_HARDWARE; |
| 83 | pe.size = sizeof(struct perf_event_attr); |
| 84 | pe.config = PERF_COUNT_HW_INSTRUCTIONS; |
| 85 | pe.disabled = 1; |
| 86 | pe.exclude_kernel = 1; |
| 87 | pe.exclude_hv = 1; |
| 88 | pe.read_format = PERF_FORMAT_GROUP | |
| 89 | PERF_FORMAT_ID | |
| 90 | PERF_FORMAT_TOTAL_TIME_ENABLED | |
| 91 | PERF_FORMAT_TOTAL_TIME_RUNNING; |
| 92 | |
| 93 | uint8_t count = 0; |
| 94 | int fd = perf_event_open(&pe, 0, -1, -1, 0); |
| 95 | if (fd >= 0) { |
| 96 | const int groupFd = fd; |
| 97 | mIds[INSTRUCTIONS] = count++; |
| 98 | mCountersFd[INSTRUCTIONS] = fd; |
| 99 | |
| 100 | pe.read_format = PERF_FORMAT_GROUP | PERF_FORMAT_ID; |
| 101 | |
| 102 | if (eventMask & EV_CPU_CYCLES) { |
| 103 | pe.type = PERF_TYPE_HARDWARE; |
| 104 | pe.config = PERF_COUNT_HW_CPU_CYCLES; |
| 105 | mCountersFd[CPU_CYCLES] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 106 | if (mCountersFd[CPU_CYCLES] > 0) { |
| 107 | mIds[CPU_CYCLES] = count++; |
| 108 | mEnabledEvents |= EV_CPU_CYCLES; |
| 109 | } |
| 110 | } |
| 111 | |
| 112 | if (eventMask & EV_L1D_REFS) { |
| 113 | pe.type = PERF_TYPE_HARDWARE; |
| 114 | pe.config = PERF_COUNT_HW_CACHE_REFERENCES; |
| 115 | mCountersFd[DCACHE_REFS] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 116 | if (mCountersFd[DCACHE_REFS] > 0) { |
| 117 | mIds[DCACHE_REFS] = count++; |
| 118 | mEnabledEvents |= EV_L1D_REFS; |
| 119 | } |
| 120 | } |
| 121 | |
| 122 | if (eventMask & EV_L1D_MISSES) { |
| 123 | pe.type = PERF_TYPE_HARDWARE; |
| 124 | pe.config = PERF_COUNT_HW_CACHE_MISSES; |
| 125 | mCountersFd[DCACHE_MISSES] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 126 | if (mCountersFd[DCACHE_MISSES] > 0) { |
| 127 | mIds[DCACHE_MISSES] = count++; |
| 128 | mEnabledEvents |= EV_L1D_MISSES; |
| 129 | } |
| 130 | } |
| 131 | |
| 132 | if (eventMask & EV_BPU_REFS) { |
| 133 | pe.type = PERF_TYPE_HARDWARE; |
| 134 | pe.config = PERF_COUNT_HW_BRANCH_INSTRUCTIONS; |
| 135 | mCountersFd[BRANCHES] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 136 | if (mCountersFd[BRANCHES] > 0) { |
| 137 | mIds[BRANCHES] = count++; |
| 138 | mEnabledEvents |= EV_BPU_REFS; |
| 139 | } |
| 140 | } |
| 141 | |
| 142 | if (eventMask & EV_BPU_MISSES) { |
| 143 | pe.type = PERF_TYPE_HARDWARE; |
| 144 | pe.config = PERF_COUNT_HW_BRANCH_MISSES; |
| 145 | mCountersFd[BRANCH_MISSES] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 146 | if (mCountersFd[BRANCH_MISSES] > 0) { |
| 147 | mIds[BRANCH_MISSES] = count++; |
| 148 | mEnabledEvents |= EV_BPU_MISSES; |
| 149 | } |
| 150 | } |
| 151 | |
| 152 | #ifdef __ARM_ARCH |
| 153 | if (eventMask & EV_L1I_REFS) { |
| 154 | pe.type = PERF_TYPE_RAW; |
| 155 | pe.config = ARMV8_PMUV3_PERFCTR_L1_ICACHE_ACCESS; |
| 156 | mCountersFd[ICACHE_REFS] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 157 | if (mCountersFd[ICACHE_REFS] > 0) { |
| 158 | mIds[ICACHE_REFS] = count++; |
| 159 | mEnabledEvents |= EV_L1I_REFS; |
| 160 | } |
| 161 | } |
| 162 | |
| 163 | if (eventMask & EV_L1I_MISSES) { |
| 164 | pe.type = PERF_TYPE_RAW; |
| 165 | pe.config = ARMV8_PMUV3_PERFCTR_L1_ICACHE_REFILL; |
| 166 | mCountersFd[ICACHE_MISSES] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 167 | if (mCountersFd[ICACHE_MISSES] > 0) { |
| 168 | mIds[ICACHE_MISSES] = count++; |
| 169 | mEnabledEvents |= EV_L1I_MISSES; |
| 170 | } |
| 171 | } |
| 172 | #else |
| 173 | if (eventMask & EV_L1I_REFS) { |
| 174 | pe.type = PERF_TYPE_HW_CACHE; |
| 175 | pe.config = PERF_COUNT_HW_CACHE_L1I | |
| 176 | (PERF_COUNT_HW_CACHE_OP_READ<<8) | (PERF_COUNT_HW_CACHE_RESULT_ACCESS<<16); |
| 177 | mCountersFd[ICACHE_REFS] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 178 | if (mCountersFd[ICACHE_REFS] > 0) { |
| 179 | mIds[ICACHE_REFS] = count++; |
| 180 | mEnabledEvents |= EV_L1I_REFS; |
| 181 | } |
| 182 | } |
| 183 | |
| 184 | if (eventMask & EV_L1I_MISSES) { |
| 185 | pe.type = PERF_TYPE_HW_CACHE; |
| 186 | pe.config = PERF_COUNT_HW_CACHE_L1I | |
| 187 | (PERF_COUNT_HW_CACHE_OP_READ<<8) | (PERF_COUNT_HW_CACHE_RESULT_MISS<<16); |
| 188 | mCountersFd[ICACHE_MISSES] = perf_event_open(&pe, 0, -1, groupFd, 0); |
| 189 | if (mCountersFd[ICACHE_MISSES] > 0) { |
| 190 | mIds[ICACHE_MISSES] = count++; |
| 191 | mEnabledEvents |= EV_L1I_MISSES; |
| 192 | } |
| 193 | } |
| 194 | #endif |
| 195 | } |
| 196 | #endif // __linux__ |
| 197 | return mEnabledEvents; |
| 198 | } |
| 199 | |
| 200 | } // namespace utils |