|
31 | 31 |
|
32 | 32 |
|
33 | 33 |
|
34 |
| -static int sched__get_first_possible_cpu(pid_t pid, cpu_set_t *maskp) |
35 |
| -{ |
36 |
| - int i, cpu = -1, nrcpus = 1024; |
37 |
| -realloc: |
38 |
| - CPU_ZERO(maskp); |
39 |
| - |
40 |
| - if (sched_getaffinity(pid, sizeof(*maskp), maskp) == -1) { |
41 |
| - if (errno == EINVAL && nrcpus < (1024 << 8)) { |
42 |
| - nrcpus = nrcpus << 2; |
43 |
| - goto realloc; |
44 |
| - } |
45 |
| - perror("sched_getaffinity"); |
46 |
| - return -1; |
47 |
| - } |
48 |
| - |
49 |
| - for (i = 0; i < nrcpus; i++) { |
50 |
| - if (CPU_ISSET(i, maskp)) { |
51 |
| - if (cpu == -1) |
52 |
| - cpu = i; |
53 |
| - else |
54 |
| - CPU_CLR(i, maskp); |
55 |
| - } |
56 |
| - } |
57 |
| - |
58 |
| - return cpu; |
59 |
| -} |
60 |
| - |
61 |
| -static int test__PERF_RECORD(void) |
62 |
| -{ |
63 |
| - struct perf_record_opts opts = { |
64 |
| - .target = { |
65 |
| - .uid = UINT_MAX, |
66 |
| - .uses_mmap = true, |
67 |
| - }, |
68 |
| - .no_delay = true, |
69 |
| - .freq = 10, |
70 |
| - .mmap_pages = 256, |
71 |
| - }; |
72 |
| - cpu_set_t cpu_mask; |
73 |
| - size_t cpu_mask_size = sizeof(cpu_mask); |
74 |
| - struct perf_evlist *evlist = perf_evlist__new(NULL, NULL); |
75 |
| - struct perf_evsel *evsel; |
76 |
| - struct perf_sample sample; |
77 |
| - const char *cmd = "sleep"; |
78 |
| - const char *argv[] = { cmd, "1", NULL, }; |
79 |
| - char *bname; |
80 |
| - u64 prev_time = 0; |
81 |
| - bool found_cmd_mmap = false, |
82 |
| - found_libc_mmap = false, |
83 |
| - found_vdso_mmap = false, |
84 |
| - found_ld_mmap = false; |
85 |
| - int err = -1, errs = 0, i, wakeups = 0; |
86 |
| - u32 cpu; |
87 |
| - int total_events = 0, nr_events[PERF_RECORD_MAX] = { 0, }; |
88 |
| - |
89 |
| - if (evlist == NULL || argv == NULL) { |
90 |
| - pr_debug("Not enough memory to create evlist\n"); |
91 |
| - goto out; |
92 |
| - } |
93 |
| - |
94 |
| - /* |
95 |
| - * We need at least one evsel in the evlist, use the default |
96 |
| - * one: "cycles". |
97 |
| - */ |
98 |
| - err = perf_evlist__add_default(evlist); |
99 |
| - if (err < 0) { |
100 |
| - pr_debug("Not enough memory to create evsel\n"); |
101 |
| - goto out_delete_evlist; |
102 |
| - } |
103 |
| - |
104 |
| - /* |
105 |
| - * Create maps of threads and cpus to monitor. In this case |
106 |
| - * we start with all threads and cpus (-1, -1) but then in |
107 |
| - * perf_evlist__prepare_workload we'll fill in the only thread |
108 |
| - * we're monitoring, the one forked there. |
109 |
| - */ |
110 |
| - err = perf_evlist__create_maps(evlist, &opts.target); |
111 |
| - if (err < 0) { |
112 |
| - pr_debug("Not enough memory to create thread/cpu maps\n"); |
113 |
| - goto out_delete_evlist; |
114 |
| - } |
115 |
| - |
116 |
| - /* |
117 |
| - * Prepare the workload in argv[] to run, it'll fork it, and then wait |
118 |
| - * for perf_evlist__start_workload() to exec it. This is done this way |
119 |
| - * so that we have time to open the evlist (calling sys_perf_event_open |
120 |
| - * on all the fds) and then mmap them. |
121 |
| - */ |
122 |
| - err = perf_evlist__prepare_workload(evlist, &opts, argv); |
123 |
| - if (err < 0) { |
124 |
| - pr_debug("Couldn't run the workload!\n"); |
125 |
| - goto out_delete_evlist; |
126 |
| - } |
127 |
| - |
128 |
| - /* |
129 |
| - * Config the evsels, setting attr->comm on the first one, etc. |
130 |
| - */ |
131 |
| - evsel = perf_evlist__first(evlist); |
132 |
| - evsel->attr.sample_type |= PERF_SAMPLE_CPU; |
133 |
| - evsel->attr.sample_type |= PERF_SAMPLE_TID; |
134 |
| - evsel->attr.sample_type |= PERF_SAMPLE_TIME; |
135 |
| - perf_evlist__config_attrs(evlist, &opts); |
136 |
| - |
137 |
| - err = sched__get_first_possible_cpu(evlist->workload.pid, &cpu_mask); |
138 |
| - if (err < 0) { |
139 |
| - pr_debug("sched__get_first_possible_cpu: %s\n", strerror(errno)); |
140 |
| - goto out_delete_evlist; |
141 |
| - } |
142 |
| - |
143 |
| - cpu = err; |
144 |
| - |
145 |
| - /* |
146 |
| - * So that we can check perf_sample.cpu on all the samples. |
147 |
| - */ |
148 |
| - if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, &cpu_mask) < 0) { |
149 |
| - pr_debug("sched_setaffinity: %s\n", strerror(errno)); |
150 |
| - goto out_delete_evlist; |
151 |
| - } |
152 |
| - |
153 |
| - /* |
154 |
| - * Call sys_perf_event_open on all the fds on all the evsels, |
155 |
| - * grouping them if asked to. |
156 |
| - */ |
157 |
| - err = perf_evlist__open(evlist); |
158 |
| - if (err < 0) { |
159 |
| - pr_debug("perf_evlist__open: %s\n", strerror(errno)); |
160 |
| - goto out_delete_evlist; |
161 |
| - } |
162 |
| - |
163 |
| - /* |
164 |
| - * mmap the first fd on a given CPU and ask for events for the other |
165 |
| - * fds in the same CPU to be injected in the same mmap ring buffer |
166 |
| - * (using ioctl(PERF_EVENT_IOC_SET_OUTPUT)). |
167 |
| - */ |
168 |
| - err = perf_evlist__mmap(evlist, opts.mmap_pages, false); |
169 |
| - if (err < 0) { |
170 |
| - pr_debug("perf_evlist__mmap: %s\n", strerror(errno)); |
171 |
| - goto out_delete_evlist; |
172 |
| - } |
173 |
| - |
174 |
| - /* |
175 |
| - * Now that all is properly set up, enable the events, they will |
176 |
| - * count just on workload.pid, which will start... |
177 |
| - */ |
178 |
| - perf_evlist__enable(evlist); |
179 |
| - |
180 |
| - /* |
181 |
| - * Now! |
182 |
| - */ |
183 |
| - perf_evlist__start_workload(evlist); |
184 |
| - |
185 |
| - while (1) { |
186 |
| - int before = total_events; |
187 |
| - |
188 |
| - for (i = 0; i < evlist->nr_mmaps; i++) { |
189 |
| - union perf_event *event; |
190 |
| - |
191 |
| - while ((event = perf_evlist__mmap_read(evlist, i)) != NULL) { |
192 |
| - const u32 type = event->header.type; |
193 |
| - const char *name = perf_event__name(type); |
194 |
| - |
195 |
| - ++total_events; |
196 |
| - if (type < PERF_RECORD_MAX) |
197 |
| - nr_events[type]++; |
198 |
| - |
199 |
| - err = perf_evlist__parse_sample(evlist, event, &sample); |
200 |
| - if (err < 0) { |
201 |
| - if (verbose) |
202 |
| - perf_event__fprintf(event, stderr); |
203 |
| - pr_debug("Couldn't parse sample\n"); |
204 |
| - goto out_err; |
205 |
| - } |
206 |
| - |
207 |
| - if (verbose) { |
208 |
| - pr_info("%" PRIu64" %d ", sample.time, sample.cpu); |
209 |
| - perf_event__fprintf(event, stderr); |
210 |
| - } |
211 |
| - |
212 |
| - if (prev_time > sample.time) { |
213 |
| - pr_debug("%s going backwards in time, prev=%" PRIu64 ", curr=%" PRIu64 "\n", |
214 |
| - name, prev_time, sample.time); |
215 |
| - ++errs; |
216 |
| - } |
217 |
| - |
218 |
| - prev_time = sample.time; |
219 |
| - |
220 |
| - if (sample.cpu != cpu) { |
221 |
| - pr_debug("%s with unexpected cpu, expected %d, got %d\n", |
222 |
| - name, cpu, sample.cpu); |
223 |
| - ++errs; |
224 |
| - } |
225 |
| - |
226 |
| - if ((pid_t)sample.pid != evlist->workload.pid) { |
227 |
| - pr_debug("%s with unexpected pid, expected %d, got %d\n", |
228 |
| - name, evlist->workload.pid, sample.pid); |
229 |
| - ++errs; |
230 |
| - } |
231 |
| - |
232 |
| - if ((pid_t)sample.tid != evlist->workload.pid) { |
233 |
| - pr_debug("%s with unexpected tid, expected %d, got %d\n", |
234 |
| - name, evlist->workload.pid, sample.tid); |
235 |
| - ++errs; |
236 |
| - } |
237 |
| - |
238 |
| - if ((type == PERF_RECORD_COMM || |
239 |
| - type == PERF_RECORD_MMAP || |
240 |
| - type == PERF_RECORD_FORK || |
241 |
| - type == PERF_RECORD_EXIT) && |
242 |
| - (pid_t)event->comm.pid != evlist->workload.pid) { |
243 |
| - pr_debug("%s with unexpected pid/tid\n", name); |
244 |
| - ++errs; |
245 |
| - } |
246 |
| - |
247 |
| - if ((type == PERF_RECORD_COMM || |
248 |
| - type == PERF_RECORD_MMAP) && |
249 |
| - event->comm.pid != event->comm.tid) { |
250 |
| - pr_debug("%s with different pid/tid!\n", name); |
251 |
| - ++errs; |
252 |
| - } |
253 |
| - |
254 |
| - switch (type) { |
255 |
| - case PERF_RECORD_COMM: |
256 |
| - if (strcmp(event->comm.comm, cmd)) { |
257 |
| - pr_debug("%s with unexpected comm!\n", name); |
258 |
| - ++errs; |
259 |
| - } |
260 |
| - break; |
261 |
| - case PERF_RECORD_EXIT: |
262 |
| - goto found_exit; |
263 |
| - case PERF_RECORD_MMAP: |
264 |
| - bname = strrchr(event->mmap.filename, '/'); |
265 |
| - if (bname != NULL) { |
266 |
| - if (!found_cmd_mmap) |
267 |
| - found_cmd_mmap = !strcmp(bname + 1, cmd); |
268 |
| - if (!found_libc_mmap) |
269 |
| - found_libc_mmap = !strncmp(bname + 1, "libc", 4); |
270 |
| - if (!found_ld_mmap) |
271 |
| - found_ld_mmap = !strncmp(bname + 1, "ld", 2); |
272 |
| - } else if (!found_vdso_mmap) |
273 |
| - found_vdso_mmap = !strcmp(event->mmap.filename, "[vdso]"); |
274 |
| - break; |
275 |
| - |
276 |
| - case PERF_RECORD_SAMPLE: |
277 |
| - /* Just ignore samples for now */ |
278 |
| - break; |
279 |
| - default: |
280 |
| - pr_debug("Unexpected perf_event->header.type %d!\n", |
281 |
| - type); |
282 |
| - ++errs; |
283 |
| - } |
284 |
| - } |
285 |
| - } |
286 |
| - |
287 |
| - /* |
288 |
| - * We don't use poll here because at least at 3.1 times the |
289 |
| - * PERF_RECORD_{!SAMPLE} events don't honour |
290 |
| - * perf_event_attr.wakeup_events, just PERF_EVENT_SAMPLE does. |
291 |
| - */ |
292 |
| - if (total_events == before && false) |
293 |
| - poll(evlist->pollfd, evlist->nr_fds, -1); |
294 |
| - |
295 |
| - sleep(1); |
296 |
| - if (++wakeups > 5) { |
297 |
| - pr_debug("No PERF_RECORD_EXIT event!\n"); |
298 |
| - break; |
299 |
| - } |
300 |
| - } |
301 |
| - |
302 |
| -found_exit: |
303 |
| - if (nr_events[PERF_RECORD_COMM] > 1) { |
304 |
| - pr_debug("Excessive number of PERF_RECORD_COMM events!\n"); |
305 |
| - ++errs; |
306 |
| - } |
307 |
| - |
308 |
| - if (nr_events[PERF_RECORD_COMM] == 0) { |
309 |
| - pr_debug("Missing PERF_RECORD_COMM for %s!\n", cmd); |
310 |
| - ++errs; |
311 |
| - } |
312 |
| - |
313 |
| - if (!found_cmd_mmap) { |
314 |
| - pr_debug("PERF_RECORD_MMAP for %s missing!\n", cmd); |
315 |
| - ++errs; |
316 |
| - } |
317 |
| - |
318 |
| - if (!found_libc_mmap) { |
319 |
| - pr_debug("PERF_RECORD_MMAP for %s missing!\n", "libc"); |
320 |
| - ++errs; |
321 |
| - } |
322 |
| - |
323 |
| - if (!found_ld_mmap) { |
324 |
| - pr_debug("PERF_RECORD_MMAP for %s missing!\n", "ld"); |
325 |
| - ++errs; |
326 |
| - } |
327 |
| - |
328 |
| - if (!found_vdso_mmap) { |
329 |
| - pr_debug("PERF_RECORD_MMAP for %s missing!\n", "[vdso]"); |
330 |
| - ++errs; |
331 |
| - } |
332 |
| -out_err: |
333 |
| - perf_evlist__munmap(evlist); |
334 |
| -out_delete_evlist: |
335 |
| - perf_evlist__delete(evlist); |
336 |
| -out: |
337 |
| - return (err < 0 || errs > 0) ? -1 : 0; |
338 |
| -} |
339 |
| - |
340 |
| - |
341 | 34 | #if defined(__x86_64__) || defined(__i386__)
|
342 | 35 |
|
343 | 36 | #define barrier() asm volatile("" ::: "memory")
|
|
0 commit comments