~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/tools/lib/perf/evsel.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0
  2 #include <errno.h>
  3 #include <unistd.h>
  4 #include <sys/syscall.h>
  5 #include <perf/evsel.h>
  6 #include <perf/cpumap.h>
  7 #include <perf/threadmap.h>
  8 #include <linux/list.h>
  9 #include <internal/evsel.h>
 10 #include <linux/zalloc.h>
 11 #include <stdlib.h>
 12 #include <internal/xyarray.h>
 13 #include <internal/cpumap.h>
 14 #include <internal/mmap.h>
 15 #include <internal/threadmap.h>
 16 #include <internal/lib.h>
 17 #include <linux/string.h>
 18 #include <sys/ioctl.h>
 19 #include <sys/mman.h>
 20 #include <asm/bug.h>
 21 
 22 void perf_evsel__init(struct perf_evsel *evsel, struct perf_event_attr *attr,
 23                       int idx)
 24 {
 25         INIT_LIST_HEAD(&evsel->node);
 26         evsel->attr = *attr;
 27         evsel->idx  = idx;
 28         evsel->leader = evsel;
 29 }
 30 
 31 struct perf_evsel *perf_evsel__new(struct perf_event_attr *attr)
 32 {
 33         struct perf_evsel *evsel = zalloc(sizeof(*evsel));
 34 
 35         if (evsel != NULL)
 36                 perf_evsel__init(evsel, attr, 0);
 37 
 38         return evsel;
 39 }
 40 
 41 void perf_evsel__delete(struct perf_evsel *evsel)
 42 {
 43         free(evsel);
 44 }
 45 
 46 #define FD(_evsel, _cpu_map_idx, _thread)                               \
 47         ((int *)xyarray__entry(_evsel->fd, _cpu_map_idx, _thread))
 48 #define MMAP(_evsel, _cpu_map_idx, _thread)                             \
 49         (_evsel->mmap ? ((struct perf_mmap *) xyarray__entry(_evsel->mmap, _cpu_map_idx, _thread)) \
 50                       : NULL)
 51 
 52 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads)
 53 {
 54         evsel->fd = xyarray__new(ncpus, nthreads, sizeof(int));
 55 
 56         if (evsel->fd) {
 57                 int idx, thread;
 58 
 59                 for (idx = 0; idx < ncpus; idx++) {
 60                         for (thread = 0; thread < nthreads; thread++) {
 61                                 int *fd = FD(evsel, idx, thread);
 62 
 63                                 if (fd)
 64                                         *fd = -1;
 65                         }
 66                 }
 67         }
 68 
 69         return evsel->fd != NULL ? 0 : -ENOMEM;
 70 }
 71 
 72 static int perf_evsel__alloc_mmap(struct perf_evsel *evsel, int ncpus, int nthreads)
 73 {
 74         evsel->mmap = xyarray__new(ncpus, nthreads, sizeof(struct perf_mmap));
 75 
 76         return evsel->mmap != NULL ? 0 : -ENOMEM;
 77 }
 78 
 79 static int
 80 sys_perf_event_open(struct perf_event_attr *attr,
 81                     pid_t pid, struct perf_cpu cpu, int group_fd,
 82                     unsigned long flags)
 83 {
 84         return syscall(__NR_perf_event_open, attr, pid, cpu.cpu, group_fd, flags);
 85 }
 86 
 87 static int get_group_fd(struct perf_evsel *evsel, int cpu_map_idx, int thread, int *group_fd)
 88 {
 89         struct perf_evsel *leader = evsel->leader;
 90         int *fd;
 91 
 92         if (evsel == leader) {
 93                 *group_fd = -1;
 94                 return 0;
 95         }
 96 
 97         /*
 98          * Leader must be already processed/open,
 99          * if not it's a bug.
100          */
101         if (!leader->fd)
102                 return -ENOTCONN;
103 
104         fd = FD(leader, cpu_map_idx, thread);
105         if (fd == NULL || *fd == -1)
106                 return -EBADF;
107 
108         *group_fd = *fd;
109 
110         return 0;
111 }
112 
113 int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
114                      struct perf_thread_map *threads)
115 {
116         struct perf_cpu cpu;
117         int idx, thread, err = 0;
118 
119         if (cpus == NULL) {
120                 static struct perf_cpu_map *empty_cpu_map;
121 
122                 if (empty_cpu_map == NULL) {
123                         empty_cpu_map = perf_cpu_map__new_any_cpu();
124                         if (empty_cpu_map == NULL)
125                                 return -ENOMEM;
126                 }
127 
128                 cpus = empty_cpu_map;
129         }
130 
131         if (threads == NULL) {
132                 static struct perf_thread_map *empty_thread_map;
133 
134                 if (empty_thread_map == NULL) {
135                         empty_thread_map = perf_thread_map__new_dummy();
136                         if (empty_thread_map == NULL)
137                                 return -ENOMEM;
138                 }
139 
140                 threads = empty_thread_map;
141         }
142 
143         if (evsel->fd == NULL &&
144             perf_evsel__alloc_fd(evsel, perf_cpu_map__nr(cpus), threads->nr) < 0)
145                 return -ENOMEM;
146 
147         perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
148                 for (thread = 0; thread < threads->nr; thread++) {
149                         int fd, group_fd, *evsel_fd;
150 
151                         evsel_fd = FD(evsel, idx, thread);
152                         if (evsel_fd == NULL) {
153                                 err = -EINVAL;
154                                 goto out;
155                         }
156 
157                         err = get_group_fd(evsel, idx, thread, &group_fd);
158                         if (err < 0)
159                                 goto out;
160 
161                         fd = sys_perf_event_open(&evsel->attr,
162                                                  threads->map[thread].pid,
163                                                  cpu, group_fd, 0);
164 
165                         if (fd < 0) {
166                                 err = -errno;
167                                 goto out;
168                         }
169 
170                         *evsel_fd = fd;
171                 }
172         }
173 out:
174         if (err)
175                 perf_evsel__close(evsel);
176 
177         return err;
178 }
179 
180 static void perf_evsel__close_fd_cpu(struct perf_evsel *evsel, int cpu_map_idx)
181 {
182         int thread;
183 
184         for (thread = 0; thread < xyarray__max_y(evsel->fd); ++thread) {
185                 int *fd = FD(evsel, cpu_map_idx, thread);
186 
187                 if (fd && *fd >= 0) {
188                         close(*fd);
189                         *fd = -1;
190                 }
191         }
192 }
193 
194 void perf_evsel__close_fd(struct perf_evsel *evsel)
195 {
196         for (int idx = 0; idx < xyarray__max_x(evsel->fd); idx++)
197                 perf_evsel__close_fd_cpu(evsel, idx);
198 }
199 
200 void perf_evsel__free_fd(struct perf_evsel *evsel)
201 {
202         xyarray__delete(evsel->fd);
203         evsel->fd = NULL;
204 }
205 
206 void perf_evsel__close(struct perf_evsel *evsel)
207 {
208         if (evsel->fd == NULL)
209                 return;
210 
211         perf_evsel__close_fd(evsel);
212         perf_evsel__free_fd(evsel);
213 }
214 
215 void perf_evsel__close_cpu(struct perf_evsel *evsel, int cpu_map_idx)
216 {
217         if (evsel->fd == NULL)
218                 return;
219 
220         perf_evsel__close_fd_cpu(evsel, cpu_map_idx);
221 }
222 
223 void perf_evsel__munmap(struct perf_evsel *evsel)
224 {
225         int idx, thread;
226 
227         if (evsel->fd == NULL || evsel->mmap == NULL)
228                 return;
229 
230         for (idx = 0; idx < xyarray__max_x(evsel->fd); idx++) {
231                 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) {
232                         int *fd = FD(evsel, idx, thread);
233 
234                         if (fd == NULL || *fd < 0)
235                                 continue;
236 
237                         perf_mmap__munmap(MMAP(evsel, idx, thread));
238                 }
239         }
240 
241         xyarray__delete(evsel->mmap);
242         evsel->mmap = NULL;
243 }
244 
245 int perf_evsel__mmap(struct perf_evsel *evsel, int pages)
246 {
247         int ret, idx, thread;
248         struct perf_mmap_param mp = {
249                 .prot = PROT_READ | PROT_WRITE,
250                 .mask = (pages * page_size) - 1,
251         };
252 
253         if (evsel->fd == NULL || evsel->mmap)
254                 return -EINVAL;
255 
256         if (perf_evsel__alloc_mmap(evsel, xyarray__max_x(evsel->fd), xyarray__max_y(evsel->fd)) < 0)
257                 return -ENOMEM;
258 
259         for (idx = 0; idx < xyarray__max_x(evsel->fd); idx++) {
260                 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) {
261                         int *fd = FD(evsel, idx, thread);
262                         struct perf_mmap *map;
263                         struct perf_cpu cpu = perf_cpu_map__cpu(evsel->cpus, idx);
264 
265                         if (fd == NULL || *fd < 0)
266                                 continue;
267 
268                         map = MMAP(evsel, idx, thread);
269                         perf_mmap__init(map, NULL, false, NULL);
270 
271                         ret = perf_mmap__mmap(map, &mp, *fd, cpu);
272                         if (ret) {
273                                 perf_evsel__munmap(evsel);
274                                 return ret;
275                         }
276                 }
277         }
278 
279         return 0;
280 }
281 
282 void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu_map_idx, int thread)
283 {
284         int *fd = FD(evsel, cpu_map_idx, thread);
285 
286         if (fd == NULL || *fd < 0 || MMAP(evsel, cpu_map_idx, thread) == NULL)
287                 return NULL;
288 
289         return MMAP(evsel, cpu_map_idx, thread)->base;
290 }
291 
292 int perf_evsel__read_size(struct perf_evsel *evsel)
293 {
294         u64 read_format = evsel->attr.read_format;
295         int entry = sizeof(u64); /* value */
296         int size = 0;
297         int nr = 1;
298 
299         if (read_format & PERF_FORMAT_TOTAL_TIME_ENABLED)
300                 size += sizeof(u64);
301 
302         if (read_format & PERF_FORMAT_TOTAL_TIME_RUNNING)
303                 size += sizeof(u64);
304 
305         if (read_format & PERF_FORMAT_ID)
306                 entry += sizeof(u64);
307 
308         if (read_format & PERF_FORMAT_LOST)
309                 entry += sizeof(u64);
310 
311         if (read_format & PERF_FORMAT_GROUP) {
312                 nr = evsel->nr_members;
313                 size += sizeof(u64);
314         }
315 
316         size += entry * nr;
317         return size;
318 }
319 
320 /* This only reads values for the leader */
321 static int perf_evsel__read_group(struct perf_evsel *evsel, int cpu_map_idx,
322                                   int thread, struct perf_counts_values *count)
323 {
324         size_t size = perf_evsel__read_size(evsel);
325         int *fd = FD(evsel, cpu_map_idx, thread);
326         u64 read_format = evsel->attr.read_format;
327         u64 *data;
328         int idx = 1;
329 
330         if (fd == NULL || *fd < 0)
331                 return -EINVAL;
332 
333         data = calloc(1, size);
334         if (data == NULL)
335                 return -ENOMEM;
336 
337         if (readn(*fd, data, size) <= 0) {
338                 free(data);
339                 return -errno;
340         }
341 
342         /*
343          * This reads only the leader event intentionally since we don't have
344          * perf counts values for sibling events.
345          */
346         if (read_format & PERF_FORMAT_TOTAL_TIME_ENABLED)
347                 count->ena = data[idx++];
348         if (read_format & PERF_FORMAT_TOTAL_TIME_RUNNING)
349                 count->run = data[idx++];
350 
351         /* value is always available */
352         count->val = data[idx++];
353         if (read_format & PERF_FORMAT_ID)
354                 count->id = data[idx++];
355         if (read_format & PERF_FORMAT_LOST)
356                 count->lost = data[idx++];
357 
358         free(data);
359         return 0;
360 }
361 
362 /*
363  * The perf read format is very flexible.  It needs to set the proper
364  * values according to the read format.
365  */
366 static void perf_evsel__adjust_values(struct perf_evsel *evsel, u64 *buf,
367                                       struct perf_counts_values *count)
368 {
369         u64 read_format = evsel->attr.read_format;
370         int n = 0;
371 
372         count->val = buf[n++];
373 
374         if (read_format & PERF_FORMAT_TOTAL_TIME_ENABLED)
375                 count->ena = buf[n++];
376 
377         if (read_format & PERF_FORMAT_TOTAL_TIME_RUNNING)
378                 count->run = buf[n++];
379 
380         if (read_format & PERF_FORMAT_ID)
381                 count->id = buf[n++];
382 
383         if (read_format & PERF_FORMAT_LOST)
384                 count->lost = buf[n++];
385 }
386 
387 int perf_evsel__read(struct perf_evsel *evsel, int cpu_map_idx, int thread,
388                      struct perf_counts_values *count)
389 {
390         size_t size = perf_evsel__read_size(evsel);
391         int *fd = FD(evsel, cpu_map_idx, thread);
392         u64 read_format = evsel->attr.read_format;
393         struct perf_counts_values buf;
394 
395         memset(count, 0, sizeof(*count));
396 
397         if (fd == NULL || *fd < 0)
398                 return -EINVAL;
399 
400         if (read_format & PERF_FORMAT_GROUP)
401                 return perf_evsel__read_group(evsel, cpu_map_idx, thread, count);
402 
403         if (MMAP(evsel, cpu_map_idx, thread) &&
404             !(read_format & (PERF_FORMAT_ID | PERF_FORMAT_LOST)) &&
405             !perf_mmap__read_self(MMAP(evsel, cpu_map_idx, thread), count))
406                 return 0;
407 
408         if (readn(*fd, buf.values, size) <= 0)
409                 return -errno;
410 
411         perf_evsel__adjust_values(evsel, buf.values, count);
412         return 0;
413 }
414 
415 static int perf_evsel__ioctl(struct perf_evsel *evsel, int ioc, void *arg,
416                              int cpu_map_idx, int thread)
417 {
418         int *fd = FD(evsel, cpu_map_idx, thread);
419 
420         if (fd == NULL || *fd < 0)
421                 return -1;
422 
423         return ioctl(*fd, ioc, arg);
424 }
425 
426 static int perf_evsel__run_ioctl(struct perf_evsel *evsel,
427                                  int ioc,  void *arg,
428                                  int cpu_map_idx)
429 {
430         int thread;
431 
432         for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) {
433                 int err = perf_evsel__ioctl(evsel, ioc, arg, cpu_map_idx, thread);
434 
435                 if (err)
436                         return err;
437         }
438 
439         return 0;
440 }
441 
442 int perf_evsel__enable_cpu(struct perf_evsel *evsel, int cpu_map_idx)
443 {
444         return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, cpu_map_idx);
445 }
446 
447 int perf_evsel__enable_thread(struct perf_evsel *evsel, int thread)
448 {
449         struct perf_cpu cpu __maybe_unused;
450         int idx;
451         int err;
452 
453         perf_cpu_map__for_each_cpu(cpu, idx, evsel->cpus) {
454                 err = perf_evsel__ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, idx, thread);
455                 if (err)
456                         return err;
457         }
458 
459         return 0;
460 }
461 
462 int perf_evsel__enable(struct perf_evsel *evsel)
463 {
464         int i;
465         int err = 0;
466 
467         for (i = 0; i < xyarray__max_x(evsel->fd) && !err; i++)
468                 err = perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, i);
469         return err;
470 }
471 
472 int perf_evsel__disable_cpu(struct perf_evsel *evsel, int cpu_map_idx)
473 {
474         return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_DISABLE, NULL, cpu_map_idx);
475 }
476 
477 int perf_evsel__disable(struct perf_evsel *evsel)
478 {
479         int i;
480         int err = 0;
481 
482         for (i = 0; i < xyarray__max_x(evsel->fd) && !err; i++)
483                 err = perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_DISABLE, NULL, i);
484         return err;
485 }
486 
487 int perf_evsel__apply_filter(struct perf_evsel *evsel, const char *filter)
488 {
489         int err = 0, i;
490 
491         for (i = 0; i < perf_cpu_map__nr(evsel->cpus) && !err; i++)
492                 err = perf_evsel__run_ioctl(evsel,
493                                      PERF_EVENT_IOC_SET_FILTER,
494                                      (void *)filter, i);
495         return err;
496 }
497 
498 struct perf_cpu_map *perf_evsel__cpus(struct perf_evsel *evsel)
499 {
500         return evsel->cpus;
501 }
502 
503 struct perf_thread_map *perf_evsel__threads(struct perf_evsel *evsel)
504 {
505         return evsel->threads;
506 }
507 
508 struct perf_event_attr *perf_evsel__attr(struct perf_evsel *evsel)
509 {
510         return &evsel->attr;
511 }
512 
513 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads)
514 {
515         if (ncpus == 0 || nthreads == 0)
516                 return 0;
517 
518         evsel->sample_id = xyarray__new(ncpus, nthreads, sizeof(struct perf_sample_id));
519         if (evsel->sample_id == NULL)
520                 return -ENOMEM;
521 
522         evsel->id = zalloc(ncpus * nthreads * sizeof(u64));
523         if (evsel->id == NULL) {
524                 xyarray__delete(evsel->sample_id);
525                 evsel->sample_id = NULL;
526                 return -ENOMEM;
527         }
528 
529         return 0;
530 }
531 
532 void perf_evsel__free_id(struct perf_evsel *evsel)
533 {
534         xyarray__delete(evsel->sample_id);
535         evsel->sample_id = NULL;
536         zfree(&evsel->id);
537         evsel->ids = 0;
538 }
539 
540 void perf_counts_values__scale(struct perf_counts_values *count,
541                                bool scale, __s8 *pscaled)
542 {
543         s8 scaled = 0;
544 
545         if (scale) {
546                 if (count->run == 0) {
547                         scaled = -1;
548                         count->val = 0;
549                 } else if (count->run < count->ena) {
550                         scaled = 1;
551                         count->val = (u64)((double)count->val * count->ena / count->run);
552                 }
553         }
554 
555         if (pscaled)
556                 *pscaled = scaled;
557 }
558 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php