1 /* SPDX-License-Identifier: GPL-2.0 */
3 #define __PERF_EVSEL_H 1
5 #include <linux/list.h>
9 #include <linux/perf_event.h>
10 #include <linux/types.h>
11 #include <internal/evsel.h>
12 #include <perf/evsel.h>
13 #include "symbol_conf.h"
14 #include <internal/cpumap.h>
21 * Per fd, to map back from PERF_SAMPLE_ID to evsel, only used when there are
22 * more than one entry in the evlist.
24 struct perf_sample_id
{
25 struct hlist_node node
;
29 * 'idx' will be used for AUX area sampling. A sample will have AUX area
30 * data that will be queued for decoding, where there are separate
31 * queues for each CPU (per-cpu tracing) or task (per-thread tracing).
32 * The sample ID can be used to lookup 'idx' which is effectively the
39 /* Holds total ID period value for PERF_SAMPLE_READ processing. */
46 * The 'struct perf_evsel_config_term' is used to pass event
47 * specific configuration data to perf_evsel__config routine.
48 * It is allocated within event parsing and attached to
49 * perf_evsel::config_terms list head.
52 PERF_EVSEL__CONFIG_TERM_PERIOD
,
53 PERF_EVSEL__CONFIG_TERM_FREQ
,
54 PERF_EVSEL__CONFIG_TERM_TIME
,
55 PERF_EVSEL__CONFIG_TERM_CALLGRAPH
,
56 PERF_EVSEL__CONFIG_TERM_STACK_USER
,
57 PERF_EVSEL__CONFIG_TERM_INHERIT
,
58 PERF_EVSEL__CONFIG_TERM_MAX_STACK
,
59 PERF_EVSEL__CONFIG_TERM_MAX_EVENTS
,
60 PERF_EVSEL__CONFIG_TERM_OVERWRITE
,
61 PERF_EVSEL__CONFIG_TERM_DRV_CFG
,
62 PERF_EVSEL__CONFIG_TERM_BRANCH
,
63 PERF_EVSEL__CONFIG_TERM_PERCORE
,
64 PERF_EVSEL__CONFIG_TERM_AUX_OUTPUT
,
67 struct perf_evsel_config_term
{
68 struct list_head list
;
81 unsigned long max_events
;
88 struct perf_stat_evsel
;
90 typedef int (perf_evsel__sb_cb_t
)(union perf_event
*event
, void *data
);
92 enum perf_tool_event
{
94 PERF_TOOL_DURATION_TIME
= 1,
101 /** struct evsel - event selector
103 * @evlist - evlist this evsel is in, if it is in one.
104 * @core - libperf evsel object
105 * @name - Can be set to retain the original event name passed by the user,
106 * so that when showing results in tools such as 'perf stat', we
107 * show the name used, not some alias.
108 * @id_pos: the position of the event id (PERF_SAMPLE_ID or
109 * PERF_SAMPLE_IDENTIFIER) in a sample event i.e. in the array of
110 * struct perf_record_sample
111 * @is_pos: the position (counting backwards) of the event id (PERF_SAMPLE_ID or
112 * PERF_SAMPLE_IDENTIFIER) in a non-sample event i.e. if sample_id_all
113 * is used there is an id sample appended to non-sample events
114 * @priv: And what is in its containing unnamed union are tool specific
117 struct perf_evsel core
;
118 struct evlist
*evlist
;
120 struct xyarray
*sample_id
;
122 struct perf_counts
*counts
;
123 struct perf_counts
*prev_raw_counts
;
126 unsigned long max_events
;
127 unsigned long nr_events_printed
;
131 struct tep_event
*tp_format
;
133 struct perf_stat_evsel
*stats
;
138 unsigned int sample_size
;
141 enum perf_tool_event tool_event
;
142 bool uniquified_name
;
153 bool ignore_missing_thread
;
155 bool use_uncore_alias
;
156 /* parse modifier helper */
159 unsigned long *per_pkg_mask
;
160 struct evsel
*leader
;
162 bool cmdline_group_boundary
;
163 struct list_head config_terms
;
164 struct bpf_object
*bpf_obj
;
166 bool auto_merge_stats
;
168 const char * metric_expr
;
169 const char * metric_name
;
170 struct evsel
**metric_events
;
171 struct evsel
*metric_leader
;
175 const char *pmu_name
;
177 perf_evsel__sb_cb_t
*cb
;
187 struct perf_missing_features
{
202 extern struct perf_missing_features perf_missing_features
;
209 static inline struct perf_cpu_map
*evsel__cpus(struct evsel
*evsel
)
211 return perf_evsel__cpus(&evsel
->core
);
214 static inline int perf_evsel__nr_cpus(struct evsel
*evsel
)
216 return evsel__cpus(evsel
)->nr
;
219 void perf_counts_values__scale(struct perf_counts_values
*count
,
220 bool scale
, s8
*pscaled
);
222 void perf_evsel__compute_deltas(struct evsel
*evsel
, int cpu
, int thread
,
223 struct perf_counts_values
*count
);
225 int perf_evsel__object_config(size_t object_size
,
226 int (*init
)(struct evsel
*evsel
),
227 void (*fini
)(struct evsel
*evsel
));
229 struct evsel
*perf_evsel__new_idx(struct perf_event_attr
*attr
, int idx
);
231 static inline struct evsel
*evsel__new(struct perf_event_attr
*attr
)
233 return perf_evsel__new_idx(attr
, 0);
236 struct evsel
*perf_evsel__newtp_idx(const char *sys
, const char *name
, int idx
);
239 * Returns pointer with encoded error via <linux/err.h> interface.
241 static inline struct evsel
*perf_evsel__newtp(const char *sys
, const char *name
)
243 return perf_evsel__newtp_idx(sys
, name
, 0);
246 struct evsel
*perf_evsel__new_cycles(bool precise
);
248 struct tep_event
*event_format__new(const char *sys
, const char *name
);
250 void evsel__init(struct evsel
*evsel
, struct perf_event_attr
*attr
, int idx
);
251 void perf_evsel__exit(struct evsel
*evsel
);
252 void evsel__delete(struct evsel
*evsel
);
254 struct callchain_param
;
256 void perf_evsel__config(struct evsel
*evsel
,
257 struct record_opts
*opts
,
258 struct callchain_param
*callchain
);
259 void perf_evsel__config_callchain(struct evsel
*evsel
,
260 struct record_opts
*opts
,
261 struct callchain_param
*callchain
);
263 int __perf_evsel__sample_size(u64 sample_type
);
264 void perf_evsel__calc_id_pos(struct evsel
*evsel
);
266 bool perf_evsel__is_cache_op_valid(u8 type
, u8 op
);
268 #define PERF_EVSEL__MAX_ALIASES 8
270 extern const char *perf_evsel__hw_cache
[PERF_COUNT_HW_CACHE_MAX
]
271 [PERF_EVSEL__MAX_ALIASES
];
272 extern const char *perf_evsel__hw_cache_op
[PERF_COUNT_HW_CACHE_OP_MAX
]
273 [PERF_EVSEL__MAX_ALIASES
];
274 extern const char *perf_evsel__hw_cache_result
[PERF_COUNT_HW_CACHE_RESULT_MAX
]
275 [PERF_EVSEL__MAX_ALIASES
];
276 extern const char *perf_evsel__hw_names
[PERF_COUNT_HW_MAX
];
277 extern const char *perf_evsel__sw_names
[PERF_COUNT_SW_MAX
];
278 int __perf_evsel__hw_cache_type_op_res_name(u8 type
, u8 op
, u8 result
,
279 char *bf
, size_t size
);
280 const char *perf_evsel__name(struct evsel
*evsel
);
282 const char *perf_evsel__group_name(struct evsel
*evsel
);
283 int perf_evsel__group_desc(struct evsel
*evsel
, char *buf
, size_t size
);
285 int perf_evsel__alloc_id(struct evsel
*evsel
, int ncpus
, int nthreads
);
287 void __perf_evsel__set_sample_bit(struct evsel
*evsel
,
288 enum perf_event_sample_format bit
);
289 void __perf_evsel__reset_sample_bit(struct evsel
*evsel
,
290 enum perf_event_sample_format bit
);
292 #define perf_evsel__set_sample_bit(evsel, bit) \
293 __perf_evsel__set_sample_bit(evsel, PERF_SAMPLE_##bit)
295 #define perf_evsel__reset_sample_bit(evsel, bit) \
296 __perf_evsel__reset_sample_bit(evsel, PERF_SAMPLE_##bit)
298 void perf_evsel__set_sample_id(struct evsel
*evsel
,
299 bool use_sample_identifier
);
301 int perf_evsel__set_filter(struct evsel
*evsel
, const char *filter
);
302 int perf_evsel__append_tp_filter(struct evsel
*evsel
, const char *filter
);
303 int perf_evsel__append_addr_filter(struct evsel
*evsel
,
305 int evsel__enable(struct evsel
*evsel
);
306 int evsel__disable(struct evsel
*evsel
);
308 int perf_evsel__open_per_cpu(struct evsel
*evsel
,
309 struct perf_cpu_map
*cpus
);
310 int perf_evsel__open_per_thread(struct evsel
*evsel
,
311 struct perf_thread_map
*threads
);
312 int evsel__open(struct evsel
*evsel
, struct perf_cpu_map
*cpus
,
313 struct perf_thread_map
*threads
);
314 void evsel__close(struct evsel
*evsel
);
318 void *perf_evsel__rawptr(struct evsel
*evsel
, struct perf_sample
*sample
,
320 u64
perf_evsel__intval(struct evsel
*evsel
, struct perf_sample
*sample
,
323 static inline char *perf_evsel__strval(struct evsel
*evsel
,
324 struct perf_sample
*sample
,
327 return perf_evsel__rawptr(evsel
, sample
, name
);
330 struct tep_format_field
;
332 u64
format_field__intval(struct tep_format_field
*field
, struct perf_sample
*sample
, bool needs_swap
);
334 struct tep_format_field
*perf_evsel__field(struct evsel
*evsel
, const char *name
);
336 #define perf_evsel__match(evsel, t, c) \
337 (evsel->core.attr.type == PERF_TYPE_##t && \
338 evsel->core.attr.config == PERF_COUNT_##c)
340 static inline bool perf_evsel__match2(struct evsel
*e1
,
343 return (e1
->core
.attr
.type
== e2
->core
.attr
.type
) &&
344 (e1
->core
.attr
.config
== e2
->core
.attr
.config
);
347 #define perf_evsel__cmp(a, b) \
350 (a)->core.attr.type == (b)->core.attr.type && \
351 (a)->core.attr.config == (b)->core.attr.config)
353 int perf_evsel__read_counter(struct evsel
*evsel
, int cpu
, int thread
);
355 int __perf_evsel__read_on_cpu(struct evsel
*evsel
,
356 int cpu
, int thread
, bool scale
);
359 * perf_evsel__read_on_cpu - Read out the results on a CPU and thread
361 * @evsel - event selector to read value
362 * @cpu - CPU of interest
363 * @thread - thread of interest
365 static inline int perf_evsel__read_on_cpu(struct evsel
*evsel
,
368 return __perf_evsel__read_on_cpu(evsel
, cpu
, thread
, false);
372 * perf_evsel__read_on_cpu_scaled - Read out the results on a CPU and thread, scaled
374 * @evsel - event selector to read value
375 * @cpu - CPU of interest
376 * @thread - thread of interest
378 static inline int perf_evsel__read_on_cpu_scaled(struct evsel
*evsel
,
381 return __perf_evsel__read_on_cpu(evsel
, cpu
, thread
, true);
384 int perf_evsel__parse_sample(struct evsel
*evsel
, union perf_event
*event
,
385 struct perf_sample
*sample
);
387 int perf_evsel__parse_sample_timestamp(struct evsel
*evsel
,
388 union perf_event
*event
,
391 static inline struct evsel
*perf_evsel__next(struct evsel
*evsel
)
393 return list_entry(evsel
->core
.node
.next
, struct evsel
, core
.node
);
396 static inline struct evsel
*perf_evsel__prev(struct evsel
*evsel
)
398 return list_entry(evsel
->core
.node
.prev
, struct evsel
, core
.node
);
402 * perf_evsel__is_group_leader - Return whether given evsel is a leader event
404 * @evsel - evsel selector to be tested
406 * Return %true if @evsel is a group leader or a stand-alone event
408 static inline bool perf_evsel__is_group_leader(const struct evsel
*evsel
)
410 return evsel
->leader
== evsel
;
414 * perf_evsel__is_group_event - Return whether given evsel is a group event
416 * @evsel - evsel selector to be tested
418 * Return %true iff event group view is enabled and @evsel is a actual group
419 * leader which has other members in the group
421 static inline bool perf_evsel__is_group_event(struct evsel
*evsel
)
423 if (!symbol_conf
.event_group
)
426 return perf_evsel__is_group_leader(evsel
) && evsel
->core
.nr_members
> 1;
429 bool perf_evsel__is_function_event(struct evsel
*evsel
);
431 static inline bool perf_evsel__is_bpf_output(struct evsel
*evsel
)
433 return perf_evsel__match(evsel
, SOFTWARE
, SW_BPF_OUTPUT
);
436 static inline bool perf_evsel__is_clock(struct evsel
*evsel
)
438 return perf_evsel__match(evsel
, SOFTWARE
, SW_CPU_CLOCK
) ||
439 perf_evsel__match(evsel
, SOFTWARE
, SW_TASK_CLOCK
);
442 struct perf_attr_details
{
450 int perf_evsel__fprintf(struct evsel
*evsel
,
451 struct perf_attr_details
*details
, FILE *fp
);
453 #define EVSEL__PRINT_IP (1<<0)
454 #define EVSEL__PRINT_SYM (1<<1)
455 #define EVSEL__PRINT_DSO (1<<2)
456 #define EVSEL__PRINT_SYMOFFSET (1<<3)
457 #define EVSEL__PRINT_ONELINE (1<<4)
458 #define EVSEL__PRINT_SRCLINE (1<<5)
459 #define EVSEL__PRINT_UNKNOWN_AS_ADDR (1<<6)
460 #define EVSEL__PRINT_CALLCHAIN_ARROW (1<<7)
461 #define EVSEL__PRINT_SKIP_IGNORED (1<<8)
463 struct callchain_cursor
;
465 int sample__fprintf_callchain(struct perf_sample
*sample
, int left_alignment
,
466 unsigned int print_opts
,
467 struct callchain_cursor
*cursor
, FILE *fp
);
469 int sample__fprintf_sym(struct perf_sample
*sample
, struct addr_location
*al
,
470 int left_alignment
, unsigned int print_opts
,
471 struct callchain_cursor
*cursor
, FILE *fp
);
473 bool perf_evsel__fallback(struct evsel
*evsel
, int err
,
474 char *msg
, size_t msgsize
);
475 int perf_evsel__open_strerror(struct evsel
*evsel
, struct target
*target
,
476 int err
, char *msg
, size_t size
);
478 static inline int perf_evsel__group_idx(struct evsel
*evsel
)
480 return evsel
->idx
- evsel
->leader
->idx
;
483 /* Iterates group WITHOUT the leader. */
484 #define for_each_group_member(_evsel, _leader) \
485 for ((_evsel) = list_entry((_leader)->core.node.next, struct evsel, core.node); \
486 (_evsel) && (_evsel)->leader == (_leader); \
487 (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
489 /* Iterates group WITH the leader. */
490 #define for_each_group_evsel(_evsel, _leader) \
491 for ((_evsel) = _leader; \
492 (_evsel) && (_evsel)->leader == (_leader); \
493 (_evsel) = list_entry((_evsel)->core.node.next, struct evsel, core.node))
495 static inline bool perf_evsel__has_branch_callstack(const struct evsel
*evsel
)
497 return evsel
->core
.attr
.branch_sample_type
& PERF_SAMPLE_BRANCH_CALL_STACK
;
500 static inline bool evsel__has_callchain(const struct evsel
*evsel
)
502 return (evsel
->core
.attr
.sample_type
& PERF_SAMPLE_CALLCHAIN
) != 0;
505 typedef int (*attr__fprintf_f
)(FILE *, const char *, const char *, void *);
507 int perf_event_attr__fprintf(FILE *fp
, struct perf_event_attr
*attr
,
508 attr__fprintf_f attr__fprintf
, void *priv
);
510 struct perf_env
*perf_evsel__env(struct evsel
*evsel
);
512 int perf_evsel__store_ids(struct evsel
*evsel
, struct evlist
*evlist
);
513 #endif /* __PERF_EVSEL_H */