struct perf_file_section secs[HEADER_FEAT_BITS];
struct guest_session guest_session;
struct strlist *known_build_ids;
- const struct evsel *mmap_evsel;
+ struct evsel *mmap_evsel;
struct ip_callchain *raw_callchain;
};
* processing mmap events. If not stashed, search the evlist for the first mmap
* gathering event.
*/
-static const struct evsel *inject__mmap_evsel(struct perf_inject *inject)
+static struct evsel *inject__mmap_evsel(struct perf_inject *inject)
{
struct evsel *pos;
sample__for_each_callchain_node(thread, evsel, sample, PERF_MAX_STACK_DEPTH,
/*symbols=*/false, mark_dso_hit_callback, &args);
-
thread__put(thread);
repipe:
perf_event__repipe(tool, event, sample, machine);
{
struct machine *machine = perf_session__findnew_machine(inject->session, machine_pid);
struct perf_sample synth_sample = {
+ .evsel = inject__mmap_evsel(inject),
.pid = -1,
.tid = -1,
.time = -1,
bool per_thread)
{
struct perf_sample_id *sid = evlist__id2sid(evlist, v->id);
- struct evsel *evsel;
+ struct evsel *saved_evsel = sample->evsel;
u64 *storage = NULL;
+ int ret;
if (sid) {
storage = perf_sample_id__get_period_storage(sid, sample->tid, per_thread);
if (!sample->period)
return 0;
- evsel = container_of(sid->evsel, struct evsel, core);
- return tool->sample(tool, event, sample, evsel, machine);
+ sample->evsel = container_of(sid->evsel, struct evsel, core);
+ ret = tool->sample(tool, event, sample, sample->evsel, machine);
+ sample->evsel = saved_evsel;
+ return ret;
}
static int deliver_sample_group(struct evlist *evlist,
struct machine *machine)
{
struct deferred_event *de, *tmp;
- struct evsel *evsel;
int ret = 0;
if (!tool->merge_deferred_callchains) {
- evsel = evlist__id2evsel(evlist, sample->id);
- return tool->callchain_deferred(tool, event, sample,
- evsel, machine);
+ struct evsel *saved_evsel = sample->evsel;
+
+ sample->evsel = evlist__id2evsel(evlist, sample->id);
+ ret = tool->callchain_deferred(tool, event, sample,
+ sample->evsel, machine);
+ sample->evsel = saved_evsel;
+ return ret;
}
list_for_each_entry_safe(de, tmp, &evlist->deferred_samples, list) {
else
orig_sample.deferred_callchain = false;
- evsel = evlist__id2evsel(evlist, orig_sample.id);
+ orig_sample.evsel = evlist__id2evsel(evlist, orig_sample.id);
ret = evlist__deliver_sample(evlist, tool, de->event,
- &orig_sample, evsel, machine);
+ &orig_sample, orig_sample.evsel, machine);
perf_sample__exit(&orig_sample);
list_del(&de->list);
struct evlist *evlist = session->evlist;
struct machine *machine = &session->machines.host;
struct deferred_event *de, *tmp;
- struct evsel *evsel;
int ret = 0;
list_for_each_entry_safe(de, tmp, &evlist->deferred_samples, list) {
break;
}
- evsel = evlist__id2evsel(evlist, sample.id);
+ sample.evsel = evlist__id2evsel(evlist, sample.id);
ret = evlist__deliver_sample(evlist, tool, de->event,
- &sample, evsel, machine);
+ &sample, sample.evsel, machine);
perf_sample__exit(&sample);
list_del(&de->list);
dump_event(evlist, event, file_offset, sample, file_path);
- evsel = evlist__id2evsel(evlist, sample->id);
+ if (!sample->evsel)
+ sample->evsel = evlist__id2evsel(evlist, sample->id);
+ else
+ assert(sample->evsel == evlist__id2evsel(evlist, sample->id));
+ evsel = sample->evsel;
machine = machines__find_for_cpumode(machines, event, sample);
switch (event->header.type) {