Lines Matching refs:sample

1337 				       struct perf_sample *sample)  in perf_evsel__parse_id_sample()  argument
1340 const u64 *array = event->sample.array; in perf_evsel__parse_id_sample()
1348 sample->id = *array; in perf_evsel__parse_id_sample()
1360 sample->cpu = u.val32[0]; in perf_evsel__parse_id_sample()
1365 sample->stream_id = *array; in perf_evsel__parse_id_sample()
1370 sample->id = *array; in perf_evsel__parse_id_sample()
1375 sample->time = *array; in perf_evsel__parse_id_sample()
1388 sample->pid = u.val32[0]; in perf_evsel__parse_id_sample()
1389 sample->tid = u.val32[1]; in perf_evsel__parse_id_sample()
1439 array = event->sample.array; in perf_evsel__parse_sample()
1678 size_t perf_event__sample_event_size(const struct perf_sample *sample, u64 type, in perf_event__sample_event_size() argument
1718 sz = sample->read.group.nr * in perf_event__sample_event_size()
1727 sz = (sample->callchain->nr + 1) * sizeof(u64); in perf_event__sample_event_size()
1733 result += sample->raw_size; in perf_event__sample_event_size()
1737 sz = sample->branch_stack->nr * sizeof(struct branch_entry); in perf_event__sample_event_size()
1743 if (sample->user_regs.abi) { in perf_event__sample_event_size()
1745 sz = hweight_long(sample->user_regs.mask) * sizeof(u64); in perf_event__sample_event_size()
1753 sz = sample->user_stack.size; in perf_event__sample_event_size()
1771 if (sample->intr_regs.abi) { in perf_event__sample_event_size()
1773 sz = hweight_long(sample->intr_regs.mask) * sizeof(u64); in perf_event__sample_event_size()
1785 const struct perf_sample *sample, in perf_event__synthesize_sample() argument
1796 array = event->sample.array; in perf_event__synthesize_sample()
1799 *array = sample->id; in perf_event__synthesize_sample()
1804 *array = sample->ip; in perf_event__synthesize_sample()
1809 u.val32[0] = sample->pid; in perf_event__synthesize_sample()
1810 u.val32[1] = sample->tid; in perf_event__synthesize_sample()
1825 *array = sample->time; in perf_event__synthesize_sample()
1830 *array = sample->addr; in perf_event__synthesize_sample()
1835 *array = sample->id; in perf_event__synthesize_sample()
1840 *array = sample->stream_id; in perf_event__synthesize_sample()
1845 u.val32[0] = sample->cpu; in perf_event__synthesize_sample()
1858 *array = sample->period; in perf_event__synthesize_sample()
1864 *array = sample->read.group.nr; in perf_event__synthesize_sample()
1866 *array = sample->read.one.value; in perf_event__synthesize_sample()
1870 *array = sample->read.time_enabled; in perf_event__synthesize_sample()
1875 *array = sample->read.time_running; in perf_event__synthesize_sample()
1881 sz = sample->read.group.nr * in perf_event__synthesize_sample()
1883 memcpy(array, sample->read.group.values, sz); in perf_event__synthesize_sample()
1886 *array = sample->read.one.id; in perf_event__synthesize_sample()
1892 sz = (sample->callchain->nr + 1) * sizeof(u64); in perf_event__synthesize_sample()
1893 memcpy(array, sample->callchain, sz); in perf_event__synthesize_sample()
1898 u.val32[0] = sample->raw_size; in perf_event__synthesize_sample()
1911 memcpy(array, sample->raw_data, sample->raw_size); in perf_event__synthesize_sample()
1912 array = (void *)array + sample->raw_size; in perf_event__synthesize_sample()
1916 sz = sample->branch_stack->nr * sizeof(struct branch_entry); in perf_event__synthesize_sample()
1918 memcpy(array, sample->branch_stack, sz); in perf_event__synthesize_sample()
1923 if (sample->user_regs.abi) { in perf_event__synthesize_sample()
1924 *array++ = sample->user_regs.abi; in perf_event__synthesize_sample()
1925 sz = hweight_long(sample->user_regs.mask) * sizeof(u64); in perf_event__synthesize_sample()
1926 memcpy(array, sample->user_regs.regs, sz); in perf_event__synthesize_sample()
1934 sz = sample->user_stack.size; in perf_event__synthesize_sample()
1937 memcpy(array, sample->user_stack.data, sz); in perf_event__synthesize_sample()
1944 *array = sample->weight; in perf_event__synthesize_sample()
1949 *array = sample->data_src; in perf_event__synthesize_sample()
1954 *array = sample->transaction; in perf_event__synthesize_sample()
1959 if (sample->intr_regs.abi) { in perf_event__synthesize_sample()
1960 *array++ = sample->intr_regs.abi; in perf_event__synthesize_sample()
1961 sz = hweight_long(sample->intr_regs.mask) * sizeof(u64); in perf_event__synthesize_sample()
1962 memcpy(array, sample->intr_regs.regs, sz); in perf_event__synthesize_sample()
1977 void *perf_evsel__rawptr(struct perf_evsel *evsel, struct perf_sample *sample, in perf_evsel__rawptr() argument
1989 offset = *(int *)(sample->raw_data + field->offset); in perf_evsel__rawptr()
1993 return sample->raw_data + offset; in perf_evsel__rawptr()
1996 u64 perf_evsel__intval(struct perf_evsel *evsel, struct perf_sample *sample, in perf_evsel__intval() argument
2006 ptr = sample->raw_data + field->offset; in perf_evsel__intval()