|
@@ -228,7 +228,7 @@ void perf_evlist__disable(struct perf_evlist *evlist)
|
|
|
int cpu, thread;
|
|
|
struct perf_evsel *pos;
|
|
|
|
|
|
- for (cpu = 0; cpu < evlist->cpus->nr; cpu++) {
|
|
|
+ for (cpu = 0; cpu < cpu_map__nr(evlist->cpus); cpu++) {
|
|
|
list_for_each_entry(pos, &evlist->entries, node) {
|
|
|
if (!perf_evsel__is_group_leader(pos))
|
|
|
continue;
|
|
@@ -443,7 +443,7 @@ static int perf_evlist__mmap_per_cpu(struct perf_evlist *evlist, int prot, int m
|
|
|
struct perf_evsel *evsel;
|
|
|
int cpu, thread;
|
|
|
|
|
|
- for (cpu = 0; cpu < evlist->cpus->nr; cpu++) {
|
|
|
+ for (cpu = 0; cpu < cpu_map__nr(evlist->cpus); cpu++) {
|
|
|
int output = -1;
|
|
|
|
|
|
for (thread = 0; thread < evlist->threads->nr; thread++) {
|
|
@@ -470,7 +470,7 @@ static int perf_evlist__mmap_per_cpu(struct perf_evlist *evlist, int prot, int m
|
|
|
return 0;
|
|
|
|
|
|
out_unmap:
|
|
|
- for (cpu = 0; cpu < evlist->cpus->nr; cpu++) {
|
|
|
+ for (cpu = 0; cpu < cpu_map__nr(evlist->cpus); cpu++) {
|
|
|
if (evlist->mmap[cpu].base != NULL) {
|
|
|
munmap(evlist->mmap[cpu].base, evlist->mmap_len);
|
|
|
evlist->mmap[cpu].base = NULL;
|
|
@@ -725,7 +725,7 @@ int perf_evlist__open(struct perf_evlist *evlist)
|
|
|
|
|
|
return 0;
|
|
|
out_err:
|
|
|
- ncpus = evlist->cpus ? evlist->cpus->nr : 1;
|
|
|
+ ncpus = cpu_map__nr(evlist->cpus);
|
|
|
nthreads = evlist->threads ? evlist->threads->nr : 1;
|
|
|
|
|
|
list_for_each_entry_reverse(evsel, &evlist->entries, node)
|