2023-01-01 20:31:20 +08:00
|
|
|
/* Copyright (C) 2021-2023 Free Software Foundation, Inc.
|
2022-03-11 16:58:31 +08:00
|
|
|
Contributed by Oracle.
|
|
|
|
|
|
|
|
This file is part of GNU Binutils.
|
|
|
|
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
|
|
it under the terms of the GNU General Public License as published by
|
|
|
|
the Free Software Foundation; either version 3, or (at your option)
|
|
|
|
any later version.
|
|
|
|
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
GNU General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
along with this program; if not, write to the Free Software
|
|
|
|
Foundation, 51 Franklin Street - Fifth Floor, Boston,
|
|
|
|
MA 02110-1301, USA. */
|
|
|
|
|
|
|
|
#include <errno.h>
|
|
|
|
#include <unistd.h>
|
|
|
|
#include <fcntl.h>
|
|
|
|
#include <sys/mman.h>
|
|
|
|
#include <sys/ioctl.h>
|
|
|
|
#include <sys/syscall.h>
|
|
|
|
#include <linux/perf_event.h>
|
|
|
|
|
|
|
|
#include "hwcdrv.h"
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
/* macros */
|
|
|
|
#define IS_GLOBAL /* Mark global symbols */
|
|
|
|
|
|
|
|
#include "cpuid.c" /* ftns for identifying a chip */
|
|
|
|
|
|
|
|
static hdrv_pcbe_api_t hdrv_pcbe_core_api;
|
|
|
|
static hdrv_pcbe_api_t hdrv_pcbe_opteron_api;
|
|
|
|
static hdrv_pcbe_api_t *hdrv_pcbe_drivers[] = {
|
|
|
|
&hdrv_pcbe_core_api,
|
|
|
|
&hdrv_pcbe_opteron_api,
|
|
|
|
NULL
|
|
|
|
};
|
|
|
|
#include "opteron_pcbe.c" /* CPU-specific code */
|
|
|
|
#include "core_pcbe.c" /* CPU-specific code */
|
|
|
|
|
|
|
|
extern hwcdrv_api_t hwcdrv_pcl_api;
|
|
|
|
IS_GLOBAL hwcdrv_api_t *hwcdrv_drivers[] = {
|
|
|
|
&hwcdrv_pcl_api,
|
|
|
|
NULL
|
|
|
|
};
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
/* utils for drivers */
|
|
|
|
IS_GLOBAL int
|
|
|
|
hwcdrv_assign_all_regnos (Hwcentry* entries[], unsigned numctrs)
|
|
|
|
{
|
|
|
|
unsigned int pmc_assigned[MAX_PICS];
|
|
|
|
unsigned idx;
|
|
|
|
for (int ii = 0; ii < MAX_PICS; ii++)
|
|
|
|
pmc_assigned[ii] = 0;
|
|
|
|
|
|
|
|
/* assign the HWCs that we already know about */
|
|
|
|
for (idx = 0; idx < numctrs; idx++)
|
|
|
|
{
|
|
|
|
regno_t regno = entries[idx]->reg_num;
|
|
|
|
if (regno == REGNO_ANY)
|
|
|
|
{
|
|
|
|
/* check to see if list of possible registers only contains one entry */
|
|
|
|
regno = REG_LIST_SINGLE_VALID_ENTRY (entries[idx]->reg_list);
|
|
|
|
}
|
|
|
|
if (regno != REGNO_ANY)
|
|
|
|
{
|
|
|
|
if (regno < 0 || regno >= MAX_PICS || !regno_is_valid (entries[idx], regno))
|
|
|
|
{
|
|
|
|
logerr (GTXT ("For counter #%d, register %d is out of range\n"), idx + 1, regno); /*!*/
|
|
|
|
return HWCFUNCS_ERROR_HWCARGS;
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT2, "hwcfuncs_assign_regnos(): preselected: idx=%d, regno=%d\n", idx, regno);
|
|
|
|
entries[idx]->reg_num = regno; /* assigning back to entries */
|
|
|
|
pmc_assigned[regno] = 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* assign HWCs that are currently REGNO_ANY */
|
|
|
|
for (idx = 0; idx < numctrs; idx++)
|
|
|
|
{
|
|
|
|
if (entries[idx]->reg_num == REGNO_ANY)
|
|
|
|
{
|
|
|
|
int assigned = 0;
|
|
|
|
regno_t *reg_list = entries[idx]->reg_list;
|
|
|
|
for (; reg_list && *reg_list != REGNO_ANY; reg_list++)
|
|
|
|
{
|
|
|
|
regno_t regno = *reg_list;
|
|
|
|
if (regno < 0 || regno >= MAX_PICS)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("For counter #%d, register %d is out of range\n"), idx + 1, regno); /*!*/
|
|
|
|
return HWCFUNCS_ERROR_HWCARGS;
|
|
|
|
}
|
|
|
|
if (pmc_assigned[regno] == 0)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT2, "hwcfuncs_assign_regnos(): assigned: idx=%d, regno=%d\n", idx, regno);
|
|
|
|
entries[idx]->reg_num = regno; /* assigning back to entries */
|
|
|
|
pmc_assigned[regno] = 1;
|
|
|
|
assigned = 1;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (!assigned)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("Counter '%s' could not be bound to a register\n"),
|
|
|
|
entries[idx]->name ? entries[idx]->name : "<NULL>");
|
|
|
|
return HWCFUNCS_ERROR_HWCARGS;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
IS_GLOBAL int
|
|
|
|
hwcdrv_lookup_cpuver (const char * cpcN_cciname)
|
|
|
|
{
|
|
|
|
libcpc2_cpu_lookup_t *plookup;
|
|
|
|
static libcpc2_cpu_lookup_t cpu_table[] = {
|
|
|
|
LIBCPC2_CPU_LOOKUP_LIST
|
|
|
|
};
|
|
|
|
if (cpcN_cciname == NULL)
|
|
|
|
return CPUVER_UNDEFINED;
|
|
|
|
|
|
|
|
/* search table for name */
|
|
|
|
for (plookup = cpu_table; plookup->cpc2_cciname; plookup++)
|
|
|
|
{
|
|
|
|
int n = strlen (plookup->cpc2_cciname);
|
|
|
|
if (!strncmp (plookup->cpc2_cciname, cpcN_cciname, n))
|
|
|
|
return plookup->cpc2_cpuver;
|
|
|
|
}
|
|
|
|
/* unknown, but does have a descriptive string */
|
|
|
|
TprintfT (DBG_LT0, "hwcfuncs: CPC2: WARNING: Id of processor '%s' "
|
|
|
|
"could not be determined\n",
|
|
|
|
cpcN_cciname);
|
|
|
|
return CPUVER_GENERIC;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
/* utils to generate x86 register definitions on Linux */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This code is structured as though we're going to initialize the
|
|
|
|
* HWC by writing the Intel MSR register directly. That is, we
|
|
|
|
* assume the lowest 16 bits of the event number will have the event
|
|
|
|
* and that higher bits will set attributes.
|
|
|
|
*
|
|
|
|
* While SPARC is different, we can nonetheless use basically the
|
|
|
|
* same "x86"-named functions:
|
|
|
|
*
|
|
|
|
* - The event code will still be 16 bits. It will still
|
|
|
|
* be in the lowest 16 bits of the event number. Though
|
|
|
|
* perf_event_code() on SPARC will expect those bits to
|
|
|
|
* shifted, hwcdrv_pcl.c can easily perform that shift.
|
|
|
|
*
|
|
|
|
* - On SPARC we support only two attributes, "user" and "system",
|
|
|
|
* which hwcdrv_pcl.c already converts to the "exclude_user"
|
|
|
|
* and "exclude_kernel" fields expected by perf_event_open().
|
|
|
|
* "user" and "system" are stored in event bits 16 and 17.
|
|
|
|
* For M8, a 4-bit mask of supported PICs is stored in bits [23:20].
|
|
|
|
*/
|
|
|
|
|
|
|
|
IS_GLOBAL hwcdrv_get_eventnum_fn_t *hwcdrv_get_x86_eventnum = 0;
|
|
|
|
|
|
|
|
static const attr_info_t perfctr_sparc_attrs[] = {
|
|
|
|
{NTXT ("user"), 0, 0x01, 16}, //usr
|
|
|
|
{NTXT ("system"), 0, 0x01, 17}, //os
|
|
|
|
{NULL, 0, 0x00, 0},
|
|
|
|
};
|
|
|
|
static const attr_info_t perfctr_x64_attrs[] = {/* ok for Core2 & later */
|
|
|
|
{NTXT ("umask"), 0, 0xff, 8},
|
|
|
|
{NTXT ("user"), 0, 0x01, 16}, //usr
|
|
|
|
//{NTXT("nouser"), 1, 0x01, 16}, //usr (inverted)
|
|
|
|
{NTXT ("system"), 0, 0x01, 17}, //os
|
|
|
|
{NTXT ("edge"), 0, 0x01, 18},
|
|
|
|
{NTXT ("pc"), 0, 0x01, 19},
|
|
|
|
{NTXT ("inv"), 0, 0x01, 23},
|
|
|
|
{NTXT ("cmask"), 0, 0xff, 24},
|
|
|
|
{NULL, 0, 0x00, 0},
|
|
|
|
};
|
|
|
|
const attr_info_t *perfctr_attrs_table = perfctr_x64_attrs;
|
|
|
|
|
|
|
|
static const eventsel_t perfctr_evntsel_enable_bits = (0x01 << 16) | /* usr */
|
|
|
|
// (0xff << 0) | /* event*/
|
|
|
|
// (0xff << 8) | /* umask */
|
|
|
|
// (0x01 << 17) | /* os */
|
|
|
|
// (0x01 << 18) | /* edge */
|
|
|
|
// (0x01 << 19) | /* pc */
|
|
|
|
(0x01 << 20) | /* int */
|
|
|
|
// (0x01 << 21) | /* reserved */
|
|
|
|
(0x01 << 22) | /* enable */
|
|
|
|
// (0x01 << 23) | /* inv */
|
|
|
|
// (0xff << 24) | /* cmask */
|
|
|
|
0;
|
|
|
|
|
|
|
|
static int
|
|
|
|
myperfctr_get_x86_eventnum (const char *eventname, uint_t pmc,
|
|
|
|
eventsel_t *eventsel, eventsel_t *valid_umask,
|
|
|
|
uint_t *pmc_sel)
|
|
|
|
{
|
|
|
|
if (hwcdrv_get_x86_eventnum &&
|
|
|
|
!hwcdrv_get_x86_eventnum (eventname, pmc, eventsel, valid_umask, pmc_sel))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* check for numerically-specified counters */
|
|
|
|
char * endptr;
|
|
|
|
uint64_t num = strtoull (eventname, &endptr, 0);
|
|
|
|
if (*eventname && !*endptr)
|
|
|
|
{
|
|
|
|
*eventsel = EXTENDED_EVNUM_2_EVSEL (num);
|
|
|
|
*valid_umask = 0xff; /* allow any umask (unused for SPARC?) */
|
|
|
|
*pmc_sel = pmc;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* name does not specify a numeric value */
|
|
|
|
*eventsel = (eventsel_t) - 1;
|
|
|
|
*valid_umask = 0x0;
|
|
|
|
*pmc_sel = pmc;
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
mask_shift_set (eventsel_t *presult, eventsel_t invalue,
|
|
|
|
eventsel_t mask, eventsel_t shift)
|
|
|
|
{
|
|
|
|
if (invalue & ~mask)
|
|
|
|
return -1; /* invalue attempts to set bits outside of mask */
|
|
|
|
*presult &= ~(mask << shift); /* clear all the mask bits */
|
|
|
|
*presult |= (invalue << shift); /* set bits according to invalue */
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
set_x86_attr_bits (eventsel_t *result_mask, eventsel_t evnt_valid_umask,
|
|
|
|
hwcfuncs_attr_t attrs[], int nattrs, const char*nameOnly)
|
|
|
|
{
|
|
|
|
eventsel_t evntsel = *result_mask;
|
|
|
|
for (int ii = 0; ii < (int) nattrs; ii++)
|
|
|
|
{
|
|
|
|
const char *attrname = attrs[ii].ca_name;
|
|
|
|
eventsel_t attrval = (eventsel_t) attrs[ii].ca_val;
|
|
|
|
const char *tmpname;
|
|
|
|
int attr_found = 0;
|
|
|
|
for (int jj = 0; (tmpname = perfctr_attrs_table[jj].attrname); jj++)
|
|
|
|
{
|
|
|
|
if (strcmp (attrname, tmpname) == 0)
|
|
|
|
{
|
|
|
|
if (strcmp (attrname, "umask") == 0)
|
|
|
|
{
|
|
|
|
if (attrval & ~evnt_valid_umask)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("for `%s', allowable umask bits are: 0x%llx\n"),
|
|
|
|
nameOnly, (long long) evnt_valid_umask);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (mask_shift_set (&evntsel,
|
|
|
|
perfctr_attrs_table[jj].is_inverted ? (attrval^1) : attrval,
|
|
|
|
perfctr_attrs_table[jj].mask,
|
|
|
|
perfctr_attrs_table[jj].shift))
|
|
|
|
{
|
|
|
|
logerr (GTXT ("`%s' attribute `%s' could not be set to 0x%llx\n"),
|
|
|
|
nameOnly, attrname, (long long) attrval);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT2, "hwcfuncs: Counter %s, attribute %s set to 0x%llx\n",
|
|
|
|
nameOnly, attrname, (long long) attrval);
|
|
|
|
attr_found = 1;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (!attr_found)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("attribute `%s' is invalid\n"), attrname);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
*result_mask = evntsel;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
IS_GLOBAL int
|
|
|
|
hwcfuncs_get_x86_eventsel (unsigned int regno, const char *int_name,
|
|
|
|
eventsel_t *return_event, uint_t *return_pmc_sel)
|
|
|
|
{
|
|
|
|
hwcfuncs_attr_t attrs[HWCFUNCS_MAX_ATTRS + 1];
|
|
|
|
unsigned nattrs = 0;
|
|
|
|
char *nameOnly = NULL;
|
|
|
|
eventsel_t evntsel = 0; // event number
|
|
|
|
eventsel_t evnt_valid_umask = 0;
|
|
|
|
uint_t pmc_sel = 0;
|
|
|
|
int rc = -1;
|
|
|
|
*return_event = 0;
|
|
|
|
*return_pmc_sel = 0;
|
|
|
|
void *attr_mem = hwcfuncs_parse_attrs (int_name, attrs, HWCFUNCS_MAX_ATTRS,
|
|
|
|
&nattrs, NULL);
|
|
|
|
if (!attr_mem)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("out of memory, could not parse attributes\n"));
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
hwcfuncs_parse_ctr (int_name, NULL, &nameOnly, NULL, NULL, NULL);
|
|
|
|
if (regno == REGNO_ANY)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("reg# could not be determined for `%s'\n"), nameOnly);
|
|
|
|
goto attr_wrapup;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* look up evntsel */
|
|
|
|
if (myperfctr_get_x86_eventnum (nameOnly, regno,
|
|
|
|
&evntsel, &evnt_valid_umask, &pmc_sel))
|
|
|
|
{
|
|
|
|
logerr (GTXT ("counter `%s' is not valid\n"), nameOnly);
|
|
|
|
goto attr_wrapup;
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT1, "hwcfuncs: event=0x%llx pmc=0x%x '%s' nattrs = %u\n",
|
|
|
|
(long long) evntsel, pmc_sel, nameOnly, nattrs);
|
|
|
|
|
|
|
|
/* determine event attributes */
|
|
|
|
eventsel_t evnt_attrs = perfctr_evntsel_enable_bits;
|
|
|
|
if (set_x86_attr_bits (&evnt_attrs, evnt_valid_umask, attrs, nattrs, nameOnly))
|
|
|
|
goto attr_wrapup;
|
|
|
|
if (evntsel & evnt_attrs)
|
|
|
|
TprintfT (DBG_LT0, "hwcfuncs: ERROR - evntsel & enable bits overlap: 0x%llx 0x%llx 0x%llx\n",
|
|
|
|
(long long) evntsel, (long long) evnt_attrs,
|
|
|
|
(long long) (evntsel & evnt_attrs));
|
|
|
|
*return_event = evntsel | evnt_attrs;
|
|
|
|
*return_pmc_sel = pmc_sel;
|
|
|
|
rc = 0;
|
|
|
|
|
|
|
|
attr_wrapup:
|
|
|
|
free (attr_mem);
|
|
|
|
free (nameOnly);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef __x86_64__
|
|
|
|
#define syscall_instr "syscall"
|
|
|
|
#define syscall_clobber "rcx", "r11", "memory"
|
|
|
|
#endif
|
|
|
|
#ifdef __i386__
|
|
|
|
#define syscall_instr "int $0x80"
|
|
|
|
#define syscall_clobber "memory"
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
perf_event_open (struct perf_event_attr *hw_event_uptr, pid_t pid,
|
|
|
|
int cpu, int group_fd, unsigned long flags)
|
|
|
|
{
|
|
|
|
/* It seems that perf_event_open() sometimes fails spuriously,
|
|
|
|
* even while an immediate retry succeeds.
|
|
|
|
* So, let's try a few retries if the call fails just to be sure.
|
|
|
|
*/
|
|
|
|
int rc;
|
|
|
|
for (int retry = 0; retry < 5; retry++)
|
|
|
|
{
|
|
|
|
rc = syscall (__NR_perf_event_open, hw_event_uptr, pid, cpu, group_fd, flags);
|
|
|
|
if (rc != -1)
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
/* macros & fwd prototypes */
|
|
|
|
|
|
|
|
#define HWCDRV_API static /* Mark functions used by hwcdrv API */
|
|
|
|
|
|
|
|
HWCDRV_API int hwcdrv_start (void);
|
|
|
|
HWCDRV_API int hwcdrv_free_counters ();
|
|
|
|
|
|
|
|
static pid_t
|
|
|
|
hwcdrv_gettid (void)
|
|
|
|
{
|
|
|
|
#ifndef LIBCOLLECTOR_SRC
|
|
|
|
return syscall (__NR_gettid);
|
|
|
|
#elif defined(intel)
|
|
|
|
pid_t r;
|
|
|
|
__asm__ __volatile__(syscall_instr
|
|
|
|
: "=a" (r) : "0" (__NR_gettid)
|
|
|
|
: syscall_clobber);
|
|
|
|
return r;
|
|
|
|
#else
|
|
|
|
return syscall (__NR_gettid); // FIXUP_XXX_SPARC_LINUX // write gettid in asm
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
/* types */
|
|
|
|
|
|
|
|
#define NPAGES_PER_BUF 1 // number of pages to be used for perf_event samples
|
|
|
|
// must be a power of 2
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
/* typedefs */
|
|
|
|
|
|
|
|
typedef struct
|
|
|
|
{ // event (hwc) definition
|
|
|
|
unsigned int reg_num; // PMC assignment, potentially for detecting conflicts
|
|
|
|
eventsel_t eventsel; // raw event bits (Intel/AMD)
|
|
|
|
uint64_t counter_preload; // number of HWC events before signal
|
|
|
|
struct perf_event_attr hw; // perf_event definition
|
|
|
|
hrtime_t min_time; // minimum time we're targeting between events
|
|
|
|
char *name;
|
|
|
|
} perf_event_def_t;
|
|
|
|
|
|
|
|
typedef struct
|
|
|
|
{ // runtime state of perf_event buffer
|
|
|
|
void *buf; // pointer to mmapped buffer
|
|
|
|
size_t pagesz; // size of pages
|
|
|
|
} buffer_state_t;
|
|
|
|
|
|
|
|
typedef struct
|
|
|
|
{ // runtime state of counter values
|
|
|
|
uint64_t prev_ena_ts; // previous perf_event "enabled" time
|
|
|
|
uint64_t prev_run_ts; // previous perf_event "running" time
|
|
|
|
uint64_t prev_value; // previous HWC value
|
|
|
|
} counter_value_state_t;
|
|
|
|
|
|
|
|
typedef struct
|
|
|
|
{ // per-counter information
|
|
|
|
perf_event_def_t *ev_def; // global HWC definition for one counter
|
|
|
|
int fd; // perf_event fd
|
|
|
|
buffer_state_t buf_state; // perf_event buffer's state
|
|
|
|
counter_value_state_t value_state; // counter state
|
|
|
|
int needs_restart; // workaround for dbx failure to preserve si_fd
|
|
|
|
uint64_t last_overflow_period;
|
|
|
|
hrtime_t last_overflow_time;
|
|
|
|
} counter_state_t;
|
|
|
|
|
|
|
|
typedef struct
|
|
|
|
{ // per-thread context
|
|
|
|
counter_state_t *ctr_list;
|
|
|
|
int signal_fd; // fd that caused the most recent signal
|
2022-09-14 16:11:45 +08:00
|
|
|
pid_t tid; // for debugging signal delivery problems
|
2022-03-11 16:58:31 +08:00
|
|
|
} hdrv_pcl_ctx_t;
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
/* static variables */
|
|
|
|
static struct
|
|
|
|
{
|
|
|
|
int library_ok;
|
|
|
|
int internal_open_called;
|
|
|
|
hwcfuncs_tsd_get_fn_t find_vpc_ctx;
|
|
|
|
unsigned hwcdef_cnt; /* number of *active* hardware counters */
|
|
|
|
hwcdrv_get_events_fn_t *get_events;
|
|
|
|
} hdrv_pcl_state;
|
|
|
|
|
|
|
|
static hwcdrv_about_t hdrv_pcl_about = {.cpcN_cpuver = CPUVER_UNDEFINED};
|
|
|
|
static perf_event_def_t global_perf_event_def[MAX_PICS];
|
|
|
|
|
|
|
|
#define COUNTERS_ENABLED() (hdrv_pcl_state.hwcdef_cnt)
|
|
|
|
|
|
|
|
|
|
|
|
/* perf_event buffer formatting and handling */
|
|
|
|
static void
|
|
|
|
reset_buf (buffer_state_t *bufstate)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: perf_event reset_buf() called!\n");
|
|
|
|
struct perf_event_mmap_page *metadata = bufstate->buf;
|
|
|
|
if (metadata)
|
|
|
|
metadata->data_tail = metadata->data_head;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
skip_buf (buffer_state_t *bufstate, size_t sz)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: WARNING: perf_event skip_buf called!\n");
|
|
|
|
struct perf_event_mmap_page *metadata = bufstate->buf;
|
|
|
|
if (metadata == NULL)
|
|
|
|
return -1;
|
|
|
|
size_t pgsz = bufstate->pagesz;
|
|
|
|
size_t bufsz = NPAGES_PER_BUF*pgsz;
|
|
|
|
uint64_t d_tail = metadata->data_tail;
|
|
|
|
uint64_t d_head = metadata->data_head;
|
|
|
|
|
|
|
|
// validate request size
|
|
|
|
if (sz > d_head - d_tail || sz >= bufsz)
|
|
|
|
{
|
|
|
|
reset_buf (bufstate);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
metadata->data_tail = d_tail + sz; // advance tail
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
read_buf (buffer_state_t *bufstate, void *buf, size_t sz)
|
|
|
|
{
|
|
|
|
struct perf_event_mmap_page *metadata = bufstate->buf;
|
|
|
|
if (metadata == NULL)
|
|
|
|
return -1;
|
|
|
|
size_t pgsz = bufstate->pagesz;
|
|
|
|
size_t bufsz = NPAGES_PER_BUF*pgsz;
|
|
|
|
uint64_t d_tail = metadata->data_tail;
|
|
|
|
uint64_t d_head = metadata->data_head;
|
|
|
|
|
|
|
|
// validate request size
|
|
|
|
if (sz > d_head - d_tail || sz >= bufsz)
|
|
|
|
{
|
|
|
|
reset_buf (bufstate);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
char *buf_base = ((char *) metadata) + pgsz; // start of data buffer
|
|
|
|
uint64_t start_pos = d_tail & (bufsz - 1); // char offset into data buffer
|
|
|
|
size_t nbytes = sz;
|
|
|
|
if (start_pos + sz > bufsz)
|
|
|
|
{
|
|
|
|
// will wrap past end of buffer
|
|
|
|
nbytes = bufsz - start_pos;
|
|
|
|
memcpy (buf, buf_base + start_pos, nbytes);
|
|
|
|
start_pos = 0; // wrap to start
|
|
|
|
buf = (void *) (((char *) buf) + nbytes);
|
|
|
|
nbytes = sz - nbytes;
|
|
|
|
}
|
|
|
|
memcpy (buf, buf_base + start_pos, nbytes);
|
|
|
|
metadata->data_tail += sz;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
read_u64 (buffer_state_t *bufstate, uint64_t *value)
|
|
|
|
{
|
|
|
|
return read_buf (bufstate, value, sizeof (uint64_t));
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
read_sample (counter_state_t *ctr_state, int msgsz, uint64_t *rvalue,
|
|
|
|
uint64_t *rlost)
|
|
|
|
{
|
|
|
|
// returns count of bytes read
|
|
|
|
buffer_state_t *bufstate = &ctr_state->buf_state;
|
|
|
|
counter_value_state_t *cntstate = &ctr_state->value_state;
|
|
|
|
int readsz = 0;
|
|
|
|
|
|
|
|
// PERF_SAMPLE_IP
|
|
|
|
uint64_t ipc = 0;
|
|
|
|
int rc = read_u64 (bufstate, &ipc);
|
|
|
|
if (rc)
|
|
|
|
return -1;
|
|
|
|
readsz += sizeof (uint64_t);
|
|
|
|
|
|
|
|
// PERF_SAMPLE_READ: value
|
|
|
|
uint64_t value = 0;
|
|
|
|
rc = read_u64 (bufstate, &value);
|
|
|
|
if (rc)
|
|
|
|
return -2;
|
|
|
|
readsz += sizeof (uint64_t);
|
|
|
|
|
|
|
|
/* Bug 20806896
|
|
|
|
* Old Linux kernels (e.g. 2.6.32) on certain systems return enabled and
|
|
|
|
* running times in the sample data that correspond to the metadata times
|
|
|
|
* metadata->time_enabled
|
|
|
|
* metadata->time_running
|
|
|
|
* from the PREVIOUS (not current) sample. Probably just ignore this bug
|
|
|
|
* since it's on old kernels and we only use the enabled and running times
|
|
|
|
* to construct loss_estimate.
|
|
|
|
*/
|
|
|
|
// PERF_SAMPLE_READ: PERF_FORMAT_ENABLED
|
|
|
|
uint64_t enabled_time = 0;
|
|
|
|
rc = read_u64 (bufstate, &enabled_time);
|
|
|
|
if (rc)
|
|
|
|
return -3;
|
|
|
|
readsz += sizeof (uint64_t);
|
|
|
|
|
|
|
|
// PERF_SAMPLE_READ: PERF_FORMAT_RUNNING
|
|
|
|
uint64_t running_time = 0;
|
|
|
|
rc = read_u64 (bufstate, &running_time);
|
|
|
|
if (rc)
|
|
|
|
return -4;
|
|
|
|
readsz += sizeof (uint64_t);
|
|
|
|
|
|
|
|
uint64_t value_delta = value - cntstate->prev_value;
|
|
|
|
uint64_t enabled_delta = enabled_time - cntstate->prev_ena_ts;
|
|
|
|
uint64_t running_delta = running_time - cntstate->prev_run_ts;
|
|
|
|
cntstate->prev_value = value;
|
|
|
|
cntstate->prev_ena_ts = enabled_time;
|
|
|
|
cntstate->prev_run_ts = running_time;
|
|
|
|
|
|
|
|
// 24830461 need workaround for Linux anomalous HWC skid overrun
|
|
|
|
int set_error_flag = 0;
|
|
|
|
if (value_delta > 2 * ctr_state->last_overflow_period + 2000 /* HWC_SKID_TOLERANCE */)
|
|
|
|
set_error_flag = 1;
|
|
|
|
|
|
|
|
uint64_t loss_estimate = 0; // estimate loss of events caused by multiplexing
|
|
|
|
if (running_delta == enabled_delta)
|
|
|
|
{
|
|
|
|
// counter was running 100% of time, no multiplexing
|
|
|
|
}
|
|
|
|
else if (running_delta == 0)
|
|
|
|
loss_estimate = 1; // token amount to aid in debugging perfctr oddities
|
|
|
|
else if ((running_delta > enabled_delta) || (enabled_delta & 0x1000000000000000ll))
|
|
|
|
{
|
|
|
|
// running should be smaller than enabled, can't estimate
|
|
|
|
/*
|
|
|
|
* 21418391 HWC can have a negative count
|
|
|
|
*
|
|
|
|
* We've also seen enabled not only be smaller than running
|
|
|
|
* but in fact go negative. Guard against this.
|
|
|
|
*/
|
|
|
|
loss_estimate = 2; // token amount to aid in debugging perfctr oddities
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// counter was running less than 100% of time
|
|
|
|
// Example: ena=7772268 run=6775669 raw_value=316004 scaled_value=362483 loss_est=46479
|
|
|
|
uint64_t scaled_delta = (double) value_delta * enabled_delta / running_delta;
|
|
|
|
value_delta = scaled_delta;
|
|
|
|
#if 0
|
|
|
|
// We should perhaps warn the user that multiplexing is going on,
|
|
|
|
// but hwcdrv_pcl.c doesn't know about the collector_interface, SP_JCMD_COMMENT, or COL_COMMENT_* values.
|
|
|
|
// For now we simply don't report.
|
|
|
|
// Perhaps we should address the issue not here but in the caller collector_sigemt_handler(),
|
|
|
|
// but at that level "lost" has a meaning that's considerably broader than just multiplexing.
|
|
|
|
collector_interface->writeLog ("<event kind=\"%s\" id=\"%d\">%s %d -> %d</event>\n",
|
|
|
|
SP_JCMD_COMMENT, COL_COMMENT_HWCADJ, global_perf_event_def[idx].name,
|
|
|
|
ctr_list[idx].last_overflow_period, new_period);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
TprintfT ((loss_estimate || set_error_flag) ? DBG_LT1 : DBG_LT3,
|
|
|
|
"hwcdrv: '%s' ipc=0x%llx ena=%llu run=%llu "
|
|
|
|
"value_delta=%lld(0x%llx) loss_est=%llu %s error_flag='%s'\n",
|
|
|
|
ctr_state->ev_def->name, (long long) ipc,
|
|
|
|
(long long) enabled_delta, (long long) running_delta,
|
|
|
|
(long long) value_delta, (long long) value_delta,
|
|
|
|
(unsigned long long) loss_estimate,
|
|
|
|
loss_estimate ? ", WARNING - SCALED" : "",
|
|
|
|
set_error_flag ? ", ERRORFLAG" : "");
|
|
|
|
if (set_error_flag == 1)
|
|
|
|
value_delta |= (1ULL << 63) /* HWCVAL_ERR_FLAG */;
|
|
|
|
*rvalue = value_delta;
|
|
|
|
*rlost = loss_estimate;
|
|
|
|
if (readsz != msgsz)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: perf_event sample not fully parsed\n");
|
|
|
|
return -5;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
dump_perf_event_attr (struct perf_event_attr *at)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT2, "dump_perf_event_attr: size=%d type=%d sample_period=%lld\n"
|
|
|
|
" config=0x%llx config1=0x%llx config2=0x%llx wakeup_events=%lld __reserved_1=%lld\n",
|
|
|
|
(int) at->size, (int) at->type, (unsigned long long) at->sample_period,
|
|
|
|
(unsigned long long) at->config, (unsigned long long) at->config1,
|
|
|
|
(unsigned long long) at->config2, (unsigned long long) at->wakeup_events,
|
|
|
|
(unsigned long long) at->__reserved_1);
|
|
|
|
#define DUMP_F(fld) if (at->fld) TprintfT(DBG_LT2, " %-10s : %lld\n", #fld, (long long) at->fld)
|
|
|
|
DUMP_F (disabled);
|
|
|
|
DUMP_F (inherit);
|
|
|
|
DUMP_F (pinned);
|
|
|
|
DUMP_F (exclusive);
|
|
|
|
DUMP_F (exclude_user);
|
|
|
|
DUMP_F (exclude_kernel);
|
|
|
|
DUMP_F (exclude_hv);
|
|
|
|
DUMP_F (exclude_idle);
|
|
|
|
// DUMP_F(xmmap);
|
|
|
|
DUMP_F (comm);
|
|
|
|
DUMP_F (freq);
|
|
|
|
DUMP_F (inherit_stat);
|
|
|
|
DUMP_F (enable_on_exec);
|
|
|
|
DUMP_F (task);
|
|
|
|
DUMP_F (watermark);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
init_perf_event (struct perf_event_attr *hw, uint64_t event, uint64_t period)
|
|
|
|
{
|
|
|
|
memset (hw, 0, sizeof (struct perf_event_attr));
|
|
|
|
hw->size = sizeof (struct perf_event_attr); // fwd/bwd compat
|
|
|
|
|
|
|
|
#if defined(__i386__) || defined(__x86_64)
|
|
|
|
//note: Nehalem/Westmere OFFCORE_RESPONSE in upper 32 bits
|
|
|
|
hw->config = event;
|
|
|
|
hw->type = PERF_TYPE_RAW; // hw/sw/trace/raw...
|
|
|
|
#elif defined(__aarch64__)
|
|
|
|
hw->type = (event >> 24) & 7;
|
|
|
|
hw->config = event & 0xff;
|
|
|
|
#elif defined(sparc)
|
|
|
|
//SPARC needs to be shifted up 16 bits
|
|
|
|
hw->config = (event & 0xFFFF) << 16; // uint64_t event
|
|
|
|
uint64_t regs = (event >> 20) & 0xf; // see sparc_pcbe.c
|
|
|
|
hw->config |= regs << 4; // for M8, supported PICs need to be placed at bits [7:4]
|
|
|
|
hw->type = PERF_TYPE_RAW; // hw/sw/trace/raw...
|
|
|
|
#endif
|
|
|
|
|
|
|
|
hw->sample_period = period;
|
2022-09-30 04:30:42 +08:00
|
|
|
hw->sample_type = PERF_SAMPLE_IP |
|
2022-03-11 16:58:31 +08:00
|
|
|
// PERF_SAMPLE_TID |
|
|
|
|
// PERF_SAMPLE_TIME | // possibly interesting
|
|
|
|
// PERF_SAMPLE_ADDR |
|
|
|
|
PERF_SAMPLE_READ | // HWC value
|
|
|
|
// PERF_SAMPLE_CALLCHAIN | // interesting
|
|
|
|
// PERF_SAMPLE_ID |
|
|
|
|
// PERF_SAMPLE_CPU | // possibly interesting
|
|
|
|
// PERF_SAMPLE_PERIOD |
|
|
|
|
// PERF_SAMPLE_STREAM_ID |
|
|
|
|
// PERF_SAMPLE_RAW |
|
|
|
|
0;
|
|
|
|
hw->read_format =
|
|
|
|
PERF_FORMAT_TOTAL_TIME_ENABLED | // detect when hwc not scheduled
|
|
|
|
PERF_FORMAT_TOTAL_TIME_RUNNING | // detect when hwc not scheduled
|
|
|
|
// PERF_FORMAT_ID |
|
|
|
|
// PERF_FORMAT_GROUP |
|
|
|
|
0;
|
|
|
|
hw->disabled = 1; /* off by default */
|
|
|
|
|
|
|
|
// Note: the following override config.priv bits!
|
|
|
|
hw->exclude_user = (event & (1 << 16)) == 0; /* don't count user */
|
|
|
|
hw->exclude_kernel = (event & (1 << 17)) == 0; /* ditto kernel */
|
|
|
|
hw->exclude_hv = 1; /* ditto hypervisor */
|
|
|
|
hw->wakeup_events = 1; /* wakeup every n events */
|
|
|
|
dump_perf_event_attr (hw);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
start_one_ctr (int ii, size_t pgsz, hdrv_pcl_ctx_t * pctx, char *error_string)
|
|
|
|
{
|
|
|
|
// pe_attr should have been initialized in hwcdrv_create_counters()
|
|
|
|
struct perf_event_attr pe_attr;
|
|
|
|
memcpy (&pe_attr, &global_perf_event_def[ii].hw, sizeof (pe_attr));
|
|
|
|
|
|
|
|
// but we adjust the period, so make sure that pctx->ctr_list[ii].last_overflow_period has been set
|
|
|
|
pe_attr.sample_period = pctx->ctr_list[ii].last_overflow_period;
|
|
|
|
|
|
|
|
int hwc_fd = perf_event_open (&pe_attr, pctx->tid, -1, -1, 0);
|
|
|
|
if (hwc_fd == -1)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT1, "%s idx=%d perf_event_open failed, errno=%d\n",
|
|
|
|
error_string, ii, errno);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
size_t buffer_area_sz = (NPAGES_PER_BUF + 1) * pgsz; // add a page for metadata
|
|
|
|
void * buf = mmap (NULL, buffer_area_sz, //YXXX is this a safe call?
|
|
|
|
PROT_READ | PROT_WRITE, MAP_SHARED, hwc_fd, 0);
|
|
|
|
if (buf == MAP_FAILED)
|
|
|
|
{
|
|
|
|
TprintfT (0, "sz = %ld, pgsz = %ld\n err=%s idx=%d mmap failed: %s\n",
|
|
|
|
(long) buffer_area_sz, (long) pgsz, error_string, ii, strerror (errno));
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
pctx->ctr_list[ii].ev_def = &global_perf_event_def[ii]; // why do we set ev_def? we never seem to use it
|
|
|
|
pctx->ctr_list[ii].fd = hwc_fd;
|
|
|
|
pctx->ctr_list[ii].buf_state.buf = buf;
|
|
|
|
pctx->ctr_list[ii].buf_state.pagesz = pgsz;
|
|
|
|
pctx->ctr_list[ii].value_state.prev_ena_ts = 0;
|
|
|
|
pctx->ctr_list[ii].value_state.prev_run_ts = 0;
|
|
|
|
pctx->ctr_list[ii].value_state.prev_value = 0;
|
|
|
|
pctx->ctr_list[ii].last_overflow_time = gethrtime ();
|
|
|
|
|
|
|
|
/* set async mode */
|
|
|
|
long flags = fcntl (hwc_fd, F_GETFL, 0) | O_ASYNC;
|
|
|
|
int rc = fcntl (hwc_fd, F_SETFL, flags);
|
|
|
|
if (rc == -1)
|
|
|
|
{
|
|
|
|
TprintfT (0, "%s idx=%d O_ASYNC failed\n", error_string, ii);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* set lwp ownership of the fd
|
|
|
|
* See BUGS section of "man perf_event_open":
|
|
|
|
* The F_SETOWN_EX option to fcntl(2) is needed to properly get
|
|
|
|
* overflow signals in threads. This was introduced in Linux 2.6.32.
|
|
|
|
* Legacy references:
|
|
|
|
* see http://lkml.org/lkml/2009/8/4/128
|
|
|
|
* google man fcntl F_SETOWN_EX -conflict
|
|
|
|
* "From Linux 2.6.32 onward, use F_SETOWN_EX to target
|
|
|
|
* SIGIO and SIGURG signals at a particular thread."
|
|
|
|
* http://icl.cs.utk.edu/papi/docs/da/d2a/examples__v2_8x_2self__smpl__multi_8c.html
|
|
|
|
* See 2010 CSCADS presentation by Eranian
|
|
|
|
*/
|
|
|
|
struct f_owner_ex fowner_ex;
|
|
|
|
fowner_ex.type = F_OWNER_TID;
|
|
|
|
fowner_ex.pid = pctx->tid;
|
|
|
|
rc = fcntl (hwc_fd, F_SETOWN_EX, (unsigned long) &fowner_ex);
|
|
|
|
if (rc == -1)
|
|
|
|
{
|
|
|
|
TprintfT (0, "%s idx=%d F_SETOWN failed\n", error_string, ii);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Use sigio so handler can determine FD via siginfo->si_fd. */
|
|
|
|
rc = fcntl (hwc_fd, F_SETSIG, SIGIO);
|
|
|
|
if (rc == -1)
|
|
|
|
{
|
|
|
|
TprintfT (0, "%s idx=%d F_SETSIG failed\n", error_string, ii);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
stop_one_ctr (int ii, counter_state_t *ctr_list)
|
|
|
|
{
|
|
|
|
int hwc_rc = 0;
|
|
|
|
if (-1 == ioctl (ctr_list[ii].fd, PERF_EVENT_IOC_DISABLE, 1))
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: PERF_EVENT_IOC_DISABLE #%d failed: errno=%d\n", ii, errno);
|
|
|
|
hwc_rc = HWCFUNCS_ERROR_GENERIC;
|
|
|
|
}
|
|
|
|
void *buf = ctr_list[ii].buf_state.buf;
|
|
|
|
if (buf)
|
|
|
|
{
|
|
|
|
size_t bufsz = (NPAGES_PER_BUF + 1) * ctr_list[ii].buf_state.pagesz;
|
|
|
|
ctr_list[ii].buf_state.buf = NULL;
|
|
|
|
int tmprc = munmap (buf, bufsz);
|
|
|
|
if (tmprc)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: munmap() #%d failed: errno=%d\n", ii, errno);
|
|
|
|
hwc_rc = HWCFUNCS_ERROR_GENERIC;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (-1 == close (ctr_list[ii].fd))
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: close(fd) #%d failed: errno=%d\n", ii, errno);
|
|
|
|
hwc_rc = HWCFUNCS_ERROR_GENERIC;
|
|
|
|
}
|
|
|
|
return hwc_rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* HWCDRV_API for thread-specific actions */
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_lwp_init (void)
|
|
|
|
{
|
|
|
|
return hwcdrv_start ();
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API void
|
|
|
|
hwcdrv_lwp_fini (void)
|
|
|
|
{
|
|
|
|
hwcdrv_free_counters (); /* also sets pctx->ctr_list=NULL; */
|
|
|
|
}
|
|
|
|
|
|
|
|
/* open */
|
|
|
|
static int
|
|
|
|
hdrv_pcl_internal_open ()
|
|
|
|
{
|
|
|
|
if (hdrv_pcl_state.internal_open_called)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: WARNING: hdrv_pcl_internal_open: already called\n");
|
|
|
|
return HWCFUNCS_ERROR_ALREADY_CALLED;
|
|
|
|
}
|
|
|
|
|
|
|
|
// determine if PCL is available
|
|
|
|
perf_event_def_t tmp_event_def;
|
|
|
|
memset (&tmp_event_def, 0, sizeof (tmp_event_def));
|
|
|
|
struct perf_event_attr *pe_attr = &tmp_event_def.hw;
|
|
|
|
init_perf_event (pe_attr, 0, 0);
|
|
|
|
pe_attr->type = PERF_TYPE_HARDWARE; // specify abstracted HW event
|
|
|
|
pe_attr->config = PERF_COUNT_HW_INSTRUCTIONS; // specify abstracted insts
|
|
|
|
int hwc_fd = perf_event_open (pe_attr,
|
|
|
|
0, // pid/tid, 0 is self
|
|
|
|
-1, // cpu, -1 is per-thread mode
|
|
|
|
-1, // group_fd, -1 is root
|
|
|
|
0); // flags
|
|
|
|
if (hwc_fd == -1)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: WARNING: hdrv_pcl_internal_open:"
|
|
|
|
" perf_event_open() failed, errno=%d\n", errno);
|
|
|
|
goto internal_open_error;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* see if the PCL is new enough to know about F_SETOWN_EX */
|
|
|
|
struct f_owner_ex fowner_ex;
|
|
|
|
fowner_ex.type = F_OWNER_TID;
|
|
|
|
fowner_ex.pid = hwcdrv_gettid (); // "pid=tid" is correct w/F_OWNER_TID
|
|
|
|
if (fcntl (hwc_fd, F_SETOWN_EX, (unsigned long) &fowner_ex) == -1)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: WARNING: hdrv_pcl_internal_open: "
|
|
|
|
"F_SETOWN failed, errno=%d\n", errno);
|
|
|
|
close (hwc_fd);
|
|
|
|
goto internal_open_error;
|
|
|
|
}
|
|
|
|
close (hwc_fd);
|
|
|
|
|
|
|
|
hdrv_pcl_state.internal_open_called = 1;
|
|
|
|
hdrv_pcl_state.library_ok = 1; // set to non-zero to show it's initted
|
|
|
|
hdrv_pcl_about.cpcN_cpuver = CPUVER_UNDEFINED;
|
|
|
|
TprintfT (DBG_LT2, "hwcdrv: hdrv_pcl_internal_open()\n");
|
|
|
|
for (int ii = 0; hdrv_pcbe_drivers[ii]; ii++)
|
|
|
|
{
|
|
|
|
hdrv_pcbe_api_t *ppcbe = hdrv_pcbe_drivers[ii];
|
|
|
|
if (!ppcbe->hdrv_pcbe_init ())
|
|
|
|
{
|
|
|
|
hdrv_pcl_about.cpcN_cciname = ppcbe->hdrv_pcbe_impl_name ();
|
|
|
|
hdrv_pcl_about.cpcN_cpuver = hwcdrv_lookup_cpuver (hdrv_pcl_about.cpcN_cciname);
|
|
|
|
if (hdrv_pcl_about.cpcN_cpuver == CPUVER_UNDEFINED)
|
|
|
|
goto internal_open_error;
|
|
|
|
hdrv_pcl_about.cpcN_npics = ppcbe->hdrv_pcbe_ncounters ();
|
|
|
|
hdrv_pcl_about.cpcN_docref = ppcbe->hdrv_pcbe_cpuref ();
|
|
|
|
hdrv_pcl_state.get_events = ppcbe->hdrv_pcbe_get_events;
|
|
|
|
hwcdrv_get_x86_eventnum = ppcbe->hdrv_pcbe_get_eventnum;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (hdrv_pcl_about.cpcN_npics > MAX_PICS)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: WARNING: hdrv_pcl_internal_open:"
|
|
|
|
" reducing number of HWCs from %u to %u on processor '%s'\n",
|
|
|
|
hdrv_pcl_about.cpcN_npics, MAX_PICS, hdrv_pcl_about.cpcN_cciname);
|
|
|
|
hdrv_pcl_about.cpcN_npics = MAX_PICS;
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: hdrv_pcl_internal_open:"
|
|
|
|
" perf_event cpuver=%d, name='%s'\n",
|
|
|
|
hdrv_pcl_about.cpcN_cpuver, hdrv_pcl_about.cpcN_cciname);
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
internal_open_error:
|
|
|
|
hdrv_pcl_about.cpcN_cpuver = CPUVER_UNDEFINED;
|
|
|
|
hdrv_pcl_about.cpcN_npics = 0;
|
|
|
|
hdrv_pcl_about.cpcN_docref = NULL;
|
|
|
|
hdrv_pcl_about.cpcN_cciname = NULL;
|
|
|
|
return HWCFUNCS_ERROR_NOT_SUPPORTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void *
|
|
|
|
single_thread_tsd_ftn ()
|
|
|
|
{
|
|
|
|
static hdrv_pcl_ctx_t tsd_context;
|
|
|
|
return &tsd_context;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* HWCDRV_API */
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_init (hwcfuncs_abort_fn_t abort_ftn, int *tsd_sz)
|
|
|
|
{
|
|
|
|
hdrv_pcl_state.find_vpc_ctx = single_thread_tsd_ftn;
|
|
|
|
if (tsd_sz)
|
|
|
|
*tsd_sz = sizeof (hdrv_pcl_ctx_t);
|
|
|
|
|
|
|
|
if (hdrv_pcl_state.internal_open_called)
|
|
|
|
return HWCFUNCS_ERROR_ALREADY_CALLED;
|
|
|
|
return hdrv_pcl_internal_open ();
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API void
|
|
|
|
hwcdrv_get_info (int *cpuver, const char **cciname, uint_t *npics,
|
|
|
|
const char **docref, uint64_t *support)
|
|
|
|
{
|
|
|
|
if (cpuver)
|
|
|
|
*cpuver = hdrv_pcl_about.cpcN_cpuver;
|
|
|
|
if (cciname)
|
|
|
|
*cciname = hdrv_pcl_about.cpcN_cciname;
|
|
|
|
if (npics)
|
|
|
|
*npics = hdrv_pcl_about.cpcN_npics;
|
|
|
|
if (docref)
|
|
|
|
*docref = hdrv_pcl_about.cpcN_docref;
|
|
|
|
if (support)
|
|
|
|
*support = HWCFUNCS_SUPPORT_OVERFLOW_PROFILING | HWCFUNCS_SUPPORT_OVERFLOW_CTR_ID;
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_enable_mt (hwcfuncs_tsd_get_fn_t tsd_ftn)
|
|
|
|
{
|
|
|
|
if (tsd_ftn)
|
|
|
|
hdrv_pcl_state.find_vpc_ctx = tsd_ftn;
|
|
|
|
else
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: enable_mt(): tsd_ftn==NULL\n");
|
|
|
|
return HWCFUNCS_ERROR_UNAVAIL;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_get_descriptions (hwcf_hwc_cb_t *hwc_cb, hwcf_attr_cb_t *attr_cb)
|
|
|
|
{
|
|
|
|
int count = 0;
|
|
|
|
if (hwc_cb && hdrv_pcl_state.get_events)
|
|
|
|
count = hdrv_pcl_state.get_events (hwc_cb);
|
|
|
|
if (attr_cb)
|
|
|
|
for (int ii = 0; perfctr_attrs_table && perfctr_attrs_table[ii].attrname; ii++)
|
|
|
|
attr_cb (perfctr_attrs_table[ii].attrname);
|
|
|
|
if (!count)
|
|
|
|
return -1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_assign_regnos (Hwcentry* entries[], unsigned numctrs)
|
|
|
|
{
|
|
|
|
return hwcdrv_assign_all_regnos (entries, numctrs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
internal_hwc_start (int fd)
|
|
|
|
{
|
|
|
|
int rc = ioctl (fd, PERF_EVENT_IOC_REFRESH, 1);
|
|
|
|
if (rc == -1)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: ERROR: internal_hwc_start:"
|
|
|
|
" PERF_EVENT_IOC_REFRESH(fd=%d) failed: errno=%d\n", fd, errno);
|
|
|
|
return HWCFUNCS_ERROR_UNAVAIL;
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT3, "hwcdrv: internal_hwc_start(fd=%d)\n", fd);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_overflow (siginfo_t *si, hwc_event_t *eventp, hwc_event_t *lost_events)
|
|
|
|
{
|
|
|
|
/* set expired counters to overflow value and all others to 0 */
|
|
|
|
/* return 0: OK, counters should be restarted */
|
|
|
|
/* return non-zero: eventp not set, counters should not be restarted */
|
|
|
|
/* clear return values */
|
|
|
|
int ii;
|
|
|
|
for (ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
{
|
|
|
|
eventp->ce_pic[ii] = 0;
|
|
|
|
lost_events->ce_pic[ii] = 0;
|
|
|
|
}
|
|
|
|
hrtime_t sig_ts = gethrtime (); //YXXX get this from HWC event?
|
|
|
|
eventp->ce_hrt = sig_ts;
|
|
|
|
lost_events->ce_hrt = sig_ts;
|
|
|
|
|
|
|
|
/* determine source signal */
|
|
|
|
int signal_fd = -1;
|
|
|
|
switch (si->si_code)
|
|
|
|
{
|
|
|
|
case POLL_HUP: /* expected value from pcl */
|
|
|
|
/* According to Stephane Eranian:
|
|
|
|
* "expect POLL_HUP instead of POLL_IN because we are
|
|
|
|
* in one-shot mode (IOC_REFRESH)"
|
|
|
|
*/
|
|
|
|
signal_fd = si->si_fd;
|
|
|
|
break;
|
|
|
|
case SI_TKILL: /* event forwarded by tkill */
|
|
|
|
/* DBX can only forward SI_TKILL when it detects POLL_HUP
|
|
|
|
* unfortunately, this means that si->si_fd has been lost...
|
|
|
|
* We need to process the buffers, but we don't know the fd!
|
|
|
|
*/
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: sig_ts=%llu: WARNING: hwcdrv_overflow:"
|
|
|
|
" SI_TKILL detected\n", sig_ts);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
// "sometimes we see a POLL_IN (1) with very high event rates,"
|
|
|
|
// according to eranian(?)
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: sig_ts=%llu: ERROR: hwcdrv_overflow:"
|
|
|
|
" unexpected si_code 0x%x\n", sig_ts, si->si_code);
|
|
|
|
return HWCFUNCS_ERROR_GENERIC;
|
|
|
|
}
|
|
|
|
|
|
|
|
hdrv_pcl_ctx_t * pctx = hdrv_pcl_state.find_vpc_ctx ();
|
|
|
|
if (!pctx)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: sig_ts=%llu: ERROR: hwcdrv_overflow:"
|
|
|
|
" tsd context is NULL\n", sig_ts);
|
|
|
|
return HWCFUNCS_ERROR_UNEXPECTED;
|
|
|
|
}
|
|
|
|
counter_state_t * ctr_list = (counter_state_t *) pctx->ctr_list;
|
|
|
|
if (!ctr_list)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: sig_ts=%llu: WARNING: hwcdrv_overflow:"
|
|
|
|
" ctr_list is NULL\n", sig_ts);
|
|
|
|
return HWCFUNCS_ERROR_UNEXPECTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* clear needs_restart flag */
|
|
|
|
for (ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
ctr_list[ii].needs_restart = 0;
|
|
|
|
|
|
|
|
/* attempt to identify the counter to read */
|
|
|
|
int signal_idx = -1;
|
|
|
|
pctx->signal_fd = signal_fd; // save the signal provided by siginfo_t
|
|
|
|
if (signal_fd != -1)
|
|
|
|
{
|
|
|
|
for (ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
{
|
|
|
|
if (ctr_list[ii].fd == signal_fd)
|
|
|
|
{
|
|
|
|
signal_idx = ii;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (signal_idx < 0)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: sig_ts=%llu: ERROR: hwcdrv_overflow:"
|
|
|
|
" pmc not determined!\n", sig_ts);
|
|
|
|
lost_events->ce_pic[0] = 1; /* record a bogus value into experiment */
|
|
|
|
// note: bogus value may get overwritten in loop below
|
|
|
|
}
|
|
|
|
|
|
|
|
/* capture sample(s). In addition to signal_idx, check other counters. */
|
|
|
|
struct perf_event_header sheader;
|
|
|
|
int idx;
|
|
|
|
for (idx = 0; idx < hdrv_pcl_state.hwcdef_cnt; idx++)
|
|
|
|
{
|
|
|
|
int num_recs = 0;
|
|
|
|
while (1)
|
|
|
|
{
|
|
|
|
/* check for samples */
|
|
|
|
struct perf_event_mmap_page *metadata = ctr_list[idx].buf_state.buf;
|
|
|
|
if (metadata == NULL)
|
|
|
|
break; // empty
|
|
|
|
if (metadata->data_tail == metadata->data_head)
|
|
|
|
break; // empty
|
|
|
|
|
|
|
|
/* read header */
|
|
|
|
if (read_buf (&ctr_list[idx].buf_state, &sheader, sizeof (sheader)))
|
|
|
|
break;
|
|
|
|
num_recs++;
|
|
|
|
|
|
|
|
/* check for PERF_RECORD_SAMPLE */
|
|
|
|
size_t datasz = sheader.size - sizeof (struct perf_event_header);
|
|
|
|
if (sheader.type != PERF_RECORD_SAMPLE)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT2, "hwcdrv: sig_ts=%llu: WARNING: hwcdrv_overflow:"
|
|
|
|
" unexpected recd type=%d\n",
|
|
|
|
sig_ts, sheader.type);
|
|
|
|
if (skip_buf (&ctr_list[idx].buf_state, datasz))
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: sig_ts=%llu: ERROR: hwcdrv_overflow:"
|
|
|
|
" skip recd type=%d failed\n", sig_ts, sheader.type);
|
|
|
|
lost_events->ce_pic[idx] = 4; /* record a bogus value */
|
|
|
|
break; // failed to skip buffer??
|
|
|
|
}
|
|
|
|
lost_events->ce_pic[idx] = 2; /* record a bogus value */
|
|
|
|
continue; // advance to next record
|
|
|
|
}
|
|
|
|
|
|
|
|
/* type is PERF_RECORD_SAMPLE */
|
|
|
|
uint64_t value, lostv;
|
|
|
|
if (read_sample (&ctr_list[idx], datasz, &value, &lostv))
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: sig_ts=%llu: ERROR: hwcdrv_overflow:"
|
|
|
|
" read_sample() failed\n", sig_ts);
|
|
|
|
lost_events->ce_pic[idx] = 3; // record a bogus value
|
|
|
|
break; // failed to read sample data??
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT3, "hwcdrv: sig_ts=%llu: hwcdrv_overflow:"
|
|
|
|
" idx=%d value=%llu lost=%llu\n", (unsigned long long) sig_ts,
|
|
|
|
idx, (unsigned long long) value, (unsigned long long) lostv);
|
|
|
|
if (eventp->ce_pic[idx])
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT2, "hwcdrv: sig_ts=%llu: WARNING: hwcdrv_overflow:"
|
|
|
|
" idx=%d previous sample recorded as lost_event\n", sig_ts, idx);
|
|
|
|
lost_events->ce_pic[idx] += eventp->ce_pic[idx];
|
|
|
|
}
|
|
|
|
eventp->ce_pic[idx] = value;
|
|
|
|
lost_events->ce_pic[idx] += lostv;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* debug output for unexpected (but common) cases */
|
|
|
|
if (idx == signal_idx)
|
|
|
|
{
|
|
|
|
if (num_recs != 1)
|
|
|
|
TprintfT (DBG_LT2, "hwcdrv: sig_ts=%llu: WARNING: hwcdrv_overflow:"
|
|
|
|
" %d records for signal_idx=%d\n", sig_ts, num_recs, signal_idx);
|
|
|
|
}
|
|
|
|
else if (num_recs)
|
|
|
|
TprintfT (DBG_LT2, "hwcdrv: sig_ts=%llu: WARNING: hwcdrv_overflow:"
|
|
|
|
" %d unexpected record(s) for idx=%d (signal_idx=%d)\n",
|
|
|
|
sig_ts, num_recs, idx, signal_idx);
|
|
|
|
|
|
|
|
/* trigger counter restart whenever records were found */
|
|
|
|
if (num_recs)
|
|
|
|
{
|
|
|
|
/* check whether to adapt the overflow interval */
|
|
|
|
/* This is the Linux version.
|
|
|
|
* The Solaris version is in hwprofile.c collector_update_overflow_counters().
|
|
|
|
*/
|
|
|
|
hrtime_t min_time = global_perf_event_def[idx].min_time;
|
|
|
|
if (min_time > 0 // overflow interval is adaptive
|
|
|
|
&& sig_ts - ctr_list[idx].last_overflow_time < min_time) // last interval below min
|
|
|
|
{
|
|
|
|
/* pick a new overflow interval */
|
|
|
|
/* roughly doubled, but add funny numbers */
|
|
|
|
/* hopefully the result is prime or not a multiple of some # of ops/loop */
|
|
|
|
uint64_t new_period = 2 * ctr_list[idx].last_overflow_period + 37;
|
|
|
|
#if 0
|
|
|
|
// On Solaris, we report the adjustment to the log file.
|
|
|
|
// On Linux it's hard for us to do so since hwcdrv_pcl.c doesn't know about collector_interface, SP_JCMD_COMMENT, or COL_COMMENT_HWCADJ.
|
|
|
|
// For now we simply don't report.
|
|
|
|
collector_interface->writeLog ("<event kind=\"%s\" id=\"%d\">%s %d -> %d</event>\n",
|
|
|
|
SP_JCMD_COMMENT, COL_COMMENT_HWCADJ, global_perf_event_def[idx].name,
|
|
|
|
ctr_list[idx].last_overflow_period, new_period);
|
|
|
|
#endif
|
|
|
|
/* There are a variety of ways of resetting the period on Linux.
|
|
|
|
* The most elegant is
|
|
|
|
* ioctl(fd,PERF_EVENT_IOC_PERIOD,&period)
|
|
|
|
* but check the perf_event_open man page for PERF_EVENT_IOC_PERIOD:
|
|
|
|
* > Prior to Linux 2.6.36 this ioctl always failed due to a bug in the kernel.
|
|
|
|
* > Prior to Linux 3.14 (or 3.7 on ARM), the new period did not take effect
|
|
|
|
* until after the next overflow.
|
|
|
|
* So we're kind of stuck shutting the fd down and restarting it with the new period.
|
|
|
|
*/
|
|
|
|
if (stop_one_ctr (idx, ctr_list))
|
|
|
|
{
|
|
|
|
// EUGENE figure out what to do on error
|
|
|
|
}
|
|
|
|
ctr_list[idx].last_overflow_period = new_period;
|
|
|
|
if (start_one_ctr (idx, ctr_list[idx].buf_state.pagesz, pctx, "hwcdrv: ERROR: hwcdrv_overflow (readjust overflow):"))
|
|
|
|
{
|
|
|
|
// EUGENE figure out what to do on error
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ctr_list[idx].last_overflow_time = sig_ts;
|
|
|
|
#if 0
|
|
|
|
ctr_list[idx].needs_restart = 1;
|
|
|
|
#else // seems to be more reliable to restart here instead of hwcdrv_sighlr_restart()
|
|
|
|
internal_hwc_start (ctr_list[idx].fd);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0; // OK to restart counters
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_sighlr_restart (const hwc_event_t *pp)
|
|
|
|
{
|
|
|
|
#if 0 // restarting here doesn't seem to work as well as restarting in hwcdrv_overflow()
|
|
|
|
hdrv_pcl_ctx_t * pctx = hdrv_pcl_state.find_vpc_ctx ();
|
|
|
|
if (!pctx)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: ERROR: hwcdrv_sighlr_restart: find_vpc_ctx()==NULL\n");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
counter_state_t * ctr_list = (counter_state_t *) pctx->ctr_list;
|
|
|
|
if (!ctr_list)
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT0, "hwcdrv: WARNING: hwcdrv_sighlr_restart: ctr_list is NULL\n");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
int errors = 0;
|
|
|
|
for (int ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
{
|
|
|
|
if (ctr_list[ii].needs_restart)
|
|
|
|
errors |= internal_hwc_start (ctr_list[ii].fd);
|
|
|
|
ctr_list[ii].needs_restart = 0;
|
|
|
|
}
|
|
|
|
return errors;
|
|
|
|
#else
|
|
|
|
return 0;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
/* create counters based on hwcdef[] */
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_create_counters (unsigned hwcdef_cnt, Hwcentry *hwcdef)
|
|
|
|
{
|
|
|
|
if (hwcdef_cnt > hdrv_pcl_about.cpcN_npics)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("More than %d counters were specified\n"), hdrv_pcl_about.cpcN_npics); /*!*/
|
|
|
|
return HWCFUNCS_ERROR_HWCARGS;
|
|
|
|
}
|
|
|
|
if (hdrv_pcl_about.cpcN_cpuver == CPUVER_UNDEFINED)
|
|
|
|
{
|
|
|
|
logerr (GTXT ("Processor not supported\n"));
|
|
|
|
return HWCFUNCS_ERROR_HWCARGS;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* add counters */
|
|
|
|
for (unsigned idx = 0; idx < hwcdef_cnt; idx++)
|
|
|
|
{
|
|
|
|
perf_event_def_t *glb_event_def = &global_perf_event_def[idx];
|
|
|
|
memset (glb_event_def, 0, sizeof (perf_event_def_t));
|
|
|
|
unsigned int pmc_sel;
|
|
|
|
eventsel_t evntsel;
|
|
|
|
if (hwcfuncs_get_x86_eventsel (hwcdef[idx].reg_num,
|
|
|
|
hwcdef[idx].int_name, &evntsel, &pmc_sel))
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: hwcfuncs_get_x86_eventsel() failed\n");
|
|
|
|
return HWCFUNCS_ERROR_HWCARGS;
|
|
|
|
}
|
|
|
|
glb_event_def->reg_num = pmc_sel;
|
|
|
|
glb_event_def->eventsel = evntsel;
|
|
|
|
glb_event_def->counter_preload = hwcdef[idx].val;
|
|
|
|
glb_event_def->min_time = hwcdef[idx].min_time;
|
|
|
|
glb_event_def->name = strdup (hwcdef[idx].name); // memory leak??? very minor
|
|
|
|
init_perf_event (&glb_event_def->hw, glb_event_def->eventsel,
|
|
|
|
glb_event_def->counter_preload);
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: create_counters: pic=%u name='%s' interval=%lld"
|
|
|
|
"(min_time=%lld): reg_num=0x%x eventsel=0x%llx ireset=%lld usr=%lld sys=%lld\n",
|
|
|
|
idx, hwcdef[idx].int_name, (long long) glb_event_def->counter_preload,
|
|
|
|
(long long) glb_event_def->min_time, (int) glb_event_def->reg_num,
|
|
|
|
(long long) glb_event_def->eventsel,
|
|
|
|
(long long) HW_INTERVAL_PRESET (hwcdef[idx].val),
|
|
|
|
(long long) glb_event_def->hw.exclude_user,
|
|
|
|
(long long) glb_event_def->hw.exclude_kernel);
|
|
|
|
}
|
|
|
|
|
|
|
|
hdrv_pcl_state.hwcdef_cnt = hwcdef_cnt;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_free_counters () // note: only performs shutdown for this thread
|
|
|
|
{
|
|
|
|
hdrv_pcl_ctx_t * pctx;
|
|
|
|
if (!COUNTERS_ENABLED ())
|
|
|
|
return 0;
|
|
|
|
pctx = hdrv_pcl_state.find_vpc_ctx ();
|
|
|
|
if (!pctx)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: WARNING: hwcdrv_free_counters: tsd context is NULL\n");
|
|
|
|
return HWCFUNCS_ERROR_GENERIC;
|
|
|
|
}
|
|
|
|
counter_state_t *ctr_list = pctx->ctr_list;
|
|
|
|
if (!ctr_list)
|
|
|
|
{
|
|
|
|
// fork child: prolog suspends hwcs, then epilog frees them
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: WARNING: hwcdrv_free_counters: ctr_list is already NULL\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
int hwc_rc = 0;
|
|
|
|
for (int ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
if (stop_one_ctr (ii, ctr_list))
|
|
|
|
hwc_rc = HWCFUNCS_ERROR_GENERIC;
|
2022-09-14 16:11:45 +08:00
|
|
|
TprintfT (DBG_LT1, "hwcdrv: hwcdrv_free_counters(tid=0x%lx).\n", (long) pctx->tid);
|
2022-03-11 16:58:31 +08:00
|
|
|
pctx->ctr_list = NULL;
|
|
|
|
return hwc_rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_start (void) /* must be called from each thread ? */
|
|
|
|
{
|
|
|
|
hdrv_pcl_ctx_t *pctx = NULL;
|
|
|
|
if (!COUNTERS_ENABLED ())
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: WARNING: hwcdrv_start: no counters to start \n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if (!hdrv_pcl_state.library_ok)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: hwcdrv_start: library is not open\n");
|
|
|
|
return HWCFUNCS_ERROR_NOT_SUPPORTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* set up per-thread context
|
|
|
|
*/
|
|
|
|
pctx = hdrv_pcl_state.find_vpc_ctx ();
|
|
|
|
if (!pctx)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: hwcdrv_start: tsd context is NULL\n");
|
|
|
|
return HWCFUNCS_ERROR_UNEXPECTED;
|
|
|
|
}
|
|
|
|
pctx->tid = hwcdrv_gettid ();
|
2022-09-14 16:11:45 +08:00
|
|
|
TprintfT (DBG_LT1, "hwcdrv: hwcdrv_start(tid=0x%lx)\n", (long) pctx->tid);
|
2022-03-11 16:58:31 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* create per-thread counter list
|
|
|
|
*/
|
|
|
|
counter_state_t *ctr_list = (counter_state_t *) calloc (hdrv_pcl_state.hwcdef_cnt,
|
|
|
|
sizeof (counter_state_t));
|
|
|
|
if (!ctr_list)
|
|
|
|
{
|
|
|
|
TprintfT (0, "hwcdrv: ERROR: hwcdrv_start: calloc(ctr_list) failed\n");
|
|
|
|
return HWCFUNCS_ERROR_MEMORY;
|
|
|
|
}
|
|
|
|
int ii;
|
|
|
|
for (ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
ctr_list[ii].fd = -1; // invalidate fds in case we have to close prematurely
|
|
|
|
pctx->ctr_list = ctr_list;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* bind the counters
|
|
|
|
*/
|
|
|
|
size_t pgsz = sysconf (_SC_PAGESIZE);
|
|
|
|
for (ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
{
|
|
|
|
ctr_list[ii].last_overflow_period = global_perf_event_def[ii].hw.sample_period;
|
|
|
|
if (start_one_ctr (ii, pgsz, pctx, "hwcdrv: ERROR: hwcdrv_start:")) goto hwcdrv_start_cleanup;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* start the counters
|
|
|
|
*/
|
|
|
|
for (ii = 0; ii < hdrv_pcl_state.hwcdef_cnt; ii++)
|
|
|
|
{
|
|
|
|
int rc = internal_hwc_start (ctr_list[ii].fd);
|
|
|
|
if (rc < 0)
|
|
|
|
goto hwcdrv_start_cleanup;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
hwcdrv_start_cleanup:
|
|
|
|
hwcdrv_free_counters (); // PERF_EVENT_IOC_DISABLE and close() for all fds
|
|
|
|
return HWCFUNCS_ERROR_UNAVAIL;
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_lwp_suspend (void) /* must be called from each thread */
|
|
|
|
{
|
|
|
|
if (!COUNTERS_ENABLED ())
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: WARNING: hwcdrv_lwp_suspend: no counters\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: hwcdrv_lwp_suspend()\n");
|
|
|
|
return hwcdrv_free_counters ();
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_lwp_resume (void) /* must be called from each thread */
|
|
|
|
{
|
|
|
|
if (!COUNTERS_ENABLED ())
|
|
|
|
{
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: WARNING: hwcdrv_lwp_resume: no counters\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
TprintfT (DBG_LT1, "hwcdrv: hwcdrv_lwp_resume()\n");
|
|
|
|
return hwcdrv_start ();
|
|
|
|
}
|
|
|
|
|
|
|
|
HWCDRV_API int
|
|
|
|
hwcdrv_read_events (hwc_event_t *overflow_data, hwc_event_samples_t *sampled_data)
|
|
|
|
{
|
|
|
|
overflow_data->ce_hrt = 0;
|
|
|
|
for (int i = 0; i < MAX_PICS; i++)
|
|
|
|
{
|
|
|
|
overflow_data->ce_pic[i] = 0;
|
|
|
|
if (sampled_data)
|
|
|
|
HWCFUNCS_SAMPLE_RESET (&sampled_data->sample[i]);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
|
|
/* HWCDRV_API */
|
|
|
|
|
|
|
|
hwcdrv_api_t hwcdrv_pcl_api = {
|
|
|
|
hwcdrv_init,
|
|
|
|
hwcdrv_get_info,
|
|
|
|
hwcdrv_enable_mt,
|
|
|
|
hwcdrv_get_descriptions,
|
|
|
|
hwcdrv_assign_regnos,
|
|
|
|
hwcdrv_create_counters,
|
|
|
|
hwcdrv_start,
|
|
|
|
hwcdrv_overflow,
|
|
|
|
hwcdrv_read_events,
|
|
|
|
hwcdrv_sighlr_restart,
|
|
|
|
hwcdrv_lwp_suspend,
|
|
|
|
hwcdrv_lwp_resume,
|
|
|
|
hwcdrv_free_counters,
|
|
|
|
hwcdrv_lwp_init,
|
|
|
|
hwcdrv_lwp_fini,
|
|
|
|
-1 // hwcdrv_init_status
|
|
|
|
};
|