git: d9e3fe322662 - main - hpwmc_amd.c: style improvements

From: Mitchell Horne <mhorne_at_FreeBSD.org>
Date: Wed, 18 Oct 2023 18:06:00 UTC
The branch main has been updated by mhorne:

URL: https://cgit.FreeBSD.org/src/commit/?id=d9e3fe3226624c391a80b623d844bb578888cec3

commit d9e3fe3226624c391a80b623d844bb578888cec3
Author:     Mitchell Horne <mhorne@FreeBSD.org>
AuthorDate: 2023-10-18 17:53:27 +0000
Commit:     Mitchell Horne <mhorne@FreeBSD.org>
CommitDate: 2023-10-18 18:05:44 +0000

    hpwmc_amd.c: style improvements
    
     - Return style
     - Explicit value checks
     - Whitespace formatting
     - Comment formatting
     - Local variable declaration order
     - __unused annotations
    
    Reviewed by:    jkoshy
    MFC after:      1 week
    Sponsored by:   The FreeBSD Foundation
    Differential Revision:  https://reviews.freebsd.org/D41272
---
 sys/dev/hwpmc/hwpmc_amd.c | 280 +++++++++++++++++++---------------------------
 1 file changed, 116 insertions(+), 164 deletions(-)

diff --git a/sys/dev/hwpmc/hwpmc_amd.c b/sys/dev/hwpmc/hwpmc_amd.c
index 000958eb9945..d8ca2e094d02 100644
--- a/sys/dev/hwpmc/hwpmc_amd.c
+++ b/sys/dev/hwpmc/hwpmc_amd.c
@@ -383,23 +383,20 @@ const int amd_event_codes_size = nitems(amd_event_codes);
 /*
  * Per-processor information
  */
-
 struct amd_cpu {
 	struct pmc_hw	pc_amdpmcs[AMD_NPMCS];
 };
-
 static struct amd_cpu **amd_pcpu;
 
 /*
- * read a pmc register
+ * Read a PMC value from the MSR.
  */
-
 static int
 amd_read_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t *v)
 {
-	enum pmc_mode mode;
 	const struct amd_descr *pd;
 	pmc_value_t tmp;
+	enum pmc_mode mode;
 
 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
@@ -411,7 +408,8 @@ amd_read_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t *v)
 	pd = &amd_pmcdesc[ri];
 	mode = PMC_TO_MODE(pm);
 
-	PMCDBG2(MDP,REA,1,"amd-read id=%d class=%d", ri, pd->pm_descr.pd_class);
+	PMCDBG2(MDP, REA, 1, "amd-read id=%d class=%d", ri,
+	    pd->pm_descr.pd_class);
 
 #ifdef	HWPMC_DEBUG
 	KASSERT(pd->pm_descr.pd_class == amd_pmc_class,
@@ -420,7 +418,7 @@ amd_read_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t *v)
 #endif
 
 	tmp = rdmsr(pd->pm_perfctr); /* RDMSR serializes */
-	PMCDBG2(MDP,REA,2,"amd-read (pre-munge) id=%d -> %jd", ri, tmp);
+	PMCDBG2(MDP, REA, 2, "amd-read (pre-munge) id=%d -> %jd", ri, tmp);
 	if (PMC_IS_SAMPLING_MODE(mode)) {
 		/*
 		 * Clamp value to 0 if the counter just overflowed,
@@ -437,15 +435,14 @@ amd_read_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t *v)
 	}
 	*v = tmp;
 
-	PMCDBG2(MDP,REA,2,"amd-read (post-munge) id=%d -> %jd", ri, *v);
+	PMCDBG2(MDP, REA, 2, "amd-read (post-munge) id=%d -> %jd", ri, *v);
 
-	return 0;
+	return (0);
 }
 
 /*
  * Write a PMC MSR.
  */
-
 static int
 amd_write_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t v)
 {
@@ -470,24 +467,22 @@ amd_write_pmc(int cpu, int ri, struct pmc *pm, pmc_value_t v)
 	if (PMC_IS_SAMPLING_MODE(mode))
 		v = AMD_RELOAD_COUNT_TO_PERFCTR_VALUE(v);
 
-	PMCDBG3(MDP,WRI,1,"amd-write cpu=%d ri=%d v=%jx", cpu, ri, v);
+	PMCDBG3(MDP, WRI, 1, "amd-write cpu=%d ri=%d v=%jx", cpu, ri, v);
 
 	/* write the PMC value */
 	wrmsr(pd->pm_perfctr, v);
-	return 0;
+	return (0);
 }
 
 /*
- * configure hardware pmc according to the configuration recorded in
- * pmc 'pm'.
+ * Configure hardware PMC according to the configuration recorded in 'pm'.
  */
-
 static int
 amd_config_pmc(int cpu, int ri, struct pmc *pm)
 {
 	struct pmc_hw *phw;
 
-	PMCDBG3(MDP,CFG,1, "cpu=%d ri=%d pm=%p", cpu, ri, pm);
+	PMCDBG3(MDP, CFG, 1, "cpu=%d ri=%d pm=%p", cpu, ri, pm);
 
 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
@@ -501,78 +496,65 @@ amd_config_pmc(int cpu, int ri, struct pmc *pm)
 		__LINE__, pm, phw->phw_pmc));
 
 	phw->phw_pmc = pm;
-	return 0;
+	return (0);
 }
 
 /*
  * Retrieve a configured PMC pointer from hardware state.
  */
-
 static int
 amd_get_config(int cpu, int ri, struct pmc **ppm)
 {
 	*ppm = amd_pcpu[cpu]->pc_amdpmcs[ri].phw_pmc;
-
-	return 0;
+	return (0);
 }
 
 /*
- * Machine dependent actions taken during the context switch in of a
+ * Machine-dependent actions taken during the context switch in of a
  * thread.
  */
-
 static int
-amd_switch_in(struct pmc_cpu *pc, struct pmc_process *pp)
+amd_switch_in(struct pmc_cpu *pc __pmcdbg_used, struct pmc_process *pp)
 {
-	(void) pc;
-
-	PMCDBG3(MDP,SWI,1, "pc=%p pp=%p enable-msr=%d", pc, pp,
+	PMCDBG3(MDP, SWI, 1, "pc=%p pp=%p enable-msr=%d", pc, pp,
 	    (pp->pp_flags & PMC_PP_ENABLE_MSR_ACCESS) != 0);
 
 	/* enable the RDPMC instruction if needed */
 	if (pp->pp_flags & PMC_PP_ENABLE_MSR_ACCESS)
 		load_cr4(rcr4() | CR4_PCE);
 
-	return 0;
+	return (0);
 }
 
 /*
- * Machine dependent actions taken during the context switch out of a
+ * Machine-dependent actions taken during the context switch out of a
  * thread.
  */
-
 static int
-amd_switch_out(struct pmc_cpu *pc, struct pmc_process *pp)
+amd_switch_out(struct pmc_cpu *pc __pmcdbg_used,
+    struct pmc_process *pp __pmcdbg_used)
 {
-	(void) pc;
-	(void) pp;		/* can be NULL */
-
-	PMCDBG3(MDP,SWO,1, "pc=%p pp=%p enable-msr=%d", pc, pp, pp ?
+	PMCDBG3(MDP, SWO, 1, "pc=%p pp=%p enable-msr=%d", pc, pp, pp ?
 	    (pp->pp_flags & PMC_PP_ENABLE_MSR_ACCESS) == 1 : 0);
 
 	/* always turn off the RDPMC instruction */
 	load_cr4(rcr4() & ~CR4_PCE);
 
-	return 0;
+	return (0);
 }
 
 /*
- * Check if a given allocation is feasible.
+ * Check if a given PMC allocation is feasible.
  */
-
 static int
-amd_allocate_pmc(int cpu, int ri, struct pmc *pm,
+amd_allocate_pmc(int cpu __unused, int ri, struct pmc *pm,
     const struct pmc_op_pmcallocate *a)
 {
-	int i;
+	const struct pmc_descr *pd;
 	uint64_t allowed_unitmask, caps, config, unitmask;
 	enum pmc_event pe;
-	const struct pmc_descr *pd;
-
-	(void) cpu;
+	int i;
 
-	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
-	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
 	    ("[amd,%d] illegal row index %d", __LINE__, ri));
 
@@ -580,26 +562,30 @@ amd_allocate_pmc(int cpu, int ri, struct pmc *pm,
 
 	/* check class match */
 	if (pd->pd_class != a->pm_class)
-		return EINVAL;
+		return (EINVAL);
 
 	if ((a->pm_flags & PMC_F_EV_PMU) == 0)
 		return (EINVAL);
 
 	caps = pm->pm_caps;
 
-	PMCDBG2(MDP,ALL,1,"amd-allocate ri=%d caps=0x%x", ri, caps);
+	PMCDBG2(MDP, ALL, 1,"amd-allocate ri=%d caps=0x%x", ri, caps);
 
-	if((ri >= 0 && ri < 6) && !(a->pm_md.pm_amd.pm_amd_sub_class == PMC_AMD_SUB_CLASS_CORE))
-		return EINVAL;
-	if((ri >= 6 && ri < 12) && !(a->pm_md.pm_amd.pm_amd_sub_class == PMC_AMD_SUB_CLASS_L3_CACHE))
-		return EINVAL;
-	if((ri >= 12 && ri < 16) && !(a->pm_md.pm_amd.pm_amd_sub_class == PMC_AMD_SUB_CLASS_DATA_FABRIC))
-		return EINVAL;
+	/* Validate sub-class. */
+	if ((ri >= 0 && ri < 6) && a->pm_md.pm_amd.pm_amd_sub_class !=
+	    PMC_AMD_SUB_CLASS_CORE)
+		return (EINVAL);
+	if ((ri >= 6 && ri < 12) && a->pm_md.pm_amd.pm_amd_sub_class !=
+	    PMC_AMD_SUB_CLASS_L3_CACHE)
+		return (EINVAL);
+	if ((ri >= 12 && ri < 16) && a->pm_md.pm_amd.pm_amd_sub_class !=
+	    PMC_AMD_SUB_CLASS_DATA_FABRIC)
+		return (EINVAL);
 
 	if (strlen(pmc_cpuid) != 0) {
-		pm->pm_md.pm_amd.pm_amd_evsel =
-			a->pm_md.pm_amd.pm_amd_config;
-		PMCDBG2(MDP,ALL,2,"amd-allocate ri=%d -> config=0x%x", ri, a->pm_md.pm_amd.pm_amd_config);
+		pm->pm_md.pm_amd.pm_amd_evsel = a->pm_md.pm_amd.pm_amd_config;
+		PMCDBG2(MDP, ALL, 2,"amd-allocate ri=%d -> config=0x%x", ri,
+		    a->pm_md.pm_amd.pm_amd_config);
 		return (0);
 	}
 
@@ -607,7 +593,7 @@ amd_allocate_pmc(int cpu, int ri, struct pmc *pm,
 
 	/* map ev to the correct event mask code */
 	config = allowed_unitmask = 0;
-	for (i = 0; i < amd_event_codes_size; i++)
+	for (i = 0; i < amd_event_codes_size; i++) {
 		if (amd_event_codes[i].pe_ev == pe) {
 			config =
 			    AMD_PMC_TO_EVENTMASK(amd_event_codes[i].pe_code);
@@ -615,58 +601,52 @@ amd_allocate_pmc(int cpu, int ri, struct pmc *pm,
 			    AMD_PMC_TO_UNITMASK(amd_event_codes[i].pe_mask);
 			break;
 		}
+	}
 	if (i == amd_event_codes_size)
-		return EINVAL;
+		return (EINVAL);
 
 	unitmask = a->pm_md.pm_amd.pm_amd_config & AMD_PMC_UNITMASK;
-	if (unitmask & ~allowed_unitmask) /* disallow reserved bits */
-		return EINVAL;
+	if ((unitmask & ~allowed_unitmask) != 0) /* disallow reserved bits */
+		return (EINVAL);
 
-	if (unitmask && (caps & PMC_CAP_QUALIFIER))
+	if (unitmask && (caps & PMC_CAP_QUALIFIER) != 0)
 		config |= unitmask;
 
-	if (caps & PMC_CAP_THRESHOLD)
+	if ((caps & PMC_CAP_THRESHOLD) != 0)
 		config |= a->pm_md.pm_amd.pm_amd_config & AMD_PMC_COUNTERMASK;
 
-	/* set at least one of the 'usr' or 'os' caps */
-	if (caps & PMC_CAP_USER)
+	/* Set at least one of the 'usr' or 'os' caps. */
+	if ((caps & PMC_CAP_USER) != 0)
 		config |= AMD_PMC_USR;
-	if (caps & PMC_CAP_SYSTEM)
+	if ((caps & PMC_CAP_SYSTEM) != 0)
 		config |= AMD_PMC_OS;
-	if ((caps & (PMC_CAP_USER|PMC_CAP_SYSTEM)) == 0)
+	if ((caps & (PMC_CAP_USER | PMC_CAP_SYSTEM)) == 0)
 		config |= (AMD_PMC_USR|AMD_PMC_OS);
 
-	if (caps & PMC_CAP_EDGE)
+	if ((caps & PMC_CAP_EDGE) != 0)
 		config |= AMD_PMC_EDGE;
-	if (caps & PMC_CAP_INVERT)
+	if ((caps & PMC_CAP_INVERT) != 0)
 		config |= AMD_PMC_INVERT;
-	if (caps & PMC_CAP_INTERRUPT)
+	if ((caps & PMC_CAP_INTERRUPT) != 0)
 		config |= AMD_PMC_INT;
 
 	pm->pm_md.pm_amd.pm_amd_evsel = config; /* save config value */
 
-	PMCDBG2(MDP,ALL,2,"amd-allocate ri=%d -> config=0x%x", ri, config);
+	PMCDBG2(MDP, ALL, 2, "amd-allocate ri=%d -> config=0x%x", ri, config);
 
-	return 0;
+	return (0);
 }
 
 /*
  * Release machine dependent state associated with a PMC.  This is a
  * no-op on this architecture.
- *
  */
-
-/* ARGSUSED0 */
 static int
-amd_release_pmc(int cpu, int ri, struct pmc *pmc)
+amd_release_pmc(int cpu, int ri, struct pmc *pmc __unused)
 {
-#ifdef	HWPMC_DEBUG
-	const struct amd_descr *pd;
-#endif
+	const struct amd_descr *pd __pmcdbg_used;
 	struct pmc_hw *phw __diagused;
 
-	(void) pmc;
-
 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
 	KASSERT(ri >= 0 && ri < AMD_NPMCS,
@@ -677,25 +657,23 @@ amd_release_pmc(int cpu, int ri, struct pmc *pmc)
 	KASSERT(phw->phw_pmc == NULL,
 	    ("[amd,%d] PHW pmc %p non-NULL", __LINE__, phw->phw_pmc));
 
-#ifdef	HWPMC_DEBUG
+#ifdef HWPMC_DEBUG
 	pd = &amd_pmcdesc[ri];
 	if (pd->pm_descr.pd_class == amd_pmc_class)
 		KASSERT(AMD_PMC_IS_STOPPED(pd->pm_evsel),
 		    ("[amd,%d] PMC %d released while active", __LINE__, ri));
 #endif
-
-	return 0;
+	return (0);
 }
 
 /*
- * start a PMC.
+ * Start a PMC.
  */
-
 static int
-amd_start_pmc(int cpu, int ri, struct pmc *pm)
+amd_start_pmc(int cpu __diagused, int ri, struct pmc *pm)
 {
-	uint64_t config;
 	const struct amd_descr *pd;
+	uint64_t config;
 
 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
 	    ("[amd,%d] illegal CPU value %d", __LINE__, cpu));
@@ -704,7 +682,7 @@ amd_start_pmc(int cpu, int ri, struct pmc *pm)
 
 	pd = &amd_pmcdesc[ri];
 
-	PMCDBG2(MDP,STA,1,"amd-start cpu=%d ri=%d", cpu, ri);
+	PMCDBG2(MDP, STA, 1, "amd-start cpu=%d ri=%d", cpu, ri);
 
 	KASSERT(AMD_PMC_IS_STOPPED(pd->pm_evsel),
 	    ("[amd,%d] pmc%d,cpu%d: Starting active PMC \"%s\"", __LINE__,
@@ -713,18 +691,17 @@ amd_start_pmc(int cpu, int ri, struct pmc *pm)
 	/* turn on the PMC ENABLE bit */
 	config = pm->pm_md.pm_amd.pm_amd_evsel | AMD_PMC_ENABLE;
 
-	PMCDBG1(MDP,STA,2,"amd-start config=0x%x", config);
+	PMCDBG1(MDP, STA, 2, "amd-start config=0x%x", config);
 
 	wrmsr(pd->pm_evsel, config);
-	return 0;
+	return (0);
 }
 
 /*
  * Stop a PMC.
  */
-
 static int
-amd_stop_pmc(int cpu, int ri, struct pmc *pm)
+amd_stop_pmc(int cpu __diagused, int ri, struct pmc *pm)
 {
 	const struct amd_descr *pd;
 	uint64_t config;
@@ -741,7 +718,7 @@ amd_stop_pmc(int cpu, int ri, struct pmc *pm)
 	    ("[amd,%d] PMC%d, CPU%d \"%s\" already stopped",
 		__LINE__, ri, cpu, pd->pm_descr.pd_name));
 
-	PMCDBG1(MDP,STO,1,"amd-stop ri=%d", ri);
+	PMCDBG1(MDP, STO, 1, "amd-stop ri=%d", ri);
 
 	/* turn off the PMC ENABLE bit */
 	config = pm->pm_md.pm_amd.pm_amd_evsel & ~AMD_PMC_ENABLE;
@@ -761,7 +738,7 @@ amd_stop_pmc(int cpu, int ri, struct pmc *pm)
 		DELAY(1);
 	}
 
-	return 0;
+	return (0);
 }
 
 /*
@@ -770,23 +747,21 @@ amd_stop_pmc(int cpu, int ri, struct pmc *pm)
  * to sleep or do anything a 'fast' interrupt handler is not allowed
  * to do.
  */
-
 static int
 amd_intr(struct trapframe *tf)
 {
-	int i, error, retval, cpu;
-	uint64_t config, evsel, perfctr;
-	struct pmc *pm;
 	struct amd_cpu *pac;
+	struct pmc *pm;
 	pmc_value_t v;
+	uint64_t config, evsel, perfctr;
 	uint32_t active = 0, count = 0;
+	int i, error, retval, cpu;
 
 	cpu = curcpu;
 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
 	    ("[amd,%d] out of range CPU %d", __LINE__, cpu));
 
-	PMCDBG3(MDP,INT,1, "cpu=%d tf=%p um=%d", cpu, (void *) tf,
-	    TRAPF_USERMODE(tf));
+	PMCDBG3(MDP, INT, 1, "cpu=%d tf=%p um=%d", cpu, tf, TRAPF_USERMODE(tf));
 
 	retval = 0;
 
@@ -803,9 +778,7 @@ amd_intr(struct trapframe *tf)
 	 * a single interrupt. Check all the valid pmcs for
 	 * overflow.
 	 */
-
 	for (i = 0; i < AMD_CORE_NPMCS; i++) {
-
 		if ((pm = pac->pc_amdpmcs[i].phw_pmc) == NULL ||
 		    !PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm))) {
 			continue;
@@ -823,8 +796,8 @@ amd_intr(struct trapframe *tf)
 			continue;
 
 		/* Stop the PMC, reload count. */
-		evsel	= amd_pmcdesc[i].pm_evsel;
-		perfctr	= amd_pmcdesc[i].pm_perfctr;
+		evsel   = amd_pmcdesc[i].pm_evsel;
+		perfctr = amd_pmcdesc[i].pm_perfctr;
 		v       = pm->pm_sc.pm_reloadcount;
 		config  = rdmsr(evsel);
 
@@ -852,7 +825,6 @@ amd_intr(struct trapframe *tf)
 	 * if this NMI was for a pmc overflow which was serviced
 	 * in an earlier request or should be ignored.
 	 */
-
 	if (retval) {
 		DPCPU_SET(nmi_counter, min(2, active));
 	} else {
@@ -867,12 +839,12 @@ amd_intr(struct trapframe *tf)
 	else
 		counter_u64_add(pmc_stats.pm_intr_ignored, 1);
 
-	PMCDBG1(MDP,INT,2, "retval=%d", retval);
+	PMCDBG1(MDP, INT, 2, "retval=%d", retval);
 	return (retval);
 }
 
 /*
- * describe a PMC
+ * Describe a PMC.
  */
 static int
 amd_describe(int cpu, int ri, struct pmc_info *pi, struct pmc **ppmc)
@@ -891,25 +863,20 @@ amd_describe(int cpu, int ri, struct pmc_info *pi, struct pmc **ppmc)
 	strlcpy(pi->pm_name, pd->pm_descr.pd_name, sizeof(pi->pm_name));
 	pi->pm_class = pd->pm_descr.pd_class;
 
-	if (phw->phw_state & PMC_PHW_FLAG_IS_ENABLED) {
-		pi->pm_enabled = TRUE;
+	if ((phw->phw_state & PMC_PHW_FLAG_IS_ENABLED) != 0) {
+		pi->pm_enabled = true;
 		*ppmc          = phw->phw_pmc;
 	} else {
-		pi->pm_enabled = FALSE;
+		pi->pm_enabled = false;
 		*ppmc          = NULL;
 	}
 
-	return 0;
+	return (0);
 }
 
 /*
- * i386 specific entry points
+ * Return the MSR address of the given PMC.
  */
-
-/*
- * return the MSR address of the given PMC.
- */
-
 static int
 amd_get_msr(int ri, uint32_t *msr)
 {
@@ -917,29 +884,27 @@ amd_get_msr(int ri, uint32_t *msr)
 	    ("[amd,%d] ri %d out of range", __LINE__, ri));
 
 	*msr = amd_pmcdesc[ri].pm_perfctr - AMD_PMC_PERFCTR_0;
-
 	return (0);
 }
 
 /*
- * processor dependent initialization.
+ * Processor-dependent initialization.
  */
-
 static int
 amd_pcpu_init(struct pmc_mdep *md, int cpu)
 {
-	int classindex, first_ri, n;
-	struct pmc_cpu *pc;
 	struct amd_cpu *pac;
+	struct pmc_cpu *pc;
 	struct pmc_hw  *phw;
+	int classindex, first_ri, n;
 
 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
 	    ("[amd,%d] insane cpu number %d", __LINE__, cpu));
 
-	PMCDBG1(MDP,INI,1,"amd-init cpu=%d", cpu);
+	PMCDBG1(MDP, INI, 1, "amd-init cpu=%d", cpu);
 
 	amd_pcpu[cpu] = pac = malloc(sizeof(struct amd_cpu), M_PMC,
-	    M_WAITOK|M_ZERO);
+	    M_WAITOK | M_ZERO);
 
 	/*
 	 * Set the content of the hardware descriptors to a known
@@ -957,33 +922,30 @@ amd_pcpu_init(struct pmc_mdep *md, int cpu)
 	KASSERT(pc != NULL, ("[amd,%d] NULL per-cpu pointer", __LINE__));
 
 	for (n = 0, phw = pac->pc_amdpmcs; n < AMD_NPMCS; n++, phw++) {
-		phw->phw_state 	  = PMC_PHW_FLAG_IS_ENABLED |
+		phw->phw_state = PMC_PHW_FLAG_IS_ENABLED |
 		    PMC_PHW_CPU_TO_STATE(cpu) | PMC_PHW_INDEX_TO_STATE(n);
-		phw->phw_pmc	  = NULL;
-		pc->pc_hwpmcs[n + first_ri]  = phw;
+		phw->phw_pmc = NULL;
+		pc->pc_hwpmcs[n + first_ri] = phw;
 	}
 
 	return (0);
 }
 
-
 /*
- * processor dependent cleanup prior to the KLD
- * being unloaded
+ * Processor-dependent cleanup prior to the KLD being unloaded.
  */
-
 static int
 amd_pcpu_fini(struct pmc_mdep *md, int cpu)
 {
-	int classindex, first_ri, i;
-	uint32_t evsel;
-	struct pmc_cpu *pc;
 	struct amd_cpu *pac;
+	struct pmc_cpu *pc;
+	uint32_t evsel;
+	int classindex, first_ri, i;
 
 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
 	    ("[amd,%d] insane cpu number (%d)", __LINE__, cpu));
 
-	PMCDBG1(MDP,INI,1,"amd-cleanup cpu=%d", cpu);
+	PMCDBG1(MDP, INI, 1, "amd-cleanup cpu=%d", cpu);
 
 	/*
 	 * First, turn off all PMCs on this CPU.
@@ -1014,9 +976,9 @@ amd_pcpu_fini(struct pmc_mdep *md, int cpu)
 	pc = pmc_pcpu[cpu];
 	KASSERT(pc != NULL, ("[amd,%d] NULL per-cpu state", __LINE__));
 
-#if	defined(__amd64__)
+#ifdef __amd64__
 	classindex = PMC_MDEP_CLASS_INDEX_K8;
-#elif	defined(__i386__)
+#else
 	classindex = md->pmd_cputype == PMC_CPU_AMD_K8 ? PMC_MDEP_CLASS_INDEX_K8 :
 	    PMC_MDEP_CLASS_INDEX_K7;
 #endif
@@ -1025,30 +987,26 @@ amd_pcpu_fini(struct pmc_mdep *md, int cpu)
 	/*
 	 * Reset pointers in the MI 'per-cpu' state.
 	 */
-	for (i = 0; i < AMD_NPMCS; i++) {
+	for (i = 0; i < AMD_NPMCS; i++)
 		pc->pc_hwpmcs[i + first_ri] = NULL;
-	}
-
 
 	free(pac, M_PMC);
-
 	return (0);
 }
 
 /*
  * Initialize ourselves.
  */
-
 struct pmc_mdep *
 pmc_amd_initialize(void)
 {
-	int classindex, error, i, ncpus;
 	struct pmc_classdep *pcd;
-	enum pmc_cputype cputype;
 	struct pmc_mdep *pmc_mdep;
+	char *name;
+	enum pmc_cputype cputype;
 	enum pmc_class class;
+	int classindex, error, i, ncpus;
 	int family, model, stepping;
-	char *name;
 
 	/*
 	 * The presence of hardware performance counters on the AMD
@@ -1071,7 +1029,7 @@ pmc_amd_initialize(void)
 		    family, model, stepping);
 
 	switch (cpu_id & 0xF00) {
-#if	defined(__i386__)
+#ifdef __i386__
 	case 0x600:		/* Athlon(tm) processor */
 		classindex = PMC_MDEP_CLASS_INDEX_K7;
 		cputype = PMC_CPU_AMD_K7;
@@ -1087,11 +1045,12 @@ pmc_amd_initialize(void)
 		break;
 
 	default:
-		(void) printf("pmc: Unknown AMD CPU %x %d-%d.\n", cpu_id, (cpu_id & 0xF00) >> 8, model);
-		return NULL;
+		printf("pmc: Unknown AMD CPU %x %d-%d.\n", cpu_id, family,
+		    model);
+		return (NULL);
 	}
 
-#ifdef	HWPMC_DEBUG
+#ifdef HWPMC_DEBUG
 	amd_pmc_class = class;
 #endif
 
@@ -1100,7 +1059,7 @@ pmc_amd_initialize(void)
 	 * the MDEP structure used by MI code.
 	 */
 	amd_pcpu = malloc(sizeof(struct amd_cpu *) * pmc_cpu_max(), M_PMC,
-	    M_WAITOK|M_ZERO);
+	    M_WAITOK | M_ZERO);
 
 	/*
 	 * These processors have two classes of PMCs: the TSC and
@@ -1108,13 +1067,11 @@ pmc_amd_initialize(void)
 	 */
 	pmc_mdep = pmc_mdep_alloc(2);
 
-	pmc_mdep->pmd_cputype = cputype;
-
 	ncpus = pmc_cpu_max();
 
 	/* Initialize the TSC. */
 	error = pmc_tsc_initialize(pmc_mdep, ncpus);
-	if (error)
+	if (error != 0)
 		goto error;
 
 	/* Initialize AMD K7 and K8 PMC handling. */
@@ -1128,8 +1085,7 @@ pmc_amd_initialize(void)
 
 	/* fill in the correct pmc name and class */
 	for (i = 0; i < AMD_NPMCS; i++) {
-		(void) snprintf(amd_pmcdesc[i].pm_descr.pd_name,
-		    sizeof(amd_pmcdesc[i].pm_descr.pd_name), "%s-%d",
+		snprintf(amd_pmcdesc[i].pm_descr.pd_name, PMC_NAME_MAX, "%s-%d",
 		    name, i);
 		amd_pmcdesc[i].pm_descr.pd_class = class;
 	}
@@ -1147,29 +1103,25 @@ pmc_amd_initialize(void)
 	pcd->pcd_stop_pmc	= amd_stop_pmc;
 	pcd->pcd_write_pmc	= amd_write_pmc;
 
+	pmc_mdep->pmd_cputype	= cputype;
 	pmc_mdep->pmd_intr	= amd_intr;
-	pmc_mdep->pmd_switch_in = amd_switch_in;
+	pmc_mdep->pmd_switch_in	= amd_switch_in;
 	pmc_mdep->pmd_switch_out = amd_switch_out;
 
-	pmc_mdep->pmd_npmc     += AMD_NPMCS;
+	pmc_mdep->pmd_npmc	+= AMD_NPMCS;
 
-	PMCDBG0(MDP,INI,0,"amd-initialize");
+	PMCDBG0(MDP, INI, 0, "amd-initialize");
 
 	return (pmc_mdep);
 
-  error:
-	if (error) {
-		free(pmc_mdep, M_PMC);
-		pmc_mdep = NULL;
-	}
-
+error:
+	free(pmc_mdep, M_PMC);
 	return (NULL);
 }
 
 /*
  * Finalization code for AMD CPUs.
  */
-
 void
 pmc_amd_finalize(struct pmc_mdep *md)
 {