|
@@ -169,6 +169,9 @@ struct hvcall_ppp_data {
|
|
u8 unallocated_weight;
|
|
u8 unallocated_weight;
|
|
u16 active_procs_in_pool;
|
|
u16 active_procs_in_pool;
|
|
u16 active_system_procs;
|
|
u16 active_system_procs;
|
|
|
|
+ u16 phys_platform_procs;
|
|
|
|
+ u32 max_proc_cap_avail;
|
|
|
|
+ u32 entitled_proc_cap_avail;
|
|
};
|
|
};
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -190,13 +193,18 @@ struct hvcall_ppp_data {
|
|
* XX - Unallocated Variable Processor Capacity Weight.
|
|
* XX - Unallocated Variable Processor Capacity Weight.
|
|
* XXXX - Active processors in Physical Processor Pool.
|
|
* XXXX - Active processors in Physical Processor Pool.
|
|
* XXXX - Processors active on platform.
|
|
* XXXX - Processors active on platform.
|
|
|
|
+ * R8 (QQQQRRRRRRSSSSSS). if ibm,partition-performance-parameters-level >= 1
|
|
|
|
+ * XXXX - Physical platform procs allocated to virtualization.
|
|
|
|
+ * XXXXXX - Max procs capacity % available to the partitions pool.
|
|
|
|
+ * XXXXXX - Entitled procs capacity % available to the
|
|
|
|
+ * partitions pool.
|
|
*/
|
|
*/
|
|
static unsigned int h_get_ppp(struct hvcall_ppp_data *ppp_data)
|
|
static unsigned int h_get_ppp(struct hvcall_ppp_data *ppp_data)
|
|
{
|
|
{
|
|
unsigned long rc;
|
|
unsigned long rc;
|
|
- unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
|
|
+ unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
|
|
|
|
|
|
- rc = plpar_hcall(H_GET_PPP, retbuf);
|
|
|
|
|
|
+ rc = plpar_hcall9(H_GET_PPP, retbuf);
|
|
|
|
|
|
ppp_data->entitlement = retbuf[0];
|
|
ppp_data->entitlement = retbuf[0];
|
|
ppp_data->unallocated_entitlement = retbuf[1];
|
|
ppp_data->unallocated_entitlement = retbuf[1];
|
|
@@ -210,6 +218,10 @@ static unsigned int h_get_ppp(struct hvcall_ppp_data *ppp_data)
|
|
ppp_data->active_procs_in_pool = (retbuf[3] >> 2 * 8) & 0xffff;
|
|
ppp_data->active_procs_in_pool = (retbuf[3] >> 2 * 8) & 0xffff;
|
|
ppp_data->active_system_procs = retbuf[3] & 0xffff;
|
|
ppp_data->active_system_procs = retbuf[3] & 0xffff;
|
|
|
|
|
|
|
|
+ ppp_data->phys_platform_procs = retbuf[4] >> 6 * 8;
|
|
|
|
+ ppp_data->max_proc_cap_avail = (retbuf[4] >> 3 * 8) & 0xffffff;
|
|
|
|
+ ppp_data->entitled_proc_cap_avail = retbuf[4] & 0xffffff;
|
|
|
|
+
|
|
return rc;
|
|
return rc;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -234,6 +246,8 @@ static unsigned h_pic(unsigned long *pool_idle_time,
|
|
static void parse_ppp_data(struct seq_file *m)
|
|
static void parse_ppp_data(struct seq_file *m)
|
|
{
|
|
{
|
|
struct hvcall_ppp_data ppp_data;
|
|
struct hvcall_ppp_data ppp_data;
|
|
|
|
+ struct device_node *root;
|
|
|
|
+ const int *perf_level;
|
|
int rc;
|
|
int rc;
|
|
|
|
|
|
rc = h_get_ppp(&ppp_data);
|
|
rc = h_get_ppp(&ppp_data);
|
|
@@ -267,6 +281,28 @@ static void parse_ppp_data(struct seq_file *m)
|
|
seq_printf(m, "capped=%d\n", ppp_data.capped);
|
|
seq_printf(m, "capped=%d\n", ppp_data.capped);
|
|
seq_printf(m, "unallocated_capacity=%lld\n",
|
|
seq_printf(m, "unallocated_capacity=%lld\n",
|
|
ppp_data.unallocated_entitlement);
|
|
ppp_data.unallocated_entitlement);
|
|
|
|
+
|
|
|
|
+ /* The last bits of information returned from h_get_ppp are only
|
|
|
|
+ * valid if the ibm,partition-performance-parameters-level
|
|
|
|
+ * property is >= 1.
|
|
|
|
+ */
|
|
|
|
+ root = of_find_node_by_path("/");
|
|
|
|
+ if (root) {
|
|
|
|
+ perf_level = of_get_property(root,
|
|
|
|
+ "ibm,partition-performance-parameters-level",
|
|
|
|
+ NULL);
|
|
|
|
+ if (perf_level && (*perf_level >= 1)) {
|
|
|
|
+ seq_printf(m,
|
|
|
|
+ "physical_procs_allocated_to_virtualization=%d\n",
|
|
|
|
+ ppp_data.phys_platform_procs);
|
|
|
|
+ seq_printf(m, "max_proc_capacity_available=%d\n",
|
|
|
|
+ ppp_data.max_proc_cap_avail);
|
|
|
|
+ seq_printf(m, "entitled_proc_capacity_available=%d\n",
|
|
|
|
+ ppp_data.entitled_proc_cap_avail);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ of_node_put(root);
|
|
|
|
+ }
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
/**
|