[PATCH v9 2/5] powerpc/hv-24x7: Add rtas call in hv-24x7 driver to get processor details

Kajol Jain kjain at linux.ibm.com
Mon May 18 22:04:49 AEST 2020


For hv_24x7 socket/chip level events, specific chip-id to which
the data requested should be added as part of pmu events.
But number of chips/socket in the system details are not exposed.

Patch implements read_sys_info_pseries() to get system
parameter values like number of sockets, cores per chip and
chips per socket. Rtas_call with token "PROCESSOR_MODULE_INFO"
is used to get these values.

Sub-sequent patch exports these values via sysfs.

Patch also make these parameters default to 1.

Signed-off-by: Kajol Jain <kjain at linux.ibm.com>
---
 arch/powerpc/include/asm/rtas.h |  6 +++
 arch/powerpc/perf/hv-24x7.c     | 72 +++++++++++++++++++++++++++++++++
 2 files changed, 78 insertions(+)

diff --git a/arch/powerpc/include/asm/rtas.h b/arch/powerpc/include/asm/rtas.h
index 3c1887351c71..a5b302180041 100644
--- a/arch/powerpc/include/asm/rtas.h
+++ b/arch/powerpc/include/asm/rtas.h
@@ -483,5 +483,11 @@ static inline void rtas_initialize(void) { };
 
 extern int call_rtas(const char *, int, int, unsigned long *, ...);
 
+#ifdef CONFIG_HV_PERF_CTRS
+void read_sys_info_pseries(void);
+#else
+static inline void read_sys_info_pseries(void) { }
+#endif
+
 #endif /* __KERNEL__ */
 #endif /* _POWERPC_RTAS_H */
diff --git a/arch/powerpc/perf/hv-24x7.c b/arch/powerpc/perf/hv-24x7.c
index 48e8f4b17b91..b8625058c2aa 100644
--- a/arch/powerpc/perf/hv-24x7.c
+++ b/arch/powerpc/perf/hv-24x7.c
@@ -20,6 +20,7 @@
 #include <asm/io.h>
 #include <linux/byteorder/generic.h>
 
+#include <asm/rtas.h>
 #include "hv-24x7.h"
 #include "hv-24x7-catalog.h"
 #include "hv-common.h"
@@ -57,6 +58,75 @@ static bool is_physical_domain(unsigned domain)
 	}
 }
 
+/*
+ * The Processor Module Information system parameter allows transferring
+ * of certain processor module information from the platform to the OS.
+ * Refer PAPR+ document to get parameter token value as '43'.
+ */
+
+#define PROCESSOR_MODULE_INFO   43
+
+static u32 phys_sockets;	/* Physical sockets */
+static u32 phys_chipspersocket;	/* Physical chips per socket*/
+static u32 phys_coresperchip; /* Physical cores per chip */
+
+
+/*
+ * read_sys_info_pseries()
+ * Retrieve the number of sockets, chips per socket and cores per
+ * chip details through the get-system-parameter rtas call.
+ */
+void read_sys_info_pseries(void)
+{
+	int call_status, len, ntypes;
+	char *rtas_local_data_buf = kmalloc(RTAS_DATA_BUF_SIZE, GFP_KERNEL);
+
+	if (!rtas_local_data_buf) {
+		printk(KERN_ERR "%s %s kmalloc failure at line %d\n",
+		       __FILE__, __func__, __LINE__);
+		return;
+	}
+
+	/*
+	 * Making system parameter: chips per socket, sockets and cores per chip
+	 * default to 1.
+	 */
+	spin_lock(&rtas_data_buf_lock);
+	phys_sockets = 1;
+	phys_chipspersocket = 1;
+	phys_coresperchip = 1;
+	memset(rtas_data_buf, 0, RTAS_DATA_BUF_SIZE);
+
+	call_status = rtas_call(rtas_token("ibm,get-system-parameter"), 3, 1,
+				NULL,
+				PROCESSOR_MODULE_INFO,
+				__pa(rtas_data_buf),
+				RTAS_DATA_BUF_SIZE);
+
+	memcpy(rtas_local_data_buf, rtas_data_buf, RTAS_DATA_BUF_SIZE);
+	rtas_local_data_buf[RTAS_DATA_BUF_SIZE - 1] = '\0';
+
+	spin_unlock(&rtas_data_buf_lock);
+
+	if (call_status != 0)
+		pr_err("Error calling get-system-parameter %d\n", call_status);
+	else {
+		len = be16_to_cpup((__be16 *)&rtas_local_data_buf[0]);
+		if (len < 8)
+			return;
+
+		ntypes = be16_to_cpup((__be16 *)&rtas_local_data_buf[2]);
+
+		if (!ntypes)
+			return;
+		phys_sockets = be16_to_cpup((__be16 *)&rtas_local_data_buf[4]);
+		phys_chipspersocket = be16_to_cpup((__be16 *)&rtas_local_data_buf[6]);
+		phys_coresperchip = be16_to_cpup((__be16 *)&rtas_local_data_buf[8]);
+	}
+
+	kfree(rtas_local_data_buf);
+}
+
 /* Domains for which more than one result element are returned for each event. */
 static bool domain_needs_aggregation(unsigned int domain)
 {
@@ -1605,6 +1675,8 @@ static int hv_24x7_init(void)
 	if (r)
 		return r;
 
+	read_sys_info_pseries();
+
 	return 0;
 }
 
-- 
2.18.2



More information about the Linuxppc-dev mailing list