Browse Source

[IA64] run some drivers/misc/sgi-xp through scripts/Lindent

Ran patches through scripts/Lindent (part 1).

Signed-off-by: Dean Nelson <dcn@sgi.com>
Signed-off-by: Tony Luck <tony.luck@intel.com>
Dean Nelson 17 years ago
parent
commit
4a3ad2ddc0

+ 29 - 51
drivers/misc/sgi-xp/xp.h

@@ -6,30 +6,25 @@
  * Copyright (C) 2004-2008 Silicon Graphics, Inc. All rights reserved.
  * Copyright (C) 2004-2008 Silicon Graphics, Inc. All rights reserved.
  */
  */
 
 
-
 /*
 /*
  * External Cross Partition (XP) structures and defines.
  * External Cross Partition (XP) structures and defines.
  */
  */
 
 
-
 #ifndef _DRIVERS_MISC_SGIXP_XP_H
 #ifndef _DRIVERS_MISC_SGIXP_XP_H
 #define _DRIVERS_MISC_SGIXP_XP_H
 #define _DRIVERS_MISC_SGIXP_XP_H
 
 
-
 #include <linux/cache.h>
 #include <linux/cache.h>
 #include <linux/hardirq.h>
 #include <linux/hardirq.h>
 #include <linux/mutex.h>
 #include <linux/mutex.h>
 #include <asm/sn/types.h>
 #include <asm/sn/types.h>
 #include <asm/sn/bte.h>
 #include <asm/sn/bte.h>
 
 
-
 #ifdef USE_DBUG_ON
 #ifdef USE_DBUG_ON
 #define DBUG_ON(condition)	BUG_ON(condition)
 #define DBUG_ON(condition)	BUG_ON(condition)
 #else
 #else
 #define DBUG_ON(condition)
 #define DBUG_ON(condition)
 #endif
 #endif
 
 
-
 /*
 /*
  * Define the maximum number of logically defined partitions the system
  * Define the maximum number of logically defined partitions the system
  * can support. It is constrained by the maximum number of hardware
  * can support. It is constrained by the maximum number of hardware
@@ -43,7 +38,6 @@
  */
  */
 #define XP_MAX_PARTITIONS	64
 #define XP_MAX_PARTITIONS	64
 
 
-
 /*
 /*
  * Define the number of u64s required to represent all the C-brick nasids
  * Define the number of u64s required to represent all the C-brick nasids
  * as a bitmap.  The cross-partition kernel modules deal only with
  * as a bitmap.  The cross-partition kernel modules deal only with
@@ -54,7 +48,6 @@
 #define XP_NASID_MASK_BYTES	((XP_MAX_PHYSNODE_ID + 7) / 8)
 #define XP_NASID_MASK_BYTES	((XP_MAX_PHYSNODE_ID + 7) / 8)
 #define XP_NASID_MASK_WORDS	((XP_MAX_PHYSNODE_ID + 63) / 64)
 #define XP_NASID_MASK_WORDS	((XP_MAX_PHYSNODE_ID + 63) / 64)
 
 
-
 /*
 /*
  * Wrapper for bte_copy() that should it return a failure status will retry
  * Wrapper for bte_copy() that should it return a failure status will retry
  * the bte_copy() once in the hope that the failure was due to a temporary
  * the bte_copy() once in the hope that the failure was due to a temporary
@@ -74,7 +67,6 @@ xp_bte_copy(u64 src, u64 vdst, u64 len, u64 mode, void *notification)
 	bte_result_t ret;
 	bte_result_t ret;
 	u64 pdst = ia64_tpa(vdst);
 	u64 pdst = ia64_tpa(vdst);
 
 
-
 	/*
 	/*
 	 * Ensure that the physically mapped memory is contiguous.
 	 * Ensure that the physically mapped memory is contiguous.
 	 *
 	 *
@@ -96,7 +88,6 @@ xp_bte_copy(u64 src, u64 vdst, u64 len, u64 mode, void *notification)
 	return ret;
 	return ret;
 }
 }
 
 
-
 /*
 /*
  * XPC establishes channel connections between the local partition and any
  * XPC establishes channel connections between the local partition and any
  * other partition that is currently up. Over these channels, kernel-level
  * other partition that is currently up. Over these channels, kernel-level
@@ -122,7 +113,6 @@ xp_bte_copy(u64 src, u64 vdst, u64 len, u64 mode, void *notification)
 #error	XPC_NCHANNELS exceeds MAXIMUM allowed.
 #error	XPC_NCHANNELS exceeds MAXIMUM allowed.
 #endif
 #endif
 
 
-
 /*
 /*
  * The format of an XPC message is as follows:
  * The format of an XPC message is as follows:
  *
  *
@@ -160,12 +150,10 @@ struct xpc_msg {
 	u64 payload;		/* user defined portion of message */
 	u64 payload;		/* user defined portion of message */
 };
 };
 
 
-
 #define XPC_MSG_PAYLOAD_OFFSET	(u64) (&((struct xpc_msg *)0)->payload)
 #define XPC_MSG_PAYLOAD_OFFSET	(u64) (&((struct xpc_msg *)0)->payload)
 #define XPC_MSG_SIZE(_payload_size) \
 #define XPC_MSG_SIZE(_payload_size) \
 		L1_CACHE_ALIGN(XPC_MSG_PAYLOAD_OFFSET + (_payload_size))
 		L1_CACHE_ALIGN(XPC_MSG_PAYLOAD_OFFSET + (_payload_size))
 
 
-
 /*
 /*
  * Define the return values and values passed to user's callout functions.
  * Define the return values and values passed to user's callout functions.
  * (It is important to add new value codes at the end just preceding
  * (It is important to add new value codes at the end just preceding
@@ -270,7 +258,6 @@ enum xpc_retval {
 	xpcUnknownReason	/* 116: unknown reason -- must be last in list */
 	xpcUnknownReason	/* 116: unknown reason -- must be last in list */
 };
 };
 
 
-
 /*
 /*
  * Define the callout function types used by XPC to update the user on
  * Define the callout function types used by XPC to update the user on
  * connection activity and state changes (via the user function registered by
  * connection activity and state changes (via the user function registered by
@@ -375,12 +362,11 @@ enum xpc_retval {
  * =====================+================================+=====================
  * =====================+================================+=====================
  */
  */
 
 
-typedef void (*xpc_channel_func)(enum xpc_retval reason, partid_t partid,
-		int ch_number, void *data, void *key);
-
-typedef void (*xpc_notify_func)(enum xpc_retval reason, partid_t partid,
-		int ch_number, void *key);
+typedef void (*xpc_channel_func) (enum xpc_retval reason, partid_t partid,
+				  int ch_number, void *data, void *key);
 
 
+typedef void (*xpc_notify_func) (enum xpc_retval reason, partid_t partid,
+				 int ch_number, void *key);
 
 
 /*
 /*
  * The following is a registration entry. There is a global array of these,
  * The following is a registration entry. There is a global array of these,
@@ -398,50 +384,45 @@ typedef void (*xpc_notify_func)(enum xpc_retval reason, partid_t partid,
  */
  */
 struct xpc_registration {
 struct xpc_registration {
 	struct mutex mutex;
 	struct mutex mutex;
-	xpc_channel_func func;		/* function to call */
-	void *key;			/* pointer to user's key */
-	u16 nentries;			/* #of msg entries in local msg queue */
-	u16 msg_size;			/* message queue's message size */
-	u32 assigned_limit;		/* limit on #of assigned kthreads */
-	u32 idle_limit;			/* limit on #of idle kthreads */
+	xpc_channel_func func;	/* function to call */
+	void *key;		/* pointer to user's key */
+	u16 nentries;		/* #of msg entries in local msg queue */
+	u16 msg_size;		/* message queue's message size */
+	u32 assigned_limit;	/* limit on #of assigned kthreads */
+	u32 idle_limit;		/* limit on #of idle kthreads */
 } ____cacheline_aligned;
 } ____cacheline_aligned;
 
 
-
 #define XPC_CHANNEL_REGISTERED(_c)	(xpc_registrations[_c].func != NULL)
 #define XPC_CHANNEL_REGISTERED(_c)	(xpc_registrations[_c].func != NULL)
 
 
-
 /* the following are valid xpc_allocate() flags */
 /* the following are valid xpc_allocate() flags */
-#define XPC_WAIT	0		/* wait flag */
-#define XPC_NOWAIT	1		/* no wait flag */
-
+#define XPC_WAIT	0	/* wait flag */
+#define XPC_NOWAIT	1	/* no wait flag */
 
 
 struct xpc_interface {
 struct xpc_interface {
-	void (*connect)(int);
-	void (*disconnect)(int);
-	enum xpc_retval (*allocate)(partid_t, int, u32, void **);
-	enum xpc_retval (*send)(partid_t, int, void *);
-	enum xpc_retval (*send_notify)(partid_t, int, void *,
-						xpc_notify_func, void *);
-	void (*received)(partid_t, int, void *);
-	enum xpc_retval (*partid_to_nasids)(partid_t, void *);
+	void (*connect) (int);
+	void (*disconnect) (int);
+	enum xpc_retval (*allocate) (partid_t, int, u32, void **);
+	enum xpc_retval (*send) (partid_t, int, void *);
+	enum xpc_retval (*send_notify) (partid_t, int, void *,
+					xpc_notify_func, void *);
+	void (*received) (partid_t, int, void *);
+	enum xpc_retval (*partid_to_nasids) (partid_t, void *);
 };
 };
 
 
-
 extern struct xpc_interface xpc_interface;
 extern struct xpc_interface xpc_interface;
 
 
 extern void xpc_set_interface(void (*)(int),
 extern void xpc_set_interface(void (*)(int),
-		void (*)(int),
-		enum xpc_retval (*)(partid_t, int, u32, void **),
-		enum xpc_retval (*)(partid_t, int, void *),
-		enum xpc_retval (*)(partid_t, int, void *, xpc_notify_func,
-								void *),
-		void (*)(partid_t, int, void *),
-		enum xpc_retval (*)(partid_t, void *));
+			      void (*)(int),
+			      enum xpc_retval (*)(partid_t, int, u32, void **),
+			      enum xpc_retval (*)(partid_t, int, void *),
+			      enum xpc_retval (*)(partid_t, int, void *,
+						  xpc_notify_func, void *),
+			      void (*)(partid_t, int, void *),
+			      enum xpc_retval (*)(partid_t, void *));
 extern void xpc_clear_interface(void);
 extern void xpc_clear_interface(void);
 
 
-
 extern enum xpc_retval xpc_connect(int, xpc_channel_func, void *, u16,
 extern enum xpc_retval xpc_connect(int, xpc_channel_func, void *, u16,
-						u16, u32, u32);
+				   u16, u32, u32);
 extern void xpc_disconnect(int);
 extern void xpc_disconnect(int);
 
 
 static inline enum xpc_retval
 static inline enum xpc_retval
@@ -458,7 +439,7 @@ xpc_send(partid_t partid, int ch_number, void *payload)
 
 
 static inline enum xpc_retval
 static inline enum xpc_retval
 xpc_send_notify(partid_t partid, int ch_number, void *payload,
 xpc_send_notify(partid_t partid, int ch_number, void *payload,
-			xpc_notify_func func, void *key)
+		xpc_notify_func func, void *key)
 {
 {
 	return xpc_interface.send_notify(partid, ch_number, payload, func, key);
 	return xpc_interface.send_notify(partid, ch_number, payload, func, key);
 }
 }
@@ -475,11 +456,8 @@ xpc_partid_to_nasids(partid_t partid, void *nasids)
 	return xpc_interface.partid_to_nasids(partid, nasids);
 	return xpc_interface.partid_to_nasids(partid, nasids);
 }
 }
 
 
-
 extern u64 xp_nofault_PIOR_target;
 extern u64 xp_nofault_PIOR_target;
 extern int xp_nofault_PIOR(void *);
 extern int xp_nofault_PIOR(void *);
 extern int xp_error_PIOR(void);
 extern int xp_error_PIOR(void);
 
 
-
 #endif /* _DRIVERS_MISC_SGIXP_XP_H */
 #endif /* _DRIVERS_MISC_SGIXP_XP_H */
-

+ 44 - 54
drivers/misc/sgi-xp/xp_main.c

@@ -6,7 +6,6 @@
  * Copyright (c) 2004-2008 Silicon Graphics, Inc.  All Rights Reserved.
  * Copyright (c) 2004-2008 Silicon Graphics, Inc.  All Rights Reserved.
  */
  */
 
 
-
 /*
 /*
  * Cross Partition (XP) base.
  * Cross Partition (XP) base.
  *
  *
@@ -15,7 +14,6 @@
  *
  *
  */
  */
 
 
-
 #include <linux/kernel.h>
 #include <linux/kernel.h>
 #include <linux/interrupt.h>
 #include <linux/interrupt.h>
 #include <linux/module.h>
 #include <linux/module.h>
@@ -24,49 +22,49 @@
 #include <asm/sn/sn_sal.h>
 #include <asm/sn/sn_sal.h>
 #include "xp.h"
 #include "xp.h"
 
 
-
 /*
 /*
  * Target of nofault PIO read.
  * Target of nofault PIO read.
  */
  */
 u64 xp_nofault_PIOR_target;
 u64 xp_nofault_PIOR_target;
 
 
-
 /*
 /*
  * xpc_registrations[] keeps track of xpc_connect()'s done by the kernel-level
  * xpc_registrations[] keeps track of xpc_connect()'s done by the kernel-level
  * users of XPC.
  * users of XPC.
  */
  */
 struct xpc_registration xpc_registrations[XPC_NCHANNELS];
 struct xpc_registration xpc_registrations[XPC_NCHANNELS];
 
 
-
 /*
 /*
  * Initialize the XPC interface to indicate that XPC isn't loaded.
  * Initialize the XPC interface to indicate that XPC isn't loaded.
  */
  */
-static enum xpc_retval xpc_notloaded(void) { return xpcNotLoaded; }
+static enum xpc_retval
+xpc_notloaded(void)
+{
+	return xpcNotLoaded;
+}
 
 
 struct xpc_interface xpc_interface = {
 struct xpc_interface xpc_interface = {
-	(void (*)(int)) xpc_notloaded,
-	(void (*)(int)) xpc_notloaded,
-	(enum xpc_retval (*)(partid_t, int, u32, void **)) xpc_notloaded,
-	(enum xpc_retval (*)(partid_t, int, void *)) xpc_notloaded,
-	(enum xpc_retval (*)(partid_t, int, void *, xpc_notify_func, void *))
-							xpc_notloaded,
-	(void (*)(partid_t, int, void *)) xpc_notloaded,
-	(enum xpc_retval (*)(partid_t, void *)) xpc_notloaded
+	(void (*)(int))xpc_notloaded,
+	(void (*)(int))xpc_notloaded,
+	(enum xpc_retval(*)(partid_t, int, u32, void **))xpc_notloaded,
+	(enum xpc_retval(*)(partid_t, int, void *))xpc_notloaded,
+	(enum xpc_retval(*)(partid_t, int, void *, xpc_notify_func, void *))
+	    xpc_notloaded,
+	(void (*)(partid_t, int, void *))xpc_notloaded,
+	(enum xpc_retval(*)(partid_t, void *))xpc_notloaded
 };
 };
 
 
-
 /*
 /*
  * XPC calls this when it (the XPC module) has been loaded.
  * XPC calls this when it (the XPC module) has been loaded.
  */
  */
 void
 void
-xpc_set_interface(void (*connect)(int),
-		void (*disconnect)(int),
-		enum xpc_retval (*allocate)(partid_t, int, u32, void **),
-		enum xpc_retval (*send)(partid_t, int, void *),
-		enum xpc_retval (*send_notify)(partid_t, int, void *,
-						xpc_notify_func, void *),
-		void (*received)(partid_t, int, void *),
-		enum xpc_retval (*partid_to_nasids)(partid_t, void *))
+xpc_set_interface(void (*connect) (int),
+		  void (*disconnect) (int),
+		  enum xpc_retval (*allocate) (partid_t, int, u32, void **),
+		  enum xpc_retval (*send) (partid_t, int, void *),
+		  enum xpc_retval (*send_notify) (partid_t, int, void *,
+						  xpc_notify_func, void *),
+		  void (*received) (partid_t, int, void *),
+		  enum xpc_retval (*partid_to_nasids) (partid_t, void *))
 {
 {
 	xpc_interface.connect = connect;
 	xpc_interface.connect = connect;
 	xpc_interface.disconnect = disconnect;
 	xpc_interface.disconnect = disconnect;
@@ -77,28 +75,27 @@ xpc_set_interface(void (*connect)(int),
 	xpc_interface.partid_to_nasids = partid_to_nasids;
 	xpc_interface.partid_to_nasids = partid_to_nasids;
 }
 }
 
 
-
 /*
 /*
  * XPC calls this when it (the XPC module) is being unloaded.
  * XPC calls this when it (the XPC module) is being unloaded.
  */
  */
 void
 void
 xpc_clear_interface(void)
 xpc_clear_interface(void)
 {
 {
-	xpc_interface.connect = (void (*)(int)) xpc_notloaded;
-	xpc_interface.disconnect = (void (*)(int)) xpc_notloaded;
-	xpc_interface.allocate = (enum xpc_retval (*)(partid_t, int, u32,
-					void **)) xpc_notloaded;
-	xpc_interface.send = (enum xpc_retval (*)(partid_t, int, void *))
-					xpc_notloaded;
-	xpc_interface.send_notify = (enum xpc_retval (*)(partid_t, int, void *,
-				    xpc_notify_func, void *)) xpc_notloaded;
+	xpc_interface.connect = (void (*)(int))xpc_notloaded;
+	xpc_interface.disconnect = (void (*)(int))xpc_notloaded;
+	xpc_interface.allocate = (enum xpc_retval(*)(partid_t, int, u32,
+						     void **))xpc_notloaded;
+	xpc_interface.send = (enum xpc_retval(*)(partid_t, int, void *))
+	    xpc_notloaded;
+	xpc_interface.send_notify = (enum xpc_retval(*)(partid_t, int, void *,
+							xpc_notify_func,
+							void *))xpc_notloaded;
 	xpc_interface.received = (void (*)(partid_t, int, void *))
 	xpc_interface.received = (void (*)(partid_t, int, void *))
-					xpc_notloaded;
-	xpc_interface.partid_to_nasids = (enum xpc_retval (*)(partid_t, void *))
-					xpc_notloaded;
+	    xpc_notloaded;
+	xpc_interface.partid_to_nasids = (enum xpc_retval(*)(partid_t, void *))
+	    xpc_notloaded;
 }
 }
 
 
-
 /*
 /*
  * Register for automatic establishment of a channel connection whenever
  * Register for automatic establishment of a channel connection whenever
  * a partition comes up.
  * a partition comes up.
@@ -125,11 +122,10 @@ xpc_clear_interface(void)
  */
  */
 enum xpc_retval
 enum xpc_retval
 xpc_connect(int ch_number, xpc_channel_func func, void *key, u16 payload_size,
 xpc_connect(int ch_number, xpc_channel_func func, void *key, u16 payload_size,
-		u16 nentries, u32 assigned_limit, u32 idle_limit)
+	    u16 nentries, u32 assigned_limit, u32 idle_limit)
 {
 {
 	struct xpc_registration *registration;
 	struct xpc_registration *registration;
 
 
-
 	DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS);
 	DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS);
 	DBUG_ON(payload_size == 0 || nentries == 0);
 	DBUG_ON(payload_size == 0 || nentries == 0);
 	DBUG_ON(func == NULL);
 	DBUG_ON(func == NULL);
@@ -162,7 +158,6 @@ xpc_connect(int ch_number, xpc_channel_func func, void *key, u16 payload_size,
 	return xpcSuccess;
 	return xpcSuccess;
 }
 }
 
 
-
 /*
 /*
  * Remove the registration for automatic connection of the specified channel
  * Remove the registration for automatic connection of the specified channel
  * when a partition comes up.
  * when a partition comes up.
@@ -181,7 +176,6 @@ xpc_disconnect(int ch_number)
 {
 {
 	struct xpc_registration *registration;
 	struct xpc_registration *registration;
 
 
-
 	DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS);
 	DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS);
 
 
 	registration = &xpc_registrations[ch_number];
 	registration = &xpc_registrations[ch_number];
@@ -214,14 +208,12 @@ xpc_disconnect(int ch_number)
 	return;
 	return;
 }
 }
 
 
-
 int __init
 int __init
 xp_init(void)
 xp_init(void)
 {
 {
 	int ret, ch_number;
 	int ret, ch_number;
-	u64 func_addr = *(u64 *) xp_nofault_PIOR;
-	u64 err_func_addr = *(u64 *) xp_error_PIOR;
-
+	u64 func_addr = *(u64 *)xp_nofault_PIOR;
+	u64 err_func_addr = *(u64 *)xp_error_PIOR;
 
 
 	if (!ia64_platform_is("sn2")) {
 	if (!ia64_platform_is("sn2")) {
 		return -ENODEV;
 		return -ENODEV;
@@ -237,9 +229,9 @@ xp_init(void)
 	 * work around).
 	 * work around).
 	 */
 	 */
 	if ((ret = sn_register_nofault_code(func_addr, err_func_addr,
 	if ((ret = sn_register_nofault_code(func_addr, err_func_addr,
-						err_func_addr, 1, 1)) != 0) {
+					    err_func_addr, 1, 1)) != 0) {
 		printk(KERN_ERR "XP: can't register nofault code, error=%d\n",
 		printk(KERN_ERR "XP: can't register nofault code, error=%d\n",
-			ret);
+		       ret);
 	}
 	}
 	/*
 	/*
 	 * Setup the nofault PIO read target. (There is no special reason why
 	 * Setup the nofault PIO read target. (There is no special reason why
@@ -258,22 +250,21 @@ xp_init(void)
 
 
 	return 0;
 	return 0;
 }
 }
-module_init(xp_init);
 
 
+module_init(xp_init);
 
 
 void __exit
 void __exit
 xp_exit(void)
 xp_exit(void)
 {
 {
-	u64 func_addr = *(u64 *) xp_nofault_PIOR;
-	u64 err_func_addr = *(u64 *) xp_error_PIOR;
-
+	u64 func_addr = *(u64 *)xp_nofault_PIOR;
+	u64 err_func_addr = *(u64 *)xp_error_PIOR;
 
 
 	/* unregister the PIO read nofault code region */
 	/* unregister the PIO read nofault code region */
-	(void) sn_register_nofault_code(func_addr, err_func_addr,
-					err_func_addr, 1, 0);
+	(void)sn_register_nofault_code(func_addr, err_func_addr,
+				       err_func_addr, 1, 0);
 }
 }
-module_exit(xp_exit);
 
 
+module_exit(xp_exit);
 
 
 MODULE_AUTHOR("Silicon Graphics, Inc.");
 MODULE_AUTHOR("Silicon Graphics, Inc.");
 MODULE_DESCRIPTION("Cross Partition (XP) base");
 MODULE_DESCRIPTION("Cross Partition (XP) base");
@@ -287,4 +278,3 @@ EXPORT_SYMBOL(xpc_clear_interface);
 EXPORT_SYMBOL(xpc_set_interface);
 EXPORT_SYMBOL(xpc_set_interface);
 EXPORT_SYMBOL(xpc_connect);
 EXPORT_SYMBOL(xpc_connect);
 EXPORT_SYMBOL(xpc_disconnect);
 EXPORT_SYMBOL(xpc_disconnect);
-

+ 0 - 1
drivers/misc/sgi-xp/xp_nofault.S

@@ -6,7 +6,6 @@
  * Copyright (c) 2004-2008 Silicon Graphics, Inc.  All Rights Reserved.
  * Copyright (c) 2004-2008 Silicon Graphics, Inc.  All Rights Reserved.
  */
  */
 
 
-
 /*
 /*
  * The xp_nofault_PIOR function takes a pointer to a remote PIO register
  * The xp_nofault_PIOR function takes a pointer to a remote PIO register
  * and attempts to load and consume a value from it.  This function
  * and attempts to load and consume a value from it.  This function

File diff suppressed because it is too large
+ 163 - 241
drivers/misc/sgi-xp/xpc.h


+ 138 - 190
drivers/misc/sgi-xp/xpc_partition.c

@@ -6,7 +6,6 @@
  * Copyright (c) 2004-2008 Silicon Graphics, Inc.  All Rights Reserved.
  * Copyright (c) 2004-2008 Silicon Graphics, Inc.  All Rights Reserved.
  */
  */
 
 
-
 /*
 /*
  * Cross Partition Communication (XPC) partition support.
  * Cross Partition Communication (XPC) partition support.
  *
  *
@@ -16,7 +15,6 @@
  *
  *
  */
  */
 
 
-
 #include <linux/kernel.h>
 #include <linux/kernel.h>
 #include <linux/sysctl.h>
 #include <linux/sysctl.h>
 #include <linux/cache.h>
 #include <linux/cache.h>
@@ -30,11 +28,9 @@
 #include <asm/sn/addrs.h>
 #include <asm/sn/addrs.h>
 #include "xpc.h"
 #include "xpc.h"
 
 
-
 /* XPC is exiting flag */
 /* XPC is exiting flag */
 int xpc_exiting;
 int xpc_exiting;
 
 
-
 /* SH_IPI_ACCESS shub register value on startup */
 /* SH_IPI_ACCESS shub register value on startup */
 static u64 xpc_sh1_IPI_access;
 static u64 xpc_sh1_IPI_access;
 static u64 xpc_sh2_IPI_access0;
 static u64 xpc_sh2_IPI_access0;
@@ -42,11 +38,9 @@ static u64 xpc_sh2_IPI_access1;
 static u64 xpc_sh2_IPI_access2;
 static u64 xpc_sh2_IPI_access2;
 static u64 xpc_sh2_IPI_access3;
 static u64 xpc_sh2_IPI_access3;
 
 
-
 /* original protection values for each node */
 /* original protection values for each node */
 u64 xpc_prot_vec[MAX_NUMNODES];
 u64 xpc_prot_vec[MAX_NUMNODES];
 
 
-
 /* this partition's reserved page pointers */
 /* this partition's reserved page pointers */
 struct xpc_rsvd_page *xpc_rsvd_page;
 struct xpc_rsvd_page *xpc_rsvd_page;
 static u64 *xpc_part_nasids;
 static u64 *xpc_part_nasids;
@@ -57,7 +51,6 @@ struct xpc_vars_part *xpc_vars_part;
 static int xp_nasid_mask_bytes;	/* actual size in bytes of nasid mask */
 static int xp_nasid_mask_bytes;	/* actual size in bytes of nasid mask */
 static int xp_nasid_mask_words;	/* actual size in words of nasid mask */
 static int xp_nasid_mask_words;	/* actual size in words of nasid mask */
 
 
-
 /*
 /*
  * For performance reasons, each entry of xpc_partitions[] is cacheline
  * For performance reasons, each entry of xpc_partitions[] is cacheline
  * aligned. And xpc_partitions[] is padded with an additional entry at the
  * aligned. And xpc_partitions[] is padded with an additional entry at the
@@ -66,7 +59,6 @@ static int xp_nasid_mask_words;	/* actual size in words of nasid mask */
  */
  */
 struct xpc_partition xpc_partitions[XP_MAX_PARTITIONS + 1];
 struct xpc_partition xpc_partitions[XP_MAX_PARTITIONS + 1];
 
 
-
 /*
 /*
  * Generic buffer used to store a local copy of portions of a remote
  * Generic buffer used to store a local copy of portions of a remote
  * partition's reserved page (either its header and part_nasids mask,
  * partition's reserved page (either its header and part_nasids mask,
@@ -75,7 +67,6 @@ struct xpc_partition xpc_partitions[XP_MAX_PARTITIONS + 1];
 char *xpc_remote_copy_buffer;
 char *xpc_remote_copy_buffer;
 void *xpc_remote_copy_buffer_base;
 void *xpc_remote_copy_buffer_base;
 
 
-
 /*
 /*
  * Guarantee that the kmalloc'd memory is cacheline aligned.
  * Guarantee that the kmalloc'd memory is cacheline aligned.
  */
  */
@@ -87,7 +78,7 @@ xpc_kmalloc_cacheline_aligned(size_t size, gfp_t flags, void **base)
 	if (*base == NULL) {
 	if (*base == NULL) {
 		return NULL;
 		return NULL;
 	}
 	}
-	if ((u64) *base == L1_CACHE_ALIGN((u64) *base)) {
+	if ((u64)*base == L1_CACHE_ALIGN((u64)*base)) {
 		return *base;
 		return *base;
 	}
 	}
 	kfree(*base);
 	kfree(*base);
@@ -97,10 +88,9 @@ xpc_kmalloc_cacheline_aligned(size_t size, gfp_t flags, void **base)
 	if (*base == NULL) {
 	if (*base == NULL) {
 		return NULL;
 		return NULL;
 	}
 	}
-	return (void *) L1_CACHE_ALIGN((u64) *base);
+	return (void *)L1_CACHE_ALIGN((u64)*base);
 }
 }
 
 
-
 /*
 /*
  * Given a nasid, get the physical address of the  partition's reserved page
  * Given a nasid, get the physical address of the  partition's reserved page
  * for that nasid. This function returns 0 on any error.
  * for that nasid. This function returns 0 on any error.
@@ -117,11 +107,10 @@ xpc_get_rsvd_page_pa(int nasid)
 	u64 buf_len = 0;
 	u64 buf_len = 0;
 	void *buf_base = NULL;
 	void *buf_base = NULL;
 
 
-
 	while (1) {
 	while (1) {
 
 
 		status = sn_partition_reserved_page_pa(buf, &cookie, &rp_pa,
 		status = sn_partition_reserved_page_pa(buf, &cookie, &rp_pa,
-								&len);
+						       &len);
 
 
 		dev_dbg(xpc_part, "SAL returned with status=%li, cookie="
 		dev_dbg(xpc_part, "SAL returned with status=%li, cookie="
 			"0x%016lx, address=0x%016lx, len=0x%016lx\n",
 			"0x%016lx, address=0x%016lx, len=0x%016lx\n",
@@ -134,8 +123,9 @@ xpc_get_rsvd_page_pa(int nasid)
 		if (L1_CACHE_ALIGN(len) > buf_len) {
 		if (L1_CACHE_ALIGN(len) > buf_len) {
 			kfree(buf_base);
 			kfree(buf_base);
 			buf_len = L1_CACHE_ALIGN(len);
 			buf_len = L1_CACHE_ALIGN(len);
-			buf = (u64) xpc_kmalloc_cacheline_aligned(buf_len,
-							GFP_KERNEL, &buf_base);
+			buf = (u64)xpc_kmalloc_cacheline_aligned(buf_len,
+								 GFP_KERNEL,
+								 &buf_base);
 			if (buf_base == NULL) {
 			if (buf_base == NULL) {
 				dev_err(xpc_part, "unable to kmalloc "
 				dev_err(xpc_part, "unable to kmalloc "
 					"len=0x%016lx\n", buf_len);
 					"len=0x%016lx\n", buf_len);
@@ -145,7 +135,7 @@ xpc_get_rsvd_page_pa(int nasid)
 		}
 		}
 
 
 		bte_res = xp_bte_copy(rp_pa, buf, buf_len,
 		bte_res = xp_bte_copy(rp_pa, buf, buf_len,
-					(BTE_NOTIFY | BTE_WACQUIRE), NULL);
+				      (BTE_NOTIFY | BTE_WACQUIRE), NULL);
 		if (bte_res != BTE_SUCCESS) {
 		if (bte_res != BTE_SUCCESS) {
 			dev_dbg(xpc_part, "xp_bte_copy failed %i\n", bte_res);
 			dev_dbg(xpc_part, "xp_bte_copy failed %i\n", bte_res);
 			status = SALRET_ERROR;
 			status = SALRET_ERROR;
@@ -162,7 +152,6 @@ xpc_get_rsvd_page_pa(int nasid)
 	return rp_pa;
 	return rp_pa;
 }
 }
 
 
-
 /*
 /*
  * Fill the partition reserved page with the information needed by
  * Fill the partition reserved page with the information needed by
  * other partitions to discover we are alive and establish initial
  * other partitions to discover we are alive and establish initial
@@ -176,7 +165,6 @@ xpc_rsvd_page_init(void)
 	u64 rp_pa, nasid_array = 0;
 	u64 rp_pa, nasid_array = 0;
 	int i, ret;
 	int i, ret;
 
 
-
 	/* get the local reserved page's address */
 	/* get the local reserved page's address */
 
 
 	preempt_disable();
 	preempt_disable();
@@ -186,7 +174,7 @@ xpc_rsvd_page_init(void)
 		dev_err(xpc_part, "SAL failed to locate the reserved page\n");
 		dev_err(xpc_part, "SAL failed to locate the reserved page\n");
 		return NULL;
 		return NULL;
 	}
 	}
-	rp = (struct xpc_rsvd_page *) __va(rp_pa);
+	rp = (struct xpc_rsvd_page *)__va(rp_pa);
 
 
 	if (rp->partid != sn_partition_id) {
 	if (rp->partid != sn_partition_id) {
 		dev_err(xpc_part, "the reserved page's partid of %d should be "
 		dev_err(xpc_part, "the reserved page's partid of %d should be "
@@ -223,7 +211,7 @@ xpc_rsvd_page_init(void)
 	 * memory protections are never restricted.
 	 * memory protections are never restricted.
 	 */
 	 */
 	if ((amos_page = xpc_vars->amos_page) == NULL) {
 	if ((amos_page = xpc_vars->amos_page) == NULL) {
-		amos_page = (AMO_t *) TO_AMO(uncached_alloc_page(0));
+		amos_page = (AMO_t *)TO_AMO(uncached_alloc_page(0));
 		if (amos_page == NULL) {
 		if (amos_page == NULL) {
 			dev_err(xpc_part, "can't allocate page of AMOs\n");
 			dev_err(xpc_part, "can't allocate page of AMOs\n");
 			return NULL;
 			return NULL;
@@ -234,30 +222,31 @@ xpc_rsvd_page_init(void)
 		 * when xpc_allow_IPI_ops() is called via xpc_hb_init().
 		 * when xpc_allow_IPI_ops() is called via xpc_hb_init().
 		 */
 		 */
 		if (!enable_shub_wars_1_1()) {
 		if (!enable_shub_wars_1_1()) {
-			ret = sn_change_memprotect(ia64_tpa((u64) amos_page),
-					PAGE_SIZE, SN_MEMPROT_ACCESS_CLASS_1,
-					&nasid_array);
+			ret = sn_change_memprotect(ia64_tpa((u64)amos_page),
+						   PAGE_SIZE,
+						   SN_MEMPROT_ACCESS_CLASS_1,
+						   &nasid_array);
 			if (ret != 0) {
 			if (ret != 0) {
 				dev_err(xpc_part, "can't change memory "
 				dev_err(xpc_part, "can't change memory "
 					"protections\n");
 					"protections\n");
 				uncached_free_page(__IA64_UNCACHED_OFFSET |
 				uncached_free_page(__IA64_UNCACHED_OFFSET |
-						   TO_PHYS((u64) amos_page));
+						   TO_PHYS((u64)amos_page));
 				return NULL;
 				return NULL;
 			}
 			}
 		}
 		}
-	} else if (!IS_AMO_ADDRESS((u64) amos_page)) {
+	} else if (!IS_AMO_ADDRESS((u64)amos_page)) {
 		/*
 		/*
 		 * EFI's XPBOOT can also set amos_page in the reserved page,
 		 * EFI's XPBOOT can also set amos_page in the reserved page,
 		 * but it happens to leave it as an uncached physical address
 		 * but it happens to leave it as an uncached physical address
 		 * and we need it to be an uncached virtual, so we'll have to
 		 * and we need it to be an uncached virtual, so we'll have to
 		 * convert it.
 		 * convert it.
 		 */
 		 */
-		if (!IS_AMO_PHYS_ADDRESS((u64) amos_page)) {
+		if (!IS_AMO_PHYS_ADDRESS((u64)amos_page)) {
 			dev_err(xpc_part, "previously used amos_page address "
 			dev_err(xpc_part, "previously used amos_page address "
-				"is bad = 0x%p\n", (void *) amos_page);
+				"is bad = 0x%p\n", (void *)amos_page);
 			return NULL;
 			return NULL;
 		}
 		}
-		amos_page = (AMO_t *) TO_AMO((u64) amos_page);
+		amos_page = (AMO_t *)TO_AMO((u64)amos_page);
 	}
 	}
 
 
 	/* clear xpc_vars */
 	/* clear xpc_vars */
@@ -267,22 +256,21 @@ xpc_rsvd_page_init(void)
 	xpc_vars->act_nasid = cpuid_to_nasid(0);
 	xpc_vars->act_nasid = cpuid_to_nasid(0);
 	xpc_vars->act_phys_cpuid = cpu_physical_id(0);
 	xpc_vars->act_phys_cpuid = cpu_physical_id(0);
 	xpc_vars->vars_part_pa = __pa(xpc_vars_part);
 	xpc_vars->vars_part_pa = __pa(xpc_vars_part);
-	xpc_vars->amos_page_pa = ia64_tpa((u64) amos_page);
-	xpc_vars->amos_page = amos_page;  /* save for next load of XPC */
-
+	xpc_vars->amos_page_pa = ia64_tpa((u64)amos_page);
+	xpc_vars->amos_page = amos_page;	/* save for next load of XPC */
 
 
 	/* clear xpc_vars_part */
 	/* clear xpc_vars_part */
-	memset((u64 *) xpc_vars_part, 0, sizeof(struct xpc_vars_part) *
-							XP_MAX_PARTITIONS);
+	memset((u64 *)xpc_vars_part, 0, sizeof(struct xpc_vars_part) *
+	       XP_MAX_PARTITIONS);
 
 
 	/* initialize the activate IRQ related AMO variables */
 	/* initialize the activate IRQ related AMO variables */
 	for (i = 0; i < xp_nasid_mask_words; i++) {
 	for (i = 0; i < xp_nasid_mask_words; i++) {
-		(void) xpc_IPI_init(XPC_ACTIVATE_IRQ_AMOS + i);
+		(void)xpc_IPI_init(XPC_ACTIVATE_IRQ_AMOS + i);
 	}
 	}
 
 
 	/* initialize the engaged remote partitions related AMO variables */
 	/* initialize the engaged remote partitions related AMO variables */
-	(void) xpc_IPI_init(XPC_ENGAGED_PARTITIONS_AMO);
-	(void) xpc_IPI_init(XPC_DISENGAGE_REQUEST_AMO);
+	(void)xpc_IPI_init(XPC_ENGAGED_PARTITIONS_AMO);
+	(void)xpc_IPI_init(XPC_DISENGAGE_REQUEST_AMO);
 
 
 	/* timestamp of when reserved page was setup by XPC */
 	/* timestamp of when reserved page was setup by XPC */
 	rp->stamp = CURRENT_TIME;
 	rp->stamp = CURRENT_TIME;
@@ -296,7 +284,6 @@ xpc_rsvd_page_init(void)
 	return rp;
 	return rp;
 }
 }
 
 
-
 /*
 /*
  * Change protections to allow IPI operations (and AMO operations on
  * Change protections to allow IPI operations (and AMO operations on
  * Shub 1.1 systems).
  * Shub 1.1 systems).
@@ -307,39 +294,38 @@ xpc_allow_IPI_ops(void)
 	int node;
 	int node;
 	int nasid;
 	int nasid;
 
 
-
 	// >>> Change SH_IPI_ACCESS code to use SAL call once it is available.
 	// >>> Change SH_IPI_ACCESS code to use SAL call once it is available.
 
 
 	if (is_shub2()) {
 	if (is_shub2()) {
 		xpc_sh2_IPI_access0 =
 		xpc_sh2_IPI_access0 =
-			(u64) HUB_L((u64 *) LOCAL_MMR_ADDR(SH2_IPI_ACCESS0));
+		    (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS0));
 		xpc_sh2_IPI_access1 =
 		xpc_sh2_IPI_access1 =
-			(u64) HUB_L((u64 *) LOCAL_MMR_ADDR(SH2_IPI_ACCESS1));
+		    (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS1));
 		xpc_sh2_IPI_access2 =
 		xpc_sh2_IPI_access2 =
-			(u64) HUB_L((u64 *) LOCAL_MMR_ADDR(SH2_IPI_ACCESS2));
+		    (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS2));
 		xpc_sh2_IPI_access3 =
 		xpc_sh2_IPI_access3 =
-			(u64) HUB_L((u64 *) LOCAL_MMR_ADDR(SH2_IPI_ACCESS3));
+		    (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS3));
 
 
 		for_each_online_node(node) {
 		for_each_online_node(node) {
 			nasid = cnodeid_to_nasid(node);
 			nasid = cnodeid_to_nasid(node);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS0),
-								-1UL);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS1),
-								-1UL);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS2),
-								-1UL);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS3),
-								-1UL);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS0),
+			      -1UL);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS1),
+			      -1UL);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS2),
+			      -1UL);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS3),
+			      -1UL);
 		}
 		}
 
 
 	} else {
 	} else {
 		xpc_sh1_IPI_access =
 		xpc_sh1_IPI_access =
-			(u64) HUB_L((u64 *) LOCAL_MMR_ADDR(SH1_IPI_ACCESS));
+		    (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH1_IPI_ACCESS));
 
 
 		for_each_online_node(node) {
 		for_each_online_node(node) {
 			nasid = cnodeid_to_nasid(node);
 			nasid = cnodeid_to_nasid(node);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH1_IPI_ACCESS),
-								-1UL);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH1_IPI_ACCESS),
+			      -1UL);
 
 
 			/*
 			/*
 			 * Since the BIST collides with memory operations on
 			 * Since the BIST collides with memory operations on
@@ -347,21 +333,23 @@ xpc_allow_IPI_ops(void)
 			 */
 			 */
 			if (enable_shub_wars_1_1()) {
 			if (enable_shub_wars_1_1()) {
 				/* open up everything */
 				/* open up everything */
-				xpc_prot_vec[node] = (u64) HUB_L((u64 *)
-						GLOBAL_MMR_ADDR(nasid,
-						SH1_MD_DQLP_MMR_DIR_PRIVEC0));
-				HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid,
-						SH1_MD_DQLP_MMR_DIR_PRIVEC0),
-								-1UL);
-				HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid,
-						SH1_MD_DQRP_MMR_DIR_PRIVEC0),
-								-1UL);
+				xpc_prot_vec[node] = (u64)HUB_L((u64 *)
+								GLOBAL_MMR_ADDR
+								(nasid,
+								 SH1_MD_DQLP_MMR_DIR_PRIVEC0));
+				HUB_S((u64 *)
+				      GLOBAL_MMR_ADDR(nasid,
+						      SH1_MD_DQLP_MMR_DIR_PRIVEC0),
+				      -1UL);
+				HUB_S((u64 *)
+				      GLOBAL_MMR_ADDR(nasid,
+						      SH1_MD_DQRP_MMR_DIR_PRIVEC0),
+				      -1UL);
 			}
 			}
 		}
 		}
 	}
 	}
 }
 }
 
 
-
 /*
 /*
  * Restrict protections to disallow IPI operations (and AMO operations on
  * Restrict protections to disallow IPI operations (and AMO operations on
  * Shub 1.1 systems).
  * Shub 1.1 systems).
@@ -372,43 +360,41 @@ xpc_restrict_IPI_ops(void)
 	int node;
 	int node;
 	int nasid;
 	int nasid;
 
 
-
 	// >>> Change SH_IPI_ACCESS code to use SAL call once it is available.
 	// >>> Change SH_IPI_ACCESS code to use SAL call once it is available.
 
 
 	if (is_shub2()) {
 	if (is_shub2()) {
 
 
 		for_each_online_node(node) {
 		for_each_online_node(node) {
 			nasid = cnodeid_to_nasid(node);
 			nasid = cnodeid_to_nasid(node);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS0),
-							xpc_sh2_IPI_access0);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS1),
-							xpc_sh2_IPI_access1);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS2),
-							xpc_sh2_IPI_access2);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS3),
-							xpc_sh2_IPI_access3);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS0),
+			      xpc_sh2_IPI_access0);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS1),
+			      xpc_sh2_IPI_access1);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS2),
+			      xpc_sh2_IPI_access2);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS3),
+			      xpc_sh2_IPI_access3);
 		}
 		}
 
 
 	} else {
 	} else {
 
 
 		for_each_online_node(node) {
 		for_each_online_node(node) {
 			nasid = cnodeid_to_nasid(node);
 			nasid = cnodeid_to_nasid(node);
-			HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, SH1_IPI_ACCESS),
-							xpc_sh1_IPI_access);
+			HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH1_IPI_ACCESS),
+			      xpc_sh1_IPI_access);
 
 
 			if (enable_shub_wars_1_1()) {
 			if (enable_shub_wars_1_1()) {
-				HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid,
-						SH1_MD_DQLP_MMR_DIR_PRIVEC0),
-							xpc_prot_vec[node]);
-				HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid,
-						SH1_MD_DQRP_MMR_DIR_PRIVEC0),
-							xpc_prot_vec[node]);
+				HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid,
+							     SH1_MD_DQLP_MMR_DIR_PRIVEC0),
+				      xpc_prot_vec[node]);
+				HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid,
+							     SH1_MD_DQRP_MMR_DIR_PRIVEC0),
+				      xpc_prot_vec[node]);
 			}
 			}
 		}
 		}
 	}
 	}
 }
 }
 
 
-
 /*
 /*
  * At periodic intervals, scan through all active partitions and ensure
  * At periodic intervals, scan through all active partitions and ensure
  * their heartbeat is still active.  If not, the partition is deactivated.
  * their heartbeat is still active.  If not, the partition is deactivated.
@@ -421,8 +407,7 @@ xpc_check_remote_hb(void)
 	partid_t partid;
 	partid_t partid;
 	bte_result_t bres;
 	bte_result_t bres;
 
 
-
-	remote_vars = (struct xpc_vars *) xpc_remote_copy_buffer;
+	remote_vars = (struct xpc_vars *)xpc_remote_copy_buffer;
 
 
 	for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) {
 	for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) {
 
 
@@ -437,18 +422,18 @@ xpc_check_remote_hb(void)
 		part = &xpc_partitions[partid];
 		part = &xpc_partitions[partid];
 
 
 		if (part->act_state == XPC_P_INACTIVE ||
 		if (part->act_state == XPC_P_INACTIVE ||
-				part->act_state == XPC_P_DEACTIVATING) {
+		    part->act_state == XPC_P_DEACTIVATING) {
 			continue;
 			continue;
 		}
 		}
 
 
 		/* pull the remote_hb cache line */
 		/* pull the remote_hb cache line */
 		bres = xp_bte_copy(part->remote_vars_pa,
 		bres = xp_bte_copy(part->remote_vars_pa,
-					(u64) remote_vars,
-					XPC_RP_VARS_SIZE,
-					(BTE_NOTIFY | BTE_WACQUIRE), NULL);
+				   (u64)remote_vars,
+				   XPC_RP_VARS_SIZE,
+				   (BTE_NOTIFY | BTE_WACQUIRE), NULL);
 		if (bres != BTE_SUCCESS) {
 		if (bres != BTE_SUCCESS) {
 			XPC_DEACTIVATE_PARTITION(part,
 			XPC_DEACTIVATE_PARTITION(part,
-						xpc_map_bte_errors(bres));
+						 xpc_map_bte_errors(bres));
 			continue;
 			continue;
 		}
 		}
 
 
@@ -459,8 +444,8 @@ xpc_check_remote_hb(void)
 			remote_vars->heartbeating_to_mask);
 			remote_vars->heartbeating_to_mask);
 
 
 		if (((remote_vars->heartbeat == part->last_heartbeat) &&
 		if (((remote_vars->heartbeat == part->last_heartbeat) &&
-			(remote_vars->heartbeat_offline == 0)) ||
-			     !xpc_hb_allowed(sn_partition_id, remote_vars)) {
+		     (remote_vars->heartbeat_offline == 0)) ||
+		    !xpc_hb_allowed(sn_partition_id, remote_vars)) {
 
 
 			XPC_DEACTIVATE_PARTITION(part, xpcNoHeartbeat);
 			XPC_DEACTIVATE_PARTITION(part, xpcNoHeartbeat);
 			continue;
 			continue;
@@ -470,7 +455,6 @@ xpc_check_remote_hb(void)
 	}
 	}
 }
 }
 
 
-
 /*
 /*
  * Get a copy of a portion of the remote partition's rsvd page.
  * Get a copy of a portion of the remote partition's rsvd page.
  *
  *
@@ -480,11 +464,10 @@ xpc_check_remote_hb(void)
  */
  */
 static enum xpc_retval
 static enum xpc_retval
 xpc_get_remote_rp(int nasid, u64 *discovered_nasids,
 xpc_get_remote_rp(int nasid, u64 *discovered_nasids,
-		struct xpc_rsvd_page *remote_rp, u64 *remote_rp_pa)
+		  struct xpc_rsvd_page *remote_rp, u64 *remote_rp_pa)
 {
 {
 	int bres, i;
 	int bres, i;
 
 
-
 	/* get the reserved page's physical address */
 	/* get the reserved page's physical address */
 
 
 	*remote_rp_pa = xpc_get_rsvd_page_pa(nasid);
 	*remote_rp_pa = xpc_get_rsvd_page_pa(nasid);
@@ -492,30 +475,26 @@ xpc_get_remote_rp(int nasid, u64 *discovered_nasids,
 		return xpcNoRsvdPageAddr;
 		return xpcNoRsvdPageAddr;
 	}
 	}
 
 
-
 	/* pull over the reserved page header and part_nasids mask */
 	/* pull over the reserved page header and part_nasids mask */
-	bres = xp_bte_copy(*remote_rp_pa, (u64) remote_rp,
-				XPC_RP_HEADER_SIZE + xp_nasid_mask_bytes,
-				(BTE_NOTIFY | BTE_WACQUIRE), NULL);
+	bres = xp_bte_copy(*remote_rp_pa, (u64)remote_rp,
+			   XPC_RP_HEADER_SIZE + xp_nasid_mask_bytes,
+			   (BTE_NOTIFY | BTE_WACQUIRE), NULL);
 	if (bres != BTE_SUCCESS) {
 	if (bres != BTE_SUCCESS) {
 		return xpc_map_bte_errors(bres);
 		return xpc_map_bte_errors(bres);
 	}
 	}
 
 
-
 	if (discovered_nasids != NULL) {
 	if (discovered_nasids != NULL) {
 		u64 *remote_part_nasids = XPC_RP_PART_NASIDS(remote_rp);
 		u64 *remote_part_nasids = XPC_RP_PART_NASIDS(remote_rp);
 
 
-
 		for (i = 0; i < xp_nasid_mask_words; i++) {
 		for (i = 0; i < xp_nasid_mask_words; i++) {
 			discovered_nasids[i] |= remote_part_nasids[i];
 			discovered_nasids[i] |= remote_part_nasids[i];
 		}
 		}
 	}
 	}
 
 
-
 	/* check that the partid is for another partition */
 	/* check that the partid is for another partition */
 
 
 	if (remote_rp->partid < 1 ||
 	if (remote_rp->partid < 1 ||
-				remote_rp->partid > (XP_MAX_PARTITIONS - 1)) {
+	    remote_rp->partid > (XP_MAX_PARTITIONS - 1)) {
 		return xpcInvalidPartid;
 		return xpcInvalidPartid;
 	}
 	}
 
 
@@ -523,16 +502,14 @@ xpc_get_remote_rp(int nasid, u64 *discovered_nasids,
 		return xpcLocalPartid;
 		return xpcLocalPartid;
 	}
 	}
 
 
-
 	if (XPC_VERSION_MAJOR(remote_rp->version) !=
 	if (XPC_VERSION_MAJOR(remote_rp->version) !=
-					XPC_VERSION_MAJOR(XPC_RP_VERSION)) {
+	    XPC_VERSION_MAJOR(XPC_RP_VERSION)) {
 		return xpcBadVersion;
 		return xpcBadVersion;
 	}
 	}
 
 
 	return xpcSuccess;
 	return xpcSuccess;
 }
 }
 
 
-
 /*
 /*
  * Get a copy of the remote partition's XPC variables from the reserved page.
  * Get a copy of the remote partition's XPC variables from the reserved page.
  *
  *
@@ -544,34 +521,32 @@ xpc_get_remote_vars(u64 remote_vars_pa, struct xpc_vars *remote_vars)
 {
 {
 	int bres;
 	int bres;
 
 
-
 	if (remote_vars_pa == 0) {
 	if (remote_vars_pa == 0) {
 		return xpcVarsNotSet;
 		return xpcVarsNotSet;
 	}
 	}
 
 
 	/* pull over the cross partition variables */
 	/* pull over the cross partition variables */
-	bres = xp_bte_copy(remote_vars_pa, (u64) remote_vars, XPC_RP_VARS_SIZE,
-				(BTE_NOTIFY | BTE_WACQUIRE), NULL);
+	bres = xp_bte_copy(remote_vars_pa, (u64)remote_vars, XPC_RP_VARS_SIZE,
+			   (BTE_NOTIFY | BTE_WACQUIRE), NULL);
 	if (bres != BTE_SUCCESS) {
 	if (bres != BTE_SUCCESS) {
 		return xpc_map_bte_errors(bres);
 		return xpc_map_bte_errors(bres);
 	}
 	}
 
 
 	if (XPC_VERSION_MAJOR(remote_vars->version) !=
 	if (XPC_VERSION_MAJOR(remote_vars->version) !=
-					XPC_VERSION_MAJOR(XPC_V_VERSION)) {
+	    XPC_VERSION_MAJOR(XPC_V_VERSION)) {
 		return xpcBadVersion;
 		return xpcBadVersion;
 	}
 	}
 
 
 	return xpcSuccess;
 	return xpcSuccess;
 }
 }
 
 
-
 /*
 /*
  * Update the remote partition's info.
  * Update the remote partition's info.
  */
  */
 static void
 static void
 xpc_update_partition_info(struct xpc_partition *part, u8 remote_rp_version,
 xpc_update_partition_info(struct xpc_partition *part, u8 remote_rp_version,
-		struct timespec *remote_rp_stamp, u64 remote_rp_pa,
-		u64 remote_vars_pa, struct xpc_vars *remote_vars)
+			  struct timespec *remote_rp_stamp, u64 remote_rp_pa,
+			  u64 remote_vars_pa, struct xpc_vars *remote_vars)
 {
 {
 	part->remote_rp_version = remote_rp_version;
 	part->remote_rp_version = remote_rp_version;
 	dev_dbg(xpc_part, "  remote_rp_version = 0x%016x\n",
 	dev_dbg(xpc_part, "  remote_rp_version = 0x%016x\n",
@@ -613,7 +588,6 @@ xpc_update_partition_info(struct xpc_partition *part, u8 remote_rp_version,
 		part->remote_vars_version);
 		part->remote_vars_version);
 }
 }
 
 
-
 /*
 /*
  * Prior code has determined the nasid which generated an IPI.  Inspect
  * Prior code has determined the nasid which generated an IPI.  Inspect
  * that nasid to determine if its partition needs to be activated or
  * that nasid to determine if its partition needs to be activated or
@@ -643,15 +617,14 @@ xpc_identify_act_IRQ_req(int nasid)
 	struct xpc_partition *part;
 	struct xpc_partition *part;
 	enum xpc_retval ret;
 	enum xpc_retval ret;
 
 
-
 	/* pull over the reserved page structure */
 	/* pull over the reserved page structure */
 
 
-	remote_rp = (struct xpc_rsvd_page *) xpc_remote_copy_buffer;
+	remote_rp = (struct xpc_rsvd_page *)xpc_remote_copy_buffer;
 
 
 	ret = xpc_get_remote_rp(nasid, NULL, remote_rp, &remote_rp_pa);
 	ret = xpc_get_remote_rp(nasid, NULL, remote_rp, &remote_rp_pa);
 	if (ret != xpcSuccess) {
 	if (ret != xpcSuccess) {
 		dev_warn(xpc_part, "unable to get reserved page from nasid %d, "
 		dev_warn(xpc_part, "unable to get reserved page from nasid %d, "
-			"which sent interrupt, reason=%d\n", nasid, ret);
+			 "which sent interrupt, reason=%d\n", nasid, ret);
 		return;
 		return;
 	}
 	}
 
 
@@ -663,34 +636,31 @@ xpc_identify_act_IRQ_req(int nasid)
 	partid = remote_rp->partid;
 	partid = remote_rp->partid;
 	part = &xpc_partitions[partid];
 	part = &xpc_partitions[partid];
 
 
-
 	/* pull over the cross partition variables */
 	/* pull over the cross partition variables */
 
 
-	remote_vars = (struct xpc_vars *) xpc_remote_copy_buffer;
+	remote_vars = (struct xpc_vars *)xpc_remote_copy_buffer;
 
 
 	ret = xpc_get_remote_vars(remote_vars_pa, remote_vars);
 	ret = xpc_get_remote_vars(remote_vars_pa, remote_vars);
 	if (ret != xpcSuccess) {
 	if (ret != xpcSuccess) {
 
 
 		dev_warn(xpc_part, "unable to get XPC variables from nasid %d, "
 		dev_warn(xpc_part, "unable to get XPC variables from nasid %d, "
-			"which sent interrupt, reason=%d\n", nasid, ret);
+			 "which sent interrupt, reason=%d\n", nasid, ret);
 
 
 		XPC_DEACTIVATE_PARTITION(part, ret);
 		XPC_DEACTIVATE_PARTITION(part, ret);
 		return;
 		return;
 	}
 	}
 
 
-
 	part->act_IRQ_rcvd++;
 	part->act_IRQ_rcvd++;
 
 
 	dev_dbg(xpc_part, "partid for nasid %d is %d; IRQs = %d; HB = "
 	dev_dbg(xpc_part, "partid for nasid %d is %d; IRQs = %d; HB = "
-		"%ld:0x%lx\n", (int) nasid, (int) partid, part->act_IRQ_rcvd,
+		"%ld:0x%lx\n", (int)nasid, (int)partid, part->act_IRQ_rcvd,
 		remote_vars->heartbeat, remote_vars->heartbeating_to_mask);
 		remote_vars->heartbeat, remote_vars->heartbeating_to_mask);
 
 
-	if (xpc_partition_disengaged(part) &&
-					part->act_state == XPC_P_INACTIVE) {
+	if (xpc_partition_disengaged(part) && part->act_state == XPC_P_INACTIVE) {
 
 
 		xpc_update_partition_info(part, remote_rp_version,
 		xpc_update_partition_info(part, remote_rp_version,
-					&remote_rp_stamp, remote_rp_pa,
-					remote_vars_pa, remote_vars);
+					  &remote_rp_stamp, remote_rp_pa,
+					  remote_vars_pa, remote_vars);
 
 
 		if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version)) {
 		if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version)) {
 			if (xpc_partition_disengage_requested(1UL << partid)) {
 			if (xpc_partition_disengage_requested(1UL << partid)) {
@@ -714,16 +684,15 @@ xpc_identify_act_IRQ_req(int nasid)
 
 
 	if (!XPC_SUPPORTS_RP_STAMP(part->remote_rp_version)) {
 	if (!XPC_SUPPORTS_RP_STAMP(part->remote_rp_version)) {
 		DBUG_ON(XPC_SUPPORTS_DISENGAGE_REQUEST(part->
 		DBUG_ON(XPC_SUPPORTS_DISENGAGE_REQUEST(part->
-							remote_vars_version));
+						       remote_vars_version));
 
 
 		if (!XPC_SUPPORTS_RP_STAMP(remote_rp_version)) {
 		if (!XPC_SUPPORTS_RP_STAMP(remote_rp_version)) {
 			DBUG_ON(XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->
 			DBUG_ON(XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->
-								version));
+							       version));
 			/* see if the other side rebooted */
 			/* see if the other side rebooted */
 			if (part->remote_amos_page_pa ==
 			if (part->remote_amos_page_pa ==
-				remote_vars->amos_page_pa &&
-					xpc_hb_allowed(sn_partition_id,
-								remote_vars)) {
+			    remote_vars->amos_page_pa &&
+			    xpc_hb_allowed(sn_partition_id, remote_vars)) {
 				/* doesn't look that way, so ignore the IPI */
 				/* doesn't look that way, so ignore the IPI */
 				return;
 				return;
 			}
 			}
@@ -735,8 +704,8 @@ xpc_identify_act_IRQ_req(int nasid)
 		 */
 		 */
 
 
 		xpc_update_partition_info(part, remote_rp_version,
 		xpc_update_partition_info(part, remote_rp_version,
-						&remote_rp_stamp, remote_rp_pa,
-						remote_vars_pa, remote_vars);
+					  &remote_rp_stamp, remote_rp_pa,
+					  remote_vars_pa, remote_vars);
 		part->reactivate_nasid = nasid;
 		part->reactivate_nasid = nasid;
 		XPC_DEACTIVATE_PARTITION(part, xpcReactivating);
 		XPC_DEACTIVATE_PARTITION(part, xpcReactivating);
 		return;
 		return;
@@ -756,15 +725,15 @@ xpc_identify_act_IRQ_req(int nasid)
 		xpc_clear_partition_disengage_request(1UL << partid);
 		xpc_clear_partition_disengage_request(1UL << partid);
 
 
 		xpc_update_partition_info(part, remote_rp_version,
 		xpc_update_partition_info(part, remote_rp_version,
-						&remote_rp_stamp, remote_rp_pa,
-						remote_vars_pa, remote_vars);
+					  &remote_rp_stamp, remote_rp_pa,
+					  remote_vars_pa, remote_vars);
 		reactivate = 1;
 		reactivate = 1;
 
 
 	} else {
 	} else {
 		DBUG_ON(!XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->version));
 		DBUG_ON(!XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->version));
 
 
 		stamp_diff = xpc_compare_stamps(&part->remote_rp_stamp,
 		stamp_diff = xpc_compare_stamps(&part->remote_rp_stamp,
-							&remote_rp_stamp);
+						&remote_rp_stamp);
 		if (stamp_diff != 0) {
 		if (stamp_diff != 0) {
 			DBUG_ON(stamp_diff >= 0);
 			DBUG_ON(stamp_diff >= 0);
 
 
@@ -775,17 +744,18 @@ xpc_identify_act_IRQ_req(int nasid)
 
 
 			DBUG_ON(xpc_partition_engaged(1UL << partid));
 			DBUG_ON(xpc_partition_engaged(1UL << partid));
 			DBUG_ON(xpc_partition_disengage_requested(1UL <<
 			DBUG_ON(xpc_partition_disengage_requested(1UL <<
-								partid));
+								  partid));
 
 
 			xpc_update_partition_info(part, remote_rp_version,
 			xpc_update_partition_info(part, remote_rp_version,
-						&remote_rp_stamp, remote_rp_pa,
-						remote_vars_pa, remote_vars);
+						  &remote_rp_stamp,
+						  remote_rp_pa, remote_vars_pa,
+						  remote_vars);
 			reactivate = 1;
 			reactivate = 1;
 		}
 		}
 	}
 	}
 
 
 	if (part->disengage_request_timeout > 0 &&
 	if (part->disengage_request_timeout > 0 &&
-					!xpc_partition_disengaged(part)) {
+	    !xpc_partition_disengaged(part)) {
 		/* still waiting on other side to disengage from us */
 		/* still waiting on other side to disengage from us */
 		return;
 		return;
 	}
 	}
@@ -795,12 +765,11 @@ xpc_identify_act_IRQ_req(int nasid)
 		XPC_DEACTIVATE_PARTITION(part, xpcReactivating);
 		XPC_DEACTIVATE_PARTITION(part, xpcReactivating);
 
 
 	} else if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version) &&
 	} else if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version) &&
-			xpc_partition_disengage_requested(1UL << partid)) {
+		   xpc_partition_disengage_requested(1UL << partid)) {
 		XPC_DEACTIVATE_PARTITION(part, xpcOtherGoingDown);
 		XPC_DEACTIVATE_PARTITION(part, xpcOtherGoingDown);
 	}
 	}
 }
 }
 
 
-
 /*
 /*
  * Loop through the activation AMO variables and process any bits
  * Loop through the activation AMO variables and process any bits
  * which are set.  Each bit indicates a nasid sending a partition
  * which are set.  Each bit indicates a nasid sending a partition
@@ -813,14 +782,12 @@ xpc_identify_act_IRQ_sender(void)
 {
 {
 	int word, bit;
 	int word, bit;
 	u64 nasid_mask;
 	u64 nasid_mask;
-	u64 nasid;			/* remote nasid */
+	u64 nasid;		/* remote nasid */
 	int n_IRQs_detected = 0;
 	int n_IRQs_detected = 0;
 	AMO_t *act_amos;
 	AMO_t *act_amos;
 
 
-
 	act_amos = xpc_vars->amos_page + XPC_ACTIVATE_IRQ_AMOS;
 	act_amos = xpc_vars->amos_page + XPC_ACTIVATE_IRQ_AMOS;
 
 
-
 	/* scan through act AMO variable looking for non-zero entries */
 	/* scan through act AMO variable looking for non-zero entries */
 	for (word = 0; word < xp_nasid_mask_words; word++) {
 	for (word = 0; word < xp_nasid_mask_words; word++) {
 
 
@@ -837,7 +804,6 @@ xpc_identify_act_IRQ_sender(void)
 		dev_dbg(xpc_part, "AMO[%d] gave back 0x%lx\n", word,
 		dev_dbg(xpc_part, "AMO[%d] gave back 0x%lx\n", word,
 			nasid_mask);
 			nasid_mask);
 
 
-
 		/*
 		/*
 		 * If this nasid has been added to the machine since
 		 * If this nasid has been added to the machine since
 		 * our partition was reset, this will retain the
 		 * our partition was reset, this will retain the
@@ -846,7 +812,6 @@ xpc_identify_act_IRQ_sender(void)
 		 */
 		 */
 		xpc_mach_nasids[word] |= nasid_mask;
 		xpc_mach_nasids[word] |= nasid_mask;
 
 
-
 		/* locate the nasid(s) which sent interrupts */
 		/* locate the nasid(s) which sent interrupts */
 
 
 		for (bit = 0; bit < (8 * sizeof(u64)); bit++) {
 		for (bit = 0; bit < (8 * sizeof(u64)); bit++) {
@@ -862,7 +827,6 @@ xpc_identify_act_IRQ_sender(void)
 	return n_IRQs_detected;
 	return n_IRQs_detected;
 }
 }
 
 
-
 /*
 /*
  * See if the other side has responded to a partition disengage request
  * See if the other side has responded to a partition disengage request
  * from us.
  * from us.
@@ -873,7 +837,6 @@ xpc_partition_disengaged(struct xpc_partition *part)
 	partid_t partid = XPC_PARTID(part);
 	partid_t partid = XPC_PARTID(part);
 	int disengaged;
 	int disengaged;
 
 
-
 	disengaged = (xpc_partition_engaged(1UL << partid) == 0);
 	disengaged = (xpc_partition_engaged(1UL << partid) == 0);
 	if (part->disengage_request_timeout) {
 	if (part->disengage_request_timeout) {
 		if (!disengaged) {
 		if (!disengaged) {
@@ -888,7 +851,7 @@ xpc_partition_disengaged(struct xpc_partition *part)
 			 */
 			 */
 
 
 			dev_info(xpc_part, "disengage from remote partition %d "
 			dev_info(xpc_part, "disengage from remote partition %d "
-				"timed out\n", partid);
+				 "timed out\n", partid);
 			xpc_disengage_request_timedout = 1;
 			xpc_disengage_request_timedout = 1;
 			xpc_clear_partition_engaged(1UL << partid);
 			xpc_clear_partition_engaged(1UL << partid);
 			disengaged = 1;
 			disengaged = 1;
@@ -898,11 +861,11 @@ xpc_partition_disengaged(struct xpc_partition *part)
 		/* cancel the timer function, provided it's not us */
 		/* cancel the timer function, provided it's not us */
 		if (!in_interrupt()) {
 		if (!in_interrupt()) {
 			del_singleshot_timer_sync(&part->
 			del_singleshot_timer_sync(&part->
-						      disengage_request_timer);
+						  disengage_request_timer);
 		}
 		}
 
 
 		DBUG_ON(part->act_state != XPC_P_DEACTIVATING &&
 		DBUG_ON(part->act_state != XPC_P_DEACTIVATING &&
-					part->act_state != XPC_P_INACTIVE);
+			part->act_state != XPC_P_INACTIVE);
 		if (part->act_state != XPC_P_INACTIVE) {
 		if (part->act_state != XPC_P_INACTIVE) {
 			xpc_wakeup_channel_mgr(part);
 			xpc_wakeup_channel_mgr(part);
 		}
 		}
@@ -914,7 +877,6 @@ xpc_partition_disengaged(struct xpc_partition *part)
 	return disengaged;
 	return disengaged;
 }
 }
 
 
-
 /*
 /*
  * Mark specified partition as active.
  * Mark specified partition as active.
  */
  */
@@ -924,7 +886,6 @@ xpc_mark_partition_active(struct xpc_partition *part)
 	unsigned long irq_flags;
 	unsigned long irq_flags;
 	enum xpc_retval ret;
 	enum xpc_retval ret;
 
 
-
 	dev_dbg(xpc_part, "setting partition %d to ACTIVE\n", XPC_PARTID(part));
 	dev_dbg(xpc_part, "setting partition %d to ACTIVE\n", XPC_PARTID(part));
 
 
 	spin_lock_irqsave(&part->act_lock, irq_flags);
 	spin_lock_irqsave(&part->act_lock, irq_flags);
@@ -940,17 +901,15 @@ xpc_mark_partition_active(struct xpc_partition *part)
 	return ret;
 	return ret;
 }
 }
 
 
-
 /*
 /*
  * Notify XPC that the partition is down.
  * Notify XPC that the partition is down.
  */
  */
 void
 void
 xpc_deactivate_partition(const int line, struct xpc_partition *part,
 xpc_deactivate_partition(const int line, struct xpc_partition *part,
-				enum xpc_retval reason)
+			 enum xpc_retval reason)
 {
 {
 	unsigned long irq_flags;
 	unsigned long irq_flags;
 
 
-
 	spin_lock_irqsave(&part->act_lock, irq_flags);
 	spin_lock_irqsave(&part->act_lock, irq_flags);
 
 
 	if (part->act_state == XPC_P_INACTIVE) {
 	if (part->act_state == XPC_P_INACTIVE) {
@@ -964,7 +923,7 @@ xpc_deactivate_partition(const int line, struct xpc_partition *part,
 	}
 	}
 	if (part->act_state == XPC_P_DEACTIVATING) {
 	if (part->act_state == XPC_P_DEACTIVATING) {
 		if ((part->reason == xpcUnloading && reason != xpcUnloading) ||
 		if ((part->reason == xpcUnloading && reason != xpcUnloading) ||
-					reason == xpcReactivating) {
+		    reason == xpcReactivating) {
 			XPC_SET_REASON(part, reason, line);
 			XPC_SET_REASON(part, reason, line);
 		}
 		}
 		spin_unlock_irqrestore(&part->act_lock, irq_flags);
 		spin_unlock_irqrestore(&part->act_lock, irq_flags);
@@ -982,9 +941,9 @@ xpc_deactivate_partition(const int line, struct xpc_partition *part,
 
 
 		/* set a timelimit on the disengage request */
 		/* set a timelimit on the disengage request */
 		part->disengage_request_timeout = jiffies +
 		part->disengage_request_timeout = jiffies +
-					(xpc_disengage_request_timelimit * HZ);
+		    (xpc_disengage_request_timelimit * HZ);
 		part->disengage_request_timer.expires =
 		part->disengage_request_timer.expires =
-					part->disengage_request_timeout;
+		    part->disengage_request_timeout;
 		add_timer(&part->disengage_request_timer);
 		add_timer(&part->disengage_request_timer);
 	}
 	}
 
 
@@ -994,7 +953,6 @@ xpc_deactivate_partition(const int line, struct xpc_partition *part,
 	xpc_partition_going_down(part, reason);
 	xpc_partition_going_down(part, reason);
 }
 }
 
 
-
 /*
 /*
  * Mark specified partition as inactive.
  * Mark specified partition as inactive.
  */
  */
@@ -1003,7 +961,6 @@ xpc_mark_partition_inactive(struct xpc_partition *part)
 {
 {
 	unsigned long irq_flags;
 	unsigned long irq_flags;
 
 
-
 	dev_dbg(xpc_part, "setting partition %d to INACTIVE\n",
 	dev_dbg(xpc_part, "setting partition %d to INACTIVE\n",
 		XPC_PARTID(part));
 		XPC_PARTID(part));
 
 
@@ -1013,7 +970,6 @@ xpc_mark_partition_inactive(struct xpc_partition *part)
 	part->remote_rp_pa = 0;
 	part->remote_rp_pa = 0;
 }
 }
 
 
-
 /*
 /*
  * SAL has provided a partition and machine mask.  The partition mask
  * SAL has provided a partition and machine mask.  The partition mask
  * contains a bit for each even nasid in our partition.  The machine
  * contains a bit for each even nasid in our partition.  The machine
@@ -1041,24 +997,22 @@ xpc_discovery(void)
 	u64 *discovered_nasids;
 	u64 *discovered_nasids;
 	enum xpc_retval ret;
 	enum xpc_retval ret;
 
 
-
 	remote_rp = xpc_kmalloc_cacheline_aligned(XPC_RP_HEADER_SIZE +
 	remote_rp = xpc_kmalloc_cacheline_aligned(XPC_RP_HEADER_SIZE +
-						xp_nasid_mask_bytes,
-						GFP_KERNEL, &remote_rp_base);
+						  xp_nasid_mask_bytes,
+						  GFP_KERNEL, &remote_rp_base);
 	if (remote_rp == NULL) {
 	if (remote_rp == NULL) {
 		return;
 		return;
 	}
 	}
-	remote_vars = (struct xpc_vars *) remote_rp;
-
+	remote_vars = (struct xpc_vars *)remote_rp;
 
 
 	discovered_nasids = kzalloc(sizeof(u64) * xp_nasid_mask_words,
 	discovered_nasids = kzalloc(sizeof(u64) * xp_nasid_mask_words,
-							GFP_KERNEL);
+				    GFP_KERNEL);
 	if (discovered_nasids == NULL) {
 	if (discovered_nasids == NULL) {
 		kfree(remote_rp_base);
 		kfree(remote_rp_base);
 		return;
 		return;
 	}
 	}
 
 
-	rp = (struct xpc_rsvd_page *) xpc_rsvd_page;
+	rp = (struct xpc_rsvd_page *)xpc_rsvd_page;
 
 
 	/*
 	/*
 	 * The term 'region' in this context refers to the minimum number of
 	 * The term 'region' in this context refers to the minimum number of
@@ -1081,23 +1035,21 @@ xpc_discovery(void)
 
 
 	for (region = 0; region < max_regions; region++) {
 	for (region = 0; region < max_regions; region++) {
 
 
-		if ((volatile int) xpc_exiting) {
+		if ((volatile int)xpc_exiting) {
 			break;
 			break;
 		}
 		}
 
 
 		dev_dbg(xpc_part, "searching region %d\n", region);
 		dev_dbg(xpc_part, "searching region %d\n", region);
 
 
 		for (nasid = (region * region_size * 2);
 		for (nasid = (region * region_size * 2);
-		     nasid < ((region + 1) * region_size * 2);
-		     nasid += 2) {
+		     nasid < ((region + 1) * region_size * 2); nasid += 2) {
 
 
-			if ((volatile int) xpc_exiting) {
+			if ((volatile int)xpc_exiting) {
 				break;
 				break;
 			}
 			}
 
 
 			dev_dbg(xpc_part, "checking nasid %d\n", nasid);
 			dev_dbg(xpc_part, "checking nasid %d\n", nasid);
 
 
-
 			if (XPC_NASID_IN_ARRAY(nasid, xpc_part_nasids)) {
 			if (XPC_NASID_IN_ARRAY(nasid, xpc_part_nasids)) {
 				dev_dbg(xpc_part, "PROM indicates Nasid %d is "
 				dev_dbg(xpc_part, "PROM indicates Nasid %d is "
 					"part of the local partition; skipping "
 					"part of the local partition; skipping "
@@ -1119,11 +1071,10 @@ xpc_discovery(void)
 				continue;
 				continue;
 			}
 			}
 
 
-
 			/* pull over the reserved page structure */
 			/* pull over the reserved page structure */
 
 
 			ret = xpc_get_remote_rp(nasid, discovered_nasids,
 			ret = xpc_get_remote_rp(nasid, discovered_nasids,
-					      remote_rp, &remote_rp_pa);
+						remote_rp, &remote_rp_pa);
 			if (ret != xpcSuccess) {
 			if (ret != xpcSuccess) {
 				dev_dbg(xpc_part, "unable to get reserved page "
 				dev_dbg(xpc_part, "unable to get reserved page "
 					"from nasid %d, reason=%d\n", nasid,
 					"from nasid %d, reason=%d\n", nasid,
@@ -1140,7 +1091,6 @@ xpc_discovery(void)
 			partid = remote_rp->partid;
 			partid = remote_rp->partid;
 			part = &xpc_partitions[partid];
 			part = &xpc_partitions[partid];
 
 
-
 			/* pull over the cross partition variables */
 			/* pull over the cross partition variables */
 
 
 			ret = xpc_get_remote_vars(remote_vars_pa, remote_vars);
 			ret = xpc_get_remote_vars(remote_vars_pa, remote_vars);
@@ -1171,15 +1121,15 @@ xpc_discovery(void)
 			 * get the same page for remote_act_amos_pa after
 			 * get the same page for remote_act_amos_pa after
 			 * module reloads and system reboots.
 			 * module reloads and system reboots.
 			 */
 			 */
-			if (sn_register_xp_addr_region(
-					    remote_vars->amos_page_pa,
-							PAGE_SIZE, 1) < 0) {
-				dev_dbg(xpc_part, "partition %d failed to "
+			if (sn_register_xp_addr_region
+			    (remote_vars->amos_page_pa, PAGE_SIZE, 1) < 0) {
+				dev_dbg(xpc_part,
+					"partition %d failed to "
 					"register xp_addr region 0x%016lx\n",
 					"register xp_addr region 0x%016lx\n",
 					partid, remote_vars->amos_page_pa);
 					partid, remote_vars->amos_page_pa);
 
 
 				XPC_SET_REASON(part, xpcPhysAddrRegFailed,
 				XPC_SET_REASON(part, xpcPhysAddrRegFailed,
-						__LINE__);
+					       __LINE__);
 				break;
 				break;
 			}
 			}
 
 
@@ -1195,9 +1145,9 @@ xpc_discovery(void)
 				remote_vars->act_phys_cpuid);
 				remote_vars->act_phys_cpuid);
 
 
 			if (XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->
 			if (XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->
-								version)) {
+							   version)) {
 				part->remote_amos_page_pa =
 				part->remote_amos_page_pa =
-						remote_vars->amos_page_pa;
+				    remote_vars->amos_page_pa;
 				xpc_mark_partition_disengaged(part);
 				xpc_mark_partition_disengaged(part);
 				xpc_cancel_partition_disengage_request(part);
 				xpc_cancel_partition_disengage_request(part);
 			}
 			}
@@ -1209,7 +1159,6 @@ xpc_discovery(void)
 	kfree(remote_rp_base);
 	kfree(remote_rp_base);
 }
 }
 
 
-
 /*
 /*
  * Given a partid, get the nasids owned by that partition from the
  * Given a partid, get the nasids owned by that partition from the
  * remote partition's reserved page.
  * remote partition's reserved page.
@@ -1221,7 +1170,6 @@ xpc_initiate_partid_to_nasids(partid_t partid, void *nasid_mask)
 	u64 part_nasid_pa;
 	u64 part_nasid_pa;
 	int bte_res;
 	int bte_res;
 
 
-
 	part = &xpc_partitions[partid];
 	part = &xpc_partitions[partid];
 	if (part->remote_rp_pa == 0) {
 	if (part->remote_rp_pa == 0) {
 		return xpcPartitionDown;
 		return xpcPartitionDown;
@@ -1229,11 +1177,11 @@ xpc_initiate_partid_to_nasids(partid_t partid, void *nasid_mask)
 
 
 	memset(nasid_mask, 0, XP_NASID_MASK_BYTES);
 	memset(nasid_mask, 0, XP_NASID_MASK_BYTES);
 
 
-	part_nasid_pa = (u64) XPC_RP_PART_NASIDS(part->remote_rp_pa);
+	part_nasid_pa = (u64)XPC_RP_PART_NASIDS(part->remote_rp_pa);
 
 
-	bte_res = xp_bte_copy(part_nasid_pa, (u64) nasid_mask,
-			xp_nasid_mask_bytes, (BTE_NOTIFY | BTE_WACQUIRE), NULL);
+	bte_res = xp_bte_copy(part_nasid_pa, (u64)nasid_mask,
+			      xp_nasid_mask_bytes, (BTE_NOTIFY | BTE_WACQUIRE),
+			      NULL);
 
 
 	return xpc_map_bte_errors(bte_res);
 	return xpc_map_bte_errors(bte_res);
 }
 }
-

Some files were not shown because too many files changed in this diff