mirror of
https://github.com/armbian/linux.git
synced 2026-01-06 10:13:00 -08:00
sgi-xp: setup the activate GRU message queue
Setup the activate GRU message queue that is used for partition activation and channel connection on UV systems. Signed-off-by: Dean Nelson <dcn@sgi.com> Cc: Jack Steiner <steiner@sgi.com> Cc: "Luck, Tony" <tony.luck@intel.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
committed by
Linus Torvalds
parent
83469b5525
commit
5b8669dfd1
@@ -208,8 +208,9 @@ enum xp_retval {
|
||||
xpNeedMoreInfo, /* 57: more info is needed by SAL */
|
||||
|
||||
xpGruCopyError, /* 58: gru_copy_gru() returned error */
|
||||
xpGruSendMqError, /* 59: gru send message queue related error */
|
||||
|
||||
xpUnknownReason /* 59: unknown reason - must be last in enum */
|
||||
xpUnknownReason /* 60: unknown reason - must be last in enum */
|
||||
};
|
||||
|
||||
/*
|
||||
|
||||
@@ -42,15 +42,25 @@ xp_remote_memcpy_uv(unsigned long dst_gpa, const unsigned long src_gpa,
|
||||
return xpGruCopyError;
|
||||
}
|
||||
|
||||
static int
|
||||
xp_cpu_to_nasid_uv(int cpuid)
|
||||
{
|
||||
/* ??? Is this same as sn2 nasid in mach/part bitmaps set up by SAL? */
|
||||
return UV_PNODE_TO_NASID(uv_cpu_to_pnode(cpuid));
|
||||
}
|
||||
|
||||
enum xp_retval
|
||||
xp_init_uv(void)
|
||||
{
|
||||
BUG_ON(!is_uv());
|
||||
|
||||
xp_max_npartitions = XP_MAX_NPARTITIONS_UV;
|
||||
xp_partition_id = 0; /* !!! not correct value */
|
||||
xp_region_size = 0; /* !!! not correct value */
|
||||
|
||||
xp_pa = xp_pa_uv;
|
||||
xp_remote_memcpy = xp_remote_memcpy_uv;
|
||||
xp_cpu_to_nasid = xp_cpu_to_nasid_uv;
|
||||
|
||||
return xpSuccess;
|
||||
}
|
||||
|
||||
@@ -164,8 +164,8 @@ struct xpc_vars_part_sn2 {
|
||||
* MAGIC2 indicates that this partition has pulled the remote partititions
|
||||
* per partition variables that pertain to this partition.
|
||||
*/
|
||||
#define XPC_VP_MAGIC1 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */
|
||||
#define XPC_VP_MAGIC2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */
|
||||
#define XPC_VP_MAGIC1_SN2 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */
|
||||
#define XPC_VP_MAGIC2_SN2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */
|
||||
|
||||
/* the reserved page sizes and offsets */
|
||||
|
||||
@@ -180,6 +180,80 @@ struct xpc_vars_part_sn2 {
|
||||
(XPC_RP_MACH_NASIDS(_rp) + \
|
||||
xpc_nasid_mask_nlongs))
|
||||
|
||||
/*
|
||||
* The activate_mq is used to send/receive messages that affect XPC's heartbeat,
|
||||
* partition active state, and channel state. This is UV only.
|
||||
*/
|
||||
struct xpc_activate_mq_msghdr_uv {
|
||||
short partid; /* sender's partid */
|
||||
u8 act_state; /* sender's act_state at time msg sent */
|
||||
u8 type; /* message's type */
|
||||
unsigned long rp_ts_jiffies; /* timestamp of sender's rp setup by XPC */
|
||||
};
|
||||
|
||||
/* activate_mq defined message types */
|
||||
#define XPC_ACTIVATE_MQ_MSG_SYNC_ACT_STATE_UV 0
|
||||
#define XPC_ACTIVATE_MQ_MSG_INC_HEARTBEAT_UV 1
|
||||
#define XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV 2
|
||||
#define XPC_ACTIVATE_MQ_MSG_ONLINE_HEARTBEAT_UV 3
|
||||
|
||||
#define XPC_ACTIVATE_MQ_MSG_ACTIVATE_REQ_UV 4
|
||||
#define XPC_ACTIVATE_MQ_MSG_DEACTIVATE_REQ_UV 5
|
||||
|
||||
#define XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREQUEST_UV 6
|
||||
#define XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREPLY_UV 7
|
||||
#define XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREQUEST_UV 8
|
||||
#define XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREPLY_UV 9
|
||||
|
||||
#define XPC_ACTIVATE_MQ_MSG_MARK_ENGAGED_UV 10
|
||||
#define XPC_ACTIVATE_MQ_MSG_MARK_DISENGAGED_UV 11
|
||||
|
||||
struct xpc_activate_mq_msg_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
};
|
||||
|
||||
struct xpc_activate_mq_msg_heartbeat_req_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
u64 heartbeat;
|
||||
};
|
||||
|
||||
struct xpc_activate_mq_msg_activate_req_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
unsigned long rp_gpa;
|
||||
unsigned long activate_mq_gpa;
|
||||
};
|
||||
|
||||
struct xpc_activate_mq_msg_deactivate_req_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
enum xp_retval reason;
|
||||
};
|
||||
|
||||
struct xpc_activate_mq_msg_chctl_closerequest_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
short ch_number;
|
||||
enum xp_retval reason;
|
||||
};
|
||||
|
||||
struct xpc_activate_mq_msg_chctl_closereply_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
short ch_number;
|
||||
};
|
||||
|
||||
struct xpc_activate_mq_msg_chctl_openrequest_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
short ch_number;
|
||||
short msg_size; /* size of notify_mq's messages */
|
||||
short local_nentries; /* ??? Is this needed? What is? */
|
||||
};
|
||||
|
||||
struct xpc_activate_mq_msg_chctl_openreply_uv {
|
||||
struct xpc_activate_mq_msghdr_uv header;
|
||||
short ch_number;
|
||||
short remote_nentries; /* ??? Is this needed? What is? */
|
||||
short local_nentries; /* ??? Is this needed? What is? */
|
||||
unsigned long local_notify_mq_gpa;
|
||||
};
|
||||
|
||||
/*
|
||||
* Functions registered by add_timer() or called by kernel_thread() only
|
||||
* allow for a single 64-bit argument. The following macros can be used to
|
||||
@@ -331,6 +405,18 @@ struct xpc_notify {
|
||||
*/
|
||||
|
||||
struct xpc_channel_sn2 {
|
||||
struct xpc_openclose_args *local_openclose_args; /* args passed on */
|
||||
/* opening or closing of channel */
|
||||
|
||||
void *local_msgqueue_base; /* base address of kmalloc'd space */
|
||||
struct xpc_msg *local_msgqueue; /* local message queue */
|
||||
void *remote_msgqueue_base; /* base address of kmalloc'd space */
|
||||
struct xpc_msg *remote_msgqueue; /* cached copy of remote partition's */
|
||||
/* local message queue */
|
||||
unsigned long remote_msgqueue_pa; /* phys addr of remote partition's */
|
||||
/* local message queue */
|
||||
|
||||
struct xpc_notify *notify_queue; /* notify queue for messages sent */
|
||||
|
||||
/* various flavors of local and remote Get/Put values */
|
||||
|
||||
@@ -344,13 +430,14 @@ struct xpc_channel_sn2 {
|
||||
};
|
||||
|
||||
struct xpc_channel_uv {
|
||||
/* !!! code is coming */
|
||||
unsigned long remote_notify_mq_gpa; /* gru phys address of remote */
|
||||
/* partition's notify mq */
|
||||
};
|
||||
|
||||
struct xpc_channel {
|
||||
short partid; /* ID of remote partition connected */
|
||||
spinlock_t lock; /* lock for updating this structure */
|
||||
u32 flags; /* general flags */
|
||||
unsigned int flags; /* general flags */
|
||||
|
||||
enum xp_retval reason; /* reason why channel is disconnect'g */
|
||||
int reason_line; /* line# disconnect initiated from */
|
||||
@@ -361,14 +448,6 @@ struct xpc_channel {
|
||||
u16 local_nentries; /* #of msg entries in local msg queue */
|
||||
u16 remote_nentries; /* #of msg entries in remote msg queue */
|
||||
|
||||
void *local_msgqueue_base; /* base address of kmalloc'd space */
|
||||
struct xpc_msg *local_msgqueue; /* local message queue */
|
||||
void *remote_msgqueue_base; /* base address of kmalloc'd space */
|
||||
struct xpc_msg *remote_msgqueue; /* cached copy of remote partition's */
|
||||
/* local message queue */
|
||||
unsigned long remote_msgqueue_pa; /* phys addr of remote partition's */
|
||||
/* local message queue */
|
||||
|
||||
atomic_t references; /* #of external references to queues */
|
||||
|
||||
atomic_t n_on_msg_allocate_wq; /* #on msg allocation wait queue */
|
||||
@@ -377,19 +456,13 @@ struct xpc_channel {
|
||||
u8 delayed_chctl_flags; /* chctl flags received, but delayed */
|
||||
/* action until channel disconnected */
|
||||
|
||||
/* queue of msg senders who want to be notified when msg received */
|
||||
|
||||
atomic_t n_to_notify; /* #of msg senders to notify */
|
||||
struct xpc_notify *notify_queue; /* notify queue for messages sent */
|
||||
|
||||
xpc_channel_func func; /* user's channel function */
|
||||
void *key; /* pointer to user's key */
|
||||
|
||||
struct completion wdisconnect_wait; /* wait for channel disconnect */
|
||||
|
||||
struct xpc_openclose_args *local_openclose_args; /* args passed on */
|
||||
/* opening or closing of channel */
|
||||
|
||||
/* kthread management related fields */
|
||||
|
||||
atomic_t kthreads_assigned; /* #of kthreads assigned to channel */
|
||||
@@ -507,6 +580,8 @@ struct xpc_partition_sn2 {
|
||||
unsigned long remote_GPs_pa; /* phys addr of remote partition's local */
|
||||
/* Get/Put values */
|
||||
|
||||
void *local_openclose_args_base; /* base address of kmalloc'd space */
|
||||
struct xpc_openclose_args *local_openclose_args; /* local's args */
|
||||
unsigned long remote_openclose_args_pa; /* phys addr of remote's args */
|
||||
|
||||
int notify_IRQ_nasid; /* nasid of where to send notify IRQs */
|
||||
@@ -520,9 +595,27 @@ struct xpc_partition_sn2 {
|
||||
};
|
||||
|
||||
struct xpc_partition_uv {
|
||||
/* !!! code is coming */
|
||||
unsigned long remote_activate_mq_gpa; /* gru phys address of remote */
|
||||
/* partition's activate mq */
|
||||
spinlock_t flags_lock; /* protect updating of flags */
|
||||
unsigned int flags; /* general flags */
|
||||
u8 remote_act_state; /* remote partition's act_state */
|
||||
u8 act_state_req; /* act_state request from remote partition */
|
||||
enum xp_retval reason; /* reason for deactivate act_state request */
|
||||
u64 heartbeat; /* incremented by remote partition */
|
||||
};
|
||||
|
||||
/* struct xpc_partition_uv flags */
|
||||
|
||||
#define XPC_P_HEARTBEAT_OFFLINE_UV 0x00000001
|
||||
#define XPC_P_ENGAGED_UV 0x00000002
|
||||
|
||||
/* struct xpc_partition_uv act_state change requests */
|
||||
|
||||
#define XPC_P_ASR_ACTIVATE_UV 0x01
|
||||
#define XPC_P_ASR_REACTIVATE_UV 0x02
|
||||
#define XPC_P_ASR_DEACTIVATE_UV 0x03
|
||||
|
||||
struct xpc_partition {
|
||||
|
||||
/* XPC HB infrastructure */
|
||||
@@ -556,8 +649,6 @@ struct xpc_partition {
|
||||
union xpc_channel_ctl_flags chctl; /* chctl flags yet to be processed */
|
||||
spinlock_t chctl_lock; /* chctl flags lock */
|
||||
|
||||
void *local_openclose_args_base; /* base address of kmalloc'd space */
|
||||
struct xpc_openclose_args *local_openclose_args; /* local's args */
|
||||
void *remote_openclose_args_base; /* base address of kmalloc'd space */
|
||||
struct xpc_openclose_args *remote_openclose_args; /* copy of remote's */
|
||||
/* args */
|
||||
@@ -616,17 +707,20 @@ extern struct device *xpc_part;
|
||||
extern struct device *xpc_chan;
|
||||
extern int xpc_disengage_timelimit;
|
||||
extern int xpc_disengage_timedout;
|
||||
extern atomic_t xpc_activate_IRQ_rcvd;
|
||||
extern int xpc_activate_IRQ_rcvd;
|
||||
extern spinlock_t xpc_activate_IRQ_rcvd_lock;
|
||||
extern wait_queue_head_t xpc_activate_IRQ_wq;
|
||||
extern void *xpc_heartbeating_to_mask;
|
||||
extern void *xpc_kzalloc_cacheline_aligned(size_t, gfp_t, void **);
|
||||
extern void xpc_activate_partition(struct xpc_partition *);
|
||||
extern void xpc_activate_kthreads(struct xpc_channel *, int);
|
||||
extern void xpc_create_kthreads(struct xpc_channel *, int, int);
|
||||
extern void xpc_disconnect_wait(int);
|
||||
extern int (*xpc_setup_partitions_sn) (void);
|
||||
extern enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *, u64 *,
|
||||
unsigned long *,
|
||||
size_t *);
|
||||
extern enum xp_retval (*xpc_rsvd_page_init) (struct xpc_rsvd_page *);
|
||||
extern int (*xpc_setup_rsvd_page_sn) (struct xpc_rsvd_page *);
|
||||
extern void (*xpc_heartbeat_init) (void);
|
||||
extern void (*xpc_heartbeat_exit) (void);
|
||||
extern void (*xpc_increment_heartbeat) (void);
|
||||
@@ -635,8 +729,8 @@ extern void (*xpc_online_heartbeat) (void);
|
||||
extern enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *);
|
||||
extern enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *);
|
||||
extern u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *);
|
||||
extern enum xp_retval (*xpc_allocate_msgqueues) (struct xpc_channel *);
|
||||
extern void (*xpc_free_msgqueues) (struct xpc_channel *);
|
||||
extern enum xp_retval (*xpc_setup_msg_structures) (struct xpc_channel *);
|
||||
extern void (*xpc_teardown_msg_structures) (struct xpc_channel *);
|
||||
extern void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *);
|
||||
extern void (*xpc_process_msg_chctl_flags) (struct xpc_partition *, int);
|
||||
extern int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *);
|
||||
@@ -647,9 +741,9 @@ extern void (*xpc_request_partition_reactivation) (struct xpc_partition *);
|
||||
extern void (*xpc_request_partition_deactivation) (struct xpc_partition *);
|
||||
extern void (*xpc_cancel_partition_deactivation_request) (
|
||||
struct xpc_partition *);
|
||||
extern void (*xpc_process_activate_IRQ_rcvd) (int);
|
||||
extern enum xp_retval (*xpc_setup_infrastructure) (struct xpc_partition *);
|
||||
extern void (*xpc_teardown_infrastructure) (struct xpc_partition *);
|
||||
extern void (*xpc_process_activate_IRQ_rcvd) (void);
|
||||
extern enum xp_retval (*xpc_setup_ch_structures_sn) (struct xpc_partition *);
|
||||
extern void (*xpc_teardown_ch_structures_sn) (struct xpc_partition *);
|
||||
|
||||
extern void (*xpc_indicate_partition_engaged) (struct xpc_partition *);
|
||||
extern int (*xpc_partition_engaged) (short);
|
||||
@@ -665,6 +759,9 @@ extern void (*xpc_send_chctl_openrequest) (struct xpc_channel *,
|
||||
unsigned long *);
|
||||
extern void (*xpc_send_chctl_openreply) (struct xpc_channel *, unsigned long *);
|
||||
|
||||
extern void (*xpc_save_remote_msgqueue_pa) (struct xpc_channel *,
|
||||
unsigned long);
|
||||
|
||||
extern enum xp_retval (*xpc_send_msg) (struct xpc_channel *, u32, void *, u16,
|
||||
u8, xpc_notify_func, void *);
|
||||
extern void (*xpc_received_msg) (struct xpc_channel *, struct xpc_msg *);
|
||||
@@ -674,7 +771,7 @@ extern int xpc_init_sn2(void);
|
||||
extern void xpc_exit_sn2(void);
|
||||
|
||||
/* found in xpc_uv.c */
|
||||
extern void xpc_init_uv(void);
|
||||
extern int xpc_init_uv(void);
|
||||
extern void xpc_exit_uv(void);
|
||||
|
||||
/* found in xpc_partition.c */
|
||||
@@ -684,7 +781,8 @@ extern struct xpc_rsvd_page *xpc_rsvd_page;
|
||||
extern unsigned long *xpc_mach_nasids;
|
||||
extern struct xpc_partition *xpc_partitions;
|
||||
extern void *xpc_kmalloc_cacheline_aligned(size_t, gfp_t, void **);
|
||||
extern struct xpc_rsvd_page *xpc_setup_rsvd_page(void);
|
||||
extern int xpc_setup_rsvd_page(void);
|
||||
extern void xpc_teardown_rsvd_page(void);
|
||||
extern int xpc_identify_activate_IRQ_sender(void);
|
||||
extern int xpc_partition_disengaged(struct xpc_partition *);
|
||||
extern enum xp_retval xpc_mark_partition_active(struct xpc_partition *);
|
||||
|
||||
@@ -39,7 +39,7 @@ xpc_process_connect(struct xpc_channel *ch, unsigned long *irq_flags)
|
||||
|
||||
if (!(ch->flags & XPC_C_SETUP)) {
|
||||
spin_unlock_irqrestore(&ch->lock, *irq_flags);
|
||||
ret = xpc_allocate_msgqueues(ch);
|
||||
ret = xpc_setup_msg_structures(ch);
|
||||
spin_lock_irqsave(&ch->lock, *irq_flags);
|
||||
|
||||
if (ret != xpSuccess)
|
||||
@@ -62,8 +62,6 @@ xpc_process_connect(struct xpc_channel *ch, unsigned long *irq_flags)
|
||||
if (!(ch->flags & XPC_C_ROPENREPLY))
|
||||
return;
|
||||
|
||||
DBUG_ON(ch->remote_msgqueue_pa == 0);
|
||||
|
||||
ch->flags = (XPC_C_CONNECTED | XPC_C_SETUP); /* clear all else */
|
||||
|
||||
dev_info(xpc_chan, "channel %d to partition %d connected\n",
|
||||
@@ -134,13 +132,23 @@ xpc_process_disconnect(struct xpc_channel *ch, unsigned long *irq_flags)
|
||||
spin_lock_irqsave(&ch->lock, *irq_flags);
|
||||
}
|
||||
|
||||
DBUG_ON(atomic_read(&ch->n_to_notify) != 0);
|
||||
|
||||
/* it's now safe to free the channel's message queues */
|
||||
xpc_free_msgqueues(ch);
|
||||
xpc_teardown_msg_structures(ch);
|
||||
|
||||
ch->func = NULL;
|
||||
ch->key = NULL;
|
||||
ch->msg_size = 0;
|
||||
ch->local_nentries = 0;
|
||||
ch->remote_nentries = 0;
|
||||
ch->kthreads_assigned_limit = 0;
|
||||
ch->kthreads_idle_limit = 0;
|
||||
|
||||
/*
|
||||
* Mark the channel disconnected and clear all other flags, including
|
||||
* XPC_C_SETUP (because of call to xpc_free_msgqueues()) but not
|
||||
* including XPC_C_WDISCONNECT (if it was set).
|
||||
* XPC_C_SETUP (because of call to xpc_teardown_msg_structures()) but
|
||||
* not including XPC_C_WDISCONNECT (if it was set).
|
||||
*/
|
||||
ch->flags = (XPC_C_DISCONNECTED | (ch->flags & XPC_C_WDISCONNECT));
|
||||
|
||||
@@ -395,7 +403,7 @@ again:
|
||||
DBUG_ON(args->remote_nentries == 0);
|
||||
|
||||
ch->flags |= XPC_C_ROPENREPLY;
|
||||
ch->remote_msgqueue_pa = args->local_msgqueue_pa;
|
||||
xpc_save_remote_msgqueue_pa(ch, args->local_msgqueue_pa);
|
||||
|
||||
if (args->local_nentries < ch->remote_nentries) {
|
||||
dev_dbg(xpc_chan, "XPC_CHCTL_OPENREPLY: new "
|
||||
|
||||
@@ -141,8 +141,9 @@ static struct ctl_table_header *xpc_sysctl;
|
||||
/* non-zero if any remote partition disengage was timed out */
|
||||
int xpc_disengage_timedout;
|
||||
|
||||
/* #of activate IRQs received */
|
||||
atomic_t xpc_activate_IRQ_rcvd = ATOMIC_INIT(0);
|
||||
/* #of activate IRQs received and not yet processed */
|
||||
int xpc_activate_IRQ_rcvd;
|
||||
DEFINE_SPINLOCK(xpc_activate_IRQ_rcvd_lock);
|
||||
|
||||
/* IRQ handler notifies this wait queue on receipt of an IRQ */
|
||||
DECLARE_WAIT_QUEUE_HEAD(xpc_activate_IRQ_wq);
|
||||
@@ -169,10 +170,11 @@ static struct notifier_block xpc_die_notifier = {
|
||||
.notifier_call = xpc_system_die,
|
||||
};
|
||||
|
||||
int (*xpc_setup_partitions_sn) (void);
|
||||
enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *buf, u64 *cookie,
|
||||
unsigned long *rp_pa,
|
||||
size_t *len);
|
||||
enum xp_retval (*xpc_rsvd_page_init) (struct xpc_rsvd_page *rp);
|
||||
int (*xpc_setup_rsvd_page_sn) (struct xpc_rsvd_page *rp);
|
||||
void (*xpc_heartbeat_init) (void);
|
||||
void (*xpc_heartbeat_exit) (void);
|
||||
void (*xpc_increment_heartbeat) (void);
|
||||
@@ -183,8 +185,8 @@ enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *part);
|
||||
enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *part);
|
||||
void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *ch);
|
||||
u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *part);
|
||||
enum xp_retval (*xpc_allocate_msgqueues) (struct xpc_channel *ch);
|
||||
void (*xpc_free_msgqueues) (struct xpc_channel *ch);
|
||||
enum xp_retval (*xpc_setup_msg_structures) (struct xpc_channel *ch);
|
||||
void (*xpc_teardown_msg_structures) (struct xpc_channel *ch);
|
||||
void (*xpc_process_msg_chctl_flags) (struct xpc_partition *part, int ch_number);
|
||||
int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *ch);
|
||||
struct xpc_msg *(*xpc_get_deliverable_msg) (struct xpc_channel *ch);
|
||||
@@ -196,9 +198,9 @@ void (*xpc_request_partition_reactivation) (struct xpc_partition *part);
|
||||
void (*xpc_request_partition_deactivation) (struct xpc_partition *part);
|
||||
void (*xpc_cancel_partition_deactivation_request) (struct xpc_partition *part);
|
||||
|
||||
void (*xpc_process_activate_IRQ_rcvd) (int n_IRQs_expected);
|
||||
enum xp_retval (*xpc_setup_infrastructure) (struct xpc_partition *part);
|
||||
void (*xpc_teardown_infrastructure) (struct xpc_partition *part);
|
||||
void (*xpc_process_activate_IRQ_rcvd) (void);
|
||||
enum xp_retval (*xpc_setup_ch_structures_sn) (struct xpc_partition *part);
|
||||
void (*xpc_teardown_ch_structures_sn) (struct xpc_partition *part);
|
||||
|
||||
void (*xpc_indicate_partition_engaged) (struct xpc_partition *part);
|
||||
int (*xpc_partition_engaged) (short partid);
|
||||
@@ -215,6 +217,9 @@ void (*xpc_send_chctl_openrequest) (struct xpc_channel *ch,
|
||||
void (*xpc_send_chctl_openreply) (struct xpc_channel *ch,
|
||||
unsigned long *irq_flags);
|
||||
|
||||
void (*xpc_save_remote_msgqueue_pa) (struct xpc_channel *ch,
|
||||
unsigned long msgqueue_pa);
|
||||
|
||||
enum xp_retval (*xpc_send_msg) (struct xpc_channel *ch, u32 flags,
|
||||
void *payload, u16 payload_size, u8 notify_type,
|
||||
xpc_notify_func func, void *key);
|
||||
@@ -308,8 +313,6 @@ xpc_check_remote_hb(void)
|
||||
static int
|
||||
xpc_hb_checker(void *ignore)
|
||||
{
|
||||
int last_IRQ_count = 0;
|
||||
int new_IRQ_count;
|
||||
int force_IRQ = 0;
|
||||
|
||||
/* this thread was marked active by xpc_hb_init() */
|
||||
@@ -325,43 +328,37 @@ xpc_hb_checker(void *ignore)
|
||||
dev_dbg(xpc_part, "woke up with %d ticks rem; %d IRQs have "
|
||||
"been received\n",
|
||||
(int)(xpc_hb_check_timeout - jiffies),
|
||||
atomic_read(&xpc_activate_IRQ_rcvd) - last_IRQ_count);
|
||||
xpc_activate_IRQ_rcvd);
|
||||
|
||||
/* checking of remote heartbeats is skewed by IRQ handling */
|
||||
if (time_is_before_eq_jiffies(xpc_hb_check_timeout)) {
|
||||
xpc_hb_check_timeout = jiffies +
|
||||
(xpc_hb_check_interval * HZ);
|
||||
|
||||
dev_dbg(xpc_part, "checking remote heartbeats\n");
|
||||
xpc_check_remote_hb();
|
||||
|
||||
/*
|
||||
* We need to periodically recheck to ensure no
|
||||
* IRQ/amo pairs have been missed. That check
|
||||
* must always reset xpc_hb_check_timeout.
|
||||
* On sn2 we need to periodically recheck to ensure no
|
||||
* IRQ/amo pairs have been missed.
|
||||
*/
|
||||
force_IRQ = 1;
|
||||
if (is_shub())
|
||||
force_IRQ = 1;
|
||||
}
|
||||
|
||||
/* check for outstanding IRQs */
|
||||
new_IRQ_count = atomic_read(&xpc_activate_IRQ_rcvd);
|
||||
if (last_IRQ_count < new_IRQ_count || force_IRQ != 0) {
|
||||
if (xpc_activate_IRQ_rcvd > 0 || force_IRQ != 0) {
|
||||
force_IRQ = 0;
|
||||
|
||||
dev_dbg(xpc_part, "found an IRQ to process; will be "
|
||||
"resetting xpc_hb_check_timeout\n");
|
||||
|
||||
xpc_process_activate_IRQ_rcvd(new_IRQ_count -
|
||||
last_IRQ_count);
|
||||
last_IRQ_count = new_IRQ_count;
|
||||
|
||||
xpc_hb_check_timeout = jiffies +
|
||||
(xpc_hb_check_interval * HZ);
|
||||
dev_dbg(xpc_part, "processing activate IRQs "
|
||||
"received\n");
|
||||
xpc_process_activate_IRQ_rcvd();
|
||||
}
|
||||
|
||||
/* wait for IRQ or timeout */
|
||||
(void)wait_event_interruptible(xpc_activate_IRQ_wq,
|
||||
(last_IRQ_count < atomic_read(
|
||||
&xpc_activate_IRQ_rcvd)
|
||||
|| time_is_before_eq_jiffies(
|
||||
(time_is_before_eq_jiffies(
|
||||
xpc_hb_check_timeout) ||
|
||||
xpc_activate_IRQ_rcvd > 0 ||
|
||||
xpc_exiting));
|
||||
}
|
||||
|
||||
@@ -436,6 +433,153 @@ xpc_channel_mgr(struct xpc_partition *part)
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Guarantee that the kzalloc'd memory is cacheline aligned.
|
||||
*/
|
||||
void *
|
||||
xpc_kzalloc_cacheline_aligned(size_t size, gfp_t flags, void **base)
|
||||
{
|
||||
/* see if kzalloc will give us cachline aligned memory by default */
|
||||
*base = kzalloc(size, flags);
|
||||
if (*base == NULL)
|
||||
return NULL;
|
||||
|
||||
if ((u64)*base == L1_CACHE_ALIGN((u64)*base))
|
||||
return *base;
|
||||
|
||||
kfree(*base);
|
||||
|
||||
/* nope, we'll have to do it ourselves */
|
||||
*base = kzalloc(size + L1_CACHE_BYTES, flags);
|
||||
if (*base == NULL)
|
||||
return NULL;
|
||||
|
||||
return (void *)L1_CACHE_ALIGN((u64)*base);
|
||||
}
|
||||
|
||||
/*
|
||||
* Setup the channel structures necessary to support XPartition Communication
|
||||
* between the specified remote partition and the local one.
|
||||
*/
|
||||
static enum xp_retval
|
||||
xpc_setup_ch_structures(struct xpc_partition *part)
|
||||
{
|
||||
enum xp_retval ret;
|
||||
int ch_number;
|
||||
struct xpc_channel *ch;
|
||||
short partid = XPC_PARTID(part);
|
||||
|
||||
/*
|
||||
* Allocate all of the channel structures as a contiguous chunk of
|
||||
* memory.
|
||||
*/
|
||||
DBUG_ON(part->channels != NULL);
|
||||
part->channels = kzalloc(sizeof(struct xpc_channel) * XPC_MAX_NCHANNELS,
|
||||
GFP_KERNEL);
|
||||
if (part->channels == NULL) {
|
||||
dev_err(xpc_chan, "can't get memory for channels\n");
|
||||
return xpNoMemory;
|
||||
}
|
||||
|
||||
/* allocate the remote open and close args */
|
||||
|
||||
part->remote_openclose_args =
|
||||
xpc_kzalloc_cacheline_aligned(XPC_OPENCLOSE_ARGS_SIZE,
|
||||
GFP_KERNEL, &part->
|
||||
remote_openclose_args_base);
|
||||
if (part->remote_openclose_args == NULL) {
|
||||
dev_err(xpc_chan, "can't get memory for remote connect args\n");
|
||||
ret = xpNoMemory;
|
||||
goto out_1;
|
||||
}
|
||||
|
||||
part->chctl.all_flags = 0;
|
||||
spin_lock_init(&part->chctl_lock);
|
||||
|
||||
atomic_set(&part->channel_mgr_requests, 1);
|
||||
init_waitqueue_head(&part->channel_mgr_wq);
|
||||
|
||||
part->nchannels = XPC_MAX_NCHANNELS;
|
||||
|
||||
atomic_set(&part->nchannels_active, 0);
|
||||
atomic_set(&part->nchannels_engaged, 0);
|
||||
|
||||
for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
|
||||
ch = &part->channels[ch_number];
|
||||
|
||||
ch->partid = partid;
|
||||
ch->number = ch_number;
|
||||
ch->flags = XPC_C_DISCONNECTED;
|
||||
|
||||
atomic_set(&ch->kthreads_assigned, 0);
|
||||
atomic_set(&ch->kthreads_idle, 0);
|
||||
atomic_set(&ch->kthreads_active, 0);
|
||||
|
||||
atomic_set(&ch->references, 0);
|
||||
atomic_set(&ch->n_to_notify, 0);
|
||||
|
||||
spin_lock_init(&ch->lock);
|
||||
init_completion(&ch->wdisconnect_wait);
|
||||
|
||||
atomic_set(&ch->n_on_msg_allocate_wq, 0);
|
||||
init_waitqueue_head(&ch->msg_allocate_wq);
|
||||
init_waitqueue_head(&ch->idle_wq);
|
||||
}
|
||||
|
||||
ret = xpc_setup_ch_structures_sn(part);
|
||||
if (ret != xpSuccess)
|
||||
goto out_2;
|
||||
|
||||
/*
|
||||
* With the setting of the partition setup_state to XPC_P_SS_SETUP,
|
||||
* we're declaring that this partition is ready to go.
|
||||
*/
|
||||
part->setup_state = XPC_P_SS_SETUP;
|
||||
|
||||
return xpSuccess;
|
||||
|
||||
/* setup of ch structures failed */
|
||||
out_2:
|
||||
kfree(part->remote_openclose_args_base);
|
||||
part->remote_openclose_args = NULL;
|
||||
out_1:
|
||||
kfree(part->channels);
|
||||
part->channels = NULL;
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Teardown the channel structures necessary to support XPartition Communication
|
||||
* between the specified remote partition and the local one.
|
||||
*/
|
||||
static void
|
||||
xpc_teardown_ch_structures(struct xpc_partition *part)
|
||||
{
|
||||
DBUG_ON(atomic_read(&part->nchannels_engaged) != 0);
|
||||
DBUG_ON(atomic_read(&part->nchannels_active) != 0);
|
||||
|
||||
/*
|
||||
* Make this partition inaccessible to local processes by marking it
|
||||
* as no longer setup. Then wait before proceeding with the teardown
|
||||
* until all existing references cease.
|
||||
*/
|
||||
DBUG_ON(part->setup_state != XPC_P_SS_SETUP);
|
||||
part->setup_state = XPC_P_SS_WTEARDOWN;
|
||||
|
||||
wait_event(part->teardown_wq, (atomic_read(&part->references) == 0));
|
||||
|
||||
/* now we can begin tearing down the infrastructure */
|
||||
|
||||
xpc_teardown_ch_structures_sn(part);
|
||||
|
||||
kfree(part->remote_openclose_args_base);
|
||||
part->remote_openclose_args = NULL;
|
||||
kfree(part->channels);
|
||||
part->channels = NULL;
|
||||
|
||||
part->setup_state = XPC_P_SS_TORNDOWN;
|
||||
}
|
||||
|
||||
/*
|
||||
* When XPC HB determines that a partition has come up, it will create a new
|
||||
* kthread and that kthread will call this function to attempt to set up the
|
||||
@@ -476,7 +620,7 @@ xpc_activating(void *__partid)
|
||||
|
||||
xpc_allow_hb(partid);
|
||||
|
||||
if (xpc_setup_infrastructure(part) == xpSuccess) {
|
||||
if (xpc_setup_ch_structures(part) == xpSuccess) {
|
||||
(void)xpc_part_ref(part); /* this will always succeed */
|
||||
|
||||
if (xpc_make_first_contact(part) == xpSuccess) {
|
||||
@@ -486,7 +630,7 @@ xpc_activating(void *__partid)
|
||||
}
|
||||
|
||||
xpc_part_deref(part);
|
||||
xpc_teardown_infrastructure(part);
|
||||
xpc_teardown_ch_structures(part);
|
||||
}
|
||||
|
||||
xpc_disallow_hb(partid);
|
||||
@@ -806,6 +950,56 @@ xpc_disconnect_wait(int ch_number)
|
||||
}
|
||||
}
|
||||
|
||||
static int
|
||||
xpc_setup_partitions(void)
|
||||
{
|
||||
short partid;
|
||||
struct xpc_partition *part;
|
||||
|
||||
xpc_partitions = kzalloc(sizeof(struct xpc_partition) *
|
||||
xp_max_npartitions, GFP_KERNEL);
|
||||
if (xpc_partitions == NULL) {
|
||||
dev_err(xpc_part, "can't get memory for partition structure\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
/*
|
||||
* The first few fields of each entry of xpc_partitions[] need to
|
||||
* be initialized now so that calls to xpc_connect() and
|
||||
* xpc_disconnect() can be made prior to the activation of any remote
|
||||
* partition. NOTE THAT NONE OF THE OTHER FIELDS BELONGING TO THESE
|
||||
* ENTRIES ARE MEANINGFUL UNTIL AFTER AN ENTRY'S CORRESPONDING
|
||||
* PARTITION HAS BEEN ACTIVATED.
|
||||
*/
|
||||
for (partid = 0; partid < xp_max_npartitions; partid++) {
|
||||
part = &xpc_partitions[partid];
|
||||
|
||||
DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part));
|
||||
|
||||
part->activate_IRQ_rcvd = 0;
|
||||
spin_lock_init(&part->act_lock);
|
||||
part->act_state = XPC_P_AS_INACTIVE;
|
||||
XPC_SET_REASON(part, 0, 0);
|
||||
|
||||
init_timer(&part->disengage_timer);
|
||||
part->disengage_timer.function =
|
||||
xpc_timeout_partition_disengage;
|
||||
part->disengage_timer.data = (unsigned long)part;
|
||||
|
||||
part->setup_state = XPC_P_SS_UNSET;
|
||||
init_waitqueue_head(&part->teardown_wq);
|
||||
atomic_set(&part->references, 0);
|
||||
}
|
||||
|
||||
return xpc_setup_partitions_sn();
|
||||
}
|
||||
|
||||
static void
|
||||
xpc_teardown_partitions(void)
|
||||
{
|
||||
kfree(xpc_partitions);
|
||||
}
|
||||
|
||||
static void
|
||||
xpc_do_exit(enum xp_retval reason)
|
||||
{
|
||||
@@ -892,8 +1086,7 @@ xpc_do_exit(enum xp_retval reason)
|
||||
DBUG_ON(xpc_any_partition_engaged());
|
||||
DBUG_ON(xpc_any_hbs_allowed() != 0);
|
||||
|
||||
/* a zero timestamp indicates our rsvd page is not initialized */
|
||||
xpc_rsvd_page->ts_jiffies = 0;
|
||||
xpc_teardown_rsvd_page();
|
||||
|
||||
if (reason == xpUnloading) {
|
||||
(void)unregister_die_notifier(&xpc_die_notifier);
|
||||
@@ -906,7 +1099,7 @@ xpc_do_exit(enum xp_retval reason)
|
||||
if (xpc_sysctl)
|
||||
unregister_sysctl_table(xpc_sysctl);
|
||||
|
||||
kfree(xpc_partitions);
|
||||
xpc_teardown_partitions();
|
||||
|
||||
if (is_shub())
|
||||
xpc_exit_sn2();
|
||||
@@ -1062,8 +1255,6 @@ int __init
|
||||
xpc_init(void)
|
||||
{
|
||||
int ret;
|
||||
short partid;
|
||||
struct xpc_partition *part;
|
||||
struct task_struct *kthread;
|
||||
|
||||
snprintf(xpc_part->bus_id, BUS_ID_SIZE, "part");
|
||||
@@ -1076,56 +1267,29 @@ xpc_init(void)
|
||||
* further to only support exactly 64 partitions on this
|
||||
* architecture, no less.
|
||||
*/
|
||||
if (xp_max_npartitions != 64)
|
||||
return -EINVAL;
|
||||
|
||||
ret = xpc_init_sn2();
|
||||
if (ret != 0)
|
||||
return ret;
|
||||
if (xp_max_npartitions != 64) {
|
||||
dev_err(xpc_part, "max #of partitions not set to 64\n");
|
||||
ret = -EINVAL;
|
||||
} else {
|
||||
ret = xpc_init_sn2();
|
||||
}
|
||||
|
||||
} else if (is_uv()) {
|
||||
xpc_init_uv();
|
||||
ret = xpc_init_uv();
|
||||
|
||||
} else {
|
||||
return -ENODEV;
|
||||
ret = -ENODEV;
|
||||
}
|
||||
|
||||
xpc_partitions = kzalloc(sizeof(struct xpc_partition) *
|
||||
xp_max_npartitions, GFP_KERNEL);
|
||||
if (xpc_partitions == NULL) {
|
||||
if (ret != 0)
|
||||
return ret;
|
||||
|
||||
ret = xpc_setup_partitions();
|
||||
if (ret != 0) {
|
||||
dev_err(xpc_part, "can't get memory for partition structure\n");
|
||||
ret = -ENOMEM;
|
||||
goto out_1;
|
||||
}
|
||||
|
||||
/*
|
||||
* The first few fields of each entry of xpc_partitions[] need to
|
||||
* be initialized now so that calls to xpc_connect() and
|
||||
* xpc_disconnect() can be made prior to the activation of any remote
|
||||
* partition. NOTE THAT NONE OF THE OTHER FIELDS BELONGING TO THESE
|
||||
* ENTRIES ARE MEANINGFUL UNTIL AFTER AN ENTRY'S CORRESPONDING
|
||||
* PARTITION HAS BEEN ACTIVATED.
|
||||
*/
|
||||
for (partid = 0; partid < xp_max_npartitions; partid++) {
|
||||
part = &xpc_partitions[partid];
|
||||
|
||||
DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part));
|
||||
|
||||
part->activate_IRQ_rcvd = 0;
|
||||
spin_lock_init(&part->act_lock);
|
||||
part->act_state = XPC_P_AS_INACTIVE;
|
||||
XPC_SET_REASON(part, 0, 0);
|
||||
|
||||
init_timer(&part->disengage_timer);
|
||||
part->disengage_timer.function =
|
||||
xpc_timeout_partition_disengage;
|
||||
part->disengage_timer.data = (unsigned long)part;
|
||||
|
||||
part->setup_state = XPC_P_SS_UNSET;
|
||||
init_waitqueue_head(&part->teardown_wq);
|
||||
atomic_set(&part->references, 0);
|
||||
}
|
||||
|
||||
xpc_sysctl = register_sysctl_table(xpc_sys_dir);
|
||||
|
||||
/*
|
||||
@@ -1133,10 +1297,9 @@ xpc_init(void)
|
||||
* other partitions to discover we are alive and establish initial
|
||||
* communications.
|
||||
*/
|
||||
xpc_rsvd_page = xpc_setup_rsvd_page();
|
||||
if (xpc_rsvd_page == NULL) {
|
||||
ret = xpc_setup_rsvd_page();
|
||||
if (ret != 0) {
|
||||
dev_err(xpc_part, "can't setup our reserved page\n");
|
||||
ret = -EBUSY;
|
||||
goto out_2;
|
||||
}
|
||||
|
||||
@@ -1187,15 +1350,15 @@ xpc_init(void)
|
||||
|
||||
/* initialization was not successful */
|
||||
out_3:
|
||||
/* a zero timestamp indicates our rsvd page is not initialized */
|
||||
xpc_rsvd_page->ts_jiffies = 0;
|
||||
xpc_teardown_rsvd_page();
|
||||
|
||||
(void)unregister_die_notifier(&xpc_die_notifier);
|
||||
(void)unregister_reboot_notifier(&xpc_reboot_notifier);
|
||||
out_2:
|
||||
if (xpc_sysctl)
|
||||
unregister_sysctl_table(xpc_sysctl);
|
||||
kfree(xpc_partitions);
|
||||
|
||||
xpc_teardown_partitions();
|
||||
out_1:
|
||||
if (is_shub())
|
||||
xpc_exit_sn2();
|
||||
|
||||
@@ -73,6 +73,12 @@ xpc_get_rsvd_page_pa(int nasid)
|
||||
|
||||
while (1) {
|
||||
|
||||
/* !!! rp_pa will need to be _gpa on UV.
|
||||
* ??? So do we save it into the architecture specific parts
|
||||
* ??? of the xpc_partition structure? Do we rename this
|
||||
* ??? function or have two versions? Rename rp_pa for UV to
|
||||
* ??? rp_gpa?
|
||||
*/
|
||||
ret = xpc_get_partition_rsvd_page_pa(buf, &cookie, &rp_pa,
|
||||
&len);
|
||||
|
||||
@@ -118,9 +124,10 @@ xpc_get_rsvd_page_pa(int nasid)
|
||||
* other partitions to discover we are alive and establish initial
|
||||
* communications.
|
||||
*/
|
||||
struct xpc_rsvd_page *
|
||||
int
|
||||
xpc_setup_rsvd_page(void)
|
||||
{
|
||||
int ret;
|
||||
struct xpc_rsvd_page *rp;
|
||||
unsigned long rp_pa;
|
||||
unsigned long new_ts_jiffies;
|
||||
@@ -132,7 +139,7 @@ xpc_setup_rsvd_page(void)
|
||||
preempt_enable();
|
||||
if (rp_pa == 0) {
|
||||
dev_err(xpc_part, "SAL failed to locate the reserved page\n");
|
||||
return NULL;
|
||||
return -ESRCH;
|
||||
}
|
||||
rp = (struct xpc_rsvd_page *)__va(rp_pa);
|
||||
|
||||
@@ -146,7 +153,7 @@ xpc_setup_rsvd_page(void)
|
||||
dev_err(xpc_part, "the reserved page's partid of %d is outside "
|
||||
"supported range (< 0 || >= %d)\n", rp->SAL_partid,
|
||||
xp_max_npartitions);
|
||||
return NULL;
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
rp->version = XPC_RP_VERSION;
|
||||
@@ -165,8 +172,9 @@ xpc_setup_rsvd_page(void)
|
||||
xpc_part_nasids = XPC_RP_PART_NASIDS(rp);
|
||||
xpc_mach_nasids = XPC_RP_MACH_NASIDS(rp);
|
||||
|
||||
if (xpc_rsvd_page_init(rp) != xpSuccess)
|
||||
return NULL;
|
||||
ret = xpc_setup_rsvd_page_sn(rp);
|
||||
if (ret != 0)
|
||||
return ret;
|
||||
|
||||
/*
|
||||
* Set timestamp of when reserved page was setup by XPC.
|
||||
@@ -178,7 +186,15 @@ xpc_setup_rsvd_page(void)
|
||||
new_ts_jiffies++;
|
||||
rp->ts_jiffies = new_ts_jiffies;
|
||||
|
||||
return rp;
|
||||
xpc_rsvd_page = rp;
|
||||
return 0;
|
||||
}
|
||||
|
||||
void
|
||||
xpc_teardown_rsvd_page(void)
|
||||
{
|
||||
/* a zero timestamp indicates our rsvd page is not initialized */
|
||||
xpc_rsvd_page->ts_jiffies = 0;
|
||||
}
|
||||
|
||||
/*
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
Reference in New Issue
Block a user