ACK: [PATCH Trusty SRU] Revert "ipmi: simplify locking"
Chris J Arges
chris.j.arges at canonical.com
Tue Nov 18 21:24:17 UTC 2014
I think this is the safest option for fixing this in 3.13.
--chris
On 11/18/2014 02:43 PM, tim.gardner at canonical.com wrote:
> From: Tim Gardner <tim.gardner at canonical.com>
>
> BugLink: http://bugs.launchpad.net/bugs/1383921
>
> This reverts commit f60adf42ad55405d1b17e9e5c33fdb63f1eb8861.
>
> This commit was originally intended as a code improvement. However,
> it appears to have either opened a locking race (which I cannot find), or
> has abused HW in such a way as to cause it to stop. Given that this was
> intended as an improvemnet, the most expedient approach appears to be to
> revert it until we develop an upstream solution.
>
> Signed-off-by: Tim Gardner <tim.gardner at canonical.com>
> ---
> drivers/char/ipmi/ipmi_si_intf.c | 54 ++++++++++++++++++++++++----------------
> 1 file changed, 33 insertions(+), 21 deletions(-)
>
> diff --git a/drivers/char/ipmi/ipmi_si_intf.c b/drivers/char/ipmi/ipmi_si_intf.c
> index 4afc8b4..8c0ffdd 100644
> --- a/drivers/char/ipmi/ipmi_si_intf.c
> +++ b/drivers/char/ipmi/ipmi_si_intf.c
> @@ -175,6 +175,7 @@ struct smi_info {
> struct si_sm_handlers *handlers;
> enum si_type si_type;
> spinlock_t si_lock;
> + spinlock_t msg_lock;
> struct list_head xmit_msgs;
> struct list_head hp_xmit_msgs;
> struct ipmi_smi_msg *curr_msg;
> @@ -356,6 +357,13 @@ static enum si_sm_result start_next_msg(struct smi_info *smi_info)
> struct timeval t;
> #endif
>
> + /*
> + * No need to save flags, we aleady have interrupts off and we
> + * already hold the SMI lock.
> + */
> + if (!smi_info->run_to_completion)
> + spin_lock(&(smi_info->msg_lock));
> +
> /* Pick the high priority queue first. */
> if (!list_empty(&(smi_info->hp_xmit_msgs))) {
> entry = smi_info->hp_xmit_msgs.next;
> @@ -393,6 +401,9 @@ static enum si_sm_result start_next_msg(struct smi_info *smi_info)
> rv = SI_SM_CALL_WITHOUT_DELAY;
> }
> out:
> + if (!smi_info->run_to_completion)
> + spin_unlock(&(smi_info->msg_lock));
> +
> return rv;
> }
>
> @@ -879,6 +890,19 @@ static void sender(void *send_info,
> printk("**Enqueue: %d.%9.9d\n", t.tv_sec, t.tv_usec);
> #endif
>
> + /*
> + * last_timeout_jiffies is updated here to avoid
> + * smi_timeout() handler passing very large time_diff
> + * value to smi_event_handler() that causes
> + * the send command to abort.
> + */
> + smi_info->last_timeout_jiffies = jiffies;
> +
> + mod_timer(&smi_info->si_timer, jiffies + SI_TIMEOUT_JIFFIES);
> +
> + if (smi_info->thread)
> + wake_up_process(smi_info->thread);
> +
> if (smi_info->run_to_completion) {
> /*
> * If we are running to completion, then throw it in
> @@ -901,26 +925,15 @@ static void sender(void *send_info,
> return;
> }
>
> - spin_lock_irqsave(&smi_info->si_lock, flags);
> + spin_lock_irqsave(&smi_info->msg_lock, flags);
> if (priority > 0)
> list_add_tail(&msg->link, &smi_info->hp_xmit_msgs);
> else
> list_add_tail(&msg->link, &smi_info->xmit_msgs);
> + spin_unlock_irqrestore(&smi_info->msg_lock, flags);
>
> + spin_lock_irqsave(&smi_info->si_lock, flags);
> if (smi_info->si_state == SI_NORMAL && smi_info->curr_msg == NULL) {
> - /*
> - * last_timeout_jiffies is updated here to avoid
> - * smi_timeout() handler passing very large time_diff
> - * value to smi_event_handler() that causes
> - * the send command to abort.
> - */
> - smi_info->last_timeout_jiffies = jiffies;
> -
> - mod_timer(&smi_info->si_timer, jiffies + SI_TIMEOUT_JIFFIES);
> -
> - if (smi_info->thread)
> - wake_up_process(smi_info->thread);
> -
> start_next_msg(smi_info);
> smi_event_handler(smi_info, 0);
> }
> @@ -1024,19 +1037,16 @@ static int ipmi_thread(void *data)
> static void poll(void *send_info)
> {
> struct smi_info *smi_info = send_info;
> - unsigned long flags = 0;
> - int run_to_completion = smi_info->run_to_completion;
> + unsigned long flags;
>
> /*
> * Make sure there is some delay in the poll loop so we can
> * drive time forward and timeout things.
> */
> udelay(10);
> - if (!run_to_completion)
> - spin_lock_irqsave(&smi_info->si_lock, flags);
> + spin_lock_irqsave(&smi_info->si_lock, flags);
> smi_event_handler(smi_info, 10);
> - if (!run_to_completion)
> - spin_unlock_irqrestore(&smi_info->si_lock, flags);
> + spin_unlock_irqrestore(&smi_info->si_lock, flags);
> }
>
> static void request_events(void *send_info)
> @@ -1702,8 +1712,10 @@ static struct smi_info *smi_info_alloc(void)
> {
> struct smi_info *info = kzalloc(sizeof(*info), GFP_KERNEL);
>
> - if (info)
> + if (info) {
> spin_lock_init(&info->si_lock);
> + spin_lock_init(&info->msg_lock);
> + }
> return info;
> }
>
>
More information about the kernel-team
mailing list