[PATCH 7/7] powerpc/64s: idle POWER8 avoid full state loss recovery where possible
Gautham R Shenoy
ego at linux.vnet.ibm.com
Mon Mar 20 21:11:39 AEDT 2017
Hi Nick,
On Mon, Mar 20, 2017 at 04:01:52PM +1000, Nicholas Piggin wrote:
> If not all threads were in winkle, full state loss recovery is not
> necessary and can be avoided. A previous patch removed this optimisation
> due to some complexity with the implementation. Re-implement it by
> counting the number of threads in winkle with the per-core idle state.
> Only restore full state loss if all threads were in winkle.
>
> This has a small window of false positives right before threads execute
> winkle and just after they wake up, when the winkle count does not
> reflect the true number of threads in winkle. This is not a significant
> problem in comparison with even the minimum winkle duration. For
> correctness, a false positive is not a problem (only false negatives
> would be).
The patch looks good. Just a minor comment.
> BEGIN_FTR_SECTION
> + /*
> + * Were we in winkle?
> + * If yes, check if all threads were in winkle, decrement our
> + * winkle count, set all thread winkle bits if all were in winkle.
> + * Check if our thread has a winkle bit set, and set cr4 accordingly
> + * (to match ISA300, above). Pseudo-code for core idle state
> + * transitions for ISA207 is as follows (everything happens atomically
> + * due to store conditional and/or lock bit):
> + *
> + * nap_idle() { }
> + * nap_wake() { }
> + *
> + * sleep_idle()
> + * {
> + * core_idle_state &= ~thread_in_core
> + * }
> + *
> + * sleep_wake()
> + * {
> + * bool first_in_core, first_in_subcore;
> + *
> + * first_in_core = (core_idle_state & IDLE_THREAD_BITS) == 0;
> + * first_in_subcore = (core_idle_state & SUBCORE_SIBLING_MASK) == 0;
> + *
> + * core_idle_state |= thread_in_core;
> + * }
> + *
> + * winkle_idle()
> + * {
> + * core_idle_state &= ~thread_in_core;
> + * core_idle_state += 1 << WINKLE_COUNT_SHIFT;
> + * }
> + *
> + * winkle_wake()
> + * {
> + * bool first_in_core, first_in_subcore, winkle_state_lost;
> + *
> + * first_in_core = (core_idle_state & IDLE_THREAD_BITS) == 0;
> + * first_in_subcore = (core_idle_state & SUBCORE_SIBLING_MASK) == 0;
> + *
> + * core_idle_state |= thread_in_core;
> + *
> + * if ((core_idle_state & WINKLE_MASK) == (8 << WINKLE_COUNT_SIHFT))
> + * core_idle_state |= THREAD_WINKLE_BITS;
We also do the following decrement. I forgot this in the pseudo-code in my
earlier reply.
core_idle_state -= 1 << WINKLE_COUNT_SHIFT;
> + * winkle_state_lost = core_idle_state &
> + * (thread_in_core << WINKLE_THREAD_SHIFT);
> + * core_idle_state &= ~(thread_in_core << WINKLE_THREAD_SHIFT);
> + * }
> + *
> + */
> + cmpwi r18,PNV_THREAD_WINKLE
> + bne 2f
> + andis. r9,r15,PNV_CORE_IDLE_WINKLE_COUNT_ALL_BIT at h
> + subis r15,r15,PNV_CORE_IDLE_WINKLE_COUNT at h
> + beq 2f
> + ori r15,r15,PNV_CORE_IDLE_THREAD_WINKLE_BITS /* all were winkle */
> +2:
> + /* Shift thread bit to winkle mask, then test if this thread is set,
> + * and remove it from the winkle bits */
> + slwi r8,r7,8
> + and r8,r8,r15
> + andc r15,r15,r8
> + cmpwi cr4,r8,1 /* cr4 will be gt if our bit is set, lt if not */
> +
Looks good other wise.
Reviewed-by: Gautham R. Shenoy <ego at linux.vnet.ibm.com>
> lbz r4,PACA_SUBCORE_SIBLING_MASK(r13)
> and r4,r4,r15
> cmpwi r4,0 /* Check if first in subcore */
> --
> 2.11.0
>
More information about the Linuxppc-dev
mailing list