diff mbox series

[v3,11/20] arm64: assembler: add macros to conditionally yield the NEON under PREEMPT

Message ID 20171206194346.24393-12-ard.biesheuvel@linaro.org
State New
Headers show
Series crypto: arm64 - play nice with CONFIG_PREEMPT | expand

Commit Message

Ard Biesheuvel Dec. 6, 2017, 7:43 p.m. UTC
Add support macros to conditionally yield the NEON (and thus the CPU)
that may be called from the assembler code.

In some cases, yielding the NEON involves saving and restoring a non
trivial amount of context (especially in the CRC folding algorithms),
and so the macro is split into three, and the code in between is only
executed when the yield path is taken, allowing the context to be preserved.
The third macro takes an optional label argument that marks the resume
path after a yield has been performed.

Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>

---
 arch/arm64/include/asm/assembler.h | 51 ++++++++++++++++++++
 1 file changed, 51 insertions(+)

-- 
2.11.0

Comments

Dave Martin Dec. 7, 2017, 2:39 p.m. UTC | #1
On Wed, Dec 06, 2017 at 07:43:37PM +0000, Ard Biesheuvel wrote:
> Add support macros to conditionally yield the NEON (and thus the CPU)

> that may be called from the assembler code.

> 

> In some cases, yielding the NEON involves saving and restoring a non

> trivial amount of context (especially in the CRC folding algorithms),

> and so the macro is split into three, and the code in between is only

> executed when the yield path is taken, allowing the context to be preserved.

> The third macro takes an optional label argument that marks the resume

> path after a yield has been performed.

> 

> Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>

> ---

>  arch/arm64/include/asm/assembler.h | 51 ++++++++++++++++++++

>  1 file changed, 51 insertions(+)

> 

> diff --git a/arch/arm64/include/asm/assembler.h b/arch/arm64/include/asm/assembler.h

> index 5f61487e9f93..c54e408fd5a7 100644

> --- a/arch/arm64/include/asm/assembler.h

> +++ b/arch/arm64/include/asm/assembler.h

> @@ -572,4 +572,55 @@ alternative_else_nop_endif

>  #endif

>  	.endm

>  

> +/*

> + * Check whether to yield to another runnable task from kernel mode NEON code

> + * (which runs with preemption disabled).

> + *

> + * if_will_cond_yield_neon

> + *        // pre-yield patchup code

> + * do_cond_yield_neon

> + *        // post-yield patchup code

> + * endif_yield_neon


^ Mention the lbl argument?

> + *

> + * - Check whether the preempt count is exactly 1, in which case disabling


                                                           enabling ^

> + *   preemption once will make the task preemptible. If this is not the case,

> + *   yielding is pointless.

> + * - Check whether TIF_NEED_RESCHED is set, and if so, disable and re-enable

> + *   kernel mode NEON (which will trigger a reschedule), and branch to the

> + *   yield fixup code.


Mention that neither patchup sequence is allowed to use section-changing
directives?

For example:

	if_will_cond_yield_neon
		// some code

		.pushsection .rodata, "a"
foo: 			.quad // some literal data for some reason
		.popsection

		// some code
	do_cond_yield_neon

is not safe, because .previous is now .rodata.

(You could protect against this with
	.pushsection .text; .previous; .subsection 1; // ...
	.popsection
but it may be overkill.)

> + *

> + * This macro sequence clobbers x0, x1 and the flags register unconditionally,

> + * and may clobber x2 .. x18 if the yield path is taken.

> + */

> +

> +	.macro		cond_yield_neon, lbl

> +	if_will_cond_yield_neon

> +	do_cond_yield_neon

> +	endif_yield_neon	\lbl

> +	.endm

> +

> +	.macro		if_will_cond_yield_neon

> +#ifdef CONFIG_PREEMPT

> +	get_thread_info	x0

> +	ldr		w1, [x0, #TSK_TI_PREEMPT]

> +	ldr		x0, [x0, #TSK_TI_FLAGS]

> +	cmp		w1, #1 // == PREEMPT_OFFSET


Can we at least drop a BUILD_BUG_ON() somewhere to check this?

Maybe in kernel_neon_begin() since this is intimately kernel-mode NEON
related.

> +	csel		x0, x0, xzr, eq

> +	tbnz		x0, #TIF_NEED_RESCHED, 5555f	// needs rescheduling?

> +#endif


A comment that we will fall through to 6666f here may be helpful.

> +	.subsection	1

> +5555:

> +	.endm

> +

> +	.macro		do_cond_yield_neon

> +	bl		kernel_neon_end

> +	bl		kernel_neon_begin

> +	.endm

> +

> +	.macro		endif_yield_neon, lbl=6666f

> +	b		\lbl

> +	.previous

> +6666:


Could have slightly more random "random" labels here, but otherwise
it looks ok to me.

I might go through and replace all the random labels with something
more robust sometime, but I've never been sure it was worth the
effort...

Cheers
---Dave
Ard Biesheuvel Dec. 7, 2017, 2:50 p.m. UTC | #2
On 7 December 2017 at 14:39, Dave Martin <Dave.Martin@arm.com> wrote:
> On Wed, Dec 06, 2017 at 07:43:37PM +0000, Ard Biesheuvel wrote:

>> Add support macros to conditionally yield the NEON (and thus the CPU)

>> that may be called from the assembler code.

>>

>> In some cases, yielding the NEON involves saving and restoring a non

>> trivial amount of context (especially in the CRC folding algorithms),

>> and so the macro is split into three, and the code in between is only

>> executed when the yield path is taken, allowing the context to be preserved.

>> The third macro takes an optional label argument that marks the resume

>> path after a yield has been performed.

>>

>> Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>

>> ---

>>  arch/arm64/include/asm/assembler.h | 51 ++++++++++++++++++++

>>  1 file changed, 51 insertions(+)

>>

>> diff --git a/arch/arm64/include/asm/assembler.h b/arch/arm64/include/asm/assembler.h

>> index 5f61487e9f93..c54e408fd5a7 100644

>> --- a/arch/arm64/include/asm/assembler.h

>> +++ b/arch/arm64/include/asm/assembler.h

>> @@ -572,4 +572,55 @@ alternative_else_nop_endif

>>  #endif

>>       .endm

>>

>> +/*

>> + * Check whether to yield to another runnable task from kernel mode NEON code

>> + * (which runs with preemption disabled).

>> + *

>> + * if_will_cond_yield_neon

>> + *        // pre-yield patchup code

>> + * do_cond_yield_neon

>> + *        // post-yield patchup code

>> + * endif_yield_neon

>

> ^ Mention the lbl argument?

>


Yep will do

>> + *

>> + * - Check whether the preempt count is exactly 1, in which case disabling

>

>                                                            enabling ^

>

>> + *   preemption once will make the task preemptible. If this is not the case,

>> + *   yielding is pointless.

>> + * - Check whether TIF_NEED_RESCHED is set, and if so, disable and re-enable

>> + *   kernel mode NEON (which will trigger a reschedule), and branch to the

>> + *   yield fixup code.

>

> Mention that neither patchup sequence is allowed to use section-changing

> directives?

>

> For example:

>

>         if_will_cond_yield_neon

>                 // some code

>

>                 .pushsection .rodata, "a"

> foo:                    .quad // some literal data for some reason

>                 .popsection

>

>                 // some code

>         do_cond_yield_neon

>

> is not safe, because .previous is now .rodata.

>


Are you sure this is true?

The gas info page for .previous tells me

   In terms of the section stack, this directive swaps the current
section with the top section on the section stack.

and it seems to me that .rodata is no longer on the section stack
after .popsection. In that sense, push/pop should be safe, but
section/subsection/previous is not (I think). So yes, let's put a note
in to mention that section directives are unsupported.

> (You could protect against this with

>         .pushsection .text; .previous; .subsection 1; // ...

>         .popsection

> but it may be overkill.)

>

>> + *

>> + * This macro sequence clobbers x0, x1 and the flags register unconditionally,

>> + * and may clobber x2 .. x18 if the yield path is taken.

>> + */

>> +

>> +     .macro          cond_yield_neon, lbl

>> +     if_will_cond_yield_neon

>> +     do_cond_yield_neon

>> +     endif_yield_neon        \lbl

>> +     .endm

>> +

>> +     .macro          if_will_cond_yield_neon

>> +#ifdef CONFIG_PREEMPT

>> +     get_thread_info x0

>> +     ldr             w1, [x0, #TSK_TI_PREEMPT]

>> +     ldr             x0, [x0, #TSK_TI_FLAGS]

>> +     cmp             w1, #1 // == PREEMPT_OFFSET

>

> Can we at least drop a BUILD_BUG_ON() somewhere to check this?

>

> Maybe in kernel_neon_begin() since this is intimately kernel-mode NEON

> related.

>


Sure.

>> +     csel            x0, x0, xzr, eq

>> +     tbnz            x0, #TIF_NEED_RESCHED, 5555f    // needs rescheduling?

>> +#endif

>

> A comment that we will fall through to 6666f here may be helpful.

>


Indeed. Will add that.

>> +     .subsection     1

>> +5555:

>> +     .endm

>> +

>> +     .macro          do_cond_yield_neon

>> +     bl              kernel_neon_end

>> +     bl              kernel_neon_begin

>> +     .endm

>> +

>> +     .macro          endif_yield_neon, lbl=6666f

>> +     b               \lbl

>> +     .previous

>> +6666:

>

> Could have slightly more random "random" labels here, but otherwise

> it looks ok to me.

>


Which number did you have in mind that is more random than 6666? :-)

> I might go through and replace all the random labels with something

> more robust sometime, but I've never been sure it was worth the

> effort...

>


I guess we could invent all kinds of elaborate schemes but as you say,
having 4 digit numbers and grep'ing the source before you add a new
one has been working fine so far, so I don't think it should be a
priority.
Ard Biesheuvel Dec. 7, 2017, 3:51 p.m. UTC | #3
On 7 December 2017 at 15:47, Ard Biesheuvel <ard.biesheuvel@linaro.org> wrote:
> On 7 December 2017 at 14:50, Ard Biesheuvel <ard.biesheuvel@linaro.org> wrote:

>> On 7 December 2017 at 14:39, Dave Martin <Dave.Martin@arm.com> wrote:

>>> On Wed, Dec 06, 2017 at 07:43:37PM +0000, Ard Biesheuvel wrote:

>>>> Add support macros to conditionally yield the NEON (and thus the CPU)

>>>> that may be called from the assembler code.

>>>>

>>>> In some cases, yielding the NEON involves saving and restoring a non

>>>> trivial amount of context (especially in the CRC folding algorithms),

>>>> and so the macro is split into three, and the code in between is only

>>>> executed when the yield path is taken, allowing the context to be preserved.

>>>> The third macro takes an optional label argument that marks the resume

>>>> path after a yield has been performed.

>>>>

>>>> Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>

>>>> ---

>>>>  arch/arm64/include/asm/assembler.h | 51 ++++++++++++++++++++

>>>>  1 file changed, 51 insertions(+)

>>>>

>>>> diff --git a/arch/arm64/include/asm/assembler.h b/arch/arm64/include/asm/assembler.h

>>>> index 5f61487e9f93..c54e408fd5a7 100644

>>>> --- a/arch/arm64/include/asm/assembler.h

>>>> +++ b/arch/arm64/include/asm/assembler.h

>>>> @@ -572,4 +572,55 @@ alternative_else_nop_endif

>>>>  #endif

>>>>       .endm

>>>>

>>>> +/*

>>>> + * Check whether to yield to another runnable task from kernel mode NEON code

>>>> + * (which runs with preemption disabled).

>>>> + *

>>>> + * if_will_cond_yield_neon

>>>> + *        // pre-yield patchup code

>>>> + * do_cond_yield_neon

>>>> + *        // post-yield patchup code

>>>> + * endif_yield_neon

>>>

>>> ^ Mention the lbl argument?

>>>

>>

>> Yep will do

>>

>>>> + *

>>>> + * - Check whether the preempt count is exactly 1, in which case disabling

>>>

>>>                                                            enabling ^

>>>

>>>> + *   preemption once will make the task preemptible. If this is not the case,

>>>> + *   yielding is pointless.

>>>> + * - Check whether TIF_NEED_RESCHED is set, and if so, disable and re-enable

>>>> + *   kernel mode NEON (which will trigger a reschedule), and branch to the

>>>> + *   yield fixup code.

>>>

>>> Mention that neither patchup sequence is allowed to use section-changing

>>> directives?

>>>

>>> For example:

>>>

>>>         if_will_cond_yield_neon

>>>                 // some code

>>>

>>>                 .pushsection .rodata, "a"

>>> foo:                    .quad // some literal data for some reason

>>>                 .popsection

>>>

>>>                 // some code

>>>         do_cond_yield_neon

>>>

>>> is not safe, because .previous is now .rodata.

>>>

>>

>> Are you sure this is true?

>>

>> The gas info page for .previous tells me

>>

>>    In terms of the section stack, this directive swaps the current

>> section with the top section on the section stack.

>>

>> and it seems to me that .rodata is no longer on the section stack

>> after .popsection. In that sense, push/pop should be safe, but

>> section/subsection/previous is not (I think). So yes, let's put a note

>> in to mention that section directives are unsupported.

>>

>>> (You could protect against this with

>>>         .pushsection .text; .previous; .subsection 1; // ...

>>>         .popsection

>>> but it may be overkill.)

>>>

>>>> + *

>>>> + * This macro sequence clobbers x0, x1 and the flags register unconditionally,

>>>> + * and may clobber x2 .. x18 if the yield path is taken.

>>>> + */

>>>> +

>>>> +     .macro          cond_yield_neon, lbl

>>>> +     if_will_cond_yield_neon

>>>> +     do_cond_yield_neon

>>>> +     endif_yield_neon        \lbl

>>>> +     .endm

>>>> +

>>>> +     .macro          if_will_cond_yield_neon

>>>> +#ifdef CONFIG_PREEMPT

>>>> +     get_thread_info x0

>>>> +     ldr             w1, [x0, #TSK_TI_PREEMPT]

>>>> +     ldr             x0, [x0, #TSK_TI_FLAGS]

>>>> +     cmp             w1, #1 // == PREEMPT_OFFSET

>>>

>>> Can we at least drop a BUILD_BUG_ON() somewhere to check this?

>>>

>>> Maybe in kernel_neon_begin() since this is intimately kernel-mode NEON

>>> related.

>>>

>>

>> Sure.

>>

>

> I only just understood your asm-offsets remark earlier. I wasn't aware

> that it allows exposing random constants as well (although it is

> fairly obvious now that I do). So I will expose PREEMPT_OFFSET rather

> than open code it

>


Of course, I mean 'arbitrary' not 'random' (like 6666)
Dave Martin Dec. 7, 2017, 4:11 p.m. UTC | #4
On Thu, Dec 07, 2017 at 02:50:11PM +0000, Ard Biesheuvel wrote:
> On 7 December 2017 at 14:39, Dave Martin <Dave.Martin@arm.com> wrote:

> > On Wed, Dec 06, 2017 at 07:43:37PM +0000, Ard Biesheuvel wrote:

> >> Add support macros to conditionally yield the NEON (and thus the CPU)

> >> that may be called from the assembler code.

> >>

> >> In some cases, yielding the NEON involves saving and restoring a non

> >> trivial amount of context (especially in the CRC folding algorithms),

> >> and so the macro is split into three, and the code in between is only

> >> executed when the yield path is taken, allowing the context to be preserved.

> >> The third macro takes an optional label argument that marks the resume

> >> path after a yield has been performed.

> >>

> >> Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>

> >> ---

> >>  arch/arm64/include/asm/assembler.h | 51 ++++++++++++++++++++

> >>  1 file changed, 51 insertions(+)

> >>

> >> diff --git a/arch/arm64/include/asm/assembler.h b/arch/arm64/include/asm/assembler.h

> >> index 5f61487e9f93..c54e408fd5a7 100644

> >> --- a/arch/arm64/include/asm/assembler.h

> >> +++ b/arch/arm64/include/asm/assembler.h

> >> @@ -572,4 +572,55 @@ alternative_else_nop_endif

> >>  #endif

> >>       .endm

> >>

> >> +/*

> >> + * Check whether to yield to another runnable task from kernel mode NEON code

> >> + * (which runs with preemption disabled).

> >> + *

> >> + * if_will_cond_yield_neon

> >> + *        // pre-yield patchup code

> >> + * do_cond_yield_neon

> >> + *        // post-yield patchup code

> >> + * endif_yield_neon

> >

> > ^ Mention the lbl argument?

> >

> 

> Yep will do

> 

> >> + *

> >> + * - Check whether the preempt count is exactly 1, in which case disabling

> >

> >                                                            enabling ^

> >

> >> + *   preemption once will make the task preemptible. If this is not the case,

> >> + *   yielding is pointless.

> >> + * - Check whether TIF_NEED_RESCHED is set, and if so, disable and re-enable

> >> + *   kernel mode NEON (which will trigger a reschedule), and branch to the

> >> + *   yield fixup code.

> >

> > Mention that neither patchup sequence is allowed to use section-changing

> > directives?

> >

> > For example:

> >

> >         if_will_cond_yield_neon

> >                 // some code

> >

> >                 .pushsection .rodata, "a"

> > foo:                    .quad // some literal data for some reason

> >                 .popsection

> >

> >                 // some code

> >         do_cond_yield_neon

> >

> > is not safe, because .previous is now .rodata.

> >

> 

> Are you sure this is true?

> 

> The gas info page for .previous tells me

> 

>    In terms of the section stack, this directive swaps the current

> section with the top section on the section stack.


That statement is either misleading or wrong, but the actual behaviour
doesn't seem straightforward either.


> and it seems to me that .rodata is no longer on the section stack

> after .popsection. In that sense, push/pop should be safe, but


My suggestion does seem to work here (I've used it in the past) but
it's probably best not to rely on it unnecessarily...  One would
have to read the gas code and get the docs fixed first.

> section/subsection/previous is not (I think). So yes, let's put a note

> in to mention that section directives are unsupported.


... here I'd agree: it doesn't seem justified relying on dubious
tricks here, since there's doubt about whether my suggestion is
really safe.

> 

> > (You could protect against this with

> >         .pushsection .text; .previous; .subsection 1; // ...

> >         .popsection

> > but it may be overkill.)

> >

> >> + *

> >> + * This macro sequence clobbers x0, x1 and the flags register unconditionally,

> >> + * and may clobber x2 .. x18 if the yield path is taken.

> >> + */

> >> +

> >> +     .macro          cond_yield_neon, lbl

> >> +     if_will_cond_yield_neon

> >> +     do_cond_yield_neon

> >> +     endif_yield_neon        \lbl

> >> +     .endm

> >> +

> >> +     .macro          if_will_cond_yield_neon

> >> +#ifdef CONFIG_PREEMPT

> >> +     get_thread_info x0

> >> +     ldr             w1, [x0, #TSK_TI_PREEMPT]

> >> +     ldr             x0, [x0, #TSK_TI_FLAGS]

> >> +     cmp             w1, #1 // == PREEMPT_OFFSET

> >

> > Can we at least drop a BUILD_BUG_ON() somewhere to check this?

> >

> > Maybe in kernel_neon_begin() since this is intimately kernel-mode NEON

> > related.

> >

> 

> Sure.

> 

> >> +     csel            x0, x0, xzr, eq

> >> +     tbnz            x0, #TIF_NEED_RESCHED, 5555f    // needs rescheduling?

> >> +#endif

> >

> > A comment that we will fall through to 6666f here may be helpful.

> >

> 

> Indeed. Will add that.

> 

> >> +     .subsection     1

> >> +5555:

> >> +     .endm

> >> +

> >> +     .macro          do_cond_yield_neon

> >> +     bl              kernel_neon_end

> >> +     bl              kernel_neon_begin

> >> +     .endm

> >> +

> >> +     .macro          endif_yield_neon, lbl=6666f

> >> +     b               \lbl

> >> +     .previous

> >> +6666:

> >

> > Could have slightly more random "random" labels here, but otherwise

> > it looks ok to me.

> >

> 

> Which number did you have in mind that is more random than 6666? :-)

> 

> > I might go through and replace all the random labels with something

> > more robust sometime, but I've never been sure it was worth the

> > effort...

> >

> 

> I guess we could invent all kinds of elaborate schemes but as you say,

> having 4 digit numbers and grep'ing the source before you add a new

> one has been working fine so far, so I don't think it should be a

> priority.


You could try $RANDOM for inspiration.

Nested macro use is rare, but a scheme with only 10 possible random
numbers seems a little too optiimstic -- and in practice people don't
always remember to grep when adding new ones.

9999, 8888, 1111 and 2222 are already taken even without this patch.

Cheers
---Dave
diff mbox series

Patch

diff --git a/arch/arm64/include/asm/assembler.h b/arch/arm64/include/asm/assembler.h
index 5f61487e9f93..c54e408fd5a7 100644
--- a/arch/arm64/include/asm/assembler.h
+++ b/arch/arm64/include/asm/assembler.h
@@ -572,4 +572,55 @@  alternative_else_nop_endif
 #endif
 	.endm
 
+/*
+ * Check whether to yield to another runnable task from kernel mode NEON code
+ * (which runs with preemption disabled).
+ *
+ * if_will_cond_yield_neon
+ *        // pre-yield patchup code
+ * do_cond_yield_neon
+ *        // post-yield patchup code
+ * endif_yield_neon
+ *
+ * - Check whether the preempt count is exactly 1, in which case disabling
+ *   preemption once will make the task preemptible. If this is not the case,
+ *   yielding is pointless.
+ * - Check whether TIF_NEED_RESCHED is set, and if so, disable and re-enable
+ *   kernel mode NEON (which will trigger a reschedule), and branch to the
+ *   yield fixup code.
+ *
+ * This macro sequence clobbers x0, x1 and the flags register unconditionally,
+ * and may clobber x2 .. x18 if the yield path is taken.
+ */
+
+	.macro		cond_yield_neon, lbl
+	if_will_cond_yield_neon
+	do_cond_yield_neon
+	endif_yield_neon	\lbl
+	.endm
+
+	.macro		if_will_cond_yield_neon
+#ifdef CONFIG_PREEMPT
+	get_thread_info	x0
+	ldr		w1, [x0, #TSK_TI_PREEMPT]
+	ldr		x0, [x0, #TSK_TI_FLAGS]
+	cmp		w1, #1 // == PREEMPT_OFFSET
+	csel		x0, x0, xzr, eq
+	tbnz		x0, #TIF_NEED_RESCHED, 5555f	// needs rescheduling?
+#endif
+	.subsection	1
+5555:
+	.endm
+
+	.macro		do_cond_yield_neon
+	bl		kernel_neon_end
+	bl		kernel_neon_begin
+	.endm
+
+	.macro		endif_yield_neon, lbl=6666f
+	b		\lbl
+	.previous
+6666:
+	.endm
+
 #endif	/* __ASM_ASSEMBLER_H */