diff mbox series

target/ppc/mem_helper.c: Remove a conditional from dcbz_common()

Message ID 20240622204833.5F7C74E6000@zero.eik.bme.hu
State New
Headers show
Series target/ppc/mem_helper.c: Remove a conditional from dcbz_common() | expand

Commit Message

BALATON Zoltan June 22, 2024, 8:48 p.m. UTC
Instead of passing a bool and select a value within dcbz_common() let
the callers pass in the right value to avoid this conditional
statement. On PPC dcbz is often used to zero memory and some code uses
it a lot. This change improves the run time of a test case that copies
memory with a dcbz call in every iteration from 6.23 to 5.83 seconds.

Signed-off-by: BALATON Zoltan <balaton@eik.bme.hu>
---
This is just a small optimisation removing some of the overhead but
dcbz still seems to be the biggest issue with this test. Removing the
dcbz call it runs in 2 seconds. In a profile I see:
  Children      Self  Command   Shared Object            Symbol
-   55.01%    11.44%  qemu-ppc  qemu-ppc                 [.] dcbz_common.constprop.0
               - 43.57% dcbz_common.constprop.0
                  - probe_access
                     - page_get_flags
                          interval_tree_iter_first
               - 11.44% helper_raise_exception_err
                    cpu_loop_exit_restore
                    cpu_loop
                    cpu_exec
                    cpu_exec_setjmp.isra.0
                    cpu_exec_loop.constprop.0
                    cpu_tb_exec
                    0x7f262403636e
                    helper_raise_exception_err
                    cpu_loop_exit_restore
                    cpu_loop
                    cpu_exec
                    cpu_exec_setjmp.isra.0
                    cpu_exec_loop.constprop.0
                    cpu_tb_exec
                  - 0x7f26240386a4
                       11.20% helper_dcbz
+   43.81%    12.28%  qemu-ppc  qemu-ppc                 [.] probe_access
+   39.31%     0.00%  qemu-ppc  [JIT] tid 9969           [.] 0x00007f2624000000
+   32.45%     4.51%  qemu-ppc  qemu-ppc                 [.] page_get_flags
+   25.50%     2.10%  qemu-ppc  qemu-ppc                 [.] interval_tree_iter_first
+   24.67%    24.67%  qemu-ppc  qemu-ppc                 [.] interval_tree_subtree_search
+   16.75%     1.19%  qemu-ppc  qemu-ppc                 [.] helper_dcbz
+    4.78%     4.78%  qemu-ppc  [JIT] tid 9969           [.] 0x00007f26240386be
+    3.46%     3.46%  qemu-ppc  libc-2.32.so             [.] __memset_avx2_unaligned_erms
Any idea how this could be optimised further? (This is running with
qemu-ppc user mode emulation but I think with system it might be even
worse.) Could an inline implementation with TCG vector ops work to
avoid the helper and let it compile to efficient host code? Even if
that could work I don't know how to do that so I'd need some further
advice on this.

 target/ppc/mem_helper.c | 7 +++----
 1 file changed, 3 insertions(+), 4 deletions(-)

Comments

Richard Henderson June 23, 2024, 5:52 p.m. UTC | #1
On 6/22/24 13:48, BALATON Zoltan wrote:
> Instead of passing a bool and select a value within dcbz_common() let
> the callers pass in the right value to avoid this conditional
> statement. On PPC dcbz is often used to zero memory and some code uses
> it a lot. This change improves the run time of a test case that copies
> memory with a dcbz call in every iteration from 6.23 to 5.83 seconds.
> 
> Signed-off-by: BALATON Zoltan <balaton@eik.bme.hu>
> ---
> This is just a small optimisation removing some of the overhead but
> dcbz still seems to be the biggest issue with this test. Removing the
> dcbz call it runs in 2 seconds. In a profile I see:
>    Children      Self  Command   Shared Object            Symbol
> -   55.01%    11.44%  qemu-ppc  qemu-ppc                 [.] dcbz_common.constprop.0
>                 - 43.57% dcbz_common.constprop.0
>                    - probe_access
>                       - page_get_flags
>                            interval_tree_iter_first
>                 - 11.44% helper_raise_exception_err
>                      cpu_loop_exit_restore
>                      cpu_loop
>                      cpu_exec
>                      cpu_exec_setjmp.isra.0
>                      cpu_exec_loop.constprop.0
>                      cpu_tb_exec
>                      0x7f262403636e
>                      helper_raise_exception_err
>                      cpu_loop_exit_restore
>                      cpu_loop
>                      cpu_exec
>                      cpu_exec_setjmp.isra.0
>                      cpu_exec_loop.constprop.0
>                      cpu_tb_exec
>                    - 0x7f26240386a4
>                         11.20% helper_dcbz
> +   43.81%    12.28%  qemu-ppc  qemu-ppc                 [.] probe_access
> +   39.31%     0.00%  qemu-ppc  [JIT] tid 9969           [.] 0x00007f2624000000
> +   32.45%     4.51%  qemu-ppc  qemu-ppc                 [.] page_get_flags
> +   25.50%     2.10%  qemu-ppc  qemu-ppc                 [.] interval_tree_iter_first
> +   24.67%    24.67%  qemu-ppc  qemu-ppc                 [.] interval_tree_subtree_search
> +   16.75%     1.19%  qemu-ppc  qemu-ppc                 [.] helper_dcbz
> +    4.78%     4.78%  qemu-ppc  [JIT] tid 9969           [.] 0x00007f26240386be
> +    3.46%     3.46%  qemu-ppc  libc-2.32.so             [.] __memset_avx2_unaligned_erms
> Any idea how this could be optimised further? (This is running with
> qemu-ppc user mode emulation but I think with system it might be even
> worse.) Could an inline implementation with TCG vector ops work to
> avoid the helper and let it compile to efficient host code? Even if
> that could work I don't know how to do that so I'd need some further
> advice on this.
> 
>   target/ppc/mem_helper.c | 7 +++----
>   1 file changed, 3 insertions(+), 4 deletions(-)
> 
> diff --git a/target/ppc/mem_helper.c b/target/ppc/mem_helper.c
> index f88155ad45..361fd72226 100644
> --- a/target/ppc/mem_helper.c
> +++ b/target/ppc/mem_helper.c
> @@ -271,12 +271,11 @@ void helper_stsw(CPUPPCState *env, target_ulong addr, uint32_t nb,
>   }
>   
>   static void dcbz_common(CPUPPCState *env, target_ulong addr,
> -                        uint32_t opcode, bool epid, uintptr_t retaddr)
> +                        uint32_t opcode, int mmu_idx, uintptr_t retaddr)
>   {
>       target_ulong mask, dcbz_size = env->dcache_line_size;
>       uint32_t i;
>       void *haddr;
> -    int mmu_idx = epid ? PPC_TLB_EPID_STORE : ppc_env_mmu_index(env, false);
>   
>   #if defined(TARGET_PPC64)
>       /* Check for dcbz vs dcbzl on 970 */
> @@ -309,12 +308,12 @@ static void dcbz_common(CPUPPCState *env, target_ulong addr,
>   
>   void helper_dcbz(CPUPPCState *env, target_ulong addr, uint32_t opcode)
>   {
> -    dcbz_common(env, addr, opcode, false, GETPC());
> +    dcbz_common(env, addr, opcode, ppc_env_mmu_index(env, false), GETPC());

This is already computed in the translator: DisasContext.mem_idx.
If you pass the mmu_idx as an argument, you can unify these two helpers.


r~

>   }
>   
>   void helper_dcbzep(CPUPPCState *env, target_ulong addr, uint32_t opcode)
>   {
> -    dcbz_common(env, addr, opcode, true, GETPC());
> +    dcbz_common(env, addr, opcode, PPC_TLB_EPID_STORE, GETPC());
>   }
>   
>   void helper_icbi(CPUPPCState *env, target_ulong addr)
BALATON Zoltan June 23, 2024, 10:24 p.m. UTC | #2
On Sun, 23 Jun 2024, Richard Henderson wrote:
> On 6/22/24 13:48, BALATON Zoltan wrote:
>> Instead of passing a bool and select a value within dcbz_common() let
>> the callers pass in the right value to avoid this conditional
>> statement. On PPC dcbz is often used to zero memory and some code uses
>> it a lot. This change improves the run time of a test case that copies
>> memory with a dcbz call in every iteration from 6.23 to 5.83 seconds.
>> 
>> Signed-off-by: BALATON Zoltan <balaton@eik.bme.hu>
>> ---
>> This is just a small optimisation removing some of the overhead but
>> dcbz still seems to be the biggest issue with this test. Removing the
>> dcbz call it runs in 2 seconds. In a profile I see:
>>    Children      Self  Command   Shared Object            Symbol
>> -   55.01%    11.44%  qemu-ppc  qemu-ppc                 [.] 
>> dcbz_common.constprop.0
>>                 - 43.57% dcbz_common.constprop.0
>>                    - probe_access
>>                       - page_get_flags
>>                            interval_tree_iter_first
>>                 - 11.44% helper_raise_exception_err
>>                      cpu_loop_exit_restore
>>                      cpu_loop
>>                      cpu_exec
>>                      cpu_exec_setjmp.isra.0
>>                      cpu_exec_loop.constprop.0
>>                      cpu_tb_exec
>>                      0x7f262403636e
>>                      helper_raise_exception_err
>>                      cpu_loop_exit_restore
>>                      cpu_loop
>>                      cpu_exec
>>                      cpu_exec_setjmp.isra.0
>>                      cpu_exec_loop.constprop.0
>>                      cpu_tb_exec
>>                    - 0x7f26240386a4
>>                         11.20% helper_dcbz
>> +   43.81%    12.28%  qemu-ppc  qemu-ppc                 [.] probe_access
>> +   39.31%     0.00%  qemu-ppc  [JIT] tid 9969           [.] 
>> 0x00007f2624000000
>> +   32.45%     4.51%  qemu-ppc  qemu-ppc                 [.] page_get_flags
>> +   25.50%     2.10%  qemu-ppc  qemu-ppc                 [.] 
>> interval_tree_iter_first
>> +   24.67%    24.67%  qemu-ppc  qemu-ppc                 [.] 
>> interval_tree_subtree_search
>> +   16.75%     1.19%  qemu-ppc  qemu-ppc                 [.] helper_dcbz
>> +    4.78%     4.78%  qemu-ppc  [JIT] tid 9969           [.] 
>> 0x00007f26240386be
>> +    3.46%     3.46%  qemu-ppc  libc-2.32.so             [.] 
>> __memset_avx2_unaligned_erms
>> Any idea how this could be optimised further? (This is running with
>> qemu-ppc user mode emulation but I think with system it might be even
>> worse.) Could an inline implementation with TCG vector ops work to
>> avoid the helper and let it compile to efficient host code? Even if
>> that could work I don't know how to do that so I'd need some further
>> advice on this.
>>
>>   target/ppc/mem_helper.c | 7 +++----
>>   1 file changed, 3 insertions(+), 4 deletions(-)
>> 
>> diff --git a/target/ppc/mem_helper.c b/target/ppc/mem_helper.c
>> index f88155ad45..361fd72226 100644
>> --- a/target/ppc/mem_helper.c
>> +++ b/target/ppc/mem_helper.c
>> @@ -271,12 +271,11 @@ void helper_stsw(CPUPPCState *env, target_ulong addr, 
>> uint32_t nb,
>>   }
>>     static void dcbz_common(CPUPPCState *env, target_ulong addr,
>> -                        uint32_t opcode, bool epid, uintptr_t retaddr)
>> +                        uint32_t opcode, int mmu_idx, uintptr_t retaddr)
>>   {
>>       target_ulong mask, dcbz_size = env->dcache_line_size;
>>       uint32_t i;
>>       void *haddr;
>> -    int mmu_idx = epid ? PPC_TLB_EPID_STORE : ppc_env_mmu_index(env, 
>> false);
>>     #if defined(TARGET_PPC64)
>>       /* Check for dcbz vs dcbzl on 970 */
>> @@ -309,12 +308,12 @@ static void dcbz_common(CPUPPCState *env, 
>> target_ulong addr,
>>     void helper_dcbz(CPUPPCState *env, target_ulong addr, uint32_t opcode)
>>   {
>> -    dcbz_common(env, addr, opcode, false, GETPC());
>> +    dcbz_common(env, addr, opcode, ppc_env_mmu_index(env, false), 
>> GETPC());
>
> This is already computed in the translator: DisasContext.mem_idx.
> If you pass the mmu_idx as an argument, you can unify these two helpers.

I've tried that. It works but slower: I get 5.9 seconds vs 5.83 with this 
patch so I think I'd stay with this one. Maybe it's because making the 
helper take 4 parameters instead of 3? I can submit the patch for 
reference if it would be useful but I'd keep this one for merging for now.

Regards,
BALATON Zoltan

>
> r~
>
>>   }
>>     void helper_dcbzep(CPUPPCState *env, target_ulong addr, uint32_t 
>> opcode)
>>   {
>> -    dcbz_common(env, addr, opcode, true, GETPC());
>> +    dcbz_common(env, addr, opcode, PPC_TLB_EPID_STORE, GETPC());
>>   }
>>     void helper_icbi(CPUPPCState *env, target_ulong addr)
>
>
Richard Henderson June 24, 2024, 3:23 a.m. UTC | #3
On 6/23/24 15:24, BALATON Zoltan wrote:
> On Sun, 23 Jun 2024, Richard Henderson wrote:
>> On 6/22/24 13:48, BALATON Zoltan wrote:
>>> Instead of passing a bool and select a value within dcbz_common() let
>>> the callers pass in the right value to avoid this conditional
>>> statement. On PPC dcbz is often used to zero memory and some code uses
>>> it a lot. This change improves the run time of a test case that copies
>>> memory with a dcbz call in every iteration from 6.23 to 5.83 seconds.
>>>
>>> Signed-off-by: BALATON Zoltan <balaton@eik.bme.hu>
>>> ---
>>> This is just a small optimisation removing some of the overhead but
>>> dcbz still seems to be the biggest issue with this test. Removing the
>>> dcbz call it runs in 2 seconds. In a profile I see:
>>>    Children      Self  Command   Shared Object            Symbol
>>> -   55.01%    11.44%  qemu-ppc  qemu-ppc                 [.] dcbz_common.constprop.0
>>>                 - 43.57% dcbz_common.constprop.0
>>>                    - probe_access
>>>                       - page_get_flags
>>>                            interval_tree_iter_first
>>>                 - 11.44% helper_raise_exception_err
>>>                      cpu_loop_exit_restore
>>>                      cpu_loop
>>>                      cpu_exec
>>>                      cpu_exec_setjmp.isra.0
>>>                      cpu_exec_loop.constprop.0
>>>                      cpu_tb_exec
>>>                      0x7f262403636e
>>>                      helper_raise_exception_err
>>>                      cpu_loop_exit_restore
>>>                      cpu_loop
>>>                      cpu_exec
>>>                      cpu_exec_setjmp.isra.0
>>>                      cpu_exec_loop.constprop.0
>>>                      cpu_tb_exec
>>>                    - 0x7f26240386a4
>>>                         11.20% helper_dcbz
>>> +   43.81%    12.28%  qemu-ppc  qemu-ppc                 [.] probe_access
>>> +   39.31%     0.00%  qemu-ppc  [JIT] tid 9969           [.] 0x00007f2624000000
>>> +   32.45%     4.51%  qemu-ppc  qemu-ppc                 [.] page_get_flags
>>> +   25.50%     2.10%  qemu-ppc  qemu-ppc                 [.] interval_tree_iter_first
>>> +   24.67%    24.67%  qemu-ppc  qemu-ppc                 [.] interval_tree_subtree_search
>>> +   16.75%     1.19%  qemu-ppc  qemu-ppc                 [.] helper_dcbz
>>> +    4.78%     4.78%  qemu-ppc  [JIT] tid 9969           [.] 0x00007f26240386be
>>> +    3.46%     3.46%  qemu-ppc  libc-2.32.so             [.] __memset_avx2_unaligned_erms
>>> Any idea how this could be optimised further? (This is running with
>>> qemu-ppc user mode emulation but I think with system it might be even
>>> worse.) Could an inline implementation with TCG vector ops work to
>>> avoid the helper and let it compile to efficient host code? Even if
>>> that could work I don't know how to do that so I'd need some further
>>> advice on this.
>>>
>>>   target/ppc/mem_helper.c | 7 +++----
>>>   1 file changed, 3 insertions(+), 4 deletions(-)
>>>
>>> diff --git a/target/ppc/mem_helper.c b/target/ppc/mem_helper.c
>>> index f88155ad45..361fd72226 100644
>>> --- a/target/ppc/mem_helper.c
>>> +++ b/target/ppc/mem_helper.c
>>> @@ -271,12 +271,11 @@ void helper_stsw(CPUPPCState *env, target_ulong addr, uint32_t nb,
>>>   }
>>>     static void dcbz_common(CPUPPCState *env, target_ulong addr,
>>> -                        uint32_t opcode, bool epid, uintptr_t retaddr)
>>> +                        uint32_t opcode, int mmu_idx, uintptr_t retaddr)
>>>   {
>>>       target_ulong mask, dcbz_size = env->dcache_line_size;
>>>       uint32_t i;
>>>       void *haddr;
>>> -    int mmu_idx = epid ? PPC_TLB_EPID_STORE : ppc_env_mmu_index(env, false);
>>>     #if defined(TARGET_PPC64)
>>>       /* Check for dcbz vs dcbzl on 970 */
>>> @@ -309,12 +308,12 @@ static void dcbz_common(CPUPPCState *env, target_ulong addr,
>>>     void helper_dcbz(CPUPPCState *env, target_ulong addr, uint32_t opcode)
>>>   {
>>> -    dcbz_common(env, addr, opcode, false, GETPC());
>>> +    dcbz_common(env, addr, opcode, ppc_env_mmu_index(env, false), GETPC());
>>
>> This is already computed in the translator: DisasContext.mem_idx.
>> If you pass the mmu_idx as an argument, you can unify these two helpers.
> 
> I've tried that. It works but slower: I get 5.9 seconds vs 5.83 with this patch so I think 
> I'd stay with this one. Maybe it's because making the helper take 4 parameters instead of 
> 3? I can submit the patch for reference if it would be useful but I'd keep this one for 
> merging for now.

Interesting.  Can you share your test case?

The other thing I see is that we can split out the 970 test to a separate helper.  We can 
test for POWERPC_EXCP_970 and !(opcode & 0x00200000) during translation.  I think the 
current placement, where dcbzep tests that as well, is wrong, since dcbze is an E500 insn.
Anyway, that would eliminate opcode as an argument bringing you back to 3.


r~
BALATON Zoltan June 24, 2024, 9:19 a.m. UTC | #4
On Sun, 23 Jun 2024, Richard Henderson wrote:
> On 6/23/24 15:24, BALATON Zoltan wrote:
>> On Sun, 23 Jun 2024, Richard Henderson wrote:
>>> On 6/22/24 13:48, BALATON Zoltan wrote:
>>>> Instead of passing a bool and select a value within dcbz_common() let
>>>> the callers pass in the right value to avoid this conditional
>>>> statement. On PPC dcbz is often used to zero memory and some code uses
>>>> it a lot. This change improves the run time of a test case that copies
>>>> memory with a dcbz call in every iteration from 6.23 to 5.83 seconds.
>>>> 
>>>> Signed-off-by: BALATON Zoltan <balaton@eik.bme.hu>
>>>> ---
>>>> This is just a small optimisation removing some of the overhead but
>>>> dcbz still seems to be the biggest issue with this test. Removing the
>>>> dcbz call it runs in 2 seconds. In a profile I see:
>>>>    Children      Self  Command   Shared Object            Symbol
>>>> -   55.01%    11.44%  qemu-ppc  qemu-ppc                 [.] 
>>>> dcbz_common.constprop.0
>>>>                 - 43.57% dcbz_common.constprop.0
>>>>                    - probe_access
>>>>                       - page_get_flags
>>>>                            interval_tree_iter_first
>>>>                 - 11.44% helper_raise_exception_err
>>>>                      cpu_loop_exit_restore
>>>>                      cpu_loop
>>>>                      cpu_exec
>>>>                      cpu_exec_setjmp.isra.0
>>>>                      cpu_exec_loop.constprop.0
>>>>                      cpu_tb_exec
>>>>                      0x7f262403636e
>>>>                      helper_raise_exception_err
>>>>                      cpu_loop_exit_restore
>>>>                      cpu_loop
>>>>                      cpu_exec
>>>>                      cpu_exec_setjmp.isra.0
>>>>                      cpu_exec_loop.constprop.0
>>>>                      cpu_tb_exec
>>>>                    - 0x7f26240386a4
>>>>                         11.20% helper_dcbz
>>>> +   43.81%    12.28%  qemu-ppc  qemu-ppc                 [.] probe_access
>>>> +   39.31%     0.00%  qemu-ppc  [JIT] tid 9969           [.] 
>>>> 0x00007f2624000000
>>>> +   32.45%     4.51%  qemu-ppc  qemu-ppc                 [.] 
>>>> page_get_flags
>>>> +   25.50%     2.10%  qemu-ppc  qemu-ppc                 [.] 
>>>> interval_tree_iter_first
>>>> +   24.67%    24.67%  qemu-ppc  qemu-ppc                 [.] 
>>>> interval_tree_subtree_search
>>>> +   16.75%     1.19%  qemu-ppc  qemu-ppc                 [.] helper_dcbz
>>>> +    4.78%     4.78%  qemu-ppc  [JIT] tid 9969           [.] 
>>>> 0x00007f26240386be
>>>> +    3.46%     3.46%  qemu-ppc  libc-2.32.so             [.] 
>>>> __memset_avx2_unaligned_erms
>>>> Any idea how this could be optimised further? (This is running with
>>>> qemu-ppc user mode emulation but I think with system it might be even
>>>> worse.) Could an inline implementation with TCG vector ops work to
>>>> avoid the helper and let it compile to efficient host code? Even if
>>>> that could work I don't know how to do that so I'd need some further
>>>> advice on this.
>>>> 
>>>>   target/ppc/mem_helper.c | 7 +++----
>>>>   1 file changed, 3 insertions(+), 4 deletions(-)
>>>> 
>>>> diff --git a/target/ppc/mem_helper.c b/target/ppc/mem_helper.c
>>>> index f88155ad45..361fd72226 100644
>>>> --- a/target/ppc/mem_helper.c
>>>> +++ b/target/ppc/mem_helper.c
>>>> @@ -271,12 +271,11 @@ void helper_stsw(CPUPPCState *env, target_ulong 
>>>> addr, uint32_t nb,
>>>>   }
>>>>     static void dcbz_common(CPUPPCState *env, target_ulong addr,
>>>> -                        uint32_t opcode, bool epid, uintptr_t retaddr)
>>>> +                        uint32_t opcode, int mmu_idx, uintptr_t retaddr)
>>>>   {
>>>>       target_ulong mask, dcbz_size = env->dcache_line_size;
>>>>       uint32_t i;
>>>>       void *haddr;
>>>> -    int mmu_idx = epid ? PPC_TLB_EPID_STORE : ppc_env_mmu_index(env, 
>>>> false);
>>>>     #if defined(TARGET_PPC64)
>>>>       /* Check for dcbz vs dcbzl on 970 */
>>>> @@ -309,12 +308,12 @@ static void dcbz_common(CPUPPCState *env, 
>>>> target_ulong addr,
>>>>     void helper_dcbz(CPUPPCState *env, target_ulong addr, uint32_t 
>>>> opcode)
>>>>   {
>>>> -    dcbz_common(env, addr, opcode, false, GETPC());
>>>> +    dcbz_common(env, addr, opcode, ppc_env_mmu_index(env, false), 
>>>> GETPC());
>>> 
>>> This is already computed in the translator: DisasContext.mem_idx.
>>> If you pass the mmu_idx as an argument, you can unify these two helpers.
>> 
>> I've tried that. It works but slower: I get 5.9 seconds vs 5.83 with this 
>> patch so I think I'd stay with this one. Maybe it's because making the 
>> helper take 4 parameters instead of 3? I can submit the patch for reference 
>> if it would be useful but I'd keep this one for merging for now.
>
> Interesting.  Can you share your test case?
>
> The other thing I see is that we can split out the 970 test to a separate 
> helper.  We can test for POWERPC_EXCP_970 and !(opcode & 0x00200000) during 
> translation.  I think the current placement, where dcbzep tests that as well, 
> is wrong, since dcbze is an E500 insn.
> Anyway, that would eliminate opcode as an argument bringing you back to 3.

It'd eliminate opcode but I think I'd need to pass cache line size instead 
somehow so it might still need 4 parameters. I'll give that a try and see 
if it's better or not. But the biggest overhead is still the call to 
probe_write. Is there a way to eliminate that and maybe do the zeroing 
with vector or 128 bir ops in compiled code and not in the helper? (I've 
tried removing the test and memset but the slow path is really slow, takes 
more than 9 seconds so it's still faster to test when memset can be used.)

Regards,
BALATON Zoltan
diff mbox series

Patch

diff --git a/target/ppc/mem_helper.c b/target/ppc/mem_helper.c
index f88155ad45..361fd72226 100644
--- a/target/ppc/mem_helper.c
+++ b/target/ppc/mem_helper.c
@@ -271,12 +271,11 @@  void helper_stsw(CPUPPCState *env, target_ulong addr, uint32_t nb,
 }
 
 static void dcbz_common(CPUPPCState *env, target_ulong addr,
-                        uint32_t opcode, bool epid, uintptr_t retaddr)
+                        uint32_t opcode, int mmu_idx, uintptr_t retaddr)
 {
     target_ulong mask, dcbz_size = env->dcache_line_size;
     uint32_t i;
     void *haddr;
-    int mmu_idx = epid ? PPC_TLB_EPID_STORE : ppc_env_mmu_index(env, false);
 
 #if defined(TARGET_PPC64)
     /* Check for dcbz vs dcbzl on 970 */
@@ -309,12 +308,12 @@  static void dcbz_common(CPUPPCState *env, target_ulong addr,
 
 void helper_dcbz(CPUPPCState *env, target_ulong addr, uint32_t opcode)
 {
-    dcbz_common(env, addr, opcode, false, GETPC());
+    dcbz_common(env, addr, opcode, ppc_env_mmu_index(env, false), GETPC());
 }
 
 void helper_dcbzep(CPUPPCState *env, target_ulong addr, uint32_t opcode)
 {
-    dcbz_common(env, addr, opcode, true, GETPC());
+    dcbz_common(env, addr, opcode, PPC_TLB_EPID_STORE, GETPC());
 }
 
 void helper_icbi(CPUPPCState *env, target_ulong addr)