Re: [PATCH -mm crypto] AES: x86_64 asm implementation optimization

From: Huang, Ying
Date: Thu Apr 24 2008 - 23:06:55 EST


Hi, Sebastian,

Thank you very much for your help. From the result you sent, the biggest
performance degradation is between step 4 and step 5. In that step, one
more register is saved before and restored after encryption/decryption.
So I think the reason maybe the read/write port throughput of CPU.

I changed the patches to group the read or write together instead of
interleaving. Can you help me to test these new patches? The new patches
is attached with the mail.

Best Regards,
Huang Ying

On Thu, 2008-04-24 at 00:32 +0200, Sebastian Siewior wrote:
> * Huang, Ying | 2008-04-17 11:36:43 [+0800]:
>
> >Hi, Sebastian,
> Hi Huang,
>
> >The files attached is the separated patches, from step1 to step 7. Thank
> >you very much for your help.
> I've run the following script:
>
> |#!/bin/bash
> |check_error()
> |{
> | r=$?
> | if [ ! $r -eq 0 ]
> | then
> | exit 1
> | fi
> |}
> |
> |modprobe tcrypt mode=200
> |modprobe tcrypt mode=200
> |dmesg -c > step-0.txt
> |
> |for ((i=1; i<=7; i++))
> |do
> | quilt push step${i}.patch
> | check_error
> |
> | make
> | check_error
> |
> | rmmod aes_x86_64
> | check_error
> |
> | insmod arch/x86/crypto/aes-x86_64.ko
> | check_error
> |
> | modprobe tcrypt mode=200
> | modprobe tcrypt mode=200
> | dmesg -c > step-${i}.txt
> |done
>
> and the result is attached.
>
> >Best Regards,
> >Huang Ying
>
> Sebastian

Attachment: patches.tbz2
Description: application/bzip-compressed-tar