/* i80586 rshift * * Copyright (C) 1992, 1994, 1998, * 2001, 2002 Free Software Foundation, Inc. * * This file is part of Libgcrypt. * * Libgcrypt is free software; you can redistribute it and/or modify * it under the terms of the GNU Lesser General Public License as * published by the Free Software Foundation; either version 2.1 of * the License, or (at your option) any later version. * * Libgcrypt is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA * * Note: This code is heavily based on the GNU MP Library. * Actually it's the same code with only minor changes in the * way the data is stored; this is to support the abstraction * of an optional secure memory allocation which may be used * to avoid revealing of sensitive data due to paging etc. */ #include "sysdep.h" #include "asm-syntax.h" /******************* * mpi_limb_t * _gcry_mpih_rshift( mpi_ptr_t wp, (sp + 4) * mpi_ptr_t up, (sp + 8) * mpi_size_t usize, (sp + 12) * unsigned cnt) (sp + 16) */ .text ALIGN (3) .globl C_SYMBOL_NAME(_gcry_mpih_rshift) C_SYMBOL_NAME(_gcry_mpih_rshift:) pushl %edi pushl %esi pushl %ebx pushl %ebp movl 20(%esp),%edi /* res_ptr */ movl 24(%esp),%esi /* s_ptr */ movl 28(%esp),%ebp /* size */ movl 32(%esp),%ecx /* cnt */ /* We can use faster code for shift-by-1 under certain conditions. */ cmp $1,%ecx jne Rnormal leal 4(%edi),%eax cmpl %esi,%eax jnc Rspecial /* jump if res_ptr + 1 >= s_ptr */ leal (%edi,%ebp,4),%eax cmpl %eax,%esi jnc Rspecial /* jump if s_ptr >= res_ptr + size */ Rnormal: movl (%esi),%edx addl $4,%esi xorl %eax,%eax shrdl %cl,%edx,%eax /* compute carry limb */ pushl %eax /* push carry limb onto stack */ decl %ebp pushl %ebp shrl $3,%ebp jz Rend movl (%edi),%eax /* fetch destination cache line */ ALIGN (2) Roop: movl 28(%edi),%eax /* fetch destination cache line */ movl %edx,%ebx movl (%esi),%eax movl 4(%esi),%edx shrdl %cl,%eax,%ebx shrdl %cl,%edx,%eax movl %ebx,(%edi) movl %eax,4(%edi) movl 8(%esi),%ebx movl 12(%esi),%eax shrdl %cl,%ebx,%edx shrdl %cl,%eax,%ebx movl %edx,8(%edi) movl %ebx,12(%edi) movl 16(%esi),%edx movl 20(%esi),%ebx shrdl %cl,%edx,%eax shrdl %cl,%ebx,%edx movl %eax,16(%edi) movl %edx,20(%edi) movl 24(%esi),%eax movl 28(%esi),%edx shrdl %cl,%eax,%ebx shrdl %cl,%edx,%eax movl %ebx,24(%edi) movl %eax,28(%edi) addl $32,%esi addl $32,%edi decl %ebp jnz Roop Rend: popl %ebp andl $7,%ebp jz Rend2 Roop2: movl (%esi),%eax shrdl %cl,%eax,%edx /* compute result limb */ movl %edx,(%edi) movl %eax,%edx addl $4,%esi addl $4,%edi decl %ebp jnz Roop2 Rend2: shrl %cl,%edx /* compute most significant limb */ movl %edx,(%edi) /* store it */ popl %eax /* pop carry limb */ popl %ebp popl %ebx popl %esi popl %edi ret /* We loop from least significant end of the arrays, which is only permissable if the source and destination don't overlap, since the function is documented to work for overlapping source and destination. */ Rspecial: leal -4(%edi,%ebp,4),%edi leal -4(%esi,%ebp,4),%esi movl (%esi),%edx subl $4,%esi decl %ebp pushl %ebp shrl $3,%ebp shrl $1,%edx incl %ebp decl %ebp jz RLend movl (%edi),%eax /* fetch destination cache line */ ALIGN (2) RLoop: movl -28(%edi),%eax /* fetch destination cache line */ movl %edx,%ebx movl (%esi),%eax movl -4(%esi),%edx rcrl $1,%eax movl %ebx,(%edi) rcrl $1,%edx movl %eax,-4(%edi) movl -8(%esi),%ebx movl -12(%esi),%eax rcrl $1,%ebx movl %edx,-8(%edi) rcrl $1,%eax movl %ebx,-12(%edi) movl -16(%esi),%edx movl -20(%esi),%ebx rcrl $1,%edx movl %eax,-16(%edi) rcrl $1,%ebx movl %edx,-20(%edi) movl -24(%esi),%eax movl -28(%esi),%edx rcrl $1,%eax movl %ebx,-24(%edi) rcrl $1,%edx movl %eax,-28(%edi) leal -32(%esi),%esi /* use leal not to clobber carry */ leal -32(%edi),%edi decl %ebp jnz RLoop RLend: popl %ebp sbbl %eax,%eax /* save carry in %eax */ andl $7,%ebp jz RLend2 addl %eax,%eax /* restore carry from eax */ RLoop2: movl %edx,%ebx movl (%esi),%edx rcrl $1,%edx movl %ebx,(%edi) leal -4(%esi),%esi /* use leal not to clobber carry */ leal -4(%edi),%edi decl %ebp jnz RLoop2 jmp RL1 RLend2: addl %eax,%eax /* restore carry from eax */ RL1: movl %edx,(%edi) /* store last limb */ movl $0,%eax rcrl $1,%eax popl %ebp popl %ebx popl %esi popl %edi ret