You've already forked slimbootloader
mirror of
https://github.com/Dasharo/slimbootloader.git
synced 2026-03-06 15:26:20 -08:00
Convert the line endings stored for all text files in the repository to LF. The majority previously used DOS-style CRLF line endings. Add a .gitattributes file to enforce this and treat certain extensions as never being text files. Update PatchCheck.py to insist on LF line endings rather than CRLF. However, its other checks fail on this commit due to lots of pre-existing complaints that it only notices because the line endings have changed. Silicon/QemuSocPkg/FspBin/Patches/0001-Build-QEMU-FSP-2.0-binaries.patch needs to be treated as binary since it contains a mixture of line endings. This change has implications depending on the client platform you are using the repository from: * Windows The usual configuration for Git on Windows means that text files will be checked out to the work tree with DOS-style CRLF line endings. If that's not the case then you can configure Git to do so for the entire machine with: git config --global core.autocrlf true or for just the repository with: git config core.autocrlf true Line endings will be normalised to LF when they are committed to the repository. If you commit a text file with only LF line endings then it will be converted to CRLF line endings in your work tree. * Linux, MacOS and other Unices The usual configuration for Git on such platforms is to check files out of the repository with LF line endings. This is probably the right thing for you. In the unlikely even that you are using Git on Unix but editing or compiling on Windows for some reason then you may need to tweak your configuration to force the use of CRLF line endings as described above. * General For more information see https://docs.github.com/en/get-started/getting-started-with-git/configuring-git-to-handle-line-endings . Fixes: https://github.com/slimbootloader/slimbootloader/issues/1400 Signed-off-by: Mike Crowe <mac@mcrowe.com>
675 lines
17 KiB
C
675 lines
17 KiB
C
/*******************************************************************************
|
|
* Copyright 2002-2020 Intel Corporation
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*******************************************************************************/
|
|
|
|
/*
|
|
//
|
|
// Purpose:
|
|
// Cryptography Primitive.
|
|
// Internal Definitions of Block Cipher Tools
|
|
//
|
|
//
|
|
*/
|
|
|
|
#if !defined(_CP_TOOL_H)
|
|
#define _CP_TOOL_H
|
|
|
|
#include "pcpmask_ct.h"
|
|
|
|
#define _NEW_COPY16_
|
|
#define _NEW_XOR16_
|
|
|
|
/* copy data block */
|
|
__INLINE void CopyBlock(const void* pSrc, void* pDst, cpSize numBytes)
|
|
{
|
|
#ifdef _SLIMBOOT_OPT
|
|
CopyMem (pDst, pSrc, numBytes);
|
|
#else
|
|
const Ipp8u* s = (Ipp8u*)pSrc;
|
|
Ipp8u* d = (Ipp8u*)pDst;
|
|
cpSize k;
|
|
for(k=0; k<numBytes; k++ )
|
|
d[k] = s[k];
|
|
#endif
|
|
}
|
|
__INLINE void CopyBlock8(const void* pSrc, void* pDst)
|
|
{
|
|
int k;
|
|
for(k=0; k<8; k++ )
|
|
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
|
|
}
|
|
|
|
#if defined(_NEW_COPY16_)
|
|
__INLINE void CopyBlock16(const void* pSrc, void* pDst)
|
|
{
|
|
#if (_IPP_ARCH ==_IPP_ARCH_EM64T)
|
|
((Ipp64u*)pDst)[0] = ((Ipp64u*)pSrc)[0];
|
|
((Ipp64u*)pDst)[1] = ((Ipp64u*)pSrc)[1];
|
|
#else
|
|
((Ipp32u*)pDst)[0] = ((Ipp32u*)pSrc)[0];
|
|
((Ipp32u*)pDst)[1] = ((Ipp32u*)pSrc)[1];
|
|
((Ipp32u*)pDst)[2] = ((Ipp32u*)pSrc)[2];
|
|
((Ipp32u*)pDst)[3] = ((Ipp32u*)pSrc)[3];
|
|
#endif
|
|
}
|
|
#else
|
|
__INLINE void CopyBlock16(const void* pSrc, void* pDst)
|
|
{
|
|
int k;
|
|
for(k=0; k<16; k++ )
|
|
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
|
|
}
|
|
#endif
|
|
|
|
__INLINE void CopyBlock24(const void* pSrc, void* pDst)
|
|
{
|
|
int k;
|
|
for(k=0; k<24; k++ )
|
|
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
|
|
}
|
|
__INLINE void CopyBlock32(const void* pSrc, void* pDst)
|
|
{
|
|
int k;
|
|
for(k=0; k<32; k++ )
|
|
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
|
|
}
|
|
|
|
/*
|
|
// padding data block
|
|
*/
|
|
__INLINE void PadBlock(Ipp8u paddingByte, void* pDst, cpSize numBytes)
|
|
{
|
|
#ifdef _SLIMBOOT_OPT
|
|
SetMem (pDst, numBytes, paddingByte);
|
|
#else
|
|
Ipp8u* d = (Ipp8u*)pDst;
|
|
cpSize k;
|
|
for(k=0; k<numBytes; k++ )
|
|
d[k] = paddingByte;
|
|
#endif
|
|
}
|
|
|
|
#if !((_IPP>=_IPP_W7) || (_IPP32E>=_IPP32E_M7))
|
|
__INLINE void PurgeBlock(void* pDst, int len)
|
|
{
|
|
#ifdef _SLIMBOOT_OPT
|
|
SetMem (pDst, len, 0);
|
|
#else
|
|
int n;
|
|
for(n=0; n<len; n++) ((Ipp8u*)pDst)[n] = 0;
|
|
#endif
|
|
}
|
|
#else
|
|
#define PurgeBlock OWNAPI(PurgeBlock)
|
|
void PurgeBlock(void* pDst, int len);
|
|
#endif
|
|
|
|
/* fill block */
|
|
__INLINE void FillBlock16(Ipp8u filler, const void* pSrc, void* pDst, int len)
|
|
{
|
|
int n;
|
|
for(n=0; n<len; n++) ((Ipp8u*)pDst)[n] = ((Ipp8u*)pSrc)[n];
|
|
for(; n<16; n++) ((Ipp8u*)pDst)[n] = filler;
|
|
}
|
|
|
|
/* xor block */
|
|
__INLINE void XorBlock(const void* pSrc1, const void* pSrc2, void* pDst, int len)
|
|
{
|
|
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
|
|
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
|
|
Ipp8u* d = (Ipp8u*)pDst;
|
|
int k;
|
|
for(k=0; k<len; k++)
|
|
d[k] = (Ipp8u)(p1[k] ^p2[k]);
|
|
}
|
|
__INLINE void XorBlock8(const void* pSrc1, const void* pSrc2, void* pDst)
|
|
{
|
|
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
|
|
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
|
|
Ipp8u* d = (Ipp8u*)pDst;
|
|
int k;
|
|
for(k=0; k<8; k++ )
|
|
d[k] = (Ipp8u)(p1[k] ^p2[k]);
|
|
}
|
|
|
|
#if defined(_NEW_XOR16_)
|
|
__INLINE void XorBlock16(const void* pSrc1, const void* pSrc2, void* pDst)
|
|
{
|
|
#if (_IPP_ARCH ==_IPP_ARCH_EM64T)
|
|
((Ipp64u*)pDst)[0] = ((Ipp64u*)pSrc1)[0] ^ ((Ipp64u*)pSrc2)[0];
|
|
((Ipp64u*)pDst)[1] = ((Ipp64u*)pSrc1)[1] ^ ((Ipp64u*)pSrc2)[1];
|
|
#else
|
|
((Ipp32u*)pDst)[0] = ((Ipp32u*)pSrc1)[0] ^ ((Ipp32u*)pSrc2)[0];
|
|
((Ipp32u*)pDst)[1] = ((Ipp32u*)pSrc1)[1] ^ ((Ipp32u*)pSrc2)[1];
|
|
((Ipp32u*)pDst)[2] = ((Ipp32u*)pSrc1)[2] ^ ((Ipp32u*)pSrc2)[2];
|
|
((Ipp32u*)pDst)[3] = ((Ipp32u*)pSrc1)[3] ^ ((Ipp32u*)pSrc2)[3];
|
|
#endif
|
|
}
|
|
#else
|
|
__INLINE void XorBlock16(const void* pSrc1, const void* pSrc2, void* pDst)
|
|
{
|
|
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
|
|
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
|
|
Ipp8u* d = (Ipp8u*)pDst;
|
|
int k;
|
|
for(k=0; k<16; k++ )
|
|
d[k] = (Ipp8u)(p1[k] ^p2[k]);
|
|
}
|
|
#endif
|
|
|
|
__INLINE void XorBlock24(const void* pSrc1, const void* pSrc2, void* pDst)
|
|
{
|
|
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
|
|
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
|
|
Ipp8u* d = (Ipp8u*)pDst;
|
|
int k;
|
|
for(k=0; k<24; k++ )
|
|
d[k] = (Ipp8u)(p1[k] ^p2[k]);
|
|
}
|
|
__INLINE void XorBlock32(const void* pSrc1, const void* pSrc2, void* pDst)
|
|
{
|
|
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
|
|
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
|
|
Ipp8u* d = (Ipp8u*)pDst;
|
|
int k;
|
|
for(k=0; k<32; k++ )
|
|
d[k] = (Ipp8u)(p1[k] ^p2[k]);
|
|
}
|
|
|
|
|
|
/* compare (equivalence) */
|
|
__INLINE int EquBlock(const void* pSrc1, const void* pSrc2, int len)
|
|
{
|
|
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
|
|
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
|
|
int k;
|
|
int isNotEqu;
|
|
for(k=0, isNotEqu=0; k<len; k++)
|
|
isNotEqu |= (p1[k]^p2[k]);
|
|
return !isNotEqu;
|
|
}
|
|
|
|
|
|
/* addition (incrementation) functions for CTR mode of diffenent block ciphers */
|
|
/* constant execution time version */
|
|
__INLINE void StdIncrement(Ipp8u* pCounter, int blkBitSize, int numSize)
|
|
{
|
|
int maskPosition = (blkBitSize -numSize)/8;
|
|
Ipp8u maskVal = (Ipp8u)( 0xFF >> (blkBitSize -numSize)%8 );
|
|
|
|
int i;
|
|
Ipp32u carry = 1;
|
|
for(i=BITS2WORD8_SIZE(blkBitSize)-1; i>=0; i--) {
|
|
int d = maskPosition - i;
|
|
Ipp8u mask = (Ipp8u)(maskVal | cpIsMsb_ct((BNU_CHUNK_T)d));
|
|
|
|
Ipp32u x = pCounter[i] + carry;
|
|
Ipp8u y = pCounter[i];
|
|
pCounter[i] = (Ipp8u)((y & ~mask) | (x & mask));
|
|
|
|
maskVal &= cpIsMsb_ct((BNU_CHUNK_T)d);
|
|
|
|
carry = (x>>8) & 0x1;
|
|
}
|
|
}
|
|
|
|
/* vb */
|
|
__INLINE void ompStdIncrement64( void* pInitCtrVal, void* pCurrCtrVal,
|
|
int ctrNumBitSize, int n )
|
|
{
|
|
int k;
|
|
Ipp64u cntr;
|
|
Ipp64u temp;
|
|
Ipp64s item;
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
( ( Ipp8u* )&cntr )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
( ( Ipp8u* )&cntr )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
|
|
#endif
|
|
|
|
if( ctrNumBitSize == 64 )
|
|
{
|
|
cntr += ( Ipp64u )n;
|
|
}
|
|
else
|
|
{
|
|
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF) >> ( 64 - ctrNumBitSize );
|
|
Ipp64u save = cntr & ( ~mask );
|
|
Ipp64u bndr = ( Ipp64u )1 << ctrNumBitSize;
|
|
|
|
temp = cntr & mask;
|
|
cntr = temp + ( Ipp64u )n;
|
|
|
|
if( cntr > bndr )
|
|
{
|
|
item = ( Ipp64s )n - ( Ipp64s )( bndr - temp );
|
|
|
|
while( item > 0 )
|
|
{
|
|
cntr = ( Ipp64u )item;
|
|
item -= ( Ipp64s )bndr;
|
|
}
|
|
}
|
|
|
|
cntr = save | ( cntr & mask );
|
|
}
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&cntr )[k];
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&cntr )[k];
|
|
#endif
|
|
}
|
|
|
|
|
|
/* vb */
|
|
__INLINE void ompStdIncrement128( void* pInitCtrVal, void* pCurrCtrVal,
|
|
int ctrNumBitSize, int n )
|
|
{
|
|
int k;
|
|
Ipp64u low;
|
|
Ipp64u hgh;
|
|
Ipp64u flag;
|
|
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF);
|
|
Ipp64u save = 0;
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[15 - k];
|
|
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
|
|
}
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[8 + k];
|
|
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
|
|
}
|
|
#endif
|
|
|
|
if( ctrNumBitSize == 64 )
|
|
{
|
|
low += ( Ipp64u )n;
|
|
}
|
|
else if( ctrNumBitSize < 64 )
|
|
{
|
|
Ipp64u bndr;
|
|
Ipp64u cntr;
|
|
Ipp64s item;
|
|
|
|
mask >>= ( 64 - ctrNumBitSize );
|
|
save = low & ( ~mask );
|
|
cntr = ( low & mask ) + ( Ipp64u )n;
|
|
|
|
if( ctrNumBitSize < 31 )
|
|
{
|
|
bndr = ( Ipp64u )1 << ctrNumBitSize;
|
|
|
|
if( cntr > bndr )
|
|
{
|
|
item = ( Ipp64s )( ( Ipp64s )n - ( ( Ipp64s )bndr -
|
|
( Ipp64s )( low & mask ) ) );
|
|
|
|
while( item > 0 )
|
|
{
|
|
cntr = ( Ipp64u )item;
|
|
item -= ( Ipp64s )bndr;
|
|
}
|
|
}
|
|
}
|
|
|
|
low = save | ( cntr & mask );
|
|
}
|
|
else
|
|
{
|
|
flag = ( low >> 63 );
|
|
|
|
if( ctrNumBitSize != 128 )
|
|
{
|
|
mask >>= ( 128 - ctrNumBitSize );
|
|
save = hgh & ( ~mask );
|
|
hgh &= mask;
|
|
}
|
|
|
|
low += ( Ipp64u )n;
|
|
|
|
if( flag != ( low >> 63 ) ) hgh++;
|
|
|
|
if( ctrNumBitSize != 128 )
|
|
{
|
|
hgh = save | ( hgh & mask );
|
|
}
|
|
}
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )pCurrCtrVal )[15 - k] = ( ( Ipp8u* )&low )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&hgh )[k];
|
|
}
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )pCurrCtrVal )[8 + k] = ( ( Ipp8u* )&low )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&hgh )[k];
|
|
}
|
|
#endif
|
|
}
|
|
|
|
#if 0
|
|
/* vb */
|
|
__INLINE void ompStdIncrement192( void* pInitCtrVal, void* pCurrCtrVal,
|
|
int ctrNumBitSize, int n )
|
|
{
|
|
int k;
|
|
Ipp64u low;
|
|
Ipp64u mdl;
|
|
Ipp64u hgh;
|
|
Ipp64u flag;
|
|
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF);
|
|
Ipp64u save;
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[23 - k];
|
|
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[15 - k];
|
|
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
|
|
}
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[16 + k];
|
|
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[8 + k];
|
|
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
|
|
}
|
|
#endif
|
|
|
|
if( ctrNumBitSize == 64 )
|
|
{
|
|
low += ( Ipp64u )n;
|
|
}
|
|
else if( ctrNumBitSize == 128 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
low += ( Ipp64u )n;
|
|
if( flag != ( low >> 63 ) ) mdl++;
|
|
}
|
|
else if( ctrNumBitSize == 192 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
low += ( Ipp64u )n;
|
|
|
|
if( flag != ( low >> 63 ) )
|
|
{
|
|
flag = ( mdl >> 63 );
|
|
mdl++;
|
|
if( flag != ( mdl >> 63 ) ) hgh++;
|
|
}
|
|
}
|
|
else if( ctrNumBitSize < 64 )
|
|
{
|
|
Ipp64u bndr;
|
|
Ipp64u cntr;
|
|
Ipp64s item;
|
|
|
|
mask >>= ( 64 - ctrNumBitSize );
|
|
save = low & ( ~mask );
|
|
cntr = ( low & mask ) + ( Ipp64u )n;
|
|
|
|
if( ctrNumBitSize < 31 )
|
|
{
|
|
bndr = ( Ipp64u )1 << ctrNumBitSize;
|
|
|
|
if( cntr > bndr )
|
|
{
|
|
item = ( Ipp64s )( ( Ipp64s )n - ( ( Ipp64s )bndr -
|
|
( Ipp64s )( low & mask ) ) );
|
|
|
|
while( item > 0 )
|
|
{
|
|
cntr = ( Ipp64u )item;
|
|
item -= ( Ipp64s )bndr;
|
|
}
|
|
}
|
|
}
|
|
|
|
low = save | ( cntr & mask );
|
|
}
|
|
else if( ctrNumBitSize < 128 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
mask >>= ( 128 - ctrNumBitSize );
|
|
save = mdl & ( ~mask );
|
|
mdl &= mask;
|
|
low += ( Ipp64u )n;
|
|
if( flag != ( low >> 63 ) ) mdl++;
|
|
mdl = save | ( mdl & mask );
|
|
}
|
|
else
|
|
{
|
|
flag = ( low >> 63 );
|
|
mask >>= ( 192 - ctrNumBitSize );
|
|
save = hgh & ( ~mask );
|
|
hgh &= mask;
|
|
low += ( Ipp64u )n;
|
|
|
|
if( flag != ( low >> 63 ) )
|
|
{
|
|
flag = ( mdl >> 63 );
|
|
mdl++;
|
|
if( flag != ( mdl >> 63 ) ) hgh++;
|
|
}
|
|
|
|
hgh = save | ( hgh & mask );
|
|
}
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )pCurrCtrVal )[23 - k] = ( ( Ipp8u* )&low )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[15 - k] = ( ( Ipp8u* )&mdl )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&hgh )[k];
|
|
}
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )pCurrCtrVal )[16 + k] = ( ( Ipp8u* )&low )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[8 + k] = ( ( Ipp8u* )&mdl )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&hgh )[k];
|
|
}
|
|
#endif
|
|
}
|
|
#endif
|
|
|
|
#if 0
|
|
/* vb */
|
|
__INLINE void ompStdIncrement256( void* pInitCtrVal, void* pCurrCtrVal,
|
|
int ctrNumBitSize, int n )
|
|
{
|
|
int k;
|
|
Ipp64u low;
|
|
Ipp64u mdl;
|
|
Ipp64u mdm;
|
|
Ipp64u hgh;
|
|
Ipp64u flag;
|
|
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF);
|
|
Ipp64u save;
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[31 - k];
|
|
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[23 - k];
|
|
( ( Ipp8u* )&mdm )[k] = ( ( Ipp8u* )pInitCtrVal )[15 - k];
|
|
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
|
|
}
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[24 + k];
|
|
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[16 + k];
|
|
( ( Ipp8u* )&mdm )[k] = ( ( Ipp8u* )pInitCtrVal )[8 + k];
|
|
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
|
|
}
|
|
#endif
|
|
|
|
if( ctrNumBitSize == 64 )
|
|
{
|
|
low += ( Ipp64u )n;
|
|
}
|
|
else if( ctrNumBitSize == 128 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
low += ( Ipp64u )n;
|
|
if( flag != ( low >> 63 ) ) mdl++;
|
|
}
|
|
else if( ctrNumBitSize == 192 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
low += ( Ipp64u )n;
|
|
|
|
if( flag != ( low >> 63 ) )
|
|
{
|
|
flag = ( mdl >> 63 );
|
|
mdl++;
|
|
if( flag != ( mdl >> 63 ) ) hgh++;
|
|
}
|
|
}
|
|
else if( ctrNumBitSize == 256 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
low += ( Ipp64u )n;
|
|
|
|
if( flag != ( low >> 63 ) )
|
|
{
|
|
flag = ( mdl >> 63 );
|
|
mdl++;
|
|
|
|
if( flag != ( mdl >> 63 ) )
|
|
{
|
|
flag = ( mdm >> 63 );
|
|
mdm++;
|
|
if( flag != ( mdm >> 63 ) ) hgh++;
|
|
}
|
|
}
|
|
}
|
|
else if( ctrNumBitSize < 64 )
|
|
{
|
|
Ipp64u bndr;
|
|
Ipp64u cntr;
|
|
Ipp64s item;
|
|
|
|
mask >>= ( 64 - ctrNumBitSize );
|
|
save = low & ( ~mask );
|
|
cntr = ( low & mask ) + ( Ipp64u )n;
|
|
|
|
if( ctrNumBitSize < 31 )
|
|
{
|
|
bndr = ( Ipp64u )1 << ctrNumBitSize;
|
|
|
|
if( cntr > bndr )
|
|
{
|
|
item = ( Ipp64s )( ( Ipp64s )n - ( ( Ipp64s )bndr -
|
|
( Ipp64s )( low & mask ) ) );
|
|
|
|
while( item > 0 )
|
|
{
|
|
cntr = ( Ipp64u )item;
|
|
item -= ( Ipp64s )bndr;
|
|
}
|
|
}
|
|
}
|
|
|
|
low = save | ( cntr & mask );
|
|
}
|
|
else if( ctrNumBitSize < 128 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
mask >>= ( 128 - ctrNumBitSize );
|
|
save = mdl & ( ~mask );
|
|
mdl &= mask;
|
|
low += ( Ipp64u )n;
|
|
if( flag != ( low >> 63 ) ) mdl++;
|
|
mdl = save | ( mdl & mask );
|
|
}
|
|
else if( ctrNumBitSize < 192 )
|
|
{
|
|
flag = ( low >> 63 );
|
|
mask >>= ( 192 - ctrNumBitSize );
|
|
save = mdm & ( ~mask );
|
|
mdm &= mask;
|
|
low += ( Ipp64u )n;
|
|
|
|
if( flag != ( low >> 63 ) )
|
|
{
|
|
flag = ( mdl >> 63 );
|
|
mdl++;
|
|
if( flag != ( mdl >> 63 ) ) mdm++;
|
|
}
|
|
|
|
mdm = save | ( mdm & mask );
|
|
}
|
|
else
|
|
{
|
|
flag = ( low >> 63 );
|
|
mask >>= ( 256 - ctrNumBitSize );
|
|
save = hgh & ( ~mask );
|
|
hgh &= mask;
|
|
low += ( Ipp64u )n;
|
|
|
|
if( flag != ( low >> 63 ) )
|
|
{
|
|
flag = ( mdl >> 63 );
|
|
mdl++;
|
|
|
|
if( flag != ( mdl >> 63 ) )
|
|
{
|
|
flag = ( mdm >> 63 );
|
|
mdm++;
|
|
if( flag != ( mdm >> 63 ) ) hgh++;
|
|
}
|
|
}
|
|
|
|
hgh = save | ( hgh & mask );
|
|
}
|
|
|
|
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )pCurrCtrVal )[31 - k] = ( ( Ipp8u* )&low )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[23 - k] = ( ( Ipp8u* )&mdl )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[15 - k] = ( ( Ipp8u* )&mdm )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&hgh )[k];
|
|
}
|
|
#else
|
|
for( k = 0; k < 8; k++ )
|
|
{
|
|
( ( Ipp8u* )pCurrCtrVal )[24 + k] = ( ( Ipp8u* )&low )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[16 + k] = ( ( Ipp8u* )&mdl )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[8 + k] = ( ( Ipp8u* )&mdm )[k];
|
|
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&hgh )[k];
|
|
}
|
|
#endif
|
|
}
|
|
#endif
|
|
|
|
#endif /* _CP_TOOL_H */
|