Files
Mike Crowe 990e3e81e6 Use LF line endings in the repository
Convert the line endings stored for all text files in the repository to
LF. The majority previously used DOS-style CRLF line endings. Add a
.gitattributes file to enforce this and treat certain extensions as
never being text files.

Update PatchCheck.py to insist on LF line endings rather than CRLF.
However, its other checks fail on this commit due to lots of
pre-existing complaints that it only notices because the line endings
have changed.

Silicon/QemuSocPkg/FspBin/Patches/0001-Build-QEMU-FSP-2.0-binaries.patch
needs to be treated as binary since it contains a mixture of line
endings.

This change has implications depending on the client platform you are
using the repository from:

* Windows

The usual configuration for Git on Windows means that text files will
be checked out to the work tree with DOS-style CRLF line endings. If
that's not the case then you can configure Git to do so for the entire
machine with:

 git config --global core.autocrlf true

or for just the repository with:

 git config core.autocrlf true

Line endings will be normalised to LF when they are committed to the
repository. If you commit a text file with only LF line endings then it
will be converted to CRLF line endings in your work tree.

* Linux, MacOS and other Unices

The usual configuration for Git on such platforms is to check files out
of the repository with LF line endings. This is probably the right thing
for you. In the unlikely even that you are using Git on Unix but editing
or compiling on Windows for some reason then you may need to tweak your
configuration to force the use of CRLF line endings as described above.

* General

For more information see
https://docs.github.com/en/get-started/getting-started-with-git/configuring-git-to-handle-line-endings .

Fixes: https://github.com/slimbootloader/slimbootloader/issues/1400
Signed-off-by: Mike Crowe <mac@mcrowe.com>
2021-11-10 12:46:42 -08:00

675 lines
17 KiB
C

/*******************************************************************************
* Copyright 2002-2020 Intel Corporation
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*******************************************************************************/
/*
//
// Purpose:
// Cryptography Primitive.
// Internal Definitions of Block Cipher Tools
//
//
*/
#if !defined(_CP_TOOL_H)
#define _CP_TOOL_H
#include "pcpmask_ct.h"
#define _NEW_COPY16_
#define _NEW_XOR16_
/* copy data block */
__INLINE void CopyBlock(const void* pSrc, void* pDst, cpSize numBytes)
{
#ifdef _SLIMBOOT_OPT
CopyMem (pDst, pSrc, numBytes);
#else
const Ipp8u* s = (Ipp8u*)pSrc;
Ipp8u* d = (Ipp8u*)pDst;
cpSize k;
for(k=0; k<numBytes; k++ )
d[k] = s[k];
#endif
}
__INLINE void CopyBlock8(const void* pSrc, void* pDst)
{
int k;
for(k=0; k<8; k++ )
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
}
#if defined(_NEW_COPY16_)
__INLINE void CopyBlock16(const void* pSrc, void* pDst)
{
#if (_IPP_ARCH ==_IPP_ARCH_EM64T)
((Ipp64u*)pDst)[0] = ((Ipp64u*)pSrc)[0];
((Ipp64u*)pDst)[1] = ((Ipp64u*)pSrc)[1];
#else
((Ipp32u*)pDst)[0] = ((Ipp32u*)pSrc)[0];
((Ipp32u*)pDst)[1] = ((Ipp32u*)pSrc)[1];
((Ipp32u*)pDst)[2] = ((Ipp32u*)pSrc)[2];
((Ipp32u*)pDst)[3] = ((Ipp32u*)pSrc)[3];
#endif
}
#else
__INLINE void CopyBlock16(const void* pSrc, void* pDst)
{
int k;
for(k=0; k<16; k++ )
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
}
#endif
__INLINE void CopyBlock24(const void* pSrc, void* pDst)
{
int k;
for(k=0; k<24; k++ )
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
}
__INLINE void CopyBlock32(const void* pSrc, void* pDst)
{
int k;
for(k=0; k<32; k++ )
((Ipp8u*)pDst)[k] = ((Ipp8u*)pSrc)[k];
}
/*
// padding data block
*/
__INLINE void PadBlock(Ipp8u paddingByte, void* pDst, cpSize numBytes)
{
#ifdef _SLIMBOOT_OPT
SetMem (pDst, numBytes, paddingByte);
#else
Ipp8u* d = (Ipp8u*)pDst;
cpSize k;
for(k=0; k<numBytes; k++ )
d[k] = paddingByte;
#endif
}
#if !((_IPP>=_IPP_W7) || (_IPP32E>=_IPP32E_M7))
__INLINE void PurgeBlock(void* pDst, int len)
{
#ifdef _SLIMBOOT_OPT
SetMem (pDst, len, 0);
#else
int n;
for(n=0; n<len; n++) ((Ipp8u*)pDst)[n] = 0;
#endif
}
#else
#define PurgeBlock OWNAPI(PurgeBlock)
void PurgeBlock(void* pDst, int len);
#endif
/* fill block */
__INLINE void FillBlock16(Ipp8u filler, const void* pSrc, void* pDst, int len)
{
int n;
for(n=0; n<len; n++) ((Ipp8u*)pDst)[n] = ((Ipp8u*)pSrc)[n];
for(; n<16; n++) ((Ipp8u*)pDst)[n] = filler;
}
/* xor block */
__INLINE void XorBlock(const void* pSrc1, const void* pSrc2, void* pDst, int len)
{
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
Ipp8u* d = (Ipp8u*)pDst;
int k;
for(k=0; k<len; k++)
d[k] = (Ipp8u)(p1[k] ^p2[k]);
}
__INLINE void XorBlock8(const void* pSrc1, const void* pSrc2, void* pDst)
{
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
Ipp8u* d = (Ipp8u*)pDst;
int k;
for(k=0; k<8; k++ )
d[k] = (Ipp8u)(p1[k] ^p2[k]);
}
#if defined(_NEW_XOR16_)
__INLINE void XorBlock16(const void* pSrc1, const void* pSrc2, void* pDst)
{
#if (_IPP_ARCH ==_IPP_ARCH_EM64T)
((Ipp64u*)pDst)[0] = ((Ipp64u*)pSrc1)[0] ^ ((Ipp64u*)pSrc2)[0];
((Ipp64u*)pDst)[1] = ((Ipp64u*)pSrc1)[1] ^ ((Ipp64u*)pSrc2)[1];
#else
((Ipp32u*)pDst)[0] = ((Ipp32u*)pSrc1)[0] ^ ((Ipp32u*)pSrc2)[0];
((Ipp32u*)pDst)[1] = ((Ipp32u*)pSrc1)[1] ^ ((Ipp32u*)pSrc2)[1];
((Ipp32u*)pDst)[2] = ((Ipp32u*)pSrc1)[2] ^ ((Ipp32u*)pSrc2)[2];
((Ipp32u*)pDst)[3] = ((Ipp32u*)pSrc1)[3] ^ ((Ipp32u*)pSrc2)[3];
#endif
}
#else
__INLINE void XorBlock16(const void* pSrc1, const void* pSrc2, void* pDst)
{
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
Ipp8u* d = (Ipp8u*)pDst;
int k;
for(k=0; k<16; k++ )
d[k] = (Ipp8u)(p1[k] ^p2[k]);
}
#endif
__INLINE void XorBlock24(const void* pSrc1, const void* pSrc2, void* pDst)
{
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
Ipp8u* d = (Ipp8u*)pDst;
int k;
for(k=0; k<24; k++ )
d[k] = (Ipp8u)(p1[k] ^p2[k]);
}
__INLINE void XorBlock32(const void* pSrc1, const void* pSrc2, void* pDst)
{
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
Ipp8u* d = (Ipp8u*)pDst;
int k;
for(k=0; k<32; k++ )
d[k] = (Ipp8u)(p1[k] ^p2[k]);
}
/* compare (equivalence) */
__INLINE int EquBlock(const void* pSrc1, const void* pSrc2, int len)
{
const Ipp8u* p1 = (const Ipp8u*)pSrc1;
const Ipp8u* p2 = (const Ipp8u*)pSrc2;
int k;
int isNotEqu;
for(k=0, isNotEqu=0; k<len; k++)
isNotEqu |= (p1[k]^p2[k]);
return !isNotEqu;
}
/* addition (incrementation) functions for CTR mode of diffenent block ciphers */
/* constant execution time version */
__INLINE void StdIncrement(Ipp8u* pCounter, int blkBitSize, int numSize)
{
int maskPosition = (blkBitSize -numSize)/8;
Ipp8u maskVal = (Ipp8u)( 0xFF >> (blkBitSize -numSize)%8 );
int i;
Ipp32u carry = 1;
for(i=BITS2WORD8_SIZE(blkBitSize)-1; i>=0; i--) {
int d = maskPosition - i;
Ipp8u mask = (Ipp8u)(maskVal | cpIsMsb_ct((BNU_CHUNK_T)d));
Ipp32u x = pCounter[i] + carry;
Ipp8u y = pCounter[i];
pCounter[i] = (Ipp8u)((y & ~mask) | (x & mask));
maskVal &= cpIsMsb_ct((BNU_CHUNK_T)d);
carry = (x>>8) & 0x1;
}
}
/* vb */
__INLINE void ompStdIncrement64( void* pInitCtrVal, void* pCurrCtrVal,
int ctrNumBitSize, int n )
{
int k;
Ipp64u cntr;
Ipp64u temp;
Ipp64s item;
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
( ( Ipp8u* )&cntr )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
#else
for( k = 0; k < 8; k++ )
( ( Ipp8u* )&cntr )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
#endif
if( ctrNumBitSize == 64 )
{
cntr += ( Ipp64u )n;
}
else
{
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF) >> ( 64 - ctrNumBitSize );
Ipp64u save = cntr & ( ~mask );
Ipp64u bndr = ( Ipp64u )1 << ctrNumBitSize;
temp = cntr & mask;
cntr = temp + ( Ipp64u )n;
if( cntr > bndr )
{
item = ( Ipp64s )n - ( Ipp64s )( bndr - temp );
while( item > 0 )
{
cntr = ( Ipp64u )item;
item -= ( Ipp64s )bndr;
}
}
cntr = save | ( cntr & mask );
}
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&cntr )[k];
#else
for( k = 0; k < 8; k++ )
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&cntr )[k];
#endif
}
/* vb */
__INLINE void ompStdIncrement128( void* pInitCtrVal, void* pCurrCtrVal,
int ctrNumBitSize, int n )
{
int k;
Ipp64u low;
Ipp64u hgh;
Ipp64u flag;
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF);
Ipp64u save = 0;
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[15 - k];
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
}
#else
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[8 + k];
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
}
#endif
if( ctrNumBitSize == 64 )
{
low += ( Ipp64u )n;
}
else if( ctrNumBitSize < 64 )
{
Ipp64u bndr;
Ipp64u cntr;
Ipp64s item;
mask >>= ( 64 - ctrNumBitSize );
save = low & ( ~mask );
cntr = ( low & mask ) + ( Ipp64u )n;
if( ctrNumBitSize < 31 )
{
bndr = ( Ipp64u )1 << ctrNumBitSize;
if( cntr > bndr )
{
item = ( Ipp64s )( ( Ipp64s )n - ( ( Ipp64s )bndr -
( Ipp64s )( low & mask ) ) );
while( item > 0 )
{
cntr = ( Ipp64u )item;
item -= ( Ipp64s )bndr;
}
}
}
low = save | ( cntr & mask );
}
else
{
flag = ( low >> 63 );
if( ctrNumBitSize != 128 )
{
mask >>= ( 128 - ctrNumBitSize );
save = hgh & ( ~mask );
hgh &= mask;
}
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) ) hgh++;
if( ctrNumBitSize != 128 )
{
hgh = save | ( hgh & mask );
}
}
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )pCurrCtrVal )[15 - k] = ( ( Ipp8u* )&low )[k];
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&hgh )[k];
}
#else
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )pCurrCtrVal )[8 + k] = ( ( Ipp8u* )&low )[k];
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&hgh )[k];
}
#endif
}
#if 0
/* vb */
__INLINE void ompStdIncrement192( void* pInitCtrVal, void* pCurrCtrVal,
int ctrNumBitSize, int n )
{
int k;
Ipp64u low;
Ipp64u mdl;
Ipp64u hgh;
Ipp64u flag;
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF);
Ipp64u save;
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[23 - k];
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[15 - k];
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
}
#else
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[16 + k];
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[8 + k];
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
}
#endif
if( ctrNumBitSize == 64 )
{
low += ( Ipp64u )n;
}
else if( ctrNumBitSize == 128 )
{
flag = ( low >> 63 );
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) ) mdl++;
}
else if( ctrNumBitSize == 192 )
{
flag = ( low >> 63 );
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) )
{
flag = ( mdl >> 63 );
mdl++;
if( flag != ( mdl >> 63 ) ) hgh++;
}
}
else if( ctrNumBitSize < 64 )
{
Ipp64u bndr;
Ipp64u cntr;
Ipp64s item;
mask >>= ( 64 - ctrNumBitSize );
save = low & ( ~mask );
cntr = ( low & mask ) + ( Ipp64u )n;
if( ctrNumBitSize < 31 )
{
bndr = ( Ipp64u )1 << ctrNumBitSize;
if( cntr > bndr )
{
item = ( Ipp64s )( ( Ipp64s )n - ( ( Ipp64s )bndr -
( Ipp64s )( low & mask ) ) );
while( item > 0 )
{
cntr = ( Ipp64u )item;
item -= ( Ipp64s )bndr;
}
}
}
low = save | ( cntr & mask );
}
else if( ctrNumBitSize < 128 )
{
flag = ( low >> 63 );
mask >>= ( 128 - ctrNumBitSize );
save = mdl & ( ~mask );
mdl &= mask;
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) ) mdl++;
mdl = save | ( mdl & mask );
}
else
{
flag = ( low >> 63 );
mask >>= ( 192 - ctrNumBitSize );
save = hgh & ( ~mask );
hgh &= mask;
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) )
{
flag = ( mdl >> 63 );
mdl++;
if( flag != ( mdl >> 63 ) ) hgh++;
}
hgh = save | ( hgh & mask );
}
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )pCurrCtrVal )[23 - k] = ( ( Ipp8u* )&low )[k];
( ( Ipp8u* )pCurrCtrVal )[15 - k] = ( ( Ipp8u* )&mdl )[k];
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&hgh )[k];
}
#else
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )pCurrCtrVal )[16 + k] = ( ( Ipp8u* )&low )[k];
( ( Ipp8u* )pCurrCtrVal )[8 + k] = ( ( Ipp8u* )&mdl )[k];
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&hgh )[k];
}
#endif
}
#endif
#if 0
/* vb */
__INLINE void ompStdIncrement256( void* pInitCtrVal, void* pCurrCtrVal,
int ctrNumBitSize, int n )
{
int k;
Ipp64u low;
Ipp64u mdl;
Ipp64u mdm;
Ipp64u hgh;
Ipp64u flag;
Ipp64u mask = CONST_64(0xFFFFFFFFFFFFFFFF);
Ipp64u save;
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[31 - k];
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[23 - k];
( ( Ipp8u* )&mdm )[k] = ( ( Ipp8u* )pInitCtrVal )[15 - k];
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[7 - k];
}
#else
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )&low )[k] = ( ( Ipp8u* )pInitCtrVal )[24 + k];
( ( Ipp8u* )&mdl )[k] = ( ( Ipp8u* )pInitCtrVal )[16 + k];
( ( Ipp8u* )&mdm )[k] = ( ( Ipp8u* )pInitCtrVal )[8 + k];
( ( Ipp8u* )&hgh )[k] = ( ( Ipp8u* )pInitCtrVal )[k];
}
#endif
if( ctrNumBitSize == 64 )
{
low += ( Ipp64u )n;
}
else if( ctrNumBitSize == 128 )
{
flag = ( low >> 63 );
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) ) mdl++;
}
else if( ctrNumBitSize == 192 )
{
flag = ( low >> 63 );
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) )
{
flag = ( mdl >> 63 );
mdl++;
if( flag != ( mdl >> 63 ) ) hgh++;
}
}
else if( ctrNumBitSize == 256 )
{
flag = ( low >> 63 );
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) )
{
flag = ( mdl >> 63 );
mdl++;
if( flag != ( mdl >> 63 ) )
{
flag = ( mdm >> 63 );
mdm++;
if( flag != ( mdm >> 63 ) ) hgh++;
}
}
}
else if( ctrNumBitSize < 64 )
{
Ipp64u bndr;
Ipp64u cntr;
Ipp64s item;
mask >>= ( 64 - ctrNumBitSize );
save = low & ( ~mask );
cntr = ( low & mask ) + ( Ipp64u )n;
if( ctrNumBitSize < 31 )
{
bndr = ( Ipp64u )1 << ctrNumBitSize;
if( cntr > bndr )
{
item = ( Ipp64s )( ( Ipp64s )n - ( ( Ipp64s )bndr -
( Ipp64s )( low & mask ) ) );
while( item > 0 )
{
cntr = ( Ipp64u )item;
item -= ( Ipp64s )bndr;
}
}
}
low = save | ( cntr & mask );
}
else if( ctrNumBitSize < 128 )
{
flag = ( low >> 63 );
mask >>= ( 128 - ctrNumBitSize );
save = mdl & ( ~mask );
mdl &= mask;
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) ) mdl++;
mdl = save | ( mdl & mask );
}
else if( ctrNumBitSize < 192 )
{
flag = ( low >> 63 );
mask >>= ( 192 - ctrNumBitSize );
save = mdm & ( ~mask );
mdm &= mask;
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) )
{
flag = ( mdl >> 63 );
mdl++;
if( flag != ( mdl >> 63 ) ) mdm++;
}
mdm = save | ( mdm & mask );
}
else
{
flag = ( low >> 63 );
mask >>= ( 256 - ctrNumBitSize );
save = hgh & ( ~mask );
hgh &= mask;
low += ( Ipp64u )n;
if( flag != ( low >> 63 ) )
{
flag = ( mdl >> 63 );
mdl++;
if( flag != ( mdl >> 63 ) )
{
flag = ( mdm >> 63 );
mdm++;
if( flag != ( mdm >> 63 ) ) hgh++;
}
}
hgh = save | ( hgh & mask );
}
#if( IPP_ENDIAN == IPP_LITTLE_ENDIAN )
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )pCurrCtrVal )[31 - k] = ( ( Ipp8u* )&low )[k];
( ( Ipp8u* )pCurrCtrVal )[23 - k] = ( ( Ipp8u* )&mdl )[k];
( ( Ipp8u* )pCurrCtrVal )[15 - k] = ( ( Ipp8u* )&mdm )[k];
( ( Ipp8u* )pCurrCtrVal )[7 - k] = ( ( Ipp8u* )&hgh )[k];
}
#else
for( k = 0; k < 8; k++ )
{
( ( Ipp8u* )pCurrCtrVal )[24 + k] = ( ( Ipp8u* )&low )[k];
( ( Ipp8u* )pCurrCtrVal )[16 + k] = ( ( Ipp8u* )&mdl )[k];
( ( Ipp8u* )pCurrCtrVal )[8 + k] = ( ( Ipp8u* )&mdm )[k];
( ( Ipp8u* )pCurrCtrVal )[k] = ( ( Ipp8u* )&hgh )[k];
}
#endif
}
#endif
#endif /* _CP_TOOL_H */