/* do length NMAX blocks -- requires just one modulo operation */
while (len >= NMAX) {
len -= NMAX;
-#ifndef ADLER32_UNROLL_LESS
+#ifndef UNROLL_LESS
n = NMAX / 16; /* NMAX is divisible by 16 */
#else
n = NMAX / 8; /* NMAX is divisible by 8 */
#endif
do {
-#ifndef ADLER32_UNROLL_LESS
+#ifndef UNROLL_LESS
DO16(buf); /* 16 sums unrolled */
buf += 16;
#else
/* do remaining bytes (less than NMAX, still just one modulo) */
if (len) { /* avoid modulos if none remaining */
-#ifndef ADLER32_UNROLL_LESS
+#ifndef UNROLL_LESS
while (len >= 16) {
len -= 16;
DO16(buf);
;;
esac
- CFLAGS="${CFLAGS} -DUNALIGNED_OK -DADLER32_UNROLL_LESS -DCRC32_UNROLL_LESS -DX86_CPUID"
- SFLAGS="${SFLAGS} -DUNALIGNED_OK -DADLER32_UNROLL_LESS -DCRC32_UNROLL_LESS -DX86_CPUID"
+ CFLAGS="${CFLAGS} -DUNALIGNED_OK -DUNROLL_LESS -DX86_CPUID"
+ SFLAGS="${SFLAGS} -DUNALIGNED_OK -DUNROLL_LESS -DX86_CPUID"
# Enable arch-specific optimizations?
if test $without_optimizations -eq 0; then
case "${ARCH}" in
armv6l | armv6hl)
- # Tests done on Raspberry pi (armv6hl) indicate that UNALIGNED_OK and CRC32_UNROLL_LESS both
+ # Tests done on Raspberry pi (armv6hl) indicate that UNALIGNED_OK and UNROLL_LESS both
# provide performance improvements, totaling about 1.5% for the two.
- CFLAGS="${CFLAGS} -DUNALIGNED_OK -DADLER32_UNROLL_LESS -DCRC32_UNROLL_LESS"
- SFLAGS="${SFLAGS} -DUNALIGNED_OK -DADLER32_UNROLL_LESS -DCRC32_UNROLL_LESS"
+ CFLAGS="${CFLAGS} -DUNALIGNED_OK -DUNROLL_LESS"
+ SFLAGS="${SFLAGS} -DUNALIGNED_OK -DUNROLL_LESS"
;;
esac
}
crc = crc ^ 0xffffffff;
-#ifdef CRC32_UNROLL_LESS
+#ifdef UNROLL_LESS
while (len >= 4) {
DO4;
len -= 4;
buf4 = (const z_crc_t *)(const void *)buf;
-#ifndef CRC32_UNROLL_LESS
+#ifndef UNROLL_LESS
while (len >= 32) {
DOLIT32;
len -= 32;
buf4 = (const z_crc_t *)(const void *)buf;
buf4--;
-#ifndef CRC32_UNROLL_LESS
+#ifndef UNROLL_LESS
while (len >= 32) {
DOBIG32;
len -= 32;