412 lines
15 KiB
C
412 lines
15 KiB
C
/**
|
|
Copyright (C) powturbo 2015-2023
|
|
SPDX-License-Identifier: GPL v2 License
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 2 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License along
|
|
with this program; if not, write to the Free Software Foundation, Inc.,
|
|
51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
|
|
- email : powturbo [AT] gmail.com
|
|
- github : https://github.com/powturbo
|
|
- homepage : https://sites.google.com/site/powturbo/
|
|
- twitter : https://twitter.com/powturbo
|
|
|
|
TurboRLE - "Most efficient and fastest Run Length Encoding"
|
|
**/
|
|
#ifndef USIZE
|
|
#include <string.h>
|
|
#include "include_/conf.h"
|
|
#include "include_/trle.h"
|
|
#include "trle_.h"
|
|
|
|
#ifdef __ARM_NEON
|
|
#define PREFETCH(_ip_,_rw_)
|
|
#else
|
|
#define PREFETCH(_ip_,_rw_) __builtin_prefetch(_ip_,_rw_)
|
|
#endif
|
|
|
|
//------------------------------------- Fastet Histogram : https://github.com/powturbo/TurboHist -------------------------------------------
|
|
#define cnt_t unsigned
|
|
#define CSIZE (256 + 8)
|
|
|
|
#define CU32(u,_i_) {\
|
|
c[_i_+0][(unsigned char) u ]++;\
|
|
c[_i_+1][(unsigned char)(u>> 8)]++;\
|
|
c[_i_+2][(unsigned char)(u>>16)]++;\
|
|
c[_i_+3][ u>>24 ]++;\
|
|
}
|
|
|
|
#define OV 8
|
|
#define INC4_32(_i_) { { unsigned u = ux, v = vx; ux = ctou32(ip+_i_+OV+0); vx = ctou32(ip+_i_+OV+ 4); CU32(u,0); CU32(v,0); }\
|
|
{ unsigned u = ux, v = vx; ux = ctou32(ip+_i_+OV+8); vx = ctou32(ip+_i_+OV+12); CU32(u,0); CU32(v,0); }\
|
|
}
|
|
|
|
static unsigned cntcalc32(const unsigned char *__restrict in, unsigned inlen, cnt_t *__restrict cnt) {
|
|
cnt_t c[4][CSIZE] = {0},i;
|
|
|
|
unsigned char *ip = in;
|
|
if(inlen >= 64) {
|
|
unsigned ux = ctou32(ip), vx = ctou32(ip+4);
|
|
for(; ip != in+(inlen&~(64-1))-64; ip += 64) { INC4_32(0); INC4_32(16); INC4_32(32); INC4_32(48); __builtin_prefetch(ip+512, 0); }
|
|
}
|
|
while(ip != in+inlen)
|
|
c[0][*ip++]++;
|
|
|
|
for(i = 0; i < 256; i++)
|
|
cnt[i] = c[0][i]+c[1][i]+c[2][i]+c[3][i];
|
|
unsigned a = 256; while(a > 1 && !cnt[a-1]) a--;
|
|
return a;
|
|
}
|
|
|
|
//------------------------------ speed optimized RLE 8 with escape char. SSE/AVX2 slower than scalar in encoding ----------------------------------------
|
|
#define SRLE8 (__WORDSIZE/2)
|
|
|
|
#define USIZE 8
|
|
#include "trlec.c"
|
|
|
|
#if SRLE8
|
|
#define PUTC(_op_, _x_) *_op_++ = _x_
|
|
#define PUTE(_op_, _e_) do { PUTC(_op_, _e_); vlput32(_op_, 0); } while(0)
|
|
|
|
#define SZ64 if((z = (ctou64(ip) ^ ctou64(ip+1)))) goto a; ip += 8;
|
|
#define SZ32 if((z = (ctou32(ip) ^ ctou32(ip+1)))) goto a; ip += 4;
|
|
|
|
#define SRLEPUT8(_pp_, _ip_, _e_, _op_) do {\
|
|
unsigned _r = (_ip_ - _pp_)+1;\
|
|
if(_r >= 4) { PUTC(_op_, _e_); _r = (_r-4)+3; vlput32(_op_, _r); PUTC(_op_, pp[0]); }\
|
|
else if(pp[0] == _e_) {\
|
|
PUTC(_op_, _e_); _r -= 1; vlput32(_op_, _r); /*1-3:Escape char -> 2-6 bytes */\
|
|
} else while(_r--) PUTC(_op_, pp[0]);\
|
|
} while(0)
|
|
|
|
unsigned _srlec8(const unsigned char *__restrict in, unsigned inlen, unsigned char *__restrict out, uint8_t e) {
|
|
uint8_t *ip = in, *pp = in, *ie = in+inlen, *op = out;
|
|
|
|
if(inlen > SRLE8+1)
|
|
while(ip < ie-1-SRLE8) {
|
|
#if __WORDSIZE == 64
|
|
uint64_t z; SZ64; SZ64; SZ64; SZ64; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz64(z)>>3;
|
|
#else
|
|
uint32_t z; SZ32; SZ32; SZ32; SZ32; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz32(z)>>3;
|
|
#endif
|
|
SRLEPUT8(pp, ip, e, op);
|
|
pp = ++ip;
|
|
}
|
|
|
|
while(ip < ie-1) {
|
|
while(ip < ie-1 && ip[1] == *pp) ip++;
|
|
SRLEPUT8(pp, ip, e, op);
|
|
pp = ++ip;
|
|
}
|
|
if(ip < ie) {
|
|
unsigned c = *ip++;
|
|
if(c == e) PUTE(op,e);
|
|
else PUTC(op, c);
|
|
} //AS(ip == ie,"FatalI ip!=ie=%d ", ip-ie)
|
|
return op - out;
|
|
}
|
|
|
|
#define SRLEPUT8X(_pp_, _ip_, _e_, _op_) do {\
|
|
unsigned _r = (_ip_ - _pp_)+1, _cr = pp[0];\
|
|
if(_r >= 4 /*|| _r == 3 && _cr == ix*/) { PUTC(_op_, _e_); _r = ((_r-4)+3)<<1; if(_cr == ix) { vlput32(_op_, _r); } else { vlput32(_op_, _r|1); PUTC(_op_, pp[0]); } }\
|
|
else if(_cr == _e_) { PUTC(_op_, _e_); _r = (_r-1)<<1|1; vlput32(_op_, _r); /*1-3:Escape char -> 2-6 bytes */ } \
|
|
else while(_r--) PUTC(_op_, _cr);\
|
|
} while(0)
|
|
|
|
static inline unsigned _srlec8x(const unsigned char *__restrict in, unsigned inlen, unsigned char *__restrict out, uint8_t e, uint8_t ix) {
|
|
uint8_t *ip = in, *pp = in, *ie = in+inlen, *op = out;
|
|
|
|
if(inlen > SRLE8+1)
|
|
while(ip < ie-1-SRLE8) {
|
|
#if __WORDSIZE == 64
|
|
uint64_t z; SZ64; SZ64; SZ64; SZ64; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz64(z)>>3;
|
|
#else
|
|
uint32_t z; SZ32; SZ32; SZ32; SZ32; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz32(z)>>3;
|
|
#endif
|
|
SRLEPUT8X(pp, ip, e, op);
|
|
pp = ++ip;
|
|
}
|
|
|
|
while(ip < ie-1) {
|
|
while(ip < ie-1 && ip[1] == *pp) ip++;
|
|
SRLEPUT8X(pp, ip, e, op);
|
|
pp = ++ip;
|
|
}
|
|
if(ip < ie) {
|
|
unsigned c = *ip++;
|
|
if(c == e) PUTE(op,e);
|
|
else PUTC(op, c);
|
|
} //AS(ip == ie,"FatalI ip!=ie=%d ", ip-ie)
|
|
return op - out;
|
|
}
|
|
#endif
|
|
|
|
unsigned srlec(const unsigned char *__restrict in, unsigned inlen, unsigned char *__restrict out) { // Automatic escape char determination
|
|
unsigned cnt[256] = {0}, a, m = -1, x = 0, im = 0, i, ix, l;
|
|
if(!inlen) return 0;
|
|
|
|
a = cntcalc32(in, inlen, cnt);
|
|
if(cnt[a-1] == inlen) {
|
|
*out = *in;
|
|
return 1; // RETURN 1 = memset
|
|
}
|
|
|
|
if(a != 256) { // determine escape char
|
|
for(im = a, i = m = 0; i < a; i++)
|
|
if(cnt[i] > x) x = cnt[i],ix = i;
|
|
} else for(i = 0; i < a; i++) {
|
|
if(cnt[i] < m) m = cnt[i],im = i; // minimum for ESC char
|
|
if(cnt[i] > x) x = cnt[i],ix = i; // maximum for embeding in the run length
|
|
}
|
|
out[0] = im;
|
|
out[1] = ix;
|
|
if((l = _srlec8x(in, inlen, out+2, im, ix)+2) < inlen)
|
|
return l;
|
|
memcpy(out, in, inlen);
|
|
return inlen;
|
|
}
|
|
|
|
//------------------------------------------------- TurboRLE ------------------------------------------
|
|
#define TRLEPUT(pp, ip, m, rmap, op) do {\
|
|
int _r_ = (ip - pp)+1;\
|
|
if(_r_ >= TMIN) { \
|
|
unsigned char *q = op; /*checkpoint*/\
|
|
if(pp[0] == ix) { unsigned _r = (_r_ - TMIN)<<1|1; vlzput(op, _r, m, rmap); } else { unsigned _r = (_r_ - TMIN)<<1; vlzput(op, _r, m, rmap); *op++ = pp[0]; }\
|
|
if(op-q >= _r_) { op = q; while(_r_--) *op++ = pp[0]; } /*rollback rle*/\
|
|
} else while(_r_--) *op++ = pp[0];\
|
|
} while(0)
|
|
|
|
unsigned trlec(const unsigned char *__restrict in, unsigned inlen, unsigned char *__restrict out) {
|
|
unsigned cnt[256] = {0}, m=-1, x=0, im, i, a, c;
|
|
unsigned char rmap[256], *op=out, *ie = in+inlen, *ip = in,*pp = in, ix;
|
|
if(!inlen) return 0; // RETURN 0 = zero length
|
|
|
|
a = cntcalc32(in, inlen, cnt);
|
|
if(cnt[a-1] == inlen) {
|
|
*out = *in;
|
|
return 1; // RETURN 1 = memset
|
|
}
|
|
|
|
if(a != 256) { // determine escape char
|
|
for(im = a, i = m = 0; i < a; i++)
|
|
if(cnt[i] > x) x = cnt[i],ix = i;
|
|
} else for(i = 0; i < a; i++) {
|
|
if(cnt[i] < m) m = cnt[i],im = i; // minimum for ESC char
|
|
if(cnt[i] > x) x = cnt[i],ix = i; // maximum for embeding in the run length
|
|
}
|
|
if(m) { // no unused bytes found
|
|
PUTC(op, 0); // 0: srle mode
|
|
PUTC(op, im); // _srlec8 escape char
|
|
op += _srlec8(in, inlen, op, im);
|
|
if(op - out < inlen) return op - out; // RETURN rle/escape
|
|
memcpy(out, in, inlen); // no compression, use memcpy
|
|
return inlen; // RETURN outlen = inlen (memcpy)
|
|
}
|
|
|
|
c = (a+7)/8;
|
|
PUTC(op, c); // c = bitmap length in bytes
|
|
memset(op, 0, 32);
|
|
for(m = i = 0; i != c*8; i++) // set bitmap for unused chars
|
|
if(!cnt[i]) op[i>>3] |= 1<<(i&7), rmap[m++] = i;
|
|
op += c;
|
|
for(; i != 256; i++) rmap[m++] = i;
|
|
|
|
m--;
|
|
PUTC(op, ix);
|
|
|
|
if(inlen > SRLE8+1) // encode
|
|
while(ip < ie-1-SRLE8) {
|
|
#if __WORDSIZE == 64
|
|
uint64_t z; SZ64; SZ64; SZ64; SZ64; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz64(z)>>3;
|
|
#else
|
|
uint32_t z; SZ32; SZ32; SZ32; SZ32; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz32(z)>>3;
|
|
#endif
|
|
TRLEPUT(pp, ip, m, rmap, op);
|
|
pp = ++ip;
|
|
}
|
|
|
|
while(ip < ie-1) {
|
|
while(ip < ie-1 && ip[1] == *pp) ip++;
|
|
TRLEPUT(pp, ip, m, rmap, op);
|
|
pp = ++ip;
|
|
}
|
|
if(ip < ie) PUTC(op, *ip++);
|
|
AS(ip == ie, "Fatal ip>ie=%d ", (int)(ip-ie));
|
|
|
|
if(op - out < inlen)
|
|
return op - out; // RETURN length = rle
|
|
memcpy(out, in, inlen); // no compression, use memcpy
|
|
return inlen; // RETURN outlen = inlen (memcpy)
|
|
}
|
|
|
|
#if 0
|
|
unsigned trlelen(const unsigned char *__restrict in, unsigned inlen) {
|
|
unsigned cnt[256] = {0}, m=-1, x=0, im, i, a, c;
|
|
unsigned char rmap[256], *ie = in+inlen, *ip = in,*pp = in, ix;
|
|
if(!inlen) return 0; // RETURN 0 = zero length
|
|
|
|
a = cntcalc32(in, inlen, cnt);
|
|
if(cnt[a-1] == inlen) {
|
|
*out = *in;
|
|
return 1; // RETURN 1 = memset
|
|
}
|
|
|
|
if(a != 256) { // determine escape char
|
|
for(im = a, i = m = 0; i < a; i++)
|
|
if(cnt[i] > x) x = cnt[i],ix = i;
|
|
} else for(i = 0; i < a; i++) {
|
|
if(cnt[i] < m) m = cnt[i],im = i; // minimum for ESC char
|
|
if(cnt[i] > x) x = cnt[i],ix = i; // maximum for embeding in the run length
|
|
}
|
|
if(m) { // no unused bytes found
|
|
rlelen++; // 0: srle mode
|
|
rlelen++; // _srlec8 escape char
|
|
rlelen += _srlelen8(in, inlen, op, im);
|
|
if(rlelen < inlen) return rlelen; // RETURN rle/escape
|
|
return inlen; // RETURN outlen = inlen (memcpy)
|
|
}
|
|
|
|
c = (a+7)/8;
|
|
rlelen++; //PUTC(op, c); // c = bitmap length in bytes
|
|
//memset(op, 0, 32);
|
|
for(m = i = 0; i != c*8; i++) // set bitmap for unused chars
|
|
if(!cnt[i]) op[i>>3] |= 1<<(i&7), rmap[m++] = i;
|
|
rlelen += c; //op += c;
|
|
for(; i != 256; i++) rmap[m++] = i;
|
|
|
|
m--;
|
|
rlelen++; //PUTC(op, ix);
|
|
|
|
if(inlen > SRLE8+1) // encode
|
|
while(ip < ie-1-SRLE8) {
|
|
#if __WORDSIZE == 64
|
|
uint64_t z; SZ64; SZ64; SZ64; SZ64; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz64(z)>>3;
|
|
#else
|
|
uint32_t z; SZ32; SZ32; SZ32; SZ32; __builtin_prefetch(ip +256, 0);
|
|
continue;
|
|
a: ip += ctz32(z)>>3;
|
|
#endif
|
|
TRLEPUT(pp, ip, m, rmap, op);
|
|
pp = ++ip;
|
|
}
|
|
|
|
while(ip < ie-1) {
|
|
while(ip < ie-1 && ip[1] == *pp) ip++;
|
|
TRLEPUT(pp, ip, m, rmap, op);
|
|
pp = ++ip;
|
|
}
|
|
if(ip < ie) PUTC(op, *ip++); AS(ip == ie, "Fatal ip>ie=%d ", (int)(ip-ie));
|
|
|
|
if(op - out < inlen)
|
|
return op - out; // RETURN length = rle
|
|
memcpy(out, in, inlen); // no compression, use memcpy
|
|
return inlen; // RETURN outlen = inlen (memcpy)
|
|
}
|
|
#endif
|
|
|
|
#undef USIZE
|
|
#undef SRLE8
|
|
//------------------------------------- RLE 16, 32, 64 --------------------------------------------------
|
|
#define USIZE 16
|
|
#include "trlec.c"
|
|
#undef USIZE
|
|
|
|
#define USIZE 32
|
|
#include "trlec.c"
|
|
#undef USIZE
|
|
|
|
#define USIZE 64
|
|
#include "trlec.c"
|
|
#undef USIZE
|
|
|
|
#else // ------------------- include RLE 16, 32, 64
|
|
#define uint_t T3(uint, USIZE, _t)
|
|
#define ctout(_x_) *(uint_t *)(_x_)
|
|
|
|
#define PUTC(_op_, _x_) ctout(_op_) = _x_, _op_ += sizeof(uint_t)
|
|
#define PUTE(_op_, _e_) do { PUTC(_op_, _e_); vlput32(_op_, 0); } while(0)
|
|
|
|
#define SRLEPUT(_pp_, _ip_, _e_, _op_) do {\
|
|
unsigned _r = (_ip_ - _pp_)+1;\
|
|
if(_r >= 4) { PUTC(_op_, _e_); _r = (_r-4)+3; vlput32(_op_, _r); PUTC(_op_, pp[0]); }\
|
|
else if(pp[0] == _e_) {\
|
|
PUTC(_op_, _e_); _r -= 1; vlput32(_op_, _r);\
|
|
} else while(_r--) PUTC(_op_, pp[0]);\
|
|
} while(0)
|
|
|
|
#if !SRLE8
|
|
unsigned T2(_srlec, USIZE)(const unsigned char *__restrict cin, unsigned inlen, unsigned char *__restrict out, uint_t e) {
|
|
unsigned char *op = out;
|
|
unsigned n = inlen/sizeof(uint_t);
|
|
uint_t *in = (uint_t *)cin, *pp = in, *ip = in, *ie = in+n;
|
|
|
|
if(!inlen) return 0;
|
|
#define SZ1 if(ip[0] != ip[1]) goto a; ++ip;
|
|
if(n > 6+1)
|
|
while(ip < ie-1-6) { // fast encode
|
|
SZ1; SZ1; SZ1; SZ1; SZ1; SZ1; __builtin_prefetch(ip +128*USIZE/8, 0);
|
|
continue;
|
|
a:
|
|
SRLEPUT(pp, ip, e, op);
|
|
pp = ++ip;
|
|
}
|
|
|
|
while(ip < ie - 1) { // encode rest
|
|
while(ip < ie-1 && ip[1] == *pp) ip++;
|
|
SRLEPUT(pp, ip, e, op);
|
|
pp = ++ip;
|
|
}
|
|
if(ip < ie) { // last item
|
|
uint_t c = *ip++;
|
|
if(c == e) PUTE(op, e);
|
|
else PUTC(op, c);
|
|
} //AS(ip == ie,"FatalI ip!=ie=%d ", ip-ie)
|
|
#if USIZE > 8
|
|
{ unsigned char *p = (unsigned char *)ip; // remaining bytes inlen % USIZE/8
|
|
while(p < cin+inlen)
|
|
*op++ = *p++;
|
|
}
|
|
#endif
|
|
//AS(ip == ie,"FatalI ip!=ie=%d ", ip-ie)
|
|
return op - out;
|
|
}
|
|
#endif
|
|
#undef SRLEPUT
|
|
#undef PUTC
|
|
#undef PUTE
|
|
|
|
unsigned T2(srlec, USIZE)(const unsigned char *__restrict in, unsigned inlen, unsigned char *__restrict out, uint_t e) {
|
|
unsigned l = T2(_srlec, USIZE)(in, inlen, out, e);
|
|
|
|
if(l < inlen)
|
|
return l;
|
|
memcpy(out, in, inlen);
|
|
return inlen;
|
|
}
|
|
#endif
|
|
|