mirror of
https://github.com/torvalds/linux.git
synced 2024-11-28 07:01:32 +00:00
977d0558e3
There is lot of headers which we do not need in this file. Delete them and add what we really need. Here is list which identify why we need this header. <linux/kernel.h> // min() <linux/slab.h> // kzalloc() <linux/stddef.h> // offsetof() <linux/string.h> // memcpy(), memset() <linux/types.h> // u8, size_t, etc. "debug.h" // PtrOffset() Signed-off-by: Kari Argillander <kari.argillander@gmail.com> Signed-off-by: Konstantin Komarov <almaz.alexandrovich@paragon-software.com>
454 lines
9.4 KiB
C
454 lines
9.4 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
*
|
|
* Copyright (C) 2019-2021 Paragon Software GmbH, All rights reserved.
|
|
*
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/stddef.h>
|
|
#include <linux/string.h>
|
|
#include <linux/types.h>
|
|
|
|
#include "debug.h"
|
|
#include "ntfs_fs.h"
|
|
|
|
// clang-format off
|
|
/* Src buffer is zero. */
|
|
#define LZNT_ERROR_ALL_ZEROS 1
|
|
#define LZNT_CHUNK_SIZE 0x1000
|
|
// clang-format on
|
|
|
|
struct lznt_hash {
|
|
const u8 *p1;
|
|
const u8 *p2;
|
|
};
|
|
|
|
struct lznt {
|
|
const u8 *unc;
|
|
const u8 *unc_end;
|
|
const u8 *best_match;
|
|
size_t max_len;
|
|
bool std;
|
|
|
|
struct lznt_hash hash[LZNT_CHUNK_SIZE];
|
|
};
|
|
|
|
static inline size_t get_match_len(const u8 *ptr, const u8 *end, const u8 *prev,
|
|
size_t max_len)
|
|
{
|
|
size_t len = 0;
|
|
|
|
while (ptr + len < end && ptr[len] == prev[len] && ++len < max_len)
|
|
;
|
|
return len;
|
|
}
|
|
|
|
static size_t longest_match_std(const u8 *src, struct lznt *ctx)
|
|
{
|
|
size_t hash_index;
|
|
size_t len1 = 0, len2 = 0;
|
|
const u8 **hash;
|
|
|
|
hash_index =
|
|
((40543U * ((((src[0] << 4) ^ src[1]) << 4) ^ src[2])) >> 4) &
|
|
(LZNT_CHUNK_SIZE - 1);
|
|
|
|
hash = &(ctx->hash[hash_index].p1);
|
|
|
|
if (hash[0] >= ctx->unc && hash[0] < src && hash[0][0] == src[0] &&
|
|
hash[0][1] == src[1] && hash[0][2] == src[2]) {
|
|
len1 = 3;
|
|
if (ctx->max_len > 3)
|
|
len1 += get_match_len(src + 3, ctx->unc_end,
|
|
hash[0] + 3, ctx->max_len - 3);
|
|
}
|
|
|
|
if (hash[1] >= ctx->unc && hash[1] < src && hash[1][0] == src[0] &&
|
|
hash[1][1] == src[1] && hash[1][2] == src[2]) {
|
|
len2 = 3;
|
|
if (ctx->max_len > 3)
|
|
len2 += get_match_len(src + 3, ctx->unc_end,
|
|
hash[1] + 3, ctx->max_len - 3);
|
|
}
|
|
|
|
/* Compare two matches and select the best one. */
|
|
if (len1 < len2) {
|
|
ctx->best_match = hash[1];
|
|
len1 = len2;
|
|
} else {
|
|
ctx->best_match = hash[0];
|
|
}
|
|
|
|
hash[1] = hash[0];
|
|
hash[0] = src;
|
|
return len1;
|
|
}
|
|
|
|
static size_t longest_match_best(const u8 *src, struct lznt *ctx)
|
|
{
|
|
size_t max_len;
|
|
const u8 *ptr;
|
|
|
|
if (ctx->unc >= src || !ctx->max_len)
|
|
return 0;
|
|
|
|
max_len = 0;
|
|
for (ptr = ctx->unc; ptr < src; ++ptr) {
|
|
size_t len =
|
|
get_match_len(src, ctx->unc_end, ptr, ctx->max_len);
|
|
if (len >= max_len) {
|
|
max_len = len;
|
|
ctx->best_match = ptr;
|
|
}
|
|
}
|
|
|
|
return max_len >= 3 ? max_len : 0;
|
|
}
|
|
|
|
static const size_t s_max_len[] = {
|
|
0x1002, 0x802, 0x402, 0x202, 0x102, 0x82, 0x42, 0x22, 0x12,
|
|
};
|
|
|
|
static const size_t s_max_off[] = {
|
|
0x10, 0x20, 0x40, 0x80, 0x100, 0x200, 0x400, 0x800, 0x1000,
|
|
};
|
|
|
|
static inline u16 make_pair(size_t offset, size_t len, size_t index)
|
|
{
|
|
return ((offset - 1) << (12 - index)) |
|
|
((len - 3) & (((1 << (12 - index)) - 1)));
|
|
}
|
|
|
|
static inline size_t parse_pair(u16 pair, size_t *offset, size_t index)
|
|
{
|
|
*offset = 1 + (pair >> (12 - index));
|
|
return 3 + (pair & ((1 << (12 - index)) - 1));
|
|
}
|
|
|
|
/*
|
|
* compress_chunk
|
|
*
|
|
* Return:
|
|
* * 0 - Ok, @cmpr contains @cmpr_chunk_size bytes of compressed data.
|
|
* * 1 - Input buffer is full zero.
|
|
* * -2 - The compressed buffer is too small to hold the compressed data.
|
|
*/
|
|
static inline int compress_chunk(size_t (*match)(const u8 *, struct lznt *),
|
|
const u8 *unc, const u8 *unc_end, u8 *cmpr,
|
|
u8 *cmpr_end, size_t *cmpr_chunk_size,
|
|
struct lznt *ctx)
|
|
{
|
|
size_t cnt = 0;
|
|
size_t idx = 0;
|
|
const u8 *up = unc;
|
|
u8 *cp = cmpr + 3;
|
|
u8 *cp2 = cmpr + 2;
|
|
u8 not_zero = 0;
|
|
/* Control byte of 8-bit values: ( 0 - means byte as is, 1 - short pair ). */
|
|
u8 ohdr = 0;
|
|
u8 *last;
|
|
u16 t16;
|
|
|
|
if (unc + LZNT_CHUNK_SIZE < unc_end)
|
|
unc_end = unc + LZNT_CHUNK_SIZE;
|
|
|
|
last = min(cmpr + LZNT_CHUNK_SIZE + sizeof(short), cmpr_end);
|
|
|
|
ctx->unc = unc;
|
|
ctx->unc_end = unc_end;
|
|
ctx->max_len = s_max_len[0];
|
|
|
|
while (up < unc_end) {
|
|
size_t max_len;
|
|
|
|
while (unc + s_max_off[idx] < up)
|
|
ctx->max_len = s_max_len[++idx];
|
|
|
|
/* Find match. */
|
|
max_len = up + 3 <= unc_end ? (*match)(up, ctx) : 0;
|
|
|
|
if (!max_len) {
|
|
if (cp >= last)
|
|
goto NotCompressed;
|
|
not_zero |= *cp++ = *up++;
|
|
} else if (cp + 1 >= last) {
|
|
goto NotCompressed;
|
|
} else {
|
|
t16 = make_pair(up - ctx->best_match, max_len, idx);
|
|
*cp++ = t16;
|
|
*cp++ = t16 >> 8;
|
|
|
|
ohdr |= 1 << cnt;
|
|
up += max_len;
|
|
}
|
|
|
|
cnt = (cnt + 1) & 7;
|
|
if (!cnt) {
|
|
*cp2 = ohdr;
|
|
ohdr = 0;
|
|
cp2 = cp;
|
|
cp += 1;
|
|
}
|
|
}
|
|
|
|
if (cp2 < last)
|
|
*cp2 = ohdr;
|
|
else
|
|
cp -= 1;
|
|
|
|
*cmpr_chunk_size = cp - cmpr;
|
|
|
|
t16 = (*cmpr_chunk_size - 3) | 0xB000;
|
|
cmpr[0] = t16;
|
|
cmpr[1] = t16 >> 8;
|
|
|
|
return not_zero ? 0 : LZNT_ERROR_ALL_ZEROS;
|
|
|
|
NotCompressed:
|
|
|
|
if ((cmpr + LZNT_CHUNK_SIZE + sizeof(short)) > last)
|
|
return -2;
|
|
|
|
/*
|
|
* Copy non cmpr data.
|
|
* 0x3FFF == ((LZNT_CHUNK_SIZE + 2 - 3) | 0x3000)
|
|
*/
|
|
cmpr[0] = 0xff;
|
|
cmpr[1] = 0x3f;
|
|
|
|
memcpy(cmpr + sizeof(short), unc, LZNT_CHUNK_SIZE);
|
|
*cmpr_chunk_size = LZNT_CHUNK_SIZE + sizeof(short);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline ssize_t decompress_chunk(u8 *unc, u8 *unc_end, const u8 *cmpr,
|
|
const u8 *cmpr_end)
|
|
{
|
|
u8 *up = unc;
|
|
u8 ch = *cmpr++;
|
|
size_t bit = 0;
|
|
size_t index = 0;
|
|
u16 pair;
|
|
size_t offset, length;
|
|
|
|
/* Do decompression until pointers are inside range. */
|
|
while (up < unc_end && cmpr < cmpr_end) {
|
|
/* Correct index */
|
|
while (unc + s_max_off[index] < up)
|
|
index += 1;
|
|
|
|
/* Check the current flag for zero. */
|
|
if (!(ch & (1 << bit))) {
|
|
/* Just copy byte. */
|
|
*up++ = *cmpr++;
|
|
goto next;
|
|
}
|
|
|
|
/* Check for boundary. */
|
|
if (cmpr + 1 >= cmpr_end)
|
|
return -EINVAL;
|
|
|
|
/* Read a short from little endian stream. */
|
|
pair = cmpr[1];
|
|
pair <<= 8;
|
|
pair |= cmpr[0];
|
|
|
|
cmpr += 2;
|
|
|
|
/* Translate packed information into offset and length. */
|
|
length = parse_pair(pair, &offset, index);
|
|
|
|
/* Check offset for boundary. */
|
|
if (unc + offset > up)
|
|
return -EINVAL;
|
|
|
|
/* Truncate the length if necessary. */
|
|
if (up + length >= unc_end)
|
|
length = unc_end - up;
|
|
|
|
/* Now we copy bytes. This is the heart of LZ algorithm. */
|
|
for (; length > 0; length--, up++)
|
|
*up = *(up - offset);
|
|
|
|
next:
|
|
/* Advance flag bit value. */
|
|
bit = (bit + 1) & 7;
|
|
|
|
if (!bit) {
|
|
if (cmpr >= cmpr_end)
|
|
break;
|
|
|
|
ch = *cmpr++;
|
|
}
|
|
}
|
|
|
|
/* Return the size of uncompressed data. */
|
|
return up - unc;
|
|
}
|
|
|
|
/*
|
|
* get_lznt_ctx
|
|
* @level: 0 - Standard compression.
|
|
* !0 - Best compression, requires a lot of cpu.
|
|
*/
|
|
struct lznt *get_lznt_ctx(int level)
|
|
{
|
|
struct lznt *r = kzalloc(level ? offsetof(struct lznt, hash)
|
|
: sizeof(struct lznt),
|
|
GFP_NOFS);
|
|
|
|
if (r)
|
|
r->std = !level;
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* compress_lznt - Compresses @unc into @cmpr
|
|
*
|
|
* Return:
|
|
* * +x - Ok, @cmpr contains 'final_compressed_size' bytes of compressed data.
|
|
* * 0 - Input buffer is full zero.
|
|
*/
|
|
size_t compress_lznt(const void *unc, size_t unc_size, void *cmpr,
|
|
size_t cmpr_size, struct lznt *ctx)
|
|
{
|
|
int err;
|
|
size_t (*match)(const u8 *src, struct lznt *ctx);
|
|
u8 *p = cmpr;
|
|
u8 *end = p + cmpr_size;
|
|
const u8 *unc_chunk = unc;
|
|
const u8 *unc_end = unc_chunk + unc_size;
|
|
bool is_zero = true;
|
|
|
|
if (ctx->std) {
|
|
match = &longest_match_std;
|
|
memset(ctx->hash, 0, sizeof(ctx->hash));
|
|
} else {
|
|
match = &longest_match_best;
|
|
}
|
|
|
|
/* Compression cycle. */
|
|
for (; unc_chunk < unc_end; unc_chunk += LZNT_CHUNK_SIZE) {
|
|
cmpr_size = 0;
|
|
err = compress_chunk(match, unc_chunk, unc_end, p, end,
|
|
&cmpr_size, ctx);
|
|
if (err < 0)
|
|
return unc_size;
|
|
|
|
if (is_zero && err != LZNT_ERROR_ALL_ZEROS)
|
|
is_zero = false;
|
|
|
|
p += cmpr_size;
|
|
}
|
|
|
|
if (p <= end - 2)
|
|
p[0] = p[1] = 0;
|
|
|
|
return is_zero ? 0 : PtrOffset(cmpr, p);
|
|
}
|
|
|
|
/*
|
|
* decompress_lznt - Decompress @cmpr into @unc.
|
|
*/
|
|
ssize_t decompress_lznt(const void *cmpr, size_t cmpr_size, void *unc,
|
|
size_t unc_size)
|
|
{
|
|
const u8 *cmpr_chunk = cmpr;
|
|
const u8 *cmpr_end = cmpr_chunk + cmpr_size;
|
|
u8 *unc_chunk = unc;
|
|
u8 *unc_end = unc_chunk + unc_size;
|
|
u16 chunk_hdr;
|
|
|
|
if (cmpr_size < sizeof(short))
|
|
return -EINVAL;
|
|
|
|
/* Read chunk header. */
|
|
chunk_hdr = cmpr_chunk[1];
|
|
chunk_hdr <<= 8;
|
|
chunk_hdr |= cmpr_chunk[0];
|
|
|
|
/* Loop through decompressing chunks. */
|
|
for (;;) {
|
|
size_t chunk_size_saved;
|
|
size_t unc_use;
|
|
size_t cmpr_use = 3 + (chunk_hdr & (LZNT_CHUNK_SIZE - 1));
|
|
|
|
/* Check that the chunk actually fits the supplied buffer. */
|
|
if (cmpr_chunk + cmpr_use > cmpr_end)
|
|
return -EINVAL;
|
|
|
|
/* First make sure the chunk contains compressed data. */
|
|
if (chunk_hdr & 0x8000) {
|
|
/* Decompress a chunk and return if we get an error. */
|
|
ssize_t err =
|
|
decompress_chunk(unc_chunk, unc_end,
|
|
cmpr_chunk + sizeof(chunk_hdr),
|
|
cmpr_chunk + cmpr_use);
|
|
if (err < 0)
|
|
return err;
|
|
unc_use = err;
|
|
} else {
|
|
/* This chunk does not contain compressed data. */
|
|
unc_use = unc_chunk + LZNT_CHUNK_SIZE > unc_end
|
|
? unc_end - unc_chunk
|
|
: LZNT_CHUNK_SIZE;
|
|
|
|
if (cmpr_chunk + sizeof(chunk_hdr) + unc_use >
|
|
cmpr_end) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
memcpy(unc_chunk, cmpr_chunk + sizeof(chunk_hdr),
|
|
unc_use);
|
|
}
|
|
|
|
/* Advance pointers. */
|
|
cmpr_chunk += cmpr_use;
|
|
unc_chunk += unc_use;
|
|
|
|
/* Check for the end of unc buffer. */
|
|
if (unc_chunk >= unc_end)
|
|
break;
|
|
|
|
/* Proceed the next chunk. */
|
|
if (cmpr_chunk > cmpr_end - 2)
|
|
break;
|
|
|
|
chunk_size_saved = LZNT_CHUNK_SIZE;
|
|
|
|
/* Read chunk header. */
|
|
chunk_hdr = cmpr_chunk[1];
|
|
chunk_hdr <<= 8;
|
|
chunk_hdr |= cmpr_chunk[0];
|
|
|
|
if (!chunk_hdr)
|
|
break;
|
|
|
|
/* Check the size of unc buffer. */
|
|
if (unc_use < chunk_size_saved) {
|
|
size_t t1 = chunk_size_saved - unc_use;
|
|
u8 *t2 = unc_chunk + t1;
|
|
|
|
/* 'Zero' memory. */
|
|
if (t2 >= unc_end)
|
|
break;
|
|
|
|
memset(unc_chunk, 0, t1);
|
|
unc_chunk = t2;
|
|
}
|
|
}
|
|
|
|
/* Check compression boundary. */
|
|
if (cmpr_chunk > cmpr_end)
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* The unc size is just a difference between current
|
|
* pointer and original one.
|
|
*/
|
|
return PtrOffset(unc, unc_chunk);
|
|
}
|