mirror of
https://git.tukaani.org/xz.git
synced 2025-10-08 18:28:23 +00:00
In the C99 and C17 standards, section 6.5.6 paragraph 8 means that adding 0 to a null pointer is undefined behavior. As of writing, "clang -fsanitize=undefined" (Clang 15) diagnoses this. However, I'm not aware of any compiler that would take advantage of this when optimizing (Clang 15 included). It's good to avoid this anyway since compilers might some day infer that pointer arithmetic implies that the pointer is not NULL. That is, the following foo() would then unconditionally return 0, even for foo(NULL, 0): void bar(char *a, char *b); int foo(char *a, size_t n) { bar(a, a + n); return a == NULL; } In contrast to C, C++ explicitly allows null pointer + 0. So if the above is compiled as C++ then there is no undefined behavior in the foo(NULL, 0) call. To me it seems that changing the C standard would be the sane thing to do (just add one sentence) as it would ensure that a huge amount of old code won't break in the future. Based on web searches it seems that a large number of codebases (where null pointer + 0 occurs) are being fixed instead to be future-proof in case compilers will some day optimize based on it (like making the above foo(NULL, 0) return 0) which in the worst case will cause security bugs. Some projects don't plan to change it. For example, gnulib and thus many GNU tools currently require that null pointer + 0 is defined: https://lists.gnu.org/archive/html/bug-gnulib/2021-11/msg00000.html https://www.gnu.org/software/gnulib/manual/html_node/Other-portability-assumptions.html In XZ Utils null pointer + 0 issue should be fixed after this commit. This adds a few if-statements and thus branches to avoid null pointer + 0. These check for size > 0 instead of ptr != NULL because this way bugs where size > 0 && ptr == NULL will likely get caught quickly. None of them are in hot spots so it shouldn't matter for performance. A little less readable version would be replacing ptr + offset with offset != 0 ? ptr + offset : ptr or creating a macro for it: #define my_ptr_add(ptr, offset) \ ((offset) != 0 ? ((ptr) + (offset)) : (ptr)) Checking for offset != 0 instead of ptr != NULL allows GCC >= 8.1, Clang >= 7, and Clang-based ICX to optimize it to the very same code as ptr + offset. That is, it won't create a branch. So for hot code this could be a good solution to avoid null pointer + 0. Unfortunately other compilers like ICC 2021 or MSVC 19.33 (VS2022) will create a branch from my_ptr_add(). Thanks to Marcin Kowalczyk for reporting the problem: https://github.com/tukaani-project/xz/issues/36
344 lines
9.0 KiB
C
344 lines
9.0 KiB
C
///////////////////////////////////////////////////////////////////////////////
|
|
//
|
|
/// \file index_hash.c
|
|
/// \brief Validates Index by using a hash function
|
|
//
|
|
// Author: Lasse Collin
|
|
//
|
|
// This file has been put into the public domain.
|
|
// You can do whatever you want with this file.
|
|
//
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
#include "common.h"
|
|
#include "index.h"
|
|
#include "check.h"
|
|
|
|
|
|
typedef struct {
|
|
/// Sum of the Block sizes (including Block Padding)
|
|
lzma_vli blocks_size;
|
|
|
|
/// Sum of the Uncompressed Size fields
|
|
lzma_vli uncompressed_size;
|
|
|
|
/// Number of Records
|
|
lzma_vli count;
|
|
|
|
/// Size of the List of Index Records as bytes
|
|
lzma_vli index_list_size;
|
|
|
|
/// Check calculated from Unpadded Sizes and Uncompressed Sizes.
|
|
lzma_check_state check;
|
|
|
|
} lzma_index_hash_info;
|
|
|
|
|
|
struct lzma_index_hash_s {
|
|
enum {
|
|
SEQ_BLOCK,
|
|
SEQ_COUNT,
|
|
SEQ_UNPADDED,
|
|
SEQ_UNCOMPRESSED,
|
|
SEQ_PADDING_INIT,
|
|
SEQ_PADDING,
|
|
SEQ_CRC32,
|
|
} sequence;
|
|
|
|
/// Information collected while decoding the actual Blocks.
|
|
lzma_index_hash_info blocks;
|
|
|
|
/// Information collected from the Index field.
|
|
lzma_index_hash_info records;
|
|
|
|
/// Number of Records not fully decoded
|
|
lzma_vli remaining;
|
|
|
|
/// Unpadded Size currently being read from an Index Record.
|
|
lzma_vli unpadded_size;
|
|
|
|
/// Uncompressed Size currently being read from an Index Record.
|
|
lzma_vli uncompressed_size;
|
|
|
|
/// Position in variable-length integers when decoding them from
|
|
/// the List of Records.
|
|
size_t pos;
|
|
|
|
/// CRC32 of the Index
|
|
uint32_t crc32;
|
|
};
|
|
|
|
|
|
extern LZMA_API(lzma_index_hash *)
|
|
lzma_index_hash_init(lzma_index_hash *index_hash,
|
|
const lzma_allocator *allocator)
|
|
{
|
|
if (index_hash == NULL) {
|
|
index_hash = lzma_alloc(sizeof(lzma_index_hash), allocator);
|
|
if (index_hash == NULL)
|
|
return NULL;
|
|
}
|
|
|
|
index_hash->sequence = SEQ_BLOCK;
|
|
index_hash->blocks.blocks_size = 0;
|
|
index_hash->blocks.uncompressed_size = 0;
|
|
index_hash->blocks.count = 0;
|
|
index_hash->blocks.index_list_size = 0;
|
|
index_hash->records.blocks_size = 0;
|
|
index_hash->records.uncompressed_size = 0;
|
|
index_hash->records.count = 0;
|
|
index_hash->records.index_list_size = 0;
|
|
index_hash->unpadded_size = 0;
|
|
index_hash->uncompressed_size = 0;
|
|
index_hash->pos = 0;
|
|
index_hash->crc32 = 0;
|
|
|
|
// These cannot fail because LZMA_CHECK_BEST is known to be supported.
|
|
(void)lzma_check_init(&index_hash->blocks.check, LZMA_CHECK_BEST);
|
|
(void)lzma_check_init(&index_hash->records.check, LZMA_CHECK_BEST);
|
|
|
|
return index_hash;
|
|
}
|
|
|
|
|
|
extern LZMA_API(void)
|
|
lzma_index_hash_end(lzma_index_hash *index_hash,
|
|
const lzma_allocator *allocator)
|
|
{
|
|
lzma_free(index_hash, allocator);
|
|
return;
|
|
}
|
|
|
|
|
|
extern LZMA_API(lzma_vli)
|
|
lzma_index_hash_size(const lzma_index_hash *index_hash)
|
|
{
|
|
// Get the size of the Index from ->blocks instead of ->records for
|
|
// cases where application wants to know the Index Size before
|
|
// decoding the Index.
|
|
return index_size(index_hash->blocks.count,
|
|
index_hash->blocks.index_list_size);
|
|
}
|
|
|
|
|
|
/// Updates the sizes and the hash without any validation.
|
|
static void
|
|
hash_append(lzma_index_hash_info *info, lzma_vli unpadded_size,
|
|
lzma_vli uncompressed_size)
|
|
{
|
|
info->blocks_size += vli_ceil4(unpadded_size);
|
|
info->uncompressed_size += uncompressed_size;
|
|
info->index_list_size += lzma_vli_size(unpadded_size)
|
|
+ lzma_vli_size(uncompressed_size);
|
|
++info->count;
|
|
|
|
const lzma_vli sizes[2] = { unpadded_size, uncompressed_size };
|
|
lzma_check_update(&info->check, LZMA_CHECK_BEST,
|
|
(const uint8_t *)(sizes), sizeof(sizes));
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
extern LZMA_API(lzma_ret)
|
|
lzma_index_hash_append(lzma_index_hash *index_hash, lzma_vli unpadded_size,
|
|
lzma_vli uncompressed_size)
|
|
{
|
|
// Validate the arguments.
|
|
if (index_hash == NULL || index_hash->sequence != SEQ_BLOCK
|
|
|| unpadded_size < UNPADDED_SIZE_MIN
|
|
|| unpadded_size > UNPADDED_SIZE_MAX
|
|
|| uncompressed_size > LZMA_VLI_MAX)
|
|
return LZMA_PROG_ERROR;
|
|
|
|
// Update the hash.
|
|
hash_append(&index_hash->blocks, unpadded_size, uncompressed_size);
|
|
|
|
// Validate the properties of *info are still in allowed limits.
|
|
if (index_hash->blocks.blocks_size > LZMA_VLI_MAX
|
|
|| index_hash->blocks.uncompressed_size > LZMA_VLI_MAX
|
|
|| index_size(index_hash->blocks.count,
|
|
index_hash->blocks.index_list_size)
|
|
> LZMA_BACKWARD_SIZE_MAX
|
|
|| index_stream_size(index_hash->blocks.blocks_size,
|
|
index_hash->blocks.count,
|
|
index_hash->blocks.index_list_size)
|
|
> LZMA_VLI_MAX)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
return LZMA_OK;
|
|
}
|
|
|
|
|
|
extern LZMA_API(lzma_ret)
|
|
lzma_index_hash_decode(lzma_index_hash *index_hash, const uint8_t *in,
|
|
size_t *in_pos, size_t in_size)
|
|
{
|
|
// Catch zero input buffer here, because in contrast to Index encoder
|
|
// and decoder functions, applications call this function directly
|
|
// instead of via lzma_code(), which does the buffer checking.
|
|
if (*in_pos >= in_size)
|
|
return LZMA_BUF_ERROR;
|
|
|
|
// NOTE: This function has many similarities to index_encode() and
|
|
// index_decode() functions found from index_encoder.c and
|
|
// index_decoder.c. See the comments especially in index_encoder.c.
|
|
const size_t in_start = *in_pos;
|
|
lzma_ret ret = LZMA_OK;
|
|
|
|
while (*in_pos < in_size)
|
|
switch (index_hash->sequence) {
|
|
case SEQ_BLOCK:
|
|
// Check the Index Indicator is present.
|
|
if (in[(*in_pos)++] != INDEX_INDICATOR)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
index_hash->sequence = SEQ_COUNT;
|
|
break;
|
|
|
|
case SEQ_COUNT: {
|
|
ret = lzma_vli_decode(&index_hash->remaining,
|
|
&index_hash->pos, in, in_pos, in_size);
|
|
if (ret != LZMA_STREAM_END)
|
|
goto out;
|
|
|
|
// The count must match the count of the Blocks decoded.
|
|
if (index_hash->remaining != index_hash->blocks.count)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
ret = LZMA_OK;
|
|
index_hash->pos = 0;
|
|
|
|
// Handle the special case when there are no Blocks.
|
|
index_hash->sequence = index_hash->remaining == 0
|
|
? SEQ_PADDING_INIT : SEQ_UNPADDED;
|
|
break;
|
|
}
|
|
|
|
case SEQ_UNPADDED:
|
|
case SEQ_UNCOMPRESSED: {
|
|
lzma_vli *size = index_hash->sequence == SEQ_UNPADDED
|
|
? &index_hash->unpadded_size
|
|
: &index_hash->uncompressed_size;
|
|
|
|
ret = lzma_vli_decode(size, &index_hash->pos,
|
|
in, in_pos, in_size);
|
|
if (ret != LZMA_STREAM_END)
|
|
goto out;
|
|
|
|
ret = LZMA_OK;
|
|
index_hash->pos = 0;
|
|
|
|
if (index_hash->sequence == SEQ_UNPADDED) {
|
|
if (index_hash->unpadded_size < UNPADDED_SIZE_MIN
|
|
|| index_hash->unpadded_size
|
|
> UNPADDED_SIZE_MAX)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
index_hash->sequence = SEQ_UNCOMPRESSED;
|
|
} else {
|
|
// Update the hash.
|
|
hash_append(&index_hash->records,
|
|
index_hash->unpadded_size,
|
|
index_hash->uncompressed_size);
|
|
|
|
// Verify that we don't go over the known sizes. Note
|
|
// that this validation is simpler than the one used
|
|
// in lzma_index_hash_append(), because here we know
|
|
// that values in index_hash->blocks are already
|
|
// validated and we are fine as long as we don't
|
|
// exceed them in index_hash->records.
|
|
if (index_hash->blocks.blocks_size
|
|
< index_hash->records.blocks_size
|
|
|| index_hash->blocks.uncompressed_size
|
|
< index_hash->records.uncompressed_size
|
|
|| index_hash->blocks.index_list_size
|
|
< index_hash->records.index_list_size)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
// Check if this was the last Record.
|
|
index_hash->sequence = --index_hash->remaining == 0
|
|
? SEQ_PADDING_INIT : SEQ_UNPADDED;
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
case SEQ_PADDING_INIT:
|
|
index_hash->pos = (LZMA_VLI_C(4) - index_size_unpadded(
|
|
index_hash->records.count,
|
|
index_hash->records.index_list_size)) & 3;
|
|
index_hash->sequence = SEQ_PADDING;
|
|
|
|
// Fall through
|
|
|
|
case SEQ_PADDING:
|
|
if (index_hash->pos > 0) {
|
|
--index_hash->pos;
|
|
if (in[(*in_pos)++] != 0x00)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
break;
|
|
}
|
|
|
|
// Compare the sizes.
|
|
if (index_hash->blocks.blocks_size
|
|
!= index_hash->records.blocks_size
|
|
|| index_hash->blocks.uncompressed_size
|
|
!= index_hash->records.uncompressed_size
|
|
|| index_hash->blocks.index_list_size
|
|
!= index_hash->records.index_list_size)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
// Finish the hashes and compare them.
|
|
lzma_check_finish(&index_hash->blocks.check, LZMA_CHECK_BEST);
|
|
lzma_check_finish(&index_hash->records.check, LZMA_CHECK_BEST);
|
|
if (memcmp(index_hash->blocks.check.buffer.u8,
|
|
index_hash->records.check.buffer.u8,
|
|
lzma_check_size(LZMA_CHECK_BEST)) != 0)
|
|
return LZMA_DATA_ERROR;
|
|
|
|
// Finish the CRC32 calculation.
|
|
index_hash->crc32 = lzma_crc32(in + in_start,
|
|
*in_pos - in_start, index_hash->crc32);
|
|
|
|
index_hash->sequence = SEQ_CRC32;
|
|
|
|
// Fall through
|
|
|
|
case SEQ_CRC32:
|
|
do {
|
|
if (*in_pos == in_size)
|
|
return LZMA_OK;
|
|
|
|
if (((index_hash->crc32 >> (index_hash->pos * 8))
|
|
& 0xFF) != in[(*in_pos)++]) {
|
|
#ifndef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION
|
|
return LZMA_DATA_ERROR;
|
|
#endif
|
|
}
|
|
|
|
} while (++index_hash->pos < 4);
|
|
|
|
return LZMA_STREAM_END;
|
|
|
|
default:
|
|
assert(0);
|
|
return LZMA_PROG_ERROR;
|
|
}
|
|
|
|
out:
|
|
// Update the CRC32.
|
|
//
|
|
// Avoid null pointer + 0 (undefined behavior) in "in + in_start".
|
|
// In such a case we had no input and thus in_used == 0.
|
|
{
|
|
const size_t in_used = *in_pos - in_start;
|
|
if (in_used > 0)
|
|
index_hash->crc32 = lzma_crc32(in + in_start,
|
|
in_used, index_hash->crc32);
|
|
}
|
|
|
|
return ret;
|
|
}
|