From f5172e6b84f14a8d3306cbd7d92503281f7a3a80 Mon Sep 17 00:00:00 2001 From: Max Kellermann Date: Fri, 9 Dec 2016 22:21:31 +0100 Subject: [PATCH] decoder/wavpack: use int16_t instead of uint16_t Technically, these samples are signed. There was no practical difference, only the declaration was formally wrong. --- src/decoder/plugins/WavpackDecoderPlugin.cxx | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/decoder/plugins/WavpackDecoderPlugin.cxx b/src/decoder/plugins/WavpackDecoderPlugin.cxx index 1a8775765..fa307d5c8 100644 --- a/src/decoder/plugins/WavpackDecoderPlugin.cxx +++ b/src/decoder/plugins/WavpackDecoderPlugin.cxx @@ -104,7 +104,7 @@ format_samples_int(int bytes_per_sample, void *buffer, uint32_t count) break; } case 2: { - uint16_t *dst = (uint16_t *)buffer; + auto *dst = (int16_t *)buffer; static_assert(sizeof(*dst) <= sizeof(*src), "Wrong size"); /* pass through and align 16-bit samples */