You compressed something which is truly natively a bitfield using regular compression algorithms? That is expected to get horrible results. Much better would be something which handles it natively, say doing run length encoding on the number of repeated bits and compressing that using elias omega encoding. That is suboptimal in a few ways but has the advantage of working well both on things which are mostly zeros or mostly ones, and only performs badly on truly random bits.It isn't super clear how relevant this information is. The TXO bitfield is fairly small to begin with, and to compress the data in real time would require a special data structure which gets worse compression than straight compressing the whole thing and has slower lookups than an uncompressed version. Writing such a thing sounds like an interesting project though.On Wed, May 23, 2018 at 4:48 PM, Jim Posen via bitcoin-dev <bitcoin-dev@lists.linuxfoundation.org> wrote:I decided to look into the metrics around compression ratios of TXO bitfields, as proposed by Bram Cohen [1]. I'm specifically interested in the feasibility of committing to them with block headers. In combination with block commitments to TXOs themselves, this would enable UTXO inclusion/exclusion proofs for light clients.First, looking just at proofs of inclusion in the UTXO set, each block needs what Bram calls a "proof of position." Concretely, one such construction is a Merkle root over all of the block's newly created coins, including their output data (scriptPubKey + amount), the outpoint (txid + index), and an absolute index of the output in the entire blockchain. A Merkle branch in this tree constitutes a proof of position. Alternatively, the "position", rather than being an absolute index in the chain, could be a block hash plus an output index within the block.Let's say we use the absolute index in the chain as position. A TXO spentness bitfield can be constructed for the entire chain, which is added to when new coins are created and modified when they are spent. In order to compactly prove spentness in this bitfield to a client, one could chunk up the bitfield and construct a Merkle Mountain Range [2] over the chunks. Instead of building an MMR over outputs themselves, as proposed by Peter Todd [3], an MMR constructed over bitfield chunks grows far slower, by a large constant factor. Slower growth means faster updates.So there's the question of how much these bitfields can be compressed. We expect some decent level because patterns of spending coins are very non-random.The top graph in the attached figure shows the compression ratios possible on a TXO bitfield split into 4 KiB chunks, using gzip (level=9) and lz4. Data was collected at block height 523,303. You can see that the compression ratio is much lower for older chunks and is worse for more recent blocks. Over the entire history, gzip achieves 34.4%, lz4 54.8%, and bz2 37.6%. I'm kind of surprised that the ratios are not lower with off-the-shelf algorithms. And that gzip performs better than bz2 (it seems to be a factor of the chunk size?).Alternatively, we can look at bitfields stored separately by block, which is more compatible with constructions where an output's position is its block hash plus relative index. The per-block bitfield sizes are shown in the bottom graph. The compression ratios overall are 50% for gzip, 70% for lz4, and 61.5% for bz2.
_______________________________________________
bitcoin-dev mailing list
bitcoin-dev@lists.linuxfoundation.org
https://lists.linuxfoundation.org/mailman/listinfo/bitcoin-dev