compression

3 articles
sort: new top best
clear filter
0 2/10

This ICLR 2026 paper frames large language model training as lossy compression, demonstrating that LLMs learn optimal compressions of training data for next-sequence prediction that approach Information Bottleneck theoretical bounds. The work shows that compression quality and structure can predict downstream benchmark performance across different model families, providing an information-theoretic framework for understanding LLM learning and representational spaces.

ICLR 2026 Henry Conklin Tom Hosking Tan Yi-Chern Jonathan D. Cohen Sarah-Jane Leslie Thomas L. Griffiths Max Bartolo Seraphina Goldfarb-Tarrant OpenReview
openreview.net · pera · 1 day ago · details · hn
0 2/10

A developer discusses a tool for compressing large log files (600MB→10MB) while preserving semantic meaning for LLM analysis, addressing token limit constraints in AI-assisted log analysis.

DrTrader · 1 day ago · details · hn
0 6/10
technical-analysis

A detailed technical comparison of compression algorithms (gzip, zstd, xz, brotli, lzip, bzip2, bzip3) for optimizing code size in resource-constrained environments, demonstrating that bzip/bzip2 achieves superior compression ratios for text-like data through Burrows-Wheeler Transform rather than LZ77, while maintaining a smaller decoder footprint.

bzip2 bzip3 LibDeflate ComputerCraft zstd xz gzip brotli lzip purplesyringa
purplesyringa.moe · enz · 1 day ago · details · hn