Data Compression
Data compression and compression formats can have a significant impact on performance. Three important places to consider data compression are in MapReduce and Spark jobs, data stored in HBase, and Impala queries. For the most part, the principles are similar for each.
You must balance the processing capacity required to compress and uncompress the data, the disk IO required to read and write the data, and the network bandwidth required to send the data across the network. The correct balance of these factors depends upon the characteristics of your cluster and your data, as well as your usage patterns.
Compression is not recommended if your data is already compressed (such as images in JPEG format). In fact, the resulting file can sometimes be larger than the original.
For more information about compression algorithms in Hadoop, see the Compression section of the Hadoop I/O chapter in Hadoop: The Definitive Guide.
Compression Types
Hadoop supports the following compression types and codecs:
- gzip - org.apache.hadoop.io.compress.GzipCodec
- bzip2 - org.apache.hadoop.io.compress.BZip2Codec
- LZO - com.hadoop.compression.lzo.LzopCodec
- Snappy - org.apache.hadoop.io.compress.SnappyCodec
- Deflate - org.apache.hadoop.io.compress.DeflateCodec
Different file types and CDH components support different compression types. For details, see Avro Data Files and Parquet Files.
For guidelines on choosing compression types and configuring compression, see Choosing and Configuring Data Compression.
<< Parquet Files | ©2016 Cloudera, Inc. All rights reserved | Snappy Compression >> |
Terms and Conditions Privacy Policy |