Lossless JPEG is a 1993 addition to JPEG standard by the Joint Photographic Experts Group to enable lossless compression. However, it might be used as an umbrella term to refer to all lossless compression schemes developed by the Joint Photographic Expert group. They include JPEG 2000 and JPEGLS.
Figure 1:
DPCM encoder model
Figure 2: Simplified block diagrams for lossless mode
Figure 3: Three neighboring samples around the sample to be predicted
Lossless JPEG was developed as a late addition to JPEG in 1993, using a completely different technique from the lossy JPEG standard. It uses a predictive scheme based on the three nearest (causal) neighbors (upper, left, and upperleft), and entropy coding is used on the prediction error. The standard Independent JPEG Group libraries cannot encode or decode it, but Ken Murchison of Oceana Matrix Ltd. wrote a patch that extends the IJG library to handle Lossless JPEG. Lossless JPEG has some popularity in medical imaging, and is used in DNG and some digital cameras to compress raw images, but otherwise was never widely adopted.
Contents

Lossless mode of operation 1

JPEGLS 2

LOCOI algorithm 2.1

Decorrelation/prediction 2.1.1

Context modeling 2.1.2

Coding corrected prediction residuals 2.1.3

Run length coding in uniform areas 2.1.4

JPEG 2000 3

References 4

External links 5
Lossless mode of operation
Lossless JPEG^{[1]} is actually a mode of operation of JPEG. This mode exists because the discrete cosine transform (DCT) based form cannot guarantee that encoder input would exactly match decoder output. Unlike the lossy mode which is based on the DCT, the lossless coding process employs a simple predictive coding model called differential pulse code modulation (DPCM). This is a model in which predictions of the sample values are estimated from the neighboring samples that are already coded in the image. Most predictors take the average of the samples immediately above and to the left of the target sample. DPCM encodes the differences between the predicted samples instead of encoding each sample independently. The differences from one sample to the next are usually close to zero. A typical DPCM encoder is displayed in Fig.1. The block in the figure acts as a storage of the current sample which will later be a previous sample.
The main steps of lossless operation mode are depicted in Fig.2. In the process, the predictor combines up to three neighboring samples at A, B, and C shown in Fig.3 in order to produce a prediction of the sample value at the position labeled by X. The three neighboring samples must be already encoded samples. Any one of the predictors shown in the table below can be used to estimate the sample located at X.^{[2]} Any one of the eight predictors listed in the table can be used. Note that selections 1, 2, and 3 are onedimensional predictors and selections 4, 5, 6, and 7 are twodimensional predictors. The first selection value in the table, zero, is only used for differential coding in the hierarchical mode of operation.
Once all the samples are predicted, the differences between the samples can be obtained and entropycoded in a lossless fashion using Huffman coding or arithmetic coding.
Selectionvalue

Prediction

0

No prediction

1

A

2

B

3

C

4

A + B – C

5

A + (B – C)/2

6

B + (A – C)/2

7

(A + B)/2

Typically, compressions using lossless operation mode can achieve around 2:1 compression ratio for color images.^{[3]} This mode is quite popular in the medical imaging field, and defined as an option in DNG standard, but otherwise it is not very widely used because of complexity of doing arithmetics on 10, 12 or 14bpp values on typical embedded 32bit processor and a little resulting gain in space.
JPEGLS
JPEGLS is a lossless/nearlossless compression standard for continuoustone images.^{[4]} Its official designation is ISO144951/ITUT.87.^{[5]} It is a simple and efficient baseline algorithm which consists of two independent and distinct stages called modeling and encoding. JPEGLS was developed with the aim of providing a lowcomplexity lossless and nearlossless image compression standard that could offer better compression efficiency than lossless JPEG. It was developed because at the time, the Huffman codingbased JPEG lossless standard and other standards were limited in their compression performance. Total decorrelation cannot be achieved by first order entropy of the prediction residuals employed by these inferior standards. JPEGLS, on the other hand, can obtain good decorrelation.^{[6]}^{[7]} Part 1 of this standard was finalized in 1999. Part 2, released in 2003, introduced extensions such as arithmetic coding. The core of JPEGLS is based on the LOCOI algorithm,^{[8]} that relies on prediction, residual modeling and contextbased coding of the residuals. Most of the low complexity of this technique comes from the assumption that prediction residuals follow a twosided geometric distribution (also called a discrete Laplace distribution) and from the use of Golomblike codes, which are known to be approximately optimal for geometric distributions. Besides lossless compression, JPEGLS also provides a lossy mode ("nearlossless") where the maximum absolute error can be controlled by the encoder. Compression for JPEGLS is generally much faster than JPEG 2000 and much better than the original lossless JPEG standard.
LOCOI algorithm
Prior to encoding, there are two essential steps to be done in the modeling stage: decorrelation (prediction) and error modeling.
Decorrelation/prediction
In the LOCOI algorithm, primitive edge detection of horizontal or vertical edges is achieved by examining the neighboring pixels of the current pixel X as illustrated in Fig.3. The pixel labeled by B is used in the case of a vertical edge while the pixel located at A is used in the case of a horizontal edge. This simple predictor is called the Median Edge Detection (MED) predictor^{[9]} or LOCOI predictor. The pixel X is predicted by the LOCOI predictor according to the following guesses:

X=\left\{ \begin{align} & \min (A,B)\quad \,\mbox{if}\,C\ge \max (A,B) \\ & \max (A,B)\quad \mbox{if}\,C\le \min (A,B) \\ & A+BC\quad \,\mbox{otherwise}. \\ \end{align} \right.
The three simple predictors are selected according to the following conditions: (1) it tends to pick B in cases where a vertical edge exists left of the X, (2) A in cases of an horizontal edge above X, or (3) A + B – C if no edge is detected.
Context modeling
The JPEGLS algorithm estimates the conditional expectations of the prediction errors E\left\{ eCtx \right\} using corresponding sample means \bar{e}(C) within each context Ctx. The purpose of context modeling is that the higher order structures like texture patterns and local activity of the image can be exploited by context modeling of the prediction error. Contexts are determined by obtaining the differences of the neighboring samples which represents the local gradient:

\begin{align} & g_1 =DB \\ & g_2 =BC \\ & g_3 =CA \end{align}
The local gradient reflects the level of activities such as smoothness and edginess of the neighboring samples. Notice that these differences are closely related to the statistical behavior of prediction errors. Each one of the differences found in the above equation is then quantized into roughly equiprobable and connected regions. For JPEGLS, the differences g1, g2, and g3 are quantized into 9 regions and the region are indexed from −4 to 4. The purpose of the quantization is to maximize the mutual information between the current sample value and its context such that the highorder dependencies can be captured. One can obtain the contexts based on the assumption that

P(e\mid Ctx=[q_1,q_2,q_3]) = P(e\mid Ctx=[q_1,q_2,q_3])
After merging contexts of both positive and negative signs, the total number of contexts is ((2\times 4+1)^3+1)/2=365 contexts. A bias estimation could be obtained by dividing cumulative prediction errors within each context by a count of context occurrences. In LOCOI algorithm, this procedure is modified and improved such that the number of subtractions and additions are reduced. The divisionfree bias computation procedure is demonstrated in [3]. Prediction refinement can then be done by applying these estimates in a feedback mechanism which eliminates prediction biases in different contexts.
Coding corrected prediction residuals
In the regular mode of JPEGLS, the standard uses Golomb–Rice codes which are a way to encode nonnegative run lengths. Its special case with the optimal encoding value 2^{k} allows simpler encoding procedures.
Run length coding in uniform areas
Since Golomb–Rice codes are quite inefficient for encoding low entropy distributions because the coding rate is at least one bit per symbol, significant redundancy may be produced because the smooth regions in an image can be encoded at less than 1 bit per symbol. To avoid having excess code length over the entropy, one can use alphabet extension which codes blocks of symbols instead of coding individual symbols. This spreads out the excess coding length over many symbols. This is the “run” mode of JPEGLS and it is executed once a flat or smooth context region characterized by zero gradients is detected. A run of west symbol “a” is expected and the end of run occurs when a new symbol occurs or the end of line is reached. The total run of length is encoded and the encoder would return to the “regular” mode.
JPEG 2000
JPEG 2000 includes a lossless mode based on a special integer wavelet filter (biorthogonal 3/5). JPEG 2000's lossless mode runs more slowly and has often worse compression ratios than JPEGLS on artificial and compound images.^{[10]}^{[11]} JPEG 2000 fares better than the UBC implementation of JPEGLS on digital camera pictures.^{[12]} JPEG 2000 is also scalable, progressive, and more widely implemented.
References

^ Pennebaker, W. B. & Mitchell, J. L. (1993). JPEG Still Image Data Compression Standard. New York: Van Nostrand Reinhold.

^ ITUT. ISO DIS 109181 Digital compression and coding of continuoustone still images (JPEG). Recommendation T.81.

^ Wallace, C. K. (1991). "The JPEG still picture compression standard". Communications of the ACM 34 (4): 31–44.

^ "HP Labs LOCOI/JPEGLS Home Page". HewlettPackard Development Company. May 13, 2005. Retrieved 24 October 2011.

^ "T.87 : Information technology – Lossless and nearlossless compression of continuoustone still images – Baseline". International Telecommunication Union. Retrieved 24 October 2011.

^ M. J. Weinberger, G. Seroussi, and G. Sapiro, “LOCOI: A low complexity, contextbased, lossless image compression algorithm,” in Proc. 1996 Data Compression Conference, Snowbird, UT, Mar. 1996, pp. 140–149.

^ M. Weinberger, G. Seroussi, and G. Sapiro, “The LOCOI lossless image compression algorithm: Principles and standardization into JPEGLS,” IEEE Trans. Image Processing, vol. 9, no. 8, pp. 1309–1324, Aug. 2000, originally as HewlettPackard Laboratories Technical Report No. HPL98193R1, November 1998, revised October 1999. Available from [4].

^ http://www.hpl.hp.com/loco/HPL98193R1.pdf LOCOI algorithm

^ Memon, Nasir D.; Wu, Xiaolin; Sippy, V. & Miller, G. (1997). "Interband coding extension of the new lossless JPEG standard". Proc. SPIE Int. Soc. Opt. Eng. 3024 (47): 47–58.

^ http://www.jpeg.org/public/wg1n1816.pdf

^ http://itohws03.ee.noda.sut.ac.jp/~matsuda/mrp/

^ http://www.imagecompression.info/gralic/LPCB.html
External links

JPEG 2000 still image coding versus other standards

JPEG2000, JPEGLS and other lossless codecs on greyscale images

JPEGLS home page

LOCOI home page

Licensing terms for HP's LOCO technology in JPEGLS (free reg. req.; not sublicensable; available only to companies)

Links to Various Implementations

Singletone/grayscale JPEGLS encoder algorithm

Evaluation of JPEGLS, IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 39, NO. 10, OCTOBER 2001
This article was sourced from Creative Commons AttributionShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and USA.gov, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for USA.gov and content contributors is made possible from the U.S. Congress, EGovernment Act of 2002.
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a nonprofit organization.