this dir | view | cards | source | edit | dark top

Exam

Exam
Introduction

data compression

Introduction

measuring the performance

Introduction

limits of lossless compression

Lossless Compression

basic concepts

Lossless Compression

typical strategy

Lossless Compression

codes

Lossless Compression

prefix codes

Lossless Compression

Shannon-Fano coding

Lossless Compression

algorithm

Lossless Compression

optimality of Huffman code

Lossless Compression

can we modify Huffman algorithm to guarantee that the resulting code minimizes the maximum codeword length?

yes, when constructing the tree, we can sort the nodes not only by their frequencies but also by the depth of their subtree

Lossless Compression

what is the maximum height of a Huffman tree for an input message of length nn?

Lossless Compression

generalize the construction of binary Huffman code to the case of mm-ary coding alphabet (m>2m\gt 2)

Lossless Compression

implementation notes


Lossless Compression

adaptive compression

Lossless Compression

if we know that the code contains 2 codewords of length 2 and 4 codewords of length 3, we can construct canonical Huffman code

00, 01, 100, 101, 110, 111

Lossless Compression

adaptive Huffman code

Lossless Compression

each string sAs\in A^* associate with a subinterval Is[0,1]I_s\subseteq[0,1] such that

Lossless Compression

encoding

Lossless Compression

decoding

Lossless Compression

problem: using floating point arithmetic leads to rounding errors → we will use integers

underflow may happen that way if we use too short integers

Lossless Compression

another problem

Lossless Compression

adaptive version

Lossless Compression

Hartley's formula

minimum and average number of yes/no questions to find the answer

Lossless Compression

Shannon's formula

Lossless Compression

Kraft-McMillan theorem

Lossless Compression

entropy of a discrete random vector

Lossless Compression

analysis of arithmetic coding

F(x)\overline{F(x)} … the midpoint of the interval F(x1),F(x)F(x-1),F(x)

Lossless Compression

problems

Lossless Compression

a better probabilistic model

Lossless Compression

according to an experimental estimate

the entropy of English is 1.3 bits per symbol

Lossless Compression

methods

Lossless Compression

Prediction by Partial Matching

Lossless Compression

unary code α\alpha

Lossless Compression

binary code β\beta

Lossless Compression

Elias codes

Lossless Compression

observation

each probability distribution p1p2pnp_1\geq p_2\geq\dots\geq p_n satisfies i[n]:pi1i\forall i\in[n]:p_i\leq \frac1i

Lossless Compression

idea

Lossless Compression

problems

Lossless Compression

LZ77

Lossless Compression

LZSS

Lossless Compression

other variants

Lossless Compression

Deflate algorithm

Lossless Compression

LZ77 disadvantages

Lossless Compression

LZ78

Lossless Compression

LZW

Lossless Compression

LZC

Lossless Compression

LZMW

Lossless Compression

LZAP

instead of ST add all substrings Sp where p is a prefix of T

Lossless Compression

digital images

Lossless Compression

color spaces

Lossless Compression

representation of bitmap images

Lossless Compression

GIF

Lossless Compression

PNG

Lossless Compression

interlacing schemes

Lossless Compression

Burrows-Wheeler transform

Lossless Compression

matrix n×nn×n is not neccessary

we can use suffix array

Lossless Compression

why does BW transform work?

Lossy Compression

how to measure the loss of information

Lossy Compression

capturing audio

Lossy Compression

quantization

Lossy Compression

vector quantization

Lossy Compression

differential encoding

Lossy Compression

transform coding

Lossy Compression

JPEG

Lossy Compression

WebP

Lossy Compression

example

Lossy Compression

can we reduce the problem to image compression?

Lossy Compression

basic types of algorithms

Lossy Compression

video signal representation

Lossy Compression

MPEG-1

Lossy Compression

H.261

Hurá, máš hotovo! 🎉
Pokud ti moje kartičky pomohly, můžeš mi koupit pivo.