Implementation of huffman coding, Advanced Statistics

Assignment Help:

Input to the compress is a text le with arbitrary size, but for this assignment we will assume that the data structure of the file fits in the main memory of a computer. Output of the program is a compressed representation of the original file. You will have to save the codetable in the header of the compressed file, so that you can use the codetable for decompressing the compressed file. Input to the decompress is a compressed file, from which the program recovers the original file. For sanity

check, you should have a specific magic word at some position in the header of the compressed file, so that decompress can identify whether the given file is a valid Huffman compressed file. You should pay attention to the following issues:

The file that we will use for testing can be very large, having size in Gigabytes, so make sure that your program is bug-free and it works for large input le.

Write efficient algorithm, we will take off as much as 20 points if we feel that the program is taking unusually long time.

You must make sure that your program runs on a Linux Machine, and identically follows the formatting instructions. For formatting error, as much as 15 points can be taken off .

You must provide a Make file to compile your programs. Also, a README.txt le should be provided that will have the instruction to compile and run the programs. 


Related Discussions:- Implementation of huffman coding

Falsediscoveryrate (fdr), The approach of controlling the error rate in an ...

The approach of controlling the error rate in an exploratory analysis where number of hypotheses are tested, but where the strict control which is provided by multiple comparison p

Attack rate, Attack rate : This term frequently used for the incidence of t...

Attack rate : This term frequently used for the incidence of the disease or condition in the particular group, or during a limited interval of time, or under the special circumstan

Kaiser''s rule, Kaiser's rule is the  rule frequently used in the principa...

Kaiser's rule is the  rule frequently used in the principal components analysis for selecting the suitable the number of components. When the components are derived from correlati

Multidimensional scaling (mds), Multidimensional scaling (MDS)  is a generi...

Multidimensional scaling (MDS)  is a generic term for a class of techniques or methods which attempt to construct a low-dimensional geometrical representation of the proximity matr

Catastrophe theory, Catastrophe theory : A theory of how little is the cont...

Catastrophe theory : A theory of how little is the continuous changes in the independent variables which can have unexpected, discontinuous effects on the dependent variables. Exam

Uncertainty analysis, Uncertainty analysis is the process for assessing th...

Uncertainty analysis is the process for assessing the variability in the outcome variable that is due to the uncertainty in estimating the values of input parameters. A sensitivit

Times series plots, There is high level of fluctuation in a zigzag pattern ...

There is high level of fluctuation in a zigzag pattern in the time series for RESI1 which indicates that there is possibly negative autocorrelation present. Column C11 show

Mann whitney test, Mann Whitney test is a distribution free test which is ...

Mann Whitney test is a distribution free test which is used as an alternative to the Student's t-test for assessing that whether the two populations have the same median. The test

Ain why the simulated result doesn''t have to be exact as the, ain why the ...

ain why the simulated result doesn''t have to be exact as the theoretical calculation

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd