K-nearest neighbor for text classification, Computer Engineering

Assignment Help:

Assignment 2: K-nearest neighbor for text classification.

The goal of text classification is to identify the topic for a piece of text (news article, web-blog, etc.). Text classification has obvious utility in the age of information overload, and it has become a popular turf for applying machine learning algorithms. In this project, you will have the opportunity to implement k-nearest neighbor and apply it to text classification on the well known Reuter news collection.

1.       Download the dataset from my website, which is created from the original collection and contains a training file, a test file, the topics, and the format for train/test.

2.       Implement the k-nearest neighbor algorithm for text classification. Your goal is to predict the topic for each news article in the test set. Try the following distance or similarity measures with their corresponding representations.

a.        Hamming distance: each document is represented as a boolean vector, where each bit represents whether the corresponding word appears in the document.

b.       Euclidean distance: each document is represented as a numeric vector, where each number represents how many times the corresponding word appears in the document (it could be zero).

c.         Cosine similarity with TF-IDF weights (a popular metric in information retrieval): each document is represented by a numeric vector as in (b). However, now each number is the TF-IDF weight for the corresponding word (as defined below). The similarity between two documents is the dot product of their corresponding vectors, divided by the product of their norms.

3.        Let w be a word, d be a document, and N(d,w) be the number of occurrences of w in d (i.e., the number in the vector in (b)). TF stands for term frequency, and TF(d,w)=N(d,w)/W(d), where W(d) is the total number of words in d. IDF stands for inverted document frequency, and IDF(d,w)=log(D/C(w)), where D is the total number of documents, and C(w) is the total number of documents that contains the word w; the base for the logarithm is irrelevant, you can use e or 2. The TF-IDF weight for w in d is TF(d,w)*IDF(d,w); this is the number you should put in the vector in (c). TF-IDF is a clever heuristic to take into account of the "information content" that each word conveys, so that frequent words like "the" is discounted and document-specific ones are amplified. You can find more details about it online or in standard IR text.

4.       You should try k = 1, k = 3 and k = 5 with each of the representations above. Notice that with a distance measure, the k-nearest neighborhoods are the ones with the smallest distance from the test point, whereas with a similarity measure, they are the ones with the highest similarity scores.

 

 


Related Discussions:- K-nearest neighbor for text classification

How to join to mwseries from power builder 8.0, Using CICS Transaction Gate...

Using CICS Transaction Gateway we can join MQ Series with Power Builder 8.0

Benefits of having densely packed integrated circuits, What are benefits of...

What are benefits of having densely packed Integrated Circuits? These are stated below: Reliability: The integrated circuit interconnections are in fact more reliable

DFD, Draw the Context level DFD for the Safe home Software.

Draw the Context level DFD for the Safe home Software.

Which way do twin twist props on a boat rotate, Q. Which way do twin twist ...

Q. Which way do twin twist props on a boat rotate when propelling the boat forward? Answer:- In opposite direction however they are designed to propel the boat forward eve

Example on public divisor, Q. Example on PUBLIC DIVISOR? Linker appends...

Q. Example on PUBLIC DIVISOR? Linker appends all segments having the same name and PUBLIC directive with segment name into one segment. Their contents are pulled together in co

Explain direct broadcast & limited broadcast, Explain Direct broadcast & li...

Explain Direct broadcast & limited broadcast. Broadcast is a method to send a packet to all the stat ions on an exact network at once. Broadcast systems permit the possibility

Learning abilities of perceptrons, Learning Abilities of Perceptrons: ...

Learning Abilities of Perceptrons: Conversely computational learning theory is the study of what concepts particular learning schemes as representation and method detail as a

Determine about the raster-scan systems, Determine about the raster-scan sy...

Determine about the raster-scan systems Most of the present raster-scan systems contain a separate processor called as display processor. This processor performs graphics func

State the difference between following, Q. State the difference between fo...

Q. State the difference between following. i. RAM and ROM ii. SRAM and DRAM iii. Dynamic and static MOS memories

Op code mnemonics in assembly language, Op code mnemonics: Instruction...

Op code mnemonics: Instructions  (statements)  in  assembly  language  are  usually  very  simple,  unlike  those that in  high-level languages. Usually, an opcode is a symbol

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd