r

SUBMISSION REQUIREMENTS: Please submit a single R script file named with your “First_Last Name.R” ONLY.  Your R script code must calculate the effectivness of your classification as described below.

Similar to the classification example.  process and classify the newsgroup document data. Download this data  and save it on your computer in your R packages folder under “tm/text/”. Your code MUST access it from there!

Don't use plagiarized sources. Get Your Custom Essay on
r
Just from $13/Page
Order Essay

Note that the data is separated into one test and one train folder, each containing 20 sub folders on different subjects.

Choose these 2 subjects to analyze (sci.space and rec.autos) and 100 documents from each.

  

Consider “rec.autos” as positive and “sci.space” as negative event. Note that kNN  syntax expects (Positive First, Negative second)

Classify the Newsgroups data (by date version data set) from Blackboard:

•        Save data in your “tm/text/” folder so you can specify path using system.file()

•       Note that the data is separated into one test and one train folder, each containing 20 sub folders on different subjects.

Choose these 2 subjects to analyze (sci.space and rec.autos) and 100 documents from each.

•        For each subject select:

–       100 documents for training from the train folder

–       100 documents for testing from the test folder

•        Obtain the merged Corpus (of 400 documents), please keep the order as

–       Doc1.Train from the “sci.space” newsgroup train data

–       Doc1.Test from the “sci.space” newsgroup test data

–       Doc2.Train from the ” rec.autos” newsgroup train data

–       Doc2.Test from the ” rec.autos” newsgroup test data

•        Implement preprocessing (clearly indicate what you have used)

•        Create the Document-Term Matrix using the following arguments (word lengths of at least 2, word frequency of at least 5)

–      use: control=list(wordLengths=c(2,Inf), bounds=list(global=c(5,Inf)))

•        Split the Document-Term Matrix into proper test/train row ranges

–       train range containing rows (1:100) and (201:300)

–       test range  containing rows (101:200) and (301:400)

–       Note that knn expects the positive (“Rec”) event as first, so re-adjust your train/test range if necessary.  

•        Use the abbreviations “Positive” and “Negative” as tag factors in your classification.

–       Check if the tag order is correct using table(Tags)

–       You should get

•        Tags

•        Positive Negative

•        100      100

–       If your order is not right make proper changes.

•        Classify text using the kNN() function

•        Display classification results as a R dataframe and name the columns as:

–       “Doc”

–       “Predict”  – Tag factors of predicted subject (Positive or Negative)

–       “Prob” – The classification probability

–       “Correct’ – TRUE/FALSE

•        What is the percentage of correct (TRUE) classifications?

•        Estimate the effectiveness of your classification:

– Calculate and  clearly mark the values TP, TN, FP, FN

–       Create the confusion matrix and name the rows and columns with what is Positive/Negative event

–       Calculate Precision

–       Calculate Recall

–       Calculate F-score

Note that one way you can select only 100 documents is

> Temp1 <- DirSource(Doc1.TestPath)

> Doc1.Train <- Corpus(URISource(Temp1$filelist[1:100]),readerControl=list(reader=readPlain))

# Text Mining Exercises
rm(list=ls()); cat(“\014”) # Clear Workspace and Console
library(tm) # Load the Text Mining package
### ====== Example ====
# Example R Script: Extracting text content from text files and load them as Corpus
loremipsum <- system.file("texts", "loremipsum.txt", package = "tm") # Path to "loremipsum.txt" ovid <- system.file("texts", "txt", "ovid_1.txt", package = "tm") # Path to "ovid.txt" Docs.pth <- URISource(sprintf("file://%s", c(loremipsum, ovid))) # Specify Source corpus.txt<-VCorpus(Docs.pth) # load them as Corpus inspect(corpus.txt) corpus.txt[[2]]$content[1:3] # Examine the first 3 lines of the "ovid.txt" corpus class(corpus.txt) corpus.txt[[1]]$content # Displays content of loremipsum.txt corpus.txt[[1]]$meta # Editable Metadata corpus.txt[[1]]$meta$author <- "Ovid" ### --- Example 3: Processing - Transforming the Corpus ---- getTransformations() # The basic transforms available within tm package # Apply the function "tm_map()" and content_transformer() to the corpus docs.tranf <- tm_map( corpus.txt, content_transformer(tolower)) # Conversion to Lowercase corpus.txt[[2]]$content[1:3] # Original corpus with Uppercase docs.tranf[[2]]$content[1:3] # Transformed Corpus all Lowercase # Transforms the "@" into " ". email.texts <- c("Jon_Smith@bu.edu", "Subject: Speaker's Info", "I hope you enjoyed the lectures.", "Here is the address of the lecturer: ", "123 Main St. #25","Boston MA 02155") # Step 1: Create corpus corpus.txt <- VCorpus(VectorSource(data.frame(email.texts))) # email.df <- data.frame(doc_id=row.names(1:length(email.texts)),text=email.texts) # corpus.txt <- Corpus(DataframeSource(email.df)) # Step 2: "content transformer()" crate a function to achieve the transformation transform.chr <- content_transformer(function(x, pattern) gsub(pattern, " ", x)) docs.tranf <- tm_map(corpus.txt, transform.chr, "@") corpus.txt[[1]]$content[1] # "Jon_Smith@bu.edu" docs.tranf[[1]]$content[1] # "Jon_Smith bu.edu" ### --- Example 4: Replacing a word with another one ---- transform.wors <- content_transformer(function(x, from, to) gsub(from, to, x)) docs.tranf <- tm_map(corpus.txt, transform.wors, "Jon_Smith", "Jane_Doe") corpus.txt[[1]]$content[1] # "Jon_Smith@bu.edu" docs.tranf[[1]]$content[1] # "Jane_Doe@bu.edu" library(SnowballC) # Stemming docs.tranf <- tm_map(corpus.txt, stemDocument) # Stemm original corpus corpus.txt[[2]]$content[1] # "Subject: Speaker's Info" docs.tranf[[2]]$content[1] # "Subject: Speaker Info" ### --- Example 5: Creating a Document Term Matrix ---- Docs.pth <- system.file(file.path("doc", "tm "), package = "tm") # Path to tm Docs.corpus <- Corpus(URISource(Docs.pth),readerControl = list(reader = readPDF(engine = "xpdf"))) dtm <- DocumentTermMatrix(Docs.corpus) # Document Term Matrix freq <- colSums(as.matrix(dtm)) # Term frequencies max(freq) # Max appearance frequency of a term (84) findFreqTerms(dtm,max(freq),max(freq)) # Find the term with max appearance ("the") ord <- order(freq) # Ordering the frequencies (ord contains the indices)) freq[tail(ord)] # Most frequent terms & their frequency (most frequent term "the" appearing 85 times) # with for and corpus text the # 17 22 23 25 26 85 findFreqTerms(dtm, lowfreq=10) # List terms (alphabetically) with frequency higher than 10 # "and" "are" "can" "corpus" "documents" "for" "metadata" "mining" "terms" "text" "the" "this" "which" "with" findFreqTerms(dtm, lowfreq=17) # List terms (alphabetically) with frequency higher than 17 # "and" "can" "corpus" "for" "text" "the" "with" # Create a Histogram library(ggplot2) WordFreq <- data.frame(word=names(freq), freq=freq) p <- ggplot(subset(WordFreq, freq>7), aes(word, freq));
p <- p + geom_bar(stat="identity") p <- p + theme(axis.text.x=element_text(angle=45, hjust=1)) p # Display Histogram # Create a Word Cloud Plot library(wordcloud) set.seed(123) wordcloud(names(freq), freq, min.freq=5, colors=brewer.pal(6, "Dark2")) matdtm <- as.matrix(dtm) # Convert dtm to a matrix write.csv(matdtm, file=" dtm.csv") # Save matrix as CSV file ### --- Example 6: Creating a Matrix ---- A <- matrix(1:9, nrow=3,ncol=3) # Matrix A (3x3) B <- matrix(rep(10,6), nrow=3,ncol=2) # Matrix B (3x2) B*B # Element wise multiplication A%*% B # Matrix multiplication a <- matrix(1:3,nrow=3,ncol=1) # Row Vector a t(a)%*%B # Vector a is transposed first ### --- Example 7: String Operations ---- unlist(strsplit("a.b.c", "\\.")) # Split the elements of a character vector # "a" "b" "c" library(tau) # Using tau package tokenize("abc defghk") # "abc" " " "defghk" nchar("Web Analytics") # Counting the number of characters in a string # Detecting a pattern in a string. library(stringr) # Using stringr package string1 <- "Web Analytics" string2 <- "Data Analytics" str_detect(string1, "Analytics") # "TRUE" str_detect(string2, " Web") # "FALSE" # Date Strings Manipulations string1 <- "12 may 2014" string2 <- "1 may 2014" regexp <- "([[:digit:]]{2}) ([[:alpha:]]+) ([[:digit:]]{4})" # define Date pattern to have 2 Day digits, text as Month, 4 digit Year str_detect(string1, regexp) # "TRUE" - Using stringr package function "str_detect" grepl(pattern = regexp, x = string1) # "TRUE" - Using base R package function "grepl" grepl(pattern = regexp, x = string2) # "FALSE" - the day in the regexp was defined to have 2 digits ### --- Example 8: Fuzzy String Matching ---- # Using The Jaccard measure library(stringdist) dict1 <- "analytics" dict2 <- "analysis" query <- "analisis" A <- unlist(strsplit(dict1,"")) # Get the set of characters in dict1 B <- unlist(strsplit(dict2,"")) # Get the set of characters in dict2 Q <- unlist(strsplit(query,"")) # Get the set of characters in query UA <- union(A,Q) IA <- intersect(A,Q) JA <- length(IA)/length(UA) # The Jaccard measure sprintf("%s %f","The Jaccard Q-A measure is",JA) # Display calculated measure # "The Jaccard Q-A measure is 0.625000" UB <- union(B,Q) IB <- intersect(B,Q) JB <- length(IB)/length(UB) # The Jaccard measure sprintf("%s %f","The Jaccard Q-B measure is",JB) # Display calculated measure # "The Jaccard Q-B measure is 0.833333" ### --- Example 9: Document Clustering by Most Frequent Strings ---- # Assume that you have several (3) documents with few most frequent strings in them and you would like to group them together by similarity. library(tm) doc1 <- c("Web Analytics", "Text Analysis", "Web Mining", "Text Mining") doc2 <- c("Data Processing", "Machine Learning", "learn from data", "Big Data") doc3 <- c("bedroom furniture", "dining room furniture", "diner chair", "new chairs") doc <- c(doc1,doc2,doc3) # Merge all terms from all the Documents dtm <- as.matrix(DocumentTermMatrix(Corpus(VectorSource(doc)))) # Document term matrix colnames(dtm) # Gives All of the terms contained in the document term matrix # Using Jaccard Distance to find distance between terms and corresponding documents library(cluster) # Load similarity measures package d <- dist(dtm, method="binary") # Find distance between terms in dtm cl <- hclust(as.dist(d), method="ward.D2") # Perform clustering cl$labels=doc # Assign labels (terms used) to cluster leaves plot(cl,main="Jaccard Distance",xlab="Term Clustering Dendrogram using",sub = "...") # Set plot title # plot.new() # plot(cl, hang=-1) # groups <- cutree(cl, k=3) # "k=" defines the number of clusters you are using rect.hclust(cl, k=3, border="red") # draw dendogram with red borders around the 5 clusters # With stemmed terms corpus.temp <- tm_map(Corpus(VectorSource(doc)), stemDocument, language = "english") dtm1 <- as.matrix(DocumentTermMatrix(corpus.temp)) d <- dist(dtm1, method="binary") # Find distance between terms in dtm cls <- hclust(as.dist(d)) # Perform clustering cls$labels=doc # Assign labels (terms used) to cluster leaves plot(cls,main="Jaccard Distance with Stemming",xlab="Term Clustering Dendrogram using",sub = "...") # Set plot title rect.hclust(cls, k=5, border="red") # draw dendogram with red borders around the 5 clusters # Hierarchical clustering library(stringdist) d <- stringdistmatrix(doc, doc) # Pairwise string distances (optimal string alignment) cl2 <- hclust(as.dist(d)) # Perform hierarchical clustering cl2$labels=doc # Assign labels to cluster leaves plot(cl2,main="Optimal String Alignment",xlab="Term Clustering Dendrogram using",sub = "...") ### --- Example 10: Document kNN Text Classification ---- library("tm") Doc1 <- "I spent 10K on my car. Compared to the prices of most cars in their class it was cheap. It is a red car so I like it and it has a lot of space." Doc2 <- "I checked the car prices and I could not find a red car for under 10K. So the price was good even though it had a hole. I heard that it belonged to a movie star." Doc3 <- "I like the red color, so I would buy a red car even if the car's price is over 10K." Doc4 <- "I don't like red cars. The insurance for red cars is higher regardless of the price and I would not spend more than 10K. I like black cars." Doc5 <- "A red giant star can curve the space to form a black hole. In absence of stars the space is flat." Doc6 <- "With exception of the stars the space is filled with blackness making the black holes even harder to see." Doc7 <- "Our sun is a small star and it will not end as a black hole. It does not have enough mass to curve the space." Doc8 <- "Very few stars will end as black holes but still the space contains large number of black holes." doc <- c(Doc1,Doc2,Doc3,Doc4,Doc5,Doc6,Doc7,Doc8) # Merge all strings corpus <- Corpus(VectorSource(doc)) # Preprocessing corpus.temp <- tm_map(corpus, removePunctuation) # Remove Punctuation corpus.temp <- tm_map(corpus.temp, stemDocument, language = "english")# Perform Stemming dtm <- as.matrix(DocumentTermMatrix(corpus.temp)) # Document term matrix # Text Classification library(class) # Using kNN train <- dtm[c(1,2,5,6),] # Dataset for which classification is already known test <- dtm[c(3,4,7,8),] # Dataset you are trying to classify Tags <- factor(c(rep("cars",2), rep("space",2))) # Tags - Correct answers for the training dataset prob.test<- knn(train , test , Tags, k = 2, prob=TRUE) # k-number of neighbors considered # Display Classification Results a <- 1:length(prob.test) b <- levels(prob.test)[prob.test] c <- attributes(prob.test)$prob result <- data.frame(Doc=a, Predict=b,Prob=c) result sum(c)/length(Tags) # Overall probability sum(prob.test==Tags)/length(Tags) # % Correct Classification

What Will You Get?

We provide professional writing services to help you score straight A’s by submitting custom written assignments that mirror your guidelines.

Premium Quality

Get result-oriented writing and never worry about grades anymore. We follow the highest quality standards to make sure that you get perfect assignments.

Experienced Writers

Our writers have experience in dealing with papers of every educational level. You can surely rely on the expertise of our qualified professionals.

On-Time Delivery

Your deadline is our threshold for success and we take it very seriously. We make sure you receive your papers before your predefined time.

24/7 Customer Support

Someone from our customer support team is always here to respond to your questions. So, hit us up if you have got any ambiguity or concern.

Complete Confidentiality

Sit back and relax while we help you out with writing your papers. We have an ultimate policy for keeping your personal and order-related details a secret.

Authentic Sources

We assure you that your document will be thoroughly checked for plagiarism and grammatical errors as we use highly authentic and licit sources.

Moneyback Guarantee

Still reluctant about placing an order? Our 100% Moneyback Guarantee backs you up on rare occasions where you aren’t satisfied with the writing.

Order Tracking

You don’t have to wait for an update for hours; you can track the progress of your order any time you want. We share the status after each step.

image

Areas of Expertise

Although you can leverage our expertise for any writing task, we have a knack for creating flawless papers for the following document types.

Areas of Expertise

Although you can leverage our expertise for any writing task, we have a knack for creating flawless papers for the following document types.

image

Trusted Partner of 9650+ Students for Writing

From brainstorming your paper's outline to perfecting its grammar, we perform every step carefully to make your paper worthy of A grade.

Preferred Writer

Hire your preferred writer anytime. Simply specify if you want your preferred expert to write your paper and we’ll make that happen.

Grammar Check Report

Get an elaborate and authentic grammar check report with your work to have the grammar goodness sealed in your document.

One Page Summary

You can purchase this feature if you want our writers to sum up your paper in the form of a concise and well-articulated summary.

Plagiarism Report

You don’t have to worry about plagiarism anymore. Get a plagiarism report to certify the uniqueness of your work.

Free Features $66FREE

  • Most Qualified Writer $10FREE
  • Plagiarism Scan Report $10FREE
  • Unlimited Revisions $08FREE
  • Paper Formatting $05FREE
  • Cover Page $05FREE
  • Referencing & Bibliography $10FREE
  • Dedicated User Area $08FREE
  • 24/7 Order Tracking $05FREE
  • Periodic Email Alerts $05FREE
image

Our Services

Join us for the best experience while seeking writing assistance in your college life. A good grade is all you need to boost up your academic excellence and we are all about it.

  • On-time Delivery
  • 24/7 Order Tracking
  • Access to Authentic Sources
Academic Writing

We create perfect papers according to the guidelines.

Professional Editing

We seamlessly edit out errors from your papers.

Thorough Proofreading

We thoroughly read your final draft to identify errors.

image

Delegate Your Challenging Writing Tasks to Experienced Professionals

Work with ultimate peace of mind because we ensure that your academic work is our responsibility and your grades are a top concern for us!

Check Out Our Sample Work

Dedication. Quality. Commitment. Punctuality

Categories
All samples
Essay (any type)
Essay (any type)
The Value of a Nursing Degree
Undergrad. (yrs 3-4)
Nursing
2
View this sample

It May Not Be Much, but It’s Honest Work!

Here is what we have achieved so far. These numbers are evidence that we go the extra mile to make your college journey successful.

0+

Happy Clients

0+

Words Written This Week

0+

Ongoing Orders

0%

Customer Satisfaction Rate
image

Process as Fine as Brewed Coffee

We have the most intuitive and minimalistic process so that you can easily place an order. Just follow a few steps to unlock success.

See How We Helped 9000+ Students Achieve Success

image

We Analyze Your Problem and Offer Customized Writing

We understand your guidelines first before delivering any writing service. You can discuss your writing needs and we will have them evaluated by our dedicated team.

  • Clear elicitation of your requirements.
  • Customized writing as per your needs.

We Mirror Your Guidelines to Deliver Quality Services

We write your papers in a standardized way. We complete your work in such a way that it turns out to be a perfect description of your guidelines.

  • Proactive analysis of your writing.
  • Active communication to understand requirements.
image
image

We Handle Your Writing Tasks to Ensure Excellent Grades

We promise you excellent grades and academic excellence that you always longed for. Our writers stay in touch with you via email.

  • Thorough research and analysis for every order.
  • Deliverance of reliable writing service to improve your grades.
Place an Order Start Chat Now
image

Order your essay today and save 30% with the discount code Happy