R text mining documents from CSV file (one row per doc)

20,598

Here's a complete workflow to get what you want:

# change this file location to suit your machine
file_loc <- "C:\\Documents and Settings\\Administrator\\Desktop\\Book1.csv"
# change TRUE to FALSE if you have no column headings in the CSV
x <- read.csv(file_loc, header = TRUE)
require(tm)
corp <- Corpus(DataframeSource(x))
dtm <- DocumentTermMatrix(corp)

In the dtm object each row will be a doc, or a line of your original CSV file. Each column will be a word.

Share:
20,598
user2407054
Author by

user2407054

Updated on July 15, 2022

Comments

  • user2407054
    user2407054 almost 2 years

    I am trying to work with the tm package in R, and have a CSV file of customer feedback with each line being a different instance of feedback. I want to import all the content of this feedback into a corpus but I want each line to be a different document within the corpus, so that I can compare the feedback in a DocTerms Matrix. There are over 10,000 rows in my data set.

    Originally I did the following:

    fdbk_corpus <-Corpus(VectorSource(fdbk), readerControl = list(language="eng"), sep="\t")
    

    This creates a corpus with 1 document and >10,000 rows, and I want >10,000 docs with 1 row each.

    I imagine I could just have 10,000+ separate CSV or TXT documents within a folder and create a corpus from that... but I'm thinking there is a much simpler answer than that, reading each line as a separate document.