[R] speed issue: gsub on large data frame

Jeff Newmiller jdnewmil at dcn.davis.CA.us
Tue Nov 5 06:59:31 CET 2013


Example not reproducible. Communication fail. Please refer to Posting Guide.
---------------------------------------------------------------------------
Jeff Newmiller                        The     .....       .....  Go Live...
DCN:<jdnewmil at dcn.davis.ca.us>        Basics: ##.#.       ##.#.  Live Go...
                                      Live:   OO#.. Dead: OO#..  Playing
Research Engineer (Solar/Batteries            O.O#.       #.O#.  with
/Software/Embedded Controllers)               .OO#.       .OO#.  rocks...1k
--------------------------------------------------------------------------- 
Sent from my phone. Please excuse my brevity.

Simon Pickert <simon.pickert at t-online.de> wrote:
>Hi R’lers,
>
>I’m running into speeding issues, performing a bunch of 
>
>„gsub(patternvector, [token],dataframe$text_column)"
>
>on a data frame containing >4millionentries.
>
>(The “patternvectors“ contain up to 500 elements) 
>
>Is there any better/faster way than performing like 20 gsub commands in
>a row?
>
>
>Thanks!
>Simon
>
>______________________________________________
>R-help at r-project.org mailing list
>https://stat.ethz.ch/mailman/listinfo/r-help
>PLEASE do read the posting guide
>http://www.R-project.org/posting-guide.html
>and provide commented, minimal, self-contained, reproducible code.



More information about the R-help mailing list