Is there an inherent bug in the design of .txt that signficant performance drops can be experienced when working with relatively large files? Working with a file of over a million characters on Windows 8 and not only do editors (specifically Notepad++ and Wordpad) tend to run slowly, but frequently crash with large find and replace operations (if the find and replace conducts over 60,000 replacements for instance).
I have verified that this has nothing to do with the particular machine on which the textfile is operating (same poor performance on a similar, high spec machine).
I initially thought this was because the file in question was being edited on a network share, but copying it to a local folder still resulted in the same low performance.
The performance particularly drops when newlines are removed (working with a single word a million characters in length). Noticeable frame drop in UI despite using high end machine)
Best Answer
Try Ultra Edit Its the best in this case.
Features at a glance
Column / block editing
Multi-caret editing
Multi-select
Syntax highlighting
Integrated FTP client
Integrated SSH/telnet
Editor themes
Large file editing +4GB
File / data sorting
Powerful search
Supports regex
CSV data reformatting
Macros and scripts
File compare
Smart templates