I’m betting even Access will choke a bit on the 10 million records-ish we’re talking about. Yes, it’ll do it, but it might be painfully slow.
Here’s a MUCH easier way if the log file is a text file and you’re just trying to extract a small fraction of the lines in it according to simple per-line criteria.
Use the CMD FIND program.
The critical failing of all the typical Windows-based editors is that they try to read the whole thing, when you really just need to look at each row independently.
Open a CMD window and then enter
C:>FIND /i “msn.com” MyHuge630MBFile.log >SmallExtract.log
It’ll read through the file, one record at a time, copying only those records containing “msn.com” to the SmallExtract.log file. The one record at a time feature means the file size is limited only by disk space and the /i switch means case-insensitive comparison. It may take a few minutes to run through the 10 million-ish records you have, but it’ll do it.
Then you can use some smarter log analysis tool (or Access or a spread sheet) to look at the details of the surviving records.
If there are several different selection criteria, you can dasiy-chain finds to achieve AND logic. You can conduct parellel finds and merge the results to achieve OR logic.
All this operates only on the per-record level of selection, and it requires that the log file contain just plain ASCII text, but that’s probably all you need.
FIND was added to MS-DOS in version 2.0 and is still as useful as ever. We all tend forget sometimes about the basic tools. For basic jobs they’re often the best bet.