I really like this product. I think it's a great value just as it is, but there are a few things that would make it a lot more powerful, both on the client side and remote side.
I see a lot of questions about how to skip or include files and/or folders during the indexing process. For some this is simple as their file organization and naming conventions are well structured, which lends itself to the simple implementation of the extension inclusion or file/folder denial word lists. But some of us inherit some less than organized legacy sites that make filling out these lists pretty cumbersome. A more comprehensive approach would be to allow the spider to first prepare a site manifest as a pre-scan, which would be presented as a collapsible/expandable tree with each folder and file accompanied by a check box. On the highest level, the entire list could be set to 'include' or 'deny' whatever checkboxes are selected. The checkboxes could then be used in a highly selective manner to easily qualify the files to be indexed. New files and folders added after the initial scan could be highlighted. The nice thing about this approach is that it would really help those using the non-pro versions to allocate their file count allotment before the indexing process begins. Initially, in lieu of a visual approach, the system could simply read a user prepared concordance file.
On the remote side, a very powerful plug-in would be to track sessions so that a user's searches could be profiled and saved. The results of the searches would be stored in a backend db along with options to automatically perform profile searches on a recurring basis with the user receiving an e-mail notification when new references to search criteria are added to the site. I know this sounds daunting, but it's all pretty standard stuff. This would be invaluable for dynamic php sites, especially forum based software. The profile searches could be scheduled to occur at off-peak times.
I see a lot of questions about how to skip or include files and/or folders during the indexing process. For some this is simple as their file organization and naming conventions are well structured, which lends itself to the simple implementation of the extension inclusion or file/folder denial word lists. But some of us inherit some less than organized legacy sites that make filling out these lists pretty cumbersome. A more comprehensive approach would be to allow the spider to first prepare a site manifest as a pre-scan, which would be presented as a collapsible/expandable tree with each folder and file accompanied by a check box. On the highest level, the entire list could be set to 'include' or 'deny' whatever checkboxes are selected. The checkboxes could then be used in a highly selective manner to easily qualify the files to be indexed. New files and folders added after the initial scan could be highlighted. The nice thing about this approach is that it would really help those using the non-pro versions to allocate their file count allotment before the indexing process begins. Initially, in lieu of a visual approach, the system could simply read a user prepared concordance file.
On the remote side, a very powerful plug-in would be to track sessions so that a user's searches could be profiled and saved. The results of the searches would be stored in a backend db along with options to automatically perform profile searches on a recurring basis with the user receiving an e-mail notification when new references to search criteria are added to the site. I know this sounds daunting, but it's all pretty standard stuff. This would be invaluable for dynamic php sites, especially forum based software. The profile searches could be scheduled to occur at off-peak times.
Comment