regain 1.6.2 Ranking & Summary
regain 1.6.2 description
regain 1.6.2 brings you an effective and easy to use search engine similar to web search engines like Google, with the difference that you don't search the web, but your own files and documents. Using regain you can search through large portions of data (several gigabytes!) in split seconds! This is possible by using a search index. regain crawles through your files or webpages, extracts all text and puts it in a smart search index. All this happens in the background. So if you want to search something you get the results immediately.
There are two versions of regain: The desktop search and the server search. The desktop search is to be used on a normal desktop computer and it offers you a fast search for documents or intranet webpages. The server search you can install on web servers. It provides searching functionality for a website or for intranet fileservers.
regain is written in Java and thus applicable on all Java compatible platforms (amongst others Windows, Linux, Mac OS, Solaris). The server search works with Java Server Pages (JSPs) and a tag library, the desktop search comes with its own small webserver. regain is released under the open source license LGPL (Lesser General Public License). I.e. regain may be used for free without any temporal limit.
- regain uses the powerful search syntax of Lucene. Thus it is possible to express very specific search queries. The most important possibilities are the following:
- Boolean operators
- Phonetic search
- and much more. You can find more information about the search syntax here.
- Multi index search: Search multiple indexes with one search mask. Totally transparent for the user.
- URL-Rewriting: You can use URL-Rewriting at your search. This enables you to index documents from file://c:/www-data/intranet/docs and show them in the browser as http://intranet.murfman.de/docs.
- Advanced search: All values that are in the index for one field may now be provided as a drop down list on the search page. Particularly together with auxiliary fields this is very useful.
- File-to-http-bridge: Some browsers load for security reasons no file links from http pages. Thus all documents that are in the index are now provided over HTTP. Of corse this may switched off and at the desktop search these documents are only accessible from the local host.
- Defining the search space: Using regain you may specify very exactly what should be indexed and what should not.
- White and black list: With a white list and a black list you may isolate very exactly which documents the crawler should process. E.g. you may index all from http://www.murfman.de except for http://www.murfman.de/dynamiccontent.
- Several sources in one index: You may index documents from different file systems and/or web sites in the same search index.
- Partial indexing: Assumed your search index contains documents from a network drive (file server) and a web page. You may update only the documents from the network drive. In doing so you may update some drives every hour and others only every week.
- Hot deployment: Change on a new search index without restarting your servlet engine (e.g. Tomcat).
- Stopword list: Define words should not be indexed.
- Analysis files: If desired all intermediate steps of the indexing process may be written out as files. In doing so you see exactly what gets in the search index.
- Content extraction for HTML: Index only the actual content of your web pages. regain removes the navigation and footer for your.
- Path extraction for HTML: Show the navigation path of your web pages in the search results.
- Dead link detection: As a sort of by-product all found dead links (links to non-existing documents) are written out.
- Breakpoints: The crawler creates periodically so called breakpoint. When doing so, the current state of the search index is copied into a separate directory. If the index update should be cancelled (e.g. if the computer is shut down), the crawler will go on from the last breakpoint the next time it is started.
- Auxiliary fields: The index may be extended by auxiliary fields that are extracted from a document's URL. Example: Assumed you have a directory with a sub directory for every project. Then you can generate an auxiliary field with the project name. Doing so, you get only documents from directory of the project "otto23" when searching for "Offer project:otto23".
- Expandability and customization
- Preparators: The preparation of a certain file format is done by so-called preparators. Thus you are able to specify which preparators regain should use. In addition regain may easily be extended for more file formats.
- Tag Library for the search: regain offers a Tag Library for creating the Java Server Page for the search. Thus the adaption of the search page to your web page's design is particularly easy.
- Configuration: regain is highly adaptable. The whole configuration of the crawler is in one XML file.
- Access rights management: It is now possible to integrate an access rights management, that ensures that a user only sees results for documents he has reading rights for.
regain 1.6.2 Screenshot
regain 1.6.2 Keywords
Bookmark regain 1.6.2
regain 1.6.2 Copyright
Want to place your software product here?
Please contact us for consideration.