diff --git a/src/site/en/xdoc/2.0/admin/browserType-guide.xml b/src/site/en/xdoc/2.0/admin/browserType-guide.xml new file mode 100644 index 000000000..251bb4b32 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/browserType-guide.xml @@ -0,0 +1,19 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
+

Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

+ +

In Administrator account after logging in, click menu browser types.

+ Setting the browser type +
+ +

You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/crawl-guide.xml b/src/site/en/xdoc/2.0/admin/crawl-guide.xml new file mode 100644 index 000000000..c98ce95b4 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/crawl-guide.xml @@ -0,0 +1,100 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
+

Describes the settings related to crawling.

+ +

In Administrator account click crawl General menu after login.

+ Crawl General +

You can specify the path to a generated index and replication capabilities to enable.

+ Replication features +
+ +

You can set the interval at which the crawl for a Web site or file system. By default, the following.

+ +

Figures are from left, seconds, minutes, during the day, month, represents a day of the week. Description format is similar to the Unix cron settings. This example, and am 0 時 0 分 to crawling daily.

+

Following are examples of how to write.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
0 0 12 * *?Each day starts at 12 pm
0 15 10? * *Day 10: 15 am start
0 15 10 * *?Day 10: 15 am start
0 15 10 * *? *Day 10: 15 am start
0 15 10 * *? 2005Each of the 2009 start am, 10:15
0 * 14 * *?Every day 2:00 in the PM-2: 59 pm start every 1 minute
0 0 / 5 14 * *?Every day 2:00 in the PM-2: 59 pm start every 5 minutes
0 0 / 5 14, 18 * *?Every day 2:00 pm-2: 59 pm and 6: 00 starts every 5 minutes at the PM-6: 59 pm
0 0-5 14 * *?Every day 2:00 in the PM-2: 05 pm start every 1 minute
0 10, 44 14? 3 WEDStarts Wednesday March 2: 10 and 2: 44 pm
0 15 10? * MON-FRIMonday through Friday at 10:15 am start
+

Also check if the seconds can be set to run at intervals 60 seconds by default. If you set seconds exactly and you should customize webapps/fess/WEB-INF/classes/chronosCustomize.dicon taskScanIntervalTime value, if enough do I see in one-hour increments.

+
+ +

If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

+
+ +

To enable replication features that can apply already copied the Solr index generated. For example, you can use them if you want to search only in the search servers crawled and indexed on a different server, placed in front.

+
+ +

After the data is registered for Solr. Index to commit or to optimize the registered data becomes available. If optimize is issued the Solr index optimization, if you have chosen, you choose to commit the commit is issued.

+
+ +

Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

+
+ +

To raise the performance of the index in Fess while crawling and sends for Solr document in 20 units. For each value specified here because without committing to continue adding documents documents added in the Solr on performance, Solr issued document commits. By default, after you add documents 1000 is committed.

+
+ +

Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

+

But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

+
+ +

You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

+
+ +

Copy index information from the index directory as the snapshot path, if replication is enabled, will be applied.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/2.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..43afb4364 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/crawlingSession-guide.xml @@ -0,0 +1,34 @@ + + + + Set session information + Shinsuke Sugaya + + +
+

Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

+ +

In Administrator account after logging in, click the session information menu.

+
+ + Session information list +

You can remove all session information and click the Delete link all in the running.

+
+ + Session details +

To specify a session ID, you can see crawling content.

+
    +
  • Information about the entire crawl Cralwer *:
  • +
  • FsCrawl *: information about the file system crawling
  • +
  • WebCrawl *: crawling the Web information
  • +
  • Information issued by Solr server optimization optimize *:
  • +
  • Commit *: information about the commit was issued to the Solr server.
  • +
  • * StartTime: start time
  • +
  • * EndTime: end time
  • +
  • * ExecTime: execution time (MS)
  • +
  • * IndexSize: number of documents indexed
  • +
+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/data-guide.xml b/src/site/en/xdoc/2.0/admin/data-guide.xml new file mode 100644 index 000000000..9ea6df5eb --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/data-guide.xml @@ -0,0 +1,33 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
+

Here, describes Fess information backup and restore methods.

+ +

In Administrator account after logging in, click the menu backup and restore.

+ Backup and restore +
+ +

Click the download link and Fess information output in XML format. Saved settings information is below.

+
    +
  • The General crawl settings
  • +
  • Web crawl settings
  • +
  • File system Crawl settings
  • +
  • Path mapping
  • +
  • Web authentication
  • +
  • Compatible browsers
  • +
  • Session information
  • +
+

In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate.

+
+ +

You can restore the configuration information by uploading the XML outputted by the backup. Specify the XML file, click the restore button on the data.

+

If there is already to enable overwriting of data, the same data does update existing data.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/design-guide.xml b/src/site/en/xdoc/2.0/admin/design-guide.xml new file mode 100644 index 000000000..b4b33f370 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/design-guide.xml @@ -0,0 +1,69 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
+

Here are settings for the design of search screens.

+ +

In Administrator account after logging in, click the menu design.

+ Design +

You can edit the search screen in the screen below.

+ JSP compilation screen +
+ +

You can upload the image files to use in the search screen. Image file names are supported are jpg, gif and png.

+
+ +

If you want the file name to upload image files to use. Uploaded if you omit the file name will be used.

+
+ +

You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

+

Following are examples of how to write.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
Top page (within the Head tags)This is the express search home page head tag in JSP files. If you want to edit the meta tags, title tags, script tags, such as the change.
Top page (content)Is a JSP file to represent the body tag in the search home page.
Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
Search results page (within the Head tags)Search result is a JSP file to represent within the head tag of the list page. If you want to edit the meta tags, title tags, script tags, such as the change.
Search results page (header)Search result is a JSP file to represent the header of the list page. Include search form at the top.
Search results page (footer)Search result is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
+

You can to edit for PCs and similar portable screen.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/2.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..1f7628385 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,96 @@ + + + + Settings for crawling a file system using + Shinsuke Sugaya + + +
+

Describes the settings for crawl here, using file system.

+

Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

+ +

In Administrator account after logging in, click menu file.

+ Setting file system Crawl +
+ +

Is the name that appears on the list page.

+
+ +

You can specify multiple paths. file: in the specify starting. For example,

+ +

The so determines. Patrolling below the specified directory.

+

So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

+
+ +

By specifying regular expressions you can exclude the crawl and search for given path pattern.

+ + + + + + + + + + + + + + + + + + + +
Path to crawlCrawl the path for the specified regular expression.
The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
+

For example, the path to target if you don't crawl less than/home /

+ +

Also the path to exclude if extension of png want to exclude from

+ +

It specifies. It is possible to specify multiple line breaks in.

+

How to specify the URI handling java.io.File: Looks like:

+ file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
+ +

Specify the depth of a directory hierarchy.

+
+ +

You can specify the number of documents to retrieve crawl.

+
+ +

Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

+
+ +

Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

+

Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

+
+ +

You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

+

Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

+
+ +

Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

+
+ +

You can control only when a particular user role can appear in search results. You must roll a set before you. > For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

+
+ +

You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

+
+ +

Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/index.xml b/src/site/en/xdoc/2.0/admin/index.xml new file mode 100644 index 000000000..8ac542aab --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/index.xml @@ -0,0 +1,12 @@ + + + + Management UI Guide + Shinsuke Sugaya + + +
+

Here is a description of Fess 2.0 administrative UI.

+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/labelType-guide.xml b/src/site/en/xdoc/2.0/admin/labelType-guide.xml new file mode 100644 index 000000000..3345c7a1f --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/labelType-guide.xml @@ -0,0 +1,23 @@ + + + + Setting a label + Shinsuke Sugaya + + +
+

Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. If you register the label shown select label drop-down box to the right of the search box.

+ +

In Administrator account after logging in, click the menu label.

+ List of labels + Setting a label +
+ +

Specifies the name that is displayed when the search label drop-down select.

+
+ +

Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/log-guide.xml b/src/site/en/xdoc/2.0/admin/log-guide.xml new file mode 100644 index 000000000..f8c69efa7 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/log-guide.xml @@ -0,0 +1,19 @@ + + + + Log file download + Shinsuke Sugaya + + +
+

Describes the log files will be output in the Fess download.

+ +

In Administrator account after logging in, click log file menu.

+ Session information +
+ +

You can download the log file and click the log file name.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/2.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..4fbc9b0df --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/overlappingHost-guide.xml @@ -0,0 +1,23 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
+

Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

+ +

In Administrator account after logging in, click the menu duplicate host.

+ A list of the duplicate host + Duplicate host settings +
+ +

Specify the canonical host name. Duplicate host names replace the canonical host name.

+
+ +

Specify the host names are duplicated. Specifies the host name you want to replace.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/2.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..47641001e --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/pathMapping-guide.xml @@ -0,0 +1,26 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
+

Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

+ +

In Administrator account after logging in, click menu path mappings.

+ List of path mapping + Path mapping settings +
+ +

Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

+
+ +

Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

+
+ +

Specifies the string to replace the matched regular expression.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/2.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..32cc62f0d --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/requestHeader-guide.xml @@ -0,0 +1,26 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
+

Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

+ +

In Administrator account after logging in, click request header menu.

+ A list of request headers + Setting a request header +
+ +

Specifies the request header name to append to the request.

+
+ +

Specifies the request header value to append to the request.

+
+ +

Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/roleType-guide.xml b/src/site/en/xdoc/2.0/admin/roleType-guide.xml new file mode 100644 index 000000000..b6e3776fb --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/roleType-guide.xml @@ -0,0 +1,23 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
+

Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

+ +

In Administrator account after logging in, click menu role.

+ The list of roles + Settings for a role +
+ +

Specifies the name that appears in the list.

+
+ +

Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/system-guide.xml b/src/site/en/xdoc/2.0/admin/system-guide.xml new file mode 100644 index 000000000..1c972d2ab --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/system-guide.xml @@ -0,0 +1,28 @@ + + + + System settings + Shinsuke Sugaya + + +
+

Describes the settings related to Solr, here registration in Fess. SOLR servers are grouped by file, has been registered.

+ +

In Administrator account after logging in, click menu Solr.

+ System settings +
+ +

Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and shut down when not running Fess server to shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

+
+ +

Server group name is used to search for and update appears.

+
+ +

Server becomes unavailable and the status of disabled. For example, inaccessible to the Solr server and changes to disabled. To enable recovery after server become unavailable will become available.

+
+ +

You can publish index commit, optimize for server groups. You can also remove a specific search for the session ID.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/2.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..017c9cc34 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/webAuthentication-guide.xml @@ -0,0 +1,37 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
+

Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

+ +

In Administrator account after logging in, click menu Web authentication.

+ Configuring Web authentication +
+ +

Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

+
+ +

Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

+
+ +

Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

+
+ +

Select the authentication method. You can use BASIC authentication or DIGEST authentication.

+
+ +

Specifies the user name to log in authentication.

+
+ +

Specifies the password to log into the certification site.

+
+ +

Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/2.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..316e0cda4 --- /dev/null +++ b/src/site/en/xdoc/2.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,99 @@ + + + + Settings for crawling the Web using + Shinsuke Sugaya + + +
+

Describes the settings here, using Web crawling.

+

Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

+ +

In Administrator account after logging in, click menu Web.

+ Web crawl settings +
+ +

Is the name that appears on the list page.

+
+ +

You can specify multiple URLs. http: or https: in the specify starting. For example,

+ +

The so determines.

+
+ +

By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

+ + + + + + + + + + + + + + + + + + + +
URL to crawlCrawl the URL for the specified regular expression.
Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
+

For example, http: URL to crawl if not crawl //localhost/ less than the

+ +

Also be excluded if the extension of png want to exclude from the URL

+ +

It specifies. It is possible to specify multiple in the line for.

+
+ +

That will follow the links contained in the document in the crawl order can specify the tracing depth.

+
+ +

You can specify the number of documents to retrieve crawl.

+
+ +

You can specify the user agent to use when crawling.

+
+ +

Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

+
+ +

Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

+

Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

+
+ +

You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

+

Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

+
+ +

Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

+
+ +

You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

+
+ +

You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

+
+ +

Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

+
+
+
+ +

Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in)

+

Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

+

Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/config/index.xml b/src/site/en/xdoc/2.0/config/index.xml new file mode 100644 index 000000000..a69b85c73 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/index.xml @@ -0,0 +1,12 @@ + + + + Set up Guide + Shinsuke Sugaya + + +
+

Here is the Fess 2.0 Setup instructions.

+
+ +
diff --git a/src/site/en/xdoc/2.0/config/logging.xml b/src/site/en/xdoc/2.0/config/logging.xml new file mode 100644 index 000000000..f1c3a6dba --- /dev/null +++ b/src/site/en/xdoc/2.0/config/logging.xml @@ -0,0 +1,18 @@ + + + + Log settings + Shinsuke Sugaya + + +
+

Fess output log (Solr log output to the logs/catalina.out) will be output to webapps/fess/WEB-INF/logs/fess.out. sets the contents to output Fess.out, Webpps/Fess/Web-INF/clsses/log4j.Xml. By default output INFO level.

+

For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

+ + + +]]> +
+ +
diff --git a/src/site/en/xdoc/2.0/config/memory-config.xml b/src/site/en/xdoc/2.0/config/memory-config.xml new file mode 100644 index 000000000..9993ec062 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/memory-config.xml @@ -0,0 +1,23 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
+

If the contents of the crawl settings cause OutOfMemory error similar to the following.

+ +

Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case the maximum value set 1024M) will change to-Xmx1024m.

+ +
+ +
diff --git a/src/site/en/xdoc/2.0/config/mobile-device.xml b/src/site/en/xdoc/2.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
+

The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

+ ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
+ +
diff --git a/src/site/en/xdoc/2.0/config/protwords.xml b/src/site/en/xdoc/2.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
+

In Fess when indexing and searching the stemming process done.

+

This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

+
+
+

You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

+

In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

+
+ +
diff --git a/src/site/en/xdoc/2.0/config/proxy.xml b/src/site/en/xdoc/2.0/config/proxy.xml new file mode 100644 index 000000000..db48fd846 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/proxy.xml @@ -0,0 +1,57 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
+

If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

+
+ +

Proxy is set in to create webapps/Fess/Web-INF/classes/s9robot_client.dicon with the following contents.

+ + + + + + + + @org.apache.commons.httpclient.cookie.CookiePolicy@BROWSER_COMPATIBILITY + "プロキシホスト名" + プロキシポート + + + + + "UTF-8" + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + + + +]]> +
+ + +
diff --git a/src/site/en/xdoc/2.0/config/replication.xml b/src/site/en/xdoc/2.0/config/replication.xml new file mode 100644 index 000000000..c36e639c4 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/replication.xml @@ -0,0 +1,25 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
+

Fess can copy the path in Solr index data. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

+

You must use the replication features of Fess for Solr index file in the shared disk, such as NFS, Fess of each can be referenced from.

+
+
+ +

Fess, download and install the./ /NET/Server1/usr/local/Fess To assume you installed.

+

To register the crawl settings as well as Fess starts after the normal construction, create the index (index for Fess building instructions normal building procedures and especially remains the same) crawling.

+
+ +

Fess, download and install the./ /NET/Server2/usr/local/Fess To assume you installed.

+

To enable replication features check box in Fess starts after the management screen crawl settings the "snapshot path'. Snapshot path designates the index location for the index for Fess. In this case, the/NET/Server1/usr/local/Fess //solr/core1/data/index In the will.

+ Replication +

Time press the update button to save the data and set in Schedule performs replication of the index.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/config/role-setting.xml b/src/site/en/xdoc/2.0/config/role-setting.xml new file mode 100644 index 000000000..26c927ca8 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/role-setting.xml @@ -0,0 +1,90 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
+

You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

+

In role-based search of the Fess roll information available below.

+ +

To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

+
+
+

Describes how to set up role-based search using J2EE authentication information.

+ +

conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

+ + + + + + + + + +]]> +
+ +

sets the webapps/fess/WEB-INF/classes/app.dicon shown below.

+ + "role1" + + : + + + {"guest"} + + + : +]]> +

authenticatedRoles can describe multiple by commas (,). You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results not logged in to set user roles are required.

+
+ +

sets the webapps/fess/WEB-INF/web.xml shown below.

+ + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
+ +

Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

+
+ +

Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

+

Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

+

Also, search not logged in will be search by guest user.

+
+ +

Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/config/server-port.xml b/src/site/en/xdoc/2.0/config/server-port.xml new file mode 100644 index 000000000..c55bb3010 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/server-port.xml @@ -0,0 +1,30 @@ + + + + Ports changes + Shinsuke Sugaya + + +
+

Fess by default, you use the port 8080. Change in the following steps to change.

+ +

Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

+
    +
  • 8080: HTTP access port
  • +
  • 8005: shut down port
  • +
  • 8009: AJP port
  • +
  • : SSL HTTP access port 8443 (the default is off)
  • +
+
+ +

May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information. change the webapps/fess/WEB-INF/classes/fess_solr.dicon.

+ "http://localhost:8080/solr" +]]> +

+ Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/config/solr-failure.xml b/src/site/en/xdoc/2.0/config/solr-failure.xml new file mode 100644 index 000000000..4b487e43b --- /dev/null +++ b/src/site/en/xdoc/2.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
+

Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

+

SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/fess_solr.dicon.

+ +
    +
  • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
  • +
  • Solr server number of valid state is minActiveServer following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status maxRetryStatusCheckCount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
  • +
  • Disable Solr group is not available.
  • +
  • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
  • +
+
+ +
    +
  • Search queries can send valid Solr group.
  • +
  • Search queries will be sent only to valid Solr server.
  • +
  • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
  • +
  • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
  • +
+
+ +
    +
  • Update queries you can send valid state Solr group.
  • +
  • Update query will be sent only to valid Solr server.
  • +
  • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
  • +
  • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
  • +
+
+
+ +
diff --git a/src/site/en/xdoc/2.0/config/tokenizer.xml b/src/site/en/xdoc/2.0/config/tokenizer.xml new file mode 100644 index 000000000..4181e30ff --- /dev/null +++ b/src/site/en/xdoc/2.0/config/tokenizer.xml @@ -0,0 +1,36 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
+

You must isolate the document in order to register as the index when creating indexes for the search.

+

Tokenizer is used for this.

+

Basically, carved by the tokenizer units smaller than go find no hits.

+

For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit.

+

For selection of the tokenizer is important.

+

You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default CJKTokenizer used.

+
+ + +

Such as CJKTokenizer Japan Japanese multibyte string against bi-gram, in other words two characters create index. In this case, can't find one letter words.

+
+ + +

StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to.

+

The following example to change schema.xml so analyzer parts, you can use the StandardTokenizer.

+ + + + + : +]]> +
+ + +
diff --git a/src/site/en/xdoc/2.0/config/windows-service.xml b/src/site/en/xdoc/2.0/config/windows-service.xml new file mode 100644 index 000000000..c9aec8e24 --- /dev/null +++ b/src/site/en/xdoc/2.0/config/windows-service.xml @@ -0,0 +1,45 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
+

You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

+ +

First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-2.0.0.

+ cd C:\Java\fess-server-2.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +

Then add properties for Fess. To run the following, Tomcat Properties window appears.

+ tomcat6w.exe //ES//fess +]]> +

Set the following in the Java Options in the Java tab.

+ +

Modifies the value of the maximum memory pool to 512. Settings to save the settings and then press OK button. Please start later as normal Windows services and.

+
+
+ +
diff --git a/src/site/en/xdoc/2.0/user/index.xml b/src/site/en/xdoc/2.0/user/index.xml new file mode 100644 index 000000000..e7f49b10c --- /dev/null +++ b/src/site/en/xdoc/2.0/user/index.xml @@ -0,0 +1,12 @@ + + + + Search Guide + Shinsuke Sugaya + + +
+

Here is the instructions on how to search for Fess 2.0.

+
+ +
diff --git a/src/site/en/xdoc/2.0/user/search-field.xml b/src/site/en/xdoc/2.0/user/search-field.xml new file mode 100644 index 000000000..b30448189 --- /dev/null +++ b/src/site/en/xdoc/2.0/user/search-field.xml @@ -0,0 +1,57 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
+

In the Fess crawl results saved in the title and text fields. You can search for a field of them.

+

You can search for a the following fields by default.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
URLThe crawl URL
hostWere included in the crawl URL host name
siteSite name was included in the crawl URL
titleTitle
contentText
contentLengthYou crawl the content size
lastModifiedLast update of the content you want to crawl
mimetypeThe MIME type of the content
+

If you do not specify the fields title and content to search for.

+ +

If a field search "field name: search terms ' of so fill out the search form, the search.

+

Title against Fess the search as a search term.

+ +
+
+ +
diff --git a/src/site/en/xdoc/2.0/user/search-label.xml b/src/site/en/xdoc/2.0/user/search-label.xml new file mode 100644 index 000000000..5340d335b --- /dev/null +++ b/src/site/en/xdoc/2.0/user/search-label.xml @@ -0,0 +1,14 @@ + + + + Search by label + Shinsuke Sugaya + + +
+

By label to be registered in the management screen will enable search by labels in the search screen. You can use the label if you want to sort the search results. If you do not register the label displayed the label drop-down box.

+ Search by label +

To set the label by creating indexes, can search each crawl settings specified on the label. All results search search do not specify a label is usually the same.

+
+ +
diff --git a/src/site/en/xdoc/2.0/user/search-sort.xml b/src/site/en/xdoc/2.0/user/search-sort.xml new file mode 100644 index 000000000..bceb8e612 --- /dev/null +++ b/src/site/en/xdoc/2.0/user/search-sort.xml @@ -0,0 +1,44 @@ + + + + Search sort + Shinsuke Sugaya + + +
+

To sort the search results by specifying the fields such as search time.

+

You can sort the following fields by default.

+ + + + + + + + + + + + + + + +
TstampOn the crawl
contentLengthYou crawl the content size
lastModifiedLast update of the content you want to crawl
+ +

If you want to sort ' sort: field name ' in to fill out the search form, the search.

+

In ascending order sort the content size as a search term, Fess is below.

+ +

To sort in descending order as below.

+ +

If you sort by multiple fields separated list, shown below.

+ +
+
+ +
diff --git a/src/site/en/xdoc/3.0/admin/browserType-guide.xml b/src/site/en/xdoc/3.0/admin/browserType-guide.xml new file mode 100644 index 000000000..3a1b65470 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/browserType-guide.xml @@ -0,0 +1,19 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
+

Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

+ +

In Administrator account after logging in, click menu browser types.

+ Setting the browser type +
+ +

You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

+
+
+ +
diff --git a/src/site/en/xdoc/3.0/admin/config-wizard.xml b/src/site/en/xdoc/3.0/admin/config-wizard.xml new file mode 100644 index 000000000..ce0704cb5 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/config-wizard.xml @@ -0,0 +1,34 @@ + + + + Settings Wizard + Sone, Takaaki + + +
+

Introduction to the Configuration Wizard.

+

You can use Settings Wizard, to set you up on the fess.

+ +

In Administrator account after logging in, click menu Settings Wizard.

+ Settings Wizard +

First, setting a schedule.

+

During the time in fess is crawling and indexes.

+

By default, every day is a 0 時 0 分.

+ Setting a schedule +

The crawl settings.

+

Crawl settings is to register a URI to look for.

+

The crawl settings name please put name of any easy to identify.

+

Put the URI part de-indexed, want to search for.

+ Crawl settings +

For example, if you want search for http://example.com, below looks like.

+ Crawl settings example +

In this is the last setting.

+

Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

+ Crawl started +
+ +

Settings in the Setup Wizard you can change from crawl General, Web, file system.

+
+
+ +
diff --git a/src/site/en/xdoc/3.0/admin/crawl-guide.xml b/src/site/en/xdoc/3.0/admin/crawl-guide.xml new file mode 100644 index 000000000..29e3108ea --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/crawl-guide.xml @@ -0,0 +1,100 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
+

Describes the settings related to crawling.

+ +

In Administrator account click crawl General menu after login.

+ Crawl General +

You can specify the path to a generated index and replication capabilities to enable.

+ Replication features +
+ +

You can set the interval at which the crawl for a Web site or file system. By default, the following.

+ +

Figures are from left, seconds, minutes, during the day, month, represents a day of the week. Description format is similar to the Unix cron settings. This example, and am 0 時 0 分 to crawling daily.

+

Following are examples of how to write.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
0 0 12 * *?Each day starts at 12 pm
0 15 10? * *Day 10: 15 am start
0 15 10 * *?Day 10: 15 am start
0 15 10 * *? *Day 10: 15 am start
0 15 10 * *? 3.053.09 years every day 10:15 am start
0 * 14 * *?Daily 2:59 pm-3.00 pm start per 1 minute
0 0 / 5 14 * *?Every day 2:59 pm-3.00 pm start every 5 minutes
0 0 / 5 14, 18 * *?Daily 3.00 pm-2: 59 pm and 6: 00 starts every 5 minutes at the PM-6: 59 pm
0 0-5 14 * *?Every day at 3.00pm-3.05pm start per 1 minute
0 10, 44 14? 3 WEDStarts Wednesday March 2: 10 and 2: 44 pm
0 15 10? * MON-FRIMonday through Friday at 10:15 am start
+

Also check if the seconds can be set to run at intervals 60 seconds by default. If you set seconds exactly and you should customize webapps/fess/WEB-INF/classes/chronosCustomize.dicon taskScanIntervalTime value, if enough do I see in one-hour increments.

+
+ +

If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

+
+ +

To enable replication features that can apply already copied the Solr index generated. For example, you can use them if you want to search only in the search servers crawled and indexed on a different server, placed in front.

+
+ +

After the data is registered for Solr. Index to commit or to optimize the registered data becomes available. If optimize is issued the Solr index optimization, if you have chosen, you choose to commit the commit is issued.

+
+ +

Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

+
+ +

To raise the performance of the index in Fess while crawling and sends for Solr document in 20 units. For each value specified here because without committing to continue adding documents documents added in the Solr on performance, Solr issued document commits. By default, after you add documents 1000 is committed.

+
+ +

Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

+

But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

+
+ +

You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

+
+ +

Copy index information from the index directory as the snapshot path, if replication is enabled, will be applied.

+
+
+ +
diff --git a/src/site/en/xdoc/3.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/3.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..efde2ce79 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/crawlingSession-guide.xml @@ -0,0 +1,34 @@ + + + + Set session information + Shinsuke Sugaya + + +
+

Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

+ +

In Administrator account after logging in, click the session information menu.

+
+ + Session information list +

You can remove all session information and click the Delete link all in the running.

+
+ + Session details +

To specify a session ID, you can see crawling content.

+
    +
  • Information about the entire crawl Cralwer *:
  • +
  • FsCrawl *: information about the file system crawling
  • +
  • WebCrawl *: crawling the Web information
  • +
  • Information issued by Solr server optimization optimize *:
  • +
  • Commit *: information about the commit was issued to the Solr server.
  • +
  • * StartTime: start time
  • +
  • * EndTime: end time
  • +
  • * ExecTime: execution time (MS)
  • +
  • * IndexSize: number of documents indexed
  • +
+
+
+ +
diff --git a/src/site/en/xdoc/3.0/admin/data-guide.xml b/src/site/en/xdoc/3.0/admin/data-guide.xml new file mode 100644 index 000000000..006b784e9 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/data-guide.xml @@ -0,0 +1,33 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
+

Here, describes Fess information backup and restore methods.

+ +

In Administrator account after logging in, click the menu backup and restore.

+ Backup and restore +
+ +

Click the download link and Fess information output in XML format. Saved settings information is below.

+
    +
  • The General crawl settings
  • +
  • Web crawl settings
  • +
  • File system Crawl settings
  • +
  • Path mapping
  • +
  • Web authentication
  • +
  • Compatible browsers
  • +
  • Session information
  • +
+

In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate.

+
+ +

You can restore the configuration information by uploading the XML outputted by the backup. Specify the XML file, click the restore button on the data.

+

If there is already to enable overwriting of data, the same data does update existing data.

+
+
+ +
diff --git a/src/site/en/xdoc/3.0/admin/dataStoreCrawling-guide.xml b/src/site/en/xdoc/3.0/admin/dataStoreCrawling-guide.xml new file mode 100644 index 000000000..1ed9547ce --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/dataStoreCrawling-guide.xml @@ -0,0 +1,129 @@ + + + + Data store configuration + Sone, Takaaki + + +
+

You can crawl databases in Fess. Here are required to store settings.

+ +

In Administrator account after logging in, click menu data store.

+ Data store configuration +

As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

+ +
+ +

Parameter settings example looks like the following.

+ +

Parameter is a "key = value" format. Description of the key is as follows.

+ + + + + + + + + + + + + + + + + + + + + + + +
driverDriver class name
URLURL
usernameTo connect to the DB user name
passwordTo connect to the DB password
SQLWant to crawl to get SQL statement
+
+ +

Script configuration example looks like the following.

+ +

+ Parameter is a "key = value" format. + Description of the key is as follows.

+

+ Side of the value written in OGNL. String, tie up in double quotation marks. + Access in the database column name, its value.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
URLURLs (links appear in search results)
hostHost name
siteSite pass
titleTitle
contentContent (string index)
cacheContent cache (not indexed)
DigestDigest piece that appears in the search results
anchorLinks to content (not usually required)
contentLengthThe length of the content
lastModifiedContent last updated
+
+ +

To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

+
+
+ +
diff --git a/src/site/en/xdoc/3.0/admin/design-guide.xml b/src/site/en/xdoc/3.0/admin/design-guide.xml new file mode 100644 index 000000000..7af81069e --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/design-guide.xml @@ -0,0 +1,101 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
+

Here are settings for the design of search screens.

+ +

In Administrator account after logging in, click the menu design.

+ Design +

You can edit the search screen in the screen below.

+ JSP compilation screen +
+ +

You can upload the image files to use in the search screen. Image file names are supported are jpg, gif and png.

+
+ +

If you want the file name to upload image files to use. Uploaded if you omit the file name will be used.

+
+ +

You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

+

Following are examples of how to write.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
Top page (within the Head tags)This is the express search home page head tag in JSP files. If you want to edit the meta tags, title tags, script tags, such as the change.
Top page (content)Is a JSP file to represent the body tag in the search home page.
Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
Search results page (within the Head tags)Search result is a JSP file to represent within the head tag of the list page. If you want to edit the meta tags, title tags, script tags, such as the change.
Search results page (header)Search result is a JSP file to represent the header of the list page. Include search form at the top.
Search results page (footer)Search result is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
+

You can to edit for PCs and similar portable screen.

+
+
+
+ +

If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

+ + +<% +java.util.Map docMap = (java.util.Map)pageContext.getAttribute("doc"); +Long tstampValue = (Long)docMap.get("tstamp"); +java.util.Date tstampDate = new java.util.Date(tstampValue); +Long lastModifiedValue = (Long)docMap.get("lastModified"); +java.util.Date lastModifiedDate = new java.util.Date(lastModifiedValue); +java.text.SimpleDateFormat sdf = new java.text.SimpleDateFormat("yyyy/MM/dd HH:mm"); +%> +
  • +

    + ${f:h(doc.contentTitle)} +

    +
    +${doc.contentDescription} +
    + ${f:h(doc.site)} +
    Registered: <%= sdf.format(tstampDate) %> +
    Last Modified: <%= sdf.format(lastModifiedDate) %> + : + : +]]> +

    tstampDate will update on registration date, lastModifiedDate. Output date format is specified in SimpeDateFormat.

    + +
  • + +
    diff --git a/src/site/en/xdoc/3.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/3.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..47b371320 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,96 @@ + + + + Settings for crawling a file system using + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + +
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. > For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/index.xml b/src/site/en/xdoc/3.0/admin/index.xml new file mode 100644 index 000000000..791718de2 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/index.xml @@ -0,0 +1,12 @@ + + + + Management UI Guide + Shinsuke Sugaya + + +
    +

    Here, is the description of the Fess 3.0 management UI.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/labelType-guide.xml b/src/site/en/xdoc/3.0/admin/labelType-guide.xml new file mode 100644 index 000000000..7342c399a --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/labelType-guide.xml @@ -0,0 +1,23 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. If you register the label shown select label drop-down box to the right of the search box.

    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/log-guide.xml b/src/site/en/xdoc/3.0/admin/log-guide.xml new file mode 100644 index 000000000..23437f94b --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/log-guide.xml @@ -0,0 +1,19 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/3.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..e3f7bc2ad --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/overlappingHost-guide.xml @@ -0,0 +1,23 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/3.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..8edd530bb --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/pathMapping-guide.xml @@ -0,0 +1,26 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/3.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..e31bdf85e --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/requestHeader-guide.xml @@ -0,0 +1,26 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/roleType-guide.xml b/src/site/en/xdoc/3.0/admin/roleType-guide.xml new file mode 100644 index 000000000..0e16265d2 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/roleType-guide.xml @@ -0,0 +1,23 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/system-guide.xml b/src/site/en/xdoc/3.0/admin/system-guide.xml new file mode 100644 index 000000000..b0408f448 --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/system-guide.xml @@ -0,0 +1,28 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here registration in Fess. SOLR servers are grouped by file, has been registered.

    + +

    In Administrator account after logging in, click menu Solr.

    + System settings +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and shut down when not running Fess server to shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +
    + +

    Server group name is used to search for and update appears.

    +
    + +

    Server becomes unavailable and the status of disabled. For example, inaccessible to the Solr server and changes to disabled. To enable recovery after server become unavailable will become available.

    +
    + +

    You can publish index commit, optimize for server groups. You can also remove a specific search for the session ID.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/3.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..34037e25b --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/webAuthentication-guide.xml @@ -0,0 +1,37 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication or DIGEST authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/3.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..b55cee33b --- /dev/null +++ b/src/site/en/xdoc/3.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,99 @@ + + + + Settings for crawling the Web using + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + +
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in)

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/filesize.xml b/src/site/en/xdoc/3.0/config/filesize.xml new file mode 100644 index 000000000..dc6c6adb0 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/filesize.xml @@ -0,0 +1,28 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of defaultMaxLength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/index-backup.xml b/src/site/en/xdoc/3.0/config/index-backup.xml new file mode 100644 index 000000000..930ea9df7 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index backup and restore + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data directory. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/index.xml b/src/site/en/xdoc/3.0/config/index.xml new file mode 100644 index 000000000..064e9aeaa --- /dev/null +++ b/src/site/en/xdoc/3.0/config/index.xml @@ -0,0 +1,12 @@ + + + + Set up Guide + Shinsuke Sugaya + + +
    +

    Here is the Fess 3.0 Setup instructions.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/logging.xml b/src/site/en/xdoc/3.0/config/logging.xml new file mode 100644 index 000000000..306bdd0cd --- /dev/null +++ b/src/site/en/xdoc/3.0/config/logging.xml @@ -0,0 +1,43 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + +
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina.outLog of the Fess Server (Tomcat). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/memory-config.xml b/src/site/en/xdoc/3.0/config/memory-config.xml new file mode 100644 index 000000000..a4e5006fd --- /dev/null +++ b/src/site/en/xdoc/3.0/config/memory-config.xml @@ -0,0 +1,42 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case the maximum value set 1024M) will change to-Xmx1024m.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-Xmx1024m (in this case the maximum value set 1024M). +

    + + "fess" + "role1" + new String[] { + "-Djava.awt.headless=true", "-server", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Dpdfbox.cjk.support=true", "-Xmx1024m", + "-XX:MaxPermSize=128m" } + +]]> +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/mobile-device.xml b/src/site/en/xdoc/3.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/path-encoding.xml b/src/site/en/xdoc/3.0/config/path-encoding.xml new file mode 100644 index 000000000..093d57cb3 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    + +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/3.0/config/protwords.xml b/src/site/en/xdoc/3.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/proxy.xml b/src/site/en/xdoc/3.0/config/proxy.xml new file mode 100644 index 000000000..db48fd846 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/proxy.xml @@ -0,0 +1,57 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    + +

    Proxy is set in to create webapps/Fess/Web-INF/classes/s9robot_client.dicon with the following contents.

    + + + + + + + + @org.apache.commons.httpclient.cookie.CookiePolicy@BROWSER_COMPATIBILITY + "プロキシホスト名" + プロキシポート + + + + + "UTF-8" + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/3.0/config/replication.xml b/src/site/en/xdoc/3.0/config/replication.xml new file mode 100644 index 000000000..916e94207 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/replication.xml @@ -0,0 +1,25 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Fess can copy the path in Solr index data. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +

    You must use the replication features of Fess for Solr index file in the shared disk, such as NFS, Fess of each can be referenced from.

    +
    +
    + +

    Fess, download and install the./ /NET/Server1/usr/local/Fess To assume you installed.

    +

    To register the crawl settings as well as Fess starts after the normal construction, create the index (index for Fess building instructions normal building procedures and especially remains the same) crawling.

    +
    + +

    Fess, download and install the./ /NET/Server2/usr/local/Fess To assume you installed.

    +

    To enable replication features check box in Fess starts after the management screen crawl settings the "snapshot path'. Snapshot path designates the index location for the index for Fess. In this case, the/NET/Server1/usr/local/Fess //solr/core1/data/index In the will.

    + Replication +

    Time press the update button to save the data and set in Schedule performs replication of the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/role-setting.xml b/src/site/en/xdoc/3.0/config/role-setting.xml new file mode 100644 index 000000000..91321034a --- /dev/null +++ b/src/site/en/xdoc/3.0/config/role-setting.xml @@ -0,0 +1,97 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    + +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/app.dicon shown below.

    + + + {"guest"} + + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + "role1" + + : +]]> +

    authenticatedRoles can describe multiple by commas (,).

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/server-port.xml b/src/site/en/xdoc/3.0/config/server-port.xml new file mode 100644 index 000000000..c55bb3010 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/server-port.xml @@ -0,0 +1,30 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information. change the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + "http://localhost:8080/solr" +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/solr-dynamic-field.xml b/src/site/en/xdoc/3.0/config/solr-dynamic-field.xml new file mode 100644 index 000000000..483b0a5f9 --- /dev/null +++ b/src/site/en/xdoc/3.0/config/solr-dynamic-field.xml @@ -0,0 +1,48 @@ + + + + How to use the dynamic field of SOLR + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in solr/core1/conf/schema.xml. dynamic fields and standard fields such as title and content can be freely defined field names are defined. The dynamic fields that are available in the schema.xml Fess become. Advanced parameter values see a Solr document.

    + + + + + + + + + + + + + + + + + + + + + + +]]> +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add fields for the following in the dynamic field data out of Solr using webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/solr-failure.xml b/src/site/en/xdoc/3.0/config/solr-failure.xml new file mode 100644 index 000000000..4b487e43b --- /dev/null +++ b/src/site/en/xdoc/3.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is minActiveServer following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status maxRetryStatusCheckCount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    + +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    + +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/config/tokenizer.xml b/src/site/en/xdoc/3.0/config/tokenizer.xml new file mode 100644 index 000000000..4181e30ff --- /dev/null +++ b/src/site/en/xdoc/3.0/config/tokenizer.xml @@ -0,0 +1,36 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search.

    +

    Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits.

    +

    For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit.

    +

    For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default CJKTokenizer used.

    +
    + + +

    Such as CJKTokenizer Japan Japanese multibyte string against bi-gram, in other words two characters create index. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to.

    +

    The following example to change schema.xml so analyzer parts, you can use the StandardTokenizer.

    + + + + + : +]]> +
    + + +
    diff --git a/src/site/en/xdoc/3.0/config/windows-service.xml b/src/site/en/xdoc/3.0/config/windows-service.xml new file mode 100644 index 000000000..a277c3a1a --- /dev/null +++ b/src/site/en/xdoc/3.0/config/windows-service.xml @@ -0,0 +1,49 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as Add %JAVA_HOME%\bin to PathYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option. (No pdfbox.cjk.support options from 3.1.0)

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-3.0.0.

    + cd C:\Java\fess-server-3.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat6w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/user/index.xml b/src/site/en/xdoc/3.0/user/index.xml new file mode 100644 index 000000000..73f4da08e --- /dev/null +++ b/src/site/en/xdoc/3.0/user/index.xml @@ -0,0 +1,12 @@ + + + + Search Guide + Shinsuke Sugaya + + +
    +

    Here is the instructions on how to search for Fess 3.0.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/user/search-field.xml b/src/site/en/xdoc/3.0/user/search-field.xml new file mode 100644 index 000000000..b30448189 --- /dev/null +++ b/src/site/en/xdoc/3.0/user/search-field.xml @@ -0,0 +1,57 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    In the Fess crawl results saved in the title and text fields. You can search for a field of them.

    +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields title and content to search for.

    + +

    If a field search "field name: search terms ' of so fill out the search form, the search.

    +

    Title against Fess the search as a search term.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/3.0/user/search-label.xml b/src/site/en/xdoc/3.0/user/search-label.xml new file mode 100644 index 000000000..c9c078bbe --- /dev/null +++ b/src/site/en/xdoc/3.0/user/search-label.xml @@ -0,0 +1,14 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    By label to be registered in the management screen will enable search by labels in the search screen. You can use the label if you want to sort the search results. If you do not register the label displayed the label drop-down box.

    + Search by label +

    To set the label by creating indexes, can search each crawl settings specified on the label. All results search search do not specify a label is usually the same.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/user/search-not.xml b/src/site/en/xdoc/3.0/user/search-not.xml new file mode 100644 index 000000000..64edf7813 --- /dev/null +++ b/src/site/en/xdoc/3.0/user/search-not.xml @@ -0,0 +1,15 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If you want to find documents that do not contain a Word can NOT find. + Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, searches, enter if you want to find documents that contain the search term 1 does not contain a search term 2 search term 1 NOT search words 2.

    +

    Attention is required because NOT find expensive.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/user/search-or.xml b/src/site/en/xdoc/3.0/user/search-or.xml new file mode 100644 index 000000000..221a2ec80 --- /dev/null +++ b/src/site/en/xdoc/3.0/user/search-or.xml @@ -0,0 +1,15 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. + When describing the multiple words in the search box, by default will search. + You want OR search the case describes OR between search words. OR write in capital letters, spaces are required before and after.

    +

    For example, the search, enter if you want to search for documents that contain either search term 2 search term 1 search term 1 OR search term 2. OR between multiple languages are available.

    +
    + +
    diff --git a/src/site/en/xdoc/3.0/user/search-sort.xml b/src/site/en/xdoc/3.0/user/search-sort.xml new file mode 100644 index 000000000..bceb8e612 --- /dev/null +++ b/src/site/en/xdoc/3.0/user/search-sort.xml @@ -0,0 +1,44 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + +
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    + +

    If you want to sort ' sort: field name ' in to fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/browserType-guide.xml b/src/site/en/xdoc/4.0/admin/browserType-guide.xml new file mode 100644 index 000000000..1a88c20bc --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/browserType-guide.xml @@ -0,0 +1,19 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
    +

    Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

    + +

    In Administrator account after logging in, click menu browser types.

    + Setting the browser type +
    + +

    You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/config-wizard.xml b/src/site/en/xdoc/4.0/admin/config-wizard.xml new file mode 100644 index 000000000..875271261 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/config-wizard.xml @@ -0,0 +1,34 @@ + + + + Settings Wizard + Sone, Takaaki + + +
    +

    Introduction to the Configuration Wizard.

    +

    You can use Settings Wizard, to set you up on the fess.

    + +

    In Administrator account after logging in, click menu Settings Wizard.

    + Settings Wizard +

    First, setting a schedule.

    +

    During the time in fess is crawling and indexes.

    +

    By default, every day is a 0 時 0 分.

    + Setting a schedule +

    The crawl settings.

    +

    Crawl settings is to register a URI to look for.

    +

    The crawl settings name please put name of any easy to identify.

    +

    Put the URI part de-indexed, want to search for.

    + Crawl settings +

    For example, if you want search for http://example.com, below looks like.

    + Crawl settings example +

    In this is the last setting.

    +

    Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

    + Crawl started +
    + +

    Settings in the Setup Wizard you can change from crawl General, Web, file system.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/crawl-guide.xml b/src/site/en/xdoc/4.0/admin/crawl-guide.xml new file mode 100644 index 000000000..3f90227b6 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/crawl-guide.xml @@ -0,0 +1,139 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to crawling.

    + +

    In Administrator account click crawl General menu after login.

    + Crawl General +

    You can specify the path to a generated index and replication capabilities to enable.

    + Replication features +
    + +

    You can set the interval at which the crawl for a Web site or file system. By default, the following.

    + +

    Figures are from left, seconds, minutes, during the day, month, represents a day of the week. Description format is similar to the Unix cron settings. This example, and am 0 時 0 分 to crawling daily.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    0 0 12 * *?Each day starts at 12 pm
    0 15 10? * *Day 10: 15 am start
    0 15 10 * *?Day 10: 15 am start
    0 15 10 * *? *Day 10: 15 am start
    0 15 10 * *? 2005Each of the 2009 start am, 10:15
    0 * 14 * *?Every day 2:00 in the PM-2: 59 pm start every 1 minute
    0 0 / 5 14 * *?Every day 2:00 in the PM-2: 59 pm start every 5 minutes
    0 0 / 5 14, 18 * *?Every day 2:00 pm-2: 59 pm and 6: 00 starts every 5 minutes at the PM-6: 59 pm
    0 0-5 14 * *?Every day 2:00 in the PM-2: 05 pm start every 1 minute
    0 10, 44 14? 3 WEDStarts Wednesday March 2: 10 and 2: 44 pm
    0 15 10? * MON-FRIMonday through Friday at 10:15 am start
    +

    Also check if the seconds can be set to run at intervals 60 seconds by default. If you set seconds exactly and you should customize webapps/fess/WEB-INF/classes/chronosCustomize.dicon taskScanIntervalTime value, if enough do I see in one-hour increments.

    +
    + +

    When the user enters a search, the search the output log. If you want to get search statistics to enable.

    +
    + +

    Search results link attaches to the search term. To display the find search terms in PDF becomes possible.

    +
    + +

    Search results can be retrieved in XML format. http://localhost:8080/Fess/XML? can get access query = search term.

    +
    + +

    Search results available in JSON format. http://localhost:8080/Fess/JSON? can get access query = search term.

    +
    + +

    If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

    +
    + +

    You can specify the label to see if the label by default,. Specifies the value of the label.

    +
    + +

    You can specify whether or not to display a search screen. If you select Web unusable for mobile search screen. If not available not available search screen. And if you want to create a dedicated index server and select not available.

    +
    + +

    In JSON format often find search words becomes available. can be retrieved by accessing the http://localhost:8080/Fess/hotsearchword.

    +
    + +

    Delete a session log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Delete a search log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Specifies the Bots name Bots you want to remove from the search log logs included in the user agent by commas (,). Log is deleted by log purge once a day.

    +
    + +

    Specifies the encoding for the CSV will be available in the backup and restore.

    +
    + +

    To enable replication features that can apply already copied the Solr index generated. For example, you can use them if you want to search only in the search servers crawled and indexed on a different server, placed in front.

    +
    + +

    After the data is registered for Solr. Index to commit or to optimize the registered data becomes available. If optimize is issued the Solr index optimization, if you have chosen, you choose to commit the commit is issued.

    +
    + +

    Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

    +
    + +

    To raise the performance of the index in Fess while crawling and sends for Solr document in 20 units. For each value specified here because without committing to continue adding documents documents added in the Solr on performance, Solr issued document commits. By default, after you add documents 1000 is committed.

    +
    + +

    Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

    +

    But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

    +
    + +

    You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

    +
    + +

    Registered disabled URL URL exceeds the failure count next time you crawl to crawl out. No need to worry about disability type is crawled next time by specifying this value.

    +
    + +

    Disaster URL exceeds the number of failures will crawl out.

    +
    + +

    Copy index information from the index directory as the snapshot path, if replication is enabled, will be applied.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/4.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..fa010c7bd --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/crawlingSession-guide.xml @@ -0,0 +1,34 @@ + + + + Set session information + Shinsuke Sugaya + + +
    +

    Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

    + +

    In Administrator account after logging in, click the session information menu.

    +
    + + Session information list +

    You can remove all session information and click the Delete link all in the running.

    +
    + + Session details +

    To specify a session ID, you can see crawling content.

    +
      +
    • Information about the entire crawl Cralwer *:
    • +
    • FsCrawl *: information about the file system crawling
    • +
    • WebCrawl *: crawling the Web information
    • +
    • Information issued by Solr server optimization optimize *:
    • +
    • Commit *: information about the commit was issued to the Solr server.
    • +
    • * StartTime: start time
    • +
    • * EndTime: end time
    • +
    • * ExecTime: execution time (MS)
    • +
    • * IndexSize: number of documents indexed
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/data-guide.xml b/src/site/en/xdoc/4.0/admin/data-guide.xml new file mode 100644 index 000000000..5730aaa1e --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/data-guide.xml @@ -0,0 +1,33 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
    +

    Here, describes Fess information backup and restore methods.

    + +

    In Administrator account after logging in, click the menu backup and restore.

    + Backup and restore +
    + +

    Click the download link and Fess information output in XML format. Saved settings information is below.

    +
      +
    • The General crawl settings
    • +
    • Web crawl settings
    • +
    • File system Crawl settings
    • +
    • Path mapping
    • +
    • Web authentication
    • +
    • Compatible browsers
    • +
    +

    Session information, search log, click log is available in CSV format.

    +

    In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate.

    +
    + +

    You can restore settings information, various log in to upload XML output by backup or CSV. To specify the files, please click the restore button on the data.

    +

    If enable overwrite data in XML file configuration information specified when the same data is updating existing data.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/dataStoreCrawling-guide.xml b/src/site/en/xdoc/4.0/admin/dataStoreCrawling-guide.xml new file mode 100644 index 000000000..5edc1cd38 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/dataStoreCrawling-guide.xml @@ -0,0 +1,129 @@ + + + + Data store configuration + Sone, Takaaki + + +
    +

    You can crawl databases in Fess. Here are required to store settings.

    + +

    In Administrator account after logging in, click menu data store.

    + Data store configuration +

    As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

    + +
    + +

    Parameter settings example looks like the following.

    + +

    Parameter is a "key = value" format. Description of the key is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + +
    driverDriver class name
    URLURL
    usernameTo connect to the DB user name
    passwordTo connect to the DB password
    SQLWant to crawl to get SQL statement
    +
    + +

    Script configuration example looks like the following.

    + +

    + Parameter is a "key = value" format. + Description of the key is as follows.

    +

    + Side of the value written in OGNL. String, tie up in double quotation marks. + Access in the database column name, its value.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    URLURLs (links appear in search results)
    hostHost name
    siteSite pass
    titleTitle
    contentContent (string index)
    cacheContent cache (not indexed)
    DigestDigest piece that appears in the search results
    anchorLinks to content (not usually required)
    contentLengthThe length of the content
    lastModifiedContent last updated
    +
    + +

    To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/design-guide.xml b/src/site/en/xdoc/4.0/admin/design-guide.xml new file mode 100644 index 000000000..23ec22751 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/design-guide.xml @@ -0,0 +1,101 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
    +

    Here are settings for the design of search screens.

    + +

    In Administrator account after logging in, click the menu design.

    + Design +

    You can edit the search screen in the screen below.

    + JSP compilation screen +
    + +

    You can upload the image files to use in the search screen. Image file names are supported are jpg, gif and png.

    +
    + +

    If you want the file name to upload image files to use. Uploaded if you omit the file name will be used.

    +
    + +

    You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
    Top page (within the Head tags)This is the express search home page head tag in JSP files. If you want to edit the meta tags, title tags, script tags, such as the change.
    Top page (content)Is a JSP file to represent the body tag in the search home page.
    Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
    Search results page (within the Head tags)Search result is a JSP file to represent within the head tag of the list page. If you want to edit the meta tags, title tags, script tags, such as the change.
    Search results page (header)Search result is a JSP file to represent the header of the list page. Include search form at the top.
    Search results page (footer)Search result is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
    Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    +

    You can to edit for PCs and similar portable screen.

    +
    +
    +
    + +

    If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

    + + +<% +java.util.Map docMap = (java.util.Map)pageContext.getAttribute("doc"); +Long tstampValue = (Long)docMap.get("tstamp"); +java.util.Date tstampDate = new java.util.Date(tstampValue); +Long lastModifiedValue = (Long)docMap.get("lastModified"); +java.util.Date lastModifiedDate = new java.util.Date(lastModifiedValue); +java.text.SimpleDateFormat sdf = new java.text.SimpleDateFormat("yyyy/MM/dd HH:mm"); +%> +
  • +

    + ${f:h(doc.contentTitle)} +

    +
    +${doc.contentDescription} +
    + ${f:h(doc.site)} +
    Registered: <%= sdf.format(tstampDate) %> +
    Last Modified: <%= sdf.format(lastModifiedDate) %> + : + : +]]> +

    tstampDate will update on registration date, lastModifiedDate. Output date format is specified in SimpeDateFormat.

    + +
  • + +
    diff --git a/src/site/en/xdoc/4.0/admin/failureUrl-guide.xml b/src/site/en/xdoc/4.0/admin/failureUrl-guide.xml new file mode 100644 index 000000000..8a2469126 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/failureUrl-guide.xml @@ -0,0 +1,21 @@ + + + + Disaster URL + Shinsuke Sugaya + + +
    +

    Here the failure URL. URL could not be obtained at crawl time are recorded and confirmed as the failure URL.

    + +

    In Administrator account click menu disabled URL after login.

    + Disaster URL +

    Clicking the confirmation link failure URL displayed for more information.

    + Details of the disaster URL +
    + +

    A glance could not crawl the URL and date.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/fileAuthentication-guide.xml b/src/site/en/xdoc/4.0/admin/fileAuthentication-guide.xml new file mode 100644 index 000000000..991676a42 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/fileAuthentication-guide.xml @@ -0,0 +1,40 @@ + + + + Settings for file system authentication + Shinsuke Sugaya + + +
    +

    Crawls using file system here, describes how to set file system authentication is required. Fess is corresponding to a crawl for a shared folder in Windows.

    + +

    In Administrator account after logging in, click the menu file system authentication.

    + File system settings +
    + +

    Specifies the host name of the site that requires authentication. Is omitted, the specified file system Kroll set applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. File system Crawl settings specified in that case applies on any port.

    +
    + +

    Select the authentication method. You can use SAMBA (Windows shared folder authentication).

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. SAMBA, the set value of the domain. If you want to write as.

    + +
    + +

    Select a file name to apply the authentication settings for the above. Must be registered ago you file system CROLL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/4.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..2634dfd32 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,98 @@ + + + + Settings for crawling a file system using + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: or smb: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +

    Windows shared folder, for example, if you want to crawl to host1 share folder crawl settings for smb: (last / to) the //host1/share/. If authentication is in the shared folder on the file system authentication screen set authentication information.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + +
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. > For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/index.xml b/src/site/en/xdoc/4.0/admin/index.xml new file mode 100644 index 000000000..3acaa4207 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/index.xml @@ -0,0 +1,12 @@ + + + + Management UI Guide + Shinsuke Sugaya + + +
    +

    Here, is the description of the Fess 4.0 management UI.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/labelType-guide.xml b/src/site/en/xdoc/4.0/admin/labelType-guide.xml new file mode 100644 index 000000000..960ae181d --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/labelType-guide.xml @@ -0,0 +1,29 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. If you register the label shown select label drop-down box to the right of the search box.

    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    + +

    Specifies the role to view the label.

    +
    + +

    Specifies the order of the labels.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/log-guide.xml b/src/site/en/xdoc/4.0/admin/log-guide.xml new file mode 100644 index 000000000..b338d22bc --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/log-guide.xml @@ -0,0 +1,19 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/4.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..4d05537fb --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/overlappingHost-guide.xml @@ -0,0 +1,23 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/4.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..b34423e59 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/pathMapping-guide.xml @@ -0,0 +1,26 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/4.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..984b61752 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/requestHeader-guide.xml @@ -0,0 +1,26 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/roleType-guide.xml b/src/site/en/xdoc/4.0/admin/roleType-guide.xml new file mode 100644 index 000000000..a56c16f7c --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/roleType-guide.xml @@ -0,0 +1,23 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/search-guide.xml b/src/site/en/xdoc/4.0/admin/search-guide.xml new file mode 100644 index 000000000..84c4f7a40 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/search-guide.xml @@ -0,0 +1,19 @@ + + + + Search + Shinsuke Sugaya + + +
    +

    Here the search for management.

    + +

    In Administrator account after logging in, click the menu search.

    + Administrative search +
    + +

    You can search by criteria you specify. In the regular search screen role and browser requirements is added implicitly, but do not provide management for search. You can document a certain remove from index from the search results.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/searchLog-guide.xml b/src/site/en/xdoc/4.0/admin/searchLog-guide.xml new file mode 100644 index 000000000..81a49c155 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/searchLog-guide.xml @@ -0,0 +1,19 @@ + + + + Search log settings + Shinsuke Sugaya + + +
    +

    Here the search log. When you search in the search screen users search logs are logged. Search log search term or date is recorded. You can also record the URL, then you want the search results to.

    + +

    In Administrator account after logging in, click menu search logs.

    + Search log +
    + +

    Search language and date are listed. You can review and detailed, you click the URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/stats-guide.xml b/src/site/en/xdoc/4.0/admin/stats-guide.xml new file mode 100644 index 000000000..dc04bc572 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/stats-guide.xml @@ -0,0 +1,19 @@ + + + + Statistics + Shinsuke Sugaya + + +
    +

    Here are statistics. You can search log and click log.

    + +

    In Administrator account after logging in, click the menu statistics.

    + Statistics +
    + +

    You can select the target by selecting the type of report, to make sure. Displayed in order by the specified criteria.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/system-guide.xml b/src/site/en/xdoc/4.0/admin/system-guide.xml new file mode 100644 index 000000000..90791b167 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/system-guide.xml @@ -0,0 +1,31 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here registration in Fess. SOLR servers are grouped by file, has been registered.

    + +

    In Administrator account after logging in, click menu Solr.

    + System settings +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and shut down when not running Fess server to shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +
    + +

    Server group name is used to search for and update appears.

    +
    + +

    Server becomes unavailable and the status of disabled. For example, inaccessible to the Solr server and changes to disabled. To enable recovery after server become unavailable will become available.

    +
    + +

    You can publish index commit, optimize for server groups. You can also remove a specific search for the session ID. You can remove only the specific documents by specifying the URL.

    +
    + +

    Shown by the number of documents registered in each session. Can verify the results list by clicking the session name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/systemInfo-guide.xml b/src/site/en/xdoc/4.0/admin/systemInfo-guide.xml new file mode 100644 index 000000000..d1f28fc5f --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/systemInfo-guide.xml @@ -0,0 +1,28 @@ + + + + System information + Shinsuke Sugaya + + +
    +

    Here, you can currently check property information such as system environment variables.

    + +

    In Administrator account after logging in, click system information menu.

    + System information +
    + +

    You can list the server environment variable.

    +
    + +

    You can list the system properties on Fess.

    +
    + +

    Fess setup information available.

    +
    + +

    Is a list of properties to attach when reporting a bug. Extract the value contains no personal information.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/4.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..6187cf5ae --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/webAuthentication-guide.xml @@ -0,0 +1,44 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication, DIGEST authentication or NTLM authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. You can set the workstation and domain values for NTLM authentication. If you want to write as.

    + +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/4.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..24417d168 --- /dev/null +++ b/src/site/en/xdoc/4.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,99 @@ + + + + Settings for crawling the Web using + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + +
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in)

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/desktop-search.xml b/src/site/en/xdoc/4.0/config/desktop-search.xml new file mode 100644 index 000000000..fd01926c6 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/desktop-search.xml @@ -0,0 +1,35 @@ + + + + The desktop search settings + Shinsuke Sugaya + + +
    +

    + Increasing awareness of security in the browser environment in recent years, open a local file (for example, c:\hoge.txt) from the Web pages on. + Not to copy and paste the link from the search results, and then reopen the usability is good. + In order to respond to this in Fess and provides desktop search functionality.

    +
    +
    +

    + Desktop Search feature is turned off by default. + Please enable the following settings.

    +

    First of all, bin/setenv.bat as java.awt.headless from true to false edits.

    + +

    Then add the following to webapps/fess/WEB-INF/conf/crawler.properties.

    + +

    Start the Fess, after you set up above. How to use Basic remains especially.

    +
    +
    + +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/filesize.xml b/src/site/en/xdoc/4.0/config/filesize.xml new file mode 100644 index 000000000..dc6c6adb0 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/filesize.xml @@ -0,0 +1,28 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of defaultMaxLength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/index-backup.xml b/src/site/en/xdoc/4.0/config/index-backup.xml new file mode 100644 index 000000000..930ea9df7 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index backup and restore + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data directory. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/index.xml b/src/site/en/xdoc/4.0/config/index.xml new file mode 100644 index 000000000..de14b4810 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/index.xml @@ -0,0 +1,12 @@ + + + + Set up Guide + Shinsuke Sugaya + + +
    +

    Here is the Fess 4.0 Setup instructions.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/install-on-tomcat.xml b/src/site/en/xdoc/4.0/config/install-on-tomcat.xml new file mode 100644 index 000000000..0aec49b11 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/install-on-tomcat.xml @@ -0,0 +1,43 @@ + + + + Install to an existing Tomcat + Shinsuke Sugaya + + +
    +

    + The standard distribution of Fess Tomcat is distributed in the deployed State. + Because Fess is not dependent on Tomcat, deploying on any Java application server is available. + Describes how to deploy a Fess Tomcat here is already available.

    + +

    Fess ServerHereFrom the download.

    +
    + +

    + Expand the downloaded Fess server. + Expanded Fess Server home directory to $FESS_HOME. + $TOMCAT_HOME the top directory of an existing Tomcat 6. + Copy the Fess Server data.

    + +

    + If you have, such as changing the destination file diff commands, updates your diff only applies. +

    +
    + +

    to start the Tomcat in normal as well as in strtup.* / http://localhost:8080/Fess To access the.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/logging.xml b/src/site/en/xdoc/4.0/config/logging.xml new file mode 100644 index 000000000..553ef0bdc --- /dev/null +++ b/src/site/en/xdoc/4.0/config/logging.xml @@ -0,0 +1,43 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + +
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina,outLog of the Fess Server (Tomcat). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/memory-config.xml b/src/site/en/xdoc/4.0/config/memory-config.xml new file mode 100644 index 000000000..7ca6c9ace --- /dev/null +++ b/src/site/en/xdoc/4.0/config/memory-config.xml @@ -0,0 +1,49 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case the maximum value set 1024M) will change to-Xmx1024m.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-Xmx1024m (in this case the maximum value set 1024M). +

    + new String[] { +"-Djava.awt.headless=true", +"-server", +"-Xmx1024m", +"-XX:MaxPermSize=128m", +"-XX:-UseGCOverheadLimit", +"-XX:+UseConcMarkSweepGC", +"-XX:CMSInitiatingOccupancyFraction=75", +"-XX:+CMSIncrementalMode", +"-XX:+CMSIncrementalPacing", +"-XX:CMSIncrementalDutyCycleMin=0", +"-XX:+UseParNewGC", +"-XX:+UseStringCache", +"-XX:+UseTLAB", +"-XX:+DisableExplicitGC" +} +]]> +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/mobile-device.xml b/src/site/en/xdoc/4.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/path-encoding.xml b/src/site/en/xdoc/4.0/config/path-encoding.xml new file mode 100644 index 000000000..093d57cb3 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    + +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/4.0/config/pdf-with-password.xml b/src/site/en/xdoc/4.0/config/pdf-with-password.xml new file mode 100644 index 000000000..0bc9da638 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/pdf-with-password.xml @@ -0,0 +1,256 @@ + + + + Password-protected PDF support + Shinsuke Sugaya + + +
    +

    + You should password files to register the settings file to PDF password is configured to search for. +

    +
    +
    +

    + First of all, create the webapps/fess/WEB-INF/classes/s2robot_extractor.dicon. + This is test _ ~ is a pass that password set to a.pdf file. + If you have multiple files, multiple settings in addPassword.

    + + + + + + + + + + + + + ".*test_.*.pdf" + + "pass" + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + { +"application/xml", +"application/xhtml+xml", +"application/rdf+xml", +"text/xml", +"text/xml-external-parsed-entity" + } + xmlExtractor + + + { +"text/html" + } + xmlExtractor + + + { +"application/pdf" + } + pdfExtractor + + + { +"image/svg+xml", +"application/x-tika-msoffice", +"application/vnd.visio", +"application/vnd.ms-powerpoint", +"application/vnd.ms-excel", +"application/vnd.ms-excel.sheet.binary.macroenabled.12", +"application/msword", +"application/vnd.ms-outlook", +"application/x-tika-ooxml", +"application/vnd.openxmlformats-officedocument.presentationml.presentation", +"application/vnd.ms-powerpoint.presentation.macroenabled.12", +"application/vnd.openxmlformats-officedocument.presentationml.template", +"application/vnd.openxmlformats-officedocument.presentationml.slideshow", +"application/vnd.ms-powerpoint.slideshow.macroenabled.12", +"application/vnd.ms-powerpoint.addin.macroenabled.12", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.ms-excel.sheet.macroenabled.12", +"application/vnd.openxmlformats-officedocument.spreadsheetml.template", +"application/vnd.ms-excel.template.macroenabled.12", +"application/vnd.ms-excel.addin.macroenabled.12", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.ms-word.document.macroenabled.12", +"application/vnd.openxmlformats-officedocument.wordprocessingml.template", +"application/vnd.ms-word.template.macroenabled.12", +"application/x-asp", +"application/rtf", +"text/plain", +"application/vnd.sun.xml.writer", +"application/vnd.oasis.opendocument.text", +"application/vnd.oasis.opendocument.graphics", +"application/vnd.oasis.opendocument.presentation", +"application/vnd.oasis.opendocument.spreadsheet", +"application/vnd.oasis.opendocument.chart", +"application/vnd.oasis.opendocument.image", +"application/vnd.oasis.opendocument.formula", +"application/vnd.oasis.opendocument.text-master", +"application/vnd.oasis.opendocument.text-web", +"application/vnd.oasis.opendocument.text-template", +"application/vnd.oasis.opendocument.graphics-template", +"application/vnd.oasis.opendocument.presentation-template", +"application/vnd.oasis.opendocument.spreadsheet-template", +"application/vnd.oasis.opendocument.chart-template", +"application/vnd.oasis.opendocument.image-template", +"application/vnd.oasis.opendocument.formula-template", +"application/x-vnd.oasis.opendocument.text", +"application/x-vnd.oasis.opendocument.graphics", +"application/x-vnd.oasis.opendocument.presentation", +"application/x-vnd.oasis.opendocument.spreadsheet", +"application/x-vnd.oasis.opendocument.chart", +"application/x-vnd.oasis.opendocument.image", +"application/x-vnd.oasis.opendocument.formula", +"application/x-vnd.oasis.opendocument.text-master", +"application/x-vnd.oasis.opendocument.text-web", +"application/x-vnd.oasis.opendocument.text-template", +"application/x-vnd.oasis.opendocument.graphics-template", +"application/x-vnd.oasis.opendocument.presentation-template", +"application/x-vnd.oasis.opendocument.spreadsheet-template", +"application/x-vnd.oasis.opendocument.chart-template", +"application/x-vnd.oasis.opendocument.image-template", +"application/x-vnd.oasis.opendocument.formula-template", +"image/bmp", +"image/gif", +"image/jpeg", +"image/png", +"image/tiff", +"image/vnd.wap.wbmp", +"image/x-icon", +"image/x-psd", +"image/x-xcf", +"application/zip", +"application/x-tar", +"application/x-gtar", +"application/x-gzip", +"application/x-bzip", +"application/x-bzip2", +"application/java-vm", +"audio/mpeg", +"application/x-midi", +"audio/midi", +"audio/basic", +"audio/x-wav", +"audio/x-aiff", +"application/mbox", +"text/calendar", +"text/css", +"text/csv", +"text/directory", +"text/dns", +"text/ecmascript", +"text/enriched", +"text/example", +"text/javascript", +"text/parityfec", +"text/prs.fallenstein.rst", +"text/prs.lines.tag", +"text/red", +"text/rfc822-headers", +"text/richtext", +"text/rtf", +"text/rtp-enc-aescm128", +"text/rtx", +"text/sgml", +"text/t140", +"text/tab-separated-values", +"text/troff", +"text/ulpfec", +"text/uri-list", +"text/vnd.abc", +"text/vnd.curl", +"text/vnd.curl.dcurl", +"text/vnd.curl.mcurl", +"text/vnd.curl.scurl", +"text/vnd.dmclientscript", +"text/vnd.esmertec.theme-descriptor", +"text/vnd.fly", +"text/vnd.fmi.flexstor", +"text/vnd.graphviz", +"text/vnd.in3d.3dml", +"text/vnd.in3d.spot", +"text/vnd.iptc.newsml", +"text/vnd.iptc.nitf", +"text/vnd.latex-z", +"text/vnd.motorola.reflex", +"text/vnd.ms-mediapackage", +"text/vnd.net2phone.commcenter.command", +"text/vnd.si.uricatalogue", +"text/vnd.sun.j2me.app-descriptor", +"text/vnd.trolltech.linguist", +"text/vnd.wap.si", +"text/vnd.wap.sl", +"text/vnd.wap.wml", +"text/vnd.wap.wmlscript", +"text/x-asm", +"text/x-c", +"text/x-diff", +"text/x-fortran", +"text/x-java-source", +"text/x-pascal", +"text/x-setext", +"text/x-uuencode", +"text/x-vcalendar", +"text/x-vcard", +"application/x-sh" + } + tikaExtractor + + + + +]]> +

    Next, edit the following to webapps/fess/WEB-INF/classes/s2robot_rule.dicon.

    + + "fsFileRule" + + + fessFileTransformer + + + true + + "url" + "file:.*" + + + "mimeType" + + + "(application/xml" ++ "|application/xhtml+xml" ++ "|application/rdf+xml" ++ "|application/pdf" ++ "|text/xml" ++ "|text/xml-external-parsed-entity" ++ "|text/html)" + + + +... +]]> +

    Fess up after setting the above and please run the crawl. How to use Basic remains especially.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/protwords.xml b/src/site/en/xdoc/4.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/proxy.xml b/src/site/en/xdoc/4.0/config/proxy.xml new file mode 100644 index 000000000..61f67d09d --- /dev/null +++ b/src/site/en/xdoc/4.0/config/proxy.xml @@ -0,0 +1,95 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    + +

    Proxy is set in to create webapps/fess/WEB-INF/classes/s2robot_client.dicon with the following contents.

    + + + + + + + + @org.apache.http.client.params.CookiePolicy@BEST_MATCH + clientConnectionManager + "プロキシサーバー名(ex. 192.168.1.1)" + プロキシサーバーのポート(ex. 8080) (" は不要) + + + + + + + + + + "http" + 80 + @org.apache.http.conn.scheme.PlainSocketFactory@getSocketFactory() + + + + + + + "https" + 443 + @org.apache.http.conn.ssl.SSLSocketFactory@getSocketFactory() + + + + + + 200 + 200 + + + + + "UTF-8" + + + + "UTF-8" + + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + "smb:.*" + smbClient + + + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/4.0/config/replication.xml b/src/site/en/xdoc/4.0/config/replication.xml new file mode 100644 index 000000000..b21333444 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/replication.xml @@ -0,0 +1,25 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Fess can copy the path in Solr index data. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +

    You must use the replication features of Fess for Solr index file in the shared disk, such as NFS, Fess of each can be referenced from.

    +
    +
    + +

    Fess, download and install the./ /NET/Server1/usr/local/Fess To assume you installed.

    +

    To register the crawl settings as well as Fess starts after the normal construction, create the index (index for Fess building instructions normal building procedures and especially remains the same) crawling.

    +
    + +

    Fess, download and install the./ /NET/Server2/usr/local/Fess To assume you installed.

    +

    To enable replication features check box in Fess starts after the management screen crawl settings the "snapshot path'. Snapshot path designates the index location for the index for Fess. In this case, the/NET/Server1/usr/local/Fess //solr/core1/data/index In the will.

    + Replication +

    Time press the update button to save the data and set in Schedule performs replication of the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/role-setting.xml b/src/site/en/xdoc/4.0/config/role-setting.xml new file mode 100644 index 000000000..91321034a --- /dev/null +++ b/src/site/en/xdoc/4.0/config/role-setting.xml @@ -0,0 +1,97 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    + +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/app.dicon shown below.

    + + + {"guest"} + + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + "role1" + + : +]]> +

    authenticatedRoles can describe multiple by commas (,).

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/server-port.xml b/src/site/en/xdoc/4.0/config/server-port.xml new file mode 100644 index 000000000..8438aab1f --- /dev/null +++ b/src/site/en/xdoc/4.0/config/server-port.xml @@ -0,0 +1,31 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    • 19092: database port (use h2database)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information. change the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + "http://localhost:8080/solr" +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/solr-dynamic-field.xml b/src/site/en/xdoc/4.0/config/solr-dynamic-field.xml new file mode 100644 index 000000000..483b0a5f9 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/solr-dynamic-field.xml @@ -0,0 +1,48 @@ + + + + How to use the dynamic field of SOLR + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in solr/core1/conf/schema.xml. dynamic fields and standard fields such as title and content can be freely defined field names are defined. The dynamic fields that are available in the schema.xml Fess become. Advanced parameter values see a Solr document.

    + + + + + + + + + + + + + + + + + + + + + + +]]> +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add fields for the following in the dynamic field data out of Solr using webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/solr-failure.xml b/src/site/en/xdoc/4.0/config/solr-failure.xml new file mode 100644 index 000000000..4b487e43b --- /dev/null +++ b/src/site/en/xdoc/4.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is minActiveServer following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status maxRetryStatusCheckCount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    + +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    + +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/config/tokenizer.xml b/src/site/en/xdoc/4.0/config/tokenizer.xml new file mode 100644 index 000000000..4181e30ff --- /dev/null +++ b/src/site/en/xdoc/4.0/config/tokenizer.xml @@ -0,0 +1,36 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search.

    +

    Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits.

    +

    For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit.

    +

    For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default CJKTokenizer used.

    +
    + + +

    Such as CJKTokenizer Japan Japanese multibyte string against bi-gram, in other words two characters create index. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to.

    +

    The following example to change schema.xml so analyzer parts, you can use the StandardTokenizer.

    + + + + + : +]]> +
    + + +
    diff --git a/src/site/en/xdoc/4.0/config/windows-service.xml b/src/site/en/xdoc/4.0/config/windows-service.xml new file mode 100644 index 000000000..a0caee850 --- /dev/null +++ b/src/site/en/xdoc/4.0/config/windows-service.xml @@ -0,0 +1,54 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as To add the path %jv_home%\BinYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option.

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-4.0.0.

    + cd C:\Java\fess-server-4.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat6w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    +
    + +

    Distributed in the Fess is 32-bit binaries for Windows Tomcat builds based on. If you use 64-bit WindowsTomcat For 64 bit Windows zip, such as getting from the site and replace tomcat6.exe, tomcat6w.exe.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/index.xml b/src/site/en/xdoc/4.0/user/index.xml new file mode 100644 index 000000000..91852b3e3 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/index.xml @@ -0,0 +1,12 @@ + + + + Search Guide + Shinsuke Sugaya + + +
    +

    Here is the instructions on how to search for Fess 4.0.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/json-response.xml b/src/site/en/xdoc/4.0/user/json-response.xml new file mode 100644 index 000000000..645fb51d7 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/json-response.xml @@ -0,0 +1,182 @@ + + + + Search results for JSON output + Shinsuke Sugaya + + +
    +

    Outputs a JSON results of Fess. You must in order to emit JSON enabled keep JSON response in the General Management screen crawl settings.

    + +

    To get JSON output http://localhost:8080/Fess/JSON? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + + + + +
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    labelTypeValueThe label value. Use to specify the label.
    callbackIf you use the JSONP callback name. You do not need to specify if you do not use the JSONP.
    +
    + +

    Returns the following response.

    + \u30C6\u30B9\u30C8<\u002Fem>\u30DA\u30FC\u30B8\u3067\u3059", + "host": "speedtest.goo.ne.jp", + "lastModified": "1284739487873", + "cache": "\u9FA0-->\n\n\ngoo \u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\n<\u002Ftitle>\n<meta...\u0022>\n\n<meta name=\u0022description\u0022 content=\u0022goo \u306E\u63D0\u4F9B\u3059\u308B\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u30DA\u30FC\u30B8\u3067\u3059\u0022>\n<meta name=\u0022keywords\u0022 content=\u0022ADSL,\u30D6\u30ED\u30FC\u30C9\u30D0\u30F3\u30C9,\u30D5\u30EC...\u30C3\u30C4,BB,\u7121\u7DDALAN,CATV,ISDN,\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u0022>\n<meta name=\u0022Copyright\u0022 content=\u0022Copyright © (C)NTT-X Inc, All...:\u002F\u002Fspeedtest.goo.ne.jp\u002F\u0022><img src=\u0022\u002Fimg\u002Fh1\u002Fh1_76.gif\u0022 width=\u0022129\u0022 height=\u002222\u0022 border=\u00220\u0022 alt=\u0022\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u0022><img src...\u0022 color=\u0022#ffffff\u0022>goo \u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem><\u002Ffont><\u002Fa><\u002Fb><\u002Ftd>\n<!--\u002FNAME SERVICE-->\n<\u002Ftr>\n<\u002Ftable>\n<\u002Ftd>\n<\u002Ftr>\n<tr>\n<td", + "score": "4.98744", + "digest": "goo \u306E\u63D0\u4F9B\u3059\u308B\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u30DA\u30FC\u30B8\u3067\u3059", + "tstamp": "1284739487887", + "url": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F", + "id": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F;type=au,docomo,pc,softbank", + "mimetype": "text\u002Fhtml", + "title": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "contentTitle": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "boost": "1.0", + "contentLength": "17810", + "urlLink": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F" + }, +... + ] + } +} +]]></source> + <p>Each element is as follows.</p> + <table> + <tbody> + <tr> + <th>response</th> + <td>The root element.</td> + </tr> + <tr> + <th>version</th> + <td>Format version.</td> + </tr> + <tr> + <th>status</th> + <td>The status of the response. 0, the normal response.</td> + </tr> + <tr> + <th>query</th> + <td>Search language.</td> + </tr> + <tr> + <th>execTime</th> + <td>Response time. Unit is seconds.</td> + </tr> + <tr> + <th>pageSize</th> + <td>The number of display.</td> + </tr> + <tr> + <th>pageNumber</th> + <td>Page number.</td> + </tr> + <tr> + <th>recordCount</th> + <td>Number of hits for the search term.</td> + </tr> + <tr> + <th>pageCount</th> + <td>Page number of the number of hits for the search term.</td> + </tr> + <tr> + <th>result</th> + <td>Parent element of the search results.</td> + </tr> + <tr> + <th>site</th> + <td>The name of the site.</td> + </tr> + <tr> + <th>contentDescription</th> + <td>A description of the content.</td> + </tr> + <tr> + <th>host</th> + <td>The name of the host.</td> + </tr> + <tr> + <th>lastModified</th> + <td>Last update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.</td> + </tr> + <tr> + <th>cache</th> + <td>The contents of the content.</td> + </tr> + <tr> + <th>score</th> + <td>Score values of the document.</td> + </tr> + <tr> + <th>Digest</th> + <td>Digest a string of documents.</td> + </tr> + <tr> + <th>Tstamp</th> + <td>Document creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.</td> + </tr> + <tr> + <th>URL</th> + <td>The URL of the document.</td> + </tr> + <tr> + <th>ID</th> + <td>The ID of the document.</td> + </tr> + <tr> + <th>mimetype</th> + <td>The MIME type.</td> + </tr> + <tr> + <th>title</th> + <td>The title of the document.</td> + </tr> + <tr> + <th>contentTitle</th> + <td>The title of the document for viewing.</td> + </tr> + <tr> + <th>contentLength</th> + <td>The size of the document.</td> + </tr> + <tr> + <th>urlLink</th> + <td>The URL as the search results.</td> + </tr> + </tbody> + </table> + </subsection> + </section> + </body> +</document> diff --git a/src/site/en/xdoc/4.0/user/search-boost.xml b/src/site/en/xdoc/4.0/user/search-boost.xml new file mode 100644 index 000000000..8661056d2 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-boost.xml @@ -0,0 +1,15 @@ +<?xml version='1.0' encoding='UTF-8'?> +<document> + <properties> + <title>Boost search + Shinsuke Sugaya + + +
    +

    + Use the boost search if you want to prioritize, search for specific search terms. + For example, if you want to find the page if you want to find apples oranges contained more 'apples' ' Apple ^ 100 orange ' that to ' ^ number "that searches in the form. + Number specifies an integer greater than 1.

    +
    + + diff --git a/src/site/en/xdoc/4.0/user/search-field.xml b/src/site/en/xdoc/4.0/user/search-field.xml new file mode 100644 index 000000000..b30448189 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-field.xml @@ -0,0 +1,57 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    In the Fess crawl results saved in the title and text fields. You can search for a field of them.

    +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields title and content to search for.

    + +

    If a field search "field name: search terms ' of so fill out the search form, the search.

    +

    Title against Fess the search as a search term.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/search-fuzzy.xml b/src/site/en/xdoc/4.0/user/search-fuzzy.xml new file mode 100644 index 000000000..601376b79 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-fuzzy.xml @@ -0,0 +1,18 @@ + + + + Fuzzy search (fuzzy search) + Shinsuke Sugaya + + +
    +

    Fess is corresponding to fuzzy searches based on the Levenshtein distance (fuzzy search). + After the search word you want to apply the fuzzy search adds '~'. + For example, see Solr ~ ' that you can search for documents that contain the search string "Solr" similar to the language or ("Solar").

    +

    + Furthermore, you can specify a number between 0 and 1 ~ after the close to 1 just like in refine. + For example, in the form of 'Solr~0.8'. + Do not specify numeric default value is 0.5.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/search-label.xml b/src/site/en/xdoc/4.0/user/search-label.xml new file mode 100644 index 000000000..0f4adece6 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-label.xml @@ -0,0 +1,14 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    By label to be registered in the management screen will enable search by labels in the search screen. You can use the label if you want to sort the search results. If you do not register the label displayed the label drop-down box.

    + Search by label +

    To set the label by creating indexes, can search each crawl settings specified on the label. All results search search do not specify a label is usually the same.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/search-not.xml b/src/site/en/xdoc/4.0/user/search-not.xml new file mode 100644 index 000000000..cc733646b --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-not.xml @@ -0,0 +1,15 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If you want to find documents that do not contain a Word can NOT find. + Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, searches, enter if you want to find documents that contain the search term 1 does not contain a search term 2 search term 1 NOT search words 2.

    +

    Attention is required because NOT find expensive.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/search-or.xml b/src/site/en/xdoc/4.0/user/search-or.xml new file mode 100644 index 000000000..acb8df0bc --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-or.xml @@ -0,0 +1,15 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. + When describing the multiple words in the search box, by default will search. + You want OR search the case describes OR between search words. OR write in capital letters, spaces are required before and after.

    +

    For example, the search, enter if you want to search for documents that contain either search term 2 search term 1 search term 1 OR search term 2. OR between multiple languages are available.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/search-range.xml b/src/site/en/xdoc/4.0/user/search-range.xml new file mode 100644 index 000000000..fb9c95595 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-range.xml @@ -0,0 +1,14 @@ + + + + Range searches + Shinsuke Sugaya + + +
    +

    Range searches can be done for field. + To range search field name: value TO the search term. + For example, if you search documents contentLength field against 1 k to 10 k bytes is ' contentLength: 1000 TO 10000 ' search language and the.

    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/search-sort.xml b/src/site/en/xdoc/4.0/user/search-sort.xml new file mode 100644 index 000000000..bceb8e612 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/search-sort.xml @@ -0,0 +1,44 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + +
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    + +

    If you want to sort ' sort: field name ' in to fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/4.0/user/xml-response.xml b/src/site/en/xdoc/4.0/user/xml-response.xml new file mode 100644 index 000000000..619938da5 --- /dev/null +++ b/src/site/en/xdoc/4.0/user/xml-response.xml @@ -0,0 +1,201 @@ + + + + XML results output + Shinsuke Sugaya + + +
    +

    To emit the XML results of Fess. You must in order to emit the XML Management screen crawl settings XML response enabled by you.

    + +

    To get the output XML http://localhost:8080/Fess/XML? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + +
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    labelTypeValueThe label value. Use to specify the label.
    +
    + +

    Returns the following response.

    + + + 1 + 0 + テスト + 0.64 + 20 + 1 + 101 + 6 + + + speedtest.goo.ne.jp/ + goo の提供するスピード<em>テスト</em>ページです + speedtest.goo.ne.jp + 1284739487873 + 龠--> +<meta http-equiv="Content-Type" content="text/html; charset=EUC-JP"> +<title> +goo スピード<em>テスト</em> +</title> +<meta..."> + +<meta name="description" content="goo の提供するスピード<em>テスト</em>ページです"> +<meta name="keywords" content="ADSL,ブロードバンド,フレ...ッツ,BB,無線LAN,CATV,ISDN,スピード<em>テスト</em>"> +<meta name="Copyright" content="Copyright &copy; (C)NTT-X Inc, All...://speedtest.goo.ne.jp/"><img src="/img/h1/h1_76.gif" width="129" height="22" border="0" alt="スピード<em>テスト</em>"><img src..." color="#ffffff">goo スピード<em>テスト</em></font></a></b></td> +<!--/NAME SERVICE--> +</tr> +</table> +</td> +</tr> +<tr> +<td + 4.98744 + goo の提供するスピード<em>テスト</em>ページです + 1284739487887 + http://speedtest.goo.ne.jp/ + http://speedtest.goo.ne.jp/;type=au,docomo,pc,softbank + text/html + +goo スピードテスト + + +goo スピードテスト + + 1.0 + 17810 + http://speedtest.goo.ne.jp/ + +... + + +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    responseThe root element.
    versionFormat version.
    statusThe status of the response. 0, the normal response.
    querySearch language.
    exec-timeResponse time. Unit is seconds.
    page-sizeThe number of display.
    page-numberPage number.
    record-countNumber of hits for the search term.
    page-countPage number of the number of hits for the search term.
    resultParent element of the search results.
    docThe elements of the search results.
    siteThe name of the site.
    content-descriptionA description of the content.
    hostThe name of the host.
    last-modifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    content-titleThe title of the document for viewing.
    content-lengthThe size of the document.
    URL-linkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/browserType-guide.xml b/src/site/en/xdoc/5.0/admin/browserType-guide.xml new file mode 100644 index 000000000..8e229d1ad --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/browserType-guide.xml @@ -0,0 +1,19 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
    +

    Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

    + +

    In Administrator account after logging in, click menu browser types.

    + Setting the browser type +
    + +

    You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/config-wizard.xml b/src/site/en/xdoc/5.0/admin/config-wizard.xml new file mode 100644 index 000000000..44a6336b6 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/config-wizard.xml @@ -0,0 +1,34 @@ + + + + Settings Wizard + Sone, Takaaki + + +
    +

    Introduction to the Configuration Wizard.

    +

    You can use Settings Wizard, to set you up on the fess.

    + +

    In Administrator account after logging in, click menu Settings Wizard.

    + Settings Wizard +

    First, setting a schedule.

    +

    During the time in fess is crawling and indexes.

    +

    By default, every day is a 0 時 0 分.

    + Setting a schedule +

    The crawl settings.

    +

    Crawl settings is to register a URI to look for.

    +

    The crawl settings name please put name of any easy to identify.

    +

    Put the URI part de-indexed, want to search for.

    + Crawl settings +

    For example, if you want search for http://example.com, below looks like.

    + Crawl settings example +

    In this is the last setting.

    +

    Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

    + Crawl started +
    + +

    Settings in the Setup Wizard you can change from crawl General, Web, file system.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/crawl-guide.xml b/src/site/en/xdoc/5.0/admin/crawl-guide.xml new file mode 100644 index 000000000..ed8fc3c13 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/crawl-guide.xml @@ -0,0 +1,139 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to crawling.

    + +

    In Administrator account click crawl General menu after login.

    + Crawl General +

    You can specify the path to a generated index and replication capabilities to enable.

    + Replication features +
    + +

    You can set the interval at which the crawl for a Web site or file system. By default, the following.

    + +

    Figures are from left, seconds, minutes, during the day, month, represents a day of the week. Description format is similar to the Unix cron settings. This example, and am 0 時 0 分 to crawling daily.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    0 0 12 * *?Each day starts at 12 pm
    0 15 10? * *Day 10: 15 am start
    0 15 10 * *?Day 10: 15 am start
    0 15 10 * *? *Day 10: 15 am start
    0 15 10 * *? 2005Each of the 2009 start am, 10:15
    0 * 14 * *?Every day 2:00 in the PM-2: 59 pm start every 1 minute
    0 0 / 5 14 * *?Every day 2:00 in the PM-2: 59 pm start every 5 minutes
    0 0 / 5 14, 18 * *?Every day 2:00 pm-2: 59 pm and 6: 00 starts every 5 minutes at the PM-6: 59 pm
    0 0-5 14 * *?Every day 2:00 in the PM-2: 05 pm start every 1 minute
    0 10, 44 14? 3 WEDStarts Wednesday March 2: 10 and 2: 44 pm
    0 15 10? * MON-FRIMonday through Friday at 10:15 am start
    +

    Also check if the seconds can be set to run at intervals 60 seconds by default. If you set seconds exactly and you should customize webapps/fess/WEB-INF/classes/chronosCustomize.dicon taskScanIntervalTime value, if enough do I see in one-hour increments.

    +
    + +

    When the user enters a search, the search the output log. If you want to get search statistics to enable.

    +
    + +

    Search results link attaches to the search term. To display the find search terms in PDF becomes possible.

    +
    + +

    Search results can be retrieved in XML format. http://localhost:8080/Fess/XML? can get access query = search term.

    +
    + +

    Search results available in JSON format. http://localhost:8080/Fess/JSON? can get access query = search term.

    +
    + +

    If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

    +
    + +

    You can specify the label to see if the label by default,. Specifies the value of the label.

    +
    + +

    You can specify whether or not to display a search screen. If you select Web unusable for mobile search screen. If not available not available search screen. And if you want to create a dedicated index server and select not available.

    +
    + +

    In JSON format often find search words becomes available. can be retrieved by accessing the http://localhost:8080/Fess/hotsearchword.

    +
    + +

    Delete a session log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Delete a search log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Specifies the Bots name Bots you want to remove from the search log logs included in the user agent by commas (,). Log is deleted by log purge once a day.

    +
    + +

    Specifies the encoding for the CSV will be available in the backup and restore.

    +
    + +

    To enable replication features that can apply already copied the Solr index generated. For example, you can use them if you want to search only in the search servers crawled and indexed on a different server, placed in front.

    +
    + +

    After the data is registered for Solr. Index to commit or to optimize the registered data becomes available. If optimize is issued the Solr index optimization, if you have chosen, you choose to commit the commit is issued.

    +
    + +

    Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

    +
    + +

    To raise the performance of the index in Fess while crawling and sends for Solr document in 20 units. For each value specified here because without committing to continue adding documents documents added in the Solr on performance, Solr issued document commits. By default, after you add documents 1000 is committed.

    +
    + +

    Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

    +

    But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

    +
    + +

    You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

    +
    + +

    Registered disabled URL URL exceeds the failure count next time you crawl to crawl out. No need to worry about disability type is crawled next time by specifying this value.

    +
    + +

    Disaster URL exceeds the number of failures will crawl out.

    +
    + +

    Copy index information from the index directory as the snapshot path, if replication is enabled, will be applied.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/5.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..1a48a8ff6 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/crawlingSession-guide.xml @@ -0,0 +1,34 @@ + + + + Set session information + Shinsuke Sugaya + + +
    +

    Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

    + +

    In Administrator account after logging in, click the session information menu.

    +
    + + Session information list +

    You can remove all session information and click the Delete link all in the running.

    +
    + + Session details +

    To specify a session ID, you can see crawling content.

    +
      +
    • Information about the entire crawl Cralwer *:
    • +
    • FsCrawl *: information about the file system crawling
    • +
    • WebCrawl *: crawling the Web information
    • +
    • Information issued by Solr server optimization optimize *:
    • +
    • Commit *: information about the commit was issued to the Solr server.
    • +
    • * StartTime: start time
    • +
    • * EndTime: end time
    • +
    • * ExecTime: execution time (MS)
    • +
    • * IndexSize: number of documents indexed
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/data-guide.xml b/src/site/en/xdoc/5.0/admin/data-guide.xml new file mode 100644 index 000000000..ea76c8137 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/data-guide.xml @@ -0,0 +1,33 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
    +

    Here, describes Fess information backup and restore methods.

    + +

    In Administrator account after logging in, click the menu backup and restore.

    + Backup and restore +
    + +

    Click the download link and Fess information output in XML format. Saved settings information is below.

    +
      +
    • The General crawl settings
    • +
    • Web crawl settings
    • +
    • File system Crawl settings
    • +
    • Path mapping
    • +
    • Web authentication
    • +
    • Compatible browsers
    • +
    +

    Session information, search log, click log is available in CSV format.

    +

    In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate.

    +
    + +

    You can restore settings information, various log in to upload XML output by backup or CSV. To specify the files, please click the restore button on the data.

    +

    If enable overwrite data in XML file configuration information specified when the same data is updating existing data.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/dataStoreCrawling-guide.xml b/src/site/en/xdoc/5.0/admin/dataStoreCrawling-guide.xml new file mode 100644 index 000000000..4075637c2 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/dataStoreCrawling-guide.xml @@ -0,0 +1,153 @@ + + + + Data store configuration + Sone, Takaaki + Shinsuke Sugaya + + +
    +

    You can crawl databases in Fess. Here are required to store settings.

    + +

    In Administrator account after logging in, click menu data store.

    + Data store configuration +

    As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

    + +

    Here the data is put something like the following.

    + +
    + +

    Parameter settings example looks like the following.

    + +

    Parameter is a "key = value" format. Description of the key is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + +
    driverDriver class name
    URLURL
    usernameTo connect to the DB user name
    passwordTo connect to the DB password
    SQLWant to crawl to get SQL statement
    +
    + +

    Script configuration example looks like the following.

    + +

    + Parameter is a "key = value" format. + Description of the key is as follows.

    +

    + Side of the value written in OGNL. String, tie up in double quotation marks. + Access in the database column name, its value.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    URLURLs (links appear in search results)
    hostHost name
    siteSite pass
    titleTitle
    contentContent (string index)
    cacheContent cache (not indexed)
    DigestDigest piece that appears in the search results
    anchorLinks to content (not usually required)
    contentLengthThe length of the content
    lastModifiedContent last updated
    +
    + +

    To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

    +
    + +

    Set the following in the webapps/fess/WEB-INF/classes/app.dicon if you see the item value, such as latitude_s in the search results. After adding to $ {doc.latitude_s}, searchResults.jsp;

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "latitude_s","longitude_s" } + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/design-guide.xml b/src/site/en/xdoc/5.0/admin/design-guide.xml new file mode 100644 index 000000000..69ac3d3f5 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/design-guide.xml @@ -0,0 +1,101 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
    +

    Here are settings for the design of search screens.

    + +

    In Administrator account after logging in, click the menu design.

    + Design +

    You can edit the search screen in the screen below.

    + JSP compilation screen +
    + +

    You can upload the image files to use in the search screen. Image file names are supported are jpg, gif and png.

    +
    + +

    If you want the file name to upload image files to use. Uploaded if you omit the file name will be used.

    +
    + +

    You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
    Top page (within the Head tags)This is the express search home page head tag in JSP files. If you want to edit the meta tags, title tags, script tags, such as the change.
    Top page (content)Is a JSP file to represent the body tag in the search home page.
    Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
    Search results page (within the Head tags)Search result is a JSP file to represent within the head tag of the list page. If you want to edit the meta tags, title tags, script tags, such as the change.
    Search results page (header)Search result is a JSP file to represent the header of the list page. Include search form at the top.
    Search results page (footer)Search result is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
    Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    +

    You can to edit for PCs and similar portable screen.

    +
    +
    +
    + +

    If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

    + + +<% +java.util.Map docMap = (java.util.Map)pageContext.getAttribute("doc"); +Long tstampValue = (Long)docMap.get("tstamp"); +java.util.Date tstampDate = new java.util.Date(tstampValue); +Long lastModifiedValue = (Long)docMap.get("lastModified"); +java.util.Date lastModifiedDate = new java.util.Date(lastModifiedValue); +java.text.SimpleDateFormat sdf = new java.text.SimpleDateFormat("yyyy/MM/dd HH:mm"); +%> +
  • +

    + ${f:h(doc.contentTitle)} +

    +
    +${doc.contentDescription} +
    + ${f:h(doc.site)} +
    Registered: <%= sdf.format(tstampDate) %> +
    Last Modified: <%= sdf.format(lastModifiedDate) %> + : + : +]]> +

    tstampDate will update on registration date, lastModifiedDate. Output date format is specified in SimpeDateFormat.

    + +
  • + +
    diff --git a/src/site/en/xdoc/5.0/admin/failureUrl-guide.xml b/src/site/en/xdoc/5.0/admin/failureUrl-guide.xml new file mode 100644 index 000000000..f86a47f57 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/failureUrl-guide.xml @@ -0,0 +1,21 @@ + + + + Disaster URL + Shinsuke Sugaya + + +
    +

    Here the failure URL. URL could not be obtained at crawl time are recorded and confirmed as the failure URL.

    + +

    In Administrator account click menu disabled URL after login.

    + Disaster URL +

    Clicking the confirmation link failure URL displayed for more information.

    + Details of the disaster URL +
    + +

    A glance could not crawl the URL and date.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/fileAuthentication-guide.xml b/src/site/en/xdoc/5.0/admin/fileAuthentication-guide.xml new file mode 100644 index 000000000..5e07af7ca --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/fileAuthentication-guide.xml @@ -0,0 +1,40 @@ + + + + Settings for file system authentication + Shinsuke Sugaya + + +
    +

    Crawls using file system here, describes how to set file system authentication is required. Fess is corresponding to a crawl for a shared folder in Windows.

    + +

    In Administrator account after logging in, click the menu file system authentication.

    + File system settings +
    + +

    Specifies the host name of the site that requires authentication. Is omitted, the specified file system Kroll set applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. File system Crawl settings specified in that case applies on any port.

    +
    + +

    Select the authentication method. You can use SAMBA (Windows shared folder authentication).

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. SAMBA, the set value of the domain. If you want to write as.

    + +
    + +

    Select a file name to apply the authentication settings for the above. Must be registered ago you file system CROLL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/5.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..3356ff798 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,98 @@ + + + + Settings for crawling a file system using + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: or smb: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +

    Windows shared folder, for example, if you want to crawl to host1 share folder crawl settings for smb: (last / to) the //host1/share/. If authentication is in the shared folder on the file system authentication screen set authentication information.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + +
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. > For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/index.xml b/src/site/en/xdoc/5.0/admin/index.xml new file mode 100644 index 000000000..7ba491f74 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/index.xml @@ -0,0 +1,12 @@ + + + + Management UI Guide + Shinsuke Sugaya + + +
    +

    Here, is the description of the Fess 5.0 management UI.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/labelType-guide.xml b/src/site/en/xdoc/5.0/admin/labelType-guide.xml new file mode 100644 index 000000000..7f10d61fe --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/labelType-guide.xml @@ -0,0 +1,29 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. If you register the label shown select label drop-down box to the right of the search box.

    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    + +

    Specifies the role to view the label.

    +
    + +

    Specifies the order of the labels.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/log-guide.xml b/src/site/en/xdoc/5.0/admin/log-guide.xml new file mode 100644 index 000000000..3fbf65776 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/log-guide.xml @@ -0,0 +1,19 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/5.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..ccf1ad50b --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/overlappingHost-guide.xml @@ -0,0 +1,23 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/5.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..7e03989d3 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/pathMapping-guide.xml @@ -0,0 +1,26 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/5.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..8cc8a02ce --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/requestHeader-guide.xml @@ -0,0 +1,26 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/roleType-guide.xml b/src/site/en/xdoc/5.0/admin/roleType-guide.xml new file mode 100644 index 000000000..920a0329d --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/roleType-guide.xml @@ -0,0 +1,23 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/search-guide.xml b/src/site/en/xdoc/5.0/admin/search-guide.xml new file mode 100644 index 000000000..6e3c256be --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/search-guide.xml @@ -0,0 +1,19 @@ + + + + Search + Shinsuke Sugaya + + +
    +

    Here the search for management.

    + +

    In Administrator account after logging in, click the menu search.

    + Administrative search +
    + +

    You can search by criteria you specify. In the regular search screen role and browser requirements is added implicitly, but do not provide management for search. You can document a certain remove from index from the search results.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/searchLog-guide.xml b/src/site/en/xdoc/5.0/admin/searchLog-guide.xml new file mode 100644 index 000000000..f289ea4a9 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/searchLog-guide.xml @@ -0,0 +1,19 @@ + + + + Search log settings + Shinsuke Sugaya + + +
    +

    Here the search log. When you search in the search screen users search logs are logged. Search log search term or date is recorded. You can also record the URL, then you want the search results to.

    + +

    In Administrator account after logging in, click menu search logs.

    + Search log +
    + +

    Search language and date are listed. You can review and detailed, you click the URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/stats-guide.xml b/src/site/en/xdoc/5.0/admin/stats-guide.xml new file mode 100644 index 000000000..13ef47c9b --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/stats-guide.xml @@ -0,0 +1,19 @@ + + + + Statistics + Shinsuke Sugaya + + +
    +

    Here are statistics. You can search log and click log.

    + +

    In Administrator account after logging in, click the menu statistics.

    + Statistics +
    + +

    You can select the target by selecting the type of report, to make sure. Displayed in order by the specified criteria.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/system-guide.xml b/src/site/en/xdoc/5.0/admin/system-guide.xml new file mode 100644 index 000000000..0040705f3 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/system-guide.xml @@ -0,0 +1,31 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here registration in Fess. SOLR servers are grouped by file, has been registered.

    + +

    In Administrator account after logging in, click menu Solr.

    + System settings +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and shut down when not running Fess server to shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +
    + +

    Server group name is used to search for and update appears.

    +
    + +

    Server becomes unavailable and the status of disabled. For example, inaccessible to the Solr server and changes to disabled. To enable recovery after server become unavailable will become available.

    +
    + +

    You can publish index commit, optimize for server groups. You can also remove a specific search for the session ID. You can remove only the specific documents by specifying the URL.

    +
    + +

    Shown by the number of documents registered in each session. Can verify the results list by clicking the session name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/systemInfo-guide.xml b/src/site/en/xdoc/5.0/admin/systemInfo-guide.xml new file mode 100644 index 000000000..1268f60ce --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/systemInfo-guide.xml @@ -0,0 +1,28 @@ + + + + System information + Shinsuke Sugaya + + +
    +

    Here, you can currently check property information such as system environment variables.

    + +

    In Administrator account after logging in, click system information menu.

    + System information +
    + +

    You can list the server environment variable.

    +
    + +

    You can list the system properties on Fess.

    +
    + +

    Fess setup information available.

    +
    + +

    Is a list of properties to attach when reporting a bug. Extract the value contains no personal information.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/5.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..191ae7a39 --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/webAuthentication-guide.xml @@ -0,0 +1,44 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication, DIGEST authentication or NTLM authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. You can set the workstation and domain values for NTLM authentication. If you want to write as.

    + +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/5.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..0f196a47a --- /dev/null +++ b/src/site/en/xdoc/5.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,99 @@ + + + + Settings for crawling the Web using + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + +
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in)

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/db.xml b/src/site/en/xdoc/5.0/config/db.xml new file mode 100644 index 000000000..bcc0bc510 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/db.xml @@ -0,0 +1,118 @@ + + + + Database changes + Shinsuke Sugaya + + +
    +

    Under normal circumstances the database use the H2 Database. You can use other databases by changing settings.

    +
    +
    + +

    Expand the MySQL binaries.

    +
    + +

    Create a database.

    + create database fess_db; +mysql> grant all privileges on fess_db.* to fess_user@localhost identified by 'fess_pass'; +mysql> create database fess_robot; +mysql> grant all privileges on fess_robot.* to s2robot@localhost identified by 's2robot'; +mysql> FLUSH PRIVILEGES; +]]> +

    Create a table in the database. DDL file is located in extension/mysql.

    + +

    webapps/fess/WEB-INF/lib on mysql driver jar.

    + +

    Edit the webapps/fess/WEB-INF/classes/jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_db?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "fess_user" + "fess_pass" + + + + 600 + 10 + true + + + + + +]]> +

    Edit the webapps/fess/WEB-INF/classes/s2robot_jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_robot?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "s2robot" + "s2robot" + + + + 600 + 10 + true + @java.sql.Connection@TRANSACTION_REPEATABLE_READ + + + + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/desktop-search.xml b/src/site/en/xdoc/5.0/config/desktop-search.xml new file mode 100644 index 000000000..fd01926c6 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/desktop-search.xml @@ -0,0 +1,35 @@ + + + + The desktop search settings + Shinsuke Sugaya + + +
    +

    + Increasing awareness of security in the browser environment in recent years, open a local file (for example, c:\hoge.txt) from the Web pages on. + Not to copy and paste the link from the search results, and then reopen the usability is good. + In order to respond to this in Fess and provides desktop search functionality.

    +
    +
    +

    + Desktop Search feature is turned off by default. + Please enable the following settings.

    +

    First of all, bin/setenv.bat as java.awt.headless from true to false edits.

    + +

    Then add the following to webapps/fess/WEB-INF/conf/crawler.properties.

    + +

    Start the Fess, after you set up above. How to use Basic remains especially.

    +
    +
    +
      +
    • Please Fess inaccessible from the outside, such as (for example, 8080 port does not release).
    • +
    • because false Java.awt.headless image size conversion for mobile devices is not available.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/filesize.xml b/src/site/en/xdoc/5.0/config/filesize.xml new file mode 100644 index 000000000..dc6c6adb0 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/filesize.xml @@ -0,0 +1,28 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of defaultMaxLength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/geosearch.xml b/src/site/en/xdoc/5.0/config/geosearch.xml new file mode 100644 index 000000000..a45779216 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/geosearch.xml @@ -0,0 +1,23 @@ + + + + GEO search + Shinsuke Sugaya + + +
    +

    Together with Google maps, including document with latitude and longitude location information, GEO (GEO) you can use the search.

    +
    +
    + +

    Location is defined as a feed that contains the location information. + When generating the index in Solr latitude longitude set to location feeds in formats such as 45.17614,-93.87341, register the document. + Also sets the value as the latitude_s and longitude_s fields if you want to display latitude and longitude as a search result. * _s is available as a dynamic field of Solr string.

    +
    + +

    During the search specifies in the request parameter to latitude and longitude, the distance. + View the results in the distance (km) specified by distance-based latitude information (latitude, longitude). Latitude and longitude and distances is treated as double.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/index-backup.xml b/src/site/en/xdoc/5.0/config/index-backup.xml new file mode 100644 index 000000000..930ea9df7 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index backup and restore + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data directory. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/index.xml b/src/site/en/xdoc/5.0/config/index.xml new file mode 100644 index 000000000..76a68d7da --- /dev/null +++ b/src/site/en/xdoc/5.0/config/index.xml @@ -0,0 +1,12 @@ + + + + Set up Guide + Shinsuke Sugaya + + +
    +

    Here is the Fess 5.0 Setup instructions.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/install-on-tomcat.xml b/src/site/en/xdoc/5.0/config/install-on-tomcat.xml new file mode 100644 index 000000000..7d1afeaf2 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/install-on-tomcat.xml @@ -0,0 +1,43 @@ + + + + Install to an existing Tomcat + Shinsuke Sugaya + + +
    +

    + The standard distribution of Fess Tomcat is distributed in the deployed State. + Because Fess is not dependent on Tomcat, deploying on any Java application server is available. + Describes how to deploy a Fess Tomcat here is already available.

    + +

    Fess ServerHereFrom the download.

    +
    + +

    + Expand the downloaded Fess server. + Expanded Fess Server home directory to $FESS_HOME. + $TOMCAT_HOME the top directory of an existing Tomcat 6. + Copy the Fess Server data.

    + +

    + If you have, such as changing the destination file diff commands, updates your diff only applies. +

    +
    + +

    to start the Tomcat in normal as well as in startup.* / http://localhost:8080/Fess To access the.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/logging.xml b/src/site/en/xdoc/5.0/config/logging.xml new file mode 100644 index 000000000..306bdd0cd --- /dev/null +++ b/src/site/en/xdoc/5.0/config/logging.xml @@ -0,0 +1,43 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + +
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina.outLog of the Fess Server (Tomcat). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/memory-config.xml b/src/site/en/xdoc/5.0/config/memory-config.xml new file mode 100644 index 000000000..7ca6c9ace --- /dev/null +++ b/src/site/en/xdoc/5.0/config/memory-config.xml @@ -0,0 +1,49 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case the maximum value set 1024M) will change to-Xmx1024m.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-Xmx1024m (in this case the maximum value set 1024M). +

    + new String[] { +"-Djava.awt.headless=true", +"-server", +"-Xmx1024m", +"-XX:MaxPermSize=128m", +"-XX:-UseGCOverheadLimit", +"-XX:+UseConcMarkSweepGC", +"-XX:CMSInitiatingOccupancyFraction=75", +"-XX:+CMSIncrementalMode", +"-XX:+CMSIncrementalPacing", +"-XX:CMSIncrementalDutyCycleMin=0", +"-XX:+UseParNewGC", +"-XX:+UseStringCache", +"-XX:+UseTLAB", +"-XX:+DisableExplicitGC" +} +]]> +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/mobile-device.xml b/src/site/en/xdoc/5.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/path-encoding.xml b/src/site/en/xdoc/5.0/config/path-encoding.xml new file mode 100644 index 000000000..093d57cb3 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    + +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/5.0/config/pdf-with-password.xml b/src/site/en/xdoc/5.0/config/pdf-with-password.xml new file mode 100644 index 000000000..b4c3afcd3 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/pdf-with-password.xml @@ -0,0 +1,256 @@ + + + + Password-protected PDF support + Shinsuke Sugaya + + +
    +

    + You should password files to register the settings file to PDF password is configured to search for. +

    +
    +
    +

    + First of all, create the webapps/fess/WEB-INF/classes/s2robot_extractor.dicon. + This is test _ ~ is a pass that password set to a.pdf file. + If you have multiple files, multiple settings in addPassword.

    + + + + + + + + + + + + + ".*test_.*.pdf" + + "pass" + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + { +"application/xml", +"application/xhtml+xml", +"application/rdf+xml", +"text/xml", +"text/xml-external-parsed-entity" + } + xmlExtractor + + + { +"text/html" + } + xmlExtractor + + + { +"application/pdf" + } + pdfExtractor + + + { +"image/svg+xml", +"application/x-tika-msoffice", +"application/vnd.visio", +"application/vnd.ms-powerpoint", +"application/vnd.ms-excel", +"application/vnd.ms-excel.sheet.binary.macroenabled.12", +"application/msword", +"application/vnd.ms-outlook", +"application/x-tika-ooxml", +"application/vnd.openxmlformats-officedocument.presentationml.presentation", +"application/vnd.ms-powerpoint.presentation.macroenabled.12", +"application/vnd.openxmlformats-officedocument.presentationml.template", +"application/vnd.openxmlformats-officedocument.presentationml.slideshow", +"application/vnd.ms-powerpoint.slideshow.macroenabled.12", +"application/vnd.ms-powerpoint.addin.macroenabled.12", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.ms-excel.sheet.macroenabled.12", +"application/vnd.openxmlformats-officedocument.spreadsheetml.template", +"application/vnd.ms-excel.template.macroenabled.12", +"application/vnd.ms-excel.addin.macroenabled.12", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.ms-word.document.macroenabled.12", +"application/vnd.openxmlformats-officedocument.wordprocessingml.template", +"application/vnd.ms-word.template.macroenabled.12", +"application/x-asp", +"application/rtf", +"text/plain", +"application/vnd.sun.xml.writer", +"application/vnd.oasis.opendocument.text", +"application/vnd.oasis.opendocument.graphics", +"application/vnd.oasis.opendocument.presentation", +"application/vnd.oasis.opendocument.spreadsheet", +"application/vnd.oasis.opendocument.chart", +"application/vnd.oasis.opendocument.image", +"application/vnd.oasis.opendocument.formula", +"application/vnd.oasis.opendocument.text-master", +"application/vnd.oasis.opendocument.text-web", +"application/vnd.oasis.opendocument.text-template", +"application/vnd.oasis.opendocument.graphics-template", +"application/vnd.oasis.opendocument.presentation-template", +"application/vnd.oasis.opendocument.spreadsheet-template", +"application/vnd.oasis.opendocument.chart-template", +"application/vnd.oasis.opendocument.image-template", +"application/vnd.oasis.opendocument.formula-template", +"application/x-vnd.oasis.opendocument.text", +"application/x-vnd.oasis.opendocument.graphics", +"application/x-vnd.oasis.opendocument.presentation", +"application/x-vnd.oasis.opendocument.spreadsheet", +"application/x-vnd.oasis.opendocument.chart", +"application/x-vnd.oasis.opendocument.image", +"application/x-vnd.oasis.opendocument.formula", +"application/x-vnd.oasis.opendocument.text-master", +"application/x-vnd.oasis.opendocument.text-web", +"application/x-vnd.oasis.opendocument.text-template", +"application/x-vnd.oasis.opendocument.graphics-template", +"application/x-vnd.oasis.opendocument.presentation-template", +"application/x-vnd.oasis.opendocument.spreadsheet-template", +"application/x-vnd.oasis.opendocument.chart-template", +"application/x-vnd.oasis.opendocument.image-template", +"application/x-vnd.oasis.opendocument.formula-template", +"image/bmp", +"image/gif", +"image/jpeg", +"image/png", +"image/tiff", +"image/vnd.wap.wbmp", +"image/x-icon", +"image/x-psd", +"image/x-xcf", +"application/zip", +"application/x-tar", +"application/x-gtar", +"application/x-gzip", +"application/x-bzip", +"application/x-bzip2", +"application/java-vm", +"audio/mpeg", +"application/x-midi", +"audio/midi", +"audio/basic", +"audio/x-wav", +"audio/x-aiff", +"application/mbox", +"text/calendar", +"text/css", +"text/csv", +"text/directory", +"text/dns", +"text/ecmascript", +"text/enriched", +"text/example", +"text/javascript", +"text/parityfec", +"text/prs.fallenstein.rst", +"text/prs.lines.tag", +"text/red", +"text/rfc822-headers", +"text/richtext", +"text/rtf", +"text/rtp-enc-aescm128", +"text/rtx", +"text/sgml", +"text/t140", +"text/tab-separated-values", +"text/troff", +"text/ulpfec", +"text/uri-list", +"text/vnd.abc", +"text/vnd.curl", +"text/vnd.curl.dcurl", +"text/vnd.curl.mcurl", +"text/vnd.curl.scurl", +"text/vnd.dmclientscript", +"text/vnd.esmertec.theme-descriptor", +"text/vnd.fly", +"text/vnd.fmi.flexstor", +"text/vnd.graphviz", +"text/vnd.in3d.3dml", +"text/vnd.in3d.spot", +"text/vnd.iptc.newsml", +"text/vnd.iptc.nitf", +"text/vnd.latex-z", +"text/vnd.motorola.reflex", +"text/vnd.ms-mediapackage", +"text/vnd.net2phone.commcenter.command", +"text/vnd.si.uricatalogue", +"text/vnd.sun.j2me.app-descriptor", +"text/vnd.trolltech.linguist", +"text/vnd.wap.si", +"text/vnd.wap.sl", +"text/vnd.wap.wml", +"text/vnd.wap.wmlscript", +"text/x-asm", +"text/x-c", +"text/x-diff", +"text/x-fortran", +"text/x-java-source", +"text/x-pascal", +"text/x-setext", +"text/x-uuencode", +"text/x-vcalendar", +"text/x-vcard", +"application/x-sh" + } + tikaExtractor + + + + +]]> +

    Next, edit the following to Webpps/Fess/Web-INF/clsses/s2robot_rule.dicon.

    + + "fsFileRule" + + + fessFileTransformer + + + true + + "url" + "file:.*" + + + "mimeType" + + + "(application/xml" ++ "|application/xhtml+xml" ++ "|application/rdf+xml" ++ "|application/pdf" ++ "|text/xml" ++ "|text/xml-external-parsed-entity" ++ "|text/html)" + + + +... +]]> +

    Fess up after setting the above and please run the crawl. How to use Basic remains especially.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/protwords.xml b/src/site/en/xdoc/5.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/proxy.xml b/src/site/en/xdoc/5.0/config/proxy.xml new file mode 100644 index 000000000..eefce1426 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/proxy.xml @@ -0,0 +1,95 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    + +

    Proxy is set in to create webapps/Fess/Web-INF/classes/s9robot_client.dicon with the following contents.

    + + + + + + + + @org.apache.http.client.params.CookiePolicy@BEST_MATCH + clientConnectionManager + "プロキシサーバー名(ex. 192.168.1.1)" + プロキシサーバーのポート(ex. 8080) (" は不要) + + + + + + + + + + "http" + 80 + @org.apache.http.conn.scheme.PlainSocketFactory@getSocketFactory() + + + + + + + "https" + 443 + @org.apache.http.conn.ssl.SSLSocketFactory@getSocketFactory() + + + + + + 200 + 200 + + + + + "UTF-8" + + + + "UTF-8" + + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + "smb:.*" + smbClient + + + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/5.0/config/replication.xml b/src/site/en/xdoc/5.0/config/replication.xml new file mode 100644 index 000000000..6c721df59 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/replication.xml @@ -0,0 +1,25 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Fess can copy the path in Solr index data. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +

    You must use the replication features of Fess for Solr index file in the shared disk, such as NFS, Fess of each can be referenced from.

    +
    +
    + +

    Fess, download and install the./ /NET/Server1/usr/local/Fess To assume you installed.

    +

    To register the crawl settings as well as Fess starts after the normal construction, create the index (index for Fess building instructions normal building procedures and especially remains the same) crawling.

    +
    + +

    Fess, download and install the./ /NET/Server2/usr/local/Fess To assume you installed.

    +

    To enable replication features check box in Fess starts after the management screen crawl settings the "snapshot path'. Snapshot path designates the index location for the index for Fess. In this case, the/NET/Server1/usr/local/Fess //solr/core1/data/index In the will.

    + Replication +

    Time press the update button to save the data and set in Schedule performs replication of the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/role-setting.xml b/src/site/en/xdoc/5.0/config/role-setting.xml new file mode 100644 index 000000000..91321034a --- /dev/null +++ b/src/site/en/xdoc/5.0/config/role-setting.xml @@ -0,0 +1,97 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    +
      +
    • Request parameter
    • +
    • Request header
    • +
    • Cookies
    • +
    • J2EE authentication information
    • +
    +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/app.dicon shown below.

    + + + {"guest"} + + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + "role1" + + : +]]> +

    authenticatedRoles can describe multiple by commas (,).

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/server-port.xml b/src/site/en/xdoc/5.0/config/server-port.xml new file mode 100644 index 000000000..8438aab1f --- /dev/null +++ b/src/site/en/xdoc/5.0/config/server-port.xml @@ -0,0 +1,31 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    • 19092: database port (use h2database)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information. change the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + "http://localhost:8080/solr" +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/solr-dynamic-field.xml b/src/site/en/xdoc/5.0/config/solr-dynamic-field.xml new file mode 100644 index 000000000..483b0a5f9 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/solr-dynamic-field.xml @@ -0,0 +1,48 @@ + + + + How to use the dynamic field of SOLR + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in solr/core1/conf/schema.xml. dynamic fields and standard fields such as title and content can be freely defined field names are defined. The dynamic fields that are available in the schema.xml Fess become. Advanced parameter values see a Solr document.

    + + + + + + + + + + + + + + + + + + + + + + +]]> +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add fields for the following in the dynamic field data out of Solr using webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/solr-failure.xml b/src/site/en/xdoc/5.0/config/solr-failure.xml new file mode 100644 index 000000000..a435e9510 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is Minctiveserver following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status Mxretrysttuscheckcount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    + +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    + +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/config/tokenizer.xml b/src/site/en/xdoc/5.0/config/tokenizer.xml new file mode 100644 index 000000000..4181e30ff --- /dev/null +++ b/src/site/en/xdoc/5.0/config/tokenizer.xml @@ -0,0 +1,36 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search.

    +

    Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits.

    +

    For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit.

    +

    For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default CJKTokenizer used.

    +
    + + +

    Such as CJKTokenizer Japan Japanese multibyte string against bi-gram, in other words two characters create index. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to.

    +

    The following example to change schema.xml so analyzer parts, you can use the StandardTokenizer.

    + + + + + : +]]> +
    + + +
    diff --git a/src/site/en/xdoc/5.0/config/windows-service.xml b/src/site/en/xdoc/5.0/config/windows-service.xml new file mode 100644 index 000000000..16ce65f50 --- /dev/null +++ b/src/site/en/xdoc/5.0/config/windows-service.xml @@ -0,0 +1,54 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as Add %JAVA_HOME%\bin to PathYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option.

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-5.0.0.

    + cd C:\Java\fess-server-5.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat6w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    +
    + +

    Distributed in the Fess is 32-bit binaries for Windows Tomcat builds based on. If you use 64-bit WindowsTomcat Of the site from, such as 64-bit Windows zip, please replace tomcat6,exe, tomcat6w,exe.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/index.xml b/src/site/en/xdoc/5.0/user/index.xml new file mode 100644 index 000000000..acf289471 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/index.xml @@ -0,0 +1,12 @@ + + + + Search Guide + Shinsuke Sugaya + + +
    +

    Here is a description of Fess 5.0 search method.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/json-response.xml b/src/site/en/xdoc/5.0/user/json-response.xml new file mode 100644 index 000000000..645fb51d7 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/json-response.xml @@ -0,0 +1,182 @@ + + + + Search results for JSON output + Shinsuke Sugaya + + +
    +

    Outputs a JSON results of Fess. You must in order to emit JSON enabled keep JSON response in the General Management screen crawl settings.

    + +

    To get JSON output http://localhost:8080/Fess/JSON? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + + + + +
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    labelTypeValueThe label value. Use to specify the label.
    callbackIf you use the JSONP callback name. You do not need to specify if you do not use the JSONP.
    +
    + +

    Returns the following response.

    + \u30C6\u30B9\u30C8<\u002Fem>\u30DA\u30FC\u30B8\u3067\u3059", + "host": "speedtest.goo.ne.jp", + "lastModified": "1284739487873", + "cache": "\u9FA0-->\n\n\ngoo \u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\n<\u002Ftitle>\n<meta...\u0022>\n\n<meta name=\u0022description\u0022 content=\u0022goo \u306E\u63D0\u4F9B\u3059\u308B\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u30DA\u30FC\u30B8\u3067\u3059\u0022>\n<meta name=\u0022keywords\u0022 content=\u0022ADSL,\u30D6\u30ED\u30FC\u30C9\u30D0\u30F3\u30C9,\u30D5\u30EC...\u30C3\u30C4,BB,\u7121\u7DDALAN,CATV,ISDN,\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u0022>\n<meta name=\u0022Copyright\u0022 content=\u0022Copyright © (C)NTT-X Inc, All...:\u002F\u002Fspeedtest.goo.ne.jp\u002F\u0022><img src=\u0022\u002Fimg\u002Fh1\u002Fh1_76.gif\u0022 width=\u0022129\u0022 height=\u002222\u0022 border=\u00220\u0022 alt=\u0022\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u0022><img src...\u0022 color=\u0022#ffffff\u0022>goo \u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem><\u002Ffont><\u002Fa><\u002Fb><\u002Ftd>\n<!--\u002FNAME SERVICE-->\n<\u002Ftr>\n<\u002Ftable>\n<\u002Ftd>\n<\u002Ftr>\n<tr>\n<td", + "score": "4.98744", + "digest": "goo \u306E\u63D0\u4F9B\u3059\u308B\u30B9\u30D4\u30FC\u30C9<em>\u30C6\u30B9\u30C8<\u002Fem>\u30DA\u30FC\u30B8\u3067\u3059", + "tstamp": "1284739487887", + "url": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F", + "id": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F;type=au,docomo,pc,softbank", + "mimetype": "text\u002Fhtml", + "title": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "contentTitle": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "boost": "1.0", + "contentLength": "17810", + "urlLink": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F" + }, +... + ] + } +} +]]></source> + <p>Each element is as follows.</p> + <table> + <tbody> + <tr> + <th>response</th> + <td>The root element.</td> + </tr> + <tr> + <th>version</th> + <td>Format version.</td> + </tr> + <tr> + <th>status</th> + <td>The status of the response. 0, the normal response.</td> + </tr> + <tr> + <th>query</th> + <td>Search language.</td> + </tr> + <tr> + <th>execTime</th> + <td>Response time. Unit is seconds.</td> + </tr> + <tr> + <th>pageSize</th> + <td>The number of display.</td> + </tr> + <tr> + <th>pageNumber</th> + <td>Page number.</td> + </tr> + <tr> + <th>recordCount</th> + <td>Number of hits for the search term.</td> + </tr> + <tr> + <th>pageCount</th> + <td>Page number of the number of hits for the search term.</td> + </tr> + <tr> + <th>result</th> + <td>Parent element of the search results.</td> + </tr> + <tr> + <th>site</th> + <td>The name of the site.</td> + </tr> + <tr> + <th>contentDescription</th> + <td>A description of the content.</td> + </tr> + <tr> + <th>host</th> + <td>The name of the host.</td> + </tr> + <tr> + <th>lastModified</th> + <td>Last update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.</td> + </tr> + <tr> + <th>cache</th> + <td>The contents of the content.</td> + </tr> + <tr> + <th>score</th> + <td>Score values of the document.</td> + </tr> + <tr> + <th>Digest</th> + <td>Digest a string of documents.</td> + </tr> + <tr> + <th>Tstamp</th> + <td>Document creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.</td> + </tr> + <tr> + <th>URL</th> + <td>The URL of the document.</td> + </tr> + <tr> + <th>ID</th> + <td>The ID of the document.</td> + </tr> + <tr> + <th>mimetype</th> + <td>The MIME type.</td> + </tr> + <tr> + <th>title</th> + <td>The title of the document.</td> + </tr> + <tr> + <th>contentTitle</th> + <td>The title of the document for viewing.</td> + </tr> + <tr> + <th>contentLength</th> + <td>The size of the document.</td> + </tr> + <tr> + <th>urlLink</th> + <td>The URL as the search results.</td> + </tr> + </tbody> + </table> + </subsection> + </section> + </body> +</document> diff --git a/src/site/en/xdoc/5.0/user/search-boost.xml b/src/site/en/xdoc/5.0/user/search-boost.xml new file mode 100644 index 000000000..8661056d2 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-boost.xml @@ -0,0 +1,15 @@ +<?xml version='1.0' encoding='UTF-8'?> +<document> + <properties> + <title>Boost search + Shinsuke Sugaya + + +
    +

    + Use the boost search if you want to prioritize, search for specific search terms. + For example, if you want to find the page if you want to find apples oranges contained more 'apples' ' Apple ^ 100 orange ' that to ' ^ number "that searches in the form. + Number specifies an integer greater than 1.

    +
    + + diff --git a/src/site/en/xdoc/5.0/user/search-field.xml b/src/site/en/xdoc/5.0/user/search-field.xml new file mode 100644 index 000000000..b30448189 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-field.xml @@ -0,0 +1,57 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    In the Fess crawl results saved in the title and text fields. You can search for a field of them.

    +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields title and content to search for.

    + +

    If a field search "field name: search terms ' of so fill out the search form, the search.

    +

    Title against Fess the search as a search term.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/search-fuzzy.xml b/src/site/en/xdoc/5.0/user/search-fuzzy.xml new file mode 100644 index 000000000..601376b79 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-fuzzy.xml @@ -0,0 +1,18 @@ + + + + Fuzzy search (fuzzy search) + Shinsuke Sugaya + + +
    +

    Fess is corresponding to fuzzy searches based on the Levenshtein distance (fuzzy search). + After the search word you want to apply the fuzzy search adds '~'. + For example, see Solr ~ ' that you can search for documents that contain the search string "Solr" similar to the language or ("Solar").

    +

    + Furthermore, you can specify a number between 0 and 1 ~ after the close to 1 just like in refine. + For example, in the form of 'Solr~0.8'. + Do not specify numeric default value is 0.5.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/search-label.xml b/src/site/en/xdoc/5.0/user/search-label.xml new file mode 100644 index 000000000..b1d24bb85 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-label.xml @@ -0,0 +1,14 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    By label to be registered in the management screen will enable search by labels in the search screen. You can use the label if you want to sort the search results. If you do not register the label displayed the label drop-down box.

    + Search by label +

    To set the label by creating indexes, can search each crawl settings specified on the label. All results search search do not specify a label is usually the same.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/search-not.xml b/src/site/en/xdoc/5.0/user/search-not.xml new file mode 100644 index 000000000..cc733646b --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-not.xml @@ -0,0 +1,15 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If you want to find documents that do not contain a Word can NOT find. + Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, searches, enter if you want to find documents that contain the search term 1 does not contain a search term 2 search term 1 NOT search words 2.

    +

    Attention is required because NOT find expensive.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/search-or.xml b/src/site/en/xdoc/5.0/user/search-or.xml new file mode 100644 index 000000000..acb8df0bc --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-or.xml @@ -0,0 +1,15 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. + When describing the multiple words in the search box, by default will search. + You want OR search the case describes OR between search words. OR write in capital letters, spaces are required before and after.

    +

    For example, the search, enter if you want to search for documents that contain either search term 2 search term 1 search term 1 OR search term 2. OR between multiple languages are available.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/search-range.xml b/src/site/en/xdoc/5.0/user/search-range.xml new file mode 100644 index 000000000..fb9c95595 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-range.xml @@ -0,0 +1,14 @@ + + + + Range searches + Shinsuke Sugaya + + +
    +

    Range searches can be done for field. + To range search field name: value TO the search term. + For example, if you search documents contentLength field against 1 k to 10 k bytes is ' contentLength: 1000 TO 10000 ' search language and the.

    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/search-sort.xml b/src/site/en/xdoc/5.0/user/search-sort.xml new file mode 100644 index 000000000..bceb8e612 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/search-sort.xml @@ -0,0 +1,44 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + +
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    + +

    If you want to sort ' sort: field name ' in to fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/5.0/user/xml-response.xml b/src/site/en/xdoc/5.0/user/xml-response.xml new file mode 100644 index 000000000..619938da5 --- /dev/null +++ b/src/site/en/xdoc/5.0/user/xml-response.xml @@ -0,0 +1,201 @@ + + + + XML results output + Shinsuke Sugaya + + +
    +

    To emit the XML results of Fess. You must in order to emit the XML Management screen crawl settings XML response enabled by you.

    + +

    To get the output XML http://localhost:8080/Fess/XML? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + +
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    labelTypeValueThe label value. Use to specify the label.
    +
    + +

    Returns the following response.

    + + + 1 + 0 + テスト + 0.64 + 20 + 1 + 101 + 6 + + + speedtest.goo.ne.jp/ + goo の提供するスピード<em>テスト</em>ページです + speedtest.goo.ne.jp + 1284739487873 + 龠--> +<meta http-equiv="Content-Type" content="text/html; charset=EUC-JP"> +<title> +goo スピード<em>テスト</em> +</title> +<meta..."> + +<meta name="description" content="goo の提供するスピード<em>テスト</em>ページです"> +<meta name="keywords" content="ADSL,ブロードバンド,フレ...ッツ,BB,無線LAN,CATV,ISDN,スピード<em>テスト</em>"> +<meta name="Copyright" content="Copyright &copy; (C)NTT-X Inc, All...://speedtest.goo.ne.jp/"><img src="/img/h1/h1_76.gif" width="129" height="22" border="0" alt="スピード<em>テスト</em>"><img src..." color="#ffffff">goo スピード<em>テスト</em></font></a></b></td> +<!--/NAME SERVICE--> +</tr> +</table> +</td> +</tr> +<tr> +<td + 4.98744 + goo の提供するスピード<em>テスト</em>ページです + 1284739487887 + http://speedtest.goo.ne.jp/ + http://speedtest.goo.ne.jp/;type=au,docomo,pc,softbank + text/html + +goo スピードテスト + + +goo スピードテスト + + 1.0 + 17810 + http://speedtest.goo.ne.jp/ + +... + + +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    responseThe root element.
    versionFormat version.
    statusThe status of the response. 0, the normal response.
    querySearch language.
    exec-timeResponse time. Unit is seconds.
    page-sizeThe number of display.
    page-numberPage number.
    record-countNumber of hits for the search term.
    page-countPage number of the number of hits for the search term.
    resultParent element of the search results.
    docThe elements of the search results.
    siteThe name of the site.
    content-descriptionA description of the content.
    hostThe name of the host.
    last-modifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    content-titleThe title of the document for viewing.
    content-lengthThe size of the document.
    URL-linkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/browserType-guide.xml b/src/site/en/xdoc/6.0/admin/browserType-guide.xml new file mode 100644 index 000000000..2398086c0 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/browserType-guide.xml @@ -0,0 +1,23 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
    +

    Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

    +
    +
    + +

    In Administrator account after logging in, click menu browser types.

    + Setting the browser type +
    +
    +
    + +

    You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/crawl-guide.xml b/src/site/en/xdoc/6.0/admin/crawl-guide.xml new file mode 100644 index 000000000..33c91309b --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/crawl-guide.xml @@ -0,0 +1,143 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to crawling.

    +
    +
    + +

    In Administrator account click crawl General menu after login.

    + Crawl General +

    You can specify the path to a generated index and replication capabilities to enable.

    + Replication features +
    +
    +
    + +

    You can set the interval at which the crawl for a Web site or file system. By default, the following.

    + +

    Figures are from left, seconds, minutes, during the day, month, represents a day of the week. Description format is similar to the Unix cron settings. This example, and am 0 時 0 分 to crawling daily.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    0 0 12 * *?Each day starts at 12 pm
    0 15 10? * *Day 10: 15 am start
    0 15 10 * *?Day 10: 15 am start
    0 15 10 * *? *Day 10: 15 am start
    0 15 10 * *? 2009Each of the 2009 start am, 10:15
    0 * 14 * *?Every day 2:00 in the PM-2: 59 pm start every 1 minute
    0 0 / 5 14 * *?Every day 2:00 in the PM-2: 59 pm start every 5 minutes
    0 0 / 5 14, 18 * *?Every day 2:00 pm-2: 59 pm and 6: 00 starts every 5 minutes at the PM-6: 59 pm
    0 0-5 14 * *?Every day 2:00 in the PM-2: 05 pm start every 1 minute
    0 10, 44 14? 3 WEDStarts Wednesday March 2: 10 and 2: 44 pm
    0 15 10? * MON-FRIMonday through Friday at 10:15 am start
    +

    Also check if the seconds can be set to run at intervals 60 seconds by default. If you set seconds exactly and you should customize webapps/fess/WEB-INF/classes/chronosCustomize.dicon taskScanIntervalTime value, if enough do I see in one-hour increments.

    +
    + +

    When the user enters a search, the search the output log. If you want to get search statistics to enable.

    +
    + +

    Search results link attaches to the search term. To display the find search terms in PDF becomes possible.

    +
    + +

    Search results can be retrieved in XML format. http://localhost:8080/Fess/XML? can get access query = search term.

    +
    + +

    Search results available in JSON format. http://localhost:8080/Fess/JSON? can get access query = search term.

    +
    + +

    If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

    +
    + +

    You can specify the label to see if the label by default,. Specifies the value of the label.

    +
    + +

    You can specify whether or not to display a search screen. If you select Web unusable for mobile search screen. If not available not available search screen. And if you want to create a dedicated index server and select not available.

    +
    + +

    In JSON format often find search words becomes available. can be retrieved by accessing the http://localhost:8080/Fess/hotsearchword.

    +
    + +

    Delete a session log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Delete a search log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Specifies the Bots name Bots you want to remove from the search log logs included in the user agent by commas (,). Log is deleted by log purge once a day.

    +
    + +

    Specifies the encoding for the CSV will be available in the backup and restore.

    +
    + +

    To enable replication features that can apply already copied the Solr index generated. For example, you can use them if you want to search only in the search servers crawled and indexed on a different server, placed in front.

    +
    + +

    After the data is registered for Solr. Index to commit or to optimize the registered data becomes available. If optimize is issued the Solr index optimization, if you have chosen, you choose to commit the commit is issued.

    +
    + +

    Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

    +
    + +

    To raise the performance of the index in Fess while crawling and sends for Solr document in 20 units. For each value specified here because without committing to continue adding documents documents added in the Solr on performance, Solr issued document commits. By default, after you add documents 1000 is committed.

    +
    + +

    Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

    +

    But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

    +
    + +

    You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

    +
    + +

    Registered disabled URL URL exceeds the failure count next time you crawl to crawl out. No need to worry about disability type is crawled next time by specifying this value.

    +
    + +

    Disaster URL exceeds the number of failures will crawl out.

    +
    + +

    Copy index information from the index directory as the snapshot path, if replication is enabled, will be applied.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/6.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..316145428 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/crawlingSession-guide.xml @@ -0,0 +1,38 @@ + + + + Set session information + Shinsuke Sugaya + + +
    +

    Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

    +
    +
    + +

    In Administrator account after logging in, click the session information menu.

    +
    +
    +
    + + Session information list +

    You can remove all session information and click the Delete link all in the running.

    +
    + + Session details +

    To specify a session ID, you can see crawling content.

    +
      +
    • Information about the entire crawl Cralwer *:
    • +
    • FsCrawl *: information about the file system crawling
    • +
    • WebCrawl *: crawling the Web information
    • +
    • Information issued by Solr server optimization optimize *:
    • +
    • Commit *: information about the commit was issued to the Solr server.
    • +
    • * StartTime: start time
    • +
    • * EndTime: end time
    • +
    • * ExecTime: execution time (MS)
    • +
    • * IndexSize: number of documents indexed
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/data-guide.xml b/src/site/en/xdoc/6.0/admin/data-guide.xml new file mode 100644 index 000000000..cc73d3646 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/data-guide.xml @@ -0,0 +1,35 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
    +

    Here, describes Fess information backup and restore methods.

    +
    +
    + +

    In Administrator account after logging in, click the menu backup and restore.

    + Backup and restore +
    + +

    Click the download link and Fess information output in XML format. Saved settings information is below.

    +
      +
    • The General crawl settings
    • +
    • Web crawl settings
    • +
    • File system Crawl settings
    • +
    • Path mapping
    • +
    • Web authentication
    • +
    • Compatible browsers
    • +
    +

    Session information, search log, click log is available in CSV format.

    +

    In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate.

    +
    + +

    You can restore settings information, various log in to upload XML output by backup or CSV. To specify the files, please click the restore button on the data.

    +

    If enable overwrite data in XML file configuration information specified when the same data is updating existing data.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/dataCrawlingConfig-guide.xml b/src/site/en/xdoc/6.0/admin/dataCrawlingConfig-guide.xml new file mode 100644 index 000000000..5acee69c0 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/dataCrawlingConfig-guide.xml @@ -0,0 +1,157 @@ + + + + Settings for crawling the data store + Sone, Takaaki + Shinsuke Sugaya + + +
    +

    You can crawl databases in Fess. Here are required to store settings.

    +
    +
    + +

    In Administrator account after logging in, click menu data store.

    + Data store configuration +

    As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

    + +

    Here the data is put something like the following.

    + +
    +
    +
    + +

    Parameter settings example looks like the following.

    + +

    Parameter is a "key = value" format. Description of the key is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + +
    driverDriver class name
    URLURL
    usernameTo connect to the DB user name
    passwordTo connect to the DB password
    SQLWant to crawl to get SQL statement
    +
    + +

    Script configuration example looks like the following.

    + +

    + Parameter is a "key = value" format. + Description of the key is as follows.

    +

    + Side of the value written in OGNL. Close the string in double quotation marks. + Access in the database column name, its value.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    URLURLs (links appear in search results)
    hostHost name
    siteSite pass
    titleTitle
    contentContent (string index)
    cacheContent cache (not indexed)
    DigestDigest piece that appears in the search results
    anchorLinks to content (not usually required)
    contentLengthThe length of the content
    lastModifiedContent last updated
    +
    + +

    To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

    +
    + +

    Set the following in the webapps/fess/WEB-INF/classes/app.dicon if you see the item value, such as latitude_s in the search results. After adding to $ {doc.latitude_s}, searchResults.jsp;

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "latitude_s","longitude_s" } + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/design-guide.xml b/src/site/en/xdoc/6.0/admin/design-guide.xml new file mode 100644 index 000000000..7e5d8ce69 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/design-guide.xml @@ -0,0 +1,103 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
    +

    Here are settings for the design of search screens.

    +
    +
    + +

    In Administrator account after logging in, click the menu design.

    + Design +

    You can edit the search screen in the screen below.

    + JSP compilation screen +
    + +

    If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

    + + +<% +java.util.Map docMap = (java.util.Map)pageContext.getAttribute("doc"); +Long tstampValue = (Long)docMap.get("tstamp"); +java.util.Date tstampDate = new java.util.Date(tstampValue); +Long lastModifiedValue = (Long)docMap.get("lastModified"); +java.util.Date lastModifiedDate = new java.util.Date(lastModifiedValue); +java.text.SimpleDateFormat sdf = new java.text.SimpleDateFormat("yyyy/MM/dd HH:mm"); +%> +
  • +

    + ${f:h(doc.contentTitle)} +

    +
    +${doc.contentDescription} +
    + ${f:h(doc.site)} +
    Registered: <%= sdf.format(tstampDate) %> +
    Last Modified: <%= sdf.format(lastModifiedDate) %> + : + : +]]> +

    tstampDate will update on registration date, lastModifiedDate. Output date format is specified in SimpeDateFormat.

    + +
  • +
    + +

    You can upload the image files to use in the search screen. Image file names are supported are jpg, gif and png.

    +
    + +

    If you want the file name to upload image files to use. Uploaded if you omit the file name will be used.

    +
    + +

    You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
    Top page (within the Head tags)This is the express search home page head tag in JSP files. If you want to edit the meta tags, title tags, script tags, such as the change.
    Top page (content)Is a JSP file to represent the body tag in the search home page.
    Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
    Search results page (within the Head tags)Search result is a JSP file to represent within the head tag of the list page. If you want to edit the meta tags, title tags, script tags, such as the change.
    Search results page (header)Search result is a JSP file to represent the header of the list page. Include search form at the top.
    Search results page (footer)Search result is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
    Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    +

    You can to edit for PCs and similar portable screen.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/failureUrl-guide.xml b/src/site/en/xdoc/6.0/admin/failureUrl-guide.xml new file mode 100644 index 000000000..abc7dffcd --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/failureUrl-guide.xml @@ -0,0 +1,23 @@ + + + + Disaster URL + Shinsuke Sugaya + + +
    +

    Here the failure URL. URL could not be obtained at crawl time are recorded and confirmed as the failure URL.

    +
    +
    + +

    In Administrator account click menu disabled URL after login.

    + Disaster URL +

    Clicking the confirmation link failure URL displayed for more information.

    + Details of the disaster URL +
    + +

    A glance could not crawl the URL and date.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/fileAuthentication-guide.xml b/src/site/en/xdoc/6.0/admin/fileAuthentication-guide.xml new file mode 100644 index 000000000..747e6bf84 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/fileAuthentication-guide.xml @@ -0,0 +1,44 @@ + + + + Settings for file system authentication + Shinsuke Sugaya + + +
    +

    Crawls using file system here, describes how to set file system authentication is required. Fess is corresponding to a crawl for a shared folder in Windows.

    +
    +
    + +

    In Administrator account after logging in, click the menu file system authentication.

    + File system settings +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Is omitted, the specified file system Kroll set applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. File system Crawl settings specified in that case applies on any port.

    +
    + +

    Select the authentication method. You can use SAMBA (Windows shared folder authentication).

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. SAMBA, the set value of the domain. If you want to write as.

    + +
    + +

    Select a file name to apply the authentication settings for the above. Must be registered ago you file system CROLL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/6.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..054df2c7e --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,102 @@ + + + + Settings for file system crawling + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: or smb: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +

    Windows shared folder, for example, if you want to crawl to host1 share folder crawl settings for smb: (last / to) the //host1/share/. If authentication is in the shared folder on the file system authentication screen set authentication information.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + +
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/index.xml b/src/site/en/xdoc/6.0/admin/index.xml new file mode 100644 index 000000000..45bc80f38 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/index.xml @@ -0,0 +1,67 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for the management work of the Fess. +

    +
    +
    +

    +This document demonstrates Fess configuration management methods. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    +
    + +

    Management UI to support browser as follows.

    +
      +
    • Internet Explorer 7 or higher
    • +
    • Firefox 4 or more
    • +
    • Chrome (latest version)
    • +
    • Safari (latest version)
    • +
    • Opera 11
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/labelType-guide.xml b/src/site/en/xdoc/6.0/admin/labelType-guide.xml new file mode 100644 index 000000000..5daaf9c28 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/labelType-guide.xml @@ -0,0 +1,33 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. If you register the label shown select label drop-down box to the right of the search box.

    +
    +
    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    +
    +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    + +

    Specifies the role to view the label.

    +
    + +

    Specifies the order of the labels.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/log-guide.xml b/src/site/en/xdoc/6.0/admin/log-guide.xml new file mode 100644 index 000000000..f273e54b9 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/log-guide.xml @@ -0,0 +1,21 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    +
    +
    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/6.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..f1aa67781 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/overlappingHost-guide.xml @@ -0,0 +1,27 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    +
    +
    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    +
    +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/6.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..8568bd7ed --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/pathMapping-guide.xml @@ -0,0 +1,30 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    +
    +
    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    +
    +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/6.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..17736eadf --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/requestHeader-guide.xml @@ -0,0 +1,30 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    +
    +
    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    +
    +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/roleType-guide.xml b/src/site/en/xdoc/6.0/admin/roleType-guide.xml new file mode 100644 index 000000000..2b63648c6 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/roleType-guide.xml @@ -0,0 +1,27 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    +
    +
    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    +
    +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/searchList-guide.xml b/src/site/en/xdoc/6.0/admin/searchList-guide.xml new file mode 100644 index 000000000..32fb21140 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/searchList-guide.xml @@ -0,0 +1,21 @@ + + + + Search + Shinsuke Sugaya + + +
    +

    Here the search for management.

    +
    +
    + +

    In Administrator account after logging in, click the menu search.

    + Administrative search +
    + +

    You can search by criteria you specify. In the regular search screen role and browser requirements is added implicitly, but do not provide management for search. You can document a certain remove from index from the search results.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/searchLog-guide.xml b/src/site/en/xdoc/6.0/admin/searchLog-guide.xml new file mode 100644 index 000000000..9bf0f8784 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/searchLog-guide.xml @@ -0,0 +1,21 @@ + + + + Search log settings + Shinsuke Sugaya + + +
    +

    Here the search log. When you search in the search screen users search logs are logged. Search log search term or date is recorded. You can also record the URL, then you want the search results to.

    +
    +
    + +

    In Administrator account after logging in, click menu search logs.

    + Search log +
    + +

    Search language and date are listed. You can review and detailed, you click the URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/stats-guide.xml b/src/site/en/xdoc/6.0/admin/stats-guide.xml new file mode 100644 index 000000000..1e0e9f85a --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/stats-guide.xml @@ -0,0 +1,21 @@ + + + + Statistics + Shinsuke Sugaya + + +
    +

    Here are statistics. You can search log and click log.

    +
    +
    + +

    In Administrator account after logging in, click the menu statistics.

    + Statistics +
    + +

    You can select the target by selecting the type of report, to make sure. Displayed in order by the specified criteria.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/system-guide.xml b/src/site/en/xdoc/6.0/admin/system-guide.xml new file mode 100644 index 000000000..323e913a4 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/system-guide.xml @@ -0,0 +1,35 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here registration in Fess. SOLR servers are grouped by file, has been registered.

    +
    +
    + +

    In Administrator account after logging in, click menu Solr.

    + System settings +
    +
    +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and shut down when not running Fess server to shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +
    + +

    Server group name is used to search for and update appears.

    +
    + +

    Server becomes unavailable and the status of disabled. For example, inaccessible to the Solr server and changes to disabled. To enable recovery after server become unavailable will become available.

    +
    + +

    You can publish index commit, optimize for server groups. You can also remove a specific search for the session ID. You can remove only the specific documents by specifying the URL.

    +
    + +

    Shown by the number of documents registered in each session. Can verify the results list by clicking the session name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/systemInfo-guide.xml b/src/site/en/xdoc/6.0/admin/systemInfo-guide.xml new file mode 100644 index 000000000..e00d6d2ef --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/systemInfo-guide.xml @@ -0,0 +1,32 @@ + + + + System information + Shinsuke Sugaya + + +
    +

    Here, you can currently check property information such as system environment variables.

    +
    +
    + +

    In Administrator account after logging in, click system information menu.

    + System information +
    +
    +
    + +

    You can list the server environment variable.

    +
    + +

    You can list the system properties on Fess.

    +
    + +

    Fess setup information available.

    +
    + +

    Is a list of properties to attach when reporting a bug. Extract the value contains no personal information.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/6.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..67af4a798 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/webAuthentication-guide.xml @@ -0,0 +1,48 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    +
    +
    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication, DIGEST authentication or NTLM authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. You can set the workstation and domain values for NTLM authentication. If you want to write as.

    + +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/6.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..ad4b96254 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,103 @@ + + + + Settings for crawling Web site + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + +
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl. If you do not specify people per 100,000.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in)

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/admin/wizard-guide.xml b/src/site/en/xdoc/6.0/admin/wizard-guide.xml new file mode 100644 index 000000000..8a86ebf81 --- /dev/null +++ b/src/site/en/xdoc/6.0/admin/wizard-guide.xml @@ -0,0 +1,33 @@ + + + + Settings Wizard + Sone, Takaaki + + +
    +

    Introduction to the Configuration Wizard.

    +

    You can use Settings Wizard, to set you up on the fess.

    +
    +
    + +

    In Administrator account after logging in, click menu Settings Wizard.

    + Settings Wizard +

    First, setting a schedule. + During the time in fess is crawling and indexes. By default, every day is a 0 時 0 分.

    + Setting a schedule +

    The crawl settings. + Crawl settings is to register a URI to look for. + The crawl settings name please put name of any easy to identify. Put the URI part de-indexed, want to search for.

    + Crawl settings +

    For example, if you want and search for http://fess.codelibs.org/, less looks like.

    + Crawl settings example +

    In this is the last setting. Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

    + Crawl started +
    + +

    Settings in the Setup Wizard you can change from crawl General, Web, file system.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/db.xml b/src/site/en/xdoc/6.0/config/db.xml new file mode 100644 index 000000000..8c9bac729 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/db.xml @@ -0,0 +1,118 @@ + + + + Database changes + Shinsuke Sugaya + + +
    +

    Under normal circumstances the database use the H2 Database. You can use other databases by changing settings.

    +
    +
    + +

    Expand the MySQL binaries.

    +
    + +

    Create a database.

    + create database fess_db; +mysql> grant all privileges on fess_db.* to fess_user@localhost identified by 'fess_pass'; +mysql> create database fess_robot; +mysql> grant all privileges on fess_robot.* to s2robot@localhost identified by 's2robot'; +mysql> FLUSH PRIVILEGES; +]]> +

    Create a table in the database. DDL file is located in extension/mysql.

    + +

    webapps/fess/WEB-INF/lib on mysql driver jar.

    + +

    Edit the webapps/fess/WEB-INF/classes/jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_db?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "fess_user" + "fess_pass" + + + + 600 + 10 + true + + + + + +]]> +

    Edit the webapps/fess/WEB-INF/classes/s2robot_jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_robot?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "s2robot" + "s2robot" + + + + 600 + 10 + true + @java.sql.Connection@TRANSACTION_REPEATABLE_READ + + + + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/desktop-search.xml b/src/site/en/xdoc/6.0/config/desktop-search.xml new file mode 100644 index 000000000..029af0eb3 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/desktop-search.xml @@ -0,0 +1,35 @@ + + + + The desktop search settings + Shinsuke Sugaya + + +
    +

    + Increasing awareness of security in the browser environment in recent years, open a local file (for example, c:\hoge.txt) from the Web pages on. + Not to copy and paste the link from the search results, and then reopen the usability is good. + In order to respond to this in Fess and provides desktop search functionality.

    +
    +
    +

    + Desktop Search feature is turned off by default. + Please enable the following settings.

    +

    First of all, bin/setenv.bat as java.awt.headless from true to false edits.

    + +

    Then add the following to webapps/fess/WEB-INF/conf/crawler.properties.

    + +

    Start the Fess, after you set up above. How to use Basic remains especially.

    +
    +
    +
      +
    • Please Fess inaccessible from the outside, such as (for example, 8080 port does not release).
    • +
    • because false Java.awt.headless image size conversion for mobile devices is not available.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/filesize.xml b/src/site/en/xdoc/6.0/config/filesize.xml new file mode 100644 index 000000000..ba51e38d6 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/filesize.xml @@ -0,0 +1,28 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of defaultMaxLength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/geosearch.xml b/src/site/en/xdoc/6.0/config/geosearch.xml new file mode 100644 index 000000000..a45779216 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/geosearch.xml @@ -0,0 +1,23 @@ + + + + GEO search + Shinsuke Sugaya + + +
    +

    Together with Google maps, including document with latitude and longitude location information, GEO (GEO) you can use the search.

    +
    +
    + +

    Location is defined as a feed that contains the location information. + When generating the index in Solr latitude longitude set to location feeds in formats such as 45.17614,-93.87341, register the document. + Also sets the value as the latitude_s and longitude_s fields if you want to display latitude and longitude as a search result. * _s is available as a dynamic field of Solr string.

    +
    + +

    During the search specifies in the request parameter to latitude and longitude, the distance. + View the results in the distance (km) specified by distance-based latitude information (latitude, longitude). Latitude and longitude and distances is treated as double.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/index-backup.xml b/src/site/en/xdoc/6.0/config/index-backup.xml new file mode 100644 index 000000000..541993b34 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index management + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data directory. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/index.xml b/src/site/en/xdoc/6.0/config/index.xml new file mode 100644 index 000000000..d621553e0 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for setting work of the Fess. +

    +
    +
    +

    +This document demonstrates setting Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN9sm, Inc.To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/install-on-tomcat.xml b/src/site/en/xdoc/6.0/config/install-on-tomcat.xml new file mode 100644 index 000000000..314d28334 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/install-on-tomcat.xml @@ -0,0 +1,43 @@ + + + + Install to an existing Tomcat + Shinsuke Sugaya + + +
    +

    + The standard distribution of Fess Tomcat is distributed in the deployed State. + Because Fess is not dependent on Tomcat, deploying on any Java application server is available. + Describes how to deploy a Fess Tomcat here is already available.

    +
    +
    +

    Fess ServerHereFrom the download.

    +
    +
    +

    + Expand the downloaded Fess server. + Expanded Fess Server home directory to $FESS_HOME. + $TOMCAT_HOME the top directory of an existing Tomcat 6. + Copy the Fess Server data.

    + +

    + If you have, such as changing the destination file diff commands, updates your diff only applies. +

    +
    +
    +

    to start the Tomcat in normal as well as in startup.* / http://localhost:8080/Fess To access the.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/logging.xml b/src/site/en/xdoc/6.0/config/logging.xml new file mode 100644 index 000000000..306bdd0cd --- /dev/null +++ b/src/site/en/xdoc/6.0/config/logging.xml @@ -0,0 +1,43 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + +
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina.outLog of the Fess Server (Tomcat). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/memory-config.xml b/src/site/en/xdoc/6.0/config/memory-config.xml new file mode 100644 index 000000000..7ca6c9ace --- /dev/null +++ b/src/site/en/xdoc/6.0/config/memory-config.xml @@ -0,0 +1,49 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case the maximum value set 1024M) will change to-Xmx1024m.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-Xmx1024m (in this case the maximum value set 1024M). +

    + new String[] { +"-Djava.awt.headless=true", +"-server", +"-Xmx1024m", +"-XX:MaxPermSize=128m", +"-XX:-UseGCOverheadLimit", +"-XX:+UseConcMarkSweepGC", +"-XX:CMSInitiatingOccupancyFraction=75", +"-XX:+CMSIncrementalMode", +"-XX:+CMSIncrementalPacing", +"-XX:CMSIncrementalDutyCycleMin=0", +"-XX:+UseParNewGC", +"-XX:+UseStringCache", +"-XX:+UseTLAB", +"-XX:+DisableExplicitGC" +} +]]> +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/mobile-device.xml b/src/site/en/xdoc/6.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/path-encoding.xml b/src/site/en/xdoc/6.0/config/path-encoding.xml new file mode 100644 index 000000000..b8aa16ac6 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    +
    +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/6.0/config/pdf-with-password.xml b/src/site/en/xdoc/6.0/config/pdf-with-password.xml new file mode 100644 index 000000000..268f01d99 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/pdf-with-password.xml @@ -0,0 +1,256 @@ + + + + Password-protected PDF + Shinsuke Sugaya + + +
    +

    + You should password files to register the settings file to PDF password is configured to search for. +

    +
    +
    +

    + First of all, create the webapps/fess/WEB-INF/classes/s2robot_extractor.dicon. + This is test _ ~ is a pass that password set to a.pdf file. + If you have multiple files, multiple settings in addPassword.

    + + + + + + + + + + + + + ".*test_.*.pdf" + + "pass" + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + { +"application/xml", +"application/xhtml+xml", +"application/rdf+xml", +"text/xml", +"text/xml-external-parsed-entity" + } + xmlExtractor + + + { +"text/html" + } + xmlExtractor + + + { +"application/pdf" + } + pdfExtractor + + + { +"image/svg+xml", +"application/x-tika-msoffice", +"application/vnd.visio", +"application/vnd.ms-powerpoint", +"application/vnd.ms-excel", +"application/vnd.ms-excel.sheet.binary.macroenabled.12", +"application/msword", +"application/vnd.ms-outlook", +"application/x-tika-ooxml", +"application/vnd.openxmlformats-officedocument.presentationml.presentation", +"application/vnd.ms-powerpoint.presentation.macroenabled.12", +"application/vnd.openxmlformats-officedocument.presentationml.template", +"application/vnd.openxmlformats-officedocument.presentationml.slideshow", +"application/vnd.ms-powerpoint.slideshow.macroenabled.12", +"application/vnd.ms-powerpoint.addin.macroenabled.12", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.ms-excel.sheet.macroenabled.12", +"application/vnd.openxmlformats-officedocument.spreadsheetml.template", +"application/vnd.ms-excel.template.macroenabled.12", +"application/vnd.ms-excel.addin.macroenabled.12", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.ms-word.document.macroenabled.12", +"application/vnd.openxmlformats-officedocument.wordprocessingml.template", +"application/vnd.ms-word.template.macroenabled.12", +"application/x-asp", +"application/rtf", +"text/plain", +"application/vnd.sun.xml.writer", +"application/vnd.oasis.opendocument.text", +"application/vnd.oasis.opendocument.graphics", +"application/vnd.oasis.opendocument.presentation", +"application/vnd.oasis.opendocument.spreadsheet", +"application/vnd.oasis.opendocument.chart", +"application/vnd.oasis.opendocument.image", +"application/vnd.oasis.opendocument.formula", +"application/vnd.oasis.opendocument.text-master", +"application/vnd.oasis.opendocument.text-web", +"application/vnd.oasis.opendocument.text-template", +"application/vnd.oasis.opendocument.graphics-template", +"application/vnd.oasis.opendocument.presentation-template", +"application/vnd.oasis.opendocument.spreadsheet-template", +"application/vnd.oasis.opendocument.chart-template", +"application/vnd.oasis.opendocument.image-template", +"application/vnd.oasis.opendocument.formula-template", +"application/x-vnd.oasis.opendocument.text", +"application/x-vnd.oasis.opendocument.graphics", +"application/x-vnd.oasis.opendocument.presentation", +"application/x-vnd.oasis.opendocument.spreadsheet", +"application/x-vnd.oasis.opendocument.chart", +"application/x-vnd.oasis.opendocument.image", +"application/x-vnd.oasis.opendocument.formula", +"application/x-vnd.oasis.opendocument.text-master", +"application/x-vnd.oasis.opendocument.text-web", +"application/x-vnd.oasis.opendocument.text-template", +"application/x-vnd.oasis.opendocument.graphics-template", +"application/x-vnd.oasis.opendocument.presentation-template", +"application/x-vnd.oasis.opendocument.spreadsheet-template", +"application/x-vnd.oasis.opendocument.chart-template", +"application/x-vnd.oasis.opendocument.image-template", +"application/x-vnd.oasis.opendocument.formula-template", +"image/bmp", +"image/gif", +"image/jpeg", +"image/png", +"image/tiff", +"image/vnd.wap.wbmp", +"image/x-icon", +"image/x-psd", +"image/x-xcf", +"application/zip", +"application/x-tar", +"application/x-gtar", +"application/x-gzip", +"application/x-bzip", +"application/x-bzip2", +"application/java-vm", +"audio/mpeg", +"application/x-midi", +"audio/midi", +"audio/basic", +"audio/x-wav", +"audio/x-aiff", +"application/mbox", +"text/calendar", +"text/css", +"text/csv", +"text/directory", +"text/dns", +"text/ecmascript", +"text/enriched", +"text/example", +"text/javascript", +"text/parityfec", +"text/prs.fallenstein.rst", +"text/prs.lines.tag", +"text/red", +"text/rfc822-headers", +"text/richtext", +"text/rtf", +"text/rtp-enc-aescm128", +"text/rtx", +"text/sgml", +"text/t140", +"text/tab-separated-values", +"text/troff", +"text/ulpfec", +"text/uri-list", +"text/vnd.abc", +"text/vnd.curl", +"text/vnd.curl.dcurl", +"text/vnd.curl.mcurl", +"text/vnd.curl.scurl", +"text/vnd.dmclientscript", +"text/vnd.esmertec.theme-descriptor", +"text/vnd.fly", +"text/vnd.fmi.flexstor", +"text/vnd.graphviz", +"text/vnd.in3d.3dml", +"text/vnd.in3d.spot", +"text/vnd.iptc.newsml", +"text/vnd.iptc.nitf", +"text/vnd.latex-z", +"text/vnd.motorola.reflex", +"text/vnd.ms-mediapackage", +"text/vnd.net2phone.commcenter.command", +"text/vnd.si.uricatalogue", +"text/vnd.sun.j2me.app-descriptor", +"text/vnd.trolltech.linguist", +"text/vnd.wap.si", +"text/vnd.wap.sl", +"text/vnd.wap.wml", +"text/vnd.wap.wmlscript", +"text/x-asm", +"text/x-c", +"text/x-diff", +"text/x-fortran", +"text/x-java-source", +"text/x-pascal", +"text/x-setext", +"text/x-uuencode", +"text/x-vcalendar", +"text/x-vcard", +"application/x-sh" + } + tikaExtractor + + + + +]]> +

    Next, edit the following to webapps/fess/WEB-INF/classes/s2robot_rule.dicon.

    + + "fsFileRule" + + + fessFileTransformer + + + true + + "url" + "file:.*" + + + "mimeType" + + + "(application/xml" ++ "|application/xhtml+xml" ++ "|application/rdf+xml" ++ "|application/pdf" ++ "|text/xml" ++ "|text/xml-external-parsed-entity" ++ "|text/html)" + + + +... +]]> +

    Fess up after setting the above and please run the crawl. How to use Basic remains especially.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/protwords.xml b/src/site/en/xdoc/6.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/proxy.xml b/src/site/en/xdoc/6.0/config/proxy.xml new file mode 100644 index 000000000..bb13eddbe --- /dev/null +++ b/src/site/en/xdoc/6.0/config/proxy.xml @@ -0,0 +1,95 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    +
    +

    Proxy is set in to create webapps/fess/WEB-INF/classes/s2robot_client.dicon with the following contents.

    + + + + + + + + @org.apache.http.client.params.CookiePolicy@BEST_MATCH + clientConnectionManager + "プロキシサーバー名(ex. 192.168.1.1)" + プロキシサーバーのポート(ex. 8080) (" は不要) + + + + + + + + + + "http" + 80 + @org.apache.http.conn.scheme.PlainSocketFactory@getSocketFactory() + + + + + + + "https" + 443 + @org.apache.http.conn.ssl.SSLSocketFactory@getSocketFactory() + + + + + + 200 + 200 + + + + + "UTF-8" + + + + "UTF-8" + + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + "smb:.*" + smbClient + + + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/6.0/config/replication.xml b/src/site/en/xdoc/6.0/config/replication.xml new file mode 100644 index 000000000..acd703c71 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/replication.xml @@ -0,0 +1,25 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Fess can copy the path in Solr index data. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +

    You must use the replication features of Fess for Solr index file in the shared disk, such as NFS, Fess of each can be referenced from.

    +
    +
    + +

    Fess, download and install the./ /NET/Server1/usr/local/Fess To assume you installed.

    +

    To register the crawl settings as well as Fess starts after the normal construction, create the index (index for Fess building instructions normal building procedures and especially remains the same) crawling.

    +
    + +

    Fess, download and install the./ /NET/Server2/usr/local/Fess To assume you installed.

    +

    To enable replication features check box in Fess starts after the management screen crawl settings the "snapshot path'. Snapshot path designates the index location for the index for Fess. In this case, the/NET/Server1/usr/local/Fess //solr/core1/data/index In the will.

    + Replication +

    Time press the update button to save the data and set in Schedule performs replication of the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/role-setting.xml b/src/site/en/xdoc/6.0/config/role-setting.xml new file mode 100644 index 000000000..91321034a --- /dev/null +++ b/src/site/en/xdoc/6.0/config/role-setting.xml @@ -0,0 +1,97 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    +
      +
    • Request parameter
    • +
    • Request header
    • +
    • Cookies
    • +
    • J2EE authentication information
    • +
    +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/app.dicon shown below.

    + + + {"guest"} + + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + "role1" + + : +]]> +

    authenticatedRoles can describe multiple by commas (,).

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/server-port.xml b/src/site/en/xdoc/6.0/config/server-port.xml new file mode 100644 index 000000000..8438aab1f --- /dev/null +++ b/src/site/en/xdoc/6.0/config/server-port.xml @@ -0,0 +1,31 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    • 19092: database port (use h2database)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information. change the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + "http://localhost:8080/solr" +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/solr-dynamic-field.xml b/src/site/en/xdoc/6.0/config/solr-dynamic-field.xml new file mode 100644 index 000000000..7a58e344e --- /dev/null +++ b/src/site/en/xdoc/6.0/config/solr-dynamic-field.xml @@ -0,0 +1,48 @@ + + + + Dynamic fields + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in SOLR/core1/conf/schema,XML. dynamic fields and standard fields such as title and content can be freely defined field names are defined. The dynamic fields that are available in the schema.xml Fess become. Advanced parameter values see a Solr document.

    + + + + + + + + + + + + + + + + + + + + + + +]]> +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add fields for the following in the dynamic field data out of Solr using webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/solr-failure.xml b/src/site/en/xdoc/6.0/config/solr-failure.xml new file mode 100644 index 000000000..687dc70b6 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    +
    +
    +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is minActiveServer following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status maxRetryStatusCheckCount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    +
    +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    +
    +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/tokenizer.xml b/src/site/en/xdoc/6.0/config/tokenizer.xml new file mode 100644 index 000000000..fcdb1422e --- /dev/null +++ b/src/site/en/xdoc/6.0/config/tokenizer.xml @@ -0,0 +1,47 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search. Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits. For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit. For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default CJKTokenizer used.

    + + +

    Such as CJKTokenizer Japan Japanese multibyte string against bi-gram, in other words two characters create index. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to. However, please note that the index size increases.

    +

    The following example to change the analyzer part like solr/core1/conf/schema.xml, you can use the StandardTokenizer.

    + + + + + + + : + + + + + : +]]> +

    Also, useBigram is enabled by default in the webapps/fess/WEB-INF/classes/app.dicon change to false.

    + + true + : +]]> +

    After the restart the Fess.

    +
    +
    + + +
    diff --git a/src/site/en/xdoc/6.0/config/use-libreoffice.xml b/src/site/en/xdoc/6.0/config/use-libreoffice.xml new file mode 100644 index 000000000..fbd948604 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/use-libreoffice.xml @@ -0,0 +1,240 @@ + + + + Use of LibreOffice + Shinsuke Sugaya + + +
    +

    + It is possible to crawl using the Apache POI Fess environmental standard in MS Office system document. + You can crawl Office system document regarding LibreOffice, OpenOffice, do even more accurate text extraction from documents.

    +
    +
    +

    JodConverter Fess server install. from http://jodconverter.googlecode.com/jodconverter-core-3.0-Beta-4-Dist.zipThe download. Expand and copy the jar file to Fess server.

    + +

    Create a s2robot_extractor.dicon to the next.

    + +

    s2robot_extractor.dicon effective jodExtractor with following contents.

    + + + + + + + + + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + + officeManagerConfiguration.setOfficeHome("/usr/lib/libreoffice").buildOfficeManager() + + + + + + + { +"application/msword", +"application/vnd.ms-excel", +"application/vnd.ms-powerpoint", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.openxmlformats-officedocument.presentationml.presentation" + } + jodExtractor + + + { +"application/xml", +"application/xhtml+xml", +"application/rdf+xml", +"application/x-freemind", +"text/xml", +"text/xml-external-parsed-entity" + } + xmlExtractor + + + { +"text/html" + } + htmlExtractor + + + { +"application/pdf" + } + pdfExtractor + + + { +"image/svg+xml", +"application/x-tika-msoffice", +"application/vnd.visio", +"application/vnd.ms-excel.sheet.binary.macroenabled.12", +"application/vnd.ms-outlook", +"application/x-tika-ooxml", +"application/vnd.ms-powerpoint.presentation.macroenabled.12", +"application/vnd.openxmlformats-officedocument.presentationml.template", +"application/vnd.openxmlformats-officedocument.presentationml.slideshow", +"application/vnd.ms-powerpoint.slideshow.macroenabled.12", +"application/vnd.ms-powerpoint.addin.macroenabled.12", +"application/vnd.ms-excel.sheet.macroenabled.12", +"application/vnd.openxmlformats-officedocument.spreadsheetml.template", +"application/vnd.ms-excel.template.macroenabled.12", +"application/vnd.ms-excel.addin.macroenabled.12", +"application/vnd.ms-word.document.macroenabled.12", +"application/vnd.openxmlformats-officedocument.wordprocessingml.template", +"application/vnd.ms-word.template.macroenabled.12", +"application/x-asp", +"application/rtf", +"text/plain", +"application/vnd.sun.xml.writer", +"application/vnd.oasis.opendocument.text", +"application/vnd.oasis.opendocument.graphics", +"application/vnd.oasis.opendocument.presentation", +"application/vnd.oasis.opendocument.spreadsheet", +"application/vnd.oasis.opendocument.chart", +"application/vnd.oasis.opendocument.image", +"application/vnd.oasis.opendocument.formula", +"application/vnd.oasis.opendocument.text-master", +"application/vnd.oasis.opendocument.text-web", +"application/vnd.oasis.opendocument.text-template", +"application/vnd.oasis.opendocument.graphics-template", +"application/vnd.oasis.opendocument.presentation-template", +"application/vnd.oasis.opendocument.spreadsheet-template", +"application/vnd.oasis.opendocument.chart-template", +"application/vnd.oasis.opendocument.image-template", +"application/vnd.oasis.opendocument.formula-template", +"application/x-vnd.oasis.opendocument.text", +"application/x-vnd.oasis.opendocument.graphics", +"application/x-vnd.oasis.opendocument.presentation", +"application/x-vnd.oasis.opendocument.spreadsheet", +"application/x-vnd.oasis.opendocument.chart", +"application/x-vnd.oasis.opendocument.image", +"application/x-vnd.oasis.opendocument.formula", +"application/x-vnd.oasis.opendocument.text-master", +"application/x-vnd.oasis.opendocument.text-web", +"application/x-vnd.oasis.opendocument.text-template", +"application/x-vnd.oasis.opendocument.graphics-template", +"application/x-vnd.oasis.opendocument.presentation-template", +"application/x-vnd.oasis.opendocument.spreadsheet-template", +"application/x-vnd.oasis.opendocument.chart-template", +"application/x-vnd.oasis.opendocument.image-template", +"application/x-vnd.oasis.opendocument.formula-template", +"image/bmp", +"image/gif", +"image/jpeg", +"image/png", +"image/tiff", +"image/vnd.wap.wbmp", +"image/x-icon", +"image/x-psd", +"image/x-xcf", +"application/zip", +"application/x-tar", +"application/x-gtar", +"application/x-gzip", +"application/x-bzip", +"application/x-bzip2", +"application/java-vm", +"audio/mpeg", +"application/x-midi", +"audio/midi", +"audio/basic", +"audio/x-wav", +"audio/x-aiff", +"application/mbox", +"text/calendar", +"text/css", +"text/csv", +"text/directory", +"text/dns", +"text/ecmascript", +"text/enriched", +"text/example", +"text/javascript", +"text/parityfec", +"text/prs.fallenstein.rst", +"text/prs.lines.tag", +"text/red", +"text/rfc822-headers", +"text/richtext", +"text/rtf", +"text/rtp-enc-aescm128", +"text/rtx", +"text/sgml", +"text/t140", +"text/tab-separated-values", +"text/troff", +"text/ulpfec", +"text/uri-list", +"text/vnd.abc", +"text/vnd.curl", +"text/vnd.curl.dcurl", +"text/vnd.curl.mcurl", +"text/vnd.curl.scurl", +"text/vnd.dmclientscript", +"text/vnd.esmertec.theme-descriptor", +"text/vnd.fly", +"text/vnd.fmi.flexstor", +"text/vnd.graphviz", +"text/vnd.in3d.3dml", +"text/vnd.in3d.spot", +"text/vnd.iptc.newsml", +"text/vnd.iptc.nitf", +"text/vnd.latex-z", +"text/vnd.motorola.reflex", +"text/vnd.ms-mediapackage", +"text/vnd.net2phone.commcenter.command", +"text/vnd.si.uricatalogue", +"text/vnd.sun.j2me.app-descriptor", +"text/vnd.trolltech.linguist", +"text/vnd.wap.si", +"text/vnd.wap.sl", +"text/vnd.wap.wml", +"text/vnd.wap.wmlscript", +"text/x-asm", +"text/x-c", +"text/x-diff", +"text/x-fortran", +"text/x-java-source", +"text/x-pascal", +"text/x-setext", +"text/x-uuencode", +"text/x-vcalendar", +"text/x-vcard", +"application/x-sh" + } + tikaExtractor + + + +]]> +

    Index to generate the settings later, usually crawled into the street.

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/config/windows-service.xml b/src/site/en/xdoc/6.0/config/windows-service.xml new file mode 100644 index 000000000..3b1d40fd0 --- /dev/null +++ b/src/site/en/xdoc/6.0/config/windows-service.xml @@ -0,0 +1,54 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as Add %JAVA_HOME%\bin to PathYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option.

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-6.0.0.

    + cd C:\Java\fess-server-6.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat6w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    +
    + +

    Distributed in the Fess is 32-bit binaries for Windows Tomcat builds based on. If you use 64-bit WindowsTomcat Of the site from, such as 64-bit Windows zip, please replace tomcat6,exe, tomcat6w,exe.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/install/index.xml b/src/site/en/xdoc/6.0/install/index.xml new file mode 100644 index 000000000..6b2adcd9b --- /dev/null +++ b/src/site/en/xdoc/6.0/install/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for Fess installation and uninstallation tasks. +

    +
    +
    +

    +This document shows how to install Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/6.0/install/install.xml b/src/site/en/xdoc/6.0/install/install.xml new file mode 100644 index 000000000..e34250c29 --- /dev/null +++ b/src/site/en/xdoc/6.0/install/install.xml @@ -0,0 +1,68 @@ + + + + Installation + + + + + +
    +

    +Fess can be available in these environments. +

    +
      +
    • OS: Windows, Unix, and Java can run OS environment
    • +
    • Java: Java 6 or more
    • +
    +

    +If Java is not installed in the environment we want to Fess takes over Java 6 JDK from http://java.sun.com/, please install. +

    +
    +
    +

    +http://sourceforge.jp/projects/Fess/releases/ From the download the latest Fess package. +

    +
    +
    +

    +Expand the downloaded fess-server-x.y.zip. +If you installed in the UNIX environment, in the bin added the performing rights to a script.

    + +
    +
    +

    +Administrator account is managed by the application server. Fess Server standard available Tomcat, as well as to the user changing the Tomcat. +Modify the password for the admin account of the conf/tomcat-user.xml if you want to change.

    + +]]> +
    +
    +

    +To access the Solr into Fess server is password becomes necessary. +Change the default passwords in production, etc.

    +

    +How to change the password, you must first change the password attribute of the conf/tomcat-user.xml solradmin. +

    + + +]]> +

    +Describes the provided password webapps/fess/WEB-INF/classes/fess_solr.dicon the following points tomcat-user.xml to the next. +

    + + + "solradmin" + "solradmin" + +]]> +
    + +
    diff --git a/src/site/en/xdoc/6.0/install/run.xml b/src/site/en/xdoc/6.0/install/run.xml new file mode 100644 index 000000000..2acf4ef01 --- /dev/null +++ b/src/site/en/xdoc/6.0/install/run.xml @@ -0,0 +1,41 @@ + + + + Start and stop + + + + + +
    +

    +to run the startup script, start the Fess. +

    + + +

    +access to / http://localhost:8080/Fess ensures startup. +

    + +

    +Management UI is / http://localhost:8080/fess/admin. +Default Administrator account user name / password is admin/admin. +Administrator account is managed by the application server. +In the management UI of the Fess, authenticate with the application server in fess role determine as an administrator.

    +
    +
    +

    +Fess to stop the running shutdown scripts. +

    + + +

    +Crawl or may take a while to completely stop during the index creation if you. +

    +
    + +
    diff --git a/src/site/en/xdoc/6.0/install/uninstall.xml b/src/site/en/xdoc/6.0/install/uninstall.xml new file mode 100644 index 000000000..d95af5b33 --- /dev/null +++ b/src/site/en/xdoc/6.0/install/uninstall.xml @@ -0,0 +1,19 @@ + + + + Uninstall + + + + + +
    +

    +You can uninstall by deleting the installation stops after the Fess fess-server-x.y directory. +

    + +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/index.xml b/src/site/en/xdoc/6.0/user/index.xml new file mode 100644 index 000000000..cd354729a --- /dev/null +++ b/src/site/en/xdoc/6.0/user/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is targeted to users to Fess. +

    +
    +
    +

    +This document demonstrates how to search in the Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/json-response.xml b/src/site/en/xdoc/6.0/user/json-response.xml new file mode 100644 index 000000000..b2e5c87fc --- /dev/null +++ b/src/site/en/xdoc/6.0/user/json-response.xml @@ -0,0 +1,184 @@ + + + + Search results for JSON + Shinsuke Sugaya + + +
    +

    Outputs a JSON results of Fess. You must in order to emit JSON enabled keep JSON response in the General Management screen crawl settings.

    + +

    To get JSON output http://localhost:8080/Fess/JSON? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    labelTypeValueThe label value. Use to specify the label.
    callbackIf you use the JSONP callback name. You do not need to specify if you do not use the JSONP.
    +
    + +

    Returns the following response.

    + \n..", + "tstamp": "1284739487887", + "url": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F", + "id": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F;type=au,docomo,pc,softbank", + "mimetype": "text\u002Fhtml", + "title": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "contentTitle": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "boost": "1.0", + "contentLength": "17810", + "urlLink": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F" + }, +... + ] + } +} +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. 0, the normal response.
    querySearch language.
    execTimeResponse time. Unit is seconds.
    pageSizeThe number of display.
    pageNumberPage number.
    recordCountNumber of hits for the search term.
    pageCountPage number of the number of hits for the search term.
    resultParent element of the search results.
    siteThe name of the site.
    contentDescriptionA description of the content.
    hostThe name of the host.
    lastModifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    contentTitleThe title of the document for viewing.
    contentLengthThe size of the document.
    urlLinkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-boost.xml b/src/site/en/xdoc/6.0/user/search-boost.xml new file mode 100644 index 000000000..697a4bd94 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-boost.xml @@ -0,0 +1,20 @@ + + + + Boost search + Shinsuke Sugaya + + +
    +

    Use the boost search if you want to prioritize, search for specific search terms. Enabling search in boost search, depending on the severity of the search words.

    + +

    To boost search after the search term ' ^ boost value "that specifies the boost value (weighted) in the format.

    +

    For example, if you want to find the page if you want to find apples oranges contained more 'apples', type the following search form.

    + +

    Boost value specifies an integer greater than 1.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-field.xml b/src/site/en/xdoc/6.0/user/search-field.xml new file mode 100644 index 000000000..72500ebf0 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-field.xml @@ -0,0 +1,62 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    You crawl in Fess results are saved for each field, such as title and full text. You can search for a field of them. You can specify the search criteria in search for a field, such as document type or size small.

    + +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Field list is available
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields subject to find the content. Fields are custom fields are also available by using the dynamic field of Solr.

    +

    If HTML file and search for the title tag that string in the title field, below the body tag registered in the body field.

    +
    + +

    If a field specifying the search field name: search words in separated by colons (:) field name and search word fill in the search form, the search.

    +

    If you search the Fess as a search term for the title field, type.

    + +

    Document, the title field in Fess above search appears as a search result.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-fuzzy.xml b/src/site/en/xdoc/6.0/user/search-fuzzy.xml new file mode 100644 index 000000000..30c6ba212 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-fuzzy.xml @@ -0,0 +1,21 @@ + + + + Fuzzy search (fuzzy search) + Shinsuke Sugaya + + +
    +

    Ambiguity in the case does not match the words word search to search is available. Based on the Levenshtein distance in Fess ambiguous corresponds to the search (fuzzy search).

    + +

    After the search word you want to apply the fuzzy search adds '~'.

    +

    For example, ambiguous word "Solr", you can find documents that contain the word, near the "Solr" If you want to find, type as the search form, such as ("Solar"). +

    + +

    Furthermore, if by '~' after the number 0 and 1, 1 just like in refine. For example, in the form of 'Solr~0.8'. Do not specify numeric default value is 0.5.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-label.xml b/src/site/en/xdoc/6.0/user/search-label.xml new file mode 100644 index 000000000..b23b481c4 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-label.xml @@ -0,0 +1,17 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    Narrow your search by adding the categories to search the document for label information the label is specified when the search is possible. Label information by registering in the Administration screen, will enable search by labels in the search screen. Label information available to selected in the drop-down search. If you do not register the label displayed the label drop-down box.

    + +

    You can select the label information at search time.

    + Label search +

    To set the label by creating indexes, can search each crawl settings specified on the label. All results search search do not specify a label is usually the same. If you change the label information to update the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-not.xml b/src/site/en/xdoc/6.0/user/search-not.xml new file mode 100644 index 000000000..121973321 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-not.xml @@ -0,0 +1,20 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If the document does not contain a word search NOT searching is available.

    +

    However, attention is required because NOT find expensive.

    + +

    Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, enter the following if you want to find documents that contain the search term 1 does not contain a search term 2 and the search.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-or.xml b/src/site/en/xdoc/6.0/user/search-or.xml new file mode 100644 index 000000000..6b3ed55f1 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-or.xml @@ -0,0 +1,20 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. When describing the multiple words in the search box, by default will search.

    + +

    To use search OR search words written OR. OR write in capital letters, the space required to back and forth.

    +

    For example, if you want to search for documents that contain either search term 2 search term 1 and type the following search form.

    + +

    OR between multiple languages are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-range.xml b/src/site/en/xdoc/6.0/user/search-range.xml new file mode 100644 index 000000000..84202b944 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-range.xml @@ -0,0 +1,19 @@ + + + + Range searches + Shinsuke Sugaya + + +
    +

    If in the field containing the data in specified range, such as the number range search is possible for that field.

    + +

    To limit "field name: value TO value ' fill in the search form.

    +

    For example, type to search document contentLength field against 1 k to 10 k bytes is shown below the search form.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/search-sort.xml b/src/site/en/xdoc/6.0/user/search-sort.xml new file mode 100644 index 000000000..b823fa354 --- /dev/null +++ b/src/site/en/xdoc/6.0/user/search-sort.xml @@ -0,0 +1,50 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    + +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + + +
    Sort fields list
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    +

    Adding custom fields as sort in Customizing.

    +
    + +

    You can select the sorting criteria when you search.

    + Sorting criteria +

    Also, for sorting in the search field sort: the field name to sort and field names separated by colon (:) fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/6.0/user/xml-response.xml b/src/site/en/xdoc/6.0/user/xml-response.xml new file mode 100644 index 000000000..3f7bb9d5b --- /dev/null +++ b/src/site/en/xdoc/6.0/user/xml-response.xml @@ -0,0 +1,207 @@ + + + + XML search results + Shinsuke Sugaya + + +
    +

    To emit the XML results of Fess. You need to do this output in XML Management screen crawl settings XML response enabled by you.

    + +

    To get the output XML http://localhost:8080/Fess/XML? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    labelTypeValueThe label value. Use to specify the label.
    +
    + +

    Returns the following response.

    + + + 1 + 0 + テスト + 0.64 + 20 + 1 + 101 + 6 + + + speedtest.goo.ne.jp/ + goo の提供するスピード<em>テスト</em>ページです + speedtest.goo.ne.jp + 1284739487873 + --> +<meta http-equiv="Content-Type" content="text/html; charset=EUC-JP"> +<title> +goo スピード<em>テスト</em> +</title> +<meta..."> + +<meta name="description" content="goo の提供するスピード<em>テスト</em>ページです"> +<meta name="keywords" content="ADSL,ブロードバンド,フレ...ッツ,BB,無線LAN,CATV,ISDN,スピード<em> +テスト</em>"> +<meta name="Copyright" content="Copyright &copy; ...://speedtest.goo.ne.jp/" +><img src="/img/h1/h1_76.gif" width="129" height="22" border="0" alt="スピード<em>テスト +</em>"><img src..." color="#ffffff">goo スピード<em>テスト</em></font +></a></b></td> +<!--/NAME SERVICE--> +</tr> +</table> +</td> +</tr> +<tr> +<td + 4.98744 + goo の提供するスピード<em>テスト</em>ページです + 1284739487887 + http://speedtest.goo.ne.jp/ + http://speedtest.goo.ne.jp/;type=au,docomo,pc,softbank + text/html + +goo スピードテスト + + +goo スピードテスト + + 1.0 + 17810 + http://speedtest.goo.ne.jp/ + +... + + +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. 0, the normal response.
    querySearch language.
    exec-timeResponse time. Unit is seconds.
    page-sizeThe number of display.
    page-numberPage number.
    record-countNumber of hits for the search term.
    page-countPage number of the number of hits for the search term.
    resultParent element of the search results.
    docThe elements of the search results.
    siteThe name of the site.
    content-descriptionA description of the content.
    hostThe name of the host.
    last-modifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    content-titleThe title of the document for viewing.
    content-lengthThe size of the document.
    URL-linkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/browserType-guide.xml b/src/site/en/xdoc/7.0/admin/browserType-guide.xml new file mode 100644 index 000000000..a3452c477 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/browserType-guide.xml @@ -0,0 +1,23 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
    +

    Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

    +
    +
    + +

    In Administrator account after logging in, click menu browser types.

    + Setting the browser type +
    +
    +
    + +

    You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/crawl-guide.xml b/src/site/en/xdoc/7.0/admin/crawl-guide.xml new file mode 100644 index 000000000..cbce17847 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/crawl-guide.xml @@ -0,0 +1,149 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to crawling.

    +
    +
    + +

    In Administrator account click crawl General menu after login.

    + Crawl General +

    You can specify the path to a generated index and replication capabilities to enable.

    + Replication features +
    +
    +
    + +

    You can set the interval at which the crawl for a Web site or file system. By default, the following.

    + +

    Figures are from left, seconds, minutes, during the day, month, represents a day of the week. Description format is similar to the Unix cron settings. This example, and am 0 時 0 分 to crawling daily.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    0 0 12 * *?Each day starts at 12 pm
    0 15 10? * *Day 10: 15 am start
    0 15 10 * *?Day 10: 15 am start
    0 15 10 * *? *Day 10: 15 am start
    0 15 10 * *? 2009Each of the 2009 start am, 10:15
    0 * 14 * *?Every day 2:00 in the PM-2: 59 pm start every 1 minute
    0 0 / 5 14 * *?Every day 2:00 in the PM-2: 59 pm start every 5 minutes
    0 0 / 5 14, 18 * *?Every day 2:00 pm-2: 59 pm and 6: 00 starts every 5 minutes at the PM-6: 59 pm
    0 0-5 14 * *?Every day 2:00 in the PM-2: 05 pm start every 1 minute
    0 10, 44 14? 3 WEDStarts Wednesday March 2: 10 and 2: 44 pm
    0 15 10? * MON-FRIMonday through Friday at 10:15 am start
    +

    Also check if the seconds can be set to run at intervals 60 seconds by default. If you set seconds exactly and you should customize webapps/fess/WEB-INF/classes/chronosCustomize.dicon taskScanIntervalTime value, if enough do I see in one-hour increments.

    +
    + +

    When the user enters a search, the search the output log. If you want to get search statistics to enable.

    +
    + +

    Save the information you find. Identifying the users becomes possible.

    +
    + +

    You can collect the search result was judged good by the user. Search result voting link appears to result in list screen, so that link press made the record. You can also reflect the results collected during the crawl index.

    +
    + +

    Search results link attaches to the search term. To display the find search terms in PDF becomes possible.

    +
    + +

    Search results can be retrieved in XML format. http://localhost:8080/Fess/XML? can get access query = search term.

    +
    + +

    Search results available in JSON format. http://localhost:8080/Fess/JSON? can get access query = search term.

    +
    + +

    If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

    +
    + +

    You can specify the label to see if the label by default,. Specifies the value of the label.

    +
    + +

    You can specify whether or not to display a search screen. If you select Web unusable for mobile search screen. If not available not available search screen. And if you want to create a dedicated index server and select not available.

    +
    + +

    In JSON format often find search words becomes available. can be retrieved by accessing the http://localhost:8080/Fess/hotsearchword.

    +
    + +

    Delete a session log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Delete a search log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Specifies the Bots name Bots you want to remove from the search log logs included in the user agent by commas (,). Log is deleted by log purge once a day.

    +
    + +

    Specifies the encoding for the CSV will be available in the backup and restore.

    +
    + +

    To enable replication features that can apply already copied the Solr index generated. For example, you can use them if you want to search only in the search servers crawled and indexed on a different server, placed in front.

    +
    + +

    After the data is registered for Solr. Index to commit or to optimize the registered data becomes available. If optimize is issued the Solr index optimization, if you have chosen, you choose to commit the commit is issued.

    +
    + +

    Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

    +
    + +

    In Fess in 10 units send the document for Solr. For each value specified here Solr issued document commits. If 0 commit is performed after crawl completion.

    +
    + +

    Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

    +

    But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

    +
    + +

    You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used. If you enable incremental crawl does not delete.

    +
    + +

    Registered disabled URL URL exceeds the failure count next time you crawl to crawl out. No need to worry about disability type is crawled next time by specifying this value.

    +
    + +

    Disaster URL exceeds the number of failures will crawl out.

    +
    + +

    Copy index information from the index directory as the snapshot path, if replication is enabled, will be applied.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/7.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..5a179a718 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/crawlingSession-guide.xml @@ -0,0 +1,27 @@ + + + + Set session information + Shinsuke Sugaya + + +
    +

    Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

    +
    +
    + +

    In Administrator account after logging in, click the session information menu.

    +
    +
    +
    + + Session information list +

    You can remove all session information and click the Delete link all in the running. Session has expired will be removed at next crawl.

    +
    + +

    Sure you can crawl the contents of session ID. Crawl start and finish time, number of documents indexed and listed.

    + Session details +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/data-guide.xml b/src/site/en/xdoc/7.0/admin/data-guide.xml new file mode 100644 index 000000000..b797db263 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/data-guide.xml @@ -0,0 +1,41 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
    +

    Here, describes Fess information backup and restore methods.

    +
    +
    + +

    In Administrator account after logging in, click the menu backup and restore.

    + Backup and restore +
    + +

    Click the download link and Fess information output in XML format. Saved settings information is below.

    +
      +
    • The General crawl settings
    • +
    • Web crawl settings
    • +
    • File system Crawl settings
    • +
    • Datastore crawl settings
    • +
    • Label
    • +
    • Path mapping
    • +
    • Web authentication
    • +
    • File system authentication
    • +
    • Request header
    • +
    • Duplicate host
    • +
    • Roll
    • +
    • Compatible browsers
    • +
    +

    Session information, search log, click log is available in CSV format.

    +

    In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate. If you need to back up the SOLR index backs solr directory.

    +
    + +

    You can restore settings information, various log in to upload XML output by backup or CSV. To specify the files, please click the restore button on the data.

    +

    If enable overwrite data in XML file configuration information specified when the same data is updating existing data.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/dataCrawlingConfig-guide.xml b/src/site/en/xdoc/7.0/admin/dataCrawlingConfig-guide.xml new file mode 100644 index 000000000..012cedccd --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/dataCrawlingConfig-guide.xml @@ -0,0 +1,159 @@ + + + + Settings for crawling the data store + Sone, Takaaki + Shinsuke Sugaya + + +
    +

    You can crawl databases in Fess. Here are required to store settings.

    +
    +
    + +

    In Administrator account after logging in, click menu data store.

    + Data store configuration +

    As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

    + +

    Here the data is put something like the following.

    + +
    +
    +
    + +

    Parameter settings example looks like the following.

    + +

    Parameter is a "key = value" format. Description of the key is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    For DB configuration parameter example
    driverDriver class name
    URLURL
    usernameTo connect to the DB user name
    passwordTo connect to the DB password
    SQLWant to crawl to get SQL statement
    +
    + +

    Script configuration example looks like the following.

    + +

    + Parameter is a "key = value" format. + Description of the key is as follows.

    +

    + Side of the value written in OGNL. Close the string in double quotation marks. + Access in the database column name, its value.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Script settings
    URLURLs (links appear in search results)
    hostHost name
    siteSite pass
    titleTitle
    contentContent (string index)
    cacheContent cache (not indexed)
    DigestDigest piece that appears in the search results
    anchorLinks to content (not usually required)
    contentLengthThe length of the content
    lastModifiedContent last updated
    +
    + +

    To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

    +
    + +

    Set the following in the webapps/fess/WEB-INF/classes/app.dicon if you see the item value, such as latitude_s in the search results. After adding to $ {doc.latitude_s}, searchResults.jsp;

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "latitude_s","longitude_s" } + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/design-guide.xml b/src/site/en/xdoc/7.0/admin/design-guide.xml new file mode 100644 index 000000000..e404de733 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/design-guide.xml @@ -0,0 +1,107 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
    +

    Here are settings for the design of search screens.

    +
    +
    + +

    In Administrator account after logging in, click the menu design.

    + Design +

    You can edit the search screen in the screen below.

    + JSP compilation screen +
    + +

    If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

    + + +<% +java.util.Map docMap = (java.util.Map)pageContext.getAttribute("doc"); +Long tstampValue = (Long)docMap.get("tstamp"); +java.util.Date tstampDate = new java.util.Date(tstampValue); +Long lastModifiedValue = (Long)docMap.get("lastModified"); +java.util.Date lastModifiedDate = new java.util.Date(lastModifiedValue); +java.text.SimpleDateFormat sdf = new java.text.SimpleDateFormat("yyyy/MM/dd HH:mm"); +%> +
  • +

    + ${f:h(doc.contentTitle)} +

    +
    +${doc.contentDescription} +
    + ${f:h(doc.site)} +
    Registered: <%= sdf.format(tstampDate) %> +
    Last Modified: <%= sdf.format(lastModifiedDate) %> + : + : +]]> +

    tstampDate will update on registration date, lastModifiedDate. Output date format is specified in SimpeDateFormat.

    + +
  • +
    + +

    On the search screen files are available to download and can be removed.

    +
    + +

    You can upload the file to use in the search screen. Image file names are supported are jpg, gif, png, css, and js.

    +
    + +

    Use if you want to specify the file name file to upload. Uploaded if you omit the file name will be used.

    +
    + +

    You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    JSP file that you can edit
    Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
    Top page (within the Head tags)This is the express search home page head tag in JSP files. If you want to edit the meta tags, title tags, script tags, such as the change.
    Top page (content)Is a JSP file to represent the body tag in the search home page.
    Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
    Search results page (within the Head tags)Search result is a JSP file to represent within the head tag of the list page. If you want to edit the meta tags, title tags, script tags, such as the change.
    Search results page (header)Search result is a JSP file to represent the header of the list page. Include search form at the top.
    Search results page (footer)Search result is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
    Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    +

    You can to edit for PCs and similar portable screen.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/failureUrl-guide.xml b/src/site/en/xdoc/7.0/admin/failureUrl-guide.xml new file mode 100644 index 000000000..013622eb1 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/failureUrl-guide.xml @@ -0,0 +1,23 @@ + + + + Disaster URL + Shinsuke Sugaya + + +
    +

    Here the failure URL. URL could not be obtained at crawl time are recorded and confirmed as the failure URL.

    +
    +
    + +

    In Administrator account click menu disabled URL after login.

    + Disaster URL +

    Clicking the confirmation link failure URL displayed for more information.

    + Details of the disaster URL +
    + +

    A glance could not crawl the URL and date.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/favoriteLog-guide.xml b/src/site/en/xdoc/7.0/admin/favoriteLog-guide.xml new file mode 100644 index 000000000..76af131d1 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/favoriteLog-guide.xml @@ -0,0 +1,21 @@ + + + + Popular URL log + Shinsuke Sugaya + + +
    +

    Here are popular URL log. When the popular URL log user clicks voting link on the search screen registers as a favorite link. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click the menu popular URL.

    + Users list +
    + +

    Lists popular URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/fileAuthentication-guide.xml b/src/site/en/xdoc/7.0/admin/fileAuthentication-guide.xml new file mode 100644 index 000000000..1ddb4dc1f --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/fileAuthentication-guide.xml @@ -0,0 +1,44 @@ + + + + Settings for file system authentication + Shinsuke Sugaya + + +
    +

    Crawls using file system here, describes how to set file system authentication is required. Fess is corresponding to a crawl for a shared folder in Windows.

    +
    +
    + +

    In Administrator account after logging in, click the menu file system authentication.

    + File system settings +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Is omitted, the specified file system Kroll set applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. File system Crawl settings specified in that case applies on any port.

    +
    + +

    Select the authentication method. You can use SAMBA (Windows shared folder authentication).

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. SAMBA, the set value of the domain. If you want to write as.

    + +
    + +

    Select a file name to apply the authentication settings for the above. Must be registered ago you file system CROLL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/7.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..89f9667e2 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,106 @@ + + + + Settings for file system crawling + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: or smb: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +

    Windows shared folder, for example, if you want to crawl to host1 share folder crawl settings for smb: (last / to) the //host1/share/. If authentication is in the shared folder on the file system authentication screen set authentication information.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + + +
    IP rings contents list
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    You can specify the crawl configuration information.

    +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/index.xml b/src/site/en/xdoc/7.0/admin/index.xml new file mode 100644 index 000000000..87a95df86 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/index.xml @@ -0,0 +1,67 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for the management work of the Fess. +

    +
    +
    +

    +This document demonstrates Fess configuration management methods. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    +
    + +

    Management UI to support browser as follows.

    +
      +
    • Internet Explorer 7 or higher
    • +
    • Firefox 4 or more
    • +
    • Chrome (latest version)
    • +
    • Safari (latest version)
    • +
    • Opera 11
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/labelType-guide.xml b/src/site/en/xdoc/7.0/admin/labelType-guide.xml new file mode 100644 index 000000000..7e3b342fe --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/labelType-guide.xml @@ -0,0 +1,33 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. If you register the label shown select label drop-down box to the right of the search box.

    +
    +
    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    +
    +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    + +

    Specifies the role to view the label.

    +
    + +

    Specifies the order of the labels.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/log-guide.xml b/src/site/en/xdoc/7.0/admin/log-guide.xml new file mode 100644 index 000000000..00c2c8fd0 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/log-guide.xml @@ -0,0 +1,21 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    +
    +
    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/7.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..8b6790e7c --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/overlappingHost-guide.xml @@ -0,0 +1,27 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    +
    +
    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    +
    +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/7.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..22c1d76be --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/pathMapping-guide.xml @@ -0,0 +1,30 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    +
    +
    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    +
    +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/7.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..6b8aadf8d --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/requestHeader-guide.xml @@ -0,0 +1,30 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    +
    +
    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    +
    +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/roleType-guide.xml b/src/site/en/xdoc/7.0/admin/roleType-guide.xml new file mode 100644 index 000000000..686275f94 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/roleType-guide.xml @@ -0,0 +1,27 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    +
    +
    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    +
    +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/searchList-guide.xml b/src/site/en/xdoc/7.0/admin/searchList-guide.xml new file mode 100644 index 000000000..f72094fcf --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/searchList-guide.xml @@ -0,0 +1,21 @@ + + + + Search + Shinsuke Sugaya + + +
    +

    Here the search for management.

    +
    +
    + +

    In Administrator account after logging in, click the menu search.

    + Administrative search +
    + +

    You can search by criteria you specify. In the regular search screen role and browser requirements is added implicitly, but do not provide management for search. You can document a certain remove from index from the search results.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/searchLog-guide.xml b/src/site/en/xdoc/7.0/admin/searchLog-guide.xml new file mode 100644 index 000000000..bd3d249ad --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/searchLog-guide.xml @@ -0,0 +1,21 @@ + + + + Search log list + Shinsuke Sugaya + + +
    +

    Here the search log. When you search in the search screen users search logs are logged. Search log search term or date is recorded. You can also record the URL, then you want the search results to.

    +
    +
    + +

    In Administrator account after logging in, click menu search logs.

    + Search log +
    + +

    Search language and date are listed. You can review and detailed, you click the URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/stats-guide.xml b/src/site/en/xdoc/7.0/admin/stats-guide.xml new file mode 100644 index 000000000..25923a212 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/stats-guide.xml @@ -0,0 +1,21 @@ + + + + Statistics + Shinsuke Sugaya + + +
    +

    Here are statistics. You can search log and click log.

    +
    +
    + +

    In Administrator account after logging in, click the menu statistics.

    + Statistics +
    + +

    You can select the target by selecting the type of report, to make sure. Displayed in order by the specified criteria.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/system-guide.xml b/src/site/en/xdoc/7.0/admin/system-guide.xml new file mode 100644 index 000000000..76f14af15 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/system-guide.xml @@ -0,0 +1,35 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here registration in Fess. SOLR servers are grouped by file, has been registered.

    +
    +
    + +

    In Administrator account after logging in, click menu Solr.

    + System settings +
    +
    +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and shut down when not running Fess server to shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +
    + +

    Server group name to search for and update available will be shown.

    +
    + +

    Server becomes unavailable and the status of disabled. For example, inaccessible to the Solr server and changes to disabled. To enable recovery after server become unavailable will become available.

    +
    + +

    You can publish index commit, optimize for server groups. You can also remove a specific search for the session ID. You can remove only the specific documents by specifying the URL.

    +
    + +

    Shown by the number of documents registered in each session. Can verify the results list by clicking the session name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/systemInfo-guide.xml b/src/site/en/xdoc/7.0/admin/systemInfo-guide.xml new file mode 100644 index 000000000..27ba783b6 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/systemInfo-guide.xml @@ -0,0 +1,32 @@ + + + + System information + Shinsuke Sugaya + + +
    +

    Here, you can currently check property information such as system environment variables.

    +
    +
    + +

    In Administrator account after logging in, click system information menu.

    + System information +
    +
    +
    + +

    You can list the server environment variable.

    +
    + +

    You can list the system properties on Fess.

    +
    + +

    Fess setup information available.

    +
    + +

    Is a list of properties to attach when reporting a bug. Extract the value contains no personal information.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/userInfo-guide.xml b/src/site/en/xdoc/7.0/admin/userInfo-guide.xml new file mode 100644 index 000000000..573db0628 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/userInfo-guide.xml @@ -0,0 +1,21 @@ + + + + User log + Shinsuke Sugaya + + +
    +

    Here the user log. Identifies the user when you search in the search screen users the user log in. You can search log and popular URL information and the use. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click menu users.

    + Users list +
    + +

    Lists the ID of the user. You can select the search logs or popular URL links, to see a list of each log.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/7.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..69ae71ce7 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/webAuthentication-guide.xml @@ -0,0 +1,48 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    +
    +
    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication, DIGEST authentication or NTLM authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. You can set the workstation and domain values for NTLM authentication. If you want to write as.

    + +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/7.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..56a5d2c11 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,107 @@ + + + + Settings for crawling Web site + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + + +
    URL filtering contents list
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    You can specify the crawl configuration information.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl. If you do not specify people per 100,000.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in)

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/admin/wizard-guide.xml b/src/site/en/xdoc/7.0/admin/wizard-guide.xml new file mode 100644 index 000000000..263fdca02 --- /dev/null +++ b/src/site/en/xdoc/7.0/admin/wizard-guide.xml @@ -0,0 +1,35 @@ + + + + Settings Wizard + Sone, Takaaki + + +
    +

    Introduction to the Configuration Wizard.

    +

    You can use Settings Wizard, to set you up on the fess.

    +
    +
    + +

    In Administrator account after logging in, click menu Settings Wizard.

    + Settings Wizard +

    First, setting a schedule. + During the time in fess is crawling and indexes. + By default, every day is a 0 時 0 分. Schedules can change even the General crawl settings.

    + Setting a schedule +

    The crawl settings. + Crawl settings is to register a URI to look for. + The crawl settings name please put name of any easy to identify. Put the URI part de-indexed, want to search for.

    + Crawl settings +

    For example, if you want and search for http://fess.codelibs.org/, less looks like.

    + Crawl settings example +

    The type, such as c:\Users\taro file.

    +

    In this is the last setting. Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

    + Crawl started +
    + +

    Settings in the Setup Wizard you can change from crawl General, Web, file system.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/db.xml b/src/site/en/xdoc/7.0/config/db.xml new file mode 100644 index 000000000..d40d85e16 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/db.xml @@ -0,0 +1,128 @@ + + + + Database changes + Shinsuke Sugaya + + +
    +

    Under normal circumstances the database use the H2 Database. You can use other databases by changing settings.

    +
    +
    + +

    The MySQL character code setting. /etc/mysql/my.cnf and the added must have the following settings.

    + +
    + +

    Expand the MySQL binaries.

    +
    + +

    Create a database.

    + create database fess_db; +mysql> grant all privileges on fess_db.* to fess_user@localhost identified by 'fess_pass'; +mysql> create database fess_robot; +mysql> grant all privileges on fess_robot.* to s2robot@localhost identified by 's2robot'; +mysql> FLUSH PRIVILEGES; +]]> +

    Create a table in the database. DDL file is located in extension/mysql.

    + +

    webapps/fess/WEB-INF/lib on mysql driver jar.

    + +

    Edit the webapps/fess/WEB-INF/classes/jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_db?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "fess_user" + "fess_pass" + + + + 600 + 10 + true + + + + + +]]> +

    Edit the webapps/fess/WEB-INF/classes/s2robot_jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_robot?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "s2robot" + "s2robot" + + + + 600 + 10 + true + + @java.sql.Connection@TRANSACTION_REPEATABLE_READ + + + + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/desktop-search.xml b/src/site/en/xdoc/7.0/config/desktop-search.xml new file mode 100644 index 000000000..029af0eb3 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/desktop-search.xml @@ -0,0 +1,35 @@ + + + + The desktop search settings + Shinsuke Sugaya + + +
    +

    + Increasing awareness of security in the browser environment in recent years, open a local file (for example, c:\hoge.txt) from the Web pages on. + Not to copy and paste the link from the search results, and then reopen the usability is good. + In order to respond to this in Fess and provides desktop search functionality.

    +
    +
    +

    + Desktop Search feature is turned off by default. + Please enable the following settings.

    +

    First of all, bin/setenv.bat as java.awt.headless from true to false edits.

    + +

    Then add the following to webapps/fess/WEB-INF/conf/crawler.properties.

    + +

    Start the Fess, after you set up above. How to use Basic remains especially.

    +
    +
    +
      +
    • Please Fess inaccessible from the outside, such as (for example, 8080 port does not release).
    • +
    • because false Java.awt.headless image size conversion for mobile devices is not available.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/filesize.xml b/src/site/en/xdoc/7.0/config/filesize.xml new file mode 100644 index 000000000..1cf88e616 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/filesize.xml @@ -0,0 +1,29 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of Defultmxlength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/geosearch.xml b/src/site/en/xdoc/7.0/config/geosearch.xml new file mode 100644 index 000000000..24702a980 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/geosearch.xml @@ -0,0 +1,23 @@ + + + + GEO search + Shinsuke Sugaya + + +
    +

    Together with Google maps, including document with latitude and longitude location information, GEO (GEO) you can use the search.

    +
    +
    + +

    Location is defined as a feed that contains the location information. + When generating the index in Solr latitude longitude set to location feeds in formats such as 45.17014,-93.87341, register the document. + Also sets the value as the latitude_s and longitude_s fields if you want to display latitude and longitude as a search result. * _s is available as a dynamic field of Solr string.

    +
    + +

    During the search specifies in the request parameter to latitude and longitude, the distance. + View the results in the distance (km) specified by distance-based latitude information (latitude, longitude). Latitude and longitude and distances is treated as double.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/index-backup.xml b/src/site/en/xdoc/7.0/config/index-backup.xml new file mode 100644 index 000000000..541993b34 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index management + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data directory. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/index.xml b/src/site/en/xdoc/7.0/config/index.xml new file mode 100644 index 000000000..0ea320154 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for setting work of the Fess. +

    +
    +
    +

    +This document demonstrates setting Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/install-on-tomcat.xml b/src/site/en/xdoc/7.0/config/install-on-tomcat.xml new file mode 100644 index 000000000..314d28334 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/install-on-tomcat.xml @@ -0,0 +1,43 @@ + + + + Install to an existing Tomcat + Shinsuke Sugaya + + +
    +

    + The standard distribution of Fess Tomcat is distributed in the deployed State. + Because Fess is not dependent on Tomcat, deploying on any Java application server is available. + Describes how to deploy a Fess Tomcat here is already available.

    +
    +
    +

    Fess ServerHereFrom the download.

    +
    +
    +

    + Expand the downloaded Fess server. + Expanded Fess Server home directory to $FESS_HOME. + $TOMCAT_HOME the top directory of an existing Tomcat 6. + Copy the Fess Server data.

    + +

    + If you have, such as changing the destination file diff commands, updates your diff only applies. +

    +
    +
    +

    to start the Tomcat in normal as well as in startup.* / http://localhost:8080/Fess To access the.

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/logging.xml b/src/site/en/xdoc/7.0/config/logging.xml new file mode 100644 index 000000000..0c991dc92 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/logging.xml @@ -0,0 +1,52 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + + +
    Log file list
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina.outLog of the Fess Server (Tomcat). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    + +

    You can check by changing the logging level of the HttpClient if you want to check HTTP access when crawling content to solve the problem.

    + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/memory-config.xml b/src/site/en/xdoc/7.0/config/memory-config.xml new file mode 100644 index 000000000..ade9047e5 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/memory-config.xml @@ -0,0 +1,52 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    Set the maximum memory per process in Java. So, do not use the upper memory in the process also had 8 GB of physical memory on the server. Memory consumption depending on the number of crawl threads and interval will also change significantly. If not enough memory please change settings in the subsequent procedure.

    +
    +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case the maximum value set 1024M) will change to-Xmx1024m.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-Xmx1024m (in this case the maximum value set 1024M). +

    + new String[] { +"-Djava.awt.headless=true", +"-server", +"-Xmx1024m", +"-XX:MaxPermSize=128m", +"-XX:-UseGCOverheadLimit", +"-XX:+UseConcMarkSweepGC", +"-XX:CMSInitiatingOccupancyFraction=75", +"-XX:+CMSIncrementalMode", +"-XX:+CMSIncrementalPacing", +"-XX:CMSIncrementalDutyCycleMin=0", +"-XX:+UseParNewGC", +"-XX:+UseStringCache", +"-XX:+UseTLAB", +"-XX:+DisableExplicitGC" +} +]]> +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/mobile-device.xml b/src/site/en/xdoc/7.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/path-encoding.xml b/src/site/en/xdoc/7.0/config/path-encoding.xml new file mode 100644 index 000000000..b8aa16ac6 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    +
    +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/7.0/config/pdf-with-password.xml b/src/site/en/xdoc/7.0/config/pdf-with-password.xml new file mode 100644 index 000000000..0a8f4e317 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/pdf-with-password.xml @@ -0,0 +1,89 @@ + + + + Password-protected PDF + Shinsuke Sugaya + + +
    +

    + You should password files to register the settings file to PDF password is configured to search for. +

    +
    +
    +

    + First of all, create the webapps/fess/WEB-INF/classes/s2robot_extractor.dicon. + This is test _ ~ is a pass that password set to a.pdf file. + If you have multiple files, multiple settings in addPassword.

    + + + + + + + + + + + + + ".*test_.*.pdf" + + "pass" + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + +... +]]> +

    Next, edit the following to webapps/fess/WEB-INF/classes/s2robot_rule.dicon.

    + + "fsFileRule" + + + fessFileTransformer + + + true + + "url" + "file:.*" + + + "mimeType" + + + "(application/xml" ++ "|application/xhtml+xml" ++ "|application/rdf+xml" ++ "|application/pdf" ++ "|text/xml" ++ "|text/xml-external-parsed-entity" ++ "|text/html)" + + +
    +... +]]> +

    Fess up after setting the above and please run the crawl. How to use Basic remains especially.

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/protwords.xml b/src/site/en/xdoc/7.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/proxy.xml b/src/site/en/xdoc/7.0/config/proxy.xml new file mode 100644 index 000000000..9c0e2421b --- /dev/null +++ b/src/site/en/xdoc/7.0/config/proxy.xml @@ -0,0 +1,107 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    +
    +

    Proxy is set in to create webapps/fess/WEB-INF/classes/s2robot_client.dicon with the following contents.

    + + + + + + + + + @org.apache.http.client.params.CookiePolicy@BEST_MATCH + clientConnectionManager + "プロキシサーバー名(ex. 192.168.1.1)" + プロキシサーバーのポート(ex. 8080) (" は不要) + + + + + + + + + + "http" + 80 + + @org.apache.http.conn.scheme.PlainSocketFactory@getSocketFactory() + + + + + + + + "https" + 443 + + @org.apache.http.conn.ssl.SSLSocketFactory@getSocketFactory() + + + + + + + 200 + 200 + + +]]> + + + "UTF-8" + + + + "UTF-8" + + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + "smb:.*" + smbClient + + + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/7.0/config/replication.xml b/src/site/en/xdoc/7.0/config/replication.xml new file mode 100644 index 000000000..cd9228849 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/replication.xml @@ -0,0 +1,25 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Fess can copy the path in Solr index data. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +

    You must use the replication features of Fess for Solr index file in the shared disk, such as NFS, Fess of each can be referenced from.

    +
    +
    + +

    Fess, download and install the./ /NET/Server1/usr/local/Fess To assume you installed.

    +

    To register the crawl settings as well as Fess starts after the normal construction, create the index (index for Fess building instructions normal building procedures and especially remains the same) crawling.

    +
    + +

    Fess, download and install the./ /NET/Server2/usr/local/Fess To assume you installed.

    +

    To enable replication features check box in Fess starts after the management screen crawl settings the "snapshot path'. Snapshot path designates the index location for the index for Fess. In this case, the/NET/Server1/usr/local/Fess //solr/core1/data/index In the will.

    + Replication +

    Time press the update button to save the data and set in Schedule performs replication of the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/role-setting.xml b/src/site/en/xdoc/7.0/config/role-setting.xml new file mode 100644 index 000000000..91321034a --- /dev/null +++ b/src/site/en/xdoc/7.0/config/role-setting.xml @@ -0,0 +1,97 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    +
      +
    • Request parameter
    • +
    • Request header
    • +
    • Cookies
    • +
    • J2EE authentication information
    • +
    +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/app.dicon shown below.

    + + + {"guest"} + + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + "role1" + + : +]]> +

    authenticatedRoles can describe multiple by commas (,).

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/server-port.xml b/src/site/en/xdoc/7.0/config/server-port.xml new file mode 100644 index 000000000..8438aab1f --- /dev/null +++ b/src/site/en/xdoc/7.0/config/server-port.xml @@ -0,0 +1,31 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    • 19092: database port (use h2database)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information. change the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    + "http://localhost:8080/solr" +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/solr-dynamic-field.xml b/src/site/en/xdoc/7.0/config/solr-dynamic-field.xml new file mode 100644 index 000000000..2b5a2ddef --- /dev/null +++ b/src/site/en/xdoc/7.0/config/solr-dynamic-field.xml @@ -0,0 +1,56 @@ + + + + Dynamic fields + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in solr/core1/conf/schema.xml. dynamic fields and standard fields such as title and content can be freely defined field names are defined. The dynamic fields that are available in the schema.xml Fess become. Advanced parameter values see a Solr document.

    + + + + + + + + + + + + + + + + + + + + + + +]]> +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add a field to use to retrieve data that is stored in the dynamic field next to the webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"url", "host", "site", + "title", "content", "contentLength", "lastModified", "mimetype", + "label", "segment", "other_t" } + +]]> +

    You must add a field to use to retrieve data that is stored in the dynamic field from Solr. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/solr-failure.xml b/src/site/en/xdoc/7.0/config/solr-failure.xml new file mode 100644 index 000000000..687dc70b6 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/fess_solr.dicon.

    +
    +
    +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is minActiveServer following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status maxRetryStatusCheckCount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    +
    +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    +
    +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/tokenizer.xml b/src/site/en/xdoc/7.0/config/tokenizer.xml new file mode 100644 index 000000000..fcdb1422e --- /dev/null +++ b/src/site/en/xdoc/7.0/config/tokenizer.xml @@ -0,0 +1,47 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search. Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits. For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit. For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default CJKTokenizer used.

    + + +

    Such as CJKTokenizer Japan Japanese multibyte string against bi-gram, in other words two characters create index. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to. However, please note that the index size increases.

    +

    The following example to change the analyzer part like solr/core1/conf/schema.xml, you can use the StandardTokenizer.

    + + + + + + + : + + + + + : +]]> +

    Also, useBigram is enabled by default in the webapps/fess/WEB-INF/classes/app.dicon change to false.

    + + true + : +]]> +

    After the restart the Fess.

    +
    +
    + + +
    diff --git a/src/site/en/xdoc/7.0/config/use-libreoffice.xml b/src/site/en/xdoc/7.0/config/use-libreoffice.xml new file mode 100644 index 000000000..363c9c3f3 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/use-libreoffice.xml @@ -0,0 +1,85 @@ + + + + Use of LibreOffice + Shinsuke Sugaya + + +
    +

    + It is possible to crawl using the Apache POI Fess environmental standard in MS Office system document. + You can crawl Office system document regarding LibreOffice, OpenOffice, do even more accurate text extraction from documents.

    +
    +
    +

    JodConverter Fess server install. from http://jodconverter.googlecode.com/jodconverter-core-3.0-Beta-4-Dist.zipThe download. Expand and copy the jar file to Fess server.

    + +

    Create a s2robot_extractor.dicon to the next.

    + +

    s2robot_extractor.dicon effective jodExtractor with following contents.

    + + + + + + + + + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + + officeManagerConfiguration.setOfficeHome("/usr/lib/libreoffice") + .buildOfficeManager() + + + + + + { +"application/msword", +"application/vnd.ms-excel", +"application/vnd.ms-powerpoint", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.openxmlformats-officedocument.presentationml.presentation" + } + jodExtractor + +... +]]> +

    Index to generate the settings later, usually crawled into the street.

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/config/windows-service.xml b/src/site/en/xdoc/7.0/config/windows-service.xml new file mode 100644 index 000000000..093123730 --- /dev/null +++ b/src/site/en/xdoc/7.0/config/windows-service.xml @@ -0,0 +1,54 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as Add %JAVA_HOME%\bin to PathYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option.

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-7.0.0.

    + cd C:\Java\fess-server-7.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat6w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    +
    + +

    Distributed in the Fess is 32-bit binaries for Windows Tomcat builds based on. If you use 64-bit WindowsTomcat For 64 bit Windows zip, such as getting from the site and replace tomcat6.exe, tomcat6w.exe.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/install/index.xml b/src/site/en/xdoc/7.0/install/index.xml new file mode 100644 index 000000000..62f1d2b0c --- /dev/null +++ b/src/site/en/xdoc/7.0/install/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for Fess installation and uninstallation tasks. +

    +
    +
    +

    +This document shows how to install Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/7.0/install/install.xml b/src/site/en/xdoc/7.0/install/install.xml new file mode 100644 index 000000000..e34250c29 --- /dev/null +++ b/src/site/en/xdoc/7.0/install/install.xml @@ -0,0 +1,68 @@ + + + + Installation + + + + + +
    +

    +Fess can be available in these environments. +

    +
      +
    • OS: Windows, Unix, and Java can run OS environment
    • +
    • Java: Java 6 or more
    • +
    +

    +If Java is not installed in the environment we want to Fess takes over Java 6 JDK from http://java.sun.com/, please install. +

    +
    +
    +

    +http://sourceforge.jp/projects/Fess/releases/ From the download the latest Fess package. +

    +
    +
    +

    +Expand the downloaded fess-server-x.y.zip. +If you installed in the UNIX environment, in the bin added the performing rights to a script.

    + +
    +
    +

    +Administrator account is managed by the application server. Fess Server standard available Tomcat, as well as to the user changing the Tomcat. +Modify the password for the admin account of the conf/tomcat-user.xml if you want to change.

    + +]]> +
    +
    +

    +To access the Solr into Fess server is password becomes necessary. +Change the default passwords in production, etc.

    +

    +How to change the password, you must first change the password attribute of the conf/tomcat-user.xml solradmin. +

    + + +]]> +

    +Describes the provided password webapps/fess/WEB-INF/classes/fess_solr.dicon the following points tomcat-user.xml to the next. +

    + + + "solradmin" + "solradmin" + +]]> +
    + +
    diff --git a/src/site/en/xdoc/7.0/install/run.xml b/src/site/en/xdoc/7.0/install/run.xml new file mode 100644 index 000000000..2acf4ef01 --- /dev/null +++ b/src/site/en/xdoc/7.0/install/run.xml @@ -0,0 +1,41 @@ + + + + Start and stop + + + + + +
    +

    +to run the startup script, start the Fess. +

    + + +

    +access to / http://localhost:8080/Fess ensures startup. +

    + +

    +Management UI is / http://localhost:8080/fess/admin. +Default Administrator account user name / password is admin/admin. +Administrator account is managed by the application server. +In the management UI of the Fess, authenticate with the application server in fess role determine as an administrator.

    +
    +
    +

    +Fess to stop the running shutdown scripts. +

    + + +

    +Crawl or may take a while to completely stop during the index creation if you. +

    +
    + +
    diff --git a/src/site/en/xdoc/7.0/install/uninstall.xml b/src/site/en/xdoc/7.0/install/uninstall.xml new file mode 100644 index 000000000..d95af5b33 --- /dev/null +++ b/src/site/en/xdoc/7.0/install/uninstall.xml @@ -0,0 +1,19 @@ + + + + Uninstall + + + + + +
    +

    +You can uninstall by deleting the installation stops after the Fess fess-server-x.y directory. +

    + +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/index.xml b/src/site/en/xdoc/7.0/user/index.xml new file mode 100644 index 000000000..54b7ae84d --- /dev/null +++ b/src/site/en/xdoc/7.0/user/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is targeted to users to Fess. +

    +
    +
    +

    +This document demonstrates how to search in the Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN9sm, Inc.To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/json-response.xml b/src/site/en/xdoc/7.0/user/json-response.xml new file mode 100644 index 000000000..0d8063b8d --- /dev/null +++ b/src/site/en/xdoc/7.0/user/json-response.xml @@ -0,0 +1,184 @@ + + + + Search results for JSON + Shinsuke Sugaya + + +
    +

    Outputs a JSON results of Fess. You must in order to emit JSON enabled keep JSON response in the General Management screen crawl settings.

    + +

    To get JSON output http://localhost:8080/Fess/JSON? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    callbackIf you use the JSONP callback name. You do not need to specify if you do not use the JSONP.
    +
    + +

    Returns the following response.

    + \n..", + "tstamp": "1284739487887", + "url": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F", + "id": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F;type=au,docomo,pc,softbank", + "mimetype": "text\u002Fhtml", + "title": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "contentTitle": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "boost": "1.0", + "contentLength": "17810", + "urlLink": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F" + }, +... + ] + } +} +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. 0, the normal response.
    querySearch language.
    execTimeResponse time. Unit is seconds.
    pageSizeThe number of display.
    pageNumberPage number.
    recordCountNumber of hits for the search term.
    pageCountPage number of the number of hits for the search term.
    resultParent element of the search results.
    siteThe name of the site.
    contentDescriptionA description of the content.
    hostThe name of the host.
    lastModifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    contentTitleThe title of the document for viewing.
    contentLengthThe size of the document.
    urlLinkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-and.xml b/src/site/en/xdoc/7.0/user/search-and.xml new file mode 100644 index 000000000..f20ac20a5 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-and.xml @@ -0,0 +1,20 @@ + + + + AND search + Shinsuke Sugaya + + +
    +

    Use the search if you want to search for documents that contain all search words of more than one. When describing multiple words in the search box separated by spaces, AND skip AND search.

    + +

    If you use the search search words written AND. Write in capital letters AND the space required to back and forth. AND is possible can be omitted.

    +

    For example, if you want to find documents that contain the search terms 1 and 2 search terms, type the following search form.

    + +

    Connecting multiple words with AND are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-boost.xml b/src/site/en/xdoc/7.0/user/search-boost.xml new file mode 100644 index 000000000..697a4bd94 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-boost.xml @@ -0,0 +1,20 @@ + + + + Boost search + Shinsuke Sugaya + + +
    +

    Use the boost search if you want to prioritize, search for specific search terms. Enabling search in boost search, depending on the severity of the search words.

    + +

    To boost search after the search term ' ^ boost value "that specifies the boost value (weighted) in the format.

    +

    For example, if you want to find the page if you want to find apples oranges contained more 'apples', type the following search form.

    + +

    Boost value specifies an integer greater than 1.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-field.xml b/src/site/en/xdoc/7.0/user/search-field.xml new file mode 100644 index 000000000..3c668053c --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-field.xml @@ -0,0 +1,66 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    You crawl in Fess results are saved for each field, such as title and full text. You can search for a field of them. You can specify the search criteria in search for a field, such as document type or size small.

    + +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Field list is available
    Field nameDescription
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields subject to find the content. Fields are custom fields are also available by using the dynamic field of Solr.

    +

    If HTML file and search for the title tag that string in the title field, below the body tag registered in the body field.

    +
    + +

    If a field specifying the search field name: search words in separated by colons (:) field name and search word fill in the search form, the search.

    +

    If you search the Fess as a search term for the title field, type.

    + +

    Document, the title field in Fess above search appears as a search result.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-fuzzy.xml b/src/site/en/xdoc/7.0/user/search-fuzzy.xml new file mode 100644 index 000000000..30c6ba212 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-fuzzy.xml @@ -0,0 +1,21 @@ + + + + Fuzzy search (fuzzy search) + Shinsuke Sugaya + + +
    +

    Ambiguity in the case does not match the words word search to search is available. Based on the Levenshtein distance in Fess ambiguous corresponds to the search (fuzzy search).

    + +

    After the search word you want to apply the fuzzy search adds '~'.

    +

    For example, ambiguous word "Solr", you can find documents that contain the word, near the "Solr" If you want to find, type as the search form, such as ("Solar"). +

    + +

    Furthermore, if by '~' after the number 0 and 1, 1 just like in refine. For example, in the form of 'Solr~0.8'. Do not specify numeric default value is 0.5.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-label.xml b/src/site/en/xdoc/7.0/user/search-label.xml new file mode 100644 index 000000000..60cd469db --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-label.xml @@ -0,0 +1,17 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    Narrow your search by adding the categories to search the document for label information the label is specified when the search is possible. Label information by registering in the Administration screen, will enable search by labels in the search screen. Label information available can multiple selections in the drop-down when you search. If you do not register the label displayed the label drop-down box.

    + +

    You can select the label information at search time.

    + Label search +

    To set the label by creating indexes, can search each crawl settings specified on the label. All results search search do not specify a label is usually the same. If you change the label information to update the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-not.xml b/src/site/en/xdoc/7.0/user/search-not.xml new file mode 100644 index 000000000..3a40c9f2b --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-not.xml @@ -0,0 +1,19 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If the document does not contain a word search NOT searching is available.

    + +

    Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, enter the following if you want to find documents that contain the search term 1 does not contain a search term 2 and the search.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-or.xml b/src/site/en/xdoc/7.0/user/search-or.xml new file mode 100644 index 000000000..6b3ed55f1 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-or.xml @@ -0,0 +1,20 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. When describing the multiple words in the search box, by default will search.

    + +

    To use search OR search words written OR. OR write in capital letters, the space required to back and forth.

    +

    For example, if you want to search for documents that contain either search term 2 search term 1 and type the following search form.

    + +

    OR between multiple languages are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-range.xml b/src/site/en/xdoc/7.0/user/search-range.xml new file mode 100644 index 000000000..84202b944 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-range.xml @@ -0,0 +1,19 @@ + + + + Range searches + Shinsuke Sugaya + + +
    +

    If in the field containing the data in specified range, such as the number range search is possible for that field.

    + +

    To limit "field name: value TO value ' fill in the search form.

    +

    For example, type to search document contentLength field against 1 k to 10 k bytes is shown below the search form.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-sort.xml b/src/site/en/xdoc/7.0/user/search-sort.xml new file mode 100644 index 000000000..d08fb7be8 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-sort.xml @@ -0,0 +1,54 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    + +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + + + + + + +
    Sort fields list
    Field nameDescription
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    +

    Adding custom fields as sort in Customizing.

    +
    + +

    You can select the sorting criteria when you search.

    + Sorting criteria +

    Also, for sorting in the search field sort: the field name to sort and field names separated by colon (:) fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/search-wildcard.xml b/src/site/en/xdoc/7.0/user/search-wildcard.xml new file mode 100644 index 000000000..753e5afa5 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/search-wildcard.xml @@ -0,0 +1,31 @@ + + + + Wildcard + Shinsuke Sugaya + + +
    +

    You can use one or multiple character wildcard search terms within. The can be specified as a one-character wildcard, * is specified as the multiple-character wildcard. Wildcards are not available in the first character. You can use wildcards for words. Wildcard search for the sentence.

    + +

    If you use one character wildcard shown below? The available.

    + +

    If the above is treated as a wildcard for one character, such as text or test.

    +

    If you use the multiple character wildcard use * below

    + +

    If the above is treated as a wildcard for multiple characters, such as test, tests or tester. Also,

    + +

    The so can be also used in the search term.

    +
    + +

    The wildcard string indexed using target. Therefore, because if the index has been created, such as bi-gram be treated meaning fixed string length in Japan Japanese wildcard in Japan, not expected behavior. Use in the field, if you use a wildcard in Japan, that used morphological analysis.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/7.0/user/xml-response.xml b/src/site/en/xdoc/7.0/user/xml-response.xml new file mode 100644 index 000000000..23bdd2be3 --- /dev/null +++ b/src/site/en/xdoc/7.0/user/xml-response.xml @@ -0,0 +1,207 @@ + + + + XML search results + Shinsuke Sugaya + + +
    +

    To emit the XML results of Fess. You need to do this output in XML Management screen crawl settings XML response enabled by you.

    + +

    To get the output XML http://localhost:8080/Fess/XML? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    +
    + +

    Returns the following response.

    + + + 1 + 0 + テスト + 0.64 + 20 + 1 + 101 + 6 + + + speedtest.goo.ne.jp/ + goo の提供するスピード<em>テスト</em>ページです + speedtest.goo.ne.jp + 1284739487873 + --> +<meta http-equiv="Content-Type" content="text/html; charset=EUC-JP"> +<title> +goo スピード<em>テスト</em> +</title> +<meta..."> + +<meta name="description" content="goo の提供するスピード<em>テスト</em>ページです"> +<meta name="keywords" content="ADSL,ブロードバンド,フレ...ッツ,BB,無線LAN,CATV,ISDN,スピード<em> +テスト</em>"> +<meta name="Copyright" content="Copyright &copy; ...://speedtest.goo.ne.jp/" +><img src="/img/h1/h1_76.gif" width="129" height="22" border="0" alt="スピード<em>テスト +</em>"><img src..." color="#ffffff">goo スピード<em>テスト</em></font +></a></b></td> +<!--/NAME SERVICE--> +</tr> +</table> +</td> +</tr> +<tr> +<td + 4.98744 + goo の提供するスピード<em>テスト</em>ページです + 1284739487887 + http://speedtest.goo.ne.jp/ + http://speedtest.goo.ne.jp/;type=au,docomo,pc,softbank + text/html + +goo スピードテスト + + +goo スピードテスト + + 1.0 + 17810 + http://speedtest.goo.ne.jp/ + +... + + +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. 0, the normal response.
    querySearch language.
    exec-timeResponse time. Unit is seconds.
    page-sizeThe number of display.
    page-numberPage number.
    record-countNumber of hits for the search term.
    page-countPage number of the number of hits for the search term.
    resultParent element of the search results.
    docThe elements of the search results.
    siteThe name of the site.
    content-descriptionA description of the content.
    hostThe name of the host.
    last-modifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    content-titleThe title of the document for viewing.
    content-lengthThe size of the document.
    URL-linkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/browserType-guide.xml b/src/site/en/xdoc/8.0/admin/browserType-guide.xml new file mode 100644 index 000000000..04cde963c --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/browserType-guide.xml @@ -0,0 +1,23 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
    +

    Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

    +
    +
    + +

    In Administrator account after logging in, click menu browser types.

    + Setting the browser type +
    +
    +
    + +

    You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/crawl-guide.xml b/src/site/en/xdoc/8.0/admin/crawl-guide.xml new file mode 100644 index 000000000..cec008f07 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/crawl-guide.xml @@ -0,0 +1,147 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to crawling.

    +
    +
    + +

    In Administrator account click crawl General menu after login.

    + Crawl General +
    +
    +
    + +

    You can set the interval at which the crawl for a Web site or file system. By default, the following.

    + +

    Figures are from left, seconds, minutes, during the day, month, represents a day of the week. Description format is similar to the Unix cron settings. For this example, am 0 時 0 分 in crawling daily.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    0 0 12 * *?Each day starts at 12 pm
    0 15 10? * *Day 10: 15 am start
    0 15 10 * *?Day 10: 15 am start
    0 15 10 * *? *Day 10: 15 am start
    0 15 10 * *? 2009Each of the 2009 start am, 10:15
    0 * 14 * *?Every day 2:00 in the PM-2: 59 pm start every 1 minute
    0 0 / 5 14 * *?Every day 2:00 in the PM-2: 59 pm start every 5 minutes
    0 0 / 5 14, 18 * *?Every day 2:00 pm-2: 59 pm and 6: 00 starts every 5 minutes at the PM-6: 59 pm
    0 0-5 14 * *?Every day 2:00 in the PM-2: 05 pm start every 1 minute
    0 10, 44 14? 3 WEDStarts Wednesday March 2: 10 and 2: 44 pm
    0 15 10? * MON-FRIMonday through Friday at 10:15 am start
    +

    Also check if the seconds can be set to run at intervals 60 seconds by default. If you set seconds exactly and you should customize webapps/fess/WEB-INF/classes/chronosCustomize.dicon taskScanIntervalTime value, if enough do I see in one-hour increments.

    +
    + +

    When the user enters a search, the search the output log. If you want to get search statistics to enable.

    +
    + +

    Save the information you find. Identifying the users becomes possible.

    +
    + +

    You can collect the search result was judged good by the user. Search result voting link appears to result in list screen, so that link press made the record. You can also reflect the results collected during the crawl index.

    +
    + +

    Search results link attaches to the search term. To display the find search terms in PDF becomes possible.

    +
    + +

    Search results can be retrieved in XML format. http://localhost:8080/Fess/XML? can get access query = search term.

    +
    + +

    Search results available in JSON format. http://localhost:8080/Fess/JSON? can get access query = search term.

    +
    + +

    Suggest candidates for search suggestions can be retrieved in XML or JSON format. If you want to get the words beginning with 'test' is the http://localhost:8080/fess/json? type = suggest

    +
    + +

    Morphological analysis of the results can be retrieved in XML or JSON format. If you want to apply morphological analysis on today's weather is sunny's http://localhost:8080/fess/json? type = analysis

    +
    + +

    If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

    +
    + +

    You can specify the label to see if the label by default,. Specifies the value of the label.

    +
    + +

    You can specify whether or not to display a search screen. If you select Web unusable for mobile search screen. If not available not available search screen. And if you want to create a dedicated index server and select not available.

    +
    + +

    In JSON format often find search words becomes available. can be retrieved by accessing the http://localhost:8080/Fess/hotsearchword.

    +
    + +

    Delete a session log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Delete a search log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Specifies the Bots name Bots you want to remove from the search log logs included in the user agent by commas (,). Log is deleted by log purge once a day.

    +
    + +

    Specifies the encoding for the CSV will be available in the backup and restore.

    +
    + +

    After the data is registered for Solr. Index to commit or to optimize the registered data becomes available. If optimize is issued the Solr index optimization, if you have chosen, you choose to commit the commit is issued.

    +
    + +

    Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

    +
    + +

    In Fess in 10 units send the document for Solr. For each value specified here Solr issued document commits. If 0 commit is performed after crawl completion.

    +
    + +

    Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

    +

    But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

    +
    + +

    You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

    +
    + +

    Registered disabled URL URL exceeds the failure count next time you crawl to crawl out. Does not need to monitor the fault type is being crawled next time by specifying this value.

    +
    + +

    Disaster URL exceeds the number of failures will crawl out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/8.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..1e833ceb8 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/crawlingSession-guide.xml @@ -0,0 +1,27 @@ + + + + Set session information + Shinsuke Sugaya + + +
    +

    Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

    +
    +
    + +

    In Administrator account after logging in, click the session information menu.

    +
    +
    +
    + + Session information list +

    You can remove all session information and click the Delete link all in the running. Session has expired will be removed at next crawl.

    +
    + +

    Sure you can crawl the contents of session ID. Crawl start and finish time, number of documents indexed and listed.

    + Session details +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/data-guide.xml b/src/site/en/xdoc/8.0/admin/data-guide.xml new file mode 100644 index 000000000..12b517dfd --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/data-guide.xml @@ -0,0 +1,41 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
    +

    Here, describes Fess information backup and restore methods.

    +
    +
    + +

    In Administrator account after logging in, click the menu backup and restore.

    + Backup and restore +
    + +

    Click the download link and Fess information output in XML format. Saved settings information is below.

    +
      +
    • The General crawl settings
    • +
    • Web crawl settings
    • +
    • File system Crawl settings
    • +
    • Datastore crawl settings
    • +
    • Label
    • +
    • Path mapping
    • +
    • Web authentication
    • +
    • File system authentication
    • +
    • Request header
    • +
    • Duplicate host
    • +
    • Roll
    • +
    • Compatible browsers
    • +
    +

    Session information, search log, click log is available in CSV format.

    +

    In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate. If you need to back up the SOLR index backs solr directory.

    +
    + +

    You can restore settings information, various log in to upload XML output by backup or CSV. To specify the files, please click the restore button on the data.

    +

    If enable overwrite data in XML file configuration information specified when the same data is updating existing data.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/dataCrawlingConfig-guide.xml b/src/site/en/xdoc/8.0/admin/dataCrawlingConfig-guide.xml new file mode 100644 index 000000000..5ee6f9634 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/dataCrawlingConfig-guide.xml @@ -0,0 +1,159 @@ + + + + Settings for crawling the data store + Sone, Takaaki + Shinsuke Sugaya + + +
    +

    You can crawl databases in Fess. Here are required to store settings.

    +
    +
    + +

    In Administrator account after logging in, click menu data store.

    + Data store configuration +

    As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

    + +

    Here the data is put something like the following.

    + +
    +
    +
    + +

    Parameter settings example looks like the following.

    + +

    Parameter is a "key = value" format. Description of the key is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    For DB configuration parameter example
    driverDriver class name
    URLURL
    usernameTo connect to the DB user name
    passwordTo connect to the DB password
    SQLWant to crawl to get SQL statement
    +
    + +

    Script configuration example looks like the following.

    + +

    + Parameter is a "key = value" format. + Description of the key is as follows.

    +

    + Side of the value written in OGNL. Close the string in double quotation marks. + Access in the database column name, its value.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Script settings
    URLURLs (links appear in search results)
    hostHost name
    siteSite pass
    titleTitle
    contentContent (string index)
    cacheContent cache (not indexed)
    DigestDigest piece that appears in the search results
    anchorLinks to content (not usually required)
    contentLengthThe length of the content
    lastModifiedContent last updated
    +
    + +

    To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

    +
    + +

    Set the following in the webapps/fess/WEB-INF/classes/app.dicon if you see the item value, such as latitude_s in the search results. After adding to $ {doc.latitude_s}, searchResults.jsp;

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "latitude_s","longitude_s" } + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/design-guide.xml b/src/site/en/xdoc/8.0/admin/design-guide.xml new file mode 100644 index 000000000..13c7045b6 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/design-guide.xml @@ -0,0 +1,84 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
    +

    Here are settings for the design of search screens.

    +
    +
    + +

    In Administrator account after logging in, click the menu design.

    + Design +

    You can edit the search screen in the screen below.

    + JSP compilation screen +
    + +

    If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

    + +]]> +

    tstampDate will crawl during registration on the lastModifiedDate modified date of the document. Output date formats follow the fmt:formateDate specification.

    +
    +
    +
    + +

    On the search screen files are available to download and can be removed.

    +
    + +

    You can upload the file to use in the search screen. Image file names are supported are jpg, gif, png, css, and js.

    +
    + +

    Use if you want to specify the file name file to upload. Uploaded if you omit the file name will be used.

    +
    + +

    You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    JSP file that you can edit
    Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
    Top page (within the Head tags)This is the express search home page head tag in JSP files. If you want to edit the meta tags, title tags, script tags, such as the change.
    Top page (content)Is a JSP file to represent the body tag in the search home page.
    Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
    Search results page (within the Head tags)Search result is a JSP file to represent within the head tag of the list page. If you want to edit the meta tags, title tags, script tags, such as the change.
    Search results page (header)Search result is a JSP file to represent the header of the list page. Include search form at the top.
    Search results page (footer)Search result is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
    Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    +

    You can to edit for PCs and similar portable screen.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/document-guide.xml b/src/site/en/xdoc/8.0/admin/document-guide.xml new file mode 100644 index 000000000..822c5391f --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/document-guide.xml @@ -0,0 +1,29 @@ + + + + Index settings + Shinsuke Sugaya + + +
    +

    Here are settings for the index.

    +
    +
    + +

    In Administrator account after logging in, click the menu index.

    + Index settings +
    +
    +
    + +

    You can publish index commit, optimize for server groups.

    +
    + +

    You can remove a specific search for the session ID. You can remove only the specific documents by specifying the URL.

    +
    + +

    Shown by the number of documents registered in each session. Can verify the results list by clicking the session name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/failureUrl-guide.xml b/src/site/en/xdoc/8.0/admin/failureUrl-guide.xml new file mode 100644 index 000000000..44db82f23 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/failureUrl-guide.xml @@ -0,0 +1,23 @@ + + + + Disaster URL + Shinsuke Sugaya + + +
    +

    Here the failure URL. URL could not be obtained at crawl time are recorded and confirmed as the failure URL.

    +
    +
    + +

    In Administrator account click menu disabled URL after login.

    + Disaster URL +

    Clicking the confirmation link failure URL displayed for more information.

    + Details of the disaster URL +
    + +

    A glance could not crawl the URL and date.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/favoriteLog-guide.xml b/src/site/en/xdoc/8.0/admin/favoriteLog-guide.xml new file mode 100644 index 000000000..37cf0aba0 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/favoriteLog-guide.xml @@ -0,0 +1,21 @@ + + + + Popular URL log + Shinsuke Sugaya + + +
    +

    Here are popular URL log. When the popular URL log user clicks voting link on the search screen registers as a favorite link. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click the menu popular URL.

    + Users list +
    + +

    Lists popular URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/fileAuthentication-guide.xml b/src/site/en/xdoc/8.0/admin/fileAuthentication-guide.xml new file mode 100644 index 000000000..3ae102fbb --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/fileAuthentication-guide.xml @@ -0,0 +1,44 @@ + + + + Settings for file system authentication + Shinsuke Sugaya + + +
    +

    Crawls using file system here, describes how to set file system authentication is required. Fess is corresponding to a crawl for a shared folder in Windows.

    +
    +
    + +

    In Administrator account after logging in, click the menu file system authentication.

    + File system settings +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Is omitted, the specified file system Kroll set applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. File system Crawl settings specified in that case applies on any port.

    +
    + +

    Select the authentication method. You can use SAMBA (Windows shared folder authentication).

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. SAMBA, the set value of the domain. If you want to write as.

    + +
    + +

    Select the set name to apply the authentication settings for the above file system CROLL. Must be registered ago you file system CROLL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/8.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..934b6538f --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,106 @@ + + + + Settings for file system crawling + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: or smb: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +

    Windows shared folder, for example, if you want to crawl to host1 share folder crawl settings for smb: (last / to) the //host1/share/. If authentication is in the shared folder on the file system authentication screen set authentication information.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + + +
    IP rings contents list
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    You can specify the crawl configuration information.

    +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/index.xml b/src/site/en/xdoc/8.0/admin/index.xml new file mode 100644 index 000000000..87a95df86 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/index.xml @@ -0,0 +1,67 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for the management work of the Fess. +

    +
    +
    +

    +This document demonstrates Fess configuration management methods. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    +
    + +

    Management UI to support browser as follows.

    +
      +
    • Internet Explorer 7 or higher
    • +
    • Firefox 4 or more
    • +
    • Chrome (latest version)
    • +
    • Safari (latest version)
    • +
    • Opera 11
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/labelType-guide.xml b/src/site/en/xdoc/8.0/admin/labelType-guide.xml new file mode 100644 index 000000000..c35728bf3 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/labelType-guide.xml @@ -0,0 +1,39 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. You can pass even if you do not set the crawl settings in the settings of the label to add labels to specify regular expressions. If you register the label shown select label drop-down box to the right of the search box.

    +
    +
    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    +
    +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    + +

    Sets the path to label in the regular expression. You can specify multiple in multiple line description. Notwithstanding the crawl configuration document to match the path specified here, will be labeled.

    +
    + +

    In the path and crawled on regular expressions set from what you want to exclude. You can specify multiple in multiple line description.

    +
    + +

    Specifies the role to view the label.

    +
    + +

    Specifies the order of the labels.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/log-guide.xml b/src/site/en/xdoc/8.0/admin/log-guide.xml new file mode 100644 index 000000000..b917b7325 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/log-guide.xml @@ -0,0 +1,21 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    +
    +
    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/8.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..35852a9d7 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/overlappingHost-guide.xml @@ -0,0 +1,27 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    +
    +
    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    +
    +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/8.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..d13c7532f --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/pathMapping-guide.xml @@ -0,0 +1,30 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    +
    +
    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    +
    +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/8.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..969df64aa --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/requestHeader-guide.xml @@ -0,0 +1,30 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    +
    +
    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    +
    +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/roleType-guide.xml b/src/site/en/xdoc/8.0/admin/roleType-guide.xml new file mode 100644 index 000000000..3de5358c0 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/roleType-guide.xml @@ -0,0 +1,27 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    +
    +
    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    +
    +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/searchList-guide.xml b/src/site/en/xdoc/8.0/admin/searchList-guide.xml new file mode 100644 index 000000000..f3f4182bf --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/searchList-guide.xml @@ -0,0 +1,21 @@ + + + + Search + Shinsuke Sugaya + + +
    +

    Here the search for management.

    +
    +
    + +

    In Administrator account after logging in, click the menu search.

    + Administrative search +
    + +

    You can search by criteria you specify. In the regular search screen role and browser requirements is added implicitly, but do not provide management for search. You can document a certain remove from index from the search results.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/searchLog-guide.xml b/src/site/en/xdoc/8.0/admin/searchLog-guide.xml new file mode 100644 index 000000000..3c354d506 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/searchLog-guide.xml @@ -0,0 +1,21 @@ + + + + Search log list + Shinsuke Sugaya + + +
    +

    Here the search log. When you search in the search screen users search logs are logged. Search log search term or date is recorded. You can also record the URL, then you want the search results to.

    +
    +
    + +

    In Administrator account after logging in, click menu search logs.

    + Search log +
    + +

    Search language and date are listed. You can review and detailed, you click the URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/stats-guide.xml b/src/site/en/xdoc/8.0/admin/stats-guide.xml new file mode 100644 index 000000000..7ef52a6cc --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/stats-guide.xml @@ -0,0 +1,21 @@ + + + + Statistics + Shinsuke Sugaya + + +
    +

    Here are statistics. You can search log and click log.

    +
    +
    + +

    In Administrator account after logging in, click the menu statistics.

    + Statistics +
    + +

    You can select the target by selecting the type of report, to make sure. Displayed in order by the specified criteria.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/system-guide.xml b/src/site/en/xdoc/8.0/admin/system-guide.xml new file mode 100644 index 000000000..5ce32bf8d --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/system-guide.xml @@ -0,0 +1,33 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here are registered in the server settings for crawling and Fess. SOLR servers are grouped by file, has been registered.

    +
    +
    + +

    In Administrator account after logging in, click menu system settings.

    + System settings +
    +
    +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and Fess server to shut down is not running when shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +

    You can manually crawling under the crawl start button press stop if it is that.

    +
    + +

    Server group name to search for and update available will be shown.

    +
    + +

    In Fess Solr Server conducts a management server and index State States. Whether or not the server state can be access to the Solr Server manages. Whether or not successfully crawl index the State could manage. You can use search server status is in effect, regardless of the State of the index. The crawl Server State is enabled and can index State runs correctly if the preparation or completion. Running start crawl manually index State preparing changes automatically. Server recovery server status and auto-recovery enabled state.

    +
    + +

    You can be sure SOLR server instance state. You can also, for each instance, start, stop, reload request.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/systemInfo-guide.xml b/src/site/en/xdoc/8.0/admin/systemInfo-guide.xml new file mode 100644 index 000000000..b0fd0710a --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/systemInfo-guide.xml @@ -0,0 +1,32 @@ + + + + System information + Shinsuke Sugaya + + +
    +

    Here, you can currently check property information such as system environment variables.

    +
    +
    + +

    In Administrator account after logging in, click system information menu.

    + System information +
    +
    +
    + +

    You can list the server environment variable.

    +
    + +

    You can list the system properties on Fess.

    +
    + +

    Fess setup information available.

    +
    + +

    Is a list of properties to attach when reporting a bug. Extract the value contains no personal information.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/userInfo-guide.xml b/src/site/en/xdoc/8.0/admin/userInfo-guide.xml new file mode 100644 index 000000000..5e7dbf30c --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/userInfo-guide.xml @@ -0,0 +1,21 @@ + + + + User log + Shinsuke Sugaya + + +
    +

    Here the user log. Identifies the user when you search in the search screen users the user log in. You can search log and popular URL information and the use. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click menu users.

    + Users list +
    + +

    Lists the ID of the user. You can select the search logs or popular URL links, to see a list of each log.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/8.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..17f72de7e --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/webAuthentication-guide.xml @@ -0,0 +1,48 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    +
    +
    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication, DIGEST authentication or NTLM authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. You can set the workstation and domain values for NTLM authentication. If you want to write as.

    + +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/8.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..9c6a4eec2 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,107 @@ + + + + Settings for crawling Web site + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + + +
    URL filtering contents list
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    You can specify the crawl configuration information.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl. If you do not specify people per 100,000.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in).

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/admin/wizard-guide.xml b/src/site/en/xdoc/8.0/admin/wizard-guide.xml new file mode 100644 index 000000000..c605f5546 --- /dev/null +++ b/src/site/en/xdoc/8.0/admin/wizard-guide.xml @@ -0,0 +1,35 @@ + + + + Settings Wizard + Sone, Takaaki + + +
    +

    Introduction to the Configuration Wizard.

    +

    You can use Settings Wizard, to set you up on the Fess.

    +
    +
    + +

    In Administrator account after logging in, click menu Settings Wizard.

    + Settings Wizard +

    First, setting a schedule. + During the time in Fess is crawling and indexes. + By default, every day is a 0 時 0 分. Schedules can change even the General crawl settings.

    + Setting a schedule +

    The crawl settings. + Crawl settings is to register a URI to look for. + The crawl settings name please put name of any easy to identify. Put the URI part de-indexed, want to search for.

    + Crawl settings +

    For example, if you want and search for http://fess.codelibs.org/, less looks like.

    + Crawl settings example +

    The type, such as c:\Users\taro file.

    +

    In this is the last setting. Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

    + Crawl started +
    + +

    Settings in the Setup Wizard you can change from crawl General, Web, file system.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/db.xml b/src/site/en/xdoc/8.0/config/db.xml new file mode 100644 index 000000000..2a62426a9 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/db.xml @@ -0,0 +1,128 @@ + + + + Database changes + Shinsuke Sugaya + + +
    +

    Provides binaries to use H2 Database with MySQL database. You can use the other database in to change the settings using the source code and build it.

    +
    +
    + +

    The MySQL character code setting. /etc/mysql/my.cnf and the added must have the following settings.

    + +
    + +

    Download MySQL binaries and expand.

    +
    + +

    Create a database.

    + create database fess_db; +mysql> grant all privileges on fess_db.* to fess_user@localhost identified by 'fess_pass'; +mysql> create database fess_robot; +mysql> grant all privileges on fess_robot.* to s2robot@localhost identified by 's2robot'; +mysql> FLUSH PRIVILEGES; +]]> +

    Create a table in the database. DDL file is located in extension/mysql.

    + +

    webapps/fess/WEB-INF/lib on mysql driver jar.

    + +

    Edit the webapps/fess/WEB-INF/classes/jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_db?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "fess_user" + "fess_pass" + + + + 600 + 10 + true + + + + + +]]> +

    Edit the webapps/fess/WEB-INF/classes/s2robot_jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_robot?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "s2robot" + "s2robot" + + + + 600 + 10 + true + + @java.sql.Connection@TRANSACTION_REPEATABLE_READ + + + + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/desktop-search.xml b/src/site/en/xdoc/8.0/config/desktop-search.xml new file mode 100644 index 000000000..5d72584e0 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/desktop-search.xml @@ -0,0 +1,37 @@ + + + + The desktop search settings + Shinsuke Sugaya + + +
    +

    + Increasing awareness of security in the browser environment in recent years, open a local file (for example, c:\hoge.txt) from the Web pages on. + Standard in Fess, open a file on a file system using the Java applet. + As a Java applet and another, offer desktop search functionality. + You can use desktop environment launches a Fess on a local PC, access to the file in the file system. + In the environment of the server and client desktop search not available.

    +
    +
    +

    + Desktop Search feature is turned off by default. + Please enable the following settings.

    +

    First of all, bin/setenv.bat as java.awt.headless from true to false edits.

    + +

    Then add the following to webapps/fess/WEB-INF/conf/crawler.properties.

    + +

    Start the Fess, after you set up above. How to use Basic remains especially.

    +
    +
    +
      +
    • Please Fess inaccessible from the outside, such as (for example, 8080 port does not release).
    • +
    • because false Java.awt.headless image size conversion for mobile devices is not available.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/filesize.xml b/src/site/en/xdoc/8.0/config/filesize.xml new file mode 100644 index 000000000..ff556b6fc --- /dev/null +++ b/src/site/en/xdoc/8.0/config/filesize.xml @@ -0,0 +1,29 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of defaultMaxLength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/geosearch.xml b/src/site/en/xdoc/8.0/config/geosearch.xml new file mode 100644 index 000000000..dfdb560eb --- /dev/null +++ b/src/site/en/xdoc/8.0/config/geosearch.xml @@ -0,0 +1,23 @@ + + + + Dios arch + Shinsuke Sugaya + + +
    +

    You can document with latitude and longitude location information in conjunction with Google maps, including the use of Dios arch.

    +
    +
    + +

    Location is defined as a feed that contains the location information. + When generating the index in Solr latitude longitude set to location feeds in formats such as 45.17614,-93.87341, register the document. + Also sets the value as the latitude_s and longitude_s fields if you want to display latitude and longitude as a search result. * _s is available as a dynamic field of Solr string.

    +
    + +

    During the search specifies in the request parameter to latitude and longitude, the distance. + View the results in the distance (km) specified by distance-based latitude information (latitude, longitude). Latitude and longitude and distances is treated as double.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/index-backup.xml b/src/site/en/xdoc/8.0/config/index-backup.xml new file mode 100644 index 000000000..541993b34 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index management + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data directory. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/index.xml b/src/site/en/xdoc/8.0/config/index.xml new file mode 100644 index 000000000..0ea320154 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for setting work of the Fess. +

    +
    +
    +

    +This document demonstrates setting Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/install-on-tomcat.xml b/src/site/en/xdoc/8.0/config/install-on-tomcat.xml new file mode 100644 index 000000000..314d28334 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/install-on-tomcat.xml @@ -0,0 +1,43 @@ + + + + Install to an existing Tomcat + Shinsuke Sugaya + + +
    +

    + The standard distribution of Fess Tomcat is distributed in the deployed State. + Because Fess is not dependent on Tomcat, deploying on any Java application server is available. + Describes how to deploy a Fess Tomcat here is already available.

    +
    +
    +

    Fess ServerHereFrom the download.

    +
    +
    +

    + Expand the downloaded Fess server. + Expanded Fess Server home directory to $FESS_HOME. + $TOMCAT_HOME the top directory of an existing Tomcat 6. + Copy the Fess Server data.

    + +

    + If you have, such as changing the destination file diff commands, updates your diff only applies. +

    +
    +
    +

    to start the Tomcat in normal as well as in startup.* / http://localhost:8080/Fess To access the.

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/logging.xml b/src/site/en/xdoc/8.0/config/logging.xml new file mode 100644 index 000000000..0c991dc92 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/logging.xml @@ -0,0 +1,52 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + + +
    Log file list
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina.outLog of the Fess Server (Tomcat). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    + +

    You can check by changing the logging level of the HttpClient if you want to check HTTP access when crawling content to solve the problem.

    + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/memory-config.xml b/src/site/en/xdoc/8.0/config/memory-config.xml new file mode 100644 index 000000000..ade9047e5 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/memory-config.xml @@ -0,0 +1,52 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    Set the maximum memory per process in Java. So, do not use the upper memory in the process also had 8 GB of physical memory on the server. Memory consumption depending on the number of crawl threads and interval will also change significantly. If not enough memory please change settings in the subsequent procedure.

    +
    +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case the maximum value set 1024M) will change to-Xmx1024m.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-Xmx1024m (in this case the maximum value set 1024M). +

    + new String[] { +"-Djava.awt.headless=true", +"-server", +"-Xmx1024m", +"-XX:MaxPermSize=128m", +"-XX:-UseGCOverheadLimit", +"-XX:+UseConcMarkSweepGC", +"-XX:CMSInitiatingOccupancyFraction=75", +"-XX:+CMSIncrementalMode", +"-XX:+CMSIncrementalPacing", +"-XX:CMSIncrementalDutyCycleMin=0", +"-XX:+UseParNewGC", +"-XX:+UseStringCache", +"-XX:+UseTLAB", +"-XX:+DisableExplicitGC" +} +]]> +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/mobile-device.xml b/src/site/en/xdoc/8.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/path-encoding.xml b/src/site/en/xdoc/8.0/config/path-encoding.xml new file mode 100644 index 000000000..b8aa16ac6 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    +
    +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/8.0/config/pdf-with-password.xml b/src/site/en/xdoc/8.0/config/pdf-with-password.xml new file mode 100644 index 000000000..0a8f4e317 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/pdf-with-password.xml @@ -0,0 +1,89 @@ + + + + Password-protected PDF + Shinsuke Sugaya + + +
    +

    + You should password files to register the settings file to PDF password is configured to search for. +

    +
    +
    +

    + First of all, create the webapps/fess/WEB-INF/classes/s2robot_extractor.dicon. + This is test _ ~ is a pass that password set to a.pdf file. + If you have multiple files, multiple settings in addPassword.

    + + + + + + + + + + + + + ".*test_.*.pdf" + + "pass" + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + +... +]]> +

    Next, edit the following to webapps/fess/WEB-INF/classes/s2robot_rule.dicon.

    + + "fsFileRule" + + + fessFileTransformer + + + true + + "url" + "file:.*" + + + "mimeType" + + + "(application/xml" ++ "|application/xhtml+xml" ++ "|application/rdf+xml" ++ "|application/pdf" ++ "|text/xml" ++ "|text/xml-external-parsed-entity" ++ "|text/html)" + + +
    +... +]]> +

    Fess up after setting the above and please run the crawl. How to use Basic remains especially.

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/protwords.xml b/src/site/en/xdoc/8.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/proxy.xml b/src/site/en/xdoc/8.0/config/proxy.xml new file mode 100644 index 000000000..2128a7058 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/proxy.xml @@ -0,0 +1,120 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    +
    +

    Proxy is set in to create webapps/Fess/Web-INF/classes/s9robot_client.dicon with the following contents.

    + + + + + + + + @org.apache.http.client.params.CookiePolicy@BEST_MATCH + clientConnectionManager + "プロキシサーバー名(ex. 192.168.1.1)" + プロキシサーバーのポート(ex. 8080) (" は不要) + + + + internalHttpClient + 5 + 500 + + + + + + + + + "http" + 80 + @org.apache.http.conn.scheme.PlainSocketFactory@getSocketFactory() + + + + + + + "https" + 443 + @org.apache.http.conn.ssl.SSLSocketFactory@getSocketFactory() + + + + + + 5 + @java.util.concurrent.TimeUnit@MINUTES + + 200 + + 20 + + + + + "UTF-8" + + + internalFsClient + 5 + 500 + + + + "UTF-8" + + + + internalSmbClient + 5 + 500 + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + "smb:.*" + smbClient + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/8.0/config/replication.xml b/src/site/en/xdoc/8.0/config/replication.xml new file mode 100644 index 000000000..0cbf69304 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/replication.xml @@ -0,0 +1,56 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Sets the replication of the index Solr replication features. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +
    +
    + +

    Fess, download and install the. When you install MasterServer named host./ /opt/fess_master To assume you installed. Edit the SOLR/core1/conf/solrconfig.XML like the following.

    + + + startup + optimize + optimize + schema.xml,stopwords.txt,stopwords_ja.txt,elevate.xml, + stoptags_ja.txt,synonyms.txt,mapping_ja.txt,mapping-FoldToASCII.txt, + mapping-ISOLatin1Accent.txt,protwords.txt,compositePOS.txt,spellings.txt, + currency.xml + + 1 + +... +]]> +

    Register the crawl settings as well as Fess starts after the normal construction. Steps to building the index for Fess remains especially as normal building procedures.

    +
    + +

    Fess, download and install the./ /opt/fess_slave To assume you installed. Edit the SOLR/core1/conf/solrconfig.XML like the following.

    + + + http://MasterServer:8080/solr/core1/replication + 00:00:60 + internal + 5000 + 10000 + solradmin + solradmin + + +... +]]> +

    Start the Fess.

    +
    + +

    Until the above settings copied to search for Fess index crawl after the optimization (optimize) the index for Fess.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/role-setting.xml b/src/site/en/xdoc/8.0/config/role-setting.xml new file mode 100644 index 000000000..3c973b9cc --- /dev/null +++ b/src/site/en/xdoc/8.0/config/role-setting.xml @@ -0,0 +1,85 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    +
      +
    • Request parameter
    • +
    • Request header
    • +
    • Cookies
    • +
    • J2EE authentication information
    • +
    +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + + {"guest"} + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/server-port.xml b/src/site/en/xdoc/8.0/config/server-port.xml new file mode 100644 index 000000000..964ef1837 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/server-port.xml @@ -0,0 +1,36 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    • 19092: database port (use h2database)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information.

    +

    change the webapps/fess/WEB-INF/classes/app.dicon the following points.

    + "http://localhost:8080/manager/text/" +]]> +

    change the webapps/fess/WEB-INF/classes/solrlib.dicon the following points.

    + "http://localhost:8080/solr/core1" +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/solr-dynamic-field.xml b/src/site/en/xdoc/8.0/config/solr-dynamic-field.xml new file mode 100644 index 000000000..2b5a2ddef --- /dev/null +++ b/src/site/en/xdoc/8.0/config/solr-dynamic-field.xml @@ -0,0 +1,56 @@ + + + + Dynamic fields + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in solr/core1/conf/schema.xml. dynamic fields and standard fields such as title and content can be freely defined field names are defined. The dynamic fields that are available in the schema.xml Fess become. Advanced parameter values see a Solr document.

    + + + + + + + + + + + + + + + + + + + + + + +]]> +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add a field to use to retrieve data that is stored in the dynamic field next to the webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"url", "host", "site", + "title", "content", "contentLength", "lastModified", "mimetype", + "label", "segment", "other_t" } + +]]> +

    You must add a field to use to retrieve data that is stored in the dynamic field from Solr. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/solr-failure.xml b/src/site/en/xdoc/8.0/config/solr-failure.xml new file mode 100644 index 000000000..3a8a9dc2d --- /dev/null +++ b/src/site/en/xdoc/8.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/solrlib.dicon.

    +
    +
    +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is Minctiveserver following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status Mxretrysttuscheckcount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    +
    +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    +
    +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/tokenizer.xml b/src/site/en/xdoc/8.0/config/tokenizer.xml new file mode 100644 index 000000000..b3d65c8e8 --- /dev/null +++ b/src/site/en/xdoc/8.0/config/tokenizer.xml @@ -0,0 +1,47 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search. Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits. For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit. For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default CJKTokenizer used.

    + + +

    CJKTokenizer index bi-gram, in other words two characters to like Japan Japanese multibyte string creates. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to. However, please note that the index size increases.

    +

    The following example to change the analyzer part like solr/core1/conf/schema.xml, you can use the StandardTokenizer.

    + + + + + + + : + + + + + : +]]> +

    Also, useBigram is enabled by default in the webapps/fess/WEB-INF/classes/app.dicon change to false.

    + + true + : +]]> +

    After the restart the Fess.

    +
    +
    + + +
    diff --git a/src/site/en/xdoc/8.0/config/use-libreoffice.xml b/src/site/en/xdoc/8.0/config/use-libreoffice.xml new file mode 100644 index 000000000..edb25c54f --- /dev/null +++ b/src/site/en/xdoc/8.0/config/use-libreoffice.xml @@ -0,0 +1,85 @@ + + + + Use of LibreOffice + Shinsuke Sugaya + + +
    +

    + It is possible to crawl using the Apache POI Fess environmental standard in MS Office system document. + You can crawl Office system document regarding LibreOffice, OpenOffice, do even more accurate text extraction from documents.

    +
    +
    +

    JodConverter Fess server install. from http://jodconverter.googlecode.com/jodconverter-core-3.0-Beta-4-Dist.zipThe download. Expand and copy the jar file to Fess server.

    + +

    Create a s2robot_extractor.dicon to the next.

    + +

    s2robot_extractor.dicon effective jodExtractor with following contents.

    + + + + + + + + + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + + officeManagerConfiguration.setOfficeHome("/usr/lib/libreoffice") + .buildOfficeManager() + + + + + + { +"application/msword", +"application/vnd.ms-excel", +"application/vnd.ms-powerpoint", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.openxmlformats-officedocument.presentationml.presentation" + } + jodExtractor + +... +]]> +

    Index to generate the settings later, usually crawled into the street.

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/config/windows-service.xml b/src/site/en/xdoc/8.0/config/windows-service.xml new file mode 100644 index 000000000..5405ec93a --- /dev/null +++ b/src/site/en/xdoc/8.0/config/windows-service.xml @@ -0,0 +1,54 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as Add %JAVA_HOME%\bin to PathYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option.

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-8.0.0.

    + cd C:\Java\fess-server-8.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat7w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    +
    + +

    Distributed in the Fess is 64-bit binaries for Windows Tomcat builds based on. If you use 32-bit WindowsTomcat Of the site from, such as 32-bit Windows zip, please replace tomcat7.exe, tomcat7w.exe, tcnative-1.dll.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/install/index.xml b/src/site/en/xdoc/8.0/install/index.xml new file mode 100644 index 000000000..6b2adcd9b --- /dev/null +++ b/src/site/en/xdoc/8.0/install/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for Fess installation and uninstallation tasks. +

    +
    +
    +

    +This document shows how to install Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/8.0/install/install.xml b/src/site/en/xdoc/8.0/install/install.xml new file mode 100644 index 000000000..1f260246e --- /dev/null +++ b/src/site/en/xdoc/8.0/install/install.xml @@ -0,0 +1,106 @@ + + + + Installation + + + + + +
    +

    +Fess can be available in these environments. +

    +
      +
    • OS: Windows, Unix, and Java can run OS environment
    • +
    • Java: Java 7 or higher (Java 6 can operate in)
    • +
    +

    +If Java is not installed in the environment we want to Fess from http://java.sun.com/ get JDK for Java 7 and above, please install. +

    +
    +
    +

    +http://sourceforge.jp/projects/Fess/releases/ From the download the latest Fess package. +

    +
    +
    +

    +Expand the downloaded fess-server-x.y.zip. +If you installed in the UNIX environment, in the bin added the performing rights to a script.

    + +
    +
    +

    +Administrator account is managed by the application server. Fess Server standard available Tomcat, as well as to the user changing the Tomcat. +Modify the password for the admin account of the conf/tomcat-user.xml if you want to change.

    + +]]> +

    +see the Tomcat documentation or JAAS authentication specification to use Tomcat-user.XML file management method other than. +

    +
    +
    +

    +To access the Solr into Fess server is password becomes necessary. +Change the default passwords in production, etc.

    +

    +How to change the password, you must first change the password attribute of the conf/tomcat-user.xml solradmin. +

    + +]]> +

    +Describes the provided password webapps/fess/WEB-INF/classes/solrlib.dicon the following points tomcat-user.xml to the next. +

    + + "solradmin" + "solradmin" + +]]> +
    +
    +

    +To manage can manage Solr from Fess server deployed on Tomcat context need a password. +Change the default passwords in production, etc.

    +

    +How to change password the change manager conf/tomcat-user.xml password attributes. +

    + + +]]> +

    +Describes the provided password webapps/fess/WEB-INF/classes/app.dicon the following points tomcat-user.xml to the next. +

    + + + "solrServer1" + "http://localhost:8080/manager/text/" + "/solr" + "manager" + "manager" + +]]> +
    +
    +

    +Login at return path settings, in the use encryption and decryption. +Please change the default password, especially in production.

    +

    +How to change the change the key value in the webapps/fess/WEB-INF/classes/app.dicon. +Set the 16-character alphanumeric.

    + +"1234567890123456" +]]> +
    + +
    diff --git a/src/site/en/xdoc/8.0/install/run.xml b/src/site/en/xdoc/8.0/install/run.xml new file mode 100644 index 000000000..ba5a8b3b2 --- /dev/null +++ b/src/site/en/xdoc/8.0/install/run.xml @@ -0,0 +1,41 @@ + + + + Start and stop + + + + + +
    +

    +to run the startup script, start the Fess. +

    + + +

    +access to / http://localhost:8080/Fess ensures startup. +

    + +

    +Management UI is / http://localhost:8080/fess/admin. +Default Administrator account user name / password is admin/admin. +Administrator account is managed by the application server. +In the management UI of the Fess, authenticate with the application server in fess role available as an administrator.

    +
    +
    +

    +Fess to stop the running shutdown scripts. +

    + + +

    +Crawl or may take a while to completely stop during the index creation if you. +

    +
    + +
    diff --git a/src/site/en/xdoc/8.0/install/uninstall.xml b/src/site/en/xdoc/8.0/install/uninstall.xml new file mode 100644 index 000000000..6519561f9 --- /dev/null +++ b/src/site/en/xdoc/8.0/install/uninstall.xml @@ -0,0 +1,19 @@ + + + + Uninstall + + + + + +
    +

    +You can uninstall by deleting the installation stops after the Fess Fess-Server-x,y directory. +

    + +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/index.xml b/src/site/en/xdoc/8.0/user/index.xml new file mode 100644 index 000000000..52bdaa14c --- /dev/null +++ b/src/site/en/xdoc/8.0/user/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is targeted to users to Fess. +

    +
    +
    +

    +This document demonstrates how to search in the Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/json-response.xml b/src/site/en/xdoc/8.0/user/json-response.xml new file mode 100644 index 000000000..2f7b3b357 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/json-response.xml @@ -0,0 +1,184 @@ + + + + Search results for JSON + Shinsuke Sugaya + + +
    +

    Outputs a JSON results of Fess. You must in order to emit JSON enabled keep JSON response in the General Management screen crawl settings.

    + +

    To get JSON output http://localhost:8080/Fess/JSON? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    callbackIf you use the JSONP callback name. You do not need to specify if you do not use the JSONP.
    +
    + +

    Returns the following response.

    + \n..", + "tstamp": "1284739487887", + "url": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F", + "id": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F;type=au,docomo,pc,softbank", + "mimetype": "text\u002Fhtml", + "title": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "contentTitle": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "boost": "1.0", + "contentLength": "17810", + "urlLink": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F" + }, +... + ] + } +} +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. the status value is 0: normal, 1: search error, 2 or 3: request parameter error, 9: in service, 1: API type of error is.
    querySearch language.
    execTimeResponse time. Unit is seconds.
    pageSizeThe number of display.
    pageNumberPage number.
    recordCountNumber of hits for the search term.
    pageCountPage number of the number of hits for the search term.
    resultParent element of the search results.
    siteThe name of the site.
    contentDescriptionA description of the content.
    hostThe name of the host.
    lastModifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    contentTitleThe title of the document for viewing.
    contentLengthThe size of the document.
    urlLinkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-additional.xml b/src/site/en/xdoc/8.0/user/search-additional.xml new file mode 100644 index 000000000..ae0c2aa61 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-additional.xml @@ -0,0 +1,15 @@ + + + + Hide search criteria + Shinsuke Sugaya + + +
    +

    You can use the additional parameters if the search string is shown on the screen without the specific search criteria like persuasion. additional value is retained but in the paging screen additional value.

    + +

    Without the conditions show screen and run a search when searches are performed by appending additional values in hidden forms, such as (for example, a search form) in the paging screen transitions and also the condition holds.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-and.xml b/src/site/en/xdoc/8.0/user/search-and.xml new file mode 100644 index 000000000..f20ac20a5 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-and.xml @@ -0,0 +1,20 @@ + + + + AND search + Shinsuke Sugaya + + +
    +

    Use the search if you want to search for documents that contain all search words of more than one. When describing multiple words in the search box separated by spaces, AND skip AND search.

    + +

    If you use the search search words written AND. Write in capital letters AND the space required to back and forth. AND is possible can be omitted.

    +

    For example, if you want to find documents that contain the search terms 1 and 2 search terms, type the following search form.

    + +

    Connecting multiple words with AND are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-boost.xml b/src/site/en/xdoc/8.0/user/search-boost.xml new file mode 100644 index 000000000..697a4bd94 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-boost.xml @@ -0,0 +1,20 @@ + + + + Boost search + Shinsuke Sugaya + + +
    +

    Use the boost search if you want to prioritize, search for specific search terms. Enabling search in boost search, depending on the severity of the search words.

    + +

    To boost search after the search term ' ^ boost value "that specifies the boost value (weighted) in the format.

    +

    For example, if you want to find the page if you want to find apples oranges contained more 'apples', type the following search form.

    + +

    Boost value specifies an integer greater than 1.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-field.xml b/src/site/en/xdoc/8.0/user/search-field.xml new file mode 100644 index 000000000..3c668053c --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-field.xml @@ -0,0 +1,66 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    You crawl in Fess results are saved for each field, such as title and full text. You can search for a field of them. You can specify the search criteria in search for a field, such as document type or size small.

    + +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Field list is available
    Field nameDescription
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields subject to find the content. Fields are custom fields are also available by using the dynamic field of Solr.

    +

    If HTML file and search for the title tag that string in the title field, below the body tag registered in the body field.

    +
    + +

    If a field specifying the search field name: search words in separated by colons (:) field name and search word fill in the search form, the search.

    +

    If you search the Fess as a search term for the title field, type.

    + +

    Document, the title field in Fess above search appears as a search result.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-fuzzy.xml b/src/site/en/xdoc/8.0/user/search-fuzzy.xml new file mode 100644 index 000000000..30c6ba212 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-fuzzy.xml @@ -0,0 +1,21 @@ + + + + Fuzzy search (fuzzy search) + Shinsuke Sugaya + + +
    +

    Ambiguity in the case does not match the words word search to search is available. Based on the Levenshtein distance in Fess ambiguous corresponds to the search (fuzzy search).

    + +

    After the search word you want to apply the fuzzy search adds '~'.

    +

    For example, ambiguous word "Solr", you can find documents that contain the word, near the "Solr" If you want to find, type as the search form, such as ("Solar"). +

    + +

    Furthermore, if by '~' after the number 0 and 1, 1 just like in refine. For example, in the form of 'Solr~0.8'. Do not specify numeric default value is 0.5.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-geo.xml b/src/site/en/xdoc/8.0/user/search-geo.xml new file mode 100644 index 000000000..8dfb4282a --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-geo.xml @@ -0,0 +1,32 @@ + + + + Dios arch + Shinsuke Sugaya + + +
    +

    Search using location information when you search, adding latitude and longitude location information for each document when generating the index becomes possible.

    + +

    Following parameters is available in the standard.

    + + + + + + + + + + + + + + + + +
    Request parameter
    GEO.latitudeLatitude degrees minutes seconds specifies double.
    GEO.longitudeLongitude degrees minutes seconds specifies double.
    GEO.distanceSpecifies the distance between documents in units of the coast. Specified in the double type.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-label.xml b/src/site/en/xdoc/8.0/user/search-label.xml new file mode 100644 index 000000000..8018dfeb5 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-label.xml @@ -0,0 +1,17 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    Narrow your search by adding the categories to search the document for label information the label is specified when the search is possible. Label information by registering in the Administration screen, will enable search by labels in the search screen. Label information available can multiple selections in the drop-down when you search. If you do not register the label displayed the label drop-down box.

    + +

    You can select the label information at search time. Label information can be selected in the search options dialog by pressing the options button will appear.

    + Label search +

    You can search each document to create an index, set the label to the label. All results search search do not specify a label is usually the same. If you change the label information to update the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-not.xml b/src/site/en/xdoc/8.0/user/search-not.xml new file mode 100644 index 000000000..3a40c9f2b --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-not.xml @@ -0,0 +1,19 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If the document does not contain a word search NOT searching is available.

    + +

    Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, enter the following if you want to find documents that contain the search term 1 does not contain a search term 2 and the search.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-options.xml b/src/site/en/xdoc/8.0/user/search-options.xml new file mode 100644 index 000000000..85132468e --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-options.xml @@ -0,0 +1,37 @@ + + + + Extended parameter search + Shinsuke Sugaya + + +
    +

    You can pass any search criteria from third-party search engines move and easy to like. Pass search criteria Please implement processing in QueryHelperImpl #buildOptionQuery.

    + +

    Following parameters is available in the standard.

    + + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    options.qThis is similar to the normal query. You can specify multiple options.q. If you specify multiple is treated as a search. Pass the URL encoding.
    options.CQTreated as exact match search queries. For example, if you specify the Fess Project searches as "Fess Project". Pass the URL encoding.
    options.OQIs treated as an OR search. For example, if you specify the Fess Project search as a Fess OR Project. Pass the URL encoding.
    options.NQThe label value. Use to specify the label.Treated as NOT search. For example, if you specify 'Fess' search as NOT Fess. Pass the URL encoding.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-or.xml b/src/site/en/xdoc/8.0/user/search-or.xml new file mode 100644 index 000000000..6b3ed55f1 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-or.xml @@ -0,0 +1,20 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. When describing the multiple words in the search box, by default will search.

    + +

    To use search OR search words written OR. OR write in capital letters, the space required to back and forth.

    +

    For example, if you want to search for documents that contain either search term 2 search term 1 and type the following search form.

    + +

    OR between multiple languages are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-range.xml b/src/site/en/xdoc/8.0/user/search-range.xml new file mode 100644 index 000000000..84202b944 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-range.xml @@ -0,0 +1,19 @@ + + + + Range searches + Shinsuke Sugaya + + +
    +

    If in the field containing the data in specified range, such as the number range search is possible for that field.

    + +

    To limit "field name: value TO value ' fill in the search form.

    +

    For example, type to search document contentLength field against 1 k to 10 k bytes is shown below the search form.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-sort.xml b/src/site/en/xdoc/8.0/user/search-sort.xml new file mode 100644 index 000000000..b1e22b065 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-sort.xml @@ -0,0 +1,54 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    + +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + + + + + + +
    Sort fields list
    Field nameDescription
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    +

    Adding custom fields as sort in Customizing.

    +
    + +

    You can select the sorting criteria when you search. Sorting criteria can be selected in the search options dialog by pressing the options button will appear.

    + Sorting criteria +

    Also, for sorting in the search field sort: the field name to sort and field names separated by colon (:) fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/search-wildcard.xml b/src/site/en/xdoc/8.0/user/search-wildcard.xml new file mode 100644 index 000000000..753e5afa5 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/search-wildcard.xml @@ -0,0 +1,31 @@ + + + + Wildcard + Shinsuke Sugaya + + +
    +

    You can use one or multiple character wildcard search terms within. The can be specified as a one-character wildcard, * is specified as the multiple-character wildcard. Wildcards are not available in the first character. You can use wildcards for words. Wildcard search for the sentence.

    + +

    If you use one character wildcard shown below? The available.

    + +

    If the above is treated as a wildcard for one character, such as text or test.

    +

    If you use the multiple character wildcard use * below

    + +

    If the above is treated as a wildcard for multiple characters, such as test, tests or tester. Also,

    + +

    The so can be also used in the search term.

    +
    + +

    The wildcard string indexed using target. Therefore, because if the index has been created, such as bi-gram be treated meaning fixed string length in Japan Japanese wildcard in Japan, not expected behavior. Use in the field, if you use a wildcard in Japan, that used morphological analysis.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/8.0/user/xml-response.xml b/src/site/en/xdoc/8.0/user/xml-response.xml new file mode 100644 index 000000000..14e337d00 --- /dev/null +++ b/src/site/en/xdoc/8.0/user/xml-response.xml @@ -0,0 +1,212 @@ + + + + XML search results + Shinsuke Sugaya + + +
    +

    To emit the XML results of Fess. You need to do this output in XML Management screen crawl settings XML response enabled by you.

    + +

    To get the output XML http://localhost:8080/Fess/XML? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    +
    + +

    Returns the following response.

    + + + 1 + 0 + テスト + 0.64 + 20 + 1 + 101 + 6 + + + speedtest.goo.ne.jp/ + goo の提供するスピード<em>テスト</em> + ページです + speedtest.goo.ne.jp + 1284739487873 + --> +<meta http-equiv="Content-Type" content="text/html; charset=EUC-JP"> +<title> +goo スピード<em>テスト</em> +</title> +<meta..."> + +<meta name="description" content="goo の提供するスピード<em>テスト +</em>ページです"> +<meta name="keywords" content="ADSL,ブロードバンド,フレ...ッツ,BB,無線LAN, +CATV,ISDN,スピード<em> +テスト</em>"> +<meta name="Copyright" content="Copyright &copy; ...://speedtest.goo.ne.jp/" +><img src="/img/h1/h1_76.gif" width="129" height="22" border="0" +alt="スピード<em>テスト +</em>"><img src..." color="#ffffff">goo スピード<em>テスト +</em></font +></a></b></td> +<!--/NAME SERVICE--> +</tr> +</table> +</td> +</tr> +<tr> +<td + 4.98744 + goo の提供するスピード<em>テスト</em>ページです + 1284739487887 + http://speedtest.goo.ne.jp/ + http://speedtest.goo.ne.jp/;type=au,docomo,pc,softbank + text/html + +goo スピードテスト + + +goo スピードテスト + + 1.0 + 17810 + http://speedtest.goo.ne.jp/ + +... + + +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. the status value is 0: normal, 1: search error, 2 or 3: request parameter error, 9: in service, 1: API type of error is.
    querySearch language.
    exec-timeResponse time. Unit is seconds.
    page-sizeThe number of display.
    page-numberPage number.
    record-countNumber of hits for the search term.
    page-countPage number of the number of hits for the search term.
    resultParent element of the search results.
    docThe elements of the search results.
    siteThe name of the site.
    content-descriptionA description of the content.
    hostThe name of the host.
    last-modifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    content-titleThe title of the document for viewing.
    content-lengthThe size of the document.
    URL-linkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/browserType-guide.xml b/src/site/en/xdoc/9.0/admin/browserType-guide.xml new file mode 100644 index 000000000..a5d112263 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/browserType-guide.xml @@ -0,0 +1,23 @@ + + + + Setting the browser type + Shinsuke Sugaya + + +
    +

    Describes the settings related to the browser type. Search results are browser type can be added to the data, for each type of browser browsing search results out into.

    +
    +
    + +

    In Administrator account after logging in, click menu browser types.

    + Setting the browser type +
    +
    +
    + +

    You can set the display name and value. It is used if you want more new terminals. You do not need special customizations are used only where necessary.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/crawl-guide.xml b/src/site/en/xdoc/9.0/admin/crawl-guide.xml new file mode 100644 index 000000000..eaa445843 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/crawl-guide.xml @@ -0,0 +1,93 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to crawling.

    +
    +
    + +

    In Administrator account click crawl General menu after login.

    + Crawl General +
    +
    +
    + +

    When the user enters a search, the search the output log. If you want to get search statistics to enable.

    +
    + +

    Save the information you find. Identifying the users becomes possible.

    +
    + +

    You can collect the search result was judged good by the user. Search result voting link appears to result in list screen, so that link press made the record. You can also reflect the results collected during the crawl index.

    +
    + +

    Search results link attaches to the search term. To display the find search terms in PDF becomes possible.

    +
    + +

    Search results can be retrieved in XML format. http://localhost:8080/Fess/XML? can get access query = search term.

    +
    + +

    Search results available in JSON format. http://localhost:8080/Fess/JSON? can get access query = search term.

    +
    + +

    If theses PC website search results on mobile devices may not display correctly. And select the mobile conversion, such as if the PC site for mobile terminals, and to show that you can. You can if you choose Google Google Wireless Transcoder allows to display content on mobile phones. For example, if site for PC and mobile devices browsing the results in the search for mobile terminals search results will link in the search result link passes the Google Wireless Transcoder. You can use smooth mobile transformation in mobile search.

    +
    + +

    You can specify the label to see if the label by default,. Specifies the value of the label.

    +
    + +

    You can specify whether or not to display a search screen. If you select Web unusable for mobile search screen. If not available not available search screen. And if you want to create a dedicated index server and select not available.

    +
    + +

    In JSON format often find search words becomes available. http://localhost:8080/Fess/JSON? can be retrieved by accessing the type = hotsearchword.

    +
    + +

    Delete a search log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Delete the job days before the specified date. One day in the one log purge old log is deleted.

    +
    + +

    Delete the user information for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Specifies the Bots name Bots you want to remove from the search log logs included in the user agent by commas (,). Log is deleted by log purge once a day.

    +
    + +

    Specifies the email address to send information about crawl upon completion crawl.

    +
    + +

    Specifies the encoding for the CSV will be available in the backup and restore.

    +
    + +

    Crawl as been updated to enable incremental crawl compared lastModified field value and the target document's modification date (if the HTTP's timestamp if LAST_MODIFIED values, file).

    +
    + +

    File additional group access rights information added to the role.

    +
    + +

    Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

    +
    + +

    In Fess in 10 units send the document for Solr. For each value specified here Solr issued document commits. If 0 commit is performed after crawl completion.

    +
    + +

    Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

    +

    But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

    +
    + +

    You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

    +
    + +

    Registered disabled URL URL exceeds the failure count next time you crawl to crawl out. Does not need to monitor the fault type is being crawled next time by specifying this value.

    +
    + +

    Disaster URL exceeds the number of failures will crawl out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/crawlingSession-guide.xml b/src/site/en/xdoc/9.0/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..d1658c22c --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/crawlingSession-guide.xml @@ -0,0 +1,27 @@ + + + + Set session information + Shinsuke Sugaya + + +
    +

    Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

    +
    +
    + +

    In Administrator account after logging in, click the session information menu.

    +
    +
    +
    + + Session information list +

    You can remove all session information and click the Delete link all in the running. Session has expired will be removed at next crawl.

    +
    + +

    Sure you can crawl the contents of session ID. Crawl start and finish time, number of documents indexed and listed.

    + Session details +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/data-guide.xml b/src/site/en/xdoc/9.0/admin/data-guide.xml new file mode 100644 index 000000000..3b0e3eee8 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/data-guide.xml @@ -0,0 +1,41 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
    +

    Here, describes Fess information backup and restore methods.

    +
    +
    + +

    In Administrator account after logging in, click the menu backup and restore.

    + Backup and restore +
    + +

    Click the download link and Fess information output in XML format. Saved settings information is below.

    +
      +
    • The General crawl settings
    • +
    • Web crawl settings
    • +
    • File system Crawl settings
    • +
    • Datastore crawl settings
    • +
    • Label
    • +
    • Path mapping
    • +
    • Web authentication
    • +
    • File system authentication
    • +
    • Request header
    • +
    • Duplicate host
    • +
    • Roll
    • +
    • Compatible browsers
    • +
    +

    Session information, search log, click log is available in CSV format.

    +

    In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate. If you need to back up the SOLR index backs solr directory.

    +
    + +

    You can restore settings information, various log in to upload XML output by backup or CSV. To specify the files, please click the restore button on the data.

    +

    If enable overwrite data in XML file configuration information specified when the same data is updating existing data.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/dataCrawlingConfig-guide.xml b/src/site/en/xdoc/9.0/admin/dataCrawlingConfig-guide.xml new file mode 100644 index 000000000..32fd2d487 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/dataCrawlingConfig-guide.xml @@ -0,0 +1,159 @@ + + + + Settings for crawling the data store + Sone, Takaaki + Shinsuke Sugaya + + +
    +

    You can crawl data sources such as databases and CSV in Fess. Here are required to store settings.

    +
    +
    + +

    In Administrator account after logging in, click menu data store.

    + Data store configuration +

    As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

    + +

    Here the data is put something like the following.

    + +
    +
    +
    + +

    Parameter settings example looks like the following.

    + +

    Parameter is a "key = value" format. Description of the key is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    For DB configuration parameter example
    driverDriver class name
    URLURL
    usernameTo connect to the DB user name
    passwordTo connect to the DB password
    SQLWant to crawl to get SQL statement
    +
    + +

    Script configuration example looks like the following.

    + +

    + Parameter is a "key = value" format. + Description of the key is as follows.

    +

    + Side of the value written in OGNL. Close the string in double quotation marks. + Access in the database column name, its value.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Script settings
    URLURLs (links appear in search results)
    hostHost name
    siteSite pass
    titleTitle
    contentContent (string index)
    cacheContent cache (not indexed)
    DigestDigest piece that appears in the search results
    anchorLinks to content (not usually required)
    contentLengthThe length of the content
    lastModifiedContent last updated
    +
    + +

    To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

    +
    + +

    Set the following in the webapps/fess/WEB-INF/classes/app.dicon if you see the item value, such as latitude_s in the search results. After adding to $ {doc.latitude_s}, searchResults.jsp;

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "latitude_s","longitude_s" } + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/design-guide.xml b/src/site/en/xdoc/9.0/admin/design-guide.xml new file mode 100644 index 000000000..663a1bad1 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/design-guide.xml @@ -0,0 +1,144 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
    +

    Here are settings for the design of search screens.

    +
    +
    + +

    In Administrator account after logging in, click the menu design.

    + Design +

    You can edit the search screen in the screen below.

    + JSP compilation screen +
    + +

    If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

    + +]]> +

    tstampDate will crawl during registration on the lastModifiedDate modified date of the document. Output date formats follow the fmt:formateDate specification.

    +
    +
    +
    + +

    On the search screen files are available to download and can be removed.

    +
    + +

    You can upload the file to use in the search screen. Image file names are supported are jpg, gif, png, css, and js.

    +
    + +

    Use if you want to specify the file name file to upload. Uploaded if you omit the file name will be used.

    +
    + +

    You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    JSP file that you can edit
    Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
    HeaderIt is a JSP file's header.
    FooterThis is the footer JSP files.
    Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
    Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    Help pages (frames)Is the help page of JSP files.
    Search error pageIt is a JSP file search error page. If you want to customize the search error expression change.
    Mobile home (frames)It is a JSP file for mobile home. This JSP include JSP file of each part.
    Mobile home (within the Head tags)Is a JSP file to express within the head tag for the mobile home page. If you want to edit the meta tags, title tags, script tags, such as the change.
    Mobile home (content)Is a JSP file to represent in the body tag of the mobile home.
    Portable search results pages (frames)Search results for mobile is a JSP file's page. This JSP include JSP file of each part.
    Portable search results page (within the Head tags)Search results for mobile is a JSP file to represent within the head tags of the page. If you want to edit the meta tags, title tags, script tags, such as the change.
    Portable search results page (header)Search results for mobile is a JSP file to represent the header of the page. Include search form at the top.
    Portable search results page (footer)Search results for mobile is a JSP file that represents the footer part of the page. Contains the copyright page at the bottom.
    Portable search results pages (content)Portable search results search results page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Portable search results page (result no)Portable search results search results page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    File boot pageIs the file boot page JSP file. Is the screen used when displaying Java plug-in is enabled to display the search results by using the file system CROLL.
    Error page (header)Is a JSP file that represents the header of the page.
    Error page (footer)It is a JSP file that represents the footer part of the page.
    Error page (page not found)It is displayed if the page cannot be found error page JSP file.
    Error (System error)JSP error page that appears if the system error is.
    Error pages (redirects)This is the JSP error page displayed when an HTTP redirect occurs.
    Error (bad request)Is the error bad request appears when the JSP file.
    +

    You can to edit for PCs and similar portable screen.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/dict-guide.xml b/src/site/en/xdoc/9.0/admin/dict-guide.xml new file mode 100644 index 000000000..8fcf66fae --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/dict-guide.xml @@ -0,0 +1,49 @@ + + + + Dictionary settings + Shinsuke Sugaya + + +
    +

    Here are settings for the dictionary.

    +
    +
    + +

    In Administrator account after logging in, click the menu Dictionary. List the various dictionaries available for editing.

    + Dictionary list +
    +
    +
    +

    You can register names, nouns, terminology. Registered click path to user dictionaries and dictionary word list is displayed.

    + User dictionary list +

    Displays the Edit screen and click the word you want to edit.

    + Create a user dictionary + +

    Type a Word to search for.

    +
    + +

    Been searching the words compound word splitting the Word can to hit. For example, you can search any word by entering the full-text search engine and full-text search engine, have split.

    +
    + +

    Enter the word reading in katakana. +   Enter the split when you split. For example, enter Sembene search engine.

    +
    + +

    Enter the words you entered.

    +
    +
    +
    +

    You can register the same meaning words (GB, gigabyte, etc.). Click the path of the synonym dictionary registered on dictionary word list is displayed.

    + Synonym list +

    Displays the Edit screen and click the word you want to edit.

    + Creating a synonym + +

    Type the word being treated as synonyms.

    +
    + +

    Expand the words you enter in the source in the word after the conversion. For example, 'TV', 'TV' and 'television' If you want to convert type 'TV', 'TV' and 'TV' type after conversion.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/document-guide.xml b/src/site/en/xdoc/9.0/admin/document-guide.xml new file mode 100644 index 000000000..06cbeb622 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/document-guide.xml @@ -0,0 +1,29 @@ + + + + Index settings + Shinsuke Sugaya + + +
    +

    Here are settings for the index.

    +
    +
    + +

    In Administrator account after logging in, click the menu index.

    + Index settings +
    +
    +
    + +

    You can publish index commit, optimize for server groups.

    +
    + +

    You can remove a specific search for the session ID. You can remove only the specific documents by specifying the URL.

    +
    + +

    Shown by the number of documents registered in each session. Can verify the results list by clicking the session name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/failureUrl-guide.xml b/src/site/en/xdoc/9.0/admin/failureUrl-guide.xml new file mode 100644 index 000000000..6e2cede70 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/failureUrl-guide.xml @@ -0,0 +1,23 @@ + + + + Disaster URL + Shinsuke Sugaya + + +
    +

    Here the failure URL. URL could not be obtained at crawl time are recorded and confirmed as the failure URL.

    +
    +
    + +

    In Administrator account click menu disabled URL after login.

    + Disaster URL +

    Clicking the confirmation link failure URL displayed for more information.

    + Details of the disaster URL +
    + +

    A glance could not crawl the URL and date.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/favoriteLog-guide.xml b/src/site/en/xdoc/9.0/admin/favoriteLog-guide.xml new file mode 100644 index 000000000..b157ab4fd --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/favoriteLog-guide.xml @@ -0,0 +1,21 @@ + + + + Popular URL log + Shinsuke Sugaya + + +
    +

    Here are popular URL log. When the popular URL log user clicks voting link on the search screen registers as a favorite link. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click the menu popular URL.

    + Users list +
    + +

    Lists popular URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/fileAuthentication-guide.xml b/src/site/en/xdoc/9.0/admin/fileAuthentication-guide.xml new file mode 100644 index 000000000..c027a47b5 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/fileAuthentication-guide.xml @@ -0,0 +1,44 @@ + + + + Settings for file system authentication + Shinsuke Sugaya + + +
    +

    Crawls using file system here, describes how to set file system authentication is required. Fess is corresponding to a crawl for a shared folder in Windows.

    +
    +
    + +

    In Administrator account after logging in, click the menu file system authentication.

    + File system settings +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Is omitted, the specified file system Kroll set applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. File system Crawl settings specified in that case applies on any port.

    +
    + +

    Select the authentication method. You can use SAMBA (Windows shared folder authentication).

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. SAMBA, the set value of the domain. If you want to write as.

    + +
    + +

    Select the set name to apply the authentication settings for the above file system CROLL. Must be registered ago you file system CROLL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/9.0/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..d3cb3cc66 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,106 @@ + + + + Settings for file system crawling + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: or smb: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +

    Windows shared folder, for example, if you want to crawl to host1 share folder crawl settings for smb: (last / to) the //host1/share/. If authentication is in the shared folder on the file system authentication screen set authentication information.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + + +
    IP rings contents list
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    You can specify the crawl configuration information.

    +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/index.xml b/src/site/en/xdoc/9.0/admin/index.xml new file mode 100644 index 000000000..45bc80f38 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/index.xml @@ -0,0 +1,67 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for the management work of the Fess. +

    +
    +
    +

    +This document demonstrates Fess configuration management methods. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    +
    + +

    Management UI to support browser as follows.

    +
      +
    • Internet Explorer 7 or higher
    • +
    • Firefox 4 or more
    • +
    • Chrome (latest version)
    • +
    • Safari (latest version)
    • +
    • Opera 11
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/jobLog-guide.xml b/src/site/en/xdoc/9.0/admin/jobLog-guide.xml new file mode 100644 index 000000000..19d11432e --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/jobLog-guide.xml @@ -0,0 +1,25 @@ + + + + Job log + Shinsuke Sugaya + + +
    +

    Job performed results appear as a list.

    +
    +
    + +

    In Administrator account after logging in, click menu users.

    +
    + + List joblog +

    Lists the job run log. You can determine the job name, status, start and finish times. You can also select more information, to check the details of each log.

    +
    + +

    You can check job log contents. Job name, status, start and completion time, displays the results, such as.

    + Job log details +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/labelType-guide.xml b/src/site/en/xdoc/9.0/admin/labelType-guide.xml new file mode 100644 index 000000000..df01a6958 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/labelType-guide.xml @@ -0,0 +1,39 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. You can pass even if you do not set the crawl settings in the settings of the label to add labels to specify regular expressions. If you register the label shown select label drop-down box to the right of the search box.

    +
    +
    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    +
    +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    + +

    Sets the path to label in the regular expression. You can specify multiple in multiple line description. Notwithstanding the crawl configuration document to match the path specified here, will be labeled.

    +
    + +

    In the path and crawled on regular expressions set from what you want to exclude. You can specify multiple in multiple line description.

    +
    + +

    Specifies the role to view the label.

    +
    + +

    Specifies the order of the labels.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/log-guide.xml b/src/site/en/xdoc/9.0/admin/log-guide.xml new file mode 100644 index 000000000..887e66780 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/log-guide.xml @@ -0,0 +1,21 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    +
    +
    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/overlappingHost-guide.xml b/src/site/en/xdoc/9.0/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..66ecaf16f --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/overlappingHost-guide.xml @@ -0,0 +1,27 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    +
    +
    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    +
    +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/pathMapping-guide.xml b/src/site/en/xdoc/9.0/admin/pathMapping-guide.xml new file mode 100644 index 000000000..a5f669c52 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/pathMapping-guide.xml @@ -0,0 +1,30 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    +
    +
    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    +
    +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/requestHeader-guide.xml b/src/site/en/xdoc/9.0/admin/requestHeader-guide.xml new file mode 100644 index 000000000..adaa3a262 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/requestHeader-guide.xml @@ -0,0 +1,30 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    +
    +
    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    +
    +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/roleType-guide.xml b/src/site/en/xdoc/9.0/admin/roleType-guide.xml new file mode 100644 index 000000000..6a022b1d2 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/roleType-guide.xml @@ -0,0 +1,27 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    +
    +
    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    +
    +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/scheduledJob-guide.xml b/src/site/en/xdoc/9.0/admin/scheduledJob-guide.xml new file mode 100644 index 000000000..f43d568b5 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/scheduledJob-guide.xml @@ -0,0 +1,49 @@ + + + + Job settings + Shinsuke Sugaya + + +
    +

    Here are settings for the job.

    +
    +
    + +

    In Administrator account after logging in, click the job management.

    + Job list + Edit job +
    +
    +
    + +

    It is the name that appears in the list.

    +
    + +

    You can use as an identifier for whether or not to run when the target job command to run directly in the batch, etc.. If the crawl command execution, do not specify 'all'.

    +
    + +

    Configure schedule settings. Run jobs written in script on a schedule you set here.

    +

    Description format describes the format such as Cron seconds minutes date month day year (optional)". For example, "0 0 12? * WED ' for if the weekly Wednesday 12:00 pm job to run. About how to specify the finer "Quartz"Please see.

    +
    + +

    Specifies the script execution environment. At the moment supports only the 'groovy'.

    +
    +
    +

    Written in the language specified in how to perform job run.

    +
    +
    +

    To enable records to the job log.

    +
    + +

    In turn treated as crawl jobs. In establishing the system crawl started and stopped.

    +
    + +

    Specifies the enabled or disabled status of the job. If the job will not run.

    +
    + +

    Specifies the display order in the job list.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/searchList-guide.xml b/src/site/en/xdoc/9.0/admin/searchList-guide.xml new file mode 100644 index 000000000..db88c9e04 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/searchList-guide.xml @@ -0,0 +1,21 @@ + + + + Search + Shinsuke Sugaya + + +
    +

    Here the search for management.

    +
    +
    + +

    In Administrator account after logging in, click the menu search.

    + Administrative search +
    + +

    You can search by criteria you specify. In the regular search screen role and browser requirements is added implicitly, but do not provide management for search. You can document a certain remove from index from the search results.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/searchLog-guide.xml b/src/site/en/xdoc/9.0/admin/searchLog-guide.xml new file mode 100644 index 000000000..620ae2a71 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/searchLog-guide.xml @@ -0,0 +1,21 @@ + + + + Search log list + Shinsuke Sugaya + + +
    +

    Here the search log. When you search in the search screen users search logs are logged. Search log search term or date is recorded. You can also record the URL, then you want the search results to.

    +
    +
    + +

    In Administrator account after logging in, click menu search logs.

    + Search log +
    + +

    Search language and date are listed. You can review and detailed, you click the URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/stats-guide.xml b/src/site/en/xdoc/9.0/admin/stats-guide.xml new file mode 100644 index 000000000..4c278c9c5 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/stats-guide.xml @@ -0,0 +1,21 @@ + + + + Statistics + Shinsuke Sugaya + + +
    +

    Here are statistics. You can search log and click log.

    +
    +
    + +

    In Administrator account after logging in, click the menu statistics.

    + Statistics +
    + +

    You can select the target by selecting the type of report, to make sure. Displayed in order by the specified criteria.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/system-guide.xml b/src/site/en/xdoc/9.0/admin/system-guide.xml new file mode 100644 index 000000000..144214571 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/system-guide.xml @@ -0,0 +1,33 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here are registered in the server settings for crawling and Fess. SOLR servers are grouped by file, has been registered.

    +
    +
    + +

    In Administrator account after logging in, click menu system settings.

    + System settings +
    +
    +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and Fess server to shut down is not running when shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +

    You can manually crawling under the crawl start button press stop if it is that.

    +
    + +

    Server group name to search for and update available will be shown.

    +
    + +

    In Fess Solr Server conducts a management server and index State States. Whether or not the server state can be access to the Solr Server manages. Whether or not successfully crawl index the State could manage. You can use search server status is in effect, regardless of the State of the index. The crawl Server State is enabled and can index State runs correctly if the preparation or completion. Running start crawl manually index State preparing changes automatically. Server recovery server status and auto-recovery enabled state.

    +
    + +

    You can be sure SOLR server instance state. You can also, for each instance, start, stop, reload request.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/systemInfo-guide.xml b/src/site/en/xdoc/9.0/admin/systemInfo-guide.xml new file mode 100644 index 000000000..4c6772054 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/systemInfo-guide.xml @@ -0,0 +1,32 @@ + + + + System information + Shinsuke Sugaya + + +
    +

    Here, you can currently check property information such as system environment variables.

    +
    +
    + +

    In Administrator account after logging in, click system information menu.

    + System information +
    +
    +
    + +

    You can list the server environment variable.

    +
    + +

    You can list the system properties on Fess.

    +
    + +

    Fess setup information available.

    +
    + +

    Is a list of properties to attach when reporting a bug. Extract the value contains no personal information.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/userInfo-guide.xml b/src/site/en/xdoc/9.0/admin/userInfo-guide.xml new file mode 100644 index 000000000..18f64c910 --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/userInfo-guide.xml @@ -0,0 +1,21 @@ + + + + User log + Shinsuke Sugaya + + +
    +

    Here the user log. Identifies the user when you search in the search screen users the user log in. You can search log and popular URL information and the use. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click menu users.

    + Users list +
    + +

    Lists the ID of the user. You can select the search logs or popular URL links, to see a list of each log.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/webAuthentication-guide.xml b/src/site/en/xdoc/9.0/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..f829b03aa --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/webAuthentication-guide.xml @@ -0,0 +1,48 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    +
    +
    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication, DIGEST authentication or NTLM authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. You can set the workstation and domain values for NTLM authentication. If you want to write as.

    + +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/9.0/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..5dc15b77e --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,107 @@ + + + + Settings for crawling Web site + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + + +
    URL filtering contents list
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    You can specify the crawl configuration information.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl. If you do not specify people per 100,000.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    Register the browser type was selected as the crawled documents. Even if you select only the PC search on your mobile device not appear in results. If you want to see only specific mobile devices also available.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in).

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/admin/wizard-guide.xml b/src/site/en/xdoc/9.0/admin/wizard-guide.xml new file mode 100644 index 000000000..d9cadeeba --- /dev/null +++ b/src/site/en/xdoc/9.0/admin/wizard-guide.xml @@ -0,0 +1,31 @@ + + + + Settings Wizard + Sone, Takaaki + + +
    +

    Introduction to the Configuration Wizard.

    +

    You can use Settings Wizard, to set you up on the Fess.

    +
    +
    + +

    In Administrator account after logging in, click menu Settings Wizard.

    + Settings Wizard +

    Do the crawl settings. + Crawl settings is to register a URI to look for. + The crawl settings name please put name of any easy to identify. Put the URI part de-indexed, want to search for.

    + Crawl settings +

    For example, if you want and search for http://fess.codelibs.org/, less looks like.

    + Crawl settings example +

    The type, such as c:\Users\taro file.

    +

    In this setting is complete. Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

    + Crawl started +
    + +

    Settings in the Setup Wizard you can change from crawl General, Web, file system.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/db.xml b/src/site/en/xdoc/9.0/config/db.xml new file mode 100644 index 000000000..2a62426a9 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/db.xml @@ -0,0 +1,128 @@ + + + + Database changes + Shinsuke Sugaya + + +
    +

    Provides binaries to use H2 Database with MySQL database. You can use the other database in to change the settings using the source code and build it.

    +
    +
    + +

    The MySQL character code setting. /etc/mysql/my.cnf and the added must have the following settings.

    + +
    + +

    Download MySQL binaries and expand.

    +
    + +

    Create a database.

    + create database fess_db; +mysql> grant all privileges on fess_db.* to fess_user@localhost identified by 'fess_pass'; +mysql> create database fess_robot; +mysql> grant all privileges on fess_robot.* to s2robot@localhost identified by 's2robot'; +mysql> FLUSH PRIVILEGES; +]]> +

    Create a table in the database. DDL file is located in extension/mysql.

    + +

    webapps/fess/WEB-INF/lib on mysql driver jar.

    + +

    Edit the webapps/fess/WEB-INF/classes/jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_db?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "fess_user" + "fess_pass" + + + + 600 + 10 + true + + + + + +]]> +

    Edit the webapps/fess/WEB-INF/classes/s2robot_jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_robot?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "s2robot" + "s2robot" + + + + 600 + 10 + true + + @java.sql.Connection@TRANSACTION_REPEATABLE_READ + + + + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/filesize.xml b/src/site/en/xdoc/9.0/config/filesize.xml new file mode 100644 index 000000000..ff556b6fc --- /dev/null +++ b/src/site/en/xdoc/9.0/config/filesize.xml @@ -0,0 +1,29 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of defaultMaxLength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/geosearch.xml b/src/site/en/xdoc/9.0/config/geosearch.xml new file mode 100644 index 000000000..dfdb560eb --- /dev/null +++ b/src/site/en/xdoc/9.0/config/geosearch.xml @@ -0,0 +1,23 @@ + + + + Dios arch + Shinsuke Sugaya + + +
    +

    You can document with latitude and longitude location information in conjunction with Google maps, including the use of Dios arch.

    +
    +
    + +

    Location is defined as a feed that contains the location information. + When generating the index in Solr latitude longitude set to location feeds in formats such as 45.17614,-93.87341, register the document. + Also sets the value as the latitude_s and longitude_s fields if you want to display latitude and longitude as a search result. * _s is available as a dynamic field of Solr string.

    +
    + +

    During the search specifies in the request parameter to latitude and longitude, the distance. + View the results in the distance (km) specified by distance-based latitude information (latitude, longitude). Latitude and longitude and distances is treated as double.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/index-backup.xml b/src/site/en/xdoc/9.0/config/index-backup.xml new file mode 100644 index 000000000..5efd742f3 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index management + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data and solr/core1-suggest/data directories. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/index.xml b/src/site/en/xdoc/9.0/config/index.xml new file mode 100644 index 000000000..0ea320154 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for setting work of the Fess. +

    +
    +
    +

    +This document demonstrates setting Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/install-on-tomcat.xml b/src/site/en/xdoc/9.0/config/install-on-tomcat.xml new file mode 100644 index 000000000..17ef0f1b6 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/install-on-tomcat.xml @@ -0,0 +1,43 @@ + + + + Install to an existing Tomcat + Shinsuke Sugaya + + +
    +

    + The standard distribution of Fess Tomcat is distributed in the deployed State. + Because Fess is not dependent on Tomcat, deploying on any Java application server is available. + Describes how to deploy a Fess Tomcat here is already available.

    +
    +
    +

    Fess ServerHereFrom the download.

    +
    +
    +

    + Expand the downloaded Fess server. + Expanded Fess Server home directory to $FESS_HOME. + $TOMCAT_HOME the top directory of an existing Tomcat 7. + Copy the Fess Server data.

    + +

    + If you have, such as changing the destination file diff commands, updates your diff only applies. +

    +
    +
    +

    to start the Tomcat in normal as well as in startup.* / http://localhost:8080/Fess To access the.

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/logging.xml b/src/site/en/xdoc/9.0/config/logging.xml new file mode 100644 index 000000000..bb093f399 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/logging.xml @@ -0,0 +1,56 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    Log file list
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina,outLog of the Fess Server (Tomcat). Tomcat logs are output.
    logs/SOLR.logLog server Fess (Solr). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    + +

    You can check by changing the logging level of the HttpClient if you want to check HTTP access when crawling content to solve the problem.

    + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/memory-config.xml b/src/site/en/xdoc/9.0/config/memory-config.xml new file mode 100644 index 000000000..6d8a2a54b --- /dev/null +++ b/src/site/en/xdoc/9.0/config/memory-config.xml @@ -0,0 +1,52 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    Set the maximum memory per process in Java. So, do not use the upper memory in the process also had 8 GB of physical memory on the server. Memory consumption depending on the number of crawl threads and interval will also change significantly. If not enough memory please change settings in the subsequent procedure.

    +
    +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case maximum value set to 1 G)-xmx1g to change.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-xmx1g (in this case maximum value set to 1 G). +

    + new String[] { +"-Djava.awt.headless=true", +"-server", +"-Xmx1g", +"-XX:MaxPermSize=128m", +"-XX:-UseGCOverheadLimit", +"-XX:+UseConcMarkSweepGC", +"-XX:CMSInitiatingOccupancyFraction=75", +"-XX:+CMSIncrementalMode", +"-XX:+CMSIncrementalPacing", +"-XX:CMSIncrementalDutyCycleMin=0", +"-XX:+UseParNewGC", +"-XX:+UseStringCache", +"-XX:+UseTLAB", +"-XX:+DisableExplicitGC" +} +]]> +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/mobile-device.xml b/src/site/en/xdoc/9.0/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/path-encoding.xml b/src/site/en/xdoc/9.0/config/path-encoding.xml new file mode 100644 index 000000000..b8aa16ac6 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    +
    +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/9.0/config/pdf-with-password.xml b/src/site/en/xdoc/9.0/config/pdf-with-password.xml new file mode 100644 index 000000000..0a8f4e317 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/pdf-with-password.xml @@ -0,0 +1,89 @@ + + + + Password-protected PDF + Shinsuke Sugaya + + +
    +

    + You should password files to register the settings file to PDF password is configured to search for. +

    +
    +
    +

    + First of all, create the webapps/fess/WEB-INF/classes/s2robot_extractor.dicon. + This is test _ ~ is a pass that password set to a.pdf file. + If you have multiple files, multiple settings in addPassword.

    + + + + + + + + + + + + + ".*test_.*.pdf" + + "pass" + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + +... +]]> +

    Next, edit the following to webapps/fess/WEB-INF/classes/s2robot_rule.dicon.

    + + "fsFileRule" + + + fessFileTransformer + + + true + + "url" + "file:.*" + + + "mimeType" + + + "(application/xml" ++ "|application/xhtml+xml" ++ "|application/rdf+xml" ++ "|application/pdf" ++ "|text/xml" ++ "|text/xml-external-parsed-entity" ++ "|text/html)" + + +
    +... +]]> +

    Fess up after setting the above and please run the crawl. How to use Basic remains especially.

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/protwords.xml b/src/site/en/xdoc/9.0/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/proxy.xml b/src/site/en/xdoc/9.0/config/proxy.xml new file mode 100644 index 000000000..db1db6b1a --- /dev/null +++ b/src/site/en/xdoc/9.0/config/proxy.xml @@ -0,0 +1,120 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    +
    +

    Proxy is set in to create webapps/fess/WEB-INF/classes/s2robot_client.dicon with the following contents.

    + + + + + + + + @org.apache.http.client.params.CookiePolicy@BEST_MATCH + clientConnectionManager + "プロキシサーバー名(ex. 192.168.1.1)" + プロキシサーバーのポート(ex. 8080) (" は不要) + + + + internalHttpClient + 5 + 500 + + + + + + + + + "http" + 80 + @org.apache.http.conn.scheme.PlainSocketFactory@getSocketFactory() + + + + + + + "https" + 443 + @org.apache.http.conn.ssl.SSLSocketFactory@getSocketFactory() + + + + + + 5 + @java.util.concurrent.TimeUnit@MINUTES + + 200 + + 20 + + + + + "UTF-8" + + + internalFsClient + 5 + 500 + + + + "UTF-8" + + + + internalSmbClient + 5 + 500 + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + "smb:.*" + smbClient + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/9.0/config/replication.xml b/src/site/en/xdoc/9.0/config/replication.xml new file mode 100644 index 000000000..4a08377d0 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/replication.xml @@ -0,0 +1,56 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Sets the replication of the index Solr replication features. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +
    +
    + +

    Fess, download and install the. When you install MasterServer named host./ /opt/fess_mster To assume you installed. Edit the SOLR/core1/conf/solrconfig.XML like the following.

    + + + startup + optimize + optimize + schema.xml,stopwords.txt,stopwords_ja.txt,elevate.xml, + stoptags_ja.txt,synonyms.txt,mapping_ja.txt,mapping-FoldToASCII.txt, + mapping-ISOLatin1Accent.txt,protwords.txt,compositePOS.txt,spellings.txt, + currency.xml + + 1 + +... +]]> +

    Register the crawl settings as well as Fess starts after the normal construction. Steps to building the index for Fess remains especially as normal building procedures.

    +
    + +

    Fess, download and install the./ /opt/fess_slave To assume you installed. Edit the SOLR/core1/conf/solrconfig.XML like the following.

    + + + http://MasterServer:8080/solr/core1/replication + 00:00:60 + internal + 5000 + 10000 + solradmin + solradmin + + +... +]]> +

    Start the Fess.

    +
    + +

    Until the above settings copied to search for Fess index crawl after the optimization (optimize) the index for Fess.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/role-setting.xml b/src/site/en/xdoc/9.0/config/role-setting.xml new file mode 100644 index 000000000..3c973b9cc --- /dev/null +++ b/src/site/en/xdoc/9.0/config/role-setting.xml @@ -0,0 +1,85 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    +
      +
    • Request parameter
    • +
    • Request header
    • +
    • Cookies
    • +
    • J2EE authentication information
    • +
    +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + + {"guest"} + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/server-port.xml b/src/site/en/xdoc/9.0/config/server-port.xml new file mode 100644 index 000000000..8b86f8ec5 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/server-port.xml @@ -0,0 +1,40 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    • 19092: database port (use h2database)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information.

    +

    change the webapps/fess/WEB-INF/classes/app.dicon the following points.

    + "http://localhost:8080/manager/text/" +]]> +

    change the webapps/fess/WEB-INF/classes/solrlib.dicon the following points.

    + "http://localhost:8080/solr/core1" +]]> +

    change the SOLR/core1/conf/solrconfig,XML the following points.

    + http://localhost:8080/solr/core1-suggest +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/solr-dynamic-field.xml b/src/site/en/xdoc/9.0/config/solr-dynamic-field.xml new file mode 100644 index 000000000..073ca1d38 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/solr-dynamic-field.xml @@ -0,0 +1,32 @@ + + + + Dynamic fields + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in SOLR/core1/conf/schema,XML. dynamic fields and standard fields such as title and content can be freely defined field names are defined. Advanced parameter values see a Solr document.

    +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add a field to use to retrieve data that is stored in the dynamic field next to the webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"url", "host", "site", + "title", "content", "contentLength", "lastModified", "mimetype", + "label", "segment", "other_t" } + +]]> +

    You must add a field to use to retrieve data that is stored in the dynamic field from Solr. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/solr-failure.xml b/src/site/en/xdoc/9.0/config/solr-failure.xml new file mode 100644 index 000000000..6474fcdfb --- /dev/null +++ b/src/site/en/xdoc/9.0/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/solrlib.dicon.

    +
    +
    +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is minActiveServer following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status maxRetryStatusCheckCount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    +
    +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    +
    +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/tokenizer.xml b/src/site/en/xdoc/9.0/config/tokenizer.xml new file mode 100644 index 000000000..296a09c3b --- /dev/null +++ b/src/site/en/xdoc/9.0/config/tokenizer.xml @@ -0,0 +1,47 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search. Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits. For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit. For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default StandardTokenizer CJKBigramFilter used.

    + + +

    StandardTokenizer CJKBigramFilter index bi-gram, in other words two characters to like Japan Japanese multibyte string creates. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to. However, please note that the index size increases.

    +

    The following example to change the analyzer part like solr/core1/conf/schema.xml, you can use the StandardTokenizer.

    + + + + + + + : + + + + + : +]]> +

    Also, useBigram is enabled by default in the webapps/fess/WEB-INF/classes/app.dicon change to false.

    + + true + : +]]> +

    After the restart the Fess.

    +
    +
    + + +
    diff --git a/src/site/en/xdoc/9.0/config/use-libreoffice.xml b/src/site/en/xdoc/9.0/config/use-libreoffice.xml new file mode 100644 index 000000000..aeb4c4363 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/use-libreoffice.xml @@ -0,0 +1,85 @@ + + + + Use of LibreOffice + Shinsuke Sugaya + + +
    +

    + It is possible to crawl using the Apache POI Fess environmental standard in MS Office system document. + You can crawl Office system document regarding LibreOffice, OpenOffice, do even more accurate text extraction from documents.

    +
    +
    +

    JodConverter Fess server install. from http://jodconverter.googlecode.com/jodconverter-core-3.0-Beta-4-Dist.zipThe download. Expand and copy the jar file to Fess server.

    + +

    Create a s2robot_extractor.dicon to the next.

    + +

    s2robot_extractor,DiCon effective jodExtractor with following contents.

    + + + + + + + + + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + + officeManagerConfiguration.setOfficeHome("/usr/lib/libreoffice") + .buildOfficeManager() + + + + + + { +"application/msword", +"application/vnd.ms-excel", +"application/vnd.ms-powerpoint", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.openxmlformats-officedocument.presentationml.presentation" + } + jodExtractor + +... +]]> +

    Index to generate the settings later, usually crawled into the street.

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/config/windows-service.xml b/src/site/en/xdoc/9.0/config/windows-service.xml new file mode 100644 index 000000000..84564daa0 --- /dev/null +++ b/src/site/en/xdoc/9.0/config/windows-service.xml @@ -0,0 +1,54 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as Add %JAVA_HOME%\bin to PathYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option.

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-9.0.0.

    + cd C:\Java\fess-server-9.0.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat7w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    +
    + +

    Distributed in the Fess is 64-bit binaries for Windows Tomcat builds based on. If you use 32-bit WindowsTomcat Of the site from, such as 32-bit Windows zip, please replace tomcat7.exe, tomcat7w.exe, tcnative-1.dll.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/install/index.xml b/src/site/en/xdoc/9.0/install/index.xml new file mode 100644 index 000000000..1599de0ae --- /dev/null +++ b/src/site/en/xdoc/9.0/install/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for Fess installation and uninstallation tasks. +

    +
    +
    +

    +This document shows how to install Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN9sm, Inc.To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.0/install/install.xml b/src/site/en/xdoc/9.0/install/install.xml new file mode 100644 index 000000000..714f02c0e --- /dev/null +++ b/src/site/en/xdoc/9.0/install/install.xml @@ -0,0 +1,131 @@ + + + + Installation + + + + + +
    +

    +Fess can be available in these environments. +

    +
      +
    • OS: Windows, Unix, and Java can run OS environment
    • +
    • Java: Java 7 or more
    • +
    +

    +If Java is not installed in the environment we want to Fess from http://java.sun.com/ get JDK for Java 7 and above, please install. +

    +
    +
    +

    +http://sourceforge.jp/projects/Fess/releases/ From the download the latest Fess package. +

    +
    +
    +

    +Expand the downloaded fess-server-x.y.zip. +If you installed in the UNIX environment, in the bin added the performing rights to a script.

    + +
    +
    +

    +Administrator account is managed by the application server. Fess Server standard available Tomcat, as well as to the user changing the Tomcat. +Modify the password for the admin account of the conf/tomcat-user.xml if you want to change.

    + +]]> +

    +see the Tomcat documentation or JAAS authentication specification to use Tomcat-user.XML file management method other than. +

    +
    +
    +

    +To access the Solr into Fess server is password becomes necessary. +Change the default passwords in production, etc.

    +

    +How to change the password, you must first change the password attribute of the conf/tomcat-user.xml solradmin. +

    + +]]> +

    +Modifies the following three files webapps/fess/WEB-INF/classes/solrlib.dicon, fess_suggest.dicon and solr/core1/conf/solrconfig.xml. +Write what you specified in tomcat-user.xml to the following password.

    +

    +modify the following areas of the solrlib.dicon. +

    + + "solradmin" + "solradmin" + +]]> +

    +fess_suggest.dicon is the following. +

    + + "solradmin" + "solradmin" + +]]> +

    +where the following is the solrconfig.Xml. +

    + + + + http://localhost:8080/solr/core1-suggest + + solradmin + solradmin + +]]> +
    +
    +

    +To manage can manage Solr from Fess server deployed on Tomcat context need a password. +Change the default passwords in production, etc.

    +

    +How to change password the change manager conf/tomcat-user.xml password attributes. +

    + + +]]> +

    +Describes the provided password webapps/fess/WEB-INF/classes/app.dicon the following points tomcat-user.xml to the next. +

    + + + "solrServer1" + "http://localhost:8080/manager/text/" + "/solr" + "manager" + "manager" + +]]> +
    +
    +

    +Login at return path settings, in the use encryption and decryption. +Please change the default password, especially in production.

    +

    +How to change the change the key value in the webapps/fess/WEB-INF/classes/app.dicon. +Set the 16-character alphanumeric.

    + +"1234567890123456" +]]> +
    + +
    diff --git a/src/site/en/xdoc/9.0/install/run.xml b/src/site/en/xdoc/9.0/install/run.xml new file mode 100644 index 000000000..ba5a8b3b2 --- /dev/null +++ b/src/site/en/xdoc/9.0/install/run.xml @@ -0,0 +1,41 @@ + + + + Start and stop + + + + + +
    +

    +to run the startup script, start the Fess. +

    + + +

    +access to / http://localhost:8080/Fess ensures startup. +

    + +

    +Management UI is / http://localhost:8080/fess/admin. +Default Administrator account user name / password is admin/admin. +Administrator account is managed by the application server. +In the management UI of the Fess, authenticate with the application server in fess role available as an administrator.

    +
    +
    +

    +Fess to stop the running shutdown scripts. +

    + + +

    +Crawl or may take a while to completely stop during the index creation if you. +

    +
    + +
    diff --git a/src/site/en/xdoc/9.0/install/uninstall.xml b/src/site/en/xdoc/9.0/install/uninstall.xml new file mode 100644 index 000000000..d95af5b33 --- /dev/null +++ b/src/site/en/xdoc/9.0/install/uninstall.xml @@ -0,0 +1,19 @@ + + + + Uninstall + + + + + +
    +

    +You can uninstall by deleting the installation stops after the Fess fess-server-x.y directory. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/index.xml b/src/site/en/xdoc/9.0/user/index.xml new file mode 100644 index 000000000..52bdaa14c --- /dev/null +++ b/src/site/en/xdoc/9.0/user/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is targeted to users to Fess. +

    +
    +
    +

    +This document demonstrates how to search in the Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/json-response.xml b/src/site/en/xdoc/9.0/user/json-response.xml new file mode 100644 index 000000000..2f7b3b357 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/json-response.xml @@ -0,0 +1,184 @@ + + + + Search results for JSON + Shinsuke Sugaya + + +
    +

    Outputs a JSON results of Fess. You must in order to emit JSON enabled keep JSON response in the General Management screen crawl settings.

    + +

    To get JSON output http://localhost:8080/Fess/JSON? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    callbackIf you use the JSONP callback name. You do not need to specify if you do not use the JSONP.
    +
    + +

    Returns the following response.

    + \n..", + "tstamp": "1284739487887", + "url": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F", + "id": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F;type=au,docomo,pc,softbank", + "mimetype": "text\u002Fhtml", + "title": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "contentTitle": "\ngoo \u30B9\u30D4\u30FC\u30C9\u30C6\u30B9\u30C8\n", + "boost": "1.0", + "contentLength": "17810", + "urlLink": "http:\u002F\u002Fspeedtest.goo.ne.jp\u002F" + }, +... + ] + } +} +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. the status value is 0: normal, 1: search error, 2 or 3: request parameter error, 9: in service, 1: API type of error is.
    querySearch language.
    execTimeResponse time. Unit is seconds.
    pageSizeThe number of display.
    pageNumberPage number.
    recordCountNumber of hits for the search term.
    pageCountPage number of the number of hits for the search term.
    resultParent element of the search results.
    siteThe name of the site.
    contentDescriptionA description of the content.
    hostThe name of the host.
    lastModifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    contentTitleThe title of the document for viewing.
    contentLengthThe size of the document.
    urlLinkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-additional.xml b/src/site/en/xdoc/9.0/user/search-additional.xml new file mode 100644 index 000000000..ae0c2aa61 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-additional.xml @@ -0,0 +1,15 @@ + + + + Hide search criteria + Shinsuke Sugaya + + +
    +

    You can use the additional parameters if the search string is shown on the screen without the specific search criteria like persuasion. additional value is retained but in the paging screen additional value.

    + +

    Without the conditions show screen and run a search when searches are performed by appending additional values in hidden forms, such as (for example, a search form) in the paging screen transitions and also the condition holds.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-and.xml b/src/site/en/xdoc/9.0/user/search-and.xml new file mode 100644 index 000000000..f20ac20a5 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-and.xml @@ -0,0 +1,20 @@ + + + + AND search + Shinsuke Sugaya + + +
    +

    Use the search if you want to search for documents that contain all search words of more than one. When describing multiple words in the search box separated by spaces, AND skip AND search.

    + +

    If you use the search search words written AND. Write in capital letters AND the space required to back and forth. AND is possible can be omitted.

    +

    For example, if you want to find documents that contain the search terms 1 and 2 search terms, type the following search form.

    + +

    Connecting multiple words with AND are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-boost.xml b/src/site/en/xdoc/9.0/user/search-boost.xml new file mode 100644 index 000000000..697a4bd94 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-boost.xml @@ -0,0 +1,20 @@ + + + + Boost search + Shinsuke Sugaya + + +
    +

    Use the boost search if you want to prioritize, search for specific search terms. Enabling search in boost search, depending on the severity of the search words.

    + +

    To boost search after the search term ' ^ boost value "that specifies the boost value (weighted) in the format.

    +

    For example, if you want to find the page if you want to find apples oranges contained more 'apples', type the following search form.

    + +

    Boost value specifies an integer greater than 1.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-field.xml b/src/site/en/xdoc/9.0/user/search-field.xml new file mode 100644 index 000000000..3c668053c --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-field.xml @@ -0,0 +1,66 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    You crawl in Fess results are saved for each field, such as title and full text. You can search for a field of them. You can specify the search criteria in search for a field, such as document type or size small.

    + +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Field list is available
    Field nameDescription
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields subject to find the content. Fields are custom fields are also available by using the dynamic field of Solr.

    +

    If HTML file and search for the title tag that string in the title field, below the body tag registered in the body field.

    +
    + +

    If a field specifying the search field name: search words in separated by colons (:) field name and search word fill in the search form, the search.

    +

    If you search the Fess as a search term for the title field, type.

    + +

    Document, the title field in Fess above search appears as a search result.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-fuzzy.xml b/src/site/en/xdoc/9.0/user/search-fuzzy.xml new file mode 100644 index 000000000..30c6ba212 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-fuzzy.xml @@ -0,0 +1,21 @@ + + + + Fuzzy search (fuzzy search) + Shinsuke Sugaya + + +
    +

    Ambiguity in the case does not match the words word search to search is available. Based on the Levenshtein distance in Fess ambiguous corresponds to the search (fuzzy search).

    + +

    After the search word you want to apply the fuzzy search adds '~'.

    +

    For example, ambiguous word "Solr", you can find documents that contain the word, near the "Solr" If you want to find, type as the search form, such as ("Solar"). +

    + +

    Furthermore, if by '~' after the number 0 and 1, 1 just like in refine. For example, in the form of 'Solr~0.8'. Do not specify numeric default value is 0.5.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-geo.xml b/src/site/en/xdoc/9.0/user/search-geo.xml new file mode 100644 index 000000000..8dfb4282a --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-geo.xml @@ -0,0 +1,32 @@ + + + + Dios arch + Shinsuke Sugaya + + +
    +

    Search using location information when you search, adding latitude and longitude location information for each document when generating the index becomes possible.

    + +

    Following parameters is available in the standard.

    + + + + + + + + + + + + + + + + +
    Request parameter
    GEO.latitudeLatitude degrees minutes seconds specifies double.
    GEO.longitudeLongitude degrees minutes seconds specifies double.
    GEO.distanceSpecifies the distance between documents in units of the coast. Specified in the double type.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-label.xml b/src/site/en/xdoc/9.0/user/search-label.xml new file mode 100644 index 000000000..e79d75748 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-label.xml @@ -0,0 +1,17 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    Narrow your search by adding the categories to search the document for label information the label is specified when the search is possible. Label information by registering in the Administration screen, will enable search by labels in the search screen. Label information available can multiple selections in the drop-down when you search. If you do not register the label displayed the label drop-down box.

    + +

    You can select the label information at search time. Label information can be selected in the search options dialog by pressing the options button will appear.

    + Label search +

    You can search each document to create an index, set the label to the label. All results search search do not specify a label is usually the same. If you change the label information to update the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-not.xml b/src/site/en/xdoc/9.0/user/search-not.xml new file mode 100644 index 000000000..3a40c9f2b --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-not.xml @@ -0,0 +1,19 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If the document does not contain a word search NOT searching is available.

    + +

    Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, enter the following if you want to find documents that contain the search term 1 does not contain a search term 2 and the search.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-options.xml b/src/site/en/xdoc/9.0/user/search-options.xml new file mode 100644 index 000000000..85132468e --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-options.xml @@ -0,0 +1,37 @@ + + + + Extended parameter search + Shinsuke Sugaya + + +
    +

    You can pass any search criteria from third-party search engines move and easy to like. Pass search criteria Please implement processing in QueryHelperImpl #buildOptionQuery.

    + +

    Following parameters is available in the standard.

    + + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    options.qThis is similar to the normal query. You can specify multiple options.q. If you specify multiple is treated as a search. Pass the URL encoding.
    options.CQTreated as exact match search queries. For example, if you specify the Fess Project searches as "Fess Project". Pass the URL encoding.
    options.OQIs treated as an OR search. For example, if you specify the Fess Project search as a Fess OR Project. Pass the URL encoding.
    options.NQThe label value. Use to specify the label.Treated as NOT search. For example, if you specify 'Fess' search as NOT Fess. Pass the URL encoding.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-or.xml b/src/site/en/xdoc/9.0/user/search-or.xml new file mode 100644 index 000000000..6b3ed55f1 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-or.xml @@ -0,0 +1,20 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. When describing the multiple words in the search box, by default will search.

    + +

    To use search OR search words written OR. OR write in capital letters, the space required to back and forth.

    +

    For example, if you want to search for documents that contain either search term 2 search term 1 and type the following search form.

    + +

    OR between multiple languages are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-range.xml b/src/site/en/xdoc/9.0/user/search-range.xml new file mode 100644 index 000000000..378788f48 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-range.xml @@ -0,0 +1,37 @@ + + + + Range searches + Shinsuke Sugaya + + +
    +

    If in the field containing the data in specified range, such as the number range search is possible for that field.

    + +

    To limit "field name: value TO value ' fill in the search form.

    +

    For example, type to search document contentLength field against 1 k to 10 k bytes is shown below the search form.

    + +

    To time range specified search ' lastModified: [date 1 TO date 2] "(Re 1 [Re 2) fill out the search form.

    +

    ISO 8601 with respect to re.

    + + + + + + + + + +
    Date and time-resolved second and fractional partIf the current relative to the date
    YYYY-MM-DDThh:mm:ss.sZ( example :2013-08-02T10:45:23.5Z)NOW (the current date), YEAR (this year), MONTH (month), DAY (today)
    +

    To relative to the current date and time NOW and DAY-(Adder, and production) and can sign and (round) like.

    +

    And a symbol for round / behind unit. Even if now-1DAY/day does today what time today 00: represents the day from 00 -1, the 00: 00.

    +

    For example, if you search for lastModified field from 2/21/2012 20: (current to date) 30 days prior to the updated document, type the following search form.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-sort.xml b/src/site/en/xdoc/9.0/user/search-sort.xml new file mode 100644 index 000000000..0629b0e87 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-sort.xml @@ -0,0 +1,54 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    + +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + + + + + + +
    Sort fields list
    Field nameDescription
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    +

    Adding custom fields as sort in Customizing.

    +
    + +

    You can select the sorting criteria when you search. Sorting criteria can be selected in the search options dialog by pressing the options button will appear.

    + Sorting criteria +

    Also, for sorting in the search field sort: the field name to sort and field names separated by colon (:) fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/search-wildcard.xml b/src/site/en/xdoc/9.0/user/search-wildcard.xml new file mode 100644 index 000000000..753e5afa5 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/search-wildcard.xml @@ -0,0 +1,31 @@ + + + + Wildcard + Shinsuke Sugaya + + +
    +

    You can use one or multiple character wildcard search terms within. The can be specified as a one-character wildcard, * is specified as the multiple-character wildcard. Wildcards are not available in the first character. You can use wildcards for words. Wildcard search for the sentence.

    + +

    If you use one character wildcard shown below? The available.

    + +

    If the above is treated as a wildcard for one character, such as text or test.

    +

    If you use the multiple character wildcard use * below

    + +

    If the above is treated as a wildcard for multiple characters, such as test, tests or tester. Also,

    + +

    The so can be also used in the search term.

    +
    + +

    The wildcard string indexed using target. Therefore, because if the index has been created, such as bi-gram be treated meaning fixed string length in Japan Japanese wildcard in Japan, not expected behavior. Use in the field, if you use a wildcard in Japan, that used morphological analysis.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.0/user/xml-response.xml b/src/site/en/xdoc/9.0/user/xml-response.xml new file mode 100644 index 000000000..14e337d00 --- /dev/null +++ b/src/site/en/xdoc/9.0/user/xml-response.xml @@ -0,0 +1,212 @@ + + + + XML search results + Shinsuke Sugaya + + +
    +

    To emit the XML results of Fess. You need to do this output in XML Management screen crawl settings XML response enabled by you.

    + +

    To get the output XML http://localhost:8080/Fess/XML? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    +
    + +

    Returns the following response.

    + + + 1 + 0 + テスト + 0.64 + 20 + 1 + 101 + 6 + + + speedtest.goo.ne.jp/ + goo の提供するスピード<em>テスト</em> + ページです + speedtest.goo.ne.jp + 1284739487873 + --> +<meta http-equiv="Content-Type" content="text/html; charset=EUC-JP"> +<title> +goo スピード<em>テスト</em> +</title> +<meta..."> + +<meta name="description" content="goo の提供するスピード<em>テスト +</em>ページです"> +<meta name="keywords" content="ADSL,ブロードバンド,フレ...ッツ,BB,無線LAN, +CATV,ISDN,スピード<em> +テスト</em>"> +<meta name="Copyright" content="Copyright &copy; ...://speedtest.goo.ne.jp/" +><img src="/img/h1/h1_76.gif" width="129" height="22" border="0" +alt="スピード<em>テスト +</em>"><img src..." color="#ffffff">goo スピード<em>テスト +</em></font +></a></b></td> +<!--/NAME SERVICE--> +</tr> +</table> +</td> +</tr> +<tr> +<td + 4.98744 + goo の提供するスピード<em>テスト</em>ページです + 1284739487887 + http://speedtest.goo.ne.jp/ + http://speedtest.goo.ne.jp/;type=au,docomo,pc,softbank + text/html + +goo スピードテスト + + +goo スピードテスト + + 1.0 + 17810 + http://speedtest.goo.ne.jp/ + +... + + +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. the status value is 0: normal, 1: search error, 2 or 3: request parameter error, 9: in service, 1: API type of error is.
    querySearch language.
    exec-timeResponse time. Unit is seconds.
    page-sizeThe number of display.
    page-numberPage number.
    record-countNumber of hits for the search term.
    page-countPage number of the number of hits for the search term.
    resultParent element of the search results.
    docThe elements of the search results.
    siteThe name of the site.
    content-descriptionA description of the content.
    hostThe name of the host.
    last-modifiedLast update date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    TstampDocument creation date. 1970 / 01 / 01 milliseconds starting at 00: 00: 00.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    content-titleThe title of the document for viewing.
    content-lengthThe size of the document.
    URL-linkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/crawl-guide.xml b/src/site/en/xdoc/9.1/admin/crawl-guide.xml new file mode 100644 index 000000000..b37ec0ba8 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/crawl-guide.xml @@ -0,0 +1,87 @@ + + + + The General crawl settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to crawling.

    +
    +
    + +

    In Administrator account click crawl General menu after login.

    + Crawl General +
    +
    +
    + +

    When the user enters a search, the search the output log. If you want to get search statistics to enable.

    +
    + +

    Save the information you find. Identifying the users becomes possible.

    +
    + +

    You can collect the search result was judged good by the user. Search result voting link appears to result in list screen, so that link press made the record. You can also reflect the results collected during the crawl index.

    +
    + +

    Search results link attaches to the search term. To display the find search terms in PDF becomes possible.

    +
    + +

    Search results can be retrieved in XML format. http://localhost:8080/Fess/XML? can get access query = search term.

    +
    + +

    Search results available in JSON format. http://localhost:8080/Fess/JSON? can get access query = search term.

    +
    + +

    You can specify the label to see if the label by default,. Specifies the value of the label.

    +
    + +

    You can specify whether or not to display a search screen. If not available not available search screen. And if you want to create a dedicated index server and select not available.

    +
    + +

    In JSON format often find search words becomes available. http://localhost:8080/Fess/JSON? can be retrieved by accessing the type = hotsearchword.

    +
    + +

    Delete a search log for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Delete the job days before the specified date. One day in the one log purge old log is deleted.

    +
    + +

    Delete the user information for the specified number of days ago. One day in the one log purge old log is deleted.

    +
    + +

    Specifies the Bots name Bots you want to remove from the search log logs included in the user agent by commas (,). Log is deleted by log purge once a day.

    +
    + +

    Specifies the email address to send information about crawl upon completion crawl.

    +
    + +

    Specifies the encoding for the CSV will be available in the backup and restore.

    +
    + +

    Crawl as been updated to enable incremental crawl compared lastModified field value and the target document's modification date (if the HTTP's timestamp if LAST_MODIFIED values, file).

    +
    + +

    File additional group access rights information added to the role.

    +
    + +

    Fess can combine multiple Solr server as a group, the group can manage multiple. Solr server group for updates and search for different groups to use. For example, if you had two groups using the Group 2 for update, search for use of Group 1. After the crawl has been completed if switching server updates for Group 1, switches to group 2 for the search. It is only valid if you have registered multiple Solr server group.

    +
    + +

    Fess document crawling is done on Web crawling, and file system CROLL. You can crawl to a set number of values in each crawl specified here only to run simultaneously multiple. For example, crawl setting number of concurrent as 3 Web crawling set 1-set 10 if the crawling runs until the set 3 3 set 1-. Complete crawl of any of them, and will start the crawl settings 4. Similarly, setting 10 to complete one each in we will start one.

    +

    But you can specify the number of threads in the crawl settings simultaneously run crawl setting number is not indicates the number of threads to start. For example, if 3 in the number of concurrent crawls settings, number of threads for each crawl settings and 5 3 x 5 = 15 thread count up and crawling.

    +
    + +

    You can automatically delete data after the data has been indexed. If you select the 5, with the expiration of index register at least 5 days before and had no update is removed. If you omit data content has been removed, can be used.

    +
    + +

    Registered disabled URL URL exceeds the failure count next time you crawl to crawl out. Does not need to monitor the fault type is being crawled next time by specifying this value.

    +
    + +

    Disaster URL exceeds the number of failures will crawl out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/crawlingSession-guide.xml b/src/site/en/xdoc/9.1/admin/crawlingSession-guide.xml new file mode 100644 index 000000000..20b862334 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/crawlingSession-guide.xml @@ -0,0 +1,27 @@ + + + + Set session information + Shinsuke Sugaya + + +
    +

    Describes the settings related to the session information. One time the crawl results saved as a single session information. You can check the run time and the number of indexed.

    +
    +
    + +

    In Administrator account after logging in, click the session information menu.

    +
    +
    +
    + + Session information list +

    You can remove all session information and click the Delete link all in the running. Session has expired will be removed at next crawl.

    +
    + +

    Sure you can crawl the contents of session ID. Crawl start and finish time, number of documents indexed and listed.

    + Session details +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/data-guide.xml b/src/site/en/xdoc/9.1/admin/data-guide.xml new file mode 100644 index 000000000..de6a940e9 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/data-guide.xml @@ -0,0 +1,40 @@ + + + + Configuration backup and restore + Shinsuke Sugaya + + +
    +

    Here, describes Fess information backup and restore methods.

    +
    +
    + +

    In Administrator account after logging in, click the menu backup and restore.

    + Backup and restore +
    + +

    Click the download link and Fess information output in XML format. Saved settings information is below.

    +
      +
    • The General crawl settings
    • +
    • Web crawl settings
    • +
    • File system Crawl settings
    • +
    • Datastore crawl settings
    • +
    • Label
    • +
    • Path mapping
    • +
    • Web authentication
    • +
    • File system authentication
    • +
    • Request header
    • +
    • Duplicate host
    • +
    • Roll
    • +
    +

    Session information, search log, click log is available in CSV format.

    +

    In the SOLR index data and data being crawled is not backed up. Those data can Fess setting information to crawl after the restore, regenerate. If you need to back up the SOLR index backs solr directory.

    +
    + +

    You can restore settings information, various log in to upload XML output by backup or CSV. To specify the files, please click the restore button on the data.

    +

    If enable overwrite data in XML file configuration information specified when the same data is updating existing data.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/dataCrawlingConfig-guide.xml b/src/site/en/xdoc/9.1/admin/dataCrawlingConfig-guide.xml new file mode 100644 index 000000000..f331705ba --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/dataCrawlingConfig-guide.xml @@ -0,0 +1,159 @@ + + + + Settings for crawling the data store + Sone, Takaaki + Shinsuke Sugaya + + +
    +

    You can crawl data sources such as databases and CSV in Fess. Here are required to store settings.

    +
    +
    + +

    In Administrator account after logging in, click menu data store.

    + Data store configuration +

    As an example, the following table database named testdb MySQL, user name hoge, fuga password connection and the will to make it.

    + +

    Here the data is put something like the following.

    + +
    +
    +
    + +

    Parameter settings example looks like the following.

    + +

    Parameter is a "key = value" format. Description of the key is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    For DB configuration parameter example
    driverDriver class name
    URLURL
    usernameTo connect to the DB user name
    passwordTo connect to the DB password
    SQLWant to crawl to get SQL statement
    +
    + +

    Script configuration example looks like the following.

    + +

    + Parameter is a "key = value" format. + Description of the key is as follows.

    +

    + Side of the value written in OGNL. Close the string in double quotation marks. + Access in the database column name, its value.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Script settings
    URLURLs (links appear in search results)
    hostHost name
    siteSite pass
    titleTitle
    contentContent (string index)
    cacheContent cache (not indexed)
    DigestDigest piece that appears in the search results
    anchorLinks to content (not usually required)
    contentLengthThe length of the content
    lastModifiedContent last updated
    +
    + +

    To connect to the database driver is needed. keep the jar file in webapps/fess/WEB-INF/cmd/lib.

    +
    + +

    Set the following in the webapps/fess/WEB-INF/classes/app.dicon if you see the item value, such as latitude_s in the search results. After adding to $ {doc.latitude_s}, searchResults.jsp;

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "latitude_s","longitude_s" } + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/design-guide.xml b/src/site/en/xdoc/9.1/admin/design-guide.xml new file mode 100644 index 000000000..843ba959e --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/design-guide.xml @@ -0,0 +1,107 @@ + + + + Appearance settings + Shinsuke Sugaya + + +
    +

    Here are settings for the design of search screens.

    +
    +
    + +

    In Administrator account after logging in, click the menu design.

    + Design +

    You can edit the search screen in the screen below.

    + JSP compilation screen +
    + +

    If you want to display in the search results crawl in Fess and registered or modified files to get the search results page (content), write the following.

    + +]]> +

    tstampDate will crawl during registration on the lastModifiedDate modified date of the document. Output date formats follow the fmt:formateDate specification.

    +
    +
    +
    + +

    On the search screen files are available to download and can be removed.

    +
    + +

    You can upload the file to use in the search screen. Image file names are supported are jpg, gif, png, css, and js.

    +
    + +

    Use if you want to specify the file name file to upload. Uploaded if you omit the file name will be used.

    +
    + +

    You can edit the JSP files in the search screen. You can by pressing the Edit button of the JSP file, edit the current JSP files. And pressing the button will default to edit as a JSP file when you install. To keep with the update button in the Edit screen, changes are reflected.

    +

    Following are examples of how to write.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    JSP file that you can edit
    Top page (frame)Is a JSP file search home page. This JSP include JSP file of each part.
    HeaderIt is a JSP file's header.
    FooterThis is the footer JSP files.
    Search results pages (frames)Search result is a list page of JSP files. This JSP include JSP file of each part.
    Search results pages (content)Search results search results list page is a JSP file to represent the part. Is the search results when the JSP file. If you want to customize the search result representation change.
    Search results page (result no)Search results search results list page is a JSP file to represent the part. Is a JSP file when the search result is not used.
    Help pages (frames)Is the help page of JSP files.
    Search error pageIt is a JSP file search error page. If you want to customize the search error expression change.
    File boot pageIs the file boot page JSP file. Is the screen used when displaying Java plug-in is enabled to display the search results by using the file system CROLL.
    Error page (header)Is a JSP file that represents the header of the page.
    Error page (footer)It is a JSP file that represents the footer part of the page.
    Error page (page not found)It is displayed if the page cannot be found error page JSP file.
    Error (System error)JSP error page that appears if the system error is.
    Error pages (redirects)This is the JSP error page displayed when an HTTP redirect occurs.
    Error (bad request)Is the error bad request appears when the JSP file.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/dict-guide.xml b/src/site/en/xdoc/9.1/admin/dict-guide.xml new file mode 100644 index 000000000..b6850bc60 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/dict-guide.xml @@ -0,0 +1,49 @@ + + + + Dictionary settings + Shinsuke Sugaya + + +
    +

    Here are settings for the dictionary.

    +
    +
    + +

    In Administrator account after logging in, click the menu Dictionary. List the various dictionaries available for editing.

    + Dictionary list +
    +
    +
    +

    You can register names, nouns, terminology. Registered click path to user dictionaries and dictionary word list is displayed.

    + User dictionary list +

    Displays the Edit screen and click the word you want to edit.

    + Create a user dictionary + +

    Type a Word to search for.

    +
    + +

    Been searching the words compound word splitting the Word can to hit. For example, you can search any word by entering the full-text search engine and full-text search engine, have split.

    +
    + +

    Enter the word reading in katakana. +   Enter the split when you split. For example, enter Sembene search engine.

    +
    + +

    Enter the words you entered.

    +
    +
    +
    +

    You can register the same meaning words (GB, gigabyte, etc.). Click the path of the synonym dictionary registered on dictionary word list is displayed.

    + Synonym list +

    Displays the Edit screen and click the word you want to edit.

    + Creating a synonym + +

    Type the word being treated as synonyms.

    +
    + +

    Expand the words you enter in the source in the word after the conversion. For example, 'TV', 'TV' and 'television' If you want to convert type 'TV', 'TV' and 'TV' type after conversion.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/document-guide.xml b/src/site/en/xdoc/9.1/admin/document-guide.xml new file mode 100644 index 000000000..967b4b682 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/document-guide.xml @@ -0,0 +1,34 @@ + + + + Index settings + Shinsuke Sugaya + + +
    +

    Here are settings for the index.

    +
    +
    + +

    In Administrator account after logging in, click the menu index.

    + Index settings +
    +
    +
    + +

    You can publish index commit, optimize for server groups.

    +
      +
    • Commit: Solr issued a commit request. Removes the document index files to reflect documents added to commit in a deleted from the index file.
    • +
    • Optimization: Solr issued request optimization. Commit processing in addition, were fragmentation of index files are merged.
    • +
    +

    No problem, manually commit or optimization to do.

    +
    + +

    You can remove a specific search for the session ID. You can remove only the specific documents by specifying the URL.

    +
    + +

    Shown by the number of documents registered in each session. Can verify the results list by clicking the session name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/failureUrl-guide.xml b/src/site/en/xdoc/9.1/admin/failureUrl-guide.xml new file mode 100644 index 000000000..479ad891f --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/failureUrl-guide.xml @@ -0,0 +1,23 @@ + + + + Disaster URL + Shinsuke Sugaya + + +
    +

    Here the failure URL. URL could not be obtained at crawl time are recorded and confirmed as the failure URL.

    +
    +
    + +

    In Administrator account click menu disabled URL after login.

    + Disaster URL +

    Clicking the confirmation link failure URL displayed for more information.

    + Details of the disaster URL +
    + +

    A glance could not crawl the URL and date.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/favoriteLog-guide.xml b/src/site/en/xdoc/9.1/admin/favoriteLog-guide.xml new file mode 100644 index 000000000..e4fc91d0b --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/favoriteLog-guide.xml @@ -0,0 +1,21 @@ + + + + Popular URL log + Shinsuke Sugaya + + +
    +

    Here are popular URL log. When the popular URL log user clicks voting link on the search screen registers as a favorite link. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click the menu popular URL.

    + Users list +
    + +

    Lists popular URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/fileAuthentication-guide.xml b/src/site/en/xdoc/9.1/admin/fileAuthentication-guide.xml new file mode 100644 index 000000000..83d8c1a34 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/fileAuthentication-guide.xml @@ -0,0 +1,44 @@ + + + + Settings for file system authentication + Shinsuke Sugaya + + +
    +

    Crawls using file system here, describes how to set file system authentication is required. Fess is corresponding to a crawl for a shared folder in Windows.

    +
    +
    + +

    In Administrator account after logging in, click the menu file system authentication.

    + File system settings +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Is omitted, the specified file system Kroll set applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. File system Crawl settings specified in that case applies on any port.

    +
    + +

    Select the authentication method. You can use SAMBA (Windows shared folder authentication).

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. SAMBA, the set value of the domain. If you want to write as.

    + +
    + +

    Select the set name to apply the authentication settings for the above file system CROLL. Must be registered ago you file system CROLL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/fileCrawlingConfig-guide.xml b/src/site/en/xdoc/9.1/admin/fileCrawlingConfig-guide.xml new file mode 100644 index 000000000..c43dfeaae --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/fileCrawlingConfig-guide.xml @@ -0,0 +1,103 @@ + + + + Settings for file system crawling + Shinsuke Sugaya + + +
    +

    Describes the settings for crawl here, using file system.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu file.

    + Setting file system Crawl +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple paths. file: or smb: in the specify starting. For example,

    + +

    The so determines. Patrolling below the specified directory.

    +

    So there is need to write URI if the Windows environment path that c:\Documents\taro in file/c: /Documents/taro and specify.

    +

    Windows shared folder, for example, if you want to crawl to host1 share folder crawl settings for smb: (last / to) the //host1/share/. If authentication is in the shared folder on the file system authentication screen set authentication information.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for given path pattern.

    + + + + + + + + + + + + + + + + + + + + +
    IP rings contents list
    Path to crawlCrawl the path for the specified regular expression.
    The path to exclude from being crawledThe path for the specified regular expression does not crawl. The path you want to crawl, even WINS here.
    Path to be searchedThe path for the specified regular expression search. Even if specified path to find excluded and WINS here.
    Path to exclude from searchesNot search the path for the specified regular expression. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, the path to target if you don't crawl less than/home /

    + +

    Also the path to exclude if extension of png want to exclude from

    + +

    It specifies. It is possible to specify multiple line breaks in.

    +

    How to specify the URI handling java.io.File: Looks like:

    + file:/home/taro +c:\memo.txt -> file:/c:/memo.txt +\\server\memo.txt -> file:////server/memo.txt +]]> +
    + +

    You can specify the crawl configuration information.

    +
    + +

    Specify the depth of a directory hierarchy.

    +
    + +

    You can specify the number of documents to retrieve crawl.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the time interval to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/index.xml b/src/site/en/xdoc/9.1/admin/index.xml new file mode 100644 index 000000000..45bc80f38 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/index.xml @@ -0,0 +1,67 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for the management work of the Fess. +

    +
    +
    +

    +This document demonstrates Fess configuration management methods. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    +
    + +

    Management UI to support browser as follows.

    +
      +
    • Internet Explorer 7 or higher
    • +
    • Firefox 4 or more
    • +
    • Chrome (latest version)
    • +
    • Safari (latest version)
    • +
    • Opera 11
    • +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/jobLog-guide.xml b/src/site/en/xdoc/9.1/admin/jobLog-guide.xml new file mode 100644 index 000000000..3abbdfd10 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/jobLog-guide.xml @@ -0,0 +1,25 @@ + + + + Job log + Shinsuke Sugaya + + +
    +

    Job performed results appear as a list.

    +
    +
    + +

    In Administrator account after logging in, click menu users.

    +
    + + List joblog +

    Lists the job run log. You can determine the job name, status, start and finish times. You can also select more information, to check the details of each log.

    +
    + +

    You can check job log contents. Job name, status, start and completion time, displays the results, such as.

    + Job log details +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/labelType-guide.xml b/src/site/en/xdoc/9.1/admin/labelType-guide.xml new file mode 100644 index 000000000..1c92f9bdc --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/labelType-guide.xml @@ -0,0 +1,39 @@ + + + + Setting a label + Shinsuke Sugaya + + +
    +

    Here are settings for the label. Label can classify documents that appear in search results, select the crawl settings in. You can pass even if you do not set the crawl settings in the settings of the label to add labels to specify regular expressions. If you register the label shown select label drop-down box to the right of the search box.

    +
    +
    + +

    In Administrator account after logging in, click the menu label.

    + List of labels + Setting a label +
    +
    +
    + +

    Specifies the name that is displayed when the search label drop-down select.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    + +

    Sets the path to label in the regular expression. You can specify multiple in multiple line description. Notwithstanding the crawl configuration document to match the path specified here, will be labeled.

    +
    + +

    In the path and crawled on regular expressions set from what you want to exclude. You can specify multiple in multiple line description.

    +
    + +

    Specifies the role to view the label.

    +
    + +

    Specifies the order of the labels.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/log-guide.xml b/src/site/en/xdoc/9.1/admin/log-guide.xml new file mode 100644 index 000000000..aabc55899 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/log-guide.xml @@ -0,0 +1,21 @@ + + + + Log file download + Shinsuke Sugaya + + +
    +

    Describes the log files will be output in the Fess download.

    +
    +
    + +

    In Administrator account after logging in, click log file menu.

    + Session information +
    + +

    You can download the log file and click the log file name.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/overlappingHost-guide.xml b/src/site/en/xdoc/9.1/admin/overlappingHost-guide.xml new file mode 100644 index 000000000..23a3b9969 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/overlappingHost-guide.xml @@ -0,0 +1,27 @@ + + + + Duplicate host settings + Shinsuke Sugaya + + +
    +

    Here are settings on the duplicate host. Available when the duplicate host to be treated as the same thing crawling at a different host name. For example, if you want the same site www.example.com and example.com in available.

    +
    +
    + +

    In Administrator account after logging in, click the menu duplicate host.

    + A list of the duplicate host + Duplicate host settings +
    +
    +
    + +

    Specify the canonical host name. Duplicate host names replace the canonical host name.

    +
    + +

    Specify the host names are duplicated. Specifies the host name you want to replace.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/pathMapping-guide.xml b/src/site/en/xdoc/9.1/admin/pathMapping-guide.xml new file mode 100644 index 000000000..cc9676394 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/pathMapping-guide.xml @@ -0,0 +1,30 @@ + + + + Path mapping settings + Shinsuke Sugaya + + +
    +

    Here are settings for path mapping. You can use if you want replaced path mapping links appear in search results.

    +
    +
    + +

    In Administrator account after logging in, click menu path mappings.

    + List of path mapping + Path mapping settings +
    +
    +
    + +

    Path mapping is replaced by parts to match the specified regular expression, replace the string with. When crawling a local filesystem environment may search result links are not valid. Such cases using path mapping, you can control the search results link. You can specify multiple path mappings.

    +
    + +

    Specifies the string you want to replace. How to write aRegular expressions in Java 6To follow.

    +
    + +

    Specifies the string to replace the matched regular expression.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/requestHeader-guide.xml b/src/site/en/xdoc/9.1/admin/requestHeader-guide.xml new file mode 100644 index 000000000..7cc5e28f6 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/requestHeader-guide.xml @@ -0,0 +1,30 @@ + + + + Setting a request header + Shinsuke Sugaya + + +
    +

    Here the request header. Feature request headers request header information added to requests when you get to crawl documents. Available if, for example, to see header information in the authentication system, if certain values are logged automatically.

    +
    +
    + +

    In Administrator account after logging in, click request header menu.

    + A list of request headers + Setting a request header +
    +
    +
    + +

    Specifies the request header name to append to the request.

    +
    + +

    Specifies the request header value to append to the request.

    +
    + +

    Select a Web crawl setting name to add request headers. Only selected the crawl settings in appended to the request header.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/roleType-guide.xml b/src/site/en/xdoc/9.1/admin/roleType-guide.xml new file mode 100644 index 000000000..fd8572cec --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/roleType-guide.xml @@ -0,0 +1,27 @@ + + + + Settings for a role + Shinsuke Sugaya + + +
    +

    Here are settings for the role. Role is selected in the crawl settings, you can classify the document appears in the search results. About how to use theSettings for a rolePlease see the.

    +
    +
    + +

    In Administrator account after logging in, click menu role.

    + The list of roles + Settings for a role +
    +
    +
    + +

    Specifies the name that appears in the list.

    +
    + +

    Specifies the identifier when a classified document. This value will be sent to Solr. Must be alphanumeric characters.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/scheduledJob-guide.xml b/src/site/en/xdoc/9.1/admin/scheduledJob-guide.xml new file mode 100644 index 000000000..e83fcebea --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/scheduledJob-guide.xml @@ -0,0 +1,53 @@ + + + + Job settings + Shinsuke Sugaya + + +
    +

    Here are settings for the job.

    +
    +
    + +

    In Administrator account after logging in, click the job management.

    + Job list + Edit job +
    +
    +
    + +

    It is the name that appears in the list.

    +
    + +

    You can use as an identifier for whether or not to run when the target job command to run directly in the batch, etc.. If the crawl command execution, do not specify 'all'.

    +
    + +

    Configure schedule settings. Run jobs written in script on a schedule you set here.

    +

    Description format describes the format such as Cron seconds minutes date month day year (optional)". For example, "0 0 12? * WED ' for if the weekly Wednesday 12:00 pm job to run. About how to specify the finer "Quartz"Please see.

    +
    + +

    Specifies the script execution environment. At the moment supports only the 'groovy'.

    +
    +
    +

    Written in the language specified in how to perform job run.

    +

    For example, if you want to run crawl jobs only three crawl settings describes as (assuming Web crawl configuration ID 1 and file system Kroll set ID to 1 and 2).

    + +
    +
    +

    To enable records to the job log.

    +
    + +

    In turn treated as crawl jobs. In establishing the system crawl started and stopped.

    +
    + +

    Specifies the enabled or disabled status of the job. If the job will not run.

    +
    + +

    Specifies the display order in the job list.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/searchList-guide.xml b/src/site/en/xdoc/9.1/admin/searchList-guide.xml new file mode 100644 index 000000000..6737f7336 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/searchList-guide.xml @@ -0,0 +1,21 @@ + + + + Search + Shinsuke Sugaya + + +
    +

    Here the search for management.

    +
    +
    + +

    In Administrator account after logging in, click the menu search.

    + Administrative search +
    + +

    You can search by criteria you specify. In the regular search screen role and browser requirements is added implicitly, but do not provide management for search. You can document a certain remove from index from the search results.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/searchLog-guide.xml b/src/site/en/xdoc/9.1/admin/searchLog-guide.xml new file mode 100644 index 000000000..2670e731a --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/searchLog-guide.xml @@ -0,0 +1,21 @@ + + + + Search log list + Shinsuke Sugaya + + +
    +

    Here the search log. When you search in the search screen users search logs are logged. Search log search term or date is recorded. You can also record the URL, then you want the search results to.

    +
    +
    + +

    In Administrator account after logging in, click menu search logs.

    + Search log +
    + +

    Search language and date are listed. You can review and detailed, you click the URL.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/stats-guide.xml b/src/site/en/xdoc/9.1/admin/stats-guide.xml new file mode 100644 index 000000000..a98901b15 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/stats-guide.xml @@ -0,0 +1,21 @@ + + + + Statistics + Shinsuke Sugaya + + +
    +

    Here are statistics. You can search log and click log.

    +
    +
    + +

    In Administrator account after logging in, click the menu statistics.

    + Statistics +
    + +

    You can select the target by selecting the type of report, to make sure. Displayed in order by the specified criteria.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/system-guide.xml b/src/site/en/xdoc/9.1/admin/system-guide.xml new file mode 100644 index 000000000..91dc011ee --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/system-guide.xml @@ -0,0 +1,33 @@ + + + + System settings + Shinsuke Sugaya + + +
    +

    Describes the settings related to Solr, here are registered in the server settings for crawling and Fess. SOLR servers are grouped by file, has been registered.

    +
    +
    + +

    In Administrator account after logging in, click menu system settings.

    + System settings +
    +
    +
    + +

    Update server appears as a running if additional documents, such as the. Crawl process displays the session ID when running. You can safely shut down and Fess server to shut down is not running when shut down. If the process does not terminate if you shut a Fess is running to finish crawling process.

    +

    You can manually crawling under the crawl start button press stop if it is that.

    +
    + +

    Server group name to search for and update available will be shown.

    +
    + +

    In Fess Solr Server conducts a management server and index State States. Whether or not the server state can be access to the Solr Server manages. Whether or not successfully crawl index the State could manage. You can use search server status is in effect, regardless of the State of the index. The crawl Server State is enabled and can index State runs correctly if the preparation or completion. Running start crawl manually index State preparing changes automatically. Server recovery server status and auto-recovery enabled state.

    +
    + +

    You can be sure SOLR server instance state. You can also, for each instance, start, stop, reload request.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/systemInfo-guide.xml b/src/site/en/xdoc/9.1/admin/systemInfo-guide.xml new file mode 100644 index 000000000..1b2c03d58 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/systemInfo-guide.xml @@ -0,0 +1,32 @@ + + + + System information + Shinsuke Sugaya + + +
    +

    Here, you can currently check property information such as system environment variables.

    +
    +
    + +

    In Administrator account after logging in, click system information menu.

    + System information +
    +
    +
    + +

    You can list the server environment variable.

    +
    + +

    You can list the system properties on Fess.

    +
    + +

    Fess setup information available.

    +
    + +

    Is a list of properties to attach when reporting a bug. Extract the value contains no personal information.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/userInfo-guide.xml b/src/site/en/xdoc/9.1/admin/userInfo-guide.xml new file mode 100644 index 000000000..b5c0f470b --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/userInfo-guide.xml @@ -0,0 +1,21 @@ + + + + User log + Shinsuke Sugaya + + +
    +

    Here the user log. Identifies the user when you search in the search screen users the user log in. You can search log and popular URL information and the use. You can disable this feature in the General crawl settings.

    +
    +
    + +

    In Administrator account after logging in, click menu users.

    + Users list +
    + +

    Lists the ID of the user. You can select the search logs or popular URL links, to see a list of each log.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/webAuthentication-guide.xml b/src/site/en/xdoc/9.1/admin/webAuthentication-guide.xml new file mode 100644 index 000000000..1fde73fa0 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/webAuthentication-guide.xml @@ -0,0 +1,48 @@ + + + + Web authentication settings + Shinsuke Sugaya + + +
    +

    Describes Web authentication is required when set against here, using Web crawling. Fess is corresponding to a crawl for BASIC authentication and DIGEST authentication.

    +
    +
    + +

    In Administrator account after logging in, click menu Web authentication.

    + Configuring Web authentication +
    +
    +
    + +

    Specifies the host name of the site that requires authentication. Web crawl settings you specify if applicable in any host name.

    +
    + +

    Specifies the port of the site that requires authentication. Specify-1 to apply for all ports. Web crawl settings you specified and if applicable on any port.

    +
    + +

    Specifies the realm name of the site that requires authentication. Web crawl settings you specify if applicable in any realm name.

    +
    + +

    Select the authentication method. You can use BASIC authentication, DIGEST authentication or NTLM authentication.

    +
    + +

    Specifies the user name to log in authentication.

    +
    + +

    Specifies the password to log into the certification site.

    +
    + +

    Sets if the authentication site login required settings. You can set the workstation and domain values for NTLM authentication. If you want to write as.

    + +
    + +

    Select to apply the above authentication settings Web settings name. Must be registered in advance Web crawl settings.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/webCrawlingConfig-guide.xml b/src/site/en/xdoc/9.1/admin/webCrawlingConfig-guide.xml new file mode 100644 index 000000000..13e735509 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/webCrawlingConfig-guide.xml @@ -0,0 +1,104 @@ + + + + Settings for crawling Web site + Shinsuke Sugaya + + +
    +

    Describes the settings here, using Web crawling.

    +

    Recommends that if you want to index document number 100000 over in Fess crawl settings for one to several tens of thousands of these. One crawl setting a target number 100000 from the indexed performance degrades.

    +
    +
    + +

    In Administrator account after logging in, click menu Web.

    + Web crawl settings +
    +
    +
    + +

    Is the name that appears on the list page.

    +
    + +

    You can specify multiple URLs. http: or https: in the specify starting. For example,

    + +

    The so determines.

    +
    + +

    By specifying regular expressions you can exclude the crawl and search for specific URL pattern.

    + + + + + + + + + + + + + + + + + + + + +
    URL filtering contents list
    URL to crawlCrawl the URL for the specified regular expression.
    Excluded from the crawl URLThe URL for the specified regular expression does not crawl. The URL to crawl, even WINS here.
    To search for URLThe URL for the specified regular expression search. Even if specified and the URL to the search excluded WINS here.
    To exclude from the search URLURL for the specified regular expression search. Unable to search all links since they exclude from being crawled and crawled when the search and not just some.
    +

    For example, http: URL to crawl if not crawl //localhost/ less than the

    + +

    Also be excluded if the extension of png want to exclude from the URL

    + +

    It specifies. It is possible to specify multiple in the line for.

    +
    + +

    You can specify the crawl configuration information.

    +
    + +

    That will follow the links contained in the document in the crawl order can specify the tracing depth.

    +
    + +

    You can specify the number of documents to retrieve crawl. If you do not specify people per 100,000.

    +
    + +

    You can specify the user agent to use when crawling.

    +
    + +

    Specifies the number of threads you want to crawl. Value of 5 in 5 threads crawling the website at the same time.

    +
    + +

    Is the interval (in milliseconds) to crawl documents. 5000 when one thread is 5 seconds at intervals Gets the document.

    +

    Number of threads, 5 pieces, will be to go to and get the 5 documents per second between when 1000 millisecond interval,. Set the adequate value when crawling a website to the Web server, the load would not load.

    +
    + +

    You can search URL in this crawl setting to weight. Available in the search results on other than you want to. The standard is 1. Priority higher values, will be displayed at the top of the search results. If you want to see results other than absolutely in favor, including 10,000 sufficiently large value.

    +

    Values that can be specified is an integer greater than 0. This value is used as the boost value when adding documents to Solr.

    +
    + +

    You can control only when a particular user role can appear in search results. You must roll a set before you. For example, available by the user in the system requires a login, such as portal servers, search results out if you want.

    +
    + +

    You can label with search results. Search on each label, such as enable, in the search screen, specify the label.

    +
    + +

    Crawl crawl time, is set to enable. If you want to avoid crawling temporarily available.

    +
    +
    +
    + +

    Fess and crawls sitemap file, as defined in the URL to crawl. Sitemaphttp://www.sitemaps.org/ Of the specification. Available formats are XML Sitemaps and XML Sitemaps Index the text (URL line written in).

    +

    Site map the specified URL. Sitemap is a XML files and XML files for text, when crawling that URL of ordinary or cannot distinguish between what a sitemap. Because the file name is sitemap.*.xml, sitemap.*.gz, sitemap.*txt in the default URL as a Sitemap handles (in webapps/fess/WEB-INF/classes/s2robot_rule.dicon can be customized).

    +

    Crawls sitemap file to crawl the HTML file links will crawl the following URL in the next crawl.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/admin/wizard-guide.xml b/src/site/en/xdoc/9.1/admin/wizard-guide.xml new file mode 100644 index 000000000..ba4c23969 --- /dev/null +++ b/src/site/en/xdoc/9.1/admin/wizard-guide.xml @@ -0,0 +1,31 @@ + + + + Settings Wizard + Sone, Takaaki + + +
    +

    Introduction to the Configuration Wizard.

    +

    You can use Settings Wizard, to set you up on the Fess.

    +
    +
    + +

    In Administrator account after logging in, click menu Settings Wizard.

    + Settings Wizard +

    Do the crawl settings. + Crawl settings is to register a URI to look for. + The crawl settings name please put name of any easy to identify. Put the URI part de-indexed, want to search for.

    + Crawl settings +

    For example, if you want and search for http://fess.codelibs.org/, less looks like.

    + Crawl settings example +

    The type, such as c:\Users\taro file.

    +

    In this setting is complete. Crawl start button press the start crawling. Not start until in the time specified in the scheduling settings by pressing the Finish button if the crawl.

    + Crawl started +
    + +

    Settings in the Setup Wizard you can change from crawl General, Web, file system.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/db.xml b/src/site/en/xdoc/9.1/config/db.xml new file mode 100644 index 000000000..2a62426a9 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/db.xml @@ -0,0 +1,128 @@ + + + + Database changes + Shinsuke Sugaya + + +
    +

    Provides binaries to use H2 Database with MySQL database. You can use the other database in to change the settings using the source code and build it.

    +
    +
    + +

    The MySQL character code setting. /etc/mysql/my.cnf and the added must have the following settings.

    + +
    + +

    Download MySQL binaries and expand.

    +
    + +

    Create a database.

    + create database fess_db; +mysql> grant all privileges on fess_db.* to fess_user@localhost identified by 'fess_pass'; +mysql> create database fess_robot; +mysql> grant all privileges on fess_robot.* to s2robot@localhost identified by 's2robot'; +mysql> FLUSH PRIVILEGES; +]]> +

    Create a table in the database. DDL file is located in extension/mysql.

    + +

    webapps/fess/WEB-INF/lib on mysql driver jar.

    + +

    Edit the webapps/fess/WEB-INF/classes/jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_db?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "fess_user" + "fess_pass" + + + + 600 + 10 + true + + + + + +]]> +

    Edit the webapps/fess/WEB-INF/classes/s2robot_jdbc.dicon.

    + + + + + + + + + "com.mysql.jdbc.Driver" + + + "jdbc:mysql://localhost:3306/fess_robot?" + + "noDatetimeStringSync=true&" + + "zeroDateTimeBehavior=convertToNull&" + + "useUnicode=true&characterEncoding=UTF-8&" + + "autoReconnect=true" + + "s2robot" + "s2robot" + + + + 600 + 10 + true + + @java.sql.Connection@TRANSACTION_REPEATABLE_READ + + + + + + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/filesize.xml b/src/site/en/xdoc/9.1/config/filesize.xml new file mode 100644 index 000000000..ff556b6fc --- /dev/null +++ b/src/site/en/xdoc/9.1/config/filesize.xml @@ -0,0 +1,29 @@ + + + + File size you want to crawl settings + Shinsuke Sugaya + + +
    +

    You can specify the file size limit crawl of Fess. In the default HTML file is 2.5 MB, otherwise handles up to 10 m bytes. Edit the webapps/fess/WEB-INF/classes/s2robot_contentlength.dicon if you want to change the file size handling. Standard s2robot_contentlength.dicon is as follows.

    + + + + + 10485760L + + "text/html" + 2621440L + + + +]]> +

    Change the value of defaultMaxLength if you want to change the default value. Dealing with file size can be specified for each content type. Describes the maximum file size to handle text/HTML and HTML files.

    +

    Note the amount of heap memory to use when changing the maximum allowed file size handling. About how to set upMemory-relatedPlease see the.

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/geosearch.xml b/src/site/en/xdoc/9.1/config/geosearch.xml new file mode 100644 index 000000000..dfdb560eb --- /dev/null +++ b/src/site/en/xdoc/9.1/config/geosearch.xml @@ -0,0 +1,23 @@ + + + + Dios arch + Shinsuke Sugaya + + +
    +

    You can document with latitude and longitude location information in conjunction with Google maps, including the use of Dios arch.

    +
    +
    + +

    Location is defined as a feed that contains the location information. + When generating the index in Solr latitude longitude set to location feeds in formats such as 45.17614,-93.87341, register the document. + Also sets the value as the latitude_s and longitude_s fields if you want to display latitude and longitude as a search result. * _s is available as a dynamic field of Solr string.

    +
    + +

    During the search specifies in the request parameter to latitude and longitude, the distance. + View the results in the distance (km) specified by distance-based latitude information (latitude, longitude). Latitude and longitude and distances is treated as double.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/index-backup.xml b/src/site/en/xdoc/9.1/config/index-backup.xml new file mode 100644 index 000000000..5efd742f3 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/index-backup.xml @@ -0,0 +1,13 @@ + + + + Index management + Shinsuke Sugaya + + +
    +

    The index data is managed by Solr. Backup from the Administration screen of the Fess, and cases will be in the size and number of Gigabit can not index data.

    +

    If you need to index data backup stopped the Fess from back solr/core1/data and solr/core1-suggest/data directories. Also, index data backed up to restore to undo.

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/index.xml b/src/site/en/xdoc/9.1/config/index.xml new file mode 100644 index 000000000..b95ee4fd9 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for setting work of the Fess. +

    +
    +
    +

    +This document demonstrates setting Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/install-on-tomcat.xml b/src/site/en/xdoc/9.1/config/install-on-tomcat.xml new file mode 100644 index 000000000..17ef0f1b6 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/install-on-tomcat.xml @@ -0,0 +1,43 @@ + + + + Install to an existing Tomcat + Shinsuke Sugaya + + +
    +

    + The standard distribution of Fess Tomcat is distributed in the deployed State. + Because Fess is not dependent on Tomcat, deploying on any Java application server is available. + Describes how to deploy a Fess Tomcat here is already available.

    +
    +
    +

    Fess ServerHereFrom the download.

    +
    +
    +

    + Expand the downloaded Fess server. + Expanded Fess Server home directory to $FESS_HOME. + $TOMCAT_HOME the top directory of an existing Tomcat 7. + Copy the Fess Server data.

    + +

    + If you have, such as changing the destination file diff commands, updates your diff only applies. +

    +
    +
    +

    to start the Tomcat in normal as well as in startup.* / http://localhost:8080/Fess To access the.

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/logging.xml b/src/site/en/xdoc/9.1/config/logging.xml new file mode 100644 index 000000000..7f583fbb9 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/logging.xml @@ -0,0 +1,61 @@ + + + + Log settings + Shinsuke Sugaya + + +
    +

    The following summarizes the log file to output the Fess.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    Log file list
    File nameContents
    webapps/fess/WEB-INF/logs/fess.outFess server log. Output logging operation in the management and search screens, etc.
    webapps/fess/WEB-INF/logs/fess_crawler.outCrawl log. Crawling log output.
    logs/Catalina.outLog of the Fess Server (Tomcat). Tomcat logs are output.
    logs/SOLR.logLog server Fess (Solr). SOLR relevant log output.
    +

    If you encounter problems to work check the log.

    +
    +
    +

    Sets the output log information is at webapps/fess/WEB-INF/classes/log4j.xml. By default output INFO level.

    +

    For example, better Fess up to document for Solr log if you want to output in log4j.xml disconnect the commented-out section below.

    + + + +]]> +

    See the Log4J documentation if you need detailed settings for the log output.

    + +

    You can check by changing the logging level of the HttpClient if you want to check HTTP access when crawling content to solve the problem.

    + + + +]]> +
    + +

    to crawl, and change the priority tag from info to debug the output debug log of all things.

    + ]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/memory-config.xml b/src/site/en/xdoc/9.1/config/memory-config.xml new file mode 100644 index 000000000..6d8a2a54b --- /dev/null +++ b/src/site/en/xdoc/9.1/config/memory-config.xml @@ -0,0 +1,52 @@ + + + + Use memory-related settings + Shinsuke Sugaya + + +
    +

    Set the maximum memory per process in Java. So, do not use the upper memory in the process also had 8 GB of physical memory on the server. Memory consumption depending on the number of crawl threads and interval will also change significantly. If not enough memory please change settings in the subsequent procedure.

    +
    +
    +

    If the contents of the crawl settings cause OutOfMemory error similar to the following.

    + +

    Increase the maximum heap memory occur. bin/setenv. [sh | bat] to (in this case maximum value set to 1 G)-xmx1g to change.

    + +
    +
    +

    + Crawler side memory maximum value can be changed. + The default is 512 m.

    +

    + Unplug the commented out webapps/fess/WEB-INF/classes/fess.dicon crawlerJavaOptions to change, change the-xmx1g (in this case maximum value set to 1 G). +

    + new String[] { +"-Djava.awt.headless=true", +"-server", +"-Xmx1g", +"-XX:MaxPermSize=128m", +"-XX:-UseGCOverheadLimit", +"-XX:+UseConcMarkSweepGC", +"-XX:CMSInitiatingOccupancyFraction=75", +"-XX:+CMSIncrementalMode", +"-XX:+CMSIncrementalPacing", +"-XX:CMSIncrementalDutyCycleMin=0", +"-XX:+UseParNewGC", +"-XX:+UseStringCache", +"-XX:+UseTLAB", +"-XX:+DisableExplicitGC" +} +]]> +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/mobile-device.xml b/src/site/en/xdoc/9.1/config/mobile-device.xml new file mode 100644 index 000000000..98235ae27 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/mobile-device.xml @@ -0,0 +1,17 @@ + + + + Mobile device information settings + Shinsuke Sugaya + + +
    +

    The mobile device informationValueEngine Inc.That provided more available. If you want to use the latest mobile device information downloaded device profile save the removed _YYYY-MM-DD and webapps/fess/WEB-INF/classes/device. After the restart to enable change.

    + ProfileData.csv +UserAgent_YYYY-MM-DD.csv -> UserAgent.csv +DisplayInfo_YYYY-MM-DD.csv -> DisplayInfo.csv +]]> +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/path-encoding.xml b/src/site/en/xdoc/9.1/config/path-encoding.xml new file mode 100644 index 000000000..b8aa16ac6 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/path-encoding.xml @@ -0,0 +1,24 @@ + + + + Path encoding change + Sone, Takaaki + + +
    +

    in non-HTML file, different character set precedents and the file name if the search result link text is garbled.

    +

    For example, test.txt file contents are written in UTF-8, file name Shift_JIS, the link text is garbled.

    +
    +
    +

    For example by revising the webapps/fess/WEB-INF/classes/s2robot_transformer.dicon as shown below, to resolve paths in Shift_JIS.

    + + "fessFileTransformer" + true + "Shift_JIS" + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/9.1/config/pdf-with-password.xml b/src/site/en/xdoc/9.1/config/pdf-with-password.xml new file mode 100644 index 000000000..0a8f4e317 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/pdf-with-password.xml @@ -0,0 +1,89 @@ + + + + Password-protected PDF + Shinsuke Sugaya + + +
    +

    + You should password files to register the settings file to PDF password is configured to search for. +

    +
    +
    +

    + First of all, create the webapps/fess/WEB-INF/classes/s2robot_extractor.dicon. + This is test _ ~ is a pass that password set to a.pdf file. + If you have multiple files, multiple settings in addPassword.

    + + + + + + + + + + + + + ".*test_.*.pdf" + + "pass" + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + +... +]]> +

    Next, edit the following to webapps/fess/WEB-INF/classes/s2robot_rule.dicon.

    + + "fsFileRule" + + + fessFileTransformer + + + true + + "url" + "file:.*" + + + "mimeType" + + + "(application/xml" ++ "|application/xhtml+xml" ++ "|application/rdf+xml" ++ "|application/pdf" ++ "|text/xml" ++ "|text/xml-external-parsed-entity" ++ "|text/html)" + + +
    +... +]]> +

    Fess up after setting the above and please run the crawl. How to use Basic remains especially.

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/protwords.xml b/src/site/en/xdoc/9.1/config/protwords.xml new file mode 100644 index 000000000..3a34b20b4 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/protwords.xml @@ -0,0 +1,17 @@ + + + + Stemming settings + Sone, Takaaki + + +
    +

    In Fess when indexing and searching the stemming process done.

    +

    This is to normalize the English word processing, for example, words such as recharging and rechargable is normalized to form recharg. Hit and even if you search by recharging the word this word rechargable, less search leakage is expected.

    +
    +
    +

    You may not intended for the stemming process basic rule-based processing, normalization is done. For example, Maine (state name) Word will be normalized in the main.

    +

    In this case, by adding Maine to protwords.txt, you can exclude the stemming process.

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/proxy.xml b/src/site/en/xdoc/9.1/config/proxy.xml new file mode 100644 index 000000000..c0963c8c9 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/proxy.xml @@ -0,0 +1,87 @@ + + + + Proxy settings + Shinsuke Sugaya + + + +
    +

    If you like crawling external sites from within the intranet firewall may end up blocked crawl. Set the proxy for the crawler in that case.

    +
    +
    +

    Proxy is set in to create webapps/Fess/Web-INF/classes/s9robot_client.dicon with the following contents.

    + + + + + + + + @org.apache.http.client.params.CookiePolicy@BEST_MATCH + clientConnectionManager + "プロキシサーバー名(ex. 192.168.1.1)" + プロキシサーバーのポート(ex. 8080) (" は不要) + + + + internalHttpClient + 5 + 500 + + + + 5 + @java.util.concurrent.TimeUnit@MINUTES + + 200 + + 20 + + + + + "UTF-8" + + + + "UTF-8" + + + + + + {"http:.*", "https:.*"} + httpClient + + + "file:.*" + fsClient + + + "smb:.*" + smbClient + + + + +]]> +
    + + +
    diff --git a/src/site/en/xdoc/9.1/config/replication.xml b/src/site/en/xdoc/9.1/config/replication.xml new file mode 100644 index 000000000..0cbf69304 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/replication.xml @@ -0,0 +1,56 @@ + + + + Setting up replication + Shinsuke Sugaya + + +
    +

    Sets the replication of the index Solr replication features. You can distribute load during indexing to build two in Fess of the crawl and index creation and search for Fess servers.

    +
    +
    + +

    Fess, download and install the. When you install MasterServer named host./ /opt/fess_master To assume you installed. Edit the SOLR/core1/conf/solrconfig.XML like the following.

    + + + startup + optimize + optimize + schema.xml,stopwords.txt,stopwords_ja.txt,elevate.xml, + stoptags_ja.txt,synonyms.txt,mapping_ja.txt,mapping-FoldToASCII.txt, + mapping-ISOLatin1Accent.txt,protwords.txt,compositePOS.txt,spellings.txt, + currency.xml + + 1 + +... +]]> +

    Register the crawl settings as well as Fess starts after the normal construction. Steps to building the index for Fess remains especially as normal building procedures.

    +
    + +

    Fess, download and install the./ /opt/fess_slave To assume you installed. Edit the SOLR/core1/conf/solrconfig.XML like the following.

    + + + http://MasterServer:8080/solr/core1/replication + 00:00:60 + internal + 5000 + 10000 + solradmin + solradmin + + +... +]]> +

    Start the Fess.

    +
    + +

    Until the above settings copied to search for Fess index crawl after the optimization (optimize) the index for Fess.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/role-setting.xml b/src/site/en/xdoc/9.1/config/role-setting.xml new file mode 100644 index 000000000..3c973b9cc --- /dev/null +++ b/src/site/en/xdoc/9.1/config/role-setting.xml @@ -0,0 +1,85 @@ + + + + Setting role-based search + Shinsuke Sugaya + + +
    +

    You can divide out search results in Fess in any authentication system authenticated users credentials to. For example, find rolls a does appears role information in search results with the roles a user a user b will not display it. By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    In role-based search of the Fess roll information available below.

    +
      +
    • Request parameter
    • +
    • Request header
    • +
    • Cookies
    • +
    • J2EE authentication information
    • +
    +

    To save authentication information in cookies for authentication when running of Fess in portal and agent-based single sign-on system domain and path that can retrieve role information. You can also reverse proxy type single sign-on system access to Fess adding authentication information in the request headers and request parameters to retrieve role information.

    +
    +
    +

    Describes how to set up role-based search using J2EE authentication information.

    + +

    conf/Tomcat-users.XML the add roles and users. This time the role1 role perform role-based search. Login to role1.

    + + + + + + + + + +]]> +
    + +

    sets the webapps/fess/WEB-INF/classes/fess.dicon shown below.

    + + + {"guest"} + + : +]]> +

    You can set the role information by setting the defaultRoleList, there is no authentication information. Do not display the search results need roles for users not logged in you.

    +
    + +

    sets the webapps/fess/WEB-INF/web.xml shown below.

    + + + Fess Authentication + /login/login + + + fess + role1 + + + : + + fess + + + + role1 + + : +]]> +
    + +

    Fess up and log in as an administrator. From the role of the menu set name Role1 (any name) and value register role at role1. After the crawl settings want to use in the user with the role1 in, crawl Crawl Settings select Role1.

    +
    + +

    Log out from the management screen. log in as user Role1. A successful login and redirect to the top of the search screen.

    +

    Only thing was the Role1 role setting in the crawl settings search as usual, and displayed.

    +

    Also, search not logged in will be search by guest user.

    +
    + +

    Whether or not logged out, logged in a non-Admin role to access http://localhost:8080/fess/admin screen appears. By pressing the logout button will log out.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/server-port.xml b/src/site/en/xdoc/9.1/config/server-port.xml new file mode 100644 index 000000000..d7a6da825 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/server-port.xml @@ -0,0 +1,40 @@ + + + + Ports changes + Shinsuke Sugaya + + +
    +

    Fess by default, you use the port 8080. Change in the following steps to change.

    + +

    Change the port Tomcat is Fess available. Modifies the following described conf/server.xml changes.

    +
      +
    • 8080: HTTP access port
    • +
    • 8005: shut down port
    • +
    • 8009: AJP port
    • +
    • : SSL HTTP access port 8443 (the default is off)
    • +
    • 19092: database port (use h2database)
    • +
    +
    + +

    May need to change if you change the Tomcat port using the settings in the standard configuration, the same Solr-Tomcat, so Fess Solr server referenced information.

    +

    change the webapps/fess/WEB-INF/classes/app.dicon the following points.

    + "http://localhost:8080/manager/text/" +]]> +

    change the webapps/fess/WEB-INF/classes/solrlib.dicon the following points.

    + "http://localhost:8080/solr/core1" +]]> +

    change the SOLR/core1/conf/solrconfig.XML the following points.

    + http://localhost:8080/solr/core1-suggest +]]> +

    + Note: to display the error on search and index update: cannot access the Solr server and do not change if you change the Tomcat port similar to the above ports. +

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/solr-dynamic-field.xml b/src/site/en/xdoc/9.1/config/solr-dynamic-field.xml new file mode 100644 index 000000000..d03dae01b --- /dev/null +++ b/src/site/en/xdoc/9.1/config/solr-dynamic-field.xml @@ -0,0 +1,32 @@ + + + + Dynamic fields + Shinsuke Sugaya + + +
    +

    SOLR is document items (fields) for each to the schema defined in order to register. Available in Fess Solr schema is defined in solr/core1/conf/schema.xml. dynamic fields and standard fields such as title and content can be freely defined field names are defined. Advanced parameter values see a Solr document.

    +
    +
    +

    I think scenes using the dynamic field of many, in database scrawl's, such as registering in datastore crawl settings. How to register dynamic fields in database scrawl by placing the script other_t = hoge hoge column data into Solr other_t field.

    +

    You need to add a field to use to retrieve data that is stored in the dynamic field next to the webapps/fess/WEB-INF/classes/app.dicon. Add the other_t.

    + + new String[]{"url", "host", "site", + "title", "content", "contentLength", "lastModified", "mimetype", + "label", "segment", "other_t" } + +]]> +

    You must add a field to use to retrieve data that is stored in the dynamic field from Solr. Add the other_t.

    + + new String[]{"id", "score", "boost", + "contentLength", "host", "site", "lastModified", "mimetype", + "tstamp", "title", "digest", "url", "other_t" } + +]]> +

    Edit the JSP file has made returns from Solr in the above settings, so to display on the page. Login to the manage screen, displays the design. Display of search results the search results displayed on the page (the content), so edit the JSP file. where you want to display the other_t value in $ {f:h(doc.other_t)} and you can display the value registered in.

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/solr-failure.xml b/src/site/en/xdoc/9.1/config/solr-failure.xml new file mode 100644 index 000000000..6474fcdfb --- /dev/null +++ b/src/site/en/xdoc/9.1/config/solr-failure.xml @@ -0,0 +1,37 @@ + + + + SOLR failure operation + Shinsuke Sugaya + + +
    +

    Solr server group in the Fess, managing multiple groups. Change the status of servers and groups if the server and group information that keeps a Fess, inaccessible to the Solr server.

    +

    SOLR server state information can change in system setting. maxErrorCount, maxRetryStatusCheckCount, maxRetryUpdateQueryCount and minActiveServer can be defined in the webapps/fess/WEB-INF/classes/solrlib.dicon.

    +
    +
    +
      +
    • When SOLR group within Solr server number of valid state minActiveServer less than Solr group will be disabled.
    • +
    • Solr server number of valid state is minActiveServer following group in the SOLR Solr group into an invalid state if is not, you can access to the Solr server, disable Solr server status maxRetryStatusCheckCount check to Solr server status change from the disabled state the valid state. The valid state not changed and was able to access Solr Server index corrupted state.
    • +
    • Disable Solr group is not available.
    • +
    • SOLR group to enable States to the group in the Solr Solr server status change enabled in system settings management screen.
    • +
    +
    +
    +
      +
    • Search queries can send valid Solr group.
    • +
    • Search queries will be sent only to valid Solr server.
    • +
    • Send a search query to fewer available if you register a Solr server multiple SOLR group in the Solr server.
    • +
    • The search query was sent to the SOLR server fails maxErrorCount than Solr server modifies the disabled state.
    • +
    +
    +
    +
      +
    • Update queries you can send valid state Solr group.
    • +
    • Update query will be sent only to valid Solr server.
    • +
    • If multiple Solr servers are registered in the SOLR group in any valid state Solr server send the update query.
    • +
    • Is sent to the SOLR Server update query fails maxRetryUpdateQueryCount than Solr server modifies the index corrupted state.
    • +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/tokenizer.xml b/src/site/en/xdoc/9.1/config/tokenizer.xml new file mode 100644 index 000000000..296a09c3b --- /dev/null +++ b/src/site/en/xdoc/9.1/config/tokenizer.xml @@ -0,0 +1,47 @@ + + + + Settings for the index string extraction + Sone, Takaaki + + + +
    +

    You must isolate the document in order to register as the index when creating indexes for the search. Tokenizer is used for this.

    +

    Basically, carved by the tokenizer units smaller than go find no hits. For example, statements of living in Tokyo, Japan. Was split by the tokenizer now, this statement is in Tokyo, living and so on. In this case, in Tokyo, Word search, you will get hit. However, when performing a search with the word 'Kyoto' will not be hit. For selection of the tokenizer is important.

    +

    You can change the tokenizer by setting the schema.xml analyzer part is if the Fess in the default StandardTokenizer CJKBigramFilter used.

    + + +

    StandardTokenizer CJKBigramFilter index bi-gram, in other words two characters to like Japan Japanese multibyte string creates. In this case, can't find one letter words.

    +
    + + +

    StandardTokenizer creates index uni-gram, in other words one by one for the Japan language of multibyte-character strings. Therefore, the less search leakage. Also, with StandardTokenizer can't CJKTokenizer the search query letter to search to. However, please note that the index size increases.

    +

    The following example to change the analyzer part like solr/core1/conf/schema.xml, you can use the StandardTokenizer.

    + + + + + + + : + + + + + : +]]> +

    Also, useBigram is enabled by default in the webapps/fess/WEB-INF/classes/app.dicon change to false.

    + + true + : +]]> +

    After the restart the Fess.

    +
    +
    + + +
    diff --git a/src/site/en/xdoc/9.1/config/use-libreoffice.xml b/src/site/en/xdoc/9.1/config/use-libreoffice.xml new file mode 100644 index 000000000..e4d25514a --- /dev/null +++ b/src/site/en/xdoc/9.1/config/use-libreoffice.xml @@ -0,0 +1,85 @@ + + + + Use of LibreOffice + Shinsuke Sugaya + + +
    +

    + It is possible to crawl using the Apache POI Fess environmental standard in MS Office system document. + You can crawl Office system document regarding LibreOffice, OpenOffice, do even more accurate text extraction from documents.

    +
    +
    +

    JodConverter Fess server install. from http://jodconverter.googlecode.com/jodconverter-core-3.0-Beta-4-Dist.zipThe download. Expand and copy the jar file to Fess server.

    + +

    Create a s2robot_extractor.dicon to the next.

    + +

    s2robot_extractor.dicon effective jodExtractor with following contents.

    + + + + + + + + + + + + + + + + "http://xml.org/sax/features/namespaces" + "false" + + + + + + + officeManagerConfiguration.setOfficeHome("/usr/lib/libreoffice") + .buildOfficeManager() + + + + + + { +"application/msword", +"application/vnd.ms-excel", +"application/vnd.ms-powerpoint", +"application/vnd.openxmlformats-officedocument.wordprocessingml.document", +"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", +"application/vnd.openxmlformats-officedocument.presentationml.presentation" + } + jodExtractor + +... +]]> +

    Index to generate the settings later, usually crawled into the street.

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/config/windows-service.xml b/src/site/en/xdoc/9.1/config/windows-service.xml new file mode 100644 index 000000000..1361b4a77 --- /dev/null +++ b/src/site/en/xdoc/9.1/config/windows-service.xml @@ -0,0 +1,54 @@ + + + + Register for the Windows service + Shinsuke Sugaya + + +
    +

    You can register the Fess as a Windows service in a Windows environment. How to register a service is similar to the Tomcat.

    + +

    Because if you registered as a Windows service, the crawling process is going to see Windows system environment variablesIs Java JAVA_HOME environment variables for the system to register, As well as Add %JAVA_HOME%\bin to PathYou must.

    +
    + +

    to edit the webapps \fess\WEB-INF\classes\fess.dicon, remove the-server option.

    + + + new String[] { + "-Djava.awt.headless=true", "-XX:+UseGCOverheadLimit", + "-XX:+UseConcMarkSweepGC", "-XX:+CMSIncrementalMode", + "-XX:+UseTLAB", "-Xmx512m", "-XX:MaxPermSize=128m" + } + +]]> +
    + +

    First, after installing the Fess from the command prompt service.bat performs (such as Vista to launch as administrator you must). Fess was installed on C:\Java\fess-server-9.1.0.

    + cd C:\Java\fess-server-9.1.0\bin +> service.bat install fess +... +The service 'fess' has been installed. +]]> +
    + +

    By making the following you can review properties for Fess. To run the following, Tomcat Properties window appears.

    + tomcat7w.exe //ES//fess +]]> +
    + +

    Control Panel - to display the management tool in administrative tools - services, you can set automatic start like normal Windows services.

    +
    +
    +
    + +

    Distributed in the Fess is 64-bit binaries for Windows Tomcat builds based on. If you use 32-bit WindowsTomcat Of the site from, such as 32-bit Windows zip, please replace tomcat7.exe, tomcat7w.exe, tcnative-1.dll.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/install/index.xml b/src/site/en/xdoc/9.1/install/index.xml new file mode 100644 index 000000000..6b2adcd9b --- /dev/null +++ b/src/site/en/xdoc/9.1/install/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is subject to users responsible for Fess installation and uninstallation tasks. +

    +
    +
    +

    +This document shows how to install Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.1/install/install.xml b/src/site/en/xdoc/9.1/install/install.xml new file mode 100644 index 000000000..a6f0dc25c --- /dev/null +++ b/src/site/en/xdoc/9.1/install/install.xml @@ -0,0 +1,131 @@ + + + + Installation + + + + + +
    +

    +Fess can be available in these environments. +

    +
      +
    • OS: Windows, Unix, and Java can run OS environment
    • +
    • Java: Java 7 or more
    • +
    +

    +If Java is not installed in the environment we want to Fess from http://java.sun.com/ get JDK for Java 7 and above, please install. +

    +
    +
    +

    +http://sourceforge.jp/projects/Fess/releases/ From the download the latest Fess package. +

    +
    +
    +

    +Expand the downloaded fess-server-x.y.zip. +If you installed in the UNIX environment, in the bin added the performing rights to a script.

    + +
    +
    +

    +Administrator account is managed by the application server. Fess Server standard available Tomcat, as well as to the user changing the Tomcat. +Modify the password for the admin account of the conf/tomcat-user.xml if you want to change.

    + +]]> +

    +see the Tomcat documentation or JAAS authentication specification to use Tomcat-user.XML file management method other than. +

    +
    +
    +

    +To access the Solr into Fess server is password becomes necessary. +Change the default passwords in production, etc.

    +

    +How to change the password, you must first change the password attribute of the conf/tomcat-user.xml solradmin. +

    + +]]> +

    +Modifies the following three files webapps/fess/WEB-INF/classes/solrlib.dicon, fess_suggest.dicon and solr/core1/conf/solrconfig.xml. +Write what you specified in tomct-user.XML to the following password.

    +

    +modify the following areas of the solrlib.dicon. +

    + + "solradmin" + "solradmin" + +]]> +

    +fess_suggest.dicon is the following. +

    + + "solradmin" + "solradmin" + +]]> +

    +where the following is the solrconfig.Xml. +

    + + + + http://localhost:8080/solr/core1-suggest + + solradmin + solradmin + +]]> +
    +
    +

    +To manage can manage Solr from Fess server deployed on Tomcat context need a password. +Change the default passwords in production, etc.

    +

    +How to change password the change manager conf/tomcat-user.xml password attributes. +

    + + +]]> +

    +Describes the provided password webapps/fess/WEB-INF/classes/app.dicon the following points tomcat-user.xml to the next. +

    + + + "solrServer1" + "http://localhost:8080/manager/text/" + "/solr" + "manager" + "manager" + +]]> +
    +
    +

    +Login at return path settings, in the use encryption and decryption. +Please change the default password, especially in production.

    +

    +How to change the change the key value in the webapps/fess/WEB-INF/classes/app.dicon. +Set the 16-character alphanumeric.

    + +"1234567890123456" +]]> +
    + +
    diff --git a/src/site/en/xdoc/9.1/install/run.xml b/src/site/en/xdoc/9.1/install/run.xml new file mode 100644 index 000000000..ba5a8b3b2 --- /dev/null +++ b/src/site/en/xdoc/9.1/install/run.xml @@ -0,0 +1,41 @@ + + + + Start and stop + + + + + +
    +

    +to run the startup script, start the Fess. +

    + + +

    +access to / http://localhost:8080/Fess ensures startup. +

    + +

    +Management UI is / http://localhost:8080/fess/admin. +Default Administrator account user name / password is admin/admin. +Administrator account is managed by the application server. +In the management UI of the Fess, authenticate with the application server in fess role available as an administrator.

    +
    +
    +

    +Fess to stop the running shutdown scripts. +

    + + +

    +Crawl or may take a while to completely stop during the index creation if you. +

    +
    + +
    diff --git a/src/site/en/xdoc/9.1/install/uninstall.xml b/src/site/en/xdoc/9.1/install/uninstall.xml new file mode 100644 index 000000000..d95af5b33 --- /dev/null +++ b/src/site/en/xdoc/9.1/install/uninstall.xml @@ -0,0 +1,19 @@ + + + + Uninstall + + + + + +
    +

    +You can uninstall by deleting the installation stops after the Fess fess-server-x.y directory. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/index.xml b/src/site/en/xdoc/9.1/user/index.xml new file mode 100644 index 000000000..52bdaa14c --- /dev/null +++ b/src/site/en/xdoc/9.1/user/index.xml @@ -0,0 +1,55 @@ + + + + Introduction + Shinsuke Sugaya + + +
    +

    +This document is targeted to users to Fess. +

    +
    +
    +

    +This document demonstrates how to search in the Fess. +Basic knowledge of computer operation becomes necessary.

    +
    +
    +

    +Download, professional services, support, and other developer information, visit the following. +

    + +
    +
    +

    +Technical questions about our products, don't produce the document in solution, visit the following. +

    + + +

    +If you need commercial support, maintenance and technical support for this productN2SM, Inc....To consult. +

    +
    +
    +
    +

    +About the effectiveness of the Web site's third party in the Fess project, described in this document has no responsibility. +The Fess project through any such site or resource available content, advertising, products, services, and other documents regarding assumes no responsibility, obligations, guarantees. +For the Fess project through such sites or resources and use of available content, advertising, products, services, and other documents, or or credit, related to it caused or alleged, any injury or damage assumes no responsibility or obligation.

    +
    +
    +

    +Fess project is committed to the improvement of this document, and welcomes comments from readers, such as proposed. +

    + +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/json-response.xml b/src/site/en/xdoc/9.1/user/json-response.xml new file mode 100644 index 000000000..b6fcd3d7e --- /dev/null +++ b/src/site/en/xdoc/9.1/user/json-response.xml @@ -0,0 +1,184 @@ + + + + Search results for JSON + Shinsuke Sugaya + + +
    +

    Outputs a JSON results of Fess. You must in order to emit JSON enabled keep JSON response in the General Management screen crawl settings.

    + +

    To get JSON output http://localhost:8080/Fess/JSON? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    callbackIf you use the JSONP callback name. You do not need to specify if you do not use the JSONP.
    +
    + +

    Returns the following response.

    + fess<\u002Fem> \u5168\u822C \u5165\u9580...", + "digest":" Fess \u5168\u822C \u5165\u9580 \u30C9\u30AD\u30E5\u30E1\u30F3 Jav...", + "host":"fess.codelibs.org", + "mimetype":"text\u002Fhtml", + "contentLength":22788, + "boost":1.0, + "lastModified":"2014-05-22T11:09:07.000+0900", + "id":"http:\u002F\u002Ffess.codelibs.org\u002Fja\u002Fusers.html", + "urlLink":"http:\u002F\u002Ffess.codelibs.org\u002Fja\u002Fusers.html" + }, +... + ] + } +} +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. the status value is 0: normal, 1: search error, 2 or 3: request parameter error, 9: in service, 1: API type of error is.
    querySearch language.
    execTimeResponse time. Unit is seconds.
    pageSizeThe number of display.
    pageNumberPage number.
    recordCountNumber of hits for the search term.
    pageCountPage number of the number of hits for the search term.
    resultParent element of the search results.
    siteThe name of the site.
    contentDescriptionA description of the content.
    hostThe name of the host.
    lastModifiedLast update date.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    createdDocument creation date.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    contentTitleThe title of the document for viewing.
    contentLengthThe size of the document.
    urlLinkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-additional.xml b/src/site/en/xdoc/9.1/user/search-additional.xml new file mode 100644 index 000000000..ae0c2aa61 --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-additional.xml @@ -0,0 +1,15 @@ + + + + Hide search criteria + Shinsuke Sugaya + + +
    +

    You can use the additional parameters if the search string is shown on the screen without the specific search criteria like persuasion. additional value is retained but in the paging screen additional value.

    + +

    Without the conditions show screen and run a search when searches are performed by appending additional values in hidden forms, such as (for example, a search form) in the paging screen transitions and also the condition holds.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-and.xml b/src/site/en/xdoc/9.1/user/search-and.xml new file mode 100644 index 000000000..f20ac20a5 --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-and.xml @@ -0,0 +1,20 @@ + + + + AND search + Shinsuke Sugaya + + +
    +

    Use the search if you want to search for documents that contain all search words of more than one. When describing multiple words in the search box separated by spaces, AND skip AND search.

    + +

    If you use the search search words written AND. Write in capital letters AND the space required to back and forth. AND is possible can be omitted.

    +

    For example, if you want to find documents that contain the search terms 1 and 2 search terms, type the following search form.

    + +

    Connecting multiple words with AND are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-boost.xml b/src/site/en/xdoc/9.1/user/search-boost.xml new file mode 100644 index 000000000..697a4bd94 --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-boost.xml @@ -0,0 +1,20 @@ + + + + Boost search + Shinsuke Sugaya + + +
    +

    Use the boost search if you want to prioritize, search for specific search terms. Enabling search in boost search, depending on the severity of the search words.

    + +

    To boost search after the search term ' ^ boost value "that specifies the boost value (weighted) in the format.

    +

    For example, if you want to find the page if you want to find apples oranges contained more 'apples', type the following search form.

    + +

    Boost value specifies an integer greater than 1.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-field.xml b/src/site/en/xdoc/9.1/user/search-field.xml new file mode 100644 index 000000000..3c668053c --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-field.xml @@ -0,0 +1,66 @@ + + + + Search by specifying a search field + Shinsuke Sugaya + + +
    +

    You crawl in Fess results are saved for each field, such as title and full text. You can search for a field of them. You can specify the search criteria in search for a field, such as document type or size small.

    + +

    You can search for a the following fields by default.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Field list is available
    Field nameDescription
    URLThe crawl URL
    hostWere included in the crawl URL host name
    siteSite name was included in the crawl URL
    titleTitle
    contentText
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    mimetypeThe MIME type of the content
    +

    If you do not specify the fields subject to find the content. Fields are custom fields are also available by using the dynamic field of Solr.

    +

    If HTML file and search for the title tag that string in the title field, below the body tag registered in the body field.

    +
    + +

    If a field specifying the search field name: search words in separated by colons (:) field name and search word fill in the search form, the search.

    +

    If you search the Fess as a search term for the title field, type.

    + +

    Document, the title field in Fess above search appears as a search result.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-fuzzy.xml b/src/site/en/xdoc/9.1/user/search-fuzzy.xml new file mode 100644 index 000000000..30c6ba212 --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-fuzzy.xml @@ -0,0 +1,21 @@ + + + + Fuzzy search (fuzzy search) + Shinsuke Sugaya + + +
    +

    Ambiguity in the case does not match the words word search to search is available. Based on the Levenshtein distance in Fess ambiguous corresponds to the search (fuzzy search).

    + +

    After the search word you want to apply the fuzzy search adds '~'.

    +

    For example, ambiguous word "Solr", you can find documents that contain the word, near the "Solr" If you want to find, type as the search form, such as ("Solar"). +

    + +

    Furthermore, if by '~' after the number 0 and 1, 1 just like in refine. For example, in the form of 'Solr~0.8'. Do not specify numeric default value is 0.5.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-geo.xml b/src/site/en/xdoc/9.1/user/search-geo.xml new file mode 100644 index 000000000..8dfb4282a --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-geo.xml @@ -0,0 +1,32 @@ + + + + Dios arch + Shinsuke Sugaya + + +
    +

    Search using location information when you search, adding latitude and longitude location information for each document when generating the index becomes possible.

    + +

    Following parameters is available in the standard.

    + + + + + + + + + + + + + + + + +
    Request parameter
    GEO.latitudeLatitude degrees minutes seconds specifies double.
    GEO.longitudeLongitude degrees minutes seconds specifies double.
    GEO.distanceSpecifies the distance between documents in units of the coast. Specified in the double type.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-label.xml b/src/site/en/xdoc/9.1/user/search-label.xml new file mode 100644 index 000000000..094f732ce --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-label.xml @@ -0,0 +1,17 @@ + + + + Search by label + Shinsuke Sugaya + + +
    +

    Narrow your search by adding the categories to search the document for label information the label is specified when the search is possible. Label information by registering in the Administration screen, will enable search by labels in the search screen. Label information available can multiple selections in the drop-down when you search. If you do not register the label displayed the label drop-down box.

    + +

    You can select the label information at search time. Label information can be selected in the search options dialog by pressing the options button will appear.

    + Label search +

    You can search each document to create an index, set the label to the label. All results search search do not specify a label is usually the same. If you change the label information to update the index.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-not.xml b/src/site/en/xdoc/9.1/user/search-not.xml new file mode 100644 index 000000000..3a40c9f2b --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-not.xml @@ -0,0 +1,19 @@ + + + + NOT search + Shinsuke Sugaya + + +
    +

    If the document does not contain a word search NOT searching is available.

    + +

    Locate the NOT search as NOT in front of the Word does not contain. Is NOT in uppercase characters ago and need space.

    +

    For example, enter the following if you want to find documents that contain the search term 1 does not contain a search term 2 and the search.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-options.xml b/src/site/en/xdoc/9.1/user/search-options.xml new file mode 100644 index 000000000..4fb1cae3a --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-options.xml @@ -0,0 +1,36 @@ + + + + Extended parameter search + Shinsuke Sugaya + + +
    +

    You can pass any search criteria from third-party search engines move and easy to like. Pass search criteria Please implement processing in QueryHelperImpl #buildOptionQuery.

    + +

    Following parameters is available in the standard.

    + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    options.qThis is similar to the normal query. You can specify multiple options.q. If you specify multiple is treated as a search. Pass the URL encoding.
    options.CQTreated as exact match search queries. For example, if you specify the Fess Project searches as "Fess Project". Pass the URL encoding.
    options.OQIs treated as an OR search. For example, if you specify the Fess Project search as a Fess OR Project. Pass the URL encoding.
    options.NQTreated as NOT search. For example, if you specify 'Fess' search as NOT Fess. Pass the URL encoding.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-or.xml b/src/site/en/xdoc/9.1/user/search-or.xml new file mode 100644 index 000000000..6b3ed55f1 --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-or.xml @@ -0,0 +1,20 @@ + + + + OR search + Shinsuke Sugaya + + +
    +

    If you want to find documents that contain any of the search terms OR search use. When describing the multiple words in the search box, by default will search.

    + +

    To use search OR search words written OR. OR write in capital letters, the space required to back and forth.

    +

    For example, if you want to search for documents that contain either search term 2 search term 1 and type the following search form.

    + +

    OR between multiple languages are available.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-range.xml b/src/site/en/xdoc/9.1/user/search-range.xml new file mode 100644 index 000000000..378788f48 --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-range.xml @@ -0,0 +1,37 @@ + + + + Range searches + Shinsuke Sugaya + + +
    +

    If in the field containing the data in specified range, such as the number range search is possible for that field.

    + +

    To limit "field name: value TO value ' fill in the search form.

    +

    For example, type to search document contentLength field against 1 k to 10 k bytes is shown below the search form.

    + +

    To time range specified search ' lastModified: [date 1 TO date 2] "(Re 1 [Re 2) fill out the search form.

    +

    ISO 8601 with respect to re.

    + + + + + + + + + +
    Date and time-resolved second and fractional partIf the current relative to the date
    YYYY-MM-DDThh:mm:ss.sZ( example :2013-08-02T10:45:23.5Z)NOW (the current date), YEAR (this year), MONTH (month), DAY (today)
    +

    To relative to the current date and time NOW and DAY-(Adder, and production) and can sign and (round) like.

    +

    And a symbol for round / behind unit. Even if now-1DAY/day does today what time today 00: represents the day from 00 -1, the 00: 00.

    +

    For example, if you search for lastModified field from 2/21/2012 20: (current to date) 30 days prior to the updated document, type the following search form.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-sort.xml b/src/site/en/xdoc/9.1/user/search-sort.xml new file mode 100644 index 000000000..3a234451c --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-sort.xml @@ -0,0 +1,54 @@ + + + + Search sort + Shinsuke Sugaya + + +
    +

    To sort the search results by specifying the fields such as search time.

    + +

    You can sort the following fields by default.

    + + + + + + + + + + + + + + + + + + + + +
    Sort fields list
    Field nameDescription
    TstampOn the crawl
    contentLengthYou crawl the content size
    lastModifiedLast update of the content you want to crawl
    +

    Adding custom fields as sort in Customizing.

    +
    + +

    You can select the sorting criteria when you search. Sorting criteria can be selected in the search options dialog by pressing the options button will appear.

    + Sorting criteria +

    Also, for sorting in the search field sort: the field name to sort and field names separated by colon (:) fill out the search form, the search.

    +

    In ascending order sort the content size as a search term, Fess is below.

    + +

    To sort in descending order as below.

    + +

    If you sort by multiple fields separated list, shown below.

    + +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/search-wildcard.xml b/src/site/en/xdoc/9.1/user/search-wildcard.xml new file mode 100644 index 000000000..753e5afa5 --- /dev/null +++ b/src/site/en/xdoc/9.1/user/search-wildcard.xml @@ -0,0 +1,31 @@ + + + + Wildcard + Shinsuke Sugaya + + +
    +

    You can use one or multiple character wildcard search terms within. The can be specified as a one-character wildcard, * is specified as the multiple-character wildcard. Wildcards are not available in the first character. You can use wildcards for words. Wildcard search for the sentence.

    + +

    If you use one character wildcard shown below? The available.

    + +

    If the above is treated as a wildcard for one character, such as text or test.

    +

    If you use the multiple character wildcard use * below

    + +

    If the above is treated as a wildcard for multiple characters, such as test, tests or tester. Also,

    + +

    The so can be also used in the search term.

    +
    + +

    The wildcard string indexed using target. Therefore, because if the index has been created, such as bi-gram be treated meaning fixed string length in Japan Japanese wildcard in Japan, not expected behavior. Use in the field, if you use a wildcard in Japan, that used morphological analysis.

    +
    +
    + +
    diff --git a/src/site/en/xdoc/9.1/user/xml-response.xml b/src/site/en/xdoc/9.1/user/xml-response.xml new file mode 100644 index 000000000..f3d407e1b --- /dev/null +++ b/src/site/en/xdoc/9.1/user/xml-response.xml @@ -0,0 +1,183 @@ + + + + XML search results + Shinsuke Sugaya + + +
    +

    To emit the XML results of Fess. You need to do this output in XML Management screen crawl settings XML response enabled by you.

    + +

    To get the output XML http://localhost:8080/Fess/XML? query = search terms Of these sends a request. Request parameters are as follows.

    + + + + + + + + + + + + + + + + + + + + +
    Request parameter
    querySearch language. Pass the URL encoding.
    StartThe number of where to start. Starts at 0.
    NUMThe number of display. The default is 20. You can display up to 100.
    fields.labelThe label value. Use to specify the label.
    +
    + +

    Returns the following response.

    + + + 1 + 0 + Fess + 0.64 + 20 + 1 + 101 + 6 + + + 2014-05-24T15:13:27.692+0900 + d55275b552da44999cafe118aa8fed62 + Fessのイン... + http://fess.codelibs.org/ja/setup.html + 0.83177954 + fess.codelibs.org/ja/setup.html + html + <em>fess</em>... + オープンソ... + fess.codelibs.org + text/html + 49702 + 1.0 + 2014-05-22T11:09:07.000+0900 + http://fess.codelibs.org/ja/setup.html + http://fess.codelibs.org/ja/setup.html + +... + + +]]> +

    Each element is as follows.

    + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    Responses to information
    responseThe root element.
    versionFormat version.
    statusThe status of the response. the status value is 0: normal, 1: search error, 2 or 3: request parameter error, 9: in service, 1: API type of error is.
    querySearch language.
    exec-timeResponse time. Unit is seconds.
    page-sizeThe number of display.
    page-numberPage number.
    record-countNumber of hits for the search term.
    page-countPage number of the number of hits for the search term.
    resultParent element of the search results.
    docThe elements of the search results.
    siteThe name of the site.
    content-descriptionA description of the content.
    hostThe name of the host.
    last-modifiedLast update date.
    cacheThe contents of the content.
    scoreScore values of the document.
    DigestDigest a string of documents.
    createdDocument creation date.
    URLThe URL of the document.
    IDThe ID of the document.
    mimetypeThe MIME type.
    titleThe title of the document.
    content-titleThe title of the document for viewing.
    content-lengthThe size of the document.
    URL-linkThe URL as the search results.
    +
    +
    + +
    diff --git a/src/site/en/xdoc/articles.xml b/src/site/en/xdoc/articles.xml new file mode 100644 index 000000000..d6c25c3fb --- /dev/null +++ b/src/site/en/xdoc/articles.xml @@ -0,0 +1,29 @@ + + + + An article in the + Shinsuke Sugaya + + +
    +
    +
    2010-12-16
    +
    + [In Fess make Apache Solr-based full-text search server-role-based search] (CodeZine) +
    +
    2010-01-28
    +
    + [In Fess make Apache Solr-based full-text search Server-Mobile] (CodeZine) +
    +
    2009-11-20
    +
    + "Easy build in 5 minutes! Try the full-text search server "Fess" ' (Wall Street Journal) +
    +
    2009-11-12
    +
    + [In Fess make Apache Solr-based full-text search Server-Introduction] (CodeZine) +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/articles/article-1.xml b/src/site/en/xdoc/articles/article-1.xml new file mode 100644 index 000000000..9b35f8dd1 --- /dev/null +++ b/src/site/en/xdoc/articles/article-1.xml @@ -0,0 +1,407 @@ + + + + In Fess make Apache Solr based search server-introduction + Shinsuke Sugaya + + +
    +

    + Document management is increasing daily, is expected to effectively manage their documents. + More managed document, with specific information from its difficult to continue. + Include implementing full-text search server able to search through vast as the solution.

    +

    + Fess is a easy deployment, Java-based open-source full-text search server. + The search engine part of the Fess by using Apache Solr. + Is a very powerful search engine called SOLR index can be the 200 million document. + On the other hand, may need to implement in your own crawler parts, such as when trying to build in the Apache Solr search system. + You can use S2Robot Fess offers from the Seasar Project crawler parts, collect Web or file system on various types of documents to search.

    +

    + Therefore, this article introduces about building a search server by Fess. +

    + +
    +
    +
      +
    • Those who want to build search system
    • +
    • Those who observed to add search functionality into existing systems
    • +
    • Those who are interested in Apache Solr
    • +
    + +
    +
    +

    + Regarding the content of this article in the following environment and behavior verification. +

    +
      +
    • + Windows 7 (Service Pack1) +
    • +
    • + JDK 1.7.0_21 +
    • +
    + +
    +
    +

    + Fess is a open source Web and file system using the full text search system. The SourceForge.jpFess sitesFrom the provided in the Apache license. +

    + + + +

    + Java-based search system +

    +

    + Fess is as in the following figure, has been built using various open source products. +

    +
    +
    Fess structure
    +
    + Fess structure +
    +
    +

    + A Fess and Solr war file is deployed to the Tomcat distribution. + War file of the Fess offers search and management screens. + The Fess as a development framework Seasar2, SAStruts employs in the presentation layer. + So, by modifying the JSP if you want to customize, such as screen easy customization is possible.

    +

    + Also using the built-in database H2Database to save settings and crawl data, is accessed by using o/r Mapper DBFlute. + S9chronos is used to perform a crawl in the time specified in the Fess, scheduling framework provided by the Seasar project. + SOLR and S2Robot are discussed.

    +

    + Fess was constructed as a Java-based system, so any platform can be performed. + Provides a UI to easily set from the Web browser settings.

    + +

    + As a search engine using Apache Solr +

    +

    + Apache Solr is an enterprise search server based on Lucene, is available from the Apache Software Foundation. + Roundness that characterized the support such as faceted search, search result highlighting, multiple output formats. + Also in the Solr server configuration depends on the number of documents that can be searched for, and several hundred million documents, you can scale out to large scale site search server. + Said to search engine usage and many in Japan, has been in.

    +

    + Fess uses Apache Solr to search engines. + Distributed in the distribution of the Fess in the Solr, but cut out Fess Solr server to another server that is available. + Also, multiple Solr Server manages the Fess as a group, form a redundant configuration is possible. + Design can take advantage of scalability in this way with SOLR in Fess;

    + +

    + Available as a crawling engine S2Robot +

    +

    + S9robot is the krolaframework provided by the Seasar project. + S2Robot can collect touring the document to the documents on the Web or on the file system. + It is possible to also document collection in multiple threads simultaneously multiple documents efficiently treating. + Also, document can handle HTML, not to mention in numerous formats such as MS Office system files, such as Word and Excel, zip archive files, images and audio files, including covers (images and audio files, gets a meta-information).

    +

    + Fess by using S2Robot, touring on the Web and file system documents, collect text information. + You can can handle S2Robot file format to accommodate even those to be searched. + Etc to crawl through the S2Robot for parameter it is possible to set from the management UI of the Fess.

    + + +

    + Mobile support +

    +

    + Fess is compatible for viewing on docomo, au and Softbank Mobile phones. + You can specify when indexing documents to can be viewed in search results with what handsets. + Book and skip for viewing on your mobile device in the paper, describes the next time.

    + +
    +
    +
    +

    + Start the Fess, and describes the steps to do a search. + You can install and launch in almost similar steps in Mac OS X and Linux provides information intended to be run on Windows XP, but.

    + + +

    + http://sourceforge.jp/projects/Fess/releases/ From the download the latest package. + The most recent version at the time of writing this article ( 2013 / 06 ) 8.1.0. + Unzip the download has finished, in any directory.

    +
    +
    Download Fess
    +
    + Download Fess +
    +
    + +
    + +

    + CATALINA_HOME and JAVA_HOME environment more appropriately, please run the %CATALINA_HOME%\bin\startup.bat. + For example, if you unzip the fess-8.1.0.zip C:\fess CATALINA_HOME is C:\fess\fess-server-8.1.0.

    +
    +
    Launch of the Fess
    +
    +C:\fess\fess-Server-8.1.0 > set "JAVA_HOME = C:\Program Files \Java\jdk1.7.0_21" C:\fess\fess-server-8.1.0 > set CATALINA_HOME = C:\fess\fess-server-8.1.0 C:\fess\fess-server-8.1.0 > cd bin C:\fess\fess-server-8.1.0\bin > startup.bat
    +
    +
    +

    + In the browser / http://localhost:8080/Fess The Fess is starting and access the following screen appears, the. +

    +
    +
    Search top screen
    +
    + Search top screen +
    +
    + +
    + +

    + Please run the shutdown.bat. +

    +
    +
    Stop Fess
    +
    +C:\fess\fess-Server-8.1.0\Bin > shutdown.bat
    +
    +
    + +
    + +

    + Directory structure looks like this. +

    +
    +
    Directory configuration
    +
    +Fess-Server-8.1.0/ |--LICENSE |--NOTICE |--RELEASE-NOTES |--RUNNING.txt |--bin / |--conf / |--extension / |--lib / |--logs / |--solr /-- SOLR data directory
    +|   |--contrib / |   |--core1 / |   | |-- bin/                    --SOLR executable
    +|   | |-- conf/                   --Place the settings file in SOLR
    +|   | |-- data/                   --Contains the SOLR index
    +|   |   '--txlog / |   |--dist / |   '--lib / |--temp / |--webapps / |   |--fess and |   |   |--META-INF / |   |   |--Web-INF / |   |   | |--cachedirs /--Store for mobile image cache
    +|   |   | |--classes /--Place the classes and settings files
    +|   |   | |-- db/                 --Place the DB data
    +|   |   |   |--cmd |   |   |   |--conf / |   |   |   |--lib / |   |   |   |--orig / |   |   | |-- logs/               --Contains the log files of the Fess
    +|   |   | |-- view/               --Place the JSP UI related
    +|   |   |   |--fe.tld |   |   |   |--struts-config.xml |   |   |   |--validator-rules.xml |   |   |   '--web.xml |   | |-- css/                    --Place the CSS file
    +|   | |-- js/                     --Place the JS file
    +|   | |-- images/                 --Place the image file
    +|   |   '--jar / |   |--fess.war | |-- solr/                       --The SOLR Web apps
    +|   |--solr.war |   |--manager / |
    +
    +'--manager.war '--work /
    +
    +

    + just below the "fess-server-8.1.0" directory configuration is similar to the Tomcat 7, might be deployed Solr data directory 'solr', 'fess.war' and 'solr.war. + Is deployed 'fess.war' to 'webapps/fess/WEB-INF/view' put JSP file search and management screens. + Also, if you need to customize the screen CSS file is placed in the 'webapps/fess/css', so edit the files.

    + +
    +
    +
    +

    + Also indexed for search in the State immediately after the launch, make a search returned nothing results. + So, you must first create the index. In this case,http://Fess.codelibs.org/ja/ Create index to below, to do a search as an example. +

    + +

    + First of all, on Administration page http://localhost:8080/Fess/Admin To access, please login. + By default user name and password are both admin.

    + +
    +
    Login to the management page
    +
    + Login +
    +
    + +
    +  +

    + Then, register the crawled. + Because the Web page, select the [Web] from the left of the admin page. + For anything not registered in the initial state, select Create new.

    + +
    +
    Select the [new]
    +
    + Create a new +
    +
    + +

    + As a Web crawl settings, this ishttp://Fess.codelibs.org/ja/ That will crawl all the pages below. + In addition, results are displayed when you search from any PC or mobile phone, and then select all as the browser type.

    + +
    +
    Web crawl settings
    +
    + Crawl settings +
    +
    + +

    + Then, click the [create] on the confirmation screen that can crawl to register. + Registration is possible to change from the Edit.

    + +
    +
    Completing the registration Web crawl settings
    +
    + Crawl settings for registering +
    +
    + +
    +  +

    + Set to collect, document, crawl schedules. + Crawl schedules are set from the menu on the left of the admin page crawl General.

    +

    + Formatting is similar to the Unix Cron. + From left, seconds, minutes, time, day, month, represents a day of the week. + For example, daily 12: If you successfully crawl your 10 am ' 0 10 12 * *? ' and then I.

    + +
    +
    Crawl schedule
    +
    + Crawl schedule +
    +
    + +

    + Crawl is started and the index has been created to make from the menu on the left side, the session information that you can. + Displays the document number when the crawl is complete, the search index size of session information (Web/file).

    + +
    +
    Check the crawl status of
    +
    + Check the crawl status of +
    +
    + +
    +
    If the crawl is complete example
    +
    + If the crawl is complete example +
    +
    + +
    +  +

    + Like image below to search crawl after the results are returned. +

    + +
    +
    Search example
    +
    + Search example +
    +
    + +
    +
    +
    +

    + Here, the most viewed users, search results and search top screen shows how to customize the list screen. +

    +

    + Shows how we change the log file name. + You can change any knowledge of HTML, so if you want to change the design itself described in a simple JSP files.

    +

    + First of all, find the top screen 'webapps/fess/WEB-INF/view/index.jsp' file. +

    +
    +
    Search top screen JSP files
    +
    +<%@page pageEncoding="UTF-8" contentType="text/html; charset=UTF-8"%>
    +<html>
    +<head>
    +<meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
    +<meta http-equiv="content-style-type" content="text/css">
    +<meta http-equiv="content-script-type" content="text/javascript">
    +<title>Fess</title>
    +<link href="${f:url('/css/style.css')}" rel="stylesheet" type="text/css">
    +</head>
    +<body>
    +<div id='main'>
    +<s:form action="search">
    +  <table>
    +    <tbody>
    +      <tr>
    +        <td><img id="logo" src="${f:url('/images/ id="logo" src=""></img id="logo" src="${f:url('/images/></td></tr></tbody></table></s:form></div></body></html>logo.gif')}" alt="<bean:message key="labels.search_top_logo_alt"></bean:message>" />
    +        <td><div class="input">
    +          <html:text styleclass="query" property="query" title="Search" size="50" maxlength="1000"></html:text>
    +          <input class="btn" type="submit" value="<bean:message key=" labels.top.search"/="">"name ="search"/ ></div></td>
    +      
    +    
    +  
    +
    +
    +
    +
    +
    +

    + Change the file name to change the images that appear on the home screen search 'logo.gif' where you want to replace. + Files placed in the 'webapps/fess/images'.

    +

    + <s:form>And <bean:message>such as a JSP tag. + For example,<s:form> the actual HTML view when converted to the form tag. + Detailed description see SAStruts or for JSP sites. </s:form></bean:message></s:form>

    +

    + The search results list screen will be in the 'webapps/fess/WEB-INF/view/search.jsp' file. +

    +
    +
    Search results part of the JSP file list screen
    +
    +<div id="header">
    +  <s:form action="search">
    +    <div class="input">
    +      <s:link action="index" title="Fess Home">
    +        <img class="logo" src="${f:url('/images/ class="logo" src=""></img class="logo" src="${f:url('/images/></s:link></div></s:form></div>logo-head.gif')}" alt="<bean:message key="labels.search_header_logo_alt"></bean:message>"/>                                 
    +      <html:text styleclass="query" property="query" title="Search" size="50" maxlength="1000"></html:text>
    +      <input class="btn" type="submit" value="<bean:message key=" labels.search"/="">"name ="search"/ >
    +  
    +
    +
    +
    +

    + Results of the 'logo-head.gif' file name change to change the image that appears at the top of the screen. + similar to 'logo.gif' put in 'webapps/fess/images'.

    +

    + Edit 'Style.css' If you want to change the CSS file used in a JSP file located in the 'webapps/fess/css'. +

    + +
    +
    +

    + About the Fess in the full-text search system, from installation until search and simple customization methods discussed. + I could introduce you can easily build a search system if you have the Java runtime environment, with no special environment. + Can be introduced into an existing system site search functionality, such as if you want, so you try.

    +

    + I want to introduce the next time you support Fess mobile site search feature. +

    + + +
    +
    + +
    + +
    diff --git a/src/site/en/xdoc/articles/article-2.xml b/src/site/en/xdoc/articles/article-2.xml new file mode 100644 index 000000000..5abb5af93 --- /dev/null +++ b/src/site/en/xdoc/articles/article-2.xml @@ -0,0 +1,195 @@ + + + + In Fess make Apache Solr based search Server-Mobile Edition + Shinsuke Sugaya + + + +
    +

    + The last timeIntroduction chapterSo, how to build open-source full-text search server by Fess introduced. + Fess on docomo, au and Softbank Mobile search so that introduce how to use this time.

    +

    + In this article, explains Fess 8.1.0. About how to build a FessIntroduction chapterPlease see the. +

    + +
    +
    +
      +
    • Those who want to build mobile terminal search system
    • +
    • Observed to add search functionality on existing cell sites
    • +
    + +
    +
    +

    + Regarding the content of this article in the following environment and behavior verification. +

    +
      +
    • + Windows 7 (Service Pack1) +
    • +
    • + JDK 1.7.0_21 +
    • +
    + +
    +
    +

    + Available for mobile terminals in the full-text search system for systematically following such response will become necessary. +

    +
      +
    1. To get the mobile device information, suitable for Terminal
    2. +
    3. You can specify a crawl when you create for user agents
    4. +
    5. Can include career information index information
    6. +
    7. View converted into Web-site if the content search results
    8. +
    +

    + In the Fess corresponds to all of the above. For processing first, retrieving the phonemobyletIs adopted. + mobylet is a Java open source framework for building mobile Web applications. + You can view the results mobylet in docomo, au and Softbank handsets to identify the appropriate terminals each.

    +

    + You can with Web crawl settings set the user agent when the next crawl search in Fess. + You can crawl the user agent for career, to get their sites for mobile phones. + However, you must allow Fess server IP you are IP restrictions for mobile site, mobile phone sites displaying. + Also, viewable results carrier terminal only in select the carriers you want to display all selected Web crawl settings in the default [browser], but that becomes possible.

    +

    + (If you use the PC site Viewer, such as can be viewed is) cannot normally see on mobile devices even if the search result Web-site, PC site link appears in the search results. + It is possible to use the Google Wireless Transcoder in Fess. + Google Wireless Transcoder is a service provided by Google, Inc., PC site will convert your various mobile phone. + It is possible for simple settings in Fess, links Google Wireless Transcoder to convert search results the search results use smooth PC site.

    + +
    +
    +

    + Fess 8.1.0 installed and is running. +

    + + +

    + create a Web crawl settings to display the search results only if you searched for DoCoMo handsets. +

    +

    + Access and http://localhost:8080/fess/admin in the management page, please login. + By default user name and password are both admin. + Select the [Web] from the left of the admin page. + For anything not registered in the initial state, select Create new.

    +
    +
    Select the [new]
    +
    + Create a new +
    +
    +

    + This time the mobile phone site is not that will crawl all the pages of http://fess.codelibs.org/ja/ following. + any mobile sites that can be displayed in DoCoMo handsets instead of http://fess.codelibs.org/ja/ the site URL specifies.

    +

    + [Browser] and select only the DoCoMo so that appears only in docomo handsets. + If you want to display in AU and Softbank handset select them here.

    +

    + After the user agent be user-agent for docomo handsets. + This time we enter DoCoMo/2.0 P903i.

    +
    +
    for DoCoMo Web crawl settings
    +
    + for DoCoMo Web crawl settings +
    +
    +

    + Then, click the [create] on the confirmation screen that can crawl to register. + Registration is possible to change from the Edit.

    + +
    + +

    + Sets the search results PC if there is Google Wireless Transcoder available sites you like. + Excluding the PC site to search for the mobile site is just there if you don't need this setting.

    +

    + Select the crawl General from the left of the admin page. + In the mobile conversion, select the Google Wireless Transcoder.

    +
    +
    Mobile conversion configuration
    +
    + Mobile conversion configuration +
    +
    +

    + Saves the settings and click [update]. +

    + +
    + +

    + Mobile handset settings after the start to crawl, create a searchable index. + Select the system settings from the left of the admin page.

    +
    +
    System settings
    +
    + System settings +
    +
    +

    + Click the start crawling, initiates a search for crawling and indexing. + While complete crawl.

    + +
    + +

    + First, try searching in the PC browser, such as Internet Explorer. + visit http://localhost:8080/Fess, locate the Fess.

    +
    +
    Search in PC browser
    +
    + Search in PC browser +
    +
    +

    + You can: set Web crawl settings in the PC browser in search results is displayed. +

    +

    + The following access in docomo handsets. This time real terminal, not in FirefoxFireMobileSimulatorUse Add-ons, to see the results. + FireMobileSimulator is a Firefox Add-on that simulates the major three carriers mobile phone browser. + FireMobileSimulator installed in Firefox, from the Firefox menu select, DC Terminal docomo P903i from [tool] [FireMobileSimulator]. + When you access this setting allows Firefox to simulate the P903i handset environment. + Similarly if the PC browser and visit http://localhost:8080/fess, locate the Fess.

    +
    +
    Search in DoCoMo handsets
    +
    + Search in DoCoMo handsets +
    +
    +

    + Search for this time is specified in the Web crawl settings is displayed. +

    + +
    +
    +
    +

    + How to respond to the Fess in the full-text search system handsets introduced. + I could introduce you can provide search functionality to the handsets of three major carriers in simple settings. + Also, it is possible to respond by phone will be released new models on a regular basis, but the latest terminal information file in Fess 'webapps/fess/WEB-INF/classes/device'. + About how to update the device information file see the README in the directory.

    +

    + Next switch to display results search results depending on authentication of users, roles functions are introduced here. +

    + + +
    +
    + +
    + +
    diff --git a/src/site/en/xdoc/articles/article-3.xml b/src/site/en/xdoc/articles/article-3.xml new file mode 100644 index 000000000..c585a94d7 --- /dev/null +++ b/src/site/en/xdoc/articles/article-3.xml @@ -0,0 +1,303 @@ + + + + In Fess make Apache Solr based search server-part role-based search + Shinsuke Sugaya + + + +
    +

    + The last timeMobile EditionOr how to build mobile device-friendly search system using Fess was introduced. + Introduces role-based search feature is also distinctive features of Fess one thing in this article.

    +

    + In this article, explains Fess 8.2.0. About how to build a FessIntroduction chapterPlease see the. +

    +
    +
    +
      +
    • More authentication such as portal sites like seen in search system building
    • +
    • Want to build an environment to search for viewing permissions each
    • +
    +
    + +
    +

    + Regarding the content of this article in the following environment and behavior verification. +

    +
      +
    • + CentOS 5.5 +
    • +
    • + JDK 1.6.0_22 +
    • +
    +
    + +
    +

    + Is divided out search results user and role-based search of the Fess, authenticated in any authentication system authentication information to the original function. + For example, search for technical personnel b sales reps A sales division role shows a sales division role information in search results, sales of rolls does not appear it. + By using this feature, user login in the portal and single sign-on environment belongs to you can enable search, sector or job title.

    +

    + In Fess role-based search to retrieve role information from the following places can be. +

    +
      +
    1. Request parameter
    2. +
    3. Request header
    4. +
    5. Cookies
    6. +
    7. J2EE authentication information
    8. +
    +

    + Role information to pass to the Fess, usages, in the Portal Server and agent-based single sign-on system authentication when running Fess's domain and path to save authentication information in cookies. + By reverse proxy type single sign-on system access to Fess to add authentication information request headers and request parameters can retrieve role information in Fess. + You can thus various authentication systems and working with the search results to put out separate.

    +

    + To cope with that to provide a class to implement the jp.sf.fess.helper.RoleQueryHelper interface if you are running your own authentication system. The class 'webapps/fess/WEB-INF/classes' As in, have put in place through the classpath "webapps/fess/WEB-INF/classes/fess.dicon"In the given on behalf of the jp.sf.fess.helper.impl.RoleQueryHelperImpl. +

    +
    + +
    +

    + By installing a Fess 8.2.0. If you have not installed yet,Introduction chapterPlease install, refer to. +

    +

    + Describes the role-based search using the credentials of the J2EE (Tomcat authentication) provides the Fess existing login screen using various authentication systems but without building a separate authentication systems in Fess Tomcat authentication environment, so use this in this article. +

    + + +

    + First of all, search results separate out, to show to Tomcat users. + This time, create sales ( sales ) and engineering (eng) two rolls. + And the user adds taro and hanako users belong to the eng role belongs to the sales role. User information below to seeconf/Tomcat-users.XML"To write. +

    +
    +
    contents of Tomcat-users.XML
    +
    +<?xml version='1.0' encoding='utf-8'?>
    +<tomcat-users>
    +  <role rolename="fess"></role>
    +  <role rolename="solr"></role>
    +  <role rolename="manager"></role>
    +  <role rolename="sales"></role><!-- 追加 -->
    +  <role rolename="eng"></role><!-- 追加 -->
    +  <user username="admin" password="admin" roles="fess"></user>
    +  <user username="solradmin" password="solradmin" roles="solr"></user>
    +  <user username="manager" password="manager" roles="manager"></user>
    +  <user username="taro" password="taropass" roles="sales"></user><!-- 追加 -->
    +  <user username="hanako" password="hanakopass" roles="eng"></user><!-- 追加 -->
    +</tomcat-users>
    +
    +
    +

    + This setting is not required if the use of an existing authentication system. +

    +
    + + +

    + Here's the settings for the Fess. First of all, 'webapps/fess/WEB-INF/classes/fess.dicon"The roleQueryHelper sets how to retrieve the default roles and authentication information. By using the J2EE authentication information, so seeFess.dicon"The roleQueryHelper will set like the following. +

    +
    +
    contents of Fess.dicon
    +
    :
    +<component name="roleQueryHelper" class="jp.sf.fess.helper.impl.RoleQueryHelperImpl">
    +    <property name="defaultRoleList">
    +        {"guest"}
    +    </property>
    +</component>
    +:
    +
    +

    + Set the default role as above. + Treated as the role search to locate the default roles are not logged in. + All search results are displayed on the status not logged in and you do not specify a default search.

    +

    + About the J2EE authentication information not available if worth mentioning here. + If the authentication information from the request parameter set follows.

    +
    +
    contents of Fess.dicon
    +
    : <component name="roleQueryHelper" class="jp.sf.fess.helper.impl.RoleQueryHelperImpl"> <property name="parameterKey">"fessRoles"</property>,<property name="encryptedParameterValue">false</property> <property name="defaultRoleList"> {guest}</property> 
    +    
    +    
    +    
    +</component> :
    +
    +

    + Key request parameter specifying where fessRoles can pass the role information in the comma-separated values. For example, the URL to locate the user with the sales and admin roles 'http: //hostname/fess/search?' The so will be added fessRoles. + You can pass here encryptedParameterValue is set to false, this value to true and encrypt, Blowfish or AES in the value part of the fessRoles. + You need to set if you encrypt the value to be FessCipher components, so that you can decrypt.

    +

    + If the authentication information from the request header set shown below. +

    +
    +
    contents of Fess.dicon
    +
    : <component name="roleQueryHelper" class="jp.sf.fess.helper.impl.RoleQueryHelperImpl"> <property name="headerKey">"fessRoles"</property>,<property name="encryptedParameterValue">false</property> <property name="defaultRoleList"> {guest}</property> 
    +    
    +    
    +    
    +</component> :
    +
    +

    + You can specify fessRoles in the request header key, role information to pass in comma-separated values. +

    +

    + If the authentication information from the cookie set shown below. +

    +
    +
    contents of Fess.dicon
    +
    : <component name="roleQueryHelper" class="jp.sf.fess.helper.impl.RoleQueryHelperImpl"> <property name="cookieKey">"fessRoles"</property>,<property name="encryptedParameterValue">false</property> <property name="defaultRoleList"> {guest}</property> 
    +    
    +    
    +    
    +</component> :
    +
    +

    + You can specify fessRoles to the name of the cookie, as well as the request parameter, pass the role information in the comma-separated values. +

    +
    + + +

    + 「Fess.dicon"And to ensure you can log in as"webapps/fess/WEB-INF/web.xml"The change security-related settings. + The following settings.

    +
    +
    the content of
    +
    : <security-constraint> <web-resource-collection> <web-resource-name>Fess Authentication</web-resource-name><url-pattern>/login/login</url-pattern>
    +    
    +    
    +  </web-resource-collection><auth-constraint><role-name>fess</role-name><role-name>sales</role-name><role-name>eng</role-name> 
    +    
    +    
    +    
    +  </auth-constraint> 
    +  
    +  
    +</security-constraint> : <security-role><role-name>fess</role-name>
    +  
    +</security-role><security-role><role-name>sales</role-name>
    +  
    +</security-role><security-role><role-name>eng</role-name>
    +  
    +</security-role>
    +
    +
    +
    +

    + This setting is required if you are using authentication, such as a request parameter. +

    +
    + +
    + +
    +

    + Complete Setup so please start Fess. +

    + + +

    + Has been turned by this admin, taro, and hanako 3 user Fess to log. + Verify that you can log in to./ http://localhost:8080/Fess/AdminTo access the login in the admin user and as usual management screen appears. Then logged out admin user, again/ http://localhost:8080/Fess/AdminTo access, please login with taro and hanako users. After a successful login, and/ http://localhost:8080/FessThe search screen is displayed. When you log out/ http://localhost:8080/Fess/AdminTo access, click the logout button. +

    +
    +
    Logout screen
    +
    + Logout screen +
    +
    +
    + + +

    + logged in as admin user and lists the role role in the left menu to click. + We create three roles.

    +
    +
    Role list
    + + + + + + + + + + + + + + + + + + + +
    Role nameValue
    By defaultdefault
    Sales Departmentsales
    Technology DepartmentEng
    +
    +
    + + +

    + Create a crawl. This time the users in the sales department rolehttp://www.n9sm.NET/Only, you can search for users of the technology of rollhttp://Fess.codelibs.org/Just so that you can find. + In order to crawl settings, click on the left menu [Web], lists the Web crawl settings. + Click the [create new], please create a Web crawl settings. First of all, sales for thehttp://www.n2sm.NET/To the Sales Department, select [role] item as the crawl settings, create. In the followinghttp://Fess.codelibs.org/The create a role Select technology, in the crawl settings. +

    +
    +
    Web crawl settings roll items
    +
    + Web crawl settings roll items +
    +
    +
    + + +

    + Registration after the crawl settings, click System settings on the left menu, click the Start button in the system settings screen, starts to crawl. + While wait for crawl to complete.

    +
    + + +

    + After crawling,/ http://localhost:8080/FessTo make sure that search word, such as "fess" access, not logged in, search results are displayed. + Then logged in taro, as well as search. for taro user has a sales rolehttp://www.n9sm.NET/The only search results are displayed. +

    +
    +
    Search screen in the sales role
    +
    + Search screen in the sales role +
    +
    +

    + Taro user logout, please login with hanako users. Destination and so have eng role hanako users as well as search andhttp://Fess.codelibs.org/The only search results are displayed. +

    +
    +
    in the Eng role search screen
    +
    + in the Eng role search screen +
    +
    +
    + +
    + +
    +

    + I introduced about the security features of the Fess in role-based search. + I think various authentication systems to accommodate because mainly covers the J2EE authentication information by using role-based search, but the passing of authentication information to the Fess general implementation. + It is possible to realize systems so each attribute in the user search results out into the corporate portal site or shared folder browsing permissions per search is required.

    +

    + Next offers Fess of Ajax functions are introduced here. +

    +
    + +
    + +
    + +
    diff --git a/src/site/en/xdoc/articles/article-4.xml b/src/site/en/xdoc/articles/article-4.xml new file mode 100644 index 000000000..9e7773477 --- /dev/null +++ b/src/site/en/xdoc/articles/article-4.xml @@ -0,0 +1,268 @@ + + + + In Fess make Apache Solr based search server-ed. REST API + Shinsuke Sugaya + + + +
    +

    + The last timePart of the role-based searchThe user permission is required in how available Fess or introduced. + An introduction of how to do search and the results display in the client-side (browser side), using the REST API is now Fess. + In the REST API, Fess to use as a search server Web system of the existing, and integrate changes only the HTML is possible too.

    +

    + In this article, explains Fess 8.1.0. About how to build a FessIntroduction chapterPlease see the. +

    +
    +
    +
      +
    • Those who want to search functionality into existing Web systems
    • +
    • Those who want to build a search system with AJAX
    • +
    +
    + +
    +

    + Regarding the content of this article in the following environment and behavior verification. +

    +
      +
    • IE 10
    • +
    • Firefox 21
    • +
    +
    + +
    +

    + According to normal HTML Fess search expressions in addition to search results using XML and JSON (including JSONP) response is possible as the REST API. + Can also use the REST API should build a Fess server will contact only the search results from the existing system to easily. + I think independent development language such as XML or JSON results format, so easy-to-integrate Fess to non-Java System. + You can handle it easily even if it is XML or JSON support in JavaScript libraries, so as Ajax.

    +

    + See below about what return the REST API of Fess. +

    +
      +
    1. XML response
    2. +
    3. JSON (JSONP) response
    4. +
    +

    + Fess uses Apache Solr as an internal search engine. + SOLR also provides XML and JSON API that Fess API is different. + Benefits as well as SOLR API, API Fess might Fess API in various Fess, such as control of search log management and viewing rights of specific features are available. + Would use the Solr case developed from the structure of the document crawling their own good, but easy to add and search features that Fess, can reduce the development cost of the many.

    +
    + +
    +

    + Describes how to build a Fess REST API using this site. + Use the JSON response to Fess server interaction. + Fess server to utilize this time uses a Fess server in the Fess project released as a demo. + If the install the Fess 4.0.0 or later if you want to use the Fess server of your own. + Fess 4.0.0 or later supports JSONP.

    + + +

    + Same-Origin policy should be aware when using the AJAX security model. + You must use the JSONP can use JSON if it displays in the browser HTML output and Fess server exists in the same domain, but different. + You can use the JSONP in Fess REST API in the JSON response request parameters with callback key value by passing.

    +
    +
    Same-Origin policy. (B) the must Fess return search results (JSON) becomes another domain if you use the JSONP.
    +
    + Same-Origin policy. (B) of if Fess return search results (JSON) another domain to you for must use the JSONP. +
    +
    +

    + Advances the story in case that this time there was HTML and Fess servers are in different domains. + So, please remove the callback from the request parameter if the good example, but in the same domain by using the JSONP.

    +
    + + +

    + Implements the search process, this time using JavaScript in HTML. + As a JavaScript library using jQuery. + You can also process, such as AJAX, such as easy to implement in jQuery. + Files you create will become.

    +
      +
    • To display the search form and search results to HTML file 'index.html'
    • +
    • JS file to communicate with Fess server 'fess.js'
    • +
    +

    + In the examples of this building has implemented the following features. +

    +
      +
    • Search buttons send a search request.
    • +
    • In the search results list
    • +
    • Search results paging
    • +
    +
    + + +

    + Create the HTML search form and search results. + Easy-to-understand this time is to adjust the design with CSS, without to have simple tag structure. + In one of the following HTML file to use.

    +
    +
    the contents of the index.html
    +
    +<html>
    +<head>
    +<meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
    +<title>Search site</title>
    +</head>
    +<body>
    +<div id="header">
    +  <form id="searchForm">
    +    <input id="searchQuery" type="text" name="query" size="30">
    +    <input id="searchButton" type="submit" value="検索">
    +    <input id="searchStart" type="hidden" name="start" value="0">
    +    <input id="searchNum" type="hidden" name="num" value="20">
    +  </form>
    +</div>
    +<div id="subheader"></div>
    +<div id="result"></div>
    +<script type="text/javascript" src="http://ajax.googleapis.com/ajax/libs/jquery/1.9.0/jquery.min.js"></script>
    +<script type="text/javascript" src="fess.js"></script>
    +</body>
    +</html>
    +
    +
    +

    + We look below the body tag, and first id attribute in the header div tag where search input box and the search button located by. + Also, in the hidden form holds display start position (start) and the number (num). + Search requests submitted in JavaScript update will start and the num value. + However, show is 1 page per, and no ability to change the number displayed in this sample code, so the value of num is unchanged. + Also, you will not send this form page transitions occur if you have JavaScript enabled for on the search form submit search requests are communicated Ajax.

    +

    + Displays information such as the number of hits in the subheader for the following div tag where to search. + displays search results and paging links in the div tag of the result.

    +

    + Load the 'fess.js' recently created this jQuery JS files and JavaScript. + I try to get via Google CDN may save the jQuery JS file in the same directory as index.html, but this time.

    +
    + + +

    + Create a JS file to display the search results, and then communicates with the Fess server 'fess.js'. + And create a 'fess.js' with the following contents will put in the same directory as index.html.

    +
    +
    contents of Fess.js
    +
    +$ (function() {/ / (1) Fess URL var baseUrl = "http://search.n9sm.co.jp/JSON?callback=?
    +
    +

    + runs after the DOM of the HTML file is built "fess.js" process. + First, specify Fess server URL by 1. + In the specify Fess public demo server. + To get search results JSON data from external servers using JSONP. + If you use JSON, JSONP, without callback =? is not required.

    +

    + 2 save the jQuery object for the search button. + Maintain the variable to use the jQuery object search button several times and reuse.

    +

    + 3 defines search functions. + Contents of this function is described in the following sections.

    +

    + 20 registers events when the search form is submitted. + Search button press or when the decision was made in the search input field Enter key press occurs when registered at 20. + Search processing function doSearch call when the event occurs. + used for paging is the value passed when calling the search processing function is the value of the Navi.

    +

    + Register the event at 21 and 22 be added paging links are clicked. + You need to register event by the delegate because these links are added dynamically. + 20 Similarly, these events call the search functions.

    +
    + + +

    + Describes the search processing function doSearch 3. +

    +

    + Gets the starting position and number of 4. + The value of these in the search form in the header area saved as a hidden value. + Display start position is set the default value if other values because 0 or more, the number of intended value from 1 to 100.

    +

    + Determines the value of parameter navi passed doSearch event registration, was 5 when the fix at the start of the show. + Here,-1 is previous page move, 1 following page move, otherwise, it will change to move to the first page.

    +

    + The decision to terminate the process and run a search if entered the search input field value is 6, if empty, without anything. +

    +

    + 7 in order to prevent double sub Mitt Fess Server contact while search button to turn off in. +

    +

    + Assemble the URL to send Ajax requests in 8. + 1 URL search, starting position, and number of joins.

    +

    + Send Ajax requests in 9. + Use the JSONP so specify the jsonp on dataType. + Change the json to use JSON. + Request comes back to normal, and runs the success function. + search results from Fess server returned success argument object is passed.

    +

    + First of all, in 10 has confirmed the contents of the response status. + 0 is set if the search request was processed successfully. More information on Fess JSON responseFess sitesPlease make sure. +

    +

    + Display a message if the search request is processed correctly, did not hit the search results 11 conditional statements in with content of the subheader area empty, did not hit results in the result area. +

    +

    + If you hit search results in the condition statement in the 12 search result processing. + Sets the execution time and number of messages in the 13 subheader area. + 14 we will add a reault area search results. + Search results are stored as an array of data.response.result. + results [i]...-in by accessing search results can retrieve field values of the document.

    +

    + Page number in the 15 current page and add a link to the previous page and next page result. + With 16 saves current starting position and number of hidden in the search form. + The starting position and number of search requests at next reuse.

    +

    + Change the on-screen position of pages in 17. + For the page itself is not updated when clicked next page links to scrawlTop moves to the top of the page.

    +

    + 18 the search process after search button to enable the. + The request succeeds or fails will perform complete to be called in.

    +

    + Avoid sending form or link after a search operation functions was called 19 returns false. + From this page transitions occur.

    +
    + + +

    + access the 'index.html' in the browser. + Displays the search form:

    +
    +
    Search form
    +
    + Search form +
    +
    +

    + Enter a suitable search term and press the search button to display search results. + If there are 20 is the default display number, hit the number of search results displays links to the following pages in the bottom of the list.

    +
    +
    Search results
    +
    + Search results +
    +
    +
    + +
    + +
    +

    + Using the REST API of the Fess and tried building jQuery-based client search site. + Build system not only browser-based applications that use the REST API, use the Fess called from another application.

    +

    + I want to show you how the next time the database scrawl functionality to add full-text search capabilities to an existing database. +

    +
    + +
    + +
    + +
    diff --git a/src/site/en/xdoc/dev/getting-started.xml b/src/site/en/xdoc/dev/getting-started.xml new file mode 100644 index 000000000..9e6bc5925 --- /dev/null +++ b/src/site/en/xdoc/dev/getting-started.xml @@ -0,0 +1,73 @@ + + + + Open source full-text search server - Fess development overview + Shinsuke Sugaya + + + + + + +
    +

    Summarizes the information needed to develop the Fess.

    +
    +
    +

    Fess is developed as an application to work in the environment of the Java 7 or more. The following knowledge becomes necessary.

    +
      +
    • Java
    • +
    • Seasar 2
    • +
    • SAStruts (if you are developing a Web screen)
    • +
    • DBFlute (if you are developing around the DB)
    • +
    • SOLR (if you are developing around the search index)
    • +
    • S2Robot (if you are developing around the crawler)
    • +
    +

    When you develop using Eclipse and Maven in this scenario (to build to generate in the fess-server release of Ant becomes necessary). Also to develop and operate Fess server for download and you have installed. Dude install in advance is what you need.

    +
    +
    +

    Summarizes how to develop management and search screens. Describes how to develop with Eclipse. You must keep your development on the Eclipse WTP, etc (you installed the J2EE Edition).

    +
      +
    1. Install Java, Eclipse, Maven 3.x, Fess and prepare. A zip file of the Fess <FESS_HOME>suppose you deploy directory. </FESS_HOME>
    2. +
    3. The source code of the Fess clone from github. + +
    4. +
    5. As a Maven project into the Eclipse.
    6. +
    7. Display the servers view. If you don't see the Window > Show View > Other... In, display the dialog Server > Servers, select OK button press concludes.
    8. +
    9. Add a new server in the servers view. Tomcat v7.0 Server, select the server name is set to moderate, the Next button press concludes. The Add Configured: fess Finish button press concludes. So that server in the servers view is displayed, double click the display settings information (Overview).
    10. +
    11. Select Use Tomcat Installation Server Locations.
    12. +
    13. In the timeouts Start to 180 seconds, change the Stop to 60 seconds.
    14. +
    15. Click the General Information Open Launch Configuration. Click the tag arguments. In the VM arguments '-Dsolr.solr.home=<FESS_HOME>/solr-Dfess.log.file=<FESS_HOME>/logs/fess.out-Dsolr.log.file=<FESS_HOME>/logs/solr.log-Djava.awt.headless=true---Xmx1g-XX: UseTLAB - XX:-XX:MaxMetaspaceSize = DisableExplicitGC 128 m-XX: Compressedclssspcesize = 32 m - XX:-usegcoverhedlimit-XX: UseConcMarkSweepGC - XX: XX - XX: Cmsincrementlmode - XX: Cmsincrementlpcing - XX: Cmsincrementldutycyclemin = 0 - Cmsinititingoccupncyfrction = 75: Useprnewgc - XX: OptimizeStringConcat ' add the. <FESS_HOME>Will change depending on the environment. OK button press concludes. (If the Java 7-XX:MaxMetaspaceSize = 128 m-XX:CompressedClassSpaceSize = 32 m-XX:MaxPermSize = 128 m to replace)</FESS_HOME> </FESS_HOME> </FESS_HOME> </FESS_HOME>
    16. +
    17. Boot the server from the servers view.
    18. +
    +

    If you want to develop a HOT deploy src/main/resources/env.txt change from the product the ct. You can change the source code, without having to restart the Tomcat (Fess).

    +
    +
    +

    The crawl process Tomcat (Fess) starts a process. If you want to pursue in the debugger, such as to register as for debugging Java applications on Eclipse.

    +
      +
    1. Register as a normal Java Application can be debugged in Eclipse. main function will be to jp.sf.fess.exec.Crawler.
    2. +
    3. In settings 1, argument settings, first, passes a session ID-sessionId 20100101000000 to appropriate program settings. Expand the contents of the Fess bin/setenv.sh to the VM Settings and register.
    4. +
    5. Add /fess/src/main/webapp/WEB-INF/cmd and geronimo_servlet_9.4_spec-1.0.jar in the classpath setting.
    6. +
    7. The run.
    8. +
    +
    +
    +

    Distribution of Fess is included in Tomcat war file of the Fess and Solr. Build the distribution of Fess in SVN fess-server. In order to build Ant becomes necessary.

    + +
    +
    + +
    + +
    diff --git a/src/site/en/xdoc/donation.xml b/src/site/en/xdoc/donation.xml new file mode 100644 index 000000000..696df0b09 --- /dev/null +++ b/src/site/en/xdoc/donation.xml @@ -0,0 +1,27 @@ + + + + Open source full-text search server - ads-project assistance + Shinsuke Sugaya + + + + + + +
    +

    The management of the project N2SM, Inc.... From the subject to support. We received contribution of advertising to raise awareness of the Fess. Fess, for many people seeking donations support projects.

    +
    + + + + +
    +

    Project assistance money paid will utilize as a cost of advertising such as AdWords. Contact us on project support, such as receiptClick hereTo contact us please.

    +
    +
    +

    In the Fess project site on AdSense ads. Income generated through AdSense will Fess awareness and to disseminate will invest in AdWords intact as advertising expenses.

    +

    Please take a look at AdSense in addition to you have any good ideas about the spread of the Fess,.

    +
    + +
    diff --git a/src/site/en/xdoc/index.xml b/src/site/en/xdoc/index.xml new file mode 100644 index 000000000..84b4a1682 --- /dev/null +++ b/src/site/en/xdoc/index.xml @@ -0,0 +1,83 @@ + + + + Open source full-text search server Fess (Fess) + Shinsuke Sugaya + + + + + + +
    +

    The Fess 'Easily 5 minutes with full-text search server can be built"Is. Any operating system, it is possible to have a Java runtime environment. Fess is offered in the Apache license and offers free (freeware).

    +
    +
    +
    +
    + +
    Standard demo
    +
    +
    + +
    Sites within search demo
    +
    +
    + +
    EC-friendly product search demo
    +
    +
    +
    +
    +
    +
      +
    • Easily 5 minutes with full-text search server can be built
    • +
    • In the Apache license (free soft, so made available free) offer
    • +
    • OS-independent (Java based build)
    • +
    • Web, file system, Windows shared folder and database crawling.
    • +
    • Support for many file formats, such as MS Office and PDF
    • +
    • Apache Solr Available as a search engine
    • +
    • BASIC/DIGEST/NTLM authentication site can search
    • +
    • Log status enables out placement of search results (in an ionic environment such as ActiveDirectory available)
    • +
    • Josearch with map information
    • +
    • On the browser can crawl target setting and search screen editing
    • +
    • Sort search results by label
    • +
    • Path translation request header information added, overlapping domain settings, search results
    • +
    • Search results in XML/JSON/JSONP format output enables cooperation with external systems
    • +
    • Summary of search logs and click Log
    • +
    • Faceted drill-down support
    • +
    • AutoComplete / suggest feature
    • +
    • User dictionary and synonym dictionary editing function
    • +
    • Related documents view features
    • +
    • Search results cache feature and screen shots view features
    • +
    • Proxy function of search results
    • +
    • Click Open feature of desktop search and local files
    • +
    • Smartphone support (Responsive Web Design)
    • +
    • To the SOLR server group and for Motorsport
    • +
    • Depending on the application the flexibility to design available
    • +
    +
    +
    +
    +
    2014-06-28
    +
    + Fess 9.2 released +
    +
    2014-03-26
    +
    + Fess 9.1 released +
    +
    2014-02-07
    +
    + Fess 9.0.0 release +
    +
    +

    News of the pastClick herePlease visit.

    +
    +
    +

    Fess is Apache licensed open-source products, personal and commercial free in freely available.

    + +

    If you need and Fess customization, implementation and support servicesCommercial supportPlease visit. Performance tuning, such as, search quality and crawling slowly accommodated in commercial support.

    +
    + +
    diff --git a/src/site/en/xdoc/news.xml b/src/site/en/xdoc/news.xml new file mode 100644 index 000000000..5aaca88da --- /dev/null +++ b/src/site/en/xdoc/news.xml @@ -0,0 +1,153 @@ + + + + Open source full-text search server - news list + Shinsuke Sugaya + + + + + + +
    +
    +
    2014-06-28
    +
    + Fess 9.2 released +
    +
    2014-03-26
    +
    + Fess 9.1 released +
    +
    2014-02-07
    +
    + Fess 9.0.0 release +
    +
    2013-09-07
    +
    + Fess 8.2.0 release +
    +
    2013-05-13
    +
    + Fess 8.1.0 release +
    +
    2013-02-23
    +
    + Fess 8.0.0 release +
    +
    2012-11-27
    +
    + Fess 7.0.2 release +
    +
    2012-09-09
    +
    + Fess 7.0.1 release +
    +
    2012-08-03
    +
    + Fess 7.0.0 release +
    +
    2012-03-02
    +
    + Fess 6.0.0 release +
    +
    2011-12-30
    +
    + Fess 5.0.1 released +
    +
    2011-12-15
    +
    + Fess 5.0 release +
    +
    2011-08-12
    +
    + Java support in the Fess 4.1 version +
    +
    2011-07-24
    +
    + Fess 4.1.0 release +
    +
    2011-01-12
    +
    + 「Part of the Fess in Apache Solr based search server-REST API' Was published +
    +
    2010-12-19
    +
    + Fess 4.0.0 release +
    +
    2010-12-16
    +
    + 「Part of the Fess in Apache Solr based search server-role-based search' Was published +
    +
    2010-11-08
    +
    + Fess 4.0.0 beta release +
    +
    2010-08-31
    +
    + Fess 3.1.3 released +
    +
    2010-07-08
    +
    + Fess 3.1.2 release +
    +
    2010-07-08
    +
    + iPhone app for Fess client 1.0.0Release +
    +
    2010-06-13
    +
    + "Java Cloud Meeting Tokyo 2010 x Seasar Conference 2010In introducing the Fess. +
    +
    2010-06-09
    +
    + Fess 3.1.1 release +
    +
    2010-05-30
    +
    + Fess 3.1.0 release +
    +
    2010-03-18
    +
    + "N2 Search (Fess), introduce free seminarIn introducing the Fess. +
    +
    2010-03-11
    +
    + 2 Solr seminarIn introducing the Fess. +
    +
    2010-01-16
    +
    + Fess 2.0.0 released +
    +
    2009-12-09
    +
    + Fess 1.1.2 released +
    +
    2009-11-20
    +
    + Commentary articles published in Wall Street Journal +
    +
    2009-11-12
    +
    + Introduction article on CodeZine +
    +
    2009-10-29
    +
    + Fess 1.1.1 released +
    +
    2009-10-10
    +
    + Fess 1.1.0 released +
    +
    2009-09-10
    +
    + Introduction to Fess at Seasar Conference 2009 Autumn +
    +
    2009-09-10
    +
    + Fess 1.0.0 released +
    +
    +
    + +
    diff --git a/src/site/en/xdoc/roadmap.xml b/src/site/en/xdoc/roadmap.xml new file mode 100644 index 000000000..2409bcc2e --- /dev/null +++ b/src/site/en/xdoc/roadmap.xml @@ -0,0 +1,35 @@ + + + + Open source full text search server – roadmap + Shinsuke Sugaya + + + + + + +
    +

    Version is provided in the form X.Y.Z.

    +
      +
    • X: major version. If the information database of the Fess or Solr schema has been updated to go up version.
    • +
    • Y: minor version. If not updated information database of the Fess or Solr schema additions and bug fixes up version.
    • +
    • Z: micro version. Critical bug fixes if you go up version. (Even not fatal if requested by users ' request. Respond on a case by case basis)
    • +
    +
    +
    +
      +
    • Bug fixes
    • +
    • SOLR update
    • +
    +

    Add the other, if anything.

    +
    +
    +
      +
    • Java 8
    • +
    • SOLR 5 introduction
    • +
    +

    Add the other, if anything.

    +
    + +
    diff --git a/src/site/en/xdoc/setup.xml b/src/site/en/xdoc/setup.xml new file mode 100644 index 000000000..d32b7539a --- /dev/null +++ b/src/site/en/xdoc/setup.xml @@ -0,0 +1,216 @@ + + + + Installation of the Fess + Shinsuke Sugaya + + + + + + +
    +

    Any operating system, it is possible to Fess any Java runtime environment. Work environment is as follows.

    +
      +
    • You can run Java, including Windows and Unix OS environment
    • +
    • Java: Java 7 or more
    • +
    +
    +
    +

    If Java is not installed please install Java by following these steps.

    + +

    With JavaScript enabled for OracleJava SE Downloads pageTo access the.

    +

    * Different versions and follow these steps butJava 7u25Install (recommended). [2014 / 3 / 13]

    +

    Click the 'Download JDK' JavaSE7. (JavaScript is disabled and download is valid and must)

    + Access +
    +

    Whether JavaScript is enabled you can check below. (If the Internet Explorer 9)

    +
      +
    1. Click [tools] on the menu bar.
    2. +
    3. Click the [Internet Options].
    4. +
    5. Click the Security tab.
    6. +
    7. Click the [custom level].
    8. +
    9. Scrawl to the [scripts] section is displayed.
    10. +
    11. In the Active Scripting section radio button to enable a check to confirm.
    12. +
    13. Turn on and off if it is, click OK.
    14. +
    +
    +
    + +

    When you accept the license, read 'The Oracle Binary Code License Agreement for Java SE' turn to Accept License Agreement check.

    + License +
    + +

    Do the JDK download according to the OS you install. Windwos 64-bit if you select Windows 86 x 64 Windows, for Windows 32-bit (Windows 64-bit version of the example is below).

    +
    +

    PCs are used whether the you can check below. (For Windows 7)

    +
      +
    1. In Control Panel-&gt; system and security → [System] appears to the type of system.
    2. +
    +
    + Download +
    + +

    (XX is the downloaded update release version) run the JDK Installer (jdk-7uXX-windows-x64.exe). The following is an example Windows 64-bit Edition.

    + Run +

    Depending on settings in Windows, see allow changes to this computer to the following program? "That may receive dialog. If so, please click the [Yes] button.

    +
    + +

    The installer will launch. Press the [next] button.

    + Run +

    You can change the destination folder. Left at default without problem, press the [next] button.

    + Directory +

    Start installing the JDK, so wait.

    + JDK installation +
    + +

    Run the JRE installation after installing the JDK, JRE installed on your PC. In the JDK as well to change the destination folder is possible, but basically the same JDK folder no problem. Press the [next] button.

    + JRE +

    JRE installation starts, so wait a while.

    + JRE installation +
    + +

    The installation complete message appears. Press the [close] button.

    + JRE installation +

    Installation is complete.

    +
    +

    The components is the following two. In the following you can check. (For Windows 7)

    +
      +
    1. [Control Panel] → [programs] → [programs and features] listed will be. +
        +
      • Java SE Development Kit 7 Update XX (64-bit)
      • +
      • Java (TM) Update 7 XX (64-bit)
      • +
      +
    2. +
    +
    +
    + +

    This is the setting information and environment variables passed to the program. In order to run the JDK command at the command prompt after the Java installation, setting environment variables is required.

    +

    Windows 7 sets the following. Select the control panel → system and security → [System] → advanced system settings → environment variables.

    +Control Panel +

    Click system and security.

    +System and security +

    Click the 'system'.

    +System +

    Click the 'advanced settings'.

    +Advanced system settings +

    Click environment variables.

    +Environment variables +

    New system environment variables button at the bottom of screen.

    + +

    Enter "JAVA_HOME" variable name.

    + +

    Describes the directory where you installed the JDK to the variable value.

    +

    Open the C:\Program Files zone\scripting in the Explorer, ' jdk... ' that describes its addresses, looking for a folder.

    +

    For example if you have installed the jdk version 1.7.0_XX will be C:\Program Files \Java\jdk1.7.0_XX. (Contains the portion of the XX version)

    +

    Description and then press 'OK'.

    +

    Locate the row from a list of system environment variables, see variables 'Path'.

    + +

    And then push open the Edit button, the row at the end of the variable value "; %JAVA_HOME%\bin ' that adds a string, click the 'OK'.

    + +
    +
    +
    + +

    http://sourceforge.jp/projects/Fess/releases/ From the download the latest Fess package.

    +

    Release file list of the destination URL and then 'fess-server-9.x.y.zip'.

    + +
    + +

    Unzip the zip file you downloaded. If the Windows environment expand the zip decompression tool, in.

    +
    +

    If you installed in the UNIX environment, in the bin added the performing rights to a script.

    + +
    + +

    Open the unzipped folder by double-clicking.

    + +

    Open the bin folder by double-clicking.

    + +
    + +

    Double-click the startup.bat file in the bin folder, Fess up.

    +
    +

    Run the startup.sh for UNIX environments.

    + +
    + +

    Will start the command prompt appears. In the last sentence "Server startup... ' is Setup complete appears.

    + +
    +
    +
    +

    access to / http://localhost:8080/Fess ensures startup.

    +

    Management UI is / http://localhost:8080/fess/admin. Default Administrator account user name / password is admin/admin. Administrator account is managed by the application server. In the management UI of the Fess, authenticate with the application server in fess role determine as an administrator.

    +
    +
    + +

    stop the Fess, double-click shutdown.bat file in the bin folder.

    +
    +

    Run the shutdown.sh for UNIX environments.

    + +
    +
    + +

    Administrator account is managed by the application server. Fess Server standard available Tomcat, as well as to the user changing the Tomcat. Modify the password for the admin account of the conf/tomcat-user.xml if you want to change.

    + +]]> +
    + +

    +To access the Solr into Fess server is password becomes necessary. +Change the default passwords in production, etc.

    +

    +How to change the password, you must first change the password attribute of the conf/tomcat-user.xml solradmin. +

    + +]]> +

    +Modifies the following three files webapps/fess/WEB-INF/classes/solrlib.dicon, fess_suggest.dicon and solr/core1/conf/solrconfig.xml. +Write what you specified in tomcat-user.xml to the following password.

    +

    +modify the following areas of the solrlib.dicon. +

    + + "solradmin" + "solradmin" + +]]> +

    +fess_suggest.dicon is the following. +

    + + "solradmin" + "solradmin" + +]]> +

    +SOLR/core1/conf/solrconfig.XML is the following. +

    + + + + http://localhost:8080/solr/core1-suggest + + solradmin + solradmin + +]]> +
    +
    + +
    diff --git a/src/site/en/xdoc/supported-files.xml b/src/site/en/xdoc/supported-files.xml new file mode 100644 index 000000000..cc6a62d2a --- /dev/null +++ b/src/site/en/xdoc/supported-files.xml @@ -0,0 +1,52 @@ + + + + Search for files + Shinsuke Sugaya + + + + + + +
    +

    This is a list of the Fess found on the crawl and search file format.

    + + + + + + + + + + + + + + + + + + + + + + + + + +
    TypeExtension
    Texttxt
    XMLXML
    XMLmm
    XMLXML
    HTMLHTML
    MS OfficePPT
    MS Officedoc
    MS Officepptx
    MS Officexls
    MS Officexlsx
    MS Officedocx
    PDFPDF
    Source codeJS
    Source codec
    Source codeh
    Source codeJava
    Source codeHPP
    Source codeCPP
    Compressed filegz
    Compressed filetar
    Compressed filezip
    +

    Character string extraction from various kinds of unknown files Fess. Not listed in the above file also crawl and you can find. If you have files you want to check,Search for system test data repositoryTo please pull requests.

    +
    +
    +

    The following file corresponds with commercial support.

    +
      +
    • A taro
    • +
    • OASYS for Windows
    • +
    • DocuWorks
    • +
    • AutoCAD
    • +
    +
    + +
    diff --git a/src/site/tools/translatexml b/src/site/tools/translatexml new file mode 100755 index 000000000..50d2744a7 --- /dev/null +++ b/src/site/tools/translatexml @@ -0,0 +1,26 @@ +#!/bin/sh +CWD="${0%/*}" + +if [ $# -ne 2 ]; then + echo "Usage: $0 input_xml_dir output_xml_dir" 1>&2 + exit 1 +fi + +in_dir=$1 +out_dir=$2 + +find $in_dir -name "*.xml" -print | while read in; do + out=${in/${in_dir}/${out_dir}} + d=$(dirname $out) + mkdir -p $d + echo $out 1>&2 + ruby ${CWD}/translatexml.rb -f ja -t en -i $in | \ + sed -e ' + s/Croll/Crawl/g + s/croll/crawl/g + s/Crolling/Crawling/g + s/crolling/crawling/g + ' \ + > $out +done + diff --git a/src/site/tools/translatexml.rb b/src/site/tools/translatexml.rb new file mode 100644 index 000000000..8bd42c472 --- /dev/null +++ b/src/site/tools/translatexml.rb @@ -0,0 +1,183 @@ +# -*- coding:utf-8 -*- +require 'net/http' +#require 'uri' +require 'rexml/document' +require 'json' +require 'rest_client' +require 'optparse' + +# +# parsing command line options +# +options = {} + +optparse = OptionParser.new do |parser| + parser.on('-f lang', '--from lang', 'Input language can be \'ja\', \'en\',.. (Required)') {|v| options[:from] = v} + parser.on('-t lang', '--to lang', 'Output language can be \'ja\', \'en\',.. (Required)') {|v| options[:to] = v} + parser.on('-i [file]', '--input [file]', 'Input file (Optional)') {|v| options[:input] = v} + parser.on('-o [file]', '--output [file]', 'Output file (Optional)') {|v| options[:output] = v} + parser.on('-v', '--verbose', 'Verbose message to STDERR (Optional)') {|v| options[:verbose] = v} + parser.on('-d', '--dryrun', 'Dry-run (don\'t translate) (Optional)') {|v| options[:dryrun] = v} +end + +begin + optparse.parse! + mandatory = [:from, :to] + missing = mandatory.select{ |param| options[param].nil? } + if not missing.empty? + STDERR.puts "Missing options: #{missing.join(', ')}" + STDERR.puts optparse + exit + end +rescue OptionParser::InvalidOption, OptionParser::MissingArgument + STDERR.puts $!.to_s + STDERR.puts optparse + exit +end + +$from = options[:from] +$to = options[:to] + +if (options[:input] == nil) + $input = STDIN +else + $input = open(options[:input]) +end + +if (options[:output] == nil) + $output = STDOUT +else + $output = open(options[:output], "w") +end + +$verbose = options[:verbose] +$dryrun = options[:dryrun] + +# +# Translator class +# +class Translator + CLIENT_ID = 'Set Client ID' + CLIENT_SECRET = 'Set Client Secret Key' + AUTHORIZE_URL = 'https://datamarket.accesscontrol.windows.net/v2/OAuth2-13' + TRANSLATION_URL = 'http://api.microsofttranslator.com/V2/Http.svc/Translate' + SCOPE = 'http://api.microsofttranslator.com' + + @@access_token = nil + + def get_access_token + unless @@access_token == nil + return @@access_token + end + json = JSON.parse( + RestClient.post(AUTHORIZE_URL, + { + 'grant_type' => 'client_credentials', + 'client_id' => CLIENT_ID, + 'client_secret' => CLIENT_SECRET, + 'scope' => SCOPE, + }, + :content_type => 'application/x-www-form-urlencoded' + ) + ) + @@access_token = json['access_token'] + @@access_token + end + private:get_access_token + + def translate(text, from, to) + access_token = get_access_token + unless $dryrun + xml = REXML::Document.new( + RestClient.get("#{TRANSLATION_URL}?from=#{from}&to=#{to}&text=#{URI.escape(text)}", + 'Authorization' => "Bearer #{access_token}" + ) + ) + xml.root.text + else + "..." + end + end +end + +# +# Extends REXML::Element class +# +class REXML::Element + def has_cdata? + self.cdatas.length > 0 + end +end + +# +# translateNode +# +def translateNode(element) + + translator = Translator.new + + # Translate attributes + if (element.is_a?(REXML::Element)) + if (element.has_attributes?) + $attributes.each do |attribute| + text = element.attributes[attribute] + if /\S+/ =~ text + unless (text.nil? || text.empty?) + STDERR.puts "attributes[#{attribute}]=#{text}" if $verbose + element.attributes[attribute] = translator.translate(text, $from, $to) + STDERR.print "." unless $verbose + STDERR.puts " =>#{element.attributes[attribute]}" if $verbose + end + end + end + end + + # Translate recursively if has children + if (element.has_elements?) + element.map.each do |child| + translateNode(child) + end + return + end + + # Noting to do if CDATA + if (element.has_cdata?) + return + end + end + + # Translate the text + if (element.is_a?(REXML::Text)) + text = element.value + if /\S+/ =~ text + unless (text.nil? || text.empty?) + STDERR.puts "text=#{text}" if $verbose + element.value = translator.translate(text, $from, $to) + STDERR.print "." unless $verbose + STDERR.puts " =>#{element.value}" if $verbose + end + end + elsif (element.is_a?(REXML::Element) && element.has_text?) + text = element.text + if /\S+/ =~ text + unless (text.nil? || text.empty?) + STDERR.puts "text=#{text}" if $verbose + element.text = translator.translate(text, $from, $to) + STDERR.print "." unless $verbose + STDERR.puts " =>#{element.text}" if $verbose + end + end + end +end + + +# +# parsing xml and translate (main) +# +$attributes = Array["name", "alt", "content"] + +doc = REXML::Document.new($input) +translateNode(doc.root) +$output.puts doc.to_s + +STDERR.print "\n" unless $verbose