Ver código fonte

remove solr and db files

Shinsuke Sugaya 10 anos atrás
pai
commit
8b876ed8ad

+ 0 - 36
solr/conf/elevate.xml

@@ -1,36 +0,0 @@
-<?xml version="1.0" encoding="UTF-8" ?>
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements.  See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License.  You may obtain a copy of the License at
-
-     http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-
-<!-- If this file is found in the config directory, it will only be
-     loaded once at startup.  If it is found in Solr's data
-     directory, it will be re-loaded every commit.
--->
-
-<elevate>
- <query text="foo bar">
-  <doc id="1" />
-  <doc id="2" />
-  <doc id="3" />
- </query>
- 
- <query text="ipod">
-   <doc id="MA147LL/A" />  <!-- put the actual ipod at the top -->
-   <doc id="IW-02" exclude="true" /> <!-- exclude this cable -->
- </query>
- 
-</elevate>

+ 0 - 21
solr/conf/protwords.txt

@@ -1,21 +0,0 @@
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-#-----------------------------------------------------------------------
-# Use a protected word file to protect against the stemmer reducing two
-# unrelated words to the same base word.
-
-# Some non-words that normally won't be encountered,
-# just to test that they won't be stemmed.
-dontstems
-zwhacky
-

+ 0 - 132
solr/conf/schema.xml

@@ -1,132 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements.  See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License.  You may obtain a copy of the License at
-
-     http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-<schema name="fess" version="1.2">
-  <types>
-    <fieldType name="text" class="solr.TextField" positionIncrementGap="100">
-      <analyzer type="index">
-        <tokenizer class="solr.CJKTokenizerFactory"/>
-        <!--
-        <tokenizer class="solr.WhitespaceTokenizerFactory"/>
-        <tokenizer class="solr.StandardTokenizerFactory"/>
-        -->
-        <!-- 
-        <filter class="solr.SynonymFilterFactory" synonyms="index_synonyms.txt" ignoreCase="true" expand="false"/>
-        -->
-        <filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt" enablePositionIncrements="true" />
-        <filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0" splitOnCaseChange="1"/>
-        <filter class="solr.LowerCaseFilterFactory"/>
-        <filter class="solr.SnowballPorterFilterFactory" language="English" protected="protwords.txt"/>
-        <!-- <filter class="solr.RemoveDuplicatesTokenFilterFactory"/> -->
-      </analyzer>
-      <analyzer type="query">
-        <tokenizer class="solr.WhitespaceTokenizerFactory"/>
-        <filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="true"/>
-        <filter class="solr.StopFilterFactory"
-                ignoreCase="true"
-                words="stopwords.txt"
-                enablePositionIncrements="true"
-                />
-        <filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0" splitOnCaseChange="1"/>
-        <filter class="solr.LowerCaseFilterFactory"/>
-        <filter class="solr.SnowballPorterFilterFactory" language="English" protected="protwords.txt"/>
-        <!-- <filter class="solr.RemoveDuplicatesTokenFilterFactory"/> -->
-      </analyzer>
-    </fieldType>
-    <fieldType name="url" class="solr.TextField" positionIncrementGap="100">
-      <analyzer>
-        <tokenizer class="solr.StandardTokenizerFactory"/>
-        <filter class="solr.LowerCaseFilterFactory"/>
-        <filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1"/>
-        <filter class="solr.RemoveDuplicatesTokenFilterFactory"/>
-      </analyzer>
-    </fieldType>
-    <fieldType name="string" class="solr.StrField" sortMissingLast="true" omitNorms="true"/>
-    <fieldType name="boolean" class="solr.BoolField" sortMissingLast="true" omitNorms="true"/>
-    <fieldtype name="binary" class="solr.BinaryField"/>
-    <fieldType name="int" class="solr.TrieIntField" precisionStep="0" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="float" class="solr.TrieFloatField" precisionStep="0" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="long" class="solr.TrieLongField" precisionStep="0" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="double" class="solr.TrieDoubleField" precisionStep="0" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="tint" class="solr.TrieIntField" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="tfloat" class="solr.TrieFloatField" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="tlong" class="solr.TrieLongField" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="tdouble" class="solr.TrieDoubleField" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>
-    <fieldType name="date" class="solr.TrieDateField" omitNorms="true" precisionStep="0" positionIncrementGap="0"/>
-    <fieldType name="tdate" class="solr.TrieDateField" omitNorms="true" precisionStep="6" positionIncrementGap="0"/>
-    <fieldType name="pint" class="solr.IntField" omitNorms="true"/>
-    <fieldType name="plong" class="solr.LongField" omitNorms="true"/>
-    <fieldType name="pfloat" class="solr.FloatField" omitNorms="true"/>
-    <fieldType name="pdouble" class="solr.DoubleField" omitNorms="true"/>
-    <fieldType name="pdate" class="solr.DateField" sortMissingLast="true" omitNorms="true"/>
-    <fieldType name="sint" class="solr.SortableIntField" sortMissingLast="true" omitNorms="true"/>
-    <fieldType name="slong" class="solr.SortableLongField" sortMissingLast="true" omitNorms="true"/>
-    <fieldType name="sfloat" class="solr.SortableFloatField" sortMissingLast="true" omitNorms="true"/>
-    <fieldType name="sdouble" class="solr.SortableDoubleField" sortMissingLast="true" omitNorms="true"/>
-    <fieldType name="random" class="solr.RandomSortField" indexed="true" />
-  </types>
-  <fields>
-    <field name="id" type="string" stored="true" indexed="true"/>
-<!-- core fields -->
-    <field name="segment" type="string" stored="true" indexed="true"/>
-    <field name="digest" type="text" stored="true" indexed="false"/>
-    <field name="boost" type="float" stored="true" indexed="false"/>
-    <field name="host" type="url" stored="true" indexed="true"/>
-    <field name="site" type="string" stored="true" indexed="false"/>
-    <field name="url" type="url" stored="true" indexed="true" required="true"/>
-    <field name="content" type="text" stored="false" indexed="true"/>
-    <field name="title" type="text" stored="true" indexed="true"/>
-    <field name="cache" type="text" stored="true" indexed="false" compressed="true"/>
-    <field name="tstamp" type="slong" stored="true" indexed="true"/>
-    <field name="anchor" type="string" stored="true" indexed="true" multiValued="true"/>
-    <field name="contentLength" type="slong" stored="true" indexed="true"/>
-    <field name="lastModified" type="slong" stored="true" indexed="true"/>
-    <field name="date" type="string" stored="true" indexed="true"/>
-    <field name="lang" type="string" stored="true" indexed="true"/>
-    <field name="mimetype" type="string" stored="true" indexed="true"/>
-<!-- multi values -->
-    <field name="type" type="string" stored="true" indexed="true" multiValued="true"/>
-    <field name="label" type="string" stored="true" indexed="true" multiValued="true"/>
-    <field name="role" type="string" stored="true" indexed="true" multiValued="true"/>
-<!-- Dynamic field definitions -->
-    <dynamicField name="*_s" type="string" indexed="true" stored="true"/>
-    <dynamicField name="*_t" type="text" indexed="true" stored="true"/>
-    <dynamicField name="*_b" type="boolean" indexed="true" stored="true"/>
-    <dynamicField name="*_i" type="int" indexed="true" stored="true"/>
-    <dynamicField name="*_l" type="long" indexed="true" stored="true"/>
-    <dynamicField name="*_f" type="float" indexed="true" stored="true"/>
-    <dynamicField name="*_d" type="double" indexed="true" stored="true"/>
-    <dynamicField name="*_ti" type="tint" indexed="true" stored="true"/>
-    <dynamicField name="*_tl" type="tlong" indexed="true" stored="true"/>
-    <dynamicField name="*_tf" type="tfloat" indexed="true" stored="true"/>
-    <dynamicField name="*_td" type="tdouble" indexed="true" stored="true"/>
-    <dynamicField name="*_tdt" type="tdate" indexed="true" stored="true"/>
-    <dynamicField name="*_pi" type="pint" indexed="true" stored="true"/>
-    <dynamicField name="*_pl" type="plong" indexed="true" stored="true"/>
-    <dynamicField name="*_pf" type="pfloat" indexed="true" stored="true"/>
-    <dynamicField name="*_pd" type="pdouble" indexed="true" stored="true"/>
-    <dynamicField name="*_pdt" type="pdate" indexed="true" stored="true"/>
-    <dynamicField name="*_si" type="sint" indexed="true" stored="true"/>
-    <dynamicField name="*_sl" type="slong" indexed="true" stored="true"/>
-    <dynamicField name="*_sf" type="sfloat" indexed="true" stored="true"/>
-    <dynamicField name="*_sd" type="sdouble" indexed="true" stored="true"/>
-    <dynamicField name="*_dt" type="date" indexed="true" stored="true"/>
-  </fields>
-  <uniqueKey>id</uniqueKey>
-  <defaultSearchField>content</defaultSearchField>
-  <solrQueryParser defaultOperator="AND"/>
-</schema>

+ 0 - 1033
solr/conf/solrconfig.xml

@@ -1,1033 +0,0 @@
-<?xml version="1.0" encoding="UTF-8" ?>
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements.  See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License.  You may obtain a copy of the License at
-
-     http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-<!-- 
-     For more details about configurations options that may appear in this 
-     file, see http://wiki.apache.org/solr/SolrConfigXml.
-
-     Specifically, the Solr Config can support XInclude, which may make it easier to manage
-     the configuration.  See https://issues.apache.org/jira/browse/SOLR-1167
--->
-<config>
-  <!-- Set this to 'false' if you want solr to continue working after it has 
-       encountered an severe configuration error.  In a production environment, 
-       you may want solr to keep working even if one handler is mis-configured.
-
-       You may also set this to false using by setting the system property:
-         -Dsolr.abortOnConfigurationError=false
-     -->
-  <abortOnConfigurationError>${solr.abortOnConfigurationError:true}</abortOnConfigurationError>
-
-  <!-- lib directives can be used to instruct Solr to load an Jars identified
-       and use them to resolve any "plugins" specified in your solrconfig.xml or
-       schema.xml (ie: Analyzers, Request Handlers, etc...).
-
-       All directories and paths are resolved relative the instanceDir.
-
-       If a "./lib" directory exists in your instanceDir, all files found in it
-       are included as if you had used the following syntax...
-       
-              <lib dir="./lib" />
-    -->
-  <!-- A dir option by itself adds any files found in the directory to the
-       classpath, this is useful for including all jars in a directory.
-    -->
-  <lib dir="../../contrib/extraction/lib" />
-  <!-- When a regex is specified in addition to a directory, only the files in that
-       directory which completely match the regex (anchored on both ends)
-       will be included.
-    -->
-  <lib dir="../../dist/" regex="apache-solr-cell-\d.*\.jar" />
-  <lib dir="../../dist/" regex="apache-solr-clustering-\d.*\.jar" />
-  <!-- If a dir option (with or without a regex) is used and nothing is found
-       that matches, it will be ignored
-    -->
-  <lib dir="../../contrib/clustering/lib/downloads/" />
-  <lib dir="../../contrib/clustering/lib/" />
-  <lib dir="/total/crap/dir/ignored" /> 
-  <!-- an exact path can be used to specify a specific file.  This will cause
-       a serious error to be logged if it can't be loaded.
-  <lib path="../a-jar-that-does-not-exist.jar" /> 
-  -->
-
-  
-  <!-- Used to specify an alternate directory to hold all index data
-       other than the default ./data under the Solr home.
-       If replication is in use, this should match the replication configuration. -->
-  <dataDir>${solr.data.dir:./solr/data}</dataDir>
-
-
-  <!-- WARNING: this <indexDefaults> section only provides defaults for index writers
-       in general. See also the <mainIndex> section after that when changing parameters
-       for Solr's main Lucene index. -->
-  <indexDefaults>
-   <!-- Values here affect all index writers and act as a default unless overridden. -->
-    <useCompoundFile>false</useCompoundFile>
-
-    <mergeFactor>10</mergeFactor>
-    <!-- If both ramBufferSizeMB and maxBufferedDocs is set, then Lucene will flush
-     based on whichever limit is hit first.  -->
-    <!--<maxBufferedDocs>1000</maxBufferedDocs>-->
-
-    <!-- Sets the amount of RAM that may be used by Lucene indexing
-      for buffering added documents and deletions before they are
-      flushed to the Directory.  -->
-    <ramBufferSizeMB>32</ramBufferSizeMB>
-    <!-- <maxMergeDocs>2147483647</maxMergeDocs> -->
-    <maxFieldLength>10000</maxFieldLength>
-    <writeLockTimeout>1000</writeLockTimeout>
-    <commitLockTimeout>10000</commitLockTimeout>
-
-    <!--
-     Expert: Turn on Lucene's auto commit capability.  This causes intermediate
-     segment flushes to write a new lucene index descriptor, enabling it to be
-     opened by an external IndexReader.  This can greatly slow down indexing
-     speed.  NOTE: Despite the name, this value does not have any relation to
-     Solr's autoCommit functionality
-     -->
-    <!--<luceneAutoCommit>false</luceneAutoCommit>-->
-
-    <!--
-     Expert: The Merge Policy in Lucene controls how merging is handled by
-     Lucene.  The default in 2.3 is the LogByteSizeMergePolicy, previous
-     versions used LogDocMergePolicy.
-
-     LogByteSizeMergePolicy chooses segments to merge based on their size.  The
-     Lucene 2.2 default, LogDocMergePolicy chose when to merge based on number
-     of documents
-
-     Other implementations of MergePolicy must have a no-argument constructor
-     -->
-    <!--<mergePolicy class="org.apache.lucene.index.LogByteSizeMergePolicy"/>-->
-
-    <!--
-     Expert:
-     The Merge Scheduler in Lucene controls how merges are performed.  The
-     ConcurrentMergeScheduler (Lucene 2.3 default) can perform merges in the
-     background using separate threads.  The SerialMergeScheduler (Lucene 2.2
-     default) does not.
-     -->
-    <!--<mergeScheduler class="org.apache.lucene.index.ConcurrentMergeScheduler"/>-->
-
-	  
-    <!--
-      This option specifies which Lucene LockFactory implementation to use.
-      
-      single = SingleInstanceLockFactory - suggested for a read-only index
-               or when there is no possibility of another process trying
-               to modify the index.
-      native = NativeFSLockFactory  - uses OS native file locking
-      simple = SimpleFSLockFactory  - uses a plain file for locking
-
-      (For backwards compatibility with Solr 1.2, 'simple' is the default
-       if not specified.)
-    -->
-    <lockType>native</lockType>
-    <!--
-     Expert:
-    Controls how often Lucene loads terms into memory -->
-    <!--<termIndexInterval>256</termIndexInterval>-->
-  </indexDefaults>
-
-  <mainIndex>
-    <!-- options specific to the main on-disk lucene index -->
-    <useCompoundFile>false</useCompoundFile>
-    <ramBufferSizeMB>32</ramBufferSizeMB>
-    <mergeFactor>10</mergeFactor>
-    <!-- Deprecated -->
-    <!--<maxBufferedDocs>1000</maxBufferedDocs>-->
-    <!--<maxMergeDocs>2147483647</maxMergeDocs>-->
-
-    <!-- inherit from indexDefaults <maxFieldLength>10000</maxFieldLength> -->
-
-    <!-- If true, unlock any held write or commit locks on startup. 
-         This defeats the locking mechanism that allows multiple
-         processes to safely access a lucene index, and should be
-         used with care.
-         This is not needed if lock type is 'none' or 'single'
-     -->
-    <unlockOnStartup>false</unlockOnStartup>
-    
-    <!-- If true, IndexReaders will be reopened (often more efficient) instead
-         of closed and then opened.  -->
-    <reopenReaders>true</reopenReaders>
-
-    <!--
-     Expert:
-    Controls how often Lucene loads terms into memory.  Default is 128 and is likely good for most everyone. -->
-    <!--<termIndexInterval>256</termIndexInterval>-->
-
-    <!--
-        Custom deletion policies can specified here. The class must
-        implement org.apache.lucene.index.IndexDeletionPolicy.
-
-        http://lucene.apache.org/java/2_3_2/api/org/apache/lucene/index/IndexDeletionPolicy.html
-
-        The standard Solr IndexDeletionPolicy implementation supports deleting
-        index commit points on number of commits, age of commit point and
-        optimized status.
-
-        The latest commit point should always be preserved regardless
-        of the criteria.
-    -->
-    <deletionPolicy class="solr.SolrDeletionPolicy">
-      <!-- The number of commit points to be kept -->
-      <str name="maxCommitsToKeep">1</str>
-      <!-- The number of optimized commit points to be kept -->
-      <str name="maxOptimizedCommitsToKeep">0</str>
-      <!--
-          Delete all commit points once they have reached the given age.
-          Supports DateMathParser syntax e.g.
-          
-          <str name="maxCommitAge">30MINUTES</str>
-          <str name="maxCommitAge">1DAY</str>
-      -->
-    </deletionPolicy>
-
-    <!--  To aid in advanced debugging, you may turn on IndexWriter debug logging.
-      Setting to true will set the file that the underlying Lucene IndexWriter
-      will write its debug infostream to.  -->
-     <infoStream file="INFOSTREAM.txt">false</infoStream> 
-
-  </mainIndex>
-
-  <!--	Enables JMX if and only if an existing MBeanServer is found, use this
-    if you want to configure JMX through JVM parameters. Remove this to disable
-    exposing Solr configuration and statistics to JMX.
-
-		If you want to connect to a particular server, specify the agentId
-		e.g. <jmx agentId="myAgent" />
-
-		If you want to start a new MBeanServer, specify the serviceUrl
-		e.g <jmx serviceUrl="service:jmx:rmi:///jndi/rmi://localhost:9999/solr"/>
-
-		For more details see http://wiki.apache.org/solr/SolrJmx
-  -->
-  <jmx />
-
-  <!-- the default high-performance update handler -->
-  <updateHandler class="solr.DirectUpdateHandler2">
-    <!-- A prefix of "solr." for class names is an alias that
-         causes solr to search appropriate packages, including
-         org.apache.solr.(search|update|request|core|analysis)
-     -->
-
-    <!-- Perform a <commit/> automatically under certain conditions:
-         maxDocs - number of updates since last commit is greater than this
-         maxTime - oldest uncommited update (in ms) is this long ago
-         Instead of enabling autoCommit, consider using "commitWithin"
-         when adding documents. http://wiki.apache.org/solr/UpdateXmlMessages
-    <autoCommit> 
-      <maxDocs>10000</maxDocs>
-      <maxTime>1000</maxTime> 
-    </autoCommit>
-    -->
-
-
-    <!-- The RunExecutableListener executes an external command from a
-      hook such as postCommit or postOptimize.
-         exe - the name of the executable to run
-         dir - dir to use as the current working directory. default="."
-         wait - the calling thread waits until the executable returns. default="true"
-         args - the arguments to pass to the program.  default=nothing
-         env - environment variables to set.  default=nothing
-      -->
-    <!-- A postCommit event is fired after every commit or optimize command
-    <listener event="postCommit" class="solr.RunExecutableListener">
-      <str name="exe">solr/bin/snapshooter</str>
-      <str name="dir">.</str>
-      <bool name="wait">true</bool>
-      <arr name="args"> <str>arg1</str> <str>arg2</str> </arr>
-      <arr name="env"> <str>MYVAR=val1</str> </arr>
-    </listener>
-    -->
-    <!-- A postOptimize event is fired only after every optimize command
-    <listener event="postOptimize" class="solr.RunExecutableListener">
-      <str name="exe">snapshooter</str>
-      <str name="dir">solr/bin</str>
-      <bool name="wait">true</bool>
-    </listener>
-    -->
-
-  </updateHandler>
-  
-  <!-- Use the following format to specify a custom IndexReaderFactory - allows for alternate
-       IndexReader implementations.
-
-       ** Experimental Feature **
-       Please note - Using a custom IndexReaderFactory may prevent certain other features
-       from working. The API to IndexReaderFactory may change without warning or may even
-       be removed from future releases if the problems cannot be resolved.
-
-       ** Features that may not work with custom IndexReaderFactory **
-       The ReplicationHandler assumes a disk-resident index. Using a custom
-       IndexReader implementation may cause incompatibility with ReplicationHandler and
-       may cause replication to not work correctly. See SOLR-1366 for details.
-
-  <indexReaderFactory name="IndexReaderFactory" class="package.class">
-    Parameters as required by the implementation
-  </indexReaderFactory >
-  -->
-  <!-- To set the termInfosIndexDivisor, do this: -->
-  <!--<indexReaderFactory name="IndexReaderFactory" class="org.apache.solr.core.StandardIndexReaderFactory">
-    <int name="termInfosIndexDivisor">12</int>
-  </indexReaderFactory >-->
-
-
-  <query>
-    <!-- Maximum number of clauses in a boolean query... in the past, this affected
-        range or prefix queries that expanded to big boolean queries - built in Solr 
-        query parsers no longer create queries with this limitation. 
-        An exception is thrown if exceeded.  -->
-    <maxBooleanClauses>1024</maxBooleanClauses>
-
-
-    <!-- There are two implementations of cache available for Solr,
-         LRUCache, based on a synchronized LinkedHashMap, and
-         FastLRUCache, based on a ConcurrentHashMap.  FastLRUCache has faster gets
-         and slower puts in single threaded operation and thus is generally faster
-         than LRUCache when the hit ratio of the cache is high (> 75%), and may be
-         faster under other scenarios on multi-cpu systems. -->
-    <!-- Cache used by SolrIndexSearcher for filters (DocSets),
-         unordered sets of *all* documents that match a query.
-         When a new searcher is opened, its caches may be prepopulated
-         or "autowarmed" using data from caches in the old searcher.
-         autowarmCount is the number of items to prepopulate.  For LRUCache,
-         the autowarmed items will be the most recently accessed items.
-       Parameters:
-         class - the SolrCache implementation LRUCache or FastLRUCache
-         size - the maximum number of entries in the cache
-         initialSize - the initial capacity (number of entries) of
-           the cache.  (seel java.util.HashMap)
-         autowarmCount - the number of entries to prepopulate from
-           and old cache.
-         -->
-    <filterCache
-      class="solr.FastLRUCache"
-      size="512"
-      initialSize="512"
-      autowarmCount="0"/>
-
-    <!-- Cache used to hold field values that are quickly accessible
-         by document id.  The fieldValueCache is created by default
-         even if not configured here.
-      <fieldValueCache
-        class="solr.FastLRUCache"
-        size="512"
-        autowarmCount="128"
-        showItems="32"
-      />
-    -->
-
-   <!-- queryResultCache caches results of searches - ordered lists of
-         document ids (DocList) based on a query, a sort, and the range
-         of documents requested.  -->
-    <queryResultCache
-      class="solr.LRUCache"
-      size="512"
-      initialSize="512"
-      autowarmCount="0"/>
-
-  <!-- documentCache caches Lucene Document objects (the stored fields for each document).
-       Since Lucene internal document ids are transient, this cache will not be autowarmed.  -->
-    <documentCache
-      class="solr.LRUCache"
-      size="512"
-      initialSize="512"
-      autowarmCount="0"/>
-
-    <!-- If true, stored fields that are not requested will be loaded lazily.
-      This can result in a significant speed improvement if the usual case is to
-      not load all stored fields, especially if the skipped fields are large
-      compressed text fields.
-    -->
-    <enableLazyFieldLoading>true</enableLazyFieldLoading>
-
-    <!-- Example of a generic cache.  These caches may be accessed by name
-         through SolrIndexSearcher.getCache(),cacheLookup(), and cacheInsert().
-         The purpose is to enable easy caching of user/application level data.
-         The regenerator argument should be specified as an implementation
-         of solr.search.CacheRegenerator if autowarming is desired.  -->
-    <!--
-    <cache name="myUserCache"
-      class="solr.LRUCache"
-      size="4096"
-      initialSize="1024"
-      autowarmCount="1024"
-      regenerator="org.mycompany.mypackage.MyRegenerator"
-      />
-    -->
-
-   <!-- An optimization that attempts to use a filter to satisfy a search.
-         If the requested sort does not include score, then the filterCache
-         will be checked for a filter matching the query. If found, the filter
-         will be used as the source of document ids, and then the sort will be
-         applied to that.
-    <useFilterForSortedQuery>true</useFilterForSortedQuery>
-   -->
-
-   <!-- An optimization for use with the queryResultCache.  When a search
-         is requested, a superset of the requested number of document ids
-         are collected.  For example, if a search for a particular query
-         requests matching documents 10 through 19, and queryWindowSize is 50,
-         then documents 0 through 49 will be collected and cached.  Any further
-         requests in that range can be satisfied via the cache.  -->
-    <queryResultWindowSize>20</queryResultWindowSize>
-
-    <!-- Maximum number of documents to cache for any entry in the
-         queryResultCache. -->
-    <queryResultMaxDocsCached>200</queryResultMaxDocsCached>
-
-    <!-- a newSearcher event is fired whenever a new searcher is being prepared
-      and there is a current searcher handling requests (aka registered).
-      It can be used to prime certain caches to prevent long request times for
-      certain requests.
-    -->
-    <!-- QuerySenderListener takes an array of NamedList and executes a
-         local query request for each NamedList in sequence. -->
-    <listener event="newSearcher" class="solr.QuerySenderListener">
-      <arr name="queries">
-        <!--
-        <lst> <str name="q">solr</str> <str name="start">0</str> <str name="rows">10</str> </lst>
-        <lst> <str name="q">rocks</str> <str name="start">0</str> <str name="rows">10</str> </lst>
-        <lst><str name="q">static newSearcher warming query from solrconfig.xml</str></lst>
-        -->
-      </arr>
-    </listener>
-
-    <!-- a firstSearcher event is fired whenever a new searcher is being
-         prepared but there is no current registered searcher to handle
-         requests or to gain autowarming data from. -->
-    <listener event="firstSearcher" class="solr.QuerySenderListener">
-      <arr name="queries">
-        <lst> <str name="q">solr rocks</str><str name="start">0</str><str name="rows">10</str></lst>
-        <lst><str name="q">static firstSearcher warming query from solrconfig.xml</str></lst>
-      </arr>
-    </listener>
-
-    <!-- If a search request comes in and there is no current registered searcher,
-         then immediately register the still warming searcher and use it.  If
-         "false" then all requests will block until the first searcher is done
-         warming. -->
-    <useColdSearcher>false</useColdSearcher>
-
-    <!-- Maximum number of searchers that may be warming in the background
-      concurrently.  An error is returned if this limit is exceeded. Recommend
-      1-2 for read-only slaves, higher for masters w/o cache warming. -->
-    <maxWarmingSearchers>2</maxWarmingSearchers>
-
-  </query>
-
-  <!-- 
-    Let the dispatch filter handler /select?qt=XXX
-    handleSelect=true will use consistent error handling for /select and /update
-    handleSelect=false will use solr1.1 style error formatting
-    -->
-  <requestDispatcher handleSelect="true" >
-    <!--Make sure your system has some authentication before enabling remote streaming!  -->
-    <requestParsers enableRemoteStreaming="true" multipartUploadLimitInKB="2048000" />
-
-    <!-- Set HTTP caching related parameters (for proxy caches and clients).
-          
-         To get the behaviour of Solr 1.2 (ie: no caching related headers)
-         use the never304="true" option and do not specify a value for
-         <cacheControl>
-    -->
-    <!-- <httpCaching never304="true"> -->
-    <httpCaching lastModifiedFrom="openTime"
-                 etagSeed="Solr">
-       <!-- lastModFrom="openTime" is the default, the Last-Modified value
-            (and validation against If-Modified-Since requests) will all be
-            relative to when the current Searcher was opened.
-            You can change it to lastModFrom="dirLastMod" if you want the
-            value to exactly corrispond to when the physical index was last
-            modified.
-
-            etagSeed="..." is an option you can change to force the ETag
-            header (and validation against If-None-Match requests) to be
-            differnet even if the index has not changed (ie: when making
-            significant changes to your config file)
-
-            lastModifiedFrom and etagSeed are both ignored if you use the
-            never304="true" option.
-       -->
-       <!-- If you include a <cacheControl> directive, it will be used to
-            generate a Cache-Control header, as well as an Expires header
-            if the value contains "max-age="
-
-            By default, no Cache-Control header is generated.
-
-            You can use the <cacheControl> option even if you have set
-            never304="true"
-       -->
-       <!-- <cacheControl>max-age=30, public</cacheControl> -->
-    </httpCaching>
-  </requestDispatcher>
-
-
-  <!-- requestHandler plugins... incoming queries will be dispatched to the
-     correct handler based on the path or the qt (query type) param.
-     Names starting with a '/' are accessed with the a path equal to the 
-     registered name.  Names without a leading '/' are accessed with:
-      http://host/app/select?qt=name
-     If no qt is defined, the requestHandler that declares default="true"
-     will be used.
-  -->
-  <requestHandler name="standard" class="solr.SearchHandler" default="true">
-    <!-- default values for query parameters -->
-     <lst name="defaults">
-       <str name="echoParams">explicit</str>
-       <!--
-       <int name="rows">10</int>
-       <str name="fl">*</str>
-       <str name="version">2.1</str>
-        -->
-     </lst>
-  </requestHandler>
-
-<!-- Please refer to http://wiki.apache.org/solr/SolrReplication for details on configuring replication -->
-<!-- remove the <lst name="master"> section if this is just a slave -->
-<!-- remove  the <lst name="slave"> section if this is just a master -->
-<!--
-<requestHandler name="/replication" class="solr.ReplicationHandler" >
-    <lst name="master">
-      <str name="replicateAfter">commit</str>
-      <str name="replicateAfter">startup</str>
-      <str name="confFiles">schema.xml,stopwords.txt</str>
-    </lst>
-    <lst name="slave">
-      <str name="masterUrl">http://localhost:8983/solr/replication</str>
-      <str name="pollInterval">00:00:60</str>
-    </lst>
-</requestHandler>-->
-
-  <!-- DisMaxRequestHandler allows easy searching across multiple fields
-       for simple user-entered phrases.  It's implementation is now
-       just the standard SearchHandler with a default query type
-       of "dismax". 
-       see http://wiki.apache.org/solr/DisMaxRequestHandler
-   -->
-  <requestHandler name="dismax" class="solr.SearchHandler" >
-    <lst name="defaults">
-     <str name="defType">dismax</str>
-     <str name="echoParams">explicit</str>
-     <float name="tie">0.01</float>
-     <str name="qf">
-        text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4
-     </str>
-     <str name="pf">
-        text^0.2 features^1.1 name^1.5 manu^1.4 manu_exact^1.9
-     </str>
-     <str name="bf">
-        popularity^0.5 recip(price,1,1000,1000)^0.3
-     </str>
-     <str name="fl">
-        id,name,price,score
-     </str>
-     <str name="mm">
-        2&lt;-1 5&lt;-2 6&lt;90%
-     </str>
-     <int name="ps">100</int>
-     <str name="q.alt">*:*</str>
-     <!-- example highlighter config, enable per-query with hl=true -->
-     <str name="hl.fl">text features name</str>
-     <!-- for this field, we want no fragmenting, just highlighting -->
-     <str name="f.name.hl.fragsize">0</str>
-     <!-- instructs Solr to return the field itself if no query terms are
-          found -->
-     <str name="f.name.hl.alternateField">name</str>
-     <str name="f.text.hl.fragmenter">regex</str> <!-- defined below -->
-    </lst>
-  </requestHandler>
-
-  <!-- Note how you can register the same handler multiple times with
-       different names (and different init parameters)
-    -->
-  <requestHandler name="partitioned" class="solr.SearchHandler" >
-    <lst name="defaults">
-     <str name="defType">dismax</str>
-     <str name="echoParams">explicit</str>
-     <str name="qf">text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0</str>
-     <str name="mm">2&lt;-1 5&lt;-2 6&lt;90%</str>
-     <!-- This is an example of using Date Math to specify a constantly
-          moving date range in a config...
-       -->
-     <str name="bq">incubationdate_dt:[* TO NOW/DAY-1MONTH]^2.2</str>
-    </lst>
-    <!-- In addition to defaults, "appends" params can be specified
-         to identify values which should be appended to the list of
-         multi-val params from the query (or the existing "defaults").
-
-         In this example, the param "fq=instock:true" will be appended to
-         any query time fq params the user may specify, as a mechanism for
-         partitioning the index, independent of any user selected filtering
-         that may also be desired (perhaps as a result of faceted searching).
-
-         NOTE: there is *absolutely* nothing a client can do to prevent these
-         "appends" values from being used, so don't use this mechanism
-         unless you are sure you always want it.
-      -->
-    <lst name="appends">
-      <str name="fq">inStock:true</str>
-    </lst>
-    <!-- "invariants" are a way of letting the Solr maintainer lock down
-         the options available to Solr clients.  Any params values
-         specified here are used regardless of what values may be specified
-         in either the query, the "defaults", or the "appends" params.
-
-         In this example, the facet.field and facet.query params are fixed,
-         limiting the facets clients can use.  Faceting is not turned on by
-         default - but if the client does specify facet=true in the request,
-         these are the only facets they will be able to see counts for;
-         regardless of what other facet.field or facet.query params they
-         may specify.
-
-         NOTE: there is *absolutely* nothing a client can do to prevent these
-         "invariants" values from being used, so don't use this mechanism
-         unless you are sure you always want it.
-      -->
-    <lst name="invariants">
-      <str name="facet.field">cat</str>
-      <str name="facet.field">manu_exact</str>
-      <str name="facet.query">price:[* TO 500]</str>
-      <str name="facet.query">price:[500 TO *]</str>
-    </lst>
-  </requestHandler>
-
-
-  <!--
-   Search components are registered to SolrCore and used by Search Handlers
-   
-   By default, the following components are avaliable:
-    
-   <searchComponent name="query"     class="org.apache.solr.handler.component.QueryComponent" />
-   <searchComponent name="facet"     class="org.apache.solr.handler.component.FacetComponent" />
-   <searchComponent name="mlt"       class="org.apache.solr.handler.component.MoreLikeThisComponent" />
-   <searchComponent name="highlight" class="org.apache.solr.handler.component.HighlightComponent" />
-   <searchComponent name="stats"     class="org.apache.solr.handler.component.StatsComponent" />
-   <searchComponent name="debug"     class="org.apache.solr.handler.component.DebugComponent" />
-   
-   Default configuration in a requestHandler would look like:
-    <arr name="components">
-      <str>query</str>
-      <str>facet</str>
-      <str>mlt</str>
-      <str>highlight</str>
-      <str>stats</str>
-      <str>debug</str>
-    </arr>
-
-    If you register a searchComponent to one of the standard names, that will be used instead.
-    To insert components before or after the 'standard' components, use:
-    
-    <arr name="first-components">
-      <str>myFirstComponentName</str>
-    </arr>
-    
-    <arr name="last-components">
-      <str>myLastComponentName</str>
-    </arr>
-  -->
-
-   <!-- The spell check component can return a list of alternative spelling
-  suggestions.  -->
-  <searchComponent name="spellcheck" class="solr.SpellCheckComponent">
-
-    <str name="queryAnalyzerFieldType">textSpell</str>
-
-    <lst name="spellchecker">
-      <str name="name">default</str>
-      <str name="field">name</str>
-      <str name="spellcheckIndexDir">./spellchecker</str>
-    </lst>
-
-    <!-- a spellchecker that uses a different distance measure
-    <lst name="spellchecker">
-      <str name="name">jarowinkler</str>
-      <str name="field">spell</str>
-      <str name="distanceMeasure">org.apache.lucene.search.spell.JaroWinklerDistance</str>
-      <str name="spellcheckIndexDir">./spellchecker2</str>
-    </lst>
-     -->
-
-    <!-- a file based spell checker
-    <lst name="spellchecker">
-      <str name="classname">solr.FileBasedSpellChecker</str>
-      <str name="name">file</str>
-      <str name="sourceLocation">spellings.txt</str>
-      <str name="characterEncoding">UTF-8</str>
-      <str name="spellcheckIndexDir">./spellcheckerFile</str>
-    </lst>
-    -->
-  </searchComponent>
-
-  <!-- A request handler utilizing the spellcheck component.  
-  #############################################################################
-  NOTE: This is purely as an example.  The whole purpose of the
-  SpellCheckComponent is to hook it into the request handler that handles (i.e.
-  the standard or dismax SearchHandler) queries such that a separate request is
-  not needed to get suggestions.
-
-  IN OTHER WORDS, THERE IS REALLY GOOD CHANCE THE SETUP BELOW IS NOT WHAT YOU
-  WANT FOR YOUR PRODUCTION SYSTEM!
-  #############################################################################
-  -->
-  <requestHandler name="/spell" class="solr.SearchHandler" lazy="true">
-    <lst name="defaults">
-      <!-- omp = Only More Popular -->
-      <str name="spellcheck.onlyMorePopular">false</str>
-      <!-- exr = Extended Results -->
-      <str name="spellcheck.extendedResults">false</str>
-      <!--  The number of suggestions to return -->
-      <str name="spellcheck.count">1</str>
-    </lst>
-    <arr name="last-components">
-      <str>spellcheck</str>
-    </arr>
-  </requestHandler>
-
-  <searchComponent name="tvComponent" class="org.apache.solr.handler.component.TermVectorComponent"/>
-  <!-- A Req Handler for working with the tvComponent.  This is purely as an example.
-  You will likely want to add the component to your already specified request handlers. -->
-  <requestHandler name="tvrh" class="org.apache.solr.handler.component.SearchHandler">
-    <lst name="defaults">
-      <bool name="tv">true</bool>
-    </lst>
-    <arr name="last-components">
-      <str>tvComponent</str>
-    </arr>
-  </requestHandler>
-
-  <!-- Clustering Component
-       http://wiki.apache.org/solr/ClusteringComponent
-       This relies on third party jars which are not included in the release.
-       To use this component (and the "/clustering" handler)
-       Those jars will need to be downloaded, and you'll need to set the
-       solr.cluster.enabled system property when running solr...
-          java -Dsolr.clustering.enabled=true -jar start.jar
-    -->
-  <searchComponent
-    name="clusteringComponent"
-    enable="${solr.clustering.enabled:false}"
-    class="org.apache.solr.handler.clustering.ClusteringComponent" >
-    <!-- Declare an engine -->
-    <lst name="engine">
-      <!-- The name, only one can be named "default" -->
-      <str name="name">default</str>
-      <!-- 
-           Class name of Carrot2 clustering algorithm. Currently available algorithms are:
-           
-           * org.carrot2.clustering.lingo.LingoClusteringAlgorithm
-           * org.carrot2.clustering.stc.STCClusteringAlgorithm
-           
-           See http://project.carrot2.org/algorithms.html for the algorithm's characteristics.
-        -->
-      <str name="carrot.algorithm">org.carrot2.clustering.lingo.LingoClusteringAlgorithm</str>
-      <!-- 
-           Overriding values for Carrot2 default algorithm attributes. For a description
-           of all available attributes, see: http://download.carrot2.org/stable/manual/#chapter.components.
-           Use attribute key as name attribute of str elements below. These can be further
-           overridden for individual requests by specifying attribute key as request
-           parameter name and attribute value as parameter value.
-        -->
-      <str name="LingoClusteringAlgorithm.desiredClusterCountBase">20</str>
-    </lst>
-    <lst name="engine">
-      <str name="name">stc</str>
-      <str name="carrot.algorithm">org.carrot2.clustering.stc.STCClusteringAlgorithm</str>
-    </lst>
-  </searchComponent>
-  <requestHandler name="/clustering"
-                  enable="${solr.clustering.enabled:false}"
-                  class="solr.SearchHandler">
-     <lst name="defaults">
-       <bool name="clustering">true</bool>
-       <str name="clustering.engine">default</str>
-       <bool name="clustering.results">true</bool>
-       <!-- The title field -->
-       <str name="carrot.title">name</str>
-       <str name="carrot.url">id</str>
-       <!-- The field to cluster on -->
-       <str name="carrot.snippet">features</str>
-       <!-- produce summaries -->
-       <bool name="carrot.produceSummary">true</bool>
-       <!-- the maximum number of labels per cluster -->
-       <!--<int name="carrot.numDescriptions">5</int>-->
-       <!-- produce sub clusters -->
-       <bool name="carrot.outputSubClusters">false</bool>
-    </lst>     
-    <arr name="last-components">
-      <str>clusteringComponent</str>
-    </arr>
-  </requestHandler>
-  
-  <!-- Solr Cell: http://wiki.apache.org/solr/ExtractingRequestHandler -->
-  <requestHandler name="/update/extract" class="org.apache.solr.handler.extraction.ExtractingRequestHandler" startup="lazy">
-    <lst name="defaults">
-      <!-- All the main content goes into "text"... if you need to return
-           the extracted text or do highlighting, use a stored field. -->
-      <str name="fmap.content">text</str>
-      <str name="lowernames">true</str>
-      <str name="uprefix">ignored_</str>
-
-      <!-- capture link hrefs but ignore div attributes -->
-      <str name="captureAttr">true</str>
-      <str name="fmap.a">links</str>
-      <str name="fmap.div">ignored_</str>
-    </lst>
-  </requestHandler>
-
-
-  <!-- A component to return terms and document frequency of those terms.
-       This component does not yet support distributed search. -->
-  <searchComponent name="termsComponent" class="org.apache.solr.handler.component.TermsComponent"/>
-
-  <requestHandler name="/terms" class="org.apache.solr.handler.component.SearchHandler">
-     <lst name="defaults">
-      <bool name="terms">true</bool>
-    </lst>     
-    <arr name="components">
-      <str>termsComponent</str>
-    </arr>
-  </requestHandler>
-
-
-  <!-- a search component that enables you to configure the top results for
-       a given query regardless of the normal lucene scoring.-->
-  <searchComponent name="elevator" class="solr.QueryElevationComponent" >
-    <!-- pick a fieldType to analyze queries -->
-    <str name="queryFieldType">string</str>
-    <str name="config-file">elevate.xml</str>
-  </searchComponent>
-
-  <!-- a request handler utilizing the elevator component -->
-  <requestHandler name="/elevate" class="solr.SearchHandler" startup="lazy">
-    <lst name="defaults">
-      <str name="echoParams">explicit</str>
-    </lst>
-    <arr name="last-components">
-      <str>elevator</str>
-    </arr>
-  </requestHandler>
-
-
-  <!-- Update request handler.  
-
-       Note: Since solr1.1 requestHandlers requires a valid content type header if posted in
-       the body. For example, curl now requires: -H 'Content-type:text/xml; charset=utf-8'
-       The response format differs from solr1.1 formatting and returns a standard error code.
-       To enable solr1.1 behavior, remove the /update handler or change its path
-    -->
-  <requestHandler name="/update" class="solr.XmlUpdateRequestHandler" />
-
-
-  <requestHandler name="/update/javabin" class="solr.BinaryUpdateRequestHandler" />
-
-  <!--
-   Analysis request handler.  Since Solr 1.3.  Use to return how a document is analyzed.  Useful
-   for debugging and as a token server for other types of applications.
-
-   This is deprecated in favor of the improved DocumentAnalysisRequestHandler and FieldAnalysisRequestHandler
-
-   <requestHandler name="/analysis" class="solr.AnalysisRequestHandler" />
-   -->
-
-  <!--
-    An analysis handler that provides a breakdown of the analysis process of provided docuemnts. This handler expects a
-    (single) content stream with the following format:
-
-    <docs>
-      <doc>
-        <field name="id">1</field>
-        <field name="name">The Name</field>
-        <field name="text">The Text Value</field>
-      <doc>
-      <doc>...</doc>
-      <doc>...</doc>
-      ...
-    </docs>
-
-    Note: Each document must contain a field which serves as the unique key. This key is used in the returned
-    response to assoicate an analysis breakdown to the analyzed document.
-
-    Like the FieldAnalysisRequestHandler, this handler also supports query analysis by
-    sending either an "analysis.query" or "q" request paraemter that holds the query text to be analyized. It also
-    supports the "analysis.showmatch" parameter which when set to true, all field tokens that match the query
-    tokens will be marked as a "match".
-  -->
-  <requestHandler name="/analysis/document" class="solr.DocumentAnalysisRequestHandler" />
-
-  <!--
-    RequestHandler that provides much the same functionality as analysis.jsp. Provides the ability
-    to specify multiple field types and field names in the same request and outputs index-time and
-    query-time analysis for each of them.
-
-    Request parameters are:
-    analysis.fieldname - The field name whose analyzers are to be used
-    analysis.fieldtype - The field type whose analyzers are to be used
-    analysis.fieldvalue - The text for index-time analysis
-    q (or analysis.q) - The text for query time analysis
-    analysis.showmatch (true|false) - When set to true and when query analysis is performed, the produced
-                                      tokens of the field value analysis will be marked as "matched" for every
-                                      token that is produces by the query analysis
-   -->
-  <requestHandler name="/analysis/field" class="solr.FieldAnalysisRequestHandler" />
-
-
-  <!-- CSV update handler, loaded on demand -->
-  <requestHandler name="/update/csv" class="solr.CSVRequestHandler" startup="lazy" />
-
-
-  <!-- 
-   Admin Handlers - This will register all the standard admin RequestHandlers.  Adding 
-   this single handler is equivalent to registering:
-   
-  <requestHandler name="/admin/luke"       class="org.apache.solr.handler.admin.LukeRequestHandler" />
-  <requestHandler name="/admin/system"     class="org.apache.solr.handler.admin.SystemInfoHandler" />
-  <requestHandler name="/admin/plugins"    class="org.apache.solr.handler.admin.PluginInfoHandler" />
-  <requestHandler name="/admin/threads"    class="org.apache.solr.handler.admin.ThreadDumpHandler" />
-  <requestHandler name="/admin/properties" class="org.apache.solr.handler.admin.PropertiesRequestHandler" />
-  <requestHandler name="/admin/file"       class="org.apache.solr.handler.admin.ShowFileRequestHandler" >
-  
-  If you wish to hide files under ${solr.home}/conf, explicitly register the ShowFileRequestHandler using:
-  <requestHandler name="/admin/file" class="org.apache.solr.handler.admin.ShowFileRequestHandler" >
-    <lst name="invariants">
-     <str name="hidden">synonyms.txt</str> 
-     <str name="hidden">anotherfile.txt</str> 
-    </lst>
-  </requestHandler>
-  -->
-  <requestHandler name="/admin/" class="org.apache.solr.handler.admin.AdminHandlers" />
-
-  <!-- ping/healthcheck -->
-  <requestHandler name="/admin/ping" class="PingRequestHandler">
-    <lst name="defaults">
-      <str name="qt">standard</str>
-      <str name="q">solrpingquery</str>
-      <str name="echoParams">all</str>
-    </lst>
-  </requestHandler>
-
-  <!-- Echo the request contents back to the client -->
-  <requestHandler name="/debug/dump" class="solr.DumpRequestHandler" >
-    <lst name="defaults">
-     <str name="echoParams">explicit</str> <!-- for all params (including the default etc) use: 'all' -->
-     <str name="echoHandler">true</str>
-    </lst>
-  </requestHandler>
-
-  <highlighting>
-   <!-- Configure the standard fragmenter -->
-   <!-- This could most likely be commented out in the "default" case -->
-   <fragmenter name="gap" class="org.apache.solr.highlight.GapFragmenter" default="true">
-    <lst name="defaults">
-     <int name="hl.fragsize">100</int>
-    </lst>
-   </fragmenter>
-
-   <!-- A regular-expression-based fragmenter (f.i., for sentence extraction) -->
-   <fragmenter name="regex" class="org.apache.solr.highlight.RegexFragmenter">
-    <lst name="defaults">
-      <!-- slightly smaller fragsizes work better because of slop -->
-      <int name="hl.fragsize">70</int>
-      <!-- allow 50% slop on fragment sizes -->
-      <float name="hl.regex.slop">0.5</float>
-      <!-- a basic sentence pattern -->
-      <str name="hl.regex.pattern">[-\w ,/\n\"']{20,200}</str>
-    </lst>
-   </fragmenter>
-
-   <!-- Configure the standard formatter -->
-   <formatter name="html" class="org.apache.solr.highlight.HtmlFormatter" default="true">
-    <lst name="defaults">
-     <str name="hl.simple.pre"><![CDATA[<em>]]></str>
-     <str name="hl.simple.post"><![CDATA[</em>]]></str>
-    </lst>
-   </formatter>
-  </highlighting>
-
-  <!-- An example dedup update processor that creates the "id" field on the fly
-       based on the hash code of some other fields.  This example has overwriteDupes
-       set to false since we are using the id field as the signatureField and Solr
-       will maintain uniqueness based on that anyway. 
-       
-       You have to link the chain to an update handler above to use it ie:
-         <requestHandler name="/update "class="solr.XmlUpdateRequestHandler">
-           <lst name="defaults">
-             <str name="update.processor">dedupe</str>
-           </lst>
-         </requestHandler>  
-  -->
-  <!--
-  <updateRequestProcessorChain name="dedupe">
-    <processor class="org.apache.solr.update.processor.SignatureUpdateProcessorFactory">
-      <bool name="enabled">true</bool>
-      <str name="signatureField">id</str>
-      <bool name="overwriteDupes">false</bool>
-      <str name="fields">name,features,cat</str>
-      <str name="signatureClass">org.apache.solr.update.processor.Lookup3Signature</str>
-    </processor>
-    <processor class="solr.LogUpdateProcessorFactory" />
-    <processor class="solr.RunUpdateProcessorFactory" />
-  </updateRequestProcessorChain>
-  -->
-
-
-  <!-- queryResponseWriter plugins... query responses will be written using the
-    writer specified by the 'wt' request parameter matching the name of a registered
-    writer.
-    The "default" writer is the default and will be used if 'wt' is not specified 
-    in the request. XMLResponseWriter will be used if nothing is specified here.
-    The json, python, and ruby writers are also available by default.
-
-    <queryResponseWriter name="xml" class="org.apache.solr.request.XMLResponseWriter" default="true"/>
-    <queryResponseWriter name="json" class="org.apache.solr.request.JSONResponseWriter"/>
-    <queryResponseWriter name="python" class="org.apache.solr.request.PythonResponseWriter"/>
-    <queryResponseWriter name="ruby" class="org.apache.solr.request.RubyResponseWriter"/>
-    <queryResponseWriter name="php" class="org.apache.solr.request.PHPResponseWriter"/>
-    <queryResponseWriter name="phps" class="org.apache.solr.request.PHPSerializedResponseWriter"/>
-
-    <queryResponseWriter name="custom" class="com.example.MyResponseWriter"/>
-  -->
-
-  <!-- XSLT response writer transforms the XML output by any xslt file found
-       in Solr's conf/xslt directory.  Changes to xslt files are checked for
-       every xsltCacheLifetimeSeconds.  
-   -->
-  <queryResponseWriter name="xslt" class="org.apache.solr.request.XSLTResponseWriter">
-    <int name="xsltCacheLifetimeSeconds">5</int>
-  </queryResponseWriter>
-
-
-  <!-- example of registering a query parser
-  <queryParser name="lucene" class="org.apache.solr.search.LuceneQParserPlugin"/>
-  -->
-
-  <!-- example of registering a custom function parser 
-  <valueSourceParser name="myfunc" class="com.mycompany.MyValueSourceParser" />
-  -->
-
-  <!-- config for the admin interface -->
-  <admin>
-    <defaultQuery>solr</defaultQuery>
-
-    <!-- configure a healthcheck file for servers behind a loadbalancer
-    <healthcheck type="file">server-enabled</healthcheck>
-    -->
-  </admin>
-
-</config>

+ 0 - 57
solr/conf/stopwords.txt

@@ -1,57 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-#-----------------------------------------------------------------------
-# a couple of test stopwords to test that the words are really being
-# configured from this file:
-stopworda
-stopwordb
-
-#Standard english stop words taken from Lucene's StopAnalyzer
-an
-and
-are
-as
-at
-be
-but
-by
-for
-if
-in
-into
-is
-it
-no
-not
-of
-on
-or
-s
-such
-t
-that
-the
-their
-then
-there
-these
-they
-this
-to
-was
-will
-with
-

+ 0 - 31
solr/conf/synonyms.txt

@@ -1,31 +0,0 @@
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-#-----------------------------------------------------------------------
-#some test synonym mappings unlikely to appear in real input text
-aaa => aaaa
-bbb => bbbb1 bbbb2
-ccc => cccc1,cccc2
-a\=>a => b\=>b
-a\,a => b\,b
-fooaaa,baraaa,bazaaa
-
-# Some synonym groups specific to this example
-GB,gib,gigabyte,gigabytes
-MB,mib,megabyte,megabytes
-Television, Televisions, TV, TVs
-#notice we use "gib" instead of "GiB" so any WordDelimiterFilter coming
-#after us won't split it into two words.
-
-# Synonym mappings can be used for spelling correction too
-pixima => pixma
-

+ 0 - 572
src/main/config/h2/fess.ddl

@@ -1,572 +0,0 @@
-DROP TABLE IF EXISTS FAVORITE_LOG;
-DROP TABLE IF EXISTS SEARCH_FIELD_LOG;
-DROP TABLE IF EXISTS FILE_AUTHENTICATION;
-DROP TABLE IF EXISTS FAILURE_URL;
-DROP TABLE IF EXISTS CLICK_LOG;
-DROP TABLE IF EXISTS LABEL_TYPE_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS SEARCH_LOG;
-DROP TABLE IF EXISTS USER_INFO;
-DROP TABLE IF EXISTS DATA_CONFIG_TO_BROWSER_TYPE_MAPPING;
-DROP TABLE IF EXISTS DATA_CONFIG_TO_LABEL_TYPE_MAPPING;
-DROP TABLE IF EXISTS DATA_CONFIG_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS DATA_CRAWLING_CONFIG;
-DROP TABLE IF EXISTS WEB_CONFIG_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS FILE_CONFIG_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS ROLE_TYPE;
-DROP TABLE IF EXISTS WEB_CONFIG_TO_LABEL_TYPE_MAPPING;
-DROP TABLE IF EXISTS FILE_CONFIG_TO_LABEL_TYPE_MAPPING;
-DROP TABLE IF EXISTS LABEL_TYPE;
-DROP TABLE IF EXISTS CRAWLING_SESSION_INFO;
-DROP TABLE IF EXISTS WEB_AUTHENTICATION;
-DROP TABLE IF EXISTS KEY_MATCH;
-DROP TABLE IF EXISTS BOOST_DOCUMENT_RULE;
-DROP TABLE IF EXISTS REQUEST_HEADER;
-DROP TABLE IF EXISTS OVERLAPPING_HOST;
-DROP TABLE IF EXISTS CRAWLING_SESSION;
-DROP TABLE IF EXISTS PATH_MAPPING;
-DROP TABLE IF EXISTS JOB_LOG;
-DROP TABLE IF EXISTS SCHEDULED_JOB;
-DROP TABLE IF EXISTS FILE_CONFIG_TO_BROWSER_TYPE_MAPPING;
-DROP TABLE IF EXISTS WEB_CONFIG_TO_BROWSER_TYPE_MAPPING;
-DROP TABLE IF EXISTS FILE_CRAWLING_CONFIG;
-DROP TABLE IF EXISTS BROWSER_TYPE;
-DROP TABLE IF EXISTS WEB_CRAWLING_CONFIG;
-DROP TABLE IF EXISTS SUGGEST_BAD_WORD;
-DROP TABLE IF EXISTS SUGGEST_ELAVATE_WORD;
-
-/**********************************/
-/* Table Name: Web Crawling Config */
-/**********************************/
-CREATE TABLE WEB_CRAWLING_CONFIG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  NAME VARCHAR(200) NOT NULL,
-  URLS VARCHAR(4000) NOT NULL,
-  INCLUDED_URLS VARCHAR(4000),
-  EXCLUDED_URLS VARCHAR(4000),
-  INCLUDED_DOC_URLS VARCHAR(4000),
-  EXCLUDED_DOC_URLS VARCHAR(4000),
-  CONFIG_PARAMETER VARCHAR(4000),
-  DEPTH INTEGER,
-  MAX_ACCESS_COUNT BIGINT,
-  USER_AGENT VARCHAR(200) NOT NULL,
-  NUM_OF_THREAD INTEGER NOT NULL,
-  INTERVAL_TIME INTEGER NOT NULL,
-  BOOST FLOAT NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: File Crawling Config */
-/**********************************/
-CREATE TABLE FILE_CRAWLING_CONFIG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  NAME VARCHAR(200) NOT NULL,
-  PATHS VARCHAR(4000) NOT NULL,
-  INCLUDED_PATHS VARCHAR(4000),
-  EXCLUDED_PATHS VARCHAR(4000),
-  INCLUDED_DOC_PATHS VARCHAR(4000),
-  EXCLUDED_DOC_PATHS VARCHAR(4000),
-  CONFIG_PARAMETER VARCHAR(4000),
-  DEPTH INTEGER,
-  MAX_ACCESS_COUNT BIGINT,
-  NUM_OF_THREAD INTEGER NOT NULL,
-  INTERVAL_TIME INTEGER NOT NULL,
-  BOOST FLOAT NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Scheduled Job */
-/**********************************/
-CREATE TABLE SCHEDULED_JOB(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  NAME VARCHAR(100) NOT NULL,
-  TARGET VARCHAR(100) NOT NULL,
-  CRON_EXPRESSION VARCHAR(100) NOT NULL,
-  SCRIPT_TYPE VARCHAR(100) NOT NULL,
-  SCRIPT_DATA VARCHAR(4000),
-  CRAWLER VARCHAR(1) NOT NULL,
-  JOB_LOGGING VARCHAR(1) NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Job Log */
-/**********************************/
-CREATE TABLE JOB_LOG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  JOB_NAME VARCHAR(100) NOT NULL,
-  JOB_STATUS VARCHAR(10) NOT NULL,
-  TARGET VARCHAR(100) NOT NULL,
-  SCRIPT_TYPE VARCHAR(100) NOT NULL,
-  SCRIPT_DATA VARCHAR(4000),
-  SCRIPT_RESULT VARCHAR(4000),
-  START_TIME TIMESTAMP NOT NULL,
-  END_TIME TIMESTAMP
-);
-
-/**********************************/
-/* Table Name: Path Mapping */
-/**********************************/
-CREATE TABLE PATH_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  REGEX VARCHAR(1000) NOT NULL,
-  REPLACEMENT VARCHAR(1000) NOT NULL,
-  PROCESS_TYPE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Crawling Session */
-/**********************************/
-CREATE TABLE CRAWLING_SESSION(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SESSION_ID VARCHAR(20) NOT NULL,
-  NAME VARCHAR(20),
-  EXPIRED_TIME TIMESTAMP,
-  CREATED_TIME TIMESTAMP NOT NULL
-);
-
-/**********************************/
-/* Table Name: Overlapping Host */
-/**********************************/
-CREATE TABLE OVERLAPPING_HOST(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  REGULAR_NAME VARCHAR(1000) NOT NULL,
-  OVERLAPPING_NAME VARCHAR(1000) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Request Header */
-/**********************************/
-CREATE TABLE REQUEST_HEADER(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  NAME VARCHAR(100) NOT NULL,
-  VALUE VARCHAR(1000) NOT NULL,
-  WEB_CRAWLING_CONFIG_ID BIGINT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL,
-  FOREIGN KEY (WEB_CRAWLING_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Boost Document Rule */
-/**********************************/
-CREATE TABLE BOOST_DOCUMENT_RULE(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  URL_EXPR VARCHAR(4000) NOT NULL,
-  BOOST_EXPR VARCHAR(4000) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Key Match */
-/**********************************/
-CREATE TABLE KEY_MATCH(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  TERM VARCHAR(200) NOT NULL,
-  QUERY VARCHAR(4000) NOT NULL,
-  MAX_SIZE INTEGER NOT NULL,
-  BOOST FLOAT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Web Authentication */
-/**********************************/
-CREATE TABLE WEB_AUTHENTICATION(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  HOSTNAME VARCHAR(100),
-  PORT INTEGER NOT NULL,
-  AUTH_REALM VARCHAR(100),
-  PROTOCOL_SCHEME VARCHAR(10),
-  USERNAME VARCHAR(100) NOT NULL,
-  PASSWORD VARCHAR(100),
-  PARAMETERS VARCHAR(1000),
-  WEB_CRAWLING_CONFIG_ID BIGINT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL,
-  FOREIGN KEY (WEB_CRAWLING_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Crawling Session Info */
-/**********************************/
-CREATE TABLE CRAWLING_SESSION_INFO(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  CRAWLING_SESSION_ID BIGINT NOT NULL,
-  KEY VARCHAR(20) NOT NULL,
-  VALUE VARCHAR(100) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  FOREIGN KEY (CRAWLING_SESSION_ID) REFERENCES CRAWLING_SESSION (ID)
-);
-
-/**********************************/
-/* Table Name: Label Type */
-/**********************************/
-CREATE TABLE LABEL_TYPE(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  NAME VARCHAR(100) NOT NULL,
-  VALUE VARCHAR(20) NOT NULL,
-  INCLUDED_PATHS VARCHAR(4000),
-  EXCLUDED_PATHS VARCHAR(4000),
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: File To Label Mapping */
-/**********************************/
-CREATE TABLE FILE_CONFIG_TO_LABEL_TYPE_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  FILE_CONFIG_ID BIGINT NOT NULL,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (FILE_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Web To Label Mapping */
-/**********************************/
-CREATE TABLE WEB_CONFIG_TO_LABEL_TYPE_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  WEB_CONFIG_ID BIGINT NOT NULL,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (WEB_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Role Type */
-/**********************************/
-CREATE TABLE ROLE_TYPE(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  NAME VARCHAR(100) NOT NULL,
-  VALUE VARCHAR(20) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: File To Role Mapping */
-/**********************************/
-CREATE TABLE FILE_CONFIG_TO_ROLE_TYPE_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  FILE_CONFIG_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (FILE_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Web To Role Mapping */
-/**********************************/
-CREATE TABLE WEB_CONFIG_TO_ROLE_TYPE_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  WEB_CONFIG_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (WEB_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Data Crawling Config */
-/**********************************/
-CREATE TABLE DATA_CRAWLING_CONFIG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  NAME VARCHAR(200) NOT NULL,
-  HANDLER_NAME VARCHAR(200) NOT NULL,
-  HANDLER_PARAMETER VARCHAR(4000),
-  HANDLER_SCRIPT VARCHAR(4000),
-  BOOST FLOAT NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Data To Role Mapping */
-/**********************************/
-CREATE TABLE DATA_CONFIG_TO_ROLE_TYPE_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  DATA_CONFIG_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (DATA_CONFIG_ID) REFERENCES DATA_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Data To Label Mapping */
-/**********************************/
-CREATE TABLE DATA_CONFIG_TO_LABEL_TYPE_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  DATA_CONFIG_ID BIGINT NOT NULL,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (DATA_CONFIG_ID) REFERENCES DATA_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: User Info */
-/**********************************/
-CREATE TABLE USER_INFO(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  CODE VARCHAR(1000) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_TIME TIMESTAMP NOT NULL
-);
-
-/**********************************/
-/* Table Name: Search Log */
-/**********************************/
-CREATE TABLE SEARCH_LOG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SEARCH_WORD VARCHAR(1000),
-  REQUESTED_TIME TIMESTAMP NOT NULL,
-  RESPONSE_TIME INTEGER NOT NULL,
-  HIT_COUNT BIGINT NOT NULL,
-  QUERY_OFFSET INTEGER NOT NULL,
-  QUERY_PAGE_SIZE INTEGER NOT NULL,
-  USER_AGENT VARCHAR(255),
-  REFERER VARCHAR(1000),
-  CLIENT_IP VARCHAR(50),
-  USER_SESSION_ID VARCHAR(100),
-  ACCESS_TYPE VARCHAR(1) NOT NULL,
-  USER_ID BIGINT,
-  FOREIGN KEY (USER_ID) REFERENCES USER_INFO (ID)
-);
-
-/**********************************/
-/* Table Name: Label Type To Role Type Mapping */
-/**********************************/
-CREATE TABLE LABEL_TYPE_TO_ROLE_TYPE_MAPPING(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Click Log */
-/**********************************/
-CREATE TABLE CLICK_LOG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SEARCH_ID BIGINT NOT NULL,
-  URL VARCHAR(4000) NOT NULL,
-  REQUESTED_TIME TIMESTAMP NOT NULL,
-  FOREIGN KEY (SEARCH_ID) REFERENCES SEARCH_LOG (ID)
-);
-
-/**********************************/
-/* Table Name: Failure Url */
-/**********************************/
-CREATE TABLE FAILURE_URL(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  URL VARCHAR(4000) NOT NULL,
-  THREAD_NAME VARCHAR(30) NOT NULL,
-  ERROR_NAME VARCHAR(255),
-  ERROR_LOG VARCHAR(4000),
-  ERROR_COUNT INTEGER NOT NULL,
-  LAST_ACCESS_TIME TIMESTAMP NOT NULL,
-  CONFIG_ID VARCHAR(100)
-);
-
-/**********************************/
-/* Table Name: File Authentication */
-/**********************************/
-CREATE TABLE FILE_AUTHENTICATION(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  HOSTNAME VARCHAR(255),
-  PORT INTEGER NOT NULL,
-  PROTOCOL_SCHEME VARCHAR(10),
-  USERNAME VARCHAR(100) NOT NULL,
-  PASSWORD VARCHAR(100),
-  PARAMETERS VARCHAR(1000),
-  FILE_CRAWLING_CONFIG_ID BIGINT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL,
-  FOREIGN KEY (FILE_CRAWLING_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Search Field Log */
-/**********************************/
-CREATE TABLE SEARCH_FIELD_LOG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SEARCH_ID BIGINT NOT NULL,
-  NAME VARCHAR(255) NOT NULL,
-  VALUE VARCHAR(1000) NOT NULL,
-  FOREIGN KEY (SEARCH_ID) REFERENCES SEARCH_LOG (ID)
-);
-
-/**********************************/
-/* Table Name: Favorite Log */
-/**********************************/
-CREATE TABLE FAVORITE_LOG(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  USER_ID BIGINT NOT NULL,
-  URL VARCHAR(4000) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  FOREIGN KEY (USER_ID) REFERENCES USER_INFO (ID)
-);
-
-/**********************************/
-/* Table Name: Suggest Ng Word */
-/**********************************/
-CREATE TABLE SUGGEST_BAD_WORD(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SUGGEST_WORD VARCHAR(255) NOT NULL,
-  TARGET_ROLE VARCHAR(255),
-  TARGET_LABEL VARCHAR(255),
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Suggest Elevate word */
-/**********************************/
-CREATE TABLE SUGGEST_ELEVATE_WORD(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SUGGEST_WORD VARCHAR(255) NOT NULL,
-  READING VARCHAR(255),
-  TARGET_ROLE VARCHAR(255),
-  TARGET_LABEL VARCHAR(255),
-  BOOST FLOAT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-CREATE UNIQUE INDEX UQ_FAVORITE_LOG ON FAVORITE_LOG (USER_ID, URL);
-
-CREATE INDEX IDX_OVERLAPPING_HOST_BY_REGULAR_NAME_AND_SORT_ORDER ON OVERLAPPING_HOST (REGULAR_NAME, SORT_ORDER);
-
-CREATE INDEX IDX_FILE_CONFIG_TO_LABEL_TYPE_MAPPING_FOR_FILE_CONFIG ON FILE_CONFIG_TO_LABEL_TYPE_MAPPING (FILE_CONFIG_ID);
-
-CREATE INDEX IDX_WEB_CONFIG_TO_LABEL_TYPE_MAPPING__FOR_WEB_CONFIG ON WEB_CONFIG_TO_LABEL_TYPE_MAPPING (WEB_CONFIG_ID);
-
-CREATE INDEX IDX_FILE_CONFIG_TO_ROLE_TYPE_MAPPING_FOR_FILE_CONFIG ON FILE_CONFIG_TO_ROLE_TYPE_MAPPING (FILE_CONFIG_ID);
-
-CREATE INDEX IDX_WEB_CONFIG_TO_ROLE_TYPE_MAPPING_FOR_WEB_CONFIG ON WEB_CONFIG_TO_ROLE_TYPE_MAPPING (WEB_CONFIG_ID);
-
-CREATE INDEX IDX_DATA_CONFIG_TO_ROLE_TYPE_MAPPING_FOR_DATA_CONFIG ON DATA_CONFIG_TO_ROLE_TYPE_MAPPING (DATA_CONFIG_ID);
-
-CREATE INDEX IDX_DATA_CONFIG_TO_LABEL_TYPE_MAPPING_FOR_DATA_CONFIG ON DATA_CONFIG_TO_LABEL_TYPE_MAPPING (DATA_CONFIG_ID);
-
-CREATE INDEX IDX_SEARCH_LOG_BY_HIT_COUNT ON SEARCH_LOG (HIT_COUNT);
-CREATE INDEX IDX_SEARCH_LOG_BY_RESPONSE_TIME ON SEARCH_LOG (RESPONSE_TIME);
-CREATE INDEX IDX_SEARCH_LOG_BY_REQUESTED_TIME ON SEARCH_LOG (REQUESTED_TIME);
-CREATE INDEX IDX_SEARCH_LOG_BY_SEARCH_WORD ON SEARCH_LOG (SEARCH_WORD);
-CREATE INDEX IDX_SEARCH_LOG_BY_RTIME_USID ON SEARCH_LOG (REQUESTED_TIME, USER_SESSION_ID);
-CREATE INDEX IDX_SEARCH_LOG_BY_USER_ID ON SEARCH_LOG (USER_ID);
-
-CREATE INDEX IDX_CLICK_LOG_URL ON CLICK_LOG (URL);
-
-CREATE INDEX IDX_FAILURE_URL_FOR_LIST ON FAILURE_URL (URL, LAST_ACCESS_TIME, ERROR_NAME, ERROR_COUNT);
-CREATE INDEX IDX_FAILURE_URL_BY_CONFIG_ID ON FAILURE_URL (CONFIG_ID);
-
-CREATE INDEX IDX_SEARCH_FIELD_LOG_NAME ON SEARCH_FIELD_LOG (NAME);
-
-CREATE INDEX IDX_SESSION_NAME_EXPIRED ON CRAWLING_SESSION (NAME, EXPIRED_TIME);
-
-INSERT INTO PUBLIC.SCHEDULED_JOB(ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, DELETED_BY, DELETED_TIME, VERSION_NO) VALUES
-(1, 'Crawler', 'all', '0 0 0 * * ?', 'groovy', 'return container.getComponent("crawlJob").execute(executor);', 'T', 'T', 'T', 0, 'system', TIMESTAMP '2013-01-01 00:00:00.000', 'system', TIMESTAMP '2013-01-01 00:00:00.000', NULL, NULL, 0),
-(2, 'Minutely Tasks', 'all', '0 * * * * ?', 'groovy', 'return container.getComponent("aggregateLogJob").execute();', 'F', 'F', 'T', 10, 'system', TIMESTAMP '2013-01-01 00:00:00.000', 'system', TIMESTAMP '2013-01-01 00:00:00.000', NULL, NULL, 0),
-(3, 'Hourly Tasks', 'all', '0 0 * * * ?', 'groovy', 'return container.getComponent("updateStatsJob").execute()+container.getComponent("updateHotWordJob").execute();', 'F', 'F', 'T', 20, 'system', TIMESTAMP '2013-01-01 00:00:00.000', 'system', TIMESTAMP '2013-01-01 00:00:00.000', NULL, NULL, 0),
-(4, 'Daily Tasks', 'all', '0 0 0 * * ?', 'groovy', 'return container.getComponent("purgeLogJob").execute();', 'F', 'F', 'T', 30, 'system', TIMESTAMP '2013-01-01 00:00:00.000', 'system', TIMESTAMP '2013-01-01 00:00:00.000', NULL, NULL, 0);
-

+ 0 - 571
src/main/config/mysql/fess.ddl

@@ -1,571 +0,0 @@
-DROP TABLE IF EXISTS FAVORITE_LOG;
-DROP TABLE IF EXISTS SEARCH_FIELD_LOG;
-DROP TABLE IF EXISTS FILE_AUTHENTICATION;
-DROP TABLE IF EXISTS FAILURE_URL;
-DROP TABLE IF EXISTS CLICK_LOG;
-DROP TABLE IF EXISTS LABEL_TYPE_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS SEARCH_LOG;
-DROP TABLE IF EXISTS USER_INFO;
-DROP TABLE IF EXISTS DATA_CONFIG_TO_BROWSER_TYPE_MAPPING;
-DROP TABLE IF EXISTS DATA_CONFIG_TO_LABEL_TYPE_MAPPING;
-DROP TABLE IF EXISTS DATA_CONFIG_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS DATA_CRAWLING_CONFIG;
-DROP TABLE IF EXISTS WEB_CONFIG_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS FILE_CONFIG_TO_ROLE_TYPE_MAPPING;
-DROP TABLE IF EXISTS ROLE_TYPE;
-DROP TABLE IF EXISTS WEB_CONFIG_TO_LABEL_TYPE_MAPPING;
-DROP TABLE IF EXISTS FILE_CONFIG_TO_LABEL_TYPE_MAPPING;
-DROP TABLE IF EXISTS LABEL_TYPE;
-DROP TABLE IF EXISTS CRAWLING_SESSION_INFO;
-DROP TABLE IF EXISTS WEB_AUTHENTICATION;
-DROP TABLE IF EXISTS KEY_MATCH;
-DROP TABLE IF EXISTS BOOST_DOCUMENT_RULE;
-DROP TABLE IF EXISTS REQUEST_HEADER;
-DROP TABLE IF EXISTS OVERLAPPING_HOST;
-DROP TABLE IF EXISTS CRAWLING_SESSION;
-DROP TABLE IF EXISTS PATH_MAPPING;
-DROP TABLE IF EXISTS JOB_LOG;
-DROP TABLE IF EXISTS SCHEDULED_JOB;
-DROP TABLE IF EXISTS FILE_CONFIG_TO_BROWSER_TYPE_MAPPING;
-DROP TABLE IF EXISTS WEB_CONFIG_TO_BROWSER_TYPE_MAPPING;
-DROP TABLE IF EXISTS FILE_CRAWLING_CONFIG;
-DROP TABLE IF EXISTS BROWSER_TYPE;
-DROP TABLE IF EXISTS WEB_CRAWLING_CONFIG;
-DROP TABLE IF EXISTS SUGGEST_BAD_WORD;
-DROP TABLE IF EXISTS SUGGEST_ELAVATE_WORD;
-
-/**********************************/
-/* Table Name: Web Crawling Config */
-/**********************************/
-CREATE TABLE WEB_CRAWLING_CONFIG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  NAME VARCHAR(200) NOT NULL,
-  URLS TEXT NOT NULL,
-  INCLUDED_URLS TEXT,
-  EXCLUDED_URLS TEXT,
-  INCLUDED_DOC_URLS TEXT,
-  EXCLUDED_DOC_URLS TEXT,
-  CONFIG_PARAMETER TEXT,
-  DEPTH INTEGER,
-  MAX_ACCESS_COUNT BIGINT,
-  USER_AGENT VARCHAR(200) NOT NULL,
-  NUM_OF_THREAD INTEGER NOT NULL,
-  INTERVAL_TIME INTEGER NOT NULL,
-  BOOST FLOAT NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: File Crawling Config */
-/**********************************/
-CREATE TABLE FILE_CRAWLING_CONFIG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  NAME VARCHAR(200) NOT NULL,
-  PATHS TEXT NOT NULL,
-  INCLUDED_PATHS TEXT,
-  EXCLUDED_PATHS TEXT,
-  INCLUDED_DOC_PATHS TEXT,
-  EXCLUDED_DOC_PATHS TEXT,
-  CONFIG_PARAMETER TEXT,
-  DEPTH INTEGER,
-  MAX_ACCESS_COUNT BIGINT,
-  NUM_OF_THREAD INTEGER NOT NULL,
-  INTERVAL_TIME INTEGER NOT NULL,
-  BOOST FLOAT NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Scheduled Job */
-/**********************************/
-CREATE TABLE SCHEDULED_JOB(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  NAME VARCHAR(100) NOT NULL,
-  TARGET VARCHAR(100) NOT NULL,
-  CRON_EXPRESSION VARCHAR(100) NOT NULL,
-  SCRIPT_TYPE VARCHAR(100) NOT NULL,
-  SCRIPT_DATA TEXT,
-  CRAWLER VARCHAR(1) NOT NULL,
-  JOB_LOGGING VARCHAR(1) NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Job Log */
-/**********************************/
-CREATE TABLE JOB_LOG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  JOB_NAME VARCHAR(100) NOT NULL,
-  JOB_STATUS VARCHAR(10) NOT NULL,
-  TARGET VARCHAR(100) NOT NULL,
-  SCRIPT_TYPE VARCHAR(100) NOT NULL,
-  SCRIPT_DATA TEXT,
-  SCRIPT_RESULT TEXT,
-  START_TIME TIMESTAMP NOT NULL,
-  END_TIME TIMESTAMP
-);
-
-/**********************************/
-/* Table Name: Path Mapping */
-/**********************************/
-CREATE TABLE PATH_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  REGEX TEXT NOT NULL,
-  REPLACEMENT TEXT NOT NULL,
-  PROCESS_TYPE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Crawling Session */
-/**********************************/
-CREATE TABLE CRAWLING_SESSION(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  SESSION_ID VARCHAR(20) NOT NULL,
-  NAME VARCHAR(20),
-  EXPIRED_TIME TIMESTAMP,
-  CREATED_TIME TIMESTAMP NOT NULL
-);
-
-/**********************************/
-/* Table Name: Overlapping Host */
-/**********************************/
-CREATE TABLE OVERLAPPING_HOST(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  REGULAR_NAME TEXT NOT NULL,
-  OVERLAPPING_NAME TEXT NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Request Header */
-/**********************************/
-CREATE TABLE REQUEST_HEADER(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  NAME VARCHAR(100) NOT NULL,
-  VALUE TEXT NOT NULL,
-  WEB_CRAWLING_CONFIG_ID BIGINT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL,
-  FOREIGN KEY (WEB_CRAWLING_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Boost Document Rule */
-/**********************************/
-CREATE TABLE BOOST_DOCUMENT_RULE(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  URL_EXPR TEXT NOT NULL,
-  BOOST_EXPR TEXT NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Key Match */
-/**********************************/
-CREATE TABLE KEY_MATCH(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  TERM VARCHAR(100) NOT NULL,
-  QUERY TEXT NOT NULL,
-  MAX_SIZE INTEGER NOT NULL,
-  BOOST FLOAT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Web Authentication */
-/**********************************/
-CREATE TABLE WEB_AUTHENTICATION(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  HOSTNAME VARCHAR(100),
-  PORT INTEGER NOT NULL,
-  AUTH_REALM VARCHAR(100),
-  PROTOCOL_SCHEME VARCHAR(10),
-  USERNAME VARCHAR(100) NOT NULL,
-  PASSWORD VARCHAR(100),
-  PARAMETERS TEXT,
-  WEB_CRAWLING_CONFIG_ID BIGINT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL,
-  FOREIGN KEY (WEB_CRAWLING_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Crawling Session Info */
-/**********************************/
-CREATE TABLE CRAWLING_SESSION_INFO(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  CRAWLING_SESSION_ID BIGINT NOT NULL,
-  ID_KEY VARCHAR(20) NOT NULL,
-  VALUE VARCHAR(100) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  FOREIGN KEY (CRAWLING_SESSION_ID) REFERENCES CRAWLING_SESSION (ID)
-);
-
-/**********************************/
-/* Table Name: Label Type */
-/**********************************/
-CREATE TABLE LABEL_TYPE(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  NAME VARCHAR(100) NOT NULL,
-  VALUE VARCHAR(20) NOT NULL,
-  INCLUDED_PATHS TEXT,
-  EXCLUDED_PATHS TEXT,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: File To Label Mapping */
-/**********************************/
-CREATE TABLE FILE_CONFIG_TO_LABEL_TYPE_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  FILE_CONFIG_ID BIGINT NOT NULL,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (FILE_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Web To Label Mapping */
-/**********************************/
-CREATE TABLE WEB_CONFIG_TO_LABEL_TYPE_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  WEB_CONFIG_ID BIGINT NOT NULL,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (WEB_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Role Type */
-/**********************************/
-CREATE TABLE ROLE_TYPE(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  NAME VARCHAR(100) NOT NULL,
-  VALUE VARCHAR(20) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: File To Role Mapping */
-/**********************************/
-CREATE TABLE FILE_CONFIG_TO_ROLE_TYPE_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  FILE_CONFIG_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (FILE_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Web To Role Mapping */
-/**********************************/
-CREATE TABLE WEB_CONFIG_TO_ROLE_TYPE_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  WEB_CONFIG_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (WEB_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Data Crawling Config */
-/**********************************/
-CREATE TABLE DATA_CRAWLING_CONFIG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  NAME VARCHAR(200) NOT NULL,
-  HANDLER_NAME VARCHAR(200) NOT NULL,
-  HANDLER_PARAMETER TEXT,
-  HANDLER_SCRIPT TEXT,
-  BOOST FLOAT NOT NULL,
-  AVAILABLE VARCHAR(1) NOT NULL,
-  SORT_ORDER INTEGER NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Data To Role Mapping */
-/**********************************/
-CREATE TABLE DATA_CONFIG_TO_ROLE_TYPE_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  DATA_CONFIG_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (DATA_CONFIG_ID) REFERENCES DATA_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Data To Label Mapping */
-/**********************************/
-CREATE TABLE DATA_CONFIG_TO_LABEL_TYPE_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  DATA_CONFIG_ID BIGINT NOT NULL,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (DATA_CONFIG_ID) REFERENCES DATA_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: User Info */
-/**********************************/
-CREATE TABLE USER_INFO(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  CODE TEXT NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_TIME TIMESTAMP NOT NULL
-);
-
-/**********************************/
-/* Table Name: Search Log */
-/**********************************/
-CREATE TABLE SEARCH_LOG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  SEARCH_WORD TEXT,
-  REQUESTED_TIME TIMESTAMP NOT NULL,
-  RESPONSE_TIME INTEGER NOT NULL,
-  HIT_COUNT BIGINT NOT NULL,
-  QUERY_OFFSET INTEGER NOT NULL,
-  QUERY_PAGE_SIZE INTEGER NOT NULL,
-  USER_AGENT VARCHAR(255),
-  REFERER TEXT,
-  CLIENT_IP VARCHAR(50),
-  USER_SESSION_ID VARCHAR(100),
-  ACCESS_TYPE VARCHAR(1) NOT NULL,
-  USER_ID BIGINT,
-  FOREIGN KEY (USER_ID) REFERENCES USER_INFO (ID)
-);
-
-/**********************************/
-/* Table Name: Label Type To Role Type Mapping */
-/**********************************/
-CREATE TABLE LABEL_TYPE_TO_ROLE_TYPE_MAPPING(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  LABEL_TYPE_ID BIGINT NOT NULL,
-  ROLE_TYPE_ID BIGINT NOT NULL,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-/**********************************/
-/* Table Name: Click Log */
-/**********************************/
-CREATE TABLE CLICK_LOG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  SEARCH_ID BIGINT NOT NULL,
-  URL TEXT NOT NULL,
-  REQUESTED_TIME TIMESTAMP NOT NULL,
-  FOREIGN KEY (SEARCH_ID) REFERENCES SEARCH_LOG (ID)
-);
-
-/**********************************/
-/* Table Name: Failure Url */
-/**********************************/
-CREATE TABLE FAILURE_URL(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  URL TEXT NOT NULL,
-  THREAD_NAME VARCHAR(30) NOT NULL,
-  ERROR_NAME VARCHAR(255),
-  ERROR_LOG TEXT,
-  ERROR_COUNT INTEGER NOT NULL,
-  LAST_ACCESS_TIME TIMESTAMP NOT NULL,
-  CONFIG_ID VARCHAR(100)
-);
-
-/**********************************/
-/* Table Name: File Authentication */
-/**********************************/
-CREATE TABLE FILE_AUTHENTICATION(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  HOSTNAME VARCHAR(255),
-  PORT INTEGER NOT NULL,
-  PROTOCOL_SCHEME VARCHAR(10),
-  USERNAME VARCHAR(100) NOT NULL,
-  PASSWORD VARCHAR(100),
-  PARAMETERS TEXT,
-  FILE_CRAWLING_CONFIG_ID BIGINT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL,
-  FOREIGN KEY (FILE_CRAWLING_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID)
-);
-
-/**********************************/
-/* Table Name: Search Field Log */
-/**********************************/
-CREATE TABLE SEARCH_FIELD_LOG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  SEARCH_ID BIGINT NOT NULL,
-  NAME VARCHAR(255) NOT NULL,
-  VALUE TEXT NOT NULL,
-  FOREIGN KEY (SEARCH_ID) REFERENCES SEARCH_LOG (ID)
-);
-
-/**********************************/
-/* Table Name: Favorite Log */
-/**********************************/
-CREATE TABLE FAVORITE_LOG(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  USER_ID BIGINT NOT NULL,
-  URL TEXT NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  FOREIGN KEY (USER_ID) REFERENCES USER_INFO (ID)
-);
-
-/**********************************/
-/* Table Name: Suggest Ng Word */
-/**********************************/
-CREATE TABLE SUGGEST_BAD_WORD(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  SUGGEST_WORD VARCHAR(255) NOT NULL,
-  TARGET_ROLE VARCHAR(255),
-  TARGET_LABEL VARCHAR(255),
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-/**********************************/
-/* Table Name: Suggest Elevate word */
-/**********************************/
-CREATE TABLE SUGGEST_ELEVATE_WORD(
-  ID BIGINT NOT NULL PRIMARY KEY AUTO_INCREMENT,
-  SUGGEST_WORD VARCHAR(255) NOT NULL,
-  READING VARCHAR(255),
-  TARGET_ROLE VARCHAR(255),
-  TARGET_LABEL VARCHAR(255),
-  BOOST FLOAT NOT NULL,
-  CREATED_BY VARCHAR(255) NOT NULL,
-  CREATED_TIME TIMESTAMP NOT NULL,
-  UPDATED_BY VARCHAR(255),
-  UPDATED_TIME TIMESTAMP,
-  DELETED_BY VARCHAR(255),
-  DELETED_TIME TIMESTAMP,
-  VERSION_NO INTEGER NOT NULL
-);
-
-CREATE UNIQUE INDEX UQ_FAVORITE_LOG ON FAVORITE_LOG (USER_ID, URL(200));
-
-CREATE INDEX IDX_OVERLAPPING_HOST_BY_REGULAR_NAME_AND_SORT_ORDER ON OVERLAPPING_HOST (REGULAR_NAME(200), SORT_ORDER);
-
-CREATE INDEX IDX_FILE_CONFIG_TO_LABEL_TYPE_MAPPING_FOR_FILE_CONFIG ON FILE_CONFIG_TO_LABEL_TYPE_MAPPING (FILE_CONFIG_ID);
-
-CREATE INDEX IDX_WEB_CONFIG_TO_LABEL_TYPE_MAPPING__FOR_WEB_CONFIG ON WEB_CONFIG_TO_LABEL_TYPE_MAPPING (WEB_CONFIG_ID);
-
-CREATE INDEX IDX_FILE_CONFIG_TO_ROLE_TYPE_MAPPING_FOR_FILE_CONFIG ON FILE_CONFIG_TO_ROLE_TYPE_MAPPING (FILE_CONFIG_ID);
-
-CREATE INDEX IDX_WEB_CONFIG_TO_ROLE_TYPE_MAPPING_FOR_WEB_CONFIG ON WEB_CONFIG_TO_ROLE_TYPE_MAPPING (WEB_CONFIG_ID);
-
-CREATE INDEX IDX_DATA_CONFIG_TO_ROLE_TYPE_MAPPING_FOR_DATA_CONFIG ON DATA_CONFIG_TO_ROLE_TYPE_MAPPING (DATA_CONFIG_ID);
-
-CREATE INDEX IDX_DATA_CONFIG_TO_LABEL_TYPE_MAPPING_FOR_DATA_CONFIG ON DATA_CONFIG_TO_LABEL_TYPE_MAPPING (DATA_CONFIG_ID);
-
-CREATE INDEX IDX_SEARCH_LOG_BY_HIT_COUNT ON SEARCH_LOG (HIT_COUNT);
-CREATE INDEX IDX_SEARCH_LOG_BY_RESPONSE_TIME ON SEARCH_LOG (RESPONSE_TIME);
-CREATE INDEX IDX_SEARCH_LOG_BY_REQUESTED_TIME ON SEARCH_LOG (REQUESTED_TIME);
-CREATE INDEX IDX_SEARCH_LOG_BY_SEARCH_WORD ON SEARCH_LOG (SEARCH_WORD(255));
-CREATE INDEX IDX_SEARCH_LOG_BY_RTIME_USID ON SEARCH_LOG (REQUESTED_TIME, USER_SESSION_ID);
-CREATE INDEX IDX_SEARCH_LOG_BY_USER_ID ON SEARCH_LOG (USER_ID);
-
-CREATE INDEX IDX_CLICK_LOG_URL ON CLICK_LOG (URL(255));
-
-CREATE INDEX IDX_FAILURE_URL_FOR_LIST ON FAILURE_URL (URL(200), LAST_ACCESS_TIME, ERROR_NAME(100), ERROR_COUNT);
-CREATE INDEX IDX_FAILURE_URL_BY_CONFIG_ID ON FAILURE_URL (CONFIG_ID);
-
-CREATE INDEX IDX_SEARCH_FIELD_LOG_NAME ON SEARCH_FIELD_LOG (NAME);
-
-CREATE INDEX IDX_SESSION_NAME_EXPIRED ON CRAWLING_SESSION (NAME, EXPIRED_TIME);
-
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (1, 'Crawler', 'all', '0 0 0 * * ?', 'groovy', 'return container.getComponent("crawlJob").execute(executor);', 'T', 'T', 'T', 0, 'system', '2000-01-01 00:00:00', 'system', '2000-01-01 00:00:00', 0);
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (2, 'Minutely Tasks', 'all', '0 * * * * ?', 'groovy', 'return container.getComponent("aggregateLogJob").execute();', 'F', 'F', 'T', 10, 'system', '2000-01-01 00:00:00', 'system', '2000-01-01 00:00:00', 0);
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (3, 'Hourly Tasks', 'all', '0 0 * * * ?', 'groovy', 'return container.getComponent("updateStatsJob").execute()+container.getComponent("updateHotWordJob").execute();', 'F', 'F', 'T', 20, 'system', '2000-01-01 00:00:00', 'system', '2000-01-01 00:00:00', 0);
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (4, 'Daily Tasks', 'all', '0 0 0 * * ?', 'groovy', 'return container.getComponent("purgeLogJob").execute();', 'F', 'F', 'T', 30, 'system', '2000-01-01 00:00:00', 'system', '2000-01-01 00:00:00', 0);
-

+ 0 - 16
src/main/config/oracle/build-create.xml

@@ -1,16 +0,0 @@
-<project basedir=".">
-   <property name="sql.driver" value="oracle.jdbc.driver.OracleDriver"/>
-   <property name="sql.url" value="jdbc:oracle:thin:@localhost:1521:XE"/>
-   <property name="sql.user" value="fess"/>
-   <property name="sql.pass" value="fess"/>
-
-   <target name="create">
-      <sql driver="${sql.driver}" url="${sql.url}" userid="${sql.user}" password="${sql.pass}" onerror="continue">
-           <classpath>
-            <pathelement location="${oracle.jar.file}"/>
-           </classpath>   
-           <transaction src="fess.ddl"/>
-      </sql>
-   </target>
-
-</project>

+ 0 - 606
src/main/config/oracle/fess.ddl

@@ -1,606 +0,0 @@
-DROP SEQUENCE WEB_CRAWLING_CONFIG_SEQ;
-DROP SEQUENCE FILE_CRAWLING_CONFIG_SEQ;
-DROP SEQUENCE JOB_LOG_SEQ;
-DROP SEQUENCE SCHEDULED_JOB_SEQ;
-DROP SEQUENCE PATH_MAPPING_SEQ;
-DROP SEQUENCE CRAWLING_SESSION_SEQ;
-DROP SEQUENCE OVERLAPPING_HOST_SEQ;
-DROP SEQUENCE REQUEST_HEADER_SEQ;
-DROP SEQUENCE KEY_MATCH_SEQ;
-DROP SEQUENCE BOOST_DOCUMENT_RULE_SEQ;
-DROP SEQUENCE WEB_AUTHENTICATION_SEQ;
-DROP SEQUENCE CRAWLING_SESSION_INFO_SEQ;
-DROP SEQUENCE LABEL_TYPE_SEQ;
-DROP SEQUENCE FILE_CONFIG_TO_LABEL_TYPE_SEQ;
-DROP SEQUENCE WEB_CONFIG_TO_LABEL_TYPE_SEQ;
-DROP SEQUENCE ROLE_TYPE_SEQ;
-DROP SEQUENCE FILE_CONFIG_TO_ROLE_TYPE_SEQ;
-DROP SEQUENCE WEB_CONFIG_TO_ROLE_TYPE_SEQ;
-DROP SEQUENCE DATA_CRAWLING_CONFIG_SEQ;
-DROP SEQUENCE DATA_CONFIG_TO_ROLE_TYPE_SEQ;
-DROP SEQUENCE DATA_CONFIG_TO_LABEL_TYPE_SEQ;
-DROP SEQUENCE USER_INFO_SEQ;
-DROP SEQUENCE SEARCH_LOG_SEQ;
-DROP SEQUENCE LABEL_TYPE_TO_ROLE_TYPE_SEQ;
-DROP SEQUENCE CLICK_LOG_SEQ;
-DROP SEQUENCE FAILURE_URL_SEQ;
-DROP SEQUENCE FILE_AUTHENTICATION_SEQ;
-DROP SEQUENCE SEARCH_FIELD_LOG_SEQ;
-DROP SEQUENCE FAVORITE_LOG_SEQ;
-DROP SEQUENCE SUGGEST_BAD_WORD;
-DROP SEQUENCE SUGGEST_ELAVATE_WORD;
-
-DROP TABLE "FAVORITE_LOG";
-DROP TABLE "SEARCH_FIELD_LOG";
-DROP TABLE "FILE_AUTHENTICATION";
-DROP TABLE "FAILURE_URL";
-DROP TABLE "CLICK_LOG";
-DROP TABLE "LABEL_TYPE_TO_ROLE_TYPE";
-DROP TABLE "SEARCH_LOG";
-DROP TABLE "USER_INFO";
-DROP TABLE "DATA_CONFIG_TO_LABEL_TYPE";
-DROP TABLE "DATA_CONFIG_TO_ROLE_TYPE";
-DROP TABLE "DATA_CRAWLING_CONFIG";
-DROP TABLE "WEB_CONFIG_TO_ROLE_TYPE";
-DROP TABLE "FILE_CONFIG_TO_ROLE_TYPE";
-DROP TABLE "ROLE_TYPE";
-DROP TABLE "WEB_CONFIG_TO_LABEL_TYPE";
-DROP TABLE "FILE_CONFIG_TO_LABEL_TYPE";
-DROP TABLE "LABEL_TYPE";
-DROP TABLE "CRAWLING_SESSION_INFO";
-DROP TABLE "WEB_AUTHENTICATION";
-DROP TABLE "KEY_MATCH";
-DROP TABLE "BOOST_DOCUMENT_RULE";
-DROP TABLE "REQUEST_HEADER";
-DROP TABLE "OVERLAPPING_HOST";
-DROP TABLE "CRAWLING_SESSION";
-DROP TABLE "PATH_MAPPING";
-DROP TABLE "SCHEDULED_JOB";
-DROP TABLE "JOB_LOG";
-DROP TABLE "FILE_CRAWLING_CONFIG";
-DROP TABLE "WEB_CRAWLING_CONFIG";
-DROP TABLE "SUGGEST_BAD_WORD";
-DROP TABLE "SUGGEST_ELAVATE_WORD";
-
-CREATE TABLE "WEB_CRAWLING_CONFIG"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(200) NOT NULL,
-  "URLS" VARCHAR2(4000) NOT NULL,
-  "INCLUDED_URLS" VARCHAR2(4000),
-  "EXCLUDED_URLS" VARCHAR2(4000),
-  "INCLUDED_DOC_URLS" VARCHAR2(4000),
-  "EXCLUDED_DOC_URLS" VARCHAR2(4000),
-  "CONFIG_PARAMETER" VARCHAR2(4000),
-  "DEPTH" NUMBER(7,0),
-  "MAX_ACCESS_COUNT" NUMBER(18,0),
-  "USER_AGENT" VARCHAR2(200) NOT NULL,
-  "NUM_OF_THREAD" NUMBER(9,0) NOT NULL,
-  "INTERVAL_TIME" NUMBER(9,0) NOT NULL,
-  "BOOST" FLOAT NOT NULL,
-  "AVAILABLE" VARCHAR2(1) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "WEB_CRAWLING_CONFIG_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "FILE_CRAWLING_CONFIG"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(200) NOT NULL,
-  "PATHS" VARCHAR2(4000) NOT NULL,
-  "INCLUDED_PATHS" VARCHAR2(4000),
-  "EXCLUDED_PATHS" VARCHAR2(4000),
-  "INCLUDED_DOC_PATHS" VARCHAR2(4000),
-  "EXCLUDED_DOC_PATHS" VARCHAR2(4000),
-  "CONFIG_PARAMETER" VARCHAR2(4000),
-  "DEPTH" NUMBER(9,0),
-  "MAX_ACCESS_COUNT" NUMBER(18,0),
-  "NUM_OF_THREAD" NUMBER(9,0) NOT NULL,
-  "INTERVAL_TIME" NUMBER(9,0) NOT NULL,
-  "BOOST" FLOAT NOT NULL,
-  "AVAILABLE" VARCHAR2(1) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "FILE_CRAWLING_CONFIG_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "SCHEDULED_JOB"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(100) NOT NULL,
-  "TARGET" VARCHAR2(100) NOT NULL,
-  "CRON_EXPRESSION" VARCHAR2(100) NOT NULL,
-  "SCRIPT_TYPE" VARCHAR2(100) NOT NULL,
-  "SCRIPT_DATA" VARCHAR2(4000),
-  "CRAWLER" VARCHAR2(1) NOT NULL,
-  "JOB_LOGGING" VARCHAR2(1) NOT NULL,
-  "AVAILABLE" VARCHAR2(1) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "SCHEDULED_JOB_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "JOB_LOG"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "JOB_NAME" VARCHAR2(100) NOT NULL,
-  "JOB_STATUS" VARCHAR2(10) NOT NULL,
-  "TARGET" VARCHAR2(100) NOT NULL,
-  "SCRIPT_TYPE" VARCHAR2(100) NOT NULL,
-  "SCRIPT_DATA" VARCHAR2(4000),
-  "SCRIPT_RESULT" VARCHAR2(4000),
-  "START_TIME" TIMESTAMP NOT NULL,
-  "END_TIME" TIMESTAMP,
-
-  CONSTRAINT "JOB_LOG_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "PATH_MAPPING"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "REGEX" VARCHAR2(1000) NOT NULL,
-  "REPLACEMENT" VARCHAR2(1000) NOT NULL,
-  "PROCESS_TYPE" VARCHAR2(1) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "PATH_MAPPING_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "CRAWLING_SESSION"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "SESSION_ID" VARCHAR2(20) NOT NULL,
-  "NAME" VARCHAR2(20),
-  "EXPIRED_TIME" TIMESTAMP,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-
-  CONSTRAINT "CRAWLING_SESSION_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "OVERLAPPING_HOST"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "REGULAR_NAME" VARCHAR2(1000) NOT NULL,
-  "OVERLAPPING_NAME" VARCHAR2(1000) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "OVERLAPPING_HOST_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "REQUEST_HEADER"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(100) NOT NULL,
-  "VALUE" VARCHAR2(1000) NOT NULL,
-  "WEB_CRAWLING_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "REQUEST_HEADER_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (WEB_CRAWLING_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-CREATE TABLE "BOOST_DOCUMENT_RULE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "URL_EXPR" VARCHAR2(4000) NOT NULL,
-  "BOOST_EXPR" VARCHAR2(4000) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "BOOST_DOCUMENT_RULE_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "KEY_MATCH"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "TERM" VARCHAR2(100) NOT NULL,
-  "QUERY" VARCHAR2(4000) NOT NULL,
-  "MAX_SIZE" NUMBER(9,0) NOT NULL,
-  "BOOST" FLOAT NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "KEY_MATCH_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "WEB_AUTHENTICATION"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "HOSTNAME" VARCHAR2(100),
-  "PORT" NUMBER(9,0) NOT NULL,
-  "AUTH_REALM" VARCHAR2(100),
-  "PROTOCOL_SCHEME" VARCHAR2(10),
-  "USERNAME" VARCHAR2(100) NOT NULL,
-  "PASSWORD" VARCHAR2(100),
-  "PARAMETERS" VARCHAR2(1000),
-  "WEB_CRAWLING_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "WEB_AUTHENTICATION_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (WEB_CRAWLING_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-CREATE TABLE "CRAWLING_SESSION_INFO"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "CRAWLING_SESSION_ID" NUMBER(18,0) NOT NULL,
-  "KEY" VARCHAR2(20) NOT NULL,
-  "VALUE" VARCHAR2(100) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-
-  CONSTRAINT "CRAWLING_SESSION_INFO_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (CRAWLING_SESSION_ID) REFERENCES CRAWLING_SESSION (ID)
-);
-
-CREATE TABLE "LABEL_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(100) NOT NULL,
-  "VALUE" VARCHAR2(20) NOT NULL,
-  "INCLUDED_PATHS" VARCHAR2(4000),
-  "EXCLUDED_PATHS" VARCHAR2(4000),
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "LABEL_TYPE_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "FILE_CONFIG_TO_LABEL_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "FILE_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "LABEL_TYPE_ID" NUMBER(18,0) NOT NULL,
-
-  CONSTRAINT "FILE_CONFIG_TO_LABEL_TYPE_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (FILE_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID)
-);
-
-CREATE TABLE "WEB_CONFIG_TO_LABEL_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "WEB_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "LABEL_TYPE_ID" NUMBER(18,0) NOT NULL,
-
-  CONSTRAINT "WEB_CONFIG_TO_LABEL_TYPE_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (WEB_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID)
-);
-
-CREATE TABLE "ROLE_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(100) NOT NULL,
-  "VALUE" VARCHAR2(20) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "ROLE_TYPE_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "FILE_CONFIG_TO_ROLE_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "FILE_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "ROLE_TYPE_ID" NUMBER(18,0) NOT NULL,
-
-  CONSTRAINT "FILE_CONFIG_TO_ROLE_TYPE_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (FILE_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-CREATE TABLE "WEB_CONFIG_TO_ROLE_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "WEB_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "ROLE_TYPE_ID" NUMBER(18,0) NOT NULL,
-
-  CONSTRAINT "WEB_CONFIG_TO_ROLE_TYPE_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (WEB_CONFIG_ID) REFERENCES WEB_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-CREATE TABLE "DATA_CRAWLING_CONFIG"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(200) NOT NULL,
-  "HANDLER_NAME" VARCHAR2(200) NOT NULL,
-  "HANDLER_PARAMETER" VARCHAR2(4000),
-  "HANDLER_SCRIPT" VARCHAR2(4000),
-  "BOOST" FLOAT NOT NULL,
-  "AVAILABLE" VARCHAR2(1) NOT NULL,
-  "SORT_ORDER" NUMBER(9,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "DATA_CRAWLING_CONFIG_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "DATA_CONFIG_TO_ROLE_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "DATA_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "ROLE_TYPE_ID" NUMBER(18,0) NOT NULL,
-
-  CONSTRAINT "DATA_CONFIG_TO_ROLE_TYPE_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (DATA_CONFIG_ID) REFERENCES DATA_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-CREATE TABLE "DATA_CONFIG_TO_LABEL_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "DATA_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "LABEL_TYPE_ID" NUMBER(18,0) NOT NULL,
-
-  CONSTRAINT "DATA_CONFIG_TO_LABEL_TYPE_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (DATA_CONFIG_ID) REFERENCES DATA_CRAWLING_CONFIG (ID),
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID)
-);
-
-CREATE TABLE "USER_INFO" (
-  "ID" NUMBER(18,0) NOT NULL,
-  "CODE" VARCHAR2(1000) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_TIME" TIMESTAMP NOT NULL,
-
-  CONSTRAINT "USER_INFO_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "SEARCH_LOG"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "SEARCH_WORD" VARCHAR2(1000),
-  "REQUESTED_TIME" TIMESTAMP NOT NULL,
-  "RESPONSE_TIME" NUMBER(9,0) NOT NULL,
-  "HIT_COUNT" NUMBER(18,0) NOT NULL,
-  "QUERY_OFFSET" NUMBER(9,0) NOT NULL,
-  "QUERY_PAGE_SIZE" NUMBER(9,0) NOT NULL,
-  "USER_AGENT" VARCHAR2(255),
-  "REFERER" VARCHAR2(1000),
-  "CLIENT_IP" VARCHAR2(50),
-  "USER_SESSION_ID" VARCHAR2(100),
-  "ACCESS_TYPE" VARCHAR2(1) NOT NULL,
-  "USER_ID" NUMBER(18,0),
-
-  CONSTRAINT "SEARCH_LOG_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (USER_ID) REFERENCES USER_INFO (ID)
-);
-
-CREATE TABLE "LABEL_TYPE_TO_ROLE_TYPE"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "LABEL_TYPE_ID" NUMBER(18,0) NOT NULL,
-  "ROLE_TYPE_ID" NUMBER(18,0) NOT NULL,
-
-  CONSTRAINT "LABEL_TYPE_TO_ROLE_TYPE_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (LABEL_TYPE_ID) REFERENCES LABEL_TYPE (ID),
-  FOREIGN KEY (ROLE_TYPE_ID) REFERENCES ROLE_TYPE (ID)
-);
-
-CREATE TABLE "CLICK_LOG"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "SEARCH_ID" NUMBER(18,0) NOT NULL,
-  "URL" VARCHAR2(4000) NOT NULL,
-  "REQUESTED_TIME" TIMESTAMP NOT NULL,
-
-  CONSTRAINT "CLICK_LOG_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (SEARCH_ID) REFERENCES SEARCH_LOG (ID)
-);
-
-CREATE TABLE "FAILURE_URL"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "URL" VARCHAR2(4000) NOT NULL,
-  "THREAD_NAME" VARCHAR2(30) NOT NULL,
-  "ERROR_NAME" VARCHAR2(255),
-  "ERROR_LOG" VARCHAR2(4000),
-  "ERROR_COUNT" NUMBER(9,0) NOT NULL,
-  "LAST_ACCESS_TIME" TIMESTAMP NOT NULL,
-  "CONFIG_ID" VARCHAR2(100),
-
-  CONSTRAINT "FAILURE_URL_PK" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE TABLE "FILE_AUTHENTICATION"(
-  "ID" NUMBER(18,0) NOT NULL,
-  "HOSTNAME" VARCHAR2(255),
-  "PORT" NUMBER(9,0) NOT NULL,
-  "PROTOCOL_SCHEME" VARCHAR2(10),
-  "USERNAME" VARCHAR2(100) NOT NULL,
-  "PASSWORD" VARCHAR2(100),
-  "PARAMETERS" VARCHAR2(1000),
-  "FILE_CRAWLING_CONFIG_ID" NUMBER(18,0) NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "FILE_AUTHENTICATION_PK" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (FILE_CRAWLING_CONFIG_ID) REFERENCES FILE_CRAWLING_CONFIG (ID)
-);
-
-CREATE TABLE "SEARCH_FIELD_LOG" (
-  "ID" NUMBER(18,0) NOT NULL,
-  "SEARCH_ID" NUMBER(18,0) NOT NULL,
-  "NAME" VARCHAR2(255) NOT NULL,
-  "VALUE" VARCHAR2(4000) NOT NULL,
-
-  CONSTRAINT "SEARCH_FIELD_LOG" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (SEARCH_ID) REFERENCES SEARCH_LOG (ID)
-);
-
-CREATE TABLE "FAVORITE_LOG" (
-  "ID" NUMBER(18,0) NOT NULL,
-  "USER_ID" NUMBER(18,0) NOT NULL,
-  "URL" VARCHAR2(4000) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-
-  CONSTRAINT "FAVORITE_LOG" PRIMARY KEY ("ID") ENABLE,
-  FOREIGN KEY (USER_ID) REFERENCES USER_INFO (ID)
-);
-
-/**********************************/
-/* Table Name: Suggest Ng Word */
-/**********************************/
-CREATE TABLE "SUGGEST_BAD_WORD" (
-  "ID" NUMBER(18,0) NOT NULL,
-  "SUGGEST_WORD" VARCHAR2(255) NOT NULL,
-  "TARGET_ROLE" VARCHAR2(255),
-  "TARGET_LABEL" VARCHAR2(255),
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "SUGGEST_BAD_WORD" PRIMARY KEY ("ID") ENABLE
-);
-
-/**********************************/
-/* Table Name: Suggest Elevate word */
-/**********************************/
-CREATE TABLE "SUGGEST_ELEVATE_WORD" (
-  "ID" NUMBER(18,0) NOT NULL,
-  "SUGGEST_WORD" VARCHAR2(255) NOT NULL,
-  "READING" VARCHAR2(255),
-  "TARGET_ROLE" VARCHAR2(255),
-  "TARGET_LABEL" VARCHAR2(255),
-  "BOOST" FLOAT NOT NULL,
-  "CREATED_BY" VARCHAR2(255) NOT NULL,
-  "CREATED_TIME" TIMESTAMP NOT NULL,
-  "UPDATED_BY" VARCHAR2(255),
-  "UPDATED_TIME" TIMESTAMP,
-  "DELETED_BY" VARCHAR2(255),
-  "DELETED_TIME" TIMESTAMP,
-  "VERSION_NO" NUMBER(9,0) NOT NULL,
-
-  CONSTRAINT "SUGGEST_ELEVATE_WORD" PRIMARY KEY ("ID") ENABLE
-);
-
-CREATE UNIQUE INDEX UQ_FAVORITE_LOG ON FAVORITE_LOG (USER_ID, URL);
-
-CREATE INDEX IDX_O_H_BY_R_N_AND_S_O ON OVERLAPPING_HOST (REGULAR_NAME, SORT_ORDER);
-
-CREATE INDEX IDX_F_C_TO_L_T_FOR_F_C ON FILE_CONFIG_TO_LABEL_TYPE (FILE_CONFIG_ID);
-
-CREATE INDEX IDX_W_C_TO_L_T_FOR_W_C ON WEB_CONFIG_TO_LABEL_TYPE (WEB_CONFIG_ID);
-
-CREATE INDEX IDX_F_C_TO_R_T_FOR_F_C ON FILE_CONFIG_TO_ROLE_TYPE (FILE_CONFIG_ID);
-
-CREATE INDEX IDX_W_C_TO_R_T_FOR_W_C ON WEB_CONFIG_TO_ROLE_TYPE (WEB_CONFIG_ID);
-
-CREATE INDEX IDX_D_C_TO_R_T_FOR_D_C ON DATA_CONFIG_TO_ROLE_TYPE (DATA_CONFIG_ID);
-
-CREATE INDEX IDX_D_C_TO_L_T_FOR_D_C ON DATA_CONFIG_TO_LABEL_TYPE (DATA_CONFIG_ID);
-
-CREATE INDEX IDX_S_L_BY_H_C ON SEARCH_LOG (HIT_COUNT);
-CREATE INDEX IDX_S_L_BY_R_T ON SEARCH_LOG (RESPONSE_TIME);
-CREATE INDEX IDX_S_L_BY_RT ON SEARCH_LOG (REQUESTED_TIME);
-CREATE INDEX IDX_S_L_BY_S_W ON SEARCH_LOG (SEARCH_WORD);
-CREATE INDEX IDX_S_L_BY_RT_USID ON SEARCH_LOG (REQUESTED_TIME, USER_SESSION_ID);
-CREATE INDEX IDX_S_L_BY_USID ON SEARCH_LOG (USER_ID);
-
-CREATE INDEX IDX_C_L_URL ON CLICK_LOG (URL);
-
-CREATE INDEX IDX_F_U_FOR_L ON FAILURE_URL (URL, LAST_ACCESS_TIME, ERROR_NAME, ERROR_COUNT);
-CREATE INDEX IDX_F_U_BY_W_C_ID ON FAILURE_URL (CONFIG_ID);
-
-CREATE INDEX IDX_S_F_LOG_NAME ON SEARCH_FIELD_LOG (NAME);
-
-CREATE INDEX IDX_S_N_EXPIRED ON CRAWLING_SESSION (NAME, EXPIRED_TIME);
-
-CREATE SEQUENCE WEB_CRAWLING_CONFIG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE FILE_CRAWLING_CONFIG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE JOB_LOG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE SCHEDULED_JOB_SEQ START WITH 5 INCREMENT BY 50;
-CREATE SEQUENCE PATH_MAPPING_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE CRAWLING_SESSION_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE OVERLAPPING_HOST_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE REQUEST_HEADER_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE BOOST_DOCUMENT_RULE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE KEY_MATCH_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE WEB_AUTHENTICATION_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE CRAWLING_SESSION_INFO_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE LABEL_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE FILE_CONFIG_TO_LABEL_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE WEB_CONFIG_TO_LABEL_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE ROLE_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE FILE_CONFIG_TO_ROLE_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE WEB_CONFIG_TO_ROLE_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE DATA_CRAWLING_CONFIG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE DATA_CONFIG_TO_ROLE_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE DATA_CONFIG_TO_LABEL_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE USER_INFO_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE SEARCH_LOG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE LABEL_TYPE_TO_ROLE_TYPE_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE CLICK_LOG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE FAILURE_URL_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE FILE_AUTHENTICATION_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE SEARCH_FIELD_LOG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE FAVORITE_LOG_SEQ START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE SUGGEST_BAD_WORD START WITH 1 INCREMENT BY 50;
-CREATE SEQUENCE SUGGEST_ELEVATE_WORD START WITH 1 INCREMENT BY 50;
-
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (1, 'Crawler', 'all', '0 0 0 * * ?', 'groovy', 'return container.getComponent("crawlJob").execute(executor);', 'T', 'T', 'T', 0, 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 0);
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (2, 'Minutely Tasks', 'all', '0 * * * * ?', 'groovy', 'return container.getComponent("aggregateLogJob").execute();', 'F', 'F', 'T', 10, 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 0);
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (3, 'Hourly Tasks', 'all', '0 0 * * * ?', 'groovy', 'return container.getComponent("updateStatsJob").execute()+container.getComponent("updateHotWordJob").execute();', 'F', 'F', 'T', 20, 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 0);
-INSERT INTO SCHEDULED_JOB (ID, NAME, TARGET, CRON_EXPRESSION, SCRIPT_TYPE, SCRIPT_DATA, CRAWLER, JOB_LOGGING, AVAILABLE, SORT_ORDER, CREATED_BY, CREATED_TIME, UPDATED_BY, UPDATED_TIME, VERSION_NO) VALUES (4, 'Daily Tasks', 'all', '0 0 0 * * ?', 'groovy', 'return container.getComponent("purgeLogJob").execute();', 'F', 'F', 'T', 30, 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 'system', to_date('2000-01-01', 'yyyy-MM-dd'), 0);
-

+ 0 - 15
src/main/config/oracle/update-schema-xml.sh

@@ -1,15 +0,0 @@
-#!/bin/bash
-
-# the following is in fess-server/build.xml.
-
-FILE=dbflute_oracle/schema/project-schema-fess.xml
-
-perl -pi -e 's/name="DATA_CONFIG_TO_LABEL_TYPE"/javaName="DataConfigToLabelTypeMapping" name="DATA_CONFIG_TO_LABEL_TYPE"/' $FILE
-perl -pi -e 's/name="DATA_CONFIG_TO_ROLE_TYPE"/javaName="DataConfigToRoleTypeMapping" name="DATA_CONFIG_TO_ROLE_TYPE"/' $FILE
-perl -pi -e 's/name="FILE_CONFIG_TO_LABEL_TYPE"/javaName="FileConfigToLabelTypeMapping" name="FILE_CONFIG_TO_LABEL_TYPE"/' $FILE
-perl -pi -e 's/name="FILE_CONFIG_TO_ROLE_TYPE"/javaName="FileConfigToRoleTypeMapping" name="FILE_CONFIG_TO_ROLE_TYPE"/' $FILE
-perl -pi -e 's/name="LABEL_TYPE_TO_ROLE_TYPE"/javaName="LabelTypeToRoleTypeMapping" name="LABEL_TYPE_TO_ROLE_TYPE"/' $FILE
-perl -pi -e 's/name="WEB_CONFIG_TO_LABEL_TYPE"/javaName="WebConfigToLabelTypeMapping" name="WEB_CONFIG_TO_LABEL_TYPE"/' $FILE
-perl -pi -e 's/name="WEB_CONFIG_TO_ROLE_TYPE"/javaName="WebConfigToRoleTypeMapping" name="WEB_CONFIG_TO_ROLE_TYPE"/' $FILE
-
-perl -pi -e 's/dbType="FLOAT" javaType="java.math.BigDecimal"/dbType="FLOAT" javaType="Float"/g' $FILE

+ 0 - 0
src/main/h2/resources/.gitkeep


+ 0 - 74
src/main/h2/resources/sql/robot.ddl

@@ -1,74 +0,0 @@
-DROP TABLE IF EXISTS URL_FILTER;
-DROP TABLE IF EXISTS ACCESS_RESULT_DATA;
-DROP TABLE IF EXISTS ACCESS_RESULT;
-DROP TABLE IF EXISTS URL_QUEUE;
-
-/**********************************/
-/* Table Name: URL Queue */
-/**********************************/
-CREATE TABLE URL_QUEUE(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SESSION_ID VARCHAR(20) NOT NULL,
-  METHOD VARCHAR(10) NOT NULL,
-  URL VARCHAR(65536) NOT NULL,
-  META_DATA VARCHAR(65536),
-  ENCODING VARCHAR(20),
-  PARENT_URL VARCHAR(65536),
-  DEPTH INTEGER NOT NULL,
-  LAST_MODIFIED TIMESTAMP,
-  CREATE_TIME TIMESTAMP NOT NULL
-);
-
-/**********************************/
-/* Table Name: Access Result */
-/**********************************/
-CREATE TABLE ACCESS_RESULT(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SESSION_ID VARCHAR(20) NOT NULL,
-  RULE_ID VARCHAR(20),
-  URL VARCHAR(65536) NOT NULL,
-  PARENT_URL VARCHAR(65536),
-  STATUS INTEGER NOT NULL,
-  HTTP_STATUS_CODE INTEGER NOT NULL,
-  METHOD VARCHAR(10) NOT NULL,
-  MIME_TYPE VARCHAR(100) NOT NULL,
-  CONTENT_LENGTH BIGINT NOT NULL,
-  EXECUTION_TIME INTEGER NOT NULL,
-  LAST_MODIFIED TIMESTAMP,
-  CREATE_TIME TIMESTAMP NOT NULL
-);
-
-/**********************************/
-/* Table Name: Access Result Data */
-/**********************************/
-CREATE TABLE ACCESS_RESULT_DATA(
-  ID BIGINT(20) NOT NULL PRIMARY KEY,
-  TRANSFORMER_NAME VARCHAR(255) NOT NULL,
-  DATA BLOB,
-  ENCODING VARCHAR(20),
-  FOREIGN KEY (ID) REFERENCES ACCESS_RESULT (ID)
-);
-
-/**********************************/
-/* Table Name: URL Filter */
-/**********************************/
-CREATE TABLE URL_FILTER(
-  ID IDENTITY NOT NULL PRIMARY KEY,
-  SESSION_ID VARCHAR(20) NOT NULL,
-  URL VARCHAR(65536) NOT NULL,
-  FILTER_TYPE VARCHAR(1) NOT NULL,
-  CREATE_TIME TIMESTAMP NOT NULL
-);
-
-
-CREATE INDEX IDX_URL_QUEUE_SESSION_ID_AND_TIME ON URL_QUEUE (SESSION_ID, CREATE_TIME);
-CREATE INDEX IDX_URL_QUEUE_SESSION_ID_AND_URL ON URL_QUEUE (SESSION_ID, URL);
-CREATE INDEX IDX_URL_QUEUE_SESSION_ID ON URL_QUEUE (SESSION_ID);
-
-CREATE INDEX IDX_ACCESS_RESULT_SESSION_ID_AND_TIME ON ACCESS_RESULT (SESSION_ID, CREATE_TIME);
-CREATE INDEX IDX_ACCESS_RESULT_SESSION_ID_AND_URL ON ACCESS_RESULT (SESSION_ID, URL);
-CREATE INDEX IDX_ACCESS_RESULT_SESSION_ID ON ACCESS_RESULT (SESSION_ID);
-CREATE INDEX IDX_ACCESS_RESULT_URL_AND_TIME ON ACCESS_RESULT (URL, CREATE_TIME);
-
-CREATE INDEX IDX_URL_FILTER_SESSION_ID_AND_FILTER_TYPE ON URL_FILTER (SESSION_ID, FILTER_TYPE);
-

+ 0 - 0
src/main/h2/webapp/.gitkeep


BIN
src/main/h2/webapp/WEB-INF/db/fess.mv.db


BIN
src/main/h2/webapp/WEB-INF/db/robot.mv.db


+ 0 - 44
src/main/mysql/resources/jdbc.dicon

@@ -1,44 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!DOCTYPE components PUBLIC "-//SEASAR2.1//DTD S2Container//EN"
-	"http://www.seasar.org/dtd/components21.dtd">
-<components namespace="jdbc">
-	<include path="jta.dicon"/>
-
-	<!-- MySQL -->
-	<component name="xaDataSource"
-		class="org.seasar.extension.dbcp.impl.XADataSourceImpl">
-		<property name="driverClassName">
-			"com.mysql.jdbc.Driver"
-		</property>
-		<property name="URL">
-			"jdbc:mysql://localhost:3306/fess_db?noDatetimeStringSync=true&amp;zeroDateTimeBehavior=convertToNull&amp;useUnicode=true&amp;characterEncoding=UTF-8&amp;autoReconnect=true"
-		</property>
-		<property name="user">"fess_user"</property>
-		<property name="password">"fess_pass"</property>
-	</component>
-
-	<component name="connectionPool"
-		class="org.seasar.extension.dbcp.impl.ConnectionPoolImpl">
-		<property name="timeout">600</property>
-		<property name="maxPoolSize">10</property>
-		<property name="allowLocalTx">true</property>
-		<destroyMethod name="close"/>
-	</component>
-
-	<component name="DataSource"
-		class="org.seasar.extension.dbcp.impl.DataSourceImpl"
-	/>
-
-	<!-- from JNDI -->
-	<!--
-	<component name="DataSource"
-		class="javax.sql.DataSource">
-		@org.seasar.extension.j2ee.JndiResourceLocator@lookup("java:comp/env/jdbc/DataSource")
-	</component>
-	-->
-
-	<!--
-	<component name="dataSource"
-		class="org.seasar.extension.datasource.impl.SelectableDataSourceProxy"/>
-	-->
-</components>

+ 0 - 37
src/main/mysql/resources/s2robot_jdbc.dicon

@@ -1,37 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!DOCTYPE components PUBLIC "-//SEASAR2.1//DTD S2Container//EN"
-    "http://www.seasar.org/dtd/components21.dtd">
-<components namespace="jdbc">
-    <include path="jta.dicon"/>
-
-    <!-- for MySQL -->
-    <component name="xaDataSource"
-        class="org.seasar.extension.dbcp.impl.XADataSourceImpl">
-        <property name="driverClassName">
-            "com.mysql.jdbc.Driver"
-        </property>
-        <property name="URL">
-            "jdbc:mysql://localhost:3306/fess_robot?" +
-            "noDatetimeStringSync=true&amp;" +
-            "zeroDateTimeBehavior=convertToNull&amp;" +
-            "useUnicode=true&amp;characterEncoding=UTF-8&amp;" +
-            "autoReconnect=true"
-        </property>
-        <property name="user">"s2robot"</property>
-        <property name="password">"s2robot"</property>
-    </component>
-
-    <component name="connectionPool"
-        class="org.seasar.extension.dbcp.impl.ConnectionPoolImpl">
-        <property name="timeout">600</property>
-        <property name="maxPoolSize">10</property>
-        <property name="allowLocalTx">true</property>
-        <property name="transactionIsolationLevel">@java.sql.Connection@TRANSACTION_REPEATABLE_READ</property>
-        <destroyMethod name="close"/>
-    </component>
-
-    <component name="DataSource"
-        class="org.seasar.extension.dbcp.impl.DataSourceImpl"
-    />
-
-</components>

+ 0 - 97
src/main/mysql/webapp/WEB-INF/cmd/resources/app.dicon

@@ -1,97 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!DOCTYPE components PUBLIC "-//SEASAR//DTD S2Container 2.4//EN"
-	"http://www.seasar.org/dtd/components24.dtd">
-<components>
-	<include path="convention.dicon" />
-	<include path="aop.dicon" />
-	<include path="dbflute.dicon" />
-	<include path="fess.dicon" />
-
-	<include path="s2robot_db.dicon" />
-
-	<component name="indexingHelper" class="org.codelibs.fess.helper.IndexingHelper">
-	</component>
-	<component name="labelTypeHelper" class="org.codelibs.fess.helper.LabelTypeHelper">
-	</component>
-	<component name="webFsIndexHelper" class="org.codelibs.fess.helper.WebFsIndexHelper">
-	</component>
-	<component name="dataIndexHelper" class="org.codelibs.fess.helper.DataIndexHelper">
-	</component>
-	<component name="overlappingHostHelper" class="org.codelibs.fess.helper.OverlappingHostHelper">
-	</component>
-	<component name="intervalControlHelper" class="org.codelibs.fess.helper.IntervalControlHelper">
-		<!-- 
-		<initMethod name="addIntervalRule">
-			<arg>"5:00"</arg>
-			<arg>"10:00"</arg>
-			<arg>"2,3,4,5,6"</arg>
-			<arg>3600000</arg>
-		</initMethod>
-		 -->
-	</component>
-	<component name="sambaHelper" class="org.codelibs.fess.helper.SambaHelper">
-	</component>
-	<component name="indexUpdater" class="org.codelibs.fess.solr.IndexUpdater"
-		instance="prototype">
-		<!-- 
-		<property name="maxDocumentCacheSize">5</property>
-		<property name="unprocessedDocumentSize">100</property>
-		<property name="threadDump">false</property>
-		<initMethod name="addBoostDocumentRule">
-			<arg>
-				<component class="org.codelibs.fess.solr.BoostDocumentRule">
-					<property name="matchExpression">"url.matches(\".*fess.*\")"</property>
-					<property name="boostExpression">"1000.0"</property>
-				</component>
-			</arg>
-		</initMethod>
-		<initMethod name="addDefaultDocValue">
-			<arg>"FieldName"</arg>
-			<arg>"VALUE"</arg>
-		</initMethod>
-		 -->
-	</component>
-	<component name="fileTypeHelper" class="org.codelibs.fess.helper.FileTypeHelper">
-		<initMethod name="add">
-			<arg>"text/html"</arg>
-			<arg>"html"</arg>
-		</initMethod>
-		<initMethod name="add">
-			<arg>"application/msword"</arg>
-			<arg>"word"</arg>
-		</initMethod>
-		<initMethod name="add">
-			<arg>"application/vnd.openxmlformats-officedocument.wordprocessingml.document"</arg>
-			<arg>"word"</arg>
-		</initMethod>
-		<initMethod name="add">
-			<arg>"application/vnd.ms-excel"</arg>
-			<arg>"excel"</arg>
-		</initMethod>
-		<initMethod name="add">
-			<arg>"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"</arg>
-			<arg>"excel"</arg>
-		</initMethod>
-		<initMethod name="add">
-			<arg>"application/vnd.ms-powerpoint"</arg>
-			<arg>"powerpoint"</arg>
-		</initMethod>
-		<initMethod name="add">
-			<arg>"application/vnd.openxmlformats-officedocument.presentationml.presentation"</arg>
-			<arg>"powerpoint"</arg>
-		</initMethod>
-		<initMethod name="add">
-			<arg>"application/pdf"</arg>
-			<arg>"pdf"</arg>
-		</initMethod>
-	</component>
-	<component name="databaseHelper" class="org.codelibs.fess.helper.impl.DatabaseHelperImpl">
-		<aspect pointcut="optimize">
-			j2ee.requiresNewTx
-		</aspect>
-	</component>
-	<component name="fessCrawler" class="org.codelibs.fess.exec.Crawler"
-		instance="prototype">
-	</component>
-
-</components>

+ 0 - 338
src/main/mysql/webapp/WEB-INF/web.xml

@@ -1,338 +0,0 @@
-<?xml version="1.0"?>
-<!--
- * Copyright 2004-2006 the Seasar Foundation and the Others.
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- *      http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
- * either express or implied. See the License for the specific language
- * governing permissions and limitations under the License.
--->
-<web-app
-    xmlns="http://java.sun.com/xml/ns/j2ee"
-    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-    xsi:schemaLocation="http://java.sun.com/xml/ns/j2ee http://java.sun.com/xml/ns/j2ee/web-app_2_4.xsd"
-    version="2.4">
-
-  <context-param>
-    <param-name>sastruts.VIEW_PREFIX</param-name>
-    <param-value>/WEB-INF/view</param-value>
-  </context-param>
-
-  <filter>
-    <filter-name>encodingfilter</filter-name>
-    <filter-class>org.codelibs.fess.filter.FessEncodingFilter</filter-class>
-    <init-param>
-      <param-name>encoding</param-name>
-      <param-value>UTF-8</param-value>
-    </init-param>
-    <init-param>
-      <param-name>encodingRules</param-name>
-      <param-value>sjis:Shift_JIS,eucjp:EUC-JP</param-value>
-    </init-param>
-  </filter>
-
-  <filter>
-    <filter-name>s2filter</filter-name>
-    <filter-class>org.seasar.framework.container.filter.S2ContainerFilter</filter-class>
-  </filter>
-
-  <filter>
-    <filter-name>hotdeployfilter</filter-name>
-    <filter-class>org.seasar.framework.container.hotdeploy.HotdeployFilter</filter-class>
-  </filter>
-
-  <filter>
-    <filter-name>authenticationFilter</filter-name>
-    <filter-class>org.codelibs.fess.filter.AdminAuthFilter</filter-class>
-    <init-param>
-      <param-name>urlPatterns</param-name>
-      <param-value>/fess/admin.*</param-value>
-    </init-param>
-    <init-param>
-      <param-name>loginPath</param-name>
-      <param-value>/fess/login/</param-value>
-    </init-param>
-    <init-param>
-      <param-name>cipherName</param-name>
-      <param-value>authenticationCipher</param-value>
-    </init-param>
-  </filter>
-
-  <filter>
-    <filter-name>webApiFilter</filter-name>
-    <filter-class>org.codelibs.fess.filter.WebApiFilter</filter-class>
-  </filter>
-
-  <filter>
-    <filter-name>routingfilter</filter-name>
-    <filter-class>org.seasar.struts.filter.RoutingFilter</filter-class>
-    <init-param>
-      <param-name>jspDirectAccess</param-name>
-      <param-value>false</param-value>
-    </init-param>
-  </filter>
-
-  <!--
-  <filter>
-    <filter-name>adLoginInfoFilter</filter-name>
-    <filter-class>org.codelibs.fess.filter.AdLoginInfoFilter</filter-class>
-    <init-param>
-      <param-name>redirectLoginError</param-name>
-      <param-value>true</param-value>
-    </init-param>
-    <init-param>
-      <param-name>useTestUser</param-name>
-      <param-value>false</param-value>
-    </init-param>
-    <init-param>
-      <param-name>testUserName</param-name>
-      <param-value>testUser</param-value>
-    </init-param>
-  </filter>
-  -->
-
-  <!-- SPNEGO    -->
-  <!--
-  <filter>
-      <filter-name>SpnegoHttpFilter</filter-name>
-      <filter-class>net.sourceforge.spnego.SpnegoHttpFilter</filter-class>
-
-      <init-param>
-          <param-name>spnego.allow.basic</param-name>
-          <param-value>false</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.allow.localhost</param-name>
-          <param-value>true</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.allow.unsecure.basic</param-name>
-          <param-value>false</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.login.client.module</param-name>
-          <param-value>spnego-client</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.krb5.conf</param-name>
-          <param-value>krb5.conf</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.login.conf</param-name>
-          <param-value>login.conf</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.preauth.username</param-name>
-          <param-value>knldguser</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.preauth.password</param-name>
-          <param-value>zaq12wsx</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.login.server.module</param-name>
-          <param-value>spnego-server</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.prompt.ntlm</param-name>
-          <param-value>false</param-value>
-      </init-param>
-
-      <init-param>
-          <param-name>spnego.logger.level</param-name>
-          <param-value>1</param-value>
-      </init-param>
-  </filter>
--->
-
-  <filter-mapping>
-    <filter-name>encodingfilter</filter-name>
-    <url-pattern>/*</url-pattern>
-    <dispatcher>REQUEST</dispatcher>
-    <dispatcher>FORWARD</dispatcher>
-  </filter-mapping>
-
-  <filter-mapping>
-    <filter-name>s2filter</filter-name>
-    <url-pattern>/*</url-pattern>
-    <dispatcher>REQUEST</dispatcher>
-    <dispatcher>FORWARD</dispatcher>
-    <dispatcher>INCLUDE</dispatcher>
-  </filter-mapping>
-
-  <filter-mapping>
-    <filter-name>hotdeployfilter</filter-name>
-    <url-pattern>/*</url-pattern>
-    <dispatcher>REQUEST</dispatcher>
-    <dispatcher>FORWARD</dispatcher>
-    <dispatcher>INCLUDE</dispatcher>
-  </filter-mapping>
-
-  <filter-mapping>
-    <filter-name>authenticationFilter</filter-name>
-    <url-pattern>/*</url-pattern>
-  </filter-mapping>
-
-  <filter-mapping>
-    <filter-name>webApiFilter</filter-name>
-    <url-pattern>/*</url-pattern>
-    <dispatcher>REQUEST</dispatcher>
-  </filter-mapping>
-
-  <filter-mapping>
-    <filter-name>routingfilter</filter-name>
-    <url-pattern>/*</url-pattern>
-    <dispatcher>REQUEST</dispatcher>
-  </filter-mapping>
-
-  <!--
-  <filter-mapping>
-      <filter-name>SpnegoHttpFilter</filter-name>
-      <url-pattern>/*</url-pattern>
-  </filter-mapping>
-  <filter-mapping>
-    <filter-name>adLoginInfoFilter</filter-name>
-    <url-pattern>*.do</url-pattern>
-    <dispatcher>REQUEST</dispatcher>
-    <dispatcher>FORWARD</dispatcher>
-  </filter-mapping>
--->
-
-  <servlet>
-    <servlet-name>action</servlet-name>
-    <servlet-class>org.codelibs.sastruts.core.servlet.SSCActionServlet</servlet-class>
-    <init-param>
-      <param-name>config</param-name>
-      <param-value>/WEB-INF/struts-config.xml</param-value>
-    </init-param>
-    <init-param>
-      <param-name>configFactory</param-name>
-      <param-value>org.seasar.struts.config.S2ModuleConfigFactory</param-value>
-    </init-param>
-    <load-on-startup>1</load-on-startup>
-  </servlet>
-
-  <!--
-  <servlet>
-    <servlet-name>h2ConfigServlet</servlet-name>
-    <servlet-class>org.codelibs.fess.servlet.H2ConfigServlet</servlet-class>
-    <init-param>
-      <param-name>tcpPort</param-name>
-      <param-value>19092</param-value>
-    </init-param>
-    <load-on-startup>2</load-on-startup>
-  </servlet>
-  -->
-
-  <servlet>
-    <servlet-name>tomcat7ConfigServlet</servlet-name>
-    <servlet-class>org.codelibs.fess.servlet.Tomcat7ConfigServlet</servlet-class>
-    <load-on-startup>3</load-on-startup>
-  </servlet>
-
-  <servlet>
-    <servlet-name>s2container</servlet-name>
-    <servlet-class>org.seasar.framework.container.servlet.S2ContainerServlet</servlet-class>
-    <load-on-startup>4</load-on-startup>
-  </servlet>
-
-  <servlet-mapping>
-    <servlet-name>action</servlet-name>
-    <url-pattern>*.do</url-pattern>
-  </servlet-mapping>
-
-<!--
-  <servlet-mapping>
-    <servlet-name>s2container</servlet-name>
-    <url-pattern>/s2container</url-pattern>
-  </servlet-mapping>
--->
-
-  <listener>
-  	<listener-class>org.apache.commons.fileupload.servlet.FileCleanerCleanup</listener-class>
-  </listener>
-
-<!--
-    <session-config>
-        <session-timeout>1</session-timeout>
-    </session-config>
-	-->
-
-<!--
-    <welcome-file-list>
-    </welcome-file-list>
-	-->
-
-  <jsp-config>
-    <jsp-property-group>
-      <url-pattern>*.jsp</url-pattern>
-      <el-ignored>false</el-ignored>
-      <page-encoding>UTF-8</page-encoding>
-      <scripting-invalid>false</scripting-invalid>
-      <include-prelude>/WEB-INF/view/common/common.jsp</include-prelude>
-    </jsp-property-group>
-  </jsp-config>
-
-  <security-constraint>
-    <web-resource-collection>
-      <web-resource-name>Fess Authentication</web-resource-name>
-      <url-pattern>/login/login</url-pattern>
-    </web-resource-collection>
-    <auth-constraint>
-      <role-name>fess</role-name>
-      <!--
-      <role-name>role1</role-name>
-      -->
-    </auth-constraint>
-  </security-constraint>
-
-  <security-role>
-    <role-name>fess</role-name>
-  </security-role>
-
-  <login-config>
-    <auth-method>FORM</auth-method>
-    <form-login-config>
-      <form-login-page>/WEB-INF/view/login/index.jsp</form-login-page>
-      <form-error-page>/WEB-INF/view/login/error.jsp</form-error-page>
-    </form-login-config>
-  </login-config>
-
-  <error-page>
-  	<error-code>400</error-code>
-  	<location>/WEB-INF/view/error/redirect.jsp?type=badRequest</location>
-  </error-page>
-  <error-page>
-  	<error-code>403</error-code>
-  	<location>/WEB-INF/view/error/redirect.jsp?type=logOut</location>
-  </error-page>
-  <error-page>
-  	<error-code>404</error-code>
-  	<location>/WEB-INF/view/error/redirect.jsp?type=notFound</location>
-  </error-page>
-  <error-page>
-  	<error-code>408</error-code>
-  	<location>/WEB-INF/view/error/redirect.jsp?type=logOut</location>
-  </error-page>
-  <error-page>
-  	<error-code>500</error-code>
-  	<location>/WEB-INF/view/error/redirect.jsp?type=systemError</location>
-  </error-page>
-
-</web-app>