Print

Print


Author: [log in to unmask]
Date: Thu Sep 24 13:05:19 2015
New Revision: 3699

Log:
Cleanup crawler package.

Removed:
    java/trunk/crawler/src/main/java/org/hps/crawler/Crawler.java
    java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerFeature.java
    java/trunk/crawler/src/main/java/org/hps/crawler/JCacheManager.java
Modified:
    java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java
    java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java

Modified: java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java
 =============================================================================
--- java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java	(original)
+++ java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java	Thu Sep 24 13:05:19 2015
@@ -3,7 +3,6 @@
 import java.io.File;
 import java.text.ParseException;
 import java.text.SimpleDateFormat;
-import java.util.ArrayList;
 import java.util.Arrays;
 import java.util.Date;
 import java.util.HashSet;
@@ -14,7 +13,6 @@
 import org.hps.conditions.database.ConnectionParameters;
 import org.hps.datacat.client.DatasetFileFormat;
 import org.hps.datacat.client.DatasetSite;
-import org.hps.record.evio.EvioEventProcessor;
 
 /**
  * Full configuration information for the {@link Crawler} class.
@@ -37,11 +35,6 @@
     private Set<Integer> acceptRuns = new LinkedHashSet<Integer>();
 
     /**
-     * <code>true</code> if database updates are allowed meaning existing records can be deleted and replaced.
-     */
-    private boolean allowUpdates = false;
-
-    /**
      * The database connection parameters which must be provided by a command line argument.
      */
     private ConnectionParameters connectionParameters;
@@ -59,11 +52,6 @@
     private boolean enableMetadata;
 
     /**
-     * Set of features enabled in this configuration.
-     */
-    Set<CrawlerFeature> features = new HashSet<CrawlerFeature>();
-
-    /**
      * Set of file formats for filtering files.
      */
     Set<DatasetFileFormat> formats = new HashSet<DatasetFileFormat>();
@@ -74,16 +62,6 @@
     private Integer maxDepth = Integer.MAX_VALUE;
 
     /**
-     * The maximum number of files to accept (just used for debugging purposes).
-     */
-    private int maxFiles = -1;
-
-    /**
-     * A list of extra {@link org.hps.record.evio.EvioEventProcessor}s to run with the job.
-     */
-    private final List<EvioEventProcessor> processors = new ArrayList<EvioEventProcessor>();
-
-    /**
      * The root directory to search for files, which defaults to the current directory.
      */
     private File rootDir = new File(System.getProperty("user.dir"));
@@ -102,26 +80,6 @@
      * A file to use for getting the timestamp date.
      */
     private File timestampFile = null;
-
-    /**
-     * <code>true</code> if the data catalog should be updated (off by default).
-     */
-    private boolean updateDatacat = false;
-
-    /**
-     * <code>true</code> if the run database should be updated from results of the job.
-     */
-    private boolean updateRunLog = false;
-
-    /**
-     * <code>true</code> if file caching should be used to move files to the cache disk from tape at JLAB.
-     */
-    private boolean useFileCache = false;
-
-    /**
-     * The maximum wait time in milliseconds to allow for file caching operations.
-     */
-    private Long waitTime;
 
     /**
      * Get the set of runs that will be accepted for the job.
@@ -130,15 +88,6 @@
      */
     Set<Integer> acceptRuns() {
         return acceptRuns;
-    }
-
-    /**
-     * Add the default set of features.
-     */
-    CrawlerConfig addDefaultFeatures() {
-        final List<CrawlerFeature> defaultFeatures = Arrays.asList(CrawlerFeature.values());
-        this.features.addAll(defaultFeatures);
-        return this;
     }
 
     /**
@@ -151,16 +100,6 @@
     }
 
     /**
-     * Add a feature to enable it.
-     *
-     * @return this object
-     */
-    CrawlerConfig addFeature(final CrawlerFeature feature) {
-        this.features.add(feature);
-        return this;
-    }
-
-    /**
      * Add a file format for filtering.
      *
      * @param format the file format
@@ -169,42 +108,7 @@
         this.formats.add(format);
         return this;
     }
-
-    /**
-     * Add an {@link org.hps.record.evio.EvioEventProcessor} to the job.
-     *
-     * @param processor
-     * @return this object
-     */
-    CrawlerConfig addProcessor(final EvioEventProcessor processor) {
-        this.processors.add(processor);
-        return this;
-    }
-
-    /**
-     * Add an {@link org.hps.record.evio.EvioEventProcessor} to the job by its class name.
-     *
-     * @param processor the <code>EvioEventProcessor</code> to instantiate
-     * @return this object
-     */
-    CrawlerConfig addProcessor(final String className) {
-        try {
-            this.processors.add(EvioEventProcessor.class.cast(Class.forName(className).newInstance()));
-        } catch (final Exception e) {
-            throw new RuntimeException("Error creating EvioEventProcessor with type: " + className, e);
-        }
-        return this;
-    }
-
-    /**
-     * Return <code>true</code> if updates/deletions of existing records in the database is allowed.
-     *
-     * @return <code>true</code> if updating/deleting records in the database is allowed
-     */
-    boolean allowUpdates() {
-        return allowUpdates;
-    }
-
+    
     /**
      * Get the database connection parameters.
      *
@@ -242,15 +146,6 @@
     }
 
     /**
-     * Get the set of enabled features.
-     *
-     * @return the set of enabled features
-     */
-    Set<CrawlerFeature> getFeatures() {
-        return this.features;
-    }
-
-    /**
      * Get the file formats for filtering.
      *
      * @return the file formats for filtering
@@ -267,37 +162,7 @@
     Integer maxDepth() {
         return maxDepth;
     }
-
-    /**
-     * Get the maximum number of files that the job can process.
-     *
-     * @return the maximum number of files
-     */
-    int maxFiles() {
-        return maxFiles;
-    }
-
-    /**
-     * Get the list of extra event processors that will run with the job.
-     * <p>
-     * Required (default) processors for the job are not included here.
-     *
-     * @return the list of extra event processors
-     */
-    List<EvioEventProcessor> processors() {
-        return processors;
-    }
-
-    /**
-     * Remove a feature to disable it.
-     *
-     * @return this object
-     */
-    CrawlerConfig removeFeature(final CrawlerFeature feature) {
-        this.features.remove(feature);
-        return this;
-    }
-
+  
     /**
      * Get the root directory for the file search.
      *
@@ -319,17 +184,6 @@
     }
 
     /**
-     * Set whether database updates are allowed, i.e. replacement of existing records.
-     *
-     * @param allowUpdates <code>true</code> to allow database record deletion/updates
-     * @return this object
-     */
-    CrawlerConfig setAllowUpdates(final boolean allowUpdates) {
-        this.allowUpdates = allowUpdates;
-        return this;
-    }
-
-    /**
      * Set the database connection parameters.
      *
      * @param connectionParameters the database connection parameters
@@ -377,19 +231,6 @@
      */
     CrawlerConfig setMaxDepth(final Integer maxDepth) {
         this.maxDepth = maxDepth;
-        return this;
-    }
-
-    /**
-     * Set the maximum number of files that will be processed by the job.
-     * <p>
-     * This should only be used for debugging purposes as it results in incorrect event counts for the run.
-     *
-     * @param maxFiles the maximum number of files to process or -1 for unlimited
-     * @return this object
-     */
-    CrawlerConfig setMaxFiles(final int maxFiles) {
-        this.maxFiles = maxFiles;
         return this;
     }
 
@@ -443,56 +284,6 @@
     }
 
     /**
-     * Set to <code>true</code> to update data catalog.
-     *
-     * @param updateDatacat <code>true</code> to update data catalog
-     */
-    CrawlerConfig setUpdateDatacat(final boolean updateDatacat) {
-        this.updateDatacat = updateDatacat;
-        return this;
-    }
-
-    /**
-     * Set whether the run database should be updated in the job.
-     * <p>
-     * This will not allow replacement of existing run log records. The {@link #allowUpdates()} flag must be on for this
-     * be allowed.
-     *
-     * @param updateRunLog <code>true</code> if the run database should be updated
-     * @return this object
-     */
-    CrawlerConfig setUpdateRunLog(final boolean updateRunLog) {
-        this.updateRunLog = updateRunLog;
-        return this;
-    }
-
-    /**
-     * Set whether file caching using the 'jcache' program should be enabled.
-     * <p>
-     * This is only relevant for jobs run at JLAB.
-     *
-     * @param useFileCache <code>true</code> to allow file caching
-     * @return this object
-     */
-    CrawlerConfig setUseFileCache(final boolean useFileCache) {
-        this.useFileCache = useFileCache;
-        return this;
-    }
-
-    /**
-     * Set the max wait time in seconds for all file caching operations to complete.
-     * <p>
-     * If this time is exceeded then the job will fail with an error.
-     *
-     * @param waitTime the max wait time in seconds allowed for file caching to complete
-     * @return this object
-     */
-    CrawlerConfig setWaitTime(final long waitTime) {
-        this.waitTime = waitTime;
-        return this;
-    }
-
-    /**
      * Get the timestamp for file filtering.
      * <p>
      * Files older than this will not be included in the job.
@@ -511,40 +302,4 @@
     File timestampFile() {
         return timestampFile;
     }
-
-    /**
-     * Get whether data catalog should be updated or not.
-     *
-     * @return <code>true</code> if data catalog should be update
-     */
-    boolean updateDatacat() {
-        return this.updateDatacat;
-    }
-
-    /**
-     * Return <code>true</code> if the run database should be updated.
-     *
-     * @return <code>true</code> if the run database should be updated
-     */
-    boolean updateRunDatabase() {
-        return updateRunLog;
-    }
-
-    /**
-     * Return <code>true</code> if file caching should be enabled.
-     *
-     * @return <code>true</code> if file caching should be enabled
-     */
-    boolean useFileCache() {
-        return useFileCache;
-    }
-
-    /**
-     * Get the max wait time in seconds to allow for file caching operations to complete.
-     *
-     * @return the max wait time in seconds to allow for file caching operations to complete
-     */
-    Long waitTime() {
-        return waitTime;
-    }
 }

Modified: java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java
 =============================================================================
--- java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java	(original)
+++ java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java	Thu Sep 24 13:05:19 2015
@@ -116,7 +116,7 @@
     /**
      * Setup the logger.
      */
-    private static final Logger LOGGER = LogUtil.create(Crawler.class, new DefaultLogFormatter(), Level.CONFIG);
+    private static final Logger LOGGER = LogUtil.create(DatacatCrawler.class, new DefaultLogFormatter(), Level.CONFIG);
 
     /**
      * Command line options for the crawler.