LISTSERV mailing list manager LISTSERV 16.5

Help for HPS-SVN Archives


HPS-SVN Archives

HPS-SVN Archives


HPS-SVN@LISTSERV.SLAC.STANFORD.EDU


View:

Message:

[

First

|

Previous

|

Next

|

Last

]

By Topic:

[

First

|

Previous

|

Next

|

Last

]

By Author:

[

First

|

Previous

|

Next

|

Last

]

Font:

Proportional Font

LISTSERV Archives

LISTSERV Archives

HPS-SVN Home

HPS-SVN Home

HPS-SVN  September 2015

HPS-SVN September 2015

Subject:

r3699 - in /java/trunk/crawler/src/main/java/org/hps/crawler: Crawler.java CrawlerConfig.java CrawlerFeature.java DatacatCrawler.java JCacheManager.java

From:

[log in to unmask]

Reply-To:

Notification of commits to the hps svn repository <[log in to unmask]>

Date:

Thu, 24 Sep 2015 20:05:23 -0000

Content-Type:

text/plain

Parts/Attachments:

Parts/Attachments

text/plain (386 lines)

Author: [log in to unmask]
Date: Thu Sep 24 13:05:19 2015
New Revision: 3699

Log:
Cleanup crawler package.

Removed:
    java/trunk/crawler/src/main/java/org/hps/crawler/Crawler.java
    java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerFeature.java
    java/trunk/crawler/src/main/java/org/hps/crawler/JCacheManager.java
Modified:
    java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java
    java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java

Modified: java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java
 =============================================================================
--- java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java	(original)
+++ java/trunk/crawler/src/main/java/org/hps/crawler/CrawlerConfig.java	Thu Sep 24 13:05:19 2015
@@ -3,7 +3,6 @@
 import java.io.File;
 import java.text.ParseException;
 import java.text.SimpleDateFormat;
-import java.util.ArrayList;
 import java.util.Arrays;
 import java.util.Date;
 import java.util.HashSet;
@@ -14,7 +13,6 @@
 import org.hps.conditions.database.ConnectionParameters;
 import org.hps.datacat.client.DatasetFileFormat;
 import org.hps.datacat.client.DatasetSite;
-import org.hps.record.evio.EvioEventProcessor;
 
 /**
  * Full configuration information for the {@link Crawler} class.
@@ -37,11 +35,6 @@
     private Set<Integer> acceptRuns = new LinkedHashSet<Integer>();
 
     /**
-     * <code>true</code> if database updates are allowed meaning existing records can be deleted and replaced.
-     */
-    private boolean allowUpdates = false;
-
-    /**
      * The database connection parameters which must be provided by a command line argument.
      */
     private ConnectionParameters connectionParameters;
@@ -59,11 +52,6 @@
     private boolean enableMetadata;
 
     /**
-     * Set of features enabled in this configuration.
-     */
-    Set<CrawlerFeature> features = new HashSet<CrawlerFeature>();
-
-    /**
      * Set of file formats for filtering files.
      */
     Set<DatasetFileFormat> formats = new HashSet<DatasetFileFormat>();
@@ -74,16 +62,6 @@
     private Integer maxDepth = Integer.MAX_VALUE;
 
     /**
-     * The maximum number of files to accept (just used for debugging purposes).
-     */
-    private int maxFiles = -1;
-
-    /**
-     * A list of extra {@link org.hps.record.evio.EvioEventProcessor}s to run with the job.
-     */
-    private final List<EvioEventProcessor> processors = new ArrayList<EvioEventProcessor>();
-
-    /**
      * The root directory to search for files, which defaults to the current directory.
      */
     private File rootDir = new File(System.getProperty("user.dir"));
@@ -102,26 +80,6 @@
      * A file to use for getting the timestamp date.
      */
     private File timestampFile = null;
-
-    /**
-     * <code>true</code> if the data catalog should be updated (off by default).
-     */
-    private boolean updateDatacat = false;
-
-    /**
-     * <code>true</code> if the run database should be updated from results of the job.
-     */
-    private boolean updateRunLog = false;
-
-    /**
-     * <code>true</code> if file caching should be used to move files to the cache disk from tape at JLAB.
-     */
-    private boolean useFileCache = false;
-
-    /**
-     * The maximum wait time in milliseconds to allow for file caching operations.
-     */
-    private Long waitTime;
 
     /**
      * Get the set of runs that will be accepted for the job.
@@ -130,15 +88,6 @@
      */
     Set<Integer> acceptRuns() {
         return acceptRuns;
-    }
-
-    /**
-     * Add the default set of features.
-     */
-    CrawlerConfig addDefaultFeatures() {
-        final List<CrawlerFeature> defaultFeatures = Arrays.asList(CrawlerFeature.values());
-        this.features.addAll(defaultFeatures);
-        return this;
     }
 
     /**
@@ -151,16 +100,6 @@
     }
 
     /**
-     * Add a feature to enable it.
-     *
-     * @return this object
-     */
-    CrawlerConfig addFeature(final CrawlerFeature feature) {
-        this.features.add(feature);
-        return this;
-    }
-
-    /**
      * Add a file format for filtering.
      *
      * @param format the file format
@@ -169,42 +108,7 @@
         this.formats.add(format);
         return this;
     }
-
-    /**
-     * Add an {@link org.hps.record.evio.EvioEventProcessor} to the job.
-     *
-     * @param processor
-     * @return this object
-     */
-    CrawlerConfig addProcessor(final EvioEventProcessor processor) {
-        this.processors.add(processor);
-        return this;
-    }
-
-    /**
-     * Add an {@link org.hps.record.evio.EvioEventProcessor} to the job by its class name.
-     *
-     * @param processor the <code>EvioEventProcessor</code> to instantiate
-     * @return this object
-     */
-    CrawlerConfig addProcessor(final String className) {
-        try {
-            this.processors.add(EvioEventProcessor.class.cast(Class.forName(className).newInstance()));
-        } catch (final Exception e) {
-            throw new RuntimeException("Error creating EvioEventProcessor with type: " + className, e);
-        }
-        return this;
-    }
-
-    /**
-     * Return <code>true</code> if updates/deletions of existing records in the database is allowed.
-     *
-     * @return <code>true</code> if updating/deleting records in the database is allowed
-     */
-    boolean allowUpdates() {
-        return allowUpdates;
-    }
-
+    
     /**
      * Get the database connection parameters.
      *
@@ -242,15 +146,6 @@
     }
 
     /**
-     * Get the set of enabled features.
-     *
-     * @return the set of enabled features
-     */
-    Set<CrawlerFeature> getFeatures() {
-        return this.features;
-    }
-
-    /**
      * Get the file formats for filtering.
      *
      * @return the file formats for filtering
@@ -267,37 +162,7 @@
     Integer maxDepth() {
         return maxDepth;
     }
-
-    /**
-     * Get the maximum number of files that the job can process.
-     *
-     * @return the maximum number of files
-     */
-    int maxFiles() {
-        return maxFiles;
-    }
-
-    /**
-     * Get the list of extra event processors that will run with the job.
-     * <p>
-     * Required (default) processors for the job are not included here.
-     *
-     * @return the list of extra event processors
-     */
-    List<EvioEventProcessor> processors() {
-        return processors;
-    }
-
-    /**
-     * Remove a feature to disable it.
-     *
-     * @return this object
-     */
-    CrawlerConfig removeFeature(final CrawlerFeature feature) {
-        this.features.remove(feature);
-        return this;
-    }
-
+  
     /**
      * Get the root directory for the file search.
      *
@@ -319,17 +184,6 @@
     }
 
     /**
-     * Set whether database updates are allowed, i.e. replacement of existing records.
-     *
-     * @param allowUpdates <code>true</code> to allow database record deletion/updates
-     * @return this object
-     */
-    CrawlerConfig setAllowUpdates(final boolean allowUpdates) {
-        this.allowUpdates = allowUpdates;
-        return this;
-    }
-
-    /**
      * Set the database connection parameters.
      *
      * @param connectionParameters the database connection parameters
@@ -377,19 +231,6 @@
      */
     CrawlerConfig setMaxDepth(final Integer maxDepth) {
         this.maxDepth = maxDepth;
-        return this;
-    }
-
-    /**
-     * Set the maximum number of files that will be processed by the job.
-     * <p>
-     * This should only be used for debugging purposes as it results in incorrect event counts for the run.
-     *
-     * @param maxFiles the maximum number of files to process or -1 for unlimited
-     * @return this object
-     */
-    CrawlerConfig setMaxFiles(final int maxFiles) {
-        this.maxFiles = maxFiles;
         return this;
     }
 
@@ -443,56 +284,6 @@
     }
 
     /**
-     * Set to <code>true</code> to update data catalog.
-     *
-     * @param updateDatacat <code>true</code> to update data catalog
-     */
-    CrawlerConfig setUpdateDatacat(final boolean updateDatacat) {
-        this.updateDatacat = updateDatacat;
-        return this;
-    }
-
-    /**
-     * Set whether the run database should be updated in the job.
-     * <p>
-     * This will not allow replacement of existing run log records. The {@link #allowUpdates()} flag must be on for this
-     * be allowed.
-     *
-     * @param updateRunLog <code>true</code> if the run database should be updated
-     * @return this object
-     */
-    CrawlerConfig setUpdateRunLog(final boolean updateRunLog) {
-        this.updateRunLog = updateRunLog;
-        return this;
-    }
-
-    /**
-     * Set whether file caching using the 'jcache' program should be enabled.
-     * <p>
-     * This is only relevant for jobs run at JLAB.
-     *
-     * @param useFileCache <code>true</code> to allow file caching
-     * @return this object
-     */
-    CrawlerConfig setUseFileCache(final boolean useFileCache) {
-        this.useFileCache = useFileCache;
-        return this;
-    }
-
-    /**
-     * Set the max wait time in seconds for all file caching operations to complete.
-     * <p>
-     * If this time is exceeded then the job will fail with an error.
-     *
-     * @param waitTime the max wait time in seconds allowed for file caching to complete
-     * @return this object
-     */
-    CrawlerConfig setWaitTime(final long waitTime) {
-        this.waitTime = waitTime;
-        return this;
-    }
-
-    /**
      * Get the timestamp for file filtering.
      * <p>
      * Files older than this will not be included in the job.
@@ -511,40 +302,4 @@
     File timestampFile() {
         return timestampFile;
     }
-
-    /**
-     * Get whether data catalog should be updated or not.
-     *
-     * @return <code>true</code> if data catalog should be update
-     */
-    boolean updateDatacat() {
-        return this.updateDatacat;
-    }
-
-    /**
-     * Return <code>true</code> if the run database should be updated.
-     *
-     * @return <code>true</code> if the run database should be updated
-     */
-    boolean updateRunDatabase() {
-        return updateRunLog;
-    }
-
-    /**
-     * Return <code>true</code> if file caching should be enabled.
-     *
-     * @return <code>true</code> if file caching should be enabled
-     */
-    boolean useFileCache() {
-        return useFileCache;
-    }
-
-    /**
-     * Get the max wait time in seconds to allow for file caching operations to complete.
-     *
-     * @return the max wait time in seconds to allow for file caching operations to complete
-     */
-    Long waitTime() {
-        return waitTime;
-    }
 }

Modified: java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java
 =============================================================================
--- java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java	(original)
+++ java/trunk/crawler/src/main/java/org/hps/crawler/DatacatCrawler.java	Thu Sep 24 13:05:19 2015
@@ -116,7 +116,7 @@
     /**
      * Setup the logger.
      */
-    private static final Logger LOGGER = LogUtil.create(Crawler.class, new DefaultLogFormatter(), Level.CONFIG);
+    private static final Logger LOGGER = LogUtil.create(DatacatCrawler.class, new DefaultLogFormatter(), Level.CONFIG);
 
     /**
      * Command line options for the crawler.

Top of Message | Previous Page | Permalink

Advanced Options


Options

Log In

Log In

Get Password

Get Password


Search Archives

Search Archives


Subscribe or Unsubscribe

Subscribe or Unsubscribe


Archives

November 2017
August 2017
July 2017
January 2017
December 2016
November 2016
October 2016
September 2016
August 2016
July 2016
June 2016
May 2016
April 2016
March 2016
February 2016
January 2016
December 2015
November 2015
October 2015
September 2015
August 2015
July 2015
June 2015
May 2015
April 2015
March 2015
February 2015
January 2015
December 2014
November 2014
October 2014
September 2014
August 2014
July 2014
June 2014
May 2014
April 2014
March 2014
February 2014
January 2014
December 2013
November 2013

ATOM RSS1 RSS2



LISTSERV.SLAC.STANFORD.EDU

Secured by F-Secure Anti-Virus CataList Email List Search Powered by the LISTSERV Email List Manager

Privacy Notice, Security Notice and Terms of Use