[Hps-analysis] Test pass1 Status
Rafayel Paremuzyan
rafopar at jlab.org
Mon May 22 15:25:03 EDT 2017
> ….ARE the occupancies x2 higher???
This was just a naive assumption, also I assumed the computing time
doesn't linearly depend on occupancies,
I was more wondered, that I can not reproduce the code speed that we had
6 months ago.
Rafo
On 05/22/2017 03:19 PM, Graham, Mathew Thomas wrote:
>>
>> 2.3 GeV data recon speed is x2 slower than the 1.05 data recon speed.
>> This is concerning, but maybe there is a good reason for that, i.e.
>> might be hit occupancies are high etc,
>
>
> ….ARE the occupancies x2 higher???
>
>>
>> but the other thing, that seems puzzling, is recon speed for 2.3 GeV
>> data is now more than x2 slower
>> using the same hps-java version, same detector, same java version, as
>> we did for pass0.
>
>
> Ok, that’s legit weird.
>
>
>>
>> Rafo
>>
>>
>> On 05/22/2017 03:10 PM, Graham, Mathew Thomas wrote:
>>>
>>> Sorry to ask a dumb question…but I will anyway.
>>>
>>> This is a “2.3 GeV vs 1.05 GeV” issue? Or an hps-java version issue?
>>>
>>>> On May 22, 2017, at 12:05 PM, Omar Moreno <email at omarmoreno.net
>>>> <mailto:email at omarmoreno.net>> wrote:
>>>>
>>>> Jeremy has also profiled the recon before so if Maurik can't do it,
>>>> I'm sure he can.
>>>>
>>>> On May 22, 2017 11:46 AM, "Rafayel Paremuzyan" <rafopar at jlab.org
>>>> <mailto:rafopar at jlab.org>> wrote:
>>>>
>>>> Hi Alessandra and all,
>>>>
>>>> Yes clearly I see that pattern as well,
>>>> going to the faster machine (pumpkin1), as Maurik suggested,
>>>> recon speed increased, but 2016 recon was about x2 slower than
>>>> the 2015 recon speed.
>>>> I also tried to with -Xms=4000m -Xmx=4000m options, and without
>>>> that options,
>>>> but I didn't notice noticeable speed improvement.
>>>>
>>>> Another thing that is *concering* is I run
>>>> same jar on batch farms, centos6. and centos7, and
>>>> can't get same speed as we got during the pass0.
>>>> Note, jar is the same, detector is the same, run number and
>>>> file number is the same.
>>>> Difference is the time when it started to tun (Oct 19 2016 vs
>>>> May 21 2017)
>>>>
>>>> If someone is interested to look at job log files
>>>> The log file that I run yesterday:
>>>> /lustre/expphy/work/hallb/hps/data/physrun2016/tpass1/logs/hps_008054.27_pass0XML.err
>>>> The log file that run on Oct 19 2016 for pass0
>>>> /lustre/expphy/work/hallb/hps/data/physrun2016/pass0/logs/hps_008054.27_R3.9.err
>>>>
>>>> I have never done profiling with java.
>>>> I know Maurik is in workshop, he might not have time to do this,
>>>> if someone else is set up doing this, it will probably be useful.
>>>>
>>>> Also, I noticed /group/hps is full
>>>>
>>>> Has anyone recently put some data there?
>>>>
>>>> Rafo
>>>>
>>>>
>>>>
>>>> On 05/22/2017 04:04 AM, Alessandra Filippi wrote:
>>>>
>>>> Hi Maurik, Rafo and all,
>>>> I think that the different Java VM could impact on the
>>>> absolute speed comparing different machines and jar
>>>> versions compiled in different times... but I am running
>>>> the same jar (hps-java 3.10, compiled afresh at slac on
>>>> rhel6-64, jvm 1.7.0) on 2015 and 2016 data, and the factor
>>>> ~2 of speed decrease for newer data is striking (whichever
>>>> geometry version).
>>>> About garbage collection, I don't use any flag in
>>>> compilation, so it acts in default mode in both cases.
>>>>
>>>> Are 2016 data affected by more noisy hits that could extend
>>>> the reconstruction time over testing all different
>>>> strategies, to your knowledge?
>>>> cheers
>>>> Alessandra
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Sun, 21 May 2017, Maurik Holtrop wrote:
>>>>
>>>> Hello Rafo,
>>>> One thing that probably is different between the last
>>>> time we ran with the 3.8 jar and now
>>>> is a different version of the Java VM. It could well be
>>>> that the newer version of Java is
>>>> not faster. Also, it is tricky to compare Endeavour
>>>> with the Jlab farm computers. They are
>>>> probably not equivalent in speed. At UNH, Pumpkin has
>>>> the more modern processors, whereas
>>>> Endeavour is now ~5 years old.
>>>>
>>>> Best,
>>>> Maurik
>>>>
>>>> On May 21, 2017, at 6:54 AM, Rafayel Paremuzyan
>>>> <rafopar at jlab.org <mailto:rafopar at jlab.org>> wrote:
>>>>
>>>> Hi Alessandra, Norman, all
>>>>
>>>> thank you for replay and your tests.
>>>>
>>>> I tested both 2015 and 2016 data using v4-4 detector on
>>>> UNH computers.
>>>> I have use 3.8 JAR (the jar for 2015 pass6). 3.9 JAR
>>>> (the jar for 2016 pass0 recon),
>>>> and the NEW jar v051717 (the newest jar tag is v051717)
>>>>
>>>> Ok, I also noticed that recon of 2015 data is faster
>>>> that 2016 data.
>>>> Also seems the new jar is 20% slower than the 3.9 jar
>>>> for 2016 data, and about 60%
>>>> slower for 2015 data.
>>>> now recon speed is about 2.55 EV/S for 2015 data, This
>>>> is too slow
>>>> it cause more than 40h for a single file.
>>>>
>>>> Ths is summary of code speeds with different jar files
>>>> V4-4 Detector, UNH (endeavour), 5K events are reconstructed
>>>>
>>>> Events per second
>>>> Events per second
>>>> Events per second
>>>>
>>>> 3.8JAR (2015 recon jar)
>>>> 3.9JAR, 2016 pass0 recon jar
>>>> v051717JAR, jar vor tpass1
>>>> 2015 Data 5772, file 20
>>>> 5.07
>>>> 5.19
>>>> 3.157
>>>> 2016 Data file 25
>>>>
>>>> 3.11
>>>> 2.53
>>>>
>>>> *However* I looked into job Wall times for pass0 recon.
>>>> The recon speed is more than 7.4 Events/sec, which is
>>>> about x3 faster than with the
>>>> new JAR.
>>>>
>>>> I again checked *same 3.9 jar*, which is slower again.
>>>> I don't know why now the code speed is so low!
>>>>
>>>>
>>>> Norman, I have tried the
>>>> "-DdisableSvtAlignmentConstants", but it didn't work
>>>>
>>>> =================The command===============
>>>> java -XX:+UseSerialGC -cp
>>>> hps-distribution-3.11-v051717-bin.jar
>>>> org.hps.evio.EvioToLcio -x
>>>> /org/hps/steering/recon/PhysicsRun2016FullRecon.lcsim -r
>>>> -d HPS-PhysicsRun2016-v5-3-fieldmap_globalAlign -R 7796
>>>> -DoutputFile=out_7796_0
>>>> -DdisableSvtAlignmentConstants hps_007796.evio.25 -n 10000
>>>>
>>>> ============The error backtrache============
>>>> 2017-05-21 00:45:39 [CONFIG] org.hps.evio.EvioToLcio
>>>> parse :: using steering
>>>> resource
>>>> /org/hps/steering/recon/PhysicsRun2016FullRecon.lcsim
>>>> 2017-05-21 00:45:39 [CONFIG] org.hps.evio.EvioToLcio
>>>> parse :: set max events to
>>>> 10000
>>>> 2017-05-21 00:45:48 [INFO] org.hps.rundb.RunManager
>>>> <init> :: ConnectionParameters {
>>>> database: hps_run_db_v2, hostname: hpsdb.jlab.org
>>>> <http://hpsdb.jlab.org/>, password: darkphoton, port: 3306,
>>>> user: hpsuser }
>>>> 2017-05-21 00:45:48 [CONFIG]
>>>> org.lcsim.job.JobControlManager addVariableDefinition
>>>> :: outputFile = out_7796_0
>>>> 2017-05-21 00:45:48 [CONFIG] org.hps.evio.EvioToLcio
>>>> parse :: set steering variable:
>>>> outputFile=out_7796_0
>>>> 2017-05-21 00:45:48 [SEVERE] org.hps.evio.EvioToLcio
>>>> parse :: bad variable format:
>>>> disableSvtAlignmentConstants
>>>> java.lang.IllegalArgumentException: Bad variable format:
>>>> disableSvtAlignmentConstants
>>>> at
>>>> org.hps.evio.EvioToLcio.parse(EvioToLcio.java:393)
>>>> at org.hps.evio.EvioToLcio.main(EvioToLcio.java:97)
>>>>
>>>> Exception in thread "main"
>>>> java.lang.IllegalArgumentException: Bad variable format:
>>>> disableSvtAlignmentConstants
>>>> at
>>>> org.hps.evio.EvioToLcio.parse(EvioToLcio.java:393)
>>>> at org.hps.evio.EvioToLcio.main(EvioToLcio.java:97)
>>>>
>>>> Rafo
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On 05/20/2017 06:17 AM, Alessandra Filippi wrote:
>>>> Hi Rafo, all,
>>>> I also noticed that the reconstruction for 2016
>>>> data is about twice as
>>>> slow as compared to 2015 (whichever geometry and
>>>> reconstruction
>>>> version).
>>>> This happens when I run the aligned geometry as
>>>> well as the "current"
>>>> one (v5.0), and the geometry taken from the db as
>>>> well (the result is
>>>> the same as v5.0). I did not make any test with
>>>> v4.4, though - actually
>>>> as far as svt alignment is concerned it should be
>>>> the same as v5.0. Can
>>>> you please try and make the same short test with
>>>> the newest jar with
>>>> v4.4?
>>>> This happens to me both with hps-java 5.10 and
>>>> 5.11 (not the most
>>>> updated one).
>>>>
>>>> I would be surprised if it could be something
>>>> connected to the
>>>> alignment, unless for some reason new positions
>>>> and harder tracks
>>>> trigger some long loops in the reconstruction.
>>>> But this happens (to me)
>>>> also with the
>>>> standard geometry, so a check with the one used
>>>> with pass0 (that should
>>>> however be equivalent to v5.0) could at least
>>>> help to rule out, or blame
>>>> on, the alignment step.
>>>> Thanks, cheers
>>>> Alessandra
>>>>
>>>>
>>>> ps. make also sure that the correct fieldmap is
>>>> called in all the
>>>> compact files - you never know!
>>>>
>>>>
>>>>
>>>> On Fri, 19 May 2017, Rafayel Paremuzyan wrote:
>>>>
>>>> Hi All,
>>>>
>>>> During the testing the recon for test pass1,
>>>> I noticed the recon time is more than x2
>>>> longer wrt pass0
>>>> recon time.
>>>>
>>>> To demonstrate it
>>>> I submit 3 simple jobs with 10K events to
>>>> reconstruct, with
>>>> new pass1 xml
>>>> file (this has the new jar v051717, and the
>>>> new detector
>>>> HPS-PhysicsRun2016-v5-3-fieldmap_globalAlign),
>>>> and the old pass0 xml file (pass0 jar
>>>> release 3.9, and the
>>>> detector
>>>> HPS-PhysicsRun2016-Nominal-v4-4-fieldmap)
>>>>
>>>> Below is a printout from the jobs with a
>>>> new JAR, v051717,
>>>> the average time
>>>> for 1000 events is more than 7 minutes
>>>> ===================== LOG from the v051717 JAR
>>>> ==============================
>>>> 2017-05-19 09:36:51 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10614074 with sequence 0
>>>> 2017-05-19 09:43:13 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10615074 with sequence 1000
>>>> 2017-05-19 09:49:18 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10616074 with sequence 2000
>>>> 2017-05-19 09:55:54 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10617074 with sequence 3000
>>>> 2017-05-19 10:02:55 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10618074 with sequence 4000
>>>> 2017-05-19 10:09:57 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10619074 with sequence 5000
>>>> 2017-05-19 10:16:13 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10620074 with sequence 6000
>>>> 2017-05-19 10:25:20 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10621074 with sequence 7000
>>>> 2017-05-19 10:32:56 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10622074 with sequence 8000
>>>> 2017-05-19 10:36:19 [WARNING]
>>>> org.hps.recon.tracking.TrackerReconDriver
>>>> process :: Discarding track with bad
>>>> HelicalTrackHit
>>>> (correction distance
>>>> 0.000000, chisq penalty 0.000000)
>>>> 2017-05-19 10:42:03 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10623074 with sequence 9000
>>>> 2017-05-19 10:47:44 [INFO]
>>>> org.hps.evio.EvioToLcio run ::
>>>> maxEvents 10000
>>>> was reached
>>>> 2017-05-19 10:47:44 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> endOfData ::
>>>> 10000 events processed in job.
>>>> 2017-05-19 10:47:44 [INFO]
>>>> org.hps.evio.EvioToLcio run ::
>>>> Job finished
>>>> successfully!
>>>>
>>>>
>>>> And below is the Job log info from the
>>>> pass0 jar. The
>>>> average time for 1000
>>>> events is less than 3 minutes
>>>> ===================== LOG from the 3.9
>>>> release JAR
>>>> ==============================
>>>> 2017-05-19 13:19:46 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10614074 with sequence 0
>>>> 2017-05-19 13:23:36 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10615074 with sequence 1000
>>>> 2017-05-19 13:27:03 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10616074 with sequence 2000
>>>> 2017-05-19 13:30:40 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10617074 with sequence 3000
>>>> 2017-05-19 13:34:20 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10618074 with sequence 4000
>>>> 2017-05-19 13:38:11 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10619074 with sequence 5000
>>>> 2017-05-19 13:41:43 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10620074 with sequence 6000
>>>> 2017-05-19 13:45:54 [WARNING]
>>>> org.hps.recon.tracking.TrackerReconDriver
>>>> process :: Discarding track with bad
>>>> HelicalTrackHit
>>>> (correction distance
>>>> 0.000000, chisq penalty 0.000000)
>>>> 2017-05-19 13:46:05 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10621074 with sequence 7000
>>>> 2017-05-19 13:50:08 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10622074 with sequence 8000
>>>> 2017-05-19 13:55:03 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> process :: Event
>>>> 10623074 with sequence 9000
>>>> 2017-05-19 13:58:27 [INFO]
>>>> org.hps.evio.EvioToLcio run ::
>>>> maxEvents 10000
>>>> was reached
>>>> 2017-05-19 13:58:27 [INFO]
>>>> org.lcsim.job.EventMarkerDriver
>>>> endOfData ::
>>>> 10000 events processed in job.
>>>> 2017-05-19 13:58:27 [INFO]
>>>> org.hps.evio.EvioToLcio run ::
>>>> Job finished
>>>> successfully!
>>>>
>>>> I also tried to do reconstruction by myself
>>>> interactively,
>>>> but I am getting
>>>> error below,
>>>>
>>>> The command
>>>> /apps/scicomp/java/jdk1.7/bin/java -XX:+UseSerialGC -cp
>>>> hps-distribution-3.9-bin.jar org.hps.evio.EvioToLcio -x
>>>> /org/hps/steering/recon/PhysicsRun2016FullRecon.lcsim -r -d
>>>> HPS-PhysicsRun2016-v5-3-fieldmap_globalAlign -R 7796
>>>> -DoutputFile=out_7796_0
>>>> hps_007796.evio.0 -n 10000
>>>>
>>>> The Error traceback
>>>> 017-05-19 14:58:44 [CONFIG]
>>>> org.hps.evio.EvioToLcio parse ::
>>>> using steering
>>>> resource
>>>> /org/hps/steering/recon/PhysicsRun2016FullRecon.lcsim
>>>> 2017-05-19 14:58:44 [CONFIG]
>>>> org.hps.evio.EvioToLcio parse
>>>> :: set max events
>>>> to 10000
>>>> 2017-05-19 14:58:45 [CONFIG]
>>>> org.lcsim.job.JobControlManager
>>>> addVariableDefinition :: outputFile =
>>>> out_7796_0
>>>> 2017-05-19 14:58:45 [CONFIG]
>>>> org.hps.evio.EvioToLcio parse
>>>> :: set steering
>>>> variable: outputFile=out_7796_0
>>>> 2017-05-19 14:58:45 [CONFIG]
>>>> org.lcsim.job.JobControlManager
>>>> initializeLoop
>>>> :: initializing LCSim loop
>>>> 2017-05-19 14:58:45 [CONFIG]
>>>> org.lcsim.job.JobControlManager
>>>> initializeLoop
>>>> :: Event marker printing disabled.
>>>> 2017-05-19 14:58:45 [INFO]
>>>> org.hps.conditions.database.DatabaseConditionsManager
>>>> resetInstance ::
>>>> DatabaseConditionsManager instance is reset
>>>> Exception in thread "main"
>>>> java.lang.UnsatisfiedLinkError:
>>>> /u/apps/scicomp/java/jdk1.7.0_75/jre/lib/i386/xawt/libmawt.so:
>>>> libXext.so.6:
>>>> cannot open shared object file: No such
>>>> file or directory
>>>> at
>>>> java.lang.ClassLoader$NativeLibrary.load(Native
>>>> Method)
>>>> at
>>>> java.lang.ClassLoader.loadLibrary1(ClassLoader.java:1965)
>>>> at
>>>> java.lang.ClassLoader.loadLibrary0(ClassLoader.java:1890)
>>>> at
>>>> java.lang.ClassLoader.loadLibrary(ClassLoader.java:1851)
>>>> at
>>>> java.lang.Runtime.load0(Runtime.java:795)
>>>> at
>>>> java.lang.System.load(System.java:1062)
>>>> at
>>>> java.lang.ClassLoader$NativeLibrary.load(Native
>>>> Method)
>>>> at
>>>> java.lang.ClassLoader.loadLibrary1(ClassLoader.java:1965)
>>>> at
>>>> java.lang.ClassLoader.loadLibrary0(ClassLoader.java:1890)
>>>> at
>>>> java.lang.ClassLoader.loadLibrary(ClassLoader.java:1872)
>>>> at
>>>> java.lang.Runtime.loadLibrary0(Runtime.java:849)
>>>> at
>>>> java.lang.System.loadLibrary(System.java:1088)
>>>> at
>>>> sun.security.action.LoadLibraryAction.run(LoadLibraryAction.java:67)
>>>> at
>>>> sun.security.action.LoadLibraryAction.run(LoadLibraryAction.java:47)
>>>> at
>>>> java.security.AccessController.doPrivileged(Native Method)
>>>> at
>>>> java.awt.Toolkit.loadLibraries(Toolkit.java:1653)
>>>> at
>>>> java.awt.Toolkit.<clinit>(Toolkit.java:1682)
>>>> at
>>>> java.awt.Component.<clinit>(Component.java:595)
>>>> at
>>>> org.lcsim.util.aida.AIDA.<init>(AIDA.java:68)
>>>> at
>>>> org.lcsim.util.aida.AIDA.defaultInstance(AIDA.java:53)
>>>> at
>>>> org.hps.evio.RfFitterDriver.<init>(RfFitterDriver.java:31)
>>>> at
>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>> Method)
>>>>
>>>> atsun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAcce
>>>> ssorImpl.java:57)
>>>>
>>>> atsun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstru
>>>> ctorAccessorImpl.java:45)
>>>> at
>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>>>> at
>>>> java.lang.Class.newInstance(Class.java:379)
>>>> at
>>>> org.lcsim.job.JobControlManager.setupDrivers(JobControlManager.java:1199)
>>>>
>>>> at
>>>> org.hps.job.JobManager.setupDrivers(JobManager.java:82)
>>>> at
>>>> org.lcsim.job.JobControlManager.setup(JobControlManager.java:1052)
>>>> at
>>>> org.lcsim.job.JobControlManager.setup(JobControlManager.java:1110)
>>>> at
>>>> org.hps.evio.EvioToLcio.parse(EvioToLcio.java:407)
>>>> at
>>>> org.hps.evio.EvioToLcio.main(EvioToLcio.java:97)
>>>>
>>>>
>>>>
>>>> I see this library libXext.so.6: in
>>>> /usr/lib64, but not in
>>>> /usr/lib,
>>>> when I put /usr/lib64 in my
>>>> LD_LIBRARY_PATH, then it
>>>> complaines again (see
>>>> below)
>>>>
>>>> Exception in thread "main"
>>>> java.lang.UnsatisfiedLinkError:
>>>> /u/apps/scicomp/java/jdk1.7.0_75/jre/lib/i386/xawt/libmawt.so:
>>>> libXext.so.6:
>>>> wrong ELF class: ELFCLASS64
>>>>
>>>> I would appreciate, if I get some help on
>>>> running the
>>>> reconstruction
>>>> interactively, then I could look more
>>>> closely into logs
>>>> of the old, and new JAR files.
>>>>
>>>> Rafo
>>>>
>>>>
>>>> ____________________________________________________________________________
>>>>
>>>>
>>>> Use REPLY-ALL to reply to list
>>>>
>>>> To unsubscribe from the HPS-SOFTWARE list,
>>>> click the
>>>> following link:
>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin>
>>>>
>>>> _wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwIDaQ&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu
>>>>
>>>> 18s&r=0HDJrGO9TZQTE97J9Abt2A&m=xnbGP5VHYWRAQRWRksVgMnYvBkXWI4roLxztdJ0Tp9I&s=ppNYedSrn5DP
>>>>
>>>> aIZZJgRZu8tBDeSjroqbj_PoevFoFpI&e=
>>>>
>>>>
>>>>
>>>> ########################################################################
>>>> Use REPLY-ALL to reply to list
>>>>
>>>> To unsubscribe from the HPS-SOFTWARE list, click
>>>> the following link:
>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin>
>>>>
>>>> _wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwIDaQ&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu
>>>>
>>>> 18s&r=0HDJrGO9TZQTE97J9Abt2A&m=xnbGP5VHYWRAQRWRksVgMnYvBkXWI4roLxztdJ0Tp9I&s=ppNYedSrn5DP
>>>>
>>>> aIZZJgRZu8tBDeSjroqbj_PoevFoFpI&e=
>>>>
>>>>
>>>>
>>>> __________________________________________________________________________________________
>>>>
>>>>
>>>> Use REPLY-ALL to reply to list
>>>>
>>>> To unsubscribe from the HPS-SOFTWARE list, click the
>>>> following link:
>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin_wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwIDbA&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu18s&r=0HDJrGO9TZQTE97J9Abt2A&m=nIEvNWJSNxvmECWKfMrVKMTZKNdym3zsAls7WG1fdfU&s=cI82nYRtYBzgLs1l5g3anHrSt0kPk4B98oUzllu5KfQ&e=
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin_wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwIDbA&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu18s&r=0HDJrGO9TZQTE97J9Abt2A&m=nIEvNWJSNxvmECWKfMrVKMTZKNdym3zsAls7WG1fdfU&s=cI82nYRtYBzgLs1l5g3anHrSt0kPk4B98oUzllu5KfQ&e=>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> ########################################################################
>>>> Use REPLY-ALL to reply to list
>>>>
>>>> To unsubscribe from the HPS-SOFTWARE list, click the
>>>> following link:
>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin_wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwIDbA&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu18s&r=0HDJrGO9TZQTE97J9Abt2A&m=nIEvNWJSNxvmECWKfMrVKMTZKNdym3zsAls7WG1fdfU&s=cI82nYRtYBzgLs1l5g3anHrSt0kPk4B98oUzllu5KfQ&e=
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin_wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwIDbA&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu18s&r=0HDJrGO9TZQTE97J9Abt2A&m=nIEvNWJSNxvmECWKfMrVKMTZKNdym3zsAls7WG1fdfU&s=cI82nYRtYBzgLs1l5g3anHrSt0kPk4B98oUzllu5KfQ&e=>
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Hps-analysis mailing list
>>>> Hps-analysis at jlab.org <mailto:Hps-analysis at jlab.org>
>>>> https://mailman.jlab.org/mailman/listinfo/hps-analysis
>>>> <https://mailman.jlab.org/mailman/listinfo/hps-analysis>
>>>>
>>>>
>>>> _______________________________________________
>>>> Hps-analysis mailing list
>>>> Hps-analysis at jlab.org <mailto:Hps-analysis at jlab.org>
>>>> https://mailman.jlab.org/mailman/listinfo/hps-analysis
>>>
>>>
>>> ------------------------------------------------------------------------
>>>
>>> Use REPLY-ALL to reply to list
>>>
>>> To unsubscribe from the HPS-SOFTWARE list, click the following link:
>>> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1
>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin_wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwMGaQ&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu18s&r=0HDJrGO9TZQTE97J9Abt2A&m=yktqxWZIxrOJDj0GcxtXMdIm5S6CzoD5o_bfENc8yOA&s=9UGWhMTkreQfG7iEUol-dsVGExrJm-lRdRBfTmh68jU&e=>
>>>
>>>
>>
>
>
> ------------------------------------------------------------------------
>
> Use REPLY-ALL to reply to list
>
> To unsubscribe from the HPS-SOFTWARE list, click the following link:
> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1
> <https://urldefense.proofpoint.com/v2/url?u=https-3A__listserv.slac.stanford.edu_cgi-2Dbin_wa-3FSUBED1-3DHPS-2DSOFTWARE-26A-3D1&d=DwMGaQ&c=lz9TcOasaINaaC3U7FbMev2lsutwpI4--09aP8Lu18s&r=0HDJrGO9TZQTE97J9Abt2A&m=7kpWnc-wAM7IOppfAzLDwv1CB_CV5cve1_5MyJdB9IY&s=0GGLvbF5XbRfWvwDkVU2SxAw54dILOaQOFxPQNW0cQc&e=>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mailman.jlab.org/pipermail/hps-analysis/attachments/20170522/1cc377be/attachment-0001.html>
More information about the Hps-analysis
mailing list