Typical run procedure
Traditionally you will have a directory structure as
data/ GEBSort/ LOG_FILES/ Merged/ ROOT_FILES/
You can make this directory structure in two ways:
cd to disk you want to use tar -zxvf ~dgs/dgs_template.tgz mv template gsfmannn cd gsfmannn
where nnn is the run number. You now have a directory with all you should need. To make sure things are up to date, you should
(cd GEBSort; git pull) (cd GEBSort; make -B) (cd trackMain; git pull) (cd trackMain; make -B)
you can checkout the software as
git clone https://gitlab.phy.anl.gov/tlauritsen/trackMain.git (cd trackMain; make -B) git clone https://gitlab.phy.anl.gov/tlauritsen/GEBSort.git (cd GEBSort; make -B)
This creates the two directories: trackMain and GEBSort. Even though DGS does not need the tracking code, it is best to create it and make the links below
# make links to the crmat files which we keep in trackMain # but are needed here as well for GEBSort # (cd GEBSort; rm GANIL_AGATA_crmat.dat; ln -s ../trackMain/GANIL_AGATA_crmat.dat GANIL_AGATA_crmat.dat) (cd GEBSort; rm GSI_AGATA_crmat.dat; ln -s ../trackMain/GSI_AGATA_crmat.dat GSI_AGATA_crmat.dat) (cd GEBSort; rm crmat.LINUX; ln -s ../trackMain/crmat.LINUX crmat.LINUX)
This option does not create all the directories and script files you need.
acquire and sort data
To acquire the data, cd to the 'data' directory. You start and stop the runs as:
start_run.sh 123 stop_run.sh
To merge the data from a run, in the same directory, type
That will take the run 123 files in the data directory and make a merged file in the Merged directory and the log file in the LOG_FILES directory
Before the sort, you should look at the GEBSort.chat file. Lines you may need to change are
bin_none bin_dgs beta 0.0 dgs_MM 350 dgs_PZ dgs_pz.cal dgs_ecal dgs_ehi.cal
The cal files are the calibration files. See below for how to generate them.
To sort the data, cd to the GEBSort directory and
The root file will be placed in the ROOT_FILES directory as run123.root To look at the root file, you would do
rootn.exe dload("../ROOT_FILES/run123.root") ...display...
Calibrations for bin_dgs in GEBSort_nogeb
GEBSort_nogeb is the program that can analyze data from DGS, DFMA and GRETINA. This is how you set up the code:
To produce the PZ spectra and 2D [sum2-sum1] vs [sum1] matrices needed to determine the PZ fudge factor (FF) you must enable
#define ALL2DS 1
in bin_dgs.c and recompile . We do not always want these spectra as they take up a lot of space, but for now we need them
You now specify the PZ and ecal files in the GEBSort.chat file with these lines:
dgs_MM 350 dgs_PZ dgs_pz.cal dgs_ecal dgs_ehi.cal
Before you start sorting data, you need to check that the map.dat file is up to date and reflects the array as it is configured.
For DGS data, enable bin_dgs in the GEBSort.chat file. To find the PZ values to use, sort some data from a 207Bi source. Then extract the pz spectra in .spe format with the get_pz.cc script
GEBSort_nogeb .... rootn.exe dload("bi.root") .x get_pz.cc
Now run (you may have to compile):
dgs_pz 350 141 dgs_pz.cal 1.003
where 350 100 are the M and K values you find in the runxx.save file. Specify the values in 10 nsec units. In this case I saw these lines in the .save file:
caput GLBL:DIG:d_window 0.06 caput GLBL:DIG:k_window 0.20 caput GLBL:DIG:m_window 3.50 caput GLBL:DIG:k0_window 0.80 caput GLBL:DIG:d3_window 0.20 caput GLBL:DIG:raw_data_window 0.32 caput VME01:SDIG1:k0_window0 0.0 caput VME01:SDIG1:k0_window1 0.0 etc
for the K value: sum up all the K and D values, in this case: 0.06+0.20+0.80+0.20 = 1.26 us or 126 in 10 nsec units. Notice that what is considered the K value also includes the D values (per SZ 6/25/18) as well as a D2 which is fixed at 0.15 (per JTA 6/26/18) and not in the listing above because the user cannot set it. Thus, in total, K in this example is 1.41 us or 141 in 10 ns units. The M value is 3.50 us or 350 in 10 nsec units. The 1.003 is a modification factor that needs to be determined by looking at energy vs baseline spectra.
you already specified the M value in GEBSort.chat
After you executed dgs_pz, a d_pz.cmd file was generated. Use that cmd file in gf3 to check the pz spectra. Some might be really bad and dgs_pz might not have been able to find a good PZ value. If that is the case, to get around this problem, you may set the PZ for these detectors to the average value of the ones that had good PZ spectra. You would simply edit the dgs_pz.cal file.
Now, after the PZ file is generated, remove the energy calibration file if there is one:
so that the calibrations defaults to 0 and 1 for offset and gain and sort again using the new pz values that were extracted above. When you resort, the PZ values in dgs_pz.cal are read in and used. Extract the new clean, uncalibrated, ehi spectra as
and run the calibration program (you may have to compile)
dgs_ecal dgs_ehi.cal 207Bi 600
you can also use "88Y", "60Co" for the source. The calibration will be 1keV/ch. The last parameter specifies the lowest channel the program will search for peaks in to avoid any noise at low energies.
Next when you run GEBSort_nogeb, both the new PZ values in dgs_pz.cal and the gain and offset values in dgs_ehi.cal are read in and used. Take a good look at the spectra. Sometimes the dgs_pz and dgs_ecal programs can be fooled by noise or strange features in the spectra, so a few PZ and ecal parameters might have to be specified by hand.
The energy processing in bin_dgs follows algorithms that were developed by Shoufei Zhu.