Analysis codes

From GammaSphere DAQ
Jump to: navigation, search

Typical run procedure

Traditionally you will have a directory structure as


You can make this directory structure in two ways:

Option1 (preferred):

 cd to disk you want to use
 tar -zxvf ~dgs/dgs_template.tgz
 mv template gsfmannn
 cd gsfmannn

where nnn is the run number. You now have a directory with all you should need. To make sure things are up to date, you should

 (cd GEBSort; git pull)
 (cd GEBSort; make -B)
 (cd trackMain; git pull)
 (cd trackMain; make -B)

Option2 (manual):

you can checkout the software as

 git clone 
 (cd trackMain; make -B)
 git clone 
 (cd GEBSort; make -B)

This creates the two directories: trackMain and GEBSort. Even though DGS does not need the tracking code, it is best to create it and make the links below

 # make links to the crmat files which we keep in trackMain
 # but are needed here as well for GEBSort
 (cd GEBSort; rm GANIL_AGATA_crmat.dat; ln -s ../trackMain/GANIL_AGATA_crmat.dat GANIL_AGATA_crmat.dat)
 (cd GEBSort; rm GSI_AGATA_crmat.dat; ln -s ../trackMain/GSI_AGATA_crmat.dat GSI_AGATA_crmat.dat)
 (cd GEBSort; rm crmat.LINUX; ln -s ../trackMain/crmat.LINUX crmat.LINUX)

This option does not create all the directories and script files you need.

acquire and sort data

To acquire the data, cd to the 'data' directory. You start and stop the runs as: 123

To merge the data from a run, in the same directory, type 123

That will take the run 123 files in the data directory and make a merged file in the Merged directory and the log file in the LOG_FILES directory

Before the sort, you should look at the file. Lines you may need to change are

 beta        0.0
 dgs_MM      350

The cal files are the calibration files. See below for how to generate them.

To sort the data, cd to the GEBSort directory and 123

The root file will be placed in the ROOT_FILES directory as run123.root To look at the root file, you would do


Calibrations for bin_dgs in GEBSort_nogeb

GEBSort_nogeb is the program that can analyze data from DGS, DFMA and GRETINA. This is how you set up the code:

To produce the PZ spectra and 2D [sum2-sum1] vs [sum1] matrices needed to determine the PZ fudge factor (FF) you must enable

 #define ALL2DS 1

in bin_dgs.c and recompile . We do not always want these spectra as they take up a lot of space, but for now we need them

You now specify the PZ and ecal files in the file with these lines:

 dgs_MM      350

Before you start sorting data, you need to check that the map.dat file is up to date and reflects the array as it is configured.

For DGS data, enable bin_dgs in the file. To find the PZ values to use, sort some data from a 207Bi source. Then extract the pz spectra in .spe format with the script

  GEBSort_nogeb ....

Now run (you may have to compile):

  dgs_pz 350 141 1.003

where 350 100 are the M and K values you find in the file. Specify the values in 10 nsec units. In this case I saw these lines in the .save file:

 caput GLBL:DIG:d_window 0.06   
 caput GLBL:DIG:k_window 0.20     
 caput GLBL:DIG:m_window 3.50
 caput GLBL:DIG:k0_window 0.80
 caput GLBL:DIG:d3_window 0.20
 caput GLBL:DIG:raw_data_window 0.32
 caput VME01:SDIG1:k0_window0 0.0
 caput VME01:SDIG1:k0_window1 0.0

for the K value: sum up all the K and D values, in this case: 0.06+0.20+0.80+0.20 = 1.26 us or 126 in 10 nsec units. Notice that what is considered the K value also includes the D values (per SZ 6/25/18) as well as a D2 which is fixed at 0.15 (per JTA 6/26/18) and not in the listing above because the user cannot set it. Thus, in total, K in this example is 1.41 us or 141 in 10 ns units. The M value is 3.50 us or 350 in 10 nsec units. The 1.003 is a modification factor that needs to be determined by looking at energy vs baseline spectra.

 you already specified the M value in

After you executed dgs_pz, a d_pz.cmd file was generated. Use that cmd file in gf3 to check the pz spectra. Some might be really bad and dgs_pz might not have been able to find a good PZ value. If that is the case, to get around this problem, you may set the PZ for these detectors to the average value of the ones that had good PZ spectra. You would simply edit the file.

Now, after the PZ file is generated, remove the energy calibration file if there is one:


so that the calibrations defaults to 0 and 1 for offset and gain and sort again using the new pz values that were extracted above. When you resort, the PZ values in are read in and used. Extract the new clean, uncalibrated, ehi spectra as


and run the calibration program (you may have to compile)

  dgs_ecal 207Bi 600

you can also use "88Y", "60Co" for the source. The calibration will be 1keV/ch. The last parameter specifies the lowest channel the program will search for peaks in to avoid any noise at low energies.

Next when you run GEBSort_nogeb, both the new PZ values in and the gain and offset values in are read in and used. Take a good look at the spectra. Sometimes the dgs_pz and dgs_ecal programs can be fooled by noise or strange features in the spectra, so a few PZ and ecal parameters might have to be specified by hand.

The energy processing in bin_dgs follows algorithms that were developed by Shoufei Zhu.