Newer
Older
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
#!/bin/bash
#Script to convert in DL2 data a set of DL1 REAL data.
#Change paths according to your system
#Some variables for retrieving the input files
DATES=( 20201120 )
#DATES=( 20201118 20201119 20201120 ) # possible to analyze more dates together
SOURCE="Crab"
#CONFIG=/fefs/aswg/workspace/alice.donini/Software/cta-lstchain/lstchain/data/lstchain_standard_config.json
VERSION="v0.9.2"
CLEANING="tailcut84"
# RF files
PATH_MODELS="/fefs/aswg/workspace/alice.donini/Analysis/data/models/source_independent"
#Script (in ./cta-lstchain/lstchain/scripts/)
SCRIPT="/fefs/aswg/workspace/alice.donini/Software/cta-lstchain/lstchain/scripts/lstchain_dl1_to_dl2.py"
for date in "${DATES[@]}"
do
INPUT_DIR=/fefs/aswg/workspace/alice.donini/Analysis/data/DL1/${SOURCE}/${date}/${VERSION}/${CLEANING}/
OUTPUT_DIR=/fefs/aswg/workspace/alice.donini/Analysis/data/DL2/${SOURCE}/${date}/${VERSION}/${CLEANING}/
mkdir -p ${OUTPUT_DIR}
mkdir -p jobs/dl2/${date}/
for file in $(ls ${INPUT_DIR}/ | grep dl1_LST); do
echo "File $file"
rm -rf jobs/dl2/${date}/run_${file}.sh
echo "#!/bin/bash
#SBATCH -N 1
#SBATCH --mem 100000
ulimit -l unlimited
ulimit -s unlimited
ulimit -a
python $SCRIPT -f $INPUT_DIR$file -p $PATH_MODELS -o $OUTPUT_DIR" >> jobs/dl2/${date}/run_${file}.sh
# python $SCRIPT -f $INPUT_DIR$file -p $PATH_MODELS -o $OUTPUT_DIR -c $CONFIG" >> jobs/dl2/${date}/run_${file}.sh
chmod gu+x jobs/dl2/${date}/run_${file}.sh
sbatch jobs/dl2/${date}/run_${file}.sh
done
done