Skip to content
Snippets Groups Projects

Compare revisions

Changes are shown as if the source revision was being merged into the target revision. Learn more about comparing revisions.

Source

Select target project
No results found

Target

Select target project
  • ipsl/lmd/intro/routingpp
1 result
Show changes
Commits on Source (11)
......@@ -65,7 +65,7 @@ EOF
loc=''
hasanaconda loc
if [ loc != "unknown" ] ; then
if [ $loc != "unknown" ] ; then
#
# Make sure the right version of Python
#
......@@ -73,13 +73,13 @@ if [ loc != "unknown" ] ; then
"ipsl")
# The first case if for Spirit, the second is legagcy for Climserv.
if [[ -e /etc/os-release ]] ; then
module load anaconda3-py
module purge
module load anaconda-meso/2023.09-0
if [ ! -e ${HOME}/.conda/envs/MPIs ] ; then
conda update -n base -c defaults conda
conda create -y -n MPIs --channel astropy mpi4py numba astropy spherical-geometry netcdf4 ConfigParser
source activate MPIs
conda install -y --channel astropy matplotlib basemap
conda install -y --channel astropy cartopy
#
conda create -y -n MPIs python==3.11.5
conda install -y -n MPIs --channel astropy mpi4py numba astropy spherical-geometry netcdf4 ConfigParser matplotlib cartopy
conda activate MPIs
else
conda activate MPIs
fi
......@@ -104,12 +104,9 @@ if [ loc != "unknown" ] ; then
;;
"idris")
# Get the most recent anaconda environment
echo "Loading Anaconda module"
# Loading same Anaconda environment and Python version as on Spirit
module purge
module load anaconda-py3/2023.09
. "/gpfslocalsup/pub/anaconda-py3/2023.09/etc/profile.d/conda.sh"
conda init
conda deactivate
conda info --envs
#
# Make sure that the .conda directory is on $WORK but linked to $HOME.
#
......@@ -125,21 +122,12 @@ if [ loc != "unknown" ] ; then
# Now that all is in place install our software
#
if [ ! -e ${WORK}/.conda/envs/MPI ] ; then
##conda init
module list
echo "Which Conda env : $CONDA_EXE"
conda create -y -n MPI --channel astropy mpi4py numba astropy spherical-geometry netcdf4 ConfigParser
#
conda create -y -n MPI python==3.11.5
conda install -y -n MPI --channel astropy mpi4py numba astropy spherical-geometry netcdf4 ConfigParser matplotlib cartopy
conda activate MPI
echo "Conda Prefix : $CONDA_PREFIX"
conda install -y --channel astropy matplotlib
conda install -y --channel astropy cartopy
else
echo "We discovered that we are on IDRIS"
module list
echo "Which Conda env : $CONDA_EXE"
##conda init
conda activate MPI
echo "Conda Prefix : $CONDA_PREFIX"
fi
;;
*)
......@@ -150,6 +138,6 @@ if [ loc != "unknown" ] ; then
fi
#
# Verify we can load all the needed modules.
#
#
checkenv
......@@ -648,7 +648,7 @@ SUBROUTINE finish_inflows(nbpt, nwbas, nbasmax, inf_max, basin_count, inflow_num
INTEGER(i_std), DIMENSION(nbpt,nwbas,inf_max), INTENT(in) :: inflow_basin !!
INTEGER(i_std), DIMENSION(nbpt,nwbas,inf_max), INTENT(in) :: inflow_grid !!
!
REAL(r_std), DIMENSION(nbpt,nbasmax), INTENT(out) :: route_innum
INTEGER(i_std), DIMENSION(nbpt,nbasmax), INTENT(out) :: route_innum
REAL(r_std), DIMENSION(nbpt,nbasmax,inf_max), INTENT(out) :: route_ingrid
REAL(r_std), DIMENSION(nbpt,nbasmax,inf_max), INTENT(out) :: route_inbasin
!
......
......@@ -4,6 +4,13 @@ import inspect
import importlib
from multiprocessing import Pool
import getargs
config = getargs.SetupConfig()
log_master, log_world = getargs.getLogger(__name__)
INFO, DEBUG, ERROR = log_master.info, log_master.debug, log_world.error
INFO_ALL, DEBUG_ALL = log_world.info, log_world.debug
def localdir(follow_symlinks=True):
if getattr(sys, 'frozen', False): # py2exe, PyInstaller, cx_Freeze
path = os.path.abspath(sys.executable)
......@@ -24,15 +31,15 @@ def compilef90(mod, mpi=None) :
# Choose the right makefile
#
host=os.uname()[1]
if host.find("ipsl") >= 0:
if host.find("ipsl") >= 0 or host.find("spirit") >= 0 :
mf=""
elif host.find("jean-zay") >= 0:
mf="-f Makefile.JeanZay"
else :
ERROR("Unknown host : "+host)
sys.exit()
if not "mpi4py" in sys.modules :
print("XXXX1 compilation with option ", mf)
err=os.system("cd "+localdir()+"/F90subroutines ; make "+mf+" "+mod)
if err != 0 :
print("Compilation error in the FORTRAN interface")
......@@ -40,7 +47,6 @@ def compilef90(mod, mpi=None) :
else :
mpimod=sys.modules["mpi4py"].MPI
if mpimod.COMM_WORLD.Get_rank() == 0 :
print("XXXX2 compilation with option ", mf)
err=os.system("cd "+localdir()+"/F90subroutines ; make "+mf+" "+mod)
if err != 0 :
print("Compilation error in the FORTRAN interface")
......
......@@ -79,7 +79,10 @@ class HydroOverlap :
#
def __init__(self, nbpt, nbvmax, sub_pts, sub_index_in, sub_area_in, sub_lon_in, sub_lat_in, sub_hdom_in, part, modelgrid, hydrodata) :
#
print("np.shape(sub_hdom_in)", np.shape(sub_hdom_in))
if isinstance(sub_hdom_in, list) :
print("len(sub_hdom_in)", len(sub_hdom_in))
else :
print("np.shape(sub_hdom_in)", np.shape(sub_hdom_in))
#
# Reshape stuff so that it fits into arrays
#
......@@ -361,7 +364,8 @@ class HydroSuper :
# hydrogrid is regular
lon = hydrogrid.ncfile.variables["nav_lon"][0,istr:iend]
lat = hydrogrid.ncfile.variables["nav_lat"][jstr:jend,0]
trip = hydrogrid.ncfile.variables["trip"][jstr:jend,istr:iend].astype(np.int32)
trip = hydrogrid.ncfile.variables["trip"][jstr:jend,istr:iend].astype(np.int32).filled(98)
# trip = hydrogrid.ncfile.variables["trip"][jstr:jend,istr:iend].fill(-1)
#
bounds = [np.ma.min(lon), np.ma.max(lon), np.ma.min(lat), np.ma.max(lat)]
trip2bearing = [0., 45., 90., 135., 180., 225., 270., 315.]
......@@ -647,7 +651,7 @@ class HydroSuper :
istr, iend, jstr, jend = hydrogrid.box[:]
lon = hydrogrid.ncfile.variables["nav_lon"][0,istr:iend]
lat = hydrogrid.ncfile.variables["nav_lat"][jstr:jend,0]
trip = hydrogrid.ncfile.variables["trip"][jstr:jend,istr:iend].astype(np.int32)
trip = hydrogrid.ncfile.variables["trip"][jstr:jend,istr:iend].astype(np.int32).filled(98)
#
bounds = [np.ma.min(lon), np.ma.max(lon), np.ma.min(lat), np.ma.max(lat)]
jml,iml = trip.shape
......@@ -673,7 +677,7 @@ class HydroSuper :
orog = hydrogrid.ncfile.variables["orog"][jstr:jend,istr:iend]
disto = hydrogrid.ncfile.variables["disto"][jstr:jend,istr:iend]
tot=0; ok = 0; nook = 0; bad = 0
for pt in part.landcorelist:
for bs in range(self.basin_count[pt]):
......@@ -689,7 +693,7 @@ class HydroSuper :
j0 = np.where(lat==ilat)[0][0]
i1 = np.where(lon==olon)[0][0]
j1 = np.where(lat==olat)[0][0]
ii,jj = get_next(j0, i0, trip, grid)
if (ii == i1) and (jj == j1):
# then correct
......
......@@ -33,37 +33,73 @@ F90T.compilef90("diagst")
import diagst
host=os.uname()[1]
if host.find("ipsl") >= 0 :
if host.find("ipsl") >= 0 or host.find("spirit") >= 0 :
mpath="/bdd/ORCHIDEE_Forcing/Routing/Hydro4ORCH"
elif host.find("jean-zay") >= 0 :
mpath="/gpfswork/rech/ron/rron972/Hydro4ORCH"
else :
ERROR("Unknown host : "+host)
sys.exit()
points_path = config.get("OverAll","Points_path",fallback=mpath)
#
def readcsv(csvfile) :
def readcsv(f) :
'''
Object : Read a CSV file separated by comas "," and put the result in a dictionary
where the keys are the names in the first line of the file.
Input :
csvfile : A csv file with the first line the column names and comas to separate.
f : A csv file with the first line the column names and comas to separate.
Output :
A dictionary with the key being the name of the columns and the list of values.
'''
with open(csvfile, newline='') as csvfile:
cf = csv.reader(csvfile, delimiter=',', quotechar='|')
o={}
for row in cf:
if len(o) == 0 :
lab=[]
for l in row :
o[l]=[]
lab.append(l)
with open(f) as csvfile:
reader = csv.DictReader(csvfile, delimiter=",")
x={}
for row in reader:
for k in row.keys() :
if k not in x.keys() :
x[k]=[row[k]]
else :
x[k].append(row[k])
#
# Detect type in each list and convert appropriatly
#
ctype={}
for k in x.keys() :
if k.find("lon") >= 0 or k.find("lat") >= 0 or k.find("Err") >= 0 :
ctype[k]="float"
else :
xt=[s.replace("-","",1).replace(".", "", 1).strip() for s in x[k]]
xx=[s.replace("-","",1).strip() for s in x[k]]
if "".join(xt).isdigit() :
i="".join(xx).count(".")
if i > 0 :
ctype[k]="float"
else :
ctype[k]="int"
else :
for i,e in enumerate(row) :
l=lab[i]
o[l].append(e)
ctype[k]="str"
#
# Convert for final output dictionary
#
o={}
for k in x.keys() :
if ctype[k].find("float") >= 0 :
o[k]=[]
for s in x[k] :
if len(s) > 0 :
o[k].append(float(s))
else :
o[k].append(np.nan)
elif ctype[k].find("int") >= 0 :
o[k]=[]
for s in x[k] :
if len(s) > 0 :
o[k].append(int(s))
else :
o[k].append(-9999)
else :
o[k]=x[k]
return o
#
class Locations :
......
......@@ -2,6 +2,8 @@ import numpy as np
import sys
from numba import jit
#
import RPPtools as RPP
#
import getargs
log_master, log_world = getargs.getLogger(__name__)
INFO, DEBUG, ERROR = log_master.info, log_master.debug, log_world.error
......@@ -768,18 +770,52 @@ class partition :
# Function to gather vectors of points on different procs on root
#
def gatherbypoint(self, x) :
'''
Object : The values of all processors will be gathered into a an array whether they arrive as a list or array.
The type of the values is preserved.
'''
maxlen = self.domainmax(len(x))
scalar = False
if isinstance(x, np.ndarray) :
if x.shape[0]==0 :
dtyp=np.float32
integer=isinstance(x, np.integer)
else :
dtyp=type(x.flat[0])
integer=isinstance(x.flat[0], np.integer)
elif isinstance(x, list) :
if len(x) == 0 :
scalar=True
dtyp=type(x)
integer=isinstance(x, np.integer)
else :
dtyp=type(x[0])
integer=isinstance(x[0], np.integer)
else :
dtyp=type(x)
integer=isinstance(x, np.integer)
if self.rank == 0 :
xout = np.empty((maxlen,self.size))
xout = np.empty((maxlen,self.size), dtype=dtyp)
# Receive from all procs
for lr in range(self.size) :
if lr == 0 :
xout[:,lr] = np.pad(x, [0,maxlen-len(x)], mode='constant', constant_values=np.nan)
if integer :
xout[:,lr] = np.pad(x, [0,maxlen-len(x)], mode='constant', constant_values=RPP.IntFillValue)
else :
xout[:,lr] = np.pad(x, [0,maxlen-len(x)], mode='constant', constant_values=RPP.FillValue)
else :
xout[:,lr] = self.comm.recv(source=lr)
else :
xout = None
self.comm.send(np.pad(x, [0,maxlen-len(x)], mode='constant', constant_values=np.nan), dest=0)
if integer :
self.comm.send(np.pad(x, [0,maxlen-len(x)], mode='constant', constant_values=RPP.IntFillValue), dest=0)
else :
self.comm.send(np.pad(x, [0,maxlen-len(x)], mode='constant', constant_values=RPP.FillValue), dest=0)
#
return xout
#
......
......@@ -70,13 +70,14 @@ class HDEM_dir:
self.EarthRadius=config.getfloat("OverAll", "EarthRadius", fallback=6378000.0)
#
host=os.uname()[1]
if host.find("ipsl") >= 0:
if host.find("ipsl") >= 0 or host.find("spirit") >= 0 :
mpath="/bdd/ORCHIDEE_Forcing/Routing/Hydro4ORCH/"
elif host.find("jean-zay") >= 0:
mpath="/gpfswork/rech/ron/rron972/Hydro4ORCH/"
else :
ERROR("Unknown host : "+host)
sys.exit()
lect = config.get("OverAll", "HydroFile")
#
if lect is not None :
......
......@@ -3,9 +3,9 @@
#SBATCH --job-name=BuildHTUs_IP # Job Name
#SBATCH --output=BuildHTUs_IP.out # standard output
#SBATCH --error=BuildHTUs_IP.out # error output
#SBATCH --ntasks=32 # Number of MPI tasks
#SBATCH --ntasks=24 # Number of MPI tasks
#SBATCH --hint=nomultithread # 1 processus MPI par par physical core (no hyperthreading)
#SBATCH --time=0-2:00 # time (D-HH:MM)
#SBATCH --time=10:00:00 # time (D-HH:MM)
#
cd ${SLURM_SUBMIT_DIR}
echo "Available : ${SLURM_NTASKS}, ${SLURM_NPROCS}"
......@@ -21,7 +21,7 @@ source ../../Environment
# Run the Python code to generate the HTUs and write them into a netCDF file.
#
ulimit -s unlimited
mpiexec -n 2 python ../../RoutingPreProc.py run.def
mpiexec -n ${SLURM_NTASKS} python ../../RoutingPreProc.py run_MEDCORDEX.def
if [ $? -gt 0 ] ; then
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXZZ"
echo "X MEDCORDEX Run on Iberia failed XZZ"
......
......@@ -3,9 +3,9 @@
#SBATCH --job-name=BuildHTUs_IPregular # Job Name
#SBATCH --output=BuildHTUs_IPregular.out # standard output
#SBATCH --error=BuildHTUs_IPregular.out # error output
#SBATCH --ntasks=32 # Number of MPI tasks
#SBATCH --ntasks=12 # Number of MPI tasks
#SBATCH --hint=nomultithread # 1 processus MPI par par physical core (no hyperthreading)
#SBATCH --time=0-2:00 # time (D-HH:MM)
#SBATCH --time=5:00:00 # time (D-HH:MM)
#
cd ${SLURM_SUBMIT_DIR}
echo "Available : ${SLURM_NTASKS}, ${SLURM_NPROCS}"
......
[OverAll]
#
#
#
EarthRadius = 6378000.
#
ModelGridFile = /bdd/MEDI/workspaces/polcher/NewRouting/geo_MEDCORDEX.nc
HydroFile = HS
IrrigationsFile = NO
#
WEST_EAST = -9.75, 5.25
SOUTH_NORTH = 35.5, 43.5
#
# FORTRAN interface parameters
#
WeightFile = Weights_MEDCORDEX_Spain.nc
#
# Configuration for the graph to be generated
#
nbasmax = 35
#
# Number of operation of simplification performed together
#
numop = 100
#
# Output
#
GraphFile = MEDCORDEX_Spain_graph.nc
......@@ -7,7 +7,7 @@ EarthRadius = 6378000.
ModelGridFile = /bdd/MEDI/workspaces/polcher/NewRouting/geo_MEDCORDEX.nc
WEST_EAST = -9.75, 5.25
SOUTH_NORTH = 35.5, 43.5
HydroFile = /bdd/ORCHIDEE_Forcing/Routing/Hydro4ORCH/HydroSHEDS_MED.nc
HydroFile = HydroSHEDS
#
# FORTRAN interface parameters
#
......@@ -23,15 +23,5 @@ numop = 100
#
# Output
#
GraphFile = MEDCORDEX_Spain_graph.nc
#
# File containing infrastructures to be placed.
# Maximum error in the distance of the station in km^2
MaxDistErr = 25
# Maximum error in the upstream area in %
MaxUpstrErr = 10
StructuresFile = Structures.txt
GRDCFile = ../../Stations/Station_Metadata.nc
PointsFile= /homedata/lbaratgin/codes/hydrofiles/Point2Graph/Points_grid_MNT_France.nc
DamsFile= ../../Source_Files/GRanCFBR_Dams_Reservoirs_France.csv
StationsFile= /bdd/ORCHIDEE_Forcing/Validation/GRDC/GRDC_Monthly_Feb21_v1.nc
\ No newline at end of file
GraphFile = MEDCORDEX_MEDCORDEX_35HTU_graph.nc
#
......@@ -6,14 +6,13 @@ EarthRadius = 6378000.
#
ModelGridFile = /bdd/MEDI/workspaces/polcher/NewRouting/EM_WFDEI_CRU_2000.nc
HydroFile = HS
IrrigationsFile = NO
#
WEST_EAST = -9.75, 5.25
SOUTH_NORTH = 35.5, 43.5
#
# FORTRAN interface parameters
#
WeightFile = Weights_EM_WFDEI_Spain.nc
WeightFile = Weights_EM_WFDEI_Iberia.nc
#
# Configuration for the graph to be generated
#
......@@ -25,4 +24,4 @@ numop = 100
#
# Output
#
GraphFile = EM_WFDEI_Spain_graph.nc
GraphFile = EM_WFDEI_Iberia_32HTU_graph.nc
......@@ -3,11 +3,12 @@
#SBATCH --job-name=BuildHTUs_Mallorca # Job Name
#SBATCH --output=BuildHTUs_Mallorca.out # standard output
#SBATCH --error=BuildHTUs_Mallorca.out # error output
#SBATCH --ntasks=32 # Number of MPI tasks
#SBATCH --ntasks=12 # Number of MPI tasks
#SBATCH --hint=nomultithread # 1 processus MPI par par physical core (no hyperthreading)
#SBATCH --time=0-2:00 # time (D-HH:MM)
#SBATCH --time=10:00:00 # time (D-HH:MM)
#
cd ${SLURM_SUBMIT_DIR}
export UCX_WARN_UNUSED_ENV_VARS=n
echo "Available : ${SLURM_NTASKS}, ${SLURM_NPROCS}"
#
# Set the right Python 3 Anaconda environment
......@@ -21,7 +22,8 @@ source ../../Environment
#
# 2 Proc
#
mpiexec -n 2 python ../../RoutingPreProc.py run_MEDCORDEX.def
/bin/rm -f Weights_MEDCORDEX_MERIT.nc
mpiexec -n 2 python3 ../../RoutingPreProc.py run_MEDCORDEX.def
if [ $? -gt 0 ] ; then
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX"
echo "X Run MECORDEX on 2 Proc failed X"
......@@ -33,24 +35,22 @@ else
echo "======================================"
ls -lt
fi
#
echo "Check status of ${SLURM_JOB_ID} job:"
sacct -j ${SLURM_JOB_ID} --format=JobID,MaxVMSize,AllocCPUs,UserCPU,Elapsed
exit
#
# 4 Proc
#
/bin/rm -f run.def Out_*.txt Weight_*.txt
mpiexec -n 3 "python ../../RoutingPreProc.py run_E2OFD.def
/bin/rm -f Weights_E2OFD_Contfrac.nc
mpiexec -n 4 python3 ../../RoutingPreProc.py run_E2OFD.def
if [ $? -gt 0 ] ; then
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX"
echo "X Run E2OFD on 3 Proc failed X"
echo "X Run E2OFD on 4 Proc failed X"
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX"
exit
else
echo "=================================="
echo "= Run E2OFD on 3 Proc successful ="
echo "= Run E2OFD on 4 Proc successful ="
echo "=================================="
ls -lt
fi
......@@ -59,15 +59,16 @@ fi
#
/bin/rm -f run.def Out_*.txt Weight_*.txt
mpiexec -n 3 python ../../RoutingPreProc.py run_E2OFDMERIT.def
/bin/rm -f Weights_E2OFD_MERIT.nc
mpiexec -n 4 python ../../RoutingPreProc.py run_E2OFDMERIT.def
if [ $? -gt 0 ] ; then
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX"
echo "X Run E2OFD & MERIT on 3 Proc failed X"
echo "X Run E2OFD & MERIT on 4 Proc failed X"
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX"
exit
else
echo "========================================="
echo "= Run E2OFD& MERIT on 3 Proc successful ="
echo "= Run E2OFD& MERIT on 4 Proc successful ="
echo "========================================="
ls -lt
fi
......@@ -76,15 +77,17 @@ fi
#
/bin/rm -f run.def Out_*.txt Weight_*.txt
mpiexec -n 32 python ../../RoutingPreProc.py run_EuroSW.def
/bin/rm -f Weights_EuroSW.nc
mpiexec -n 12 python ../../RoutingPreProc.py run_EuroSW.def
if [ $? -gt 0 ] ; then
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX"
echo "X Run EuroSW on 32 Proc failed X"
echo "X Run EuroSW on 12 Proc failed X"
echo "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX"
else
echo "===================================="
echo "= Run EuroSW on 32 Proc successful ="
echo "= Run EuroSW on 12 Proc successful ="
echo "===================================="
ls -lt
fi
sstat --format=AveCPU,AvePages,AveRSS,AveVMSize,JobID -j ${SLURM_JOB_ID}
......@@ -10,7 +10,7 @@ SOUTH_NORTH = 39.00, 40.2
#
# FORTRAN interface parameters
#
WeightFile = E2OFD_MERIT_Weights.nc
WeightFile = Weights_E2OFD_MERIT.nc
#
# Configuration for the graph to be generated
#
......
......@@ -10,7 +10,7 @@ SOUTH_NORTH = 39.20, 40.2
#
# FORTRAN interface parameters
#
WeightFile = EuroSW_Weights.nc
WeightFile = Weights_EuroSW.nc
#
# Configuration for the graph to be generated
#
......