source: trunk/MESOSCALE_DEV/MANUAL/SRC/compile_exec.tex @ 608

Last change on this file since 608 was 493, checked in by aslmd, 13 years ago

MESOSCALE: final version of the user manual

File size: 15.1 KB
Line 
1\chapter{Compiling the model and running a test case}\label{compile}
2
3\vk
4This chapter is meant for first time users of the LMD Martian Mesoscale Model. We describe how to compile the program and run a test case. We start with important basics about how the model works and how it is organized.
5
6\mk
7\section{Basics}
8
9\sk
10\subsection{Necessary steps to run a simulation}\label{steps}
11
12\sk
13Any simulation that will be carried out with the LMD Martian Mesoscale Model comprises the five following steps. More details are given on these steps in the following chapters, but it is important at this stage to have this structure in mind.
14
15\sk 
16\begin{itemize}
17\item \textbf{Step 0} Compiling the model.
18\item \textbf{Step 1} Running the LMD Global Circulation Model (GCM) to provide initial and boundary conditions for the mesoscale model.
19\item \textbf{Step 2} Choosing the mesoscale limited-area domain of simulation. Running preprocessing programs to horizontally interpolate GCM meteorological fields and static data (topography, soil properties) to the chosen simulation domain.
20\item \textbf{Step 3} Running preprocessing programs to vertically interpolate GCM meteorological fields and generate the initial and boundary conditions directly used by the mesoscale model.
21\item \textbf{Step 4} Running the LMD Martian Mesoscale Model.
22\end{itemize}
23
24\sk
25In this chapter, the general method to perform steps 0 and 4 is reviewed. Other steps are reviewed in chapter~\ref{zepreproc}; here the model is compiled and run for a test case with precomputed sample files for preprocessing steps 1, 2, 3.
26
27\sk
28\subsection{Structure of the \ttt{LMD\_MM\_MARS} directory}
29
30\sk
31Please take the time to check the contents of the \ttt{LMD\_MM\_MARS} directories\footnote{If you used method~$2$, you will probably notice that other directories than~\ttt{LMD\_MM\_MARS} are present in \ttt{\$MESO}, but those are not important at this stage.} and sub-directories through the following command lines:
32\begin{verbatim}
33ls $MMM ; ls $MMM/*
34\end{verbatim}
35
36\sk
37Contents of~\ttt{LMD\_MM\_MARS} directory:
38\begin{citemize}
39\item \ttt{makemeso}: this is the \ttt{bash} script to compile the model.
40\item \ttt{SRC}: this is a directory containing the model sources.
41\item \ttt{SIMU}: this is a directory containing scripts and files for an advanced use.
42\item \ttt{WPS\_GEOG}: this is a directory containing static data used in step~2.
43\end{citemize}
44
45\sk
46Contents of~\ttt{LMD\_MM\_MARS/SRC} subdirectory:
47\begin{citemize}
48\item \ttt{SCRIPTS}: this is a directory containing useful \ttt{bash} scripts for installation.
49\item \ttt{WRFV2}: this is a directory containing main model sources (modified WRF dynamics + LMD physics in \ttt{mars\_lmd*}).
50\item \ttt{PREP\_MARS}: this is a directory containing sources for the last part of step~1.
51\item \ttt{WPS}: this is a directory containing sources for step~2.
52\item \ttt{POSTPROC}: this is a directory containing postprocessing sources.
53%\item \ttt{PYTHON}: this is a directory containing \ttt{python}-based graphical scripts.
54\item \ttt{LES} and \ttt{LESnophys\_}: these are directories containing sources for Large-Eddy Simulations.
55\end{citemize}
56
57\sk
58Contents of~\ttt{LMD\_MM\_MARS/SIMU} subdirectory:
59\begin{citemize}
60\item \ttt{callphys.def},\ttt{dustopacity.def}, \ttt{run.def}, \ttt{namelist.input\_full}, \ttt{namelist.input\_minim}, \ttt{namelist.input\_nests}, \ttt{namelist.input\_les}, \ttt{namelist.wps\_example}, \ttt{namelist.wps\_nests}, \ttt{namelist.wps.template}~: these are useful example and template files to guide you through setting up your own parameters for the LMD Martian Mesoscale Model simulations.
61\item \ttt{calendar}: this is a text file containing time management information in the model.
62\item \ttt{runmeso}: this is a \ttt{bash} script that can be used once the model and preprocessing systems are installed; it prepares and runs a mesoscale simulation by going from step~1 to~4.
63\item \ttt{RUN}: this is a directory containing various files and scripts useful for advanced simulations.
64\item \ttt{DEF}: this is a directory containing many examples of parameter files for simulations.
65\end{citemize}
66
67\sk
68\begin{finger}
69\item In pre-2011 versions of the model, the contents of the various directories listed here might differ. This has probably no impact on your use of the model if you ensure the following files and directories are present in \ttt{LMD\_MM\_MARS}:
70\begin{citemize}
71\item \ttt{makemeso}, \ttt{prepare}, \ttt{prepare\_ini}, \ttt{copy\_model}
72\item \ttt{SRC/WRFV2}, \ttt{SRC/PREP\_MARS}, \ttt{SRC/WPS}
73\item \ttt{SIMU/runmeso}, \ttt{SIMU/calendar}
74\item \ttt{WPS\_GEOG}
75\end{citemize}
76\end{finger}
77
78\mk
79\section{Main compilation step}
80\label{sc:makemeso}
81
82\sk
83\subsection{Description of the \ttt{makemeso} script}
84
85\sk
86The \ttt{bash} script which allows you to compile the LMD Martian Mesoscale Model is \ttt{makemeso}. It is an automated script which performs the following serie of tasks:
87\begin{citemize}
88\item ask the user about compilation settings;
89\item retrieve some additional information about the system;
90\item create a directory \ttt{\$MESO/LMD\_MM\_MARS/your\_compdir} which name depends\footnote{For example, a \ttt{your\_compdir} directory named \ttt{g95\_32\_single} is created if the user requested a \ttt{g95} compilation of the code for single-domain simulations on a 32 bits machine.} on the kind of compiler you are using, on whether your system is 32 or 64 bits, on whether sequential or parallel computations are planned and on the kind of simulations (idealized or real-case);
91\item generate with \ttt{copy\_model} a directory \ttt{your\_compdir/WRFV2} with links to \ttt{SRC/WRFV2} sources\footnote{A note to developers: this method ensures that any change to the model sources would be propagated to all the different \ttt{your\_compdir} installation folders.};
92\item execute the WRF \ttt{configure} script with the correct options;
93\item tweak the resulting \ttt{configure.wrf} file to include a link towards the Martian physics and various patches and specific compilation options;
94\item calculate the total number of horizontal grid points handled by the LMD physics;
95\item duplicate LMD physical sources if nesting is activated;
96\pagebreak
97\item compile the LMD physical packages with the appropriate \ttt{makegcm} command
98and collect the compiled objects in the library \ttt{liblmd.a};
99\begin{finger}
100\item This step could be a bit long, especially if you are defining more than one domain. The \ttt{makemeso} script provides you with the full path towards the text file \ttt{log\_compile\_phys} in which you can check for compilation progress and possible errors. In the end of the process, you might find at the end of~\ttt{log\_compile\_phys} an error message associated to the generation of the final executable. Please do not pay attention to this, as the compilation of the LMD sources is meant to generate a library of compiled objects called \ttt{liblmd.a} instead of an executable.
101\end{finger}
102\item compile the modified Martian ARW-WRF solver and include the \ttt{liblmd.a} library;
103\begin{finger}
104\item When it is the first time the model is compiled, this step could be quite long. The \ttt{makemeso} script provides you with a \ttt{log\_compile} text file where the progress of the compilation can be checked and a \ttt{log\_error} text file listing errors and warnings during compilation. A list of warnings related to \ttt{grib} utilities (not used in the Martian model) may appear and have no impact on the final executables.
105\end{finger}
106\item change the name of the executables in agreement with the
107settings provided by the user.
108\end{citemize}
109
110\sk
111\subsection{Use of the \ttt{makemeso} script}
112
113\sk
114To compile the model, change directory to \ttt{\$MMM} and execute the \ttt{makemeso} command:
115
116\begin{verbatim}
117cd $MMM
118./makemeso
119\end{verbatim}
120
121\sk
122You are asked a few questions by the \ttt{makemeso} script (see the list below) then it compiles the model for you. The script outputs a text file named \ttt{last} in which your answers to the questions are stored, which allows you to re-run the script without the ``questions to the user" step through the \ttt{makemeso < last} command line. In what follows, the answers given in brackets are the ones you want to use so that you will be able to run the test case proposed in the next section.
123
124\mk
125\begin{asparaenum}[1.]%[\itshape Q1\upshape)]
126\item \textbf{choice of compiler}\footnote{We advise you to compile the model on the same kind of system (computer + operating system + librairies) as the one you plan to use to run the model.}
127\item[1.bis] (mpi-based compilation) number of processors to be used
128\item \textbf{number of grid points in longitude}\footnote{When you use parallel computations, please bear in mind that with $2$ (respectively $4$, $6$, $8$, $12$, $16$, $20$, $24$, $32$, $64$, $128$) processors the whole domain would be separated into $1$ (resp. $2$, $2$, $2$$3$$4$$4$$4$$4$$8$,   $8$) tiles over the longitude direction and $2$ (resp. $2$, $3$, $4$$4$$4$$5$$6$$8$$8$$16$) tiles over the latitude direction. Thus make sure that the number of grid points minus $1$ in each direction could be divided by the aforementioned number of tiles over the considered direction. For instance a~$82 \times 109$ horizontal grid is compliant with the use of~$12$ processors.} [61]
129\item \textbf{number of grid points in latitude} [61]
130\item \textbf{number of vertical levels} [61] 
131\item \textbf{number of tracers} [1]
132\item \textbf{number of domains} [1]
133%\item[6.bis] (not the first time you use \ttt{makemeso}) a question for advanced users [press any key]
134%\item[6.ter] (new LMD physics) number of different scatterers to be used
135\end{asparaenum}
136
137\mk
138A key question that often arises when using the LMD Martian Mesoscale Model is: when does the model has to be recompiled? The set of questions asked by~\ttt{makemeso} give some hints about this. Suppose you compiled a version of the model for a given set of parameters $1$ to $6$ to run a specific compilation. If you would like to run another simulation with at least one of parameters $1$ to $6$ subject to change, the model needs to be recompiled\footnote{This necessary recompilation each time the number of grid points, tracers and domains is modified is imposed by the LMD physics code. The WRF dynamical core alone is more flexible.} with \ttt{makemeso} (cf. also chapter~\ref{zeparam}).
139
140\mk
141Note that the \ttt{makemeso -h} command lists the various options that can be used in the \ttt{makemeso} script. Most options should be used only by advanced users and some of them will be described in the following chapters. At this stage, the only option of \ttt{makemeso} which can be useful to you is \ttt{-f} which forces the model to be recompiled from scratch. If you already compiled the model succesfully, but the model fails to compile a few days later for reasons unrelated to your operations on your system or on the model file, we recommend you to use the \ttt{-f} option in \ttt{makemeso} to try to recompile the model\footnote{A more extreme solution if \ttt{makemeso -f} does not solve your problem is to remove the corresponding \ttt{your\_compdir} directory. See chapter~\ref{faq}}.
142
143\scriptsize
144\codesource{makemesohelp}
145\normalsize
146
147\mk
148\section{Running a simple test case}
149\label{sc:arsia}
150
151\sk
152We assume here that you had successfully compiled the model with \ttt{makemeso} at the end of the previous section and you had based your answers to the \ttt{makemeso} script on the indications in brackets. You should then find in the \ttt{your\_compdir} directory the \ttt{real\_x61\_y61\_z61\_d1\_t1\_p1.exe} and~\ttt{wrf\_x61\_y61\_z61\_d1\_t1\_p1.exe} executables.
153
154\sk
155In order to test the compiled executables, a ready-to-use test case (with pre-generated initial and boundary conditions) is proposed in the \ttt{LMD\_MM\_MARS\_TESTCASE.tar.gz} archive that you can download in the following FTP site \url{ftp://ftp.lmd.jussieu.fr/pub/aslmd/LMD_MM_MARS_TESTCASE.tar.gz}. This test case simulates the hydrostatic atmospheric flow around Arsia Mons (Figure~\ref{arsia}) during half a sol in springtime with constant thermal inertia, albedo and dust opacity\footnote{Though the simulation reproduces some reasonable features of the mesoscale circulation around Arsia Mons (e.g. slope winds), it should not be used for scientific purpose, for the number of grid points is unsufficient for single-domain simulation and the integration time is below the necessary spin-up time.}.
156
157\sk
158To launch the test simulation, please type the following commands, replacing the \ttt{g95\_32\_single} directory with its corresponding value on your system. In the end, the model should run and output the computed meteorological fields in netCDF files named \ttt{wrfout*}. Feel free to browse those files with \ttt{ncview} or your favorite graphical tool to check if the simulated fields look reasonable.
159%
160\begin{verbatim}
161cp LMD_MM_MARS_TESTCASE.tar.gz $MMM
162tar xzvf LMD_MM_MARS_TESTCASE.tar.gz
163cd TESTCASE
164ln -sf ../g95_32_single/wrf_x61_y61_z61_d1_t1_p1.exe wrf.exe 
165nohup wrf.exe > log_wrf &
166\end{verbatim}
167
168\sk
169The files contained in \ttt{TESTCASE} prior to launching the simulations with the \ttt{wrf.exe} command illustrate which files are needed to perform step 4, i.e. running a LMD Martian Mesoscale Model simulation\footnote{For the test case presented here, a file named \ttt{dustopacity.def} is needed because for the sake of simplicity of this test case, we set idealized uniform dust opacity. The file \ttt{namelist.wps} is included in the \ttt{TESTCASE} folder for further reference but not needed at this stage.}.
170\begin{itemize}
171\item \ttt{namelist.input}: text file containing parameters for the dynamical core
172\item \ttt{callphys.def}: text file containing parameters for the physics parameterizations
173\item \ttt{wrf.exe}: the model executable (or a link to it) as compiled by \ttt{makemeso}
174\item \ttt{wrfinput\_d01} and \ttt{wrfbdy\_d01}: data files containing initial and boundary conditions
175\end{itemize}
176
177\begin{center} 
178\begin{figure}[h!]
179\includegraphics[width=0.5\textwidth]{arsiadomain.png} 
180\includegraphics[width=0.5\textwidth]{LMD_MMM_d1_20km_HGT_UV_10m-ALS_Ls8_LT1_100.png}
181\caption{\label{arsia} [Left plot] Simulation domain defined in the test case proposed as a demonstrator for running the LMD Martian Mesoscale Model. [Right plot] Nighttime winds predicted by the model~$10$~m above the surface. Both plots have been generated by command-line scripts written in~\ttt{python + numpy + matplotlib} (see chapter~\ref{postproc}).}
182\end{figure}
183\end{center}
184
185\bk
186%\scriptsize
187\begin{finger}
188\item If you compiled the model using MPI, the command to launch a simulation is slightly different:
189%
190\begin{verbatim}
191[if several connected machines, create a file mpd.hosts with machine names...]
192[... and make sure that ssh between machines does not need authentification]
193mpirun [-f mpd.hosts] -np number_of_processors wrf.exe < /dev/null &     
194tail -20 rsl.out.000? # to check the outputs
195\end{verbatim}
196%%echo barry.lmd.jussieu.fr > ~/mpd.hosts
197%%echo white.lmd.jussieu.fr >> ~/mpd.hosts
198%%echo loves.lmd.jussieu.fr >> ~/mpd.hosts
199%%echo tapas.lmd.jussieu.fr >> ~/mpd.hosts     
200%%mpdboot  -n 4
201%%mpdtrace
202%%mpirun -l -np 16 wrf.exe < /dev/null &   # NB: mpirun is only a link to mpiexec
203\end{finger}
204%\normalsize
205
206\clearemptydoublepage
Note: See TracBrowser for help on using the repository browser.