[Vp-reproduce-subgroup] White paper revision
John Gennari
gennari at uw.edu
Mon May 17 22:10:36 PDT 2021
All: About 9 of us had a lively Zoom meeting today to chat about the
manuscript. By the end, it was a productive meeting, and I'm hoping that
this email will capture some key outputs from the meeting. I apologize
if I said some things that were a bit "inflammatory". Obviously 2 years
would be much too long to get this paper out-the door.
I saw two outcomes. First, we had some nice ideas and discussion about
re-ordering (initiated by Tomas Helikar). In the below, I'm going to
propose one possible ordering, but this is certainly a work-in-progress.
The reason that I think ordering is important is that it will give us a
much better ability to write a strong concluding section, where we talk
about themes and the larger arc of our ideas.
Second, we agreed that we should nominate "point persons" who would be
in charge of at least the initial cut of each of the subsections. As
Jacob pointed out, this information should be easy to get from older
email and history of the development of the paper. During the zoom
meeting, we associated some co-authors with some sections, but our
coverage wasn't perfect (see challenge #12). Hopefully people will
"stand up" and admit that some section of text is theirs.
So in the below, I include the original title of the section, a few
words about the content of that section, and then a name (or several
names) of co-authors who will be the "point person" to make sure that
the appropriate content is included. Obviously, all co-authors can and
should chime in on any part of the text, but the point person should
make sure that the key ideas are included.
The basic ordering idea for the dozen challenges was to follow the
life-cycle of model development, execution, sharing and integration, and
eventually implementation. So...
*********************************************
*(1) "**Data**and measurement definitions*". Before you can build a
model, you must have data. So data availability and measurement
standards is the place to start.
*People: *Hana D, Jacob B
*(2) "**The variety of modeling languages*" This is about the choice of
modeling languages, such as using SBML, CellML, or Matlab. As I said on
the phone call, this is sort of about "syntax"--how do you write down
your model?
*People:* John G, Jon K, Rahuman S.
*(3) "**The variety of modeling paradigms and scales"***Separately from
modeling syntax, we must acknowledge modeling paradigms with very
different semantics. Some clear examples are PDEs versus ODEs versus
rule-based systems (and obviously one can combine these). Certainly
semantics might impact syntax (the prior challenge), in that certain
modeling language might be appropriate only for some paradigms.
People: James G, Eric F (?)
*(4) "**Units standardization*" A common reason that models are not
reproducible are errors in units, or misunderstanding about units, or
simply a lack of information about units.
People: Jacob B, Hana D
*(5) "**A lack of annotations in models*". Once researchers publish
models, they must annotate the model so that others can understand it.
Quality annotation is essential for both search and reproducibility.
People: John G.
*(6) "**Models are hard to locate"* If your goal is to reproduce,
understand and possibly reuse or integrate some other model, one must
first find that model. This requires annotation (prior section) and
repositories (Physiome Model Repository, BioModels) and search platforms
(ModeleXchange).
People: Jon K, John G.
*(7) "**Common platforms to execute models" *A model is pretty worthless
as a static object. For folk to understand and reproduce models they
must be executable. Alas, there is no single or consistent way of
executing a model -- and of course, this interacts direction with
section #2 and #3, above: Execution platforms are usually only for one
modeling paradigm, and often for one modeling language. The
BioSimulators work goes here.
People: Jon K.
*(8) "**Credibility **and validity of models*" Once a model is
published, how do folk know it is right? Model validation is a big topic
and challenge. Credibility follows (in part) from validation, but also
requires transparency and reproducibility, etc.
People: John Rice, Jon K, Jacob B
*(9) "**Environments to adapt and integrate models*" As I see it, one of
the end-targets for this manuscript is to better enable model
integration, to build better models. There are many challenges with the
task of integrating two (or more) models. (One that has recently been
discussed is that even if model A and model B are valid and correct,
there is no guarantee that the combined model A+B is correct. I liked
what William Waites and Katherine Morse posted on this subject.) This
section is where SBML-comp and SemGen environments can be mentioned.
People: John G.
(*10) "Challenges for stochastic models" *Special challenges specific to
stochaistic modeling. An obvious point to mention is repeatability --
stochastic models don't necessarily give the same results with the same
inputs.
People: James G., Eric F
*(11) "Licensing barriers" *Issues around "open source" and CC0 licensing.
**
People: Jacob B
*(12) "Barriers to model implementations and applications"* (I might
suggest this be re-phrased for better clarity). What this section should
discuss are challenges is getting a community to actually use models for
"real-world" applications or decision making. This is more of a
cultural/societal challenge, and thus seems like a nice big-picture way
to end.
*People: ?? *I don't have any names here...
*********************************************
We didn't really talk much about it in the Zoom meeting, but there have
been ideas tossed around about a "baker's dozen", i.e., adding a 13th
challenge. We could also potentially merge some of the above.
The "point persons" listed above is obviously a subset of co-authors.
That's fine and appropriate. Just for transparency, I follow what I
think is pretty standard policy for authorship issues, and nicely
summarized by theInternational Committee of Medical Journal Editors
(ICMJE); see 2019 updated document at
http://www.icmje.org/icmje-recommendations.pdf
<http://www.icmje.org/icmje-recommendations.pdf> (Or see, below my
signature, a summary of the key points of this document).
Finally, I've made the document editable by all at
https://docs.google.com/document/d/1VvyP3YZQdQYjj8DFKOpQ4pn_0pdDGgiT/edit?ts=60a294c2
-John G.
==========================================================================
Associate Professor & Graduate Program Director <gennari at uw.edu>
Dep't of Biomedical Informatics and telephone:206-616-6641
Medical Education, box 358047
University of Washington
Seattle, WA 98109-4714 http://faculty.washington.edu/gennari/
==========================================================================
The ICMJE recommends that authorship be based on the following 4 criteria:
1.Substantial contributions to the conception or design of the work; or
the acquisition, analysis, or interpretation of data for the work; AND
2.Drafting the work or revising it critically for important intellectual
content; AND
3.Final approval of the version to be published; AND
4.Agreement to be accountable for all aspects of the work in ensuring
that questions related to the accuracy or integrity of any part of the
work are appropriately investigated and resolved.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.simtk.org/pipermail/vp-reproduce-subgroup/attachments/20210517/fb52e45c/attachment-0001.html>
More information about the Vp-reproduce-subgroup
mailing list