Modelling is a key component of science. So, what is modelling? What different types of models are there, and what is important to keep in mind when modelling? Curious to learn more, we talked to Prof. Marina Axelson-Fisk, Professor in Mathematical Statistics at Chalmers University of Technology, who shares some of her vast knowledge on mathematical modelling and how this is used in science.

A mathematical model, or a model in general, is an abstract representation of some real-life object, Prof. Axelson-Fisk says. For example, a sculpture - that is a model of whatever it is portraying. And a painting is a model of a picture. The mathematical model is similar, but in formulas. So, we are representing things in the real world with formulas. Typically, you want to relate different factors, features, or entities in the real world and see how they relate. So that is a brief explanation of the model - it is an abstract version of reality, of different features in the reality. As such, it is of course, a simplification of the real world, Prof. Axelson-Fisk says.

The mathematical model represents a phenomenon or process and in the mathematical formula, you have variables representing the features or entities that you want to measure, Prof. Axelson-Fisk says. Then, you have parameters that are constants that adjust how the variables relate to each other. Those are the typical formulations in a in a mathematical model. The variable is what you measure, i.e., these are the unknowns that you want to understand more of.

In science, you often want to either determine the relationship between features, i.e., you want to understand how different features in the real world relate to each other or affect each other, or you may want to understand how different features interact with each other, Prof. Axelson-Fisk says. So, the combination of features may have a different influence on whatever you measure compared to each of them separately. Another scenario is where you want to predict something. Say for example that you want to predict something that is difficult to measure but you have related entities that are easy to measure and that are connected. Then you make a mathematical formulation of how these entities connect to the thing you want to measure, Prof. Axelson-Fisk says.

This is very abstract, so let us take an example, for example predicting the weather. Typically, you predict the weather by looking at the weather history. The weather tomorrow is related in various ways to the weather today. In an advanced approach you could for example use satellite pictures which contain valuable information about historic weather, and these can be used as input in the model to predict the weather in the future, Prof. Axelson-Fisk says.

You use the model to predict things or to classify things, so, it can be used in many ways. In science, in applied science, you usually want to test out relationships in various ways using hypothesis testing, for instance. And then you need a model, Prof. Axelson-Fisk says.

There are many different types of models, and they can be grouped into different categories. One important distinction is whether the model is *linear* or *nonlinear*. In a linear model, there is a linear relationship between the variables. If you only have two variables that relate to each other, you can draw a line. If you have more variables, then it becomes more complex, but you still have a linear relationship. This means that if you have more than one feature affecting the outcome of something, then the different features add up. So, you have the feature times a parameter constant and then you add them up.

In a nonlinear relationship on the other hand, you can have all kinds of functions. One example is the exponential function, i.e., ‘e’ to the power of your variable. The distinction between linear and nonlinear models is important because linear relationships are much easier to handle, both to train the model and to interpret it. In many applications, even though the relationship maybe is not exactly linear, you can often get away with approximating a linear relationship, Prof. Axelson-Fisk says.

As an example of a linear relation, I can mention my own master thesis, which I did for quite a while ago, Prof. Axelson-Fisk says. Then I made a model to estimate the age of fetuses. This would be useful in a situation where you have a pregnant woman that does not know exactly which week she is in and would like to estimate the age of the fetus. I made a model where the length of the femur and the diameter of the skull were used as input. We had several data points, i.e., several fetuses where we knew the age and corresponding values of the lengths of the femurs and diameters of the skulls. Then we related those parameters in a linear way to estimate the fetus’s age. When there was a new pregnant woman, we measured these parameters to estimate the age of the fetus. So that is an example of a linear relationship, Prof. Axelson-Fisk says.

In a *deterministic* model, all your variables are set. They can be unknown, but typically the same input will always yield the same output, Prof. Axelson-Fisk says. So, it is a fixed or constant relationship. In *stochastic* models, on the other hand, you have some addition of randomness. You may have some noise in your measurements, so you have a distribution over your variables, i.e., it is not an exact relationship. That is the main distinction. And then, instead of having measured variables, i.e., determined by the measurement, you also add on the randomness or the noise through distribution of random values.

The third important distinction is whether the model is *discrete* or *continuous*, i.e., the variables are discrete or continuous, Prof. Axelson-Fisk says. Typically, if you have a measurement device, you will get something that is continuous meaning that you can get any value between any two values. In a discrete model, however, you can separate the values and count them. They can be infinitely many, but you would still be able to count them, Prof. Axelson-Fisk says.

So how do you create a model from scratch, or even just continue to work on something that exists but does not really describe the world in a good enough way? So, first you need data, Prof. Axelson-Fisk says. If you want to model the relationships, you need some data and then you need to see how they relate in various ways. The relationships can be explored by visualizing the data, plotting it in different ways. Then you would see, for instance, if it is a linear relationship or if there is some other pattern.

To start completely from scratch is difficult though. You need to have some idea of what relations and or phenomenon you are studying, so you need a lot of knowledge. Typically, you would also survey the field so you would investigate what different methods there are and pros and cons with those. And you need to determine whether you want a deterministic or stochastic model, whether it should be linear or nonlinear. You also need to consider that every model is simplification of the real world, Prof. Axelson-Fisk says. But there is a tradeoff between complexity and accuracy. The more complex your model is, the more accurate it is. But on the other hand, it will be much harder to handle, and you need a lot of data to train it. So, you must weigh the complexity towards the accuracy in some sense, Prof. Axelson-Fisk says.

You could also try to invent new models by combining existing ones. That is the kind of work I do a lot. I work typically on creating new models and we rarely develop them from scratch, but instead use existing models and tweak them or combine different parts, Prof. Axelson-Fisk says.

To adjust the model, you have to have some sort of idea of what changes to make. If you see that something does not fit, then you need to figure out why it does not fit. For instance, if you are doing a linear regression model where you have an output and then you have one or several input variables, if you have done it as a linear relationship, then you can check whether this is a good assumption. And if it is not, then you might to want to change it to adapt your function data. If you see that there seem to be dependencies between the data, you need to take care of that. Maybe there are additional features that need to be included in the model and so on, Prof. Axelson-Fisk says.

There is a tradeoff between model complexity and accuracy, so where to draw the line in terms of trying to improve the model depends on what you are doing, Prof. Axelson-Fisk says. This could be considered already from the start - how big variation can I accept? If you have a stochastic model, it is about how big variation the data is. If the variation is too big, you need to either have a better model or have a larger dataset and so on. If you are training your model iteratively, for instance, then then you can for example have some error function that tells you how far from the truth you are when you train on annotated data. You can say that when I'm within this interval from that, then my parameter estimates are good enough and I am done. Or, when the error does not change between iterations, i.e., it becomes constant or near constant, then that would be a place to stop, because then it you cannot do it better, Prof. Axelson-Fisk says. If that is still not good enough, then you need to review your data and your model. This relates to the design of experiments that we talked about in a previous pod episode. You need to think things through and try to understand if there are variables that you cannot control which are still affecting your output. Then you need to figure out if you can solve that somehow. For instance, if you have some model where you measure things, and the temperature matters, then you may try to measure under the same temperature every time, under the same circumstances and so on. So, there are many things to think of and that you can tweak, but that that can also affect your results, Prof. Axelson-Fisk says.

Listen to the full interview with Prof. Axelson-Fisk to learn more about mathematical modelling, typical challenges and pitfalls to avoid.

Topics:

Learn about what aspects to consider after you have run a QCM measurement

Get a checklist that will help you optimize the reproducibility of your QCM-D data by minimizing unintentional changes of the QCM-D parameters.

Learn about what aspects to consider when running your QCM measurement

Learn the guidelines on how to assess which method to use to quantify QCM mass.

Learn about what aspects to consider when setting up your QCM measurement

Temperature stability is critical for reliable QCM data. Here are the top four factors that will help you eliminate temperature induced artifacts.

Read about the key signatures in the QCM data that reveal if viscoelastic modelling should be used to extract the mass.

QCM-sensor regeneration can be a resource efficient way of running experiments. Learn about aspects to consider if you plan to resuse ypur sensors.

You have only scratched the surface.

- QCM-D (105)
- Technology (45)
- Contact Angle (34)
- Surface science (20)
- Surface tension (19)
- Tips & Tricks (19)
- Wettability (18)
- Adhesion (15)
- Protein Adsorption (15)
- Interfacial rheology (14)
- Biomaterials (13)
- Enhanced oil recovery (13)
- Nanoparticles (13)
- Biomembranes (12)
- Surfactants (12)
- Batteries (11)
- QSense (11)
- Biocompatibility (10)
- Dynamic contact angle (10)
- Superhydrophobic surfaces (10)
- Thin Film Deposition and Characterization (10)
- Nanoparticle deposition (9)
- Virus (9)
- Polymers (8)
- Detergent (7)
- Cleaning (6)
- Combined QCM-D and Ellipsometry (6)
- Lipids (6)
- EQCM-D (5)
- Surface Free Energy (5)
- Surface Roughness (5)
- Nanoparticle thin film (4)
- Nanotechnology (4)
- Transformer oil (4)
- protein (4)
- CMP (3)
- Food Science (3)
- Protein Aggregation (3)
- graphene (3)
- CO2 sequestration (2)
- Langmuir-Blodgett (2)
- Paper and Board (2)
- SPR (2)
- thin film (2)
- Antifouling (1)
- Attension (1)
- Biosensors (1)
- Clean tech (1)
- Critical micelle concentration (1)
- Environment (1)
- Etching (1)
- Fouling (1)
- Lipid Nanoparticles (1)
- Nanoparticle characterization (1)
- Nanotoxicology (1)

- Susanna Laurén (149)
- Malin Edvardsson (99)
- Gabriel Ohlsson (8)
- Jennie Ringberg (8)
- Erik Nilebäck (7)
- Prof. Bengt Kasemo (7)
- Eva Ekerot (6)
- Kenneth Olesen PhD (6)
- Jyrki Korpela (5)
- Anna Oom (4)
- Michael Rodal, PhD (3)
- Maiju Pöysti (3)
- Ville Jokinen (3)
- Fredrik Pettersson (3)
- Gunilla Rydén (3)
- Henni Mäntykangas (2)
- Brian Rodenhausen PhD (2)
- Fredrik Andersson (2)
- Alaric Taylor (2)
- Matthew Dixon (2)
- Déborah Rupert (2)
- Prof. Robin Ras (1)
- Eda Kaya (1)
- Farad Sagala (1)
- Dr. Stefan Guldin (1)
- Prof. Zhenghe Xu (1)
- Dr Luke A. Clifton (1)
- Olivia Uddenberg (1)

- May 2024 (3)
- April 2024 (5)
- March 2024 (4)
- February 2024 (4)
- January 2024 (5)
- December 2023 (3)
- November 2023 (4)
- October 2023 (5)
- September 2023 (4)
- August 2023 (5)
- July 2023 (4)
- June 2023 (4)
- May 2023 (5)
- April 2023 (4)
- March 2023 (4)
- February 2023 (4)
- January 2023 (5)
- December 2022 (3)
- November 2022 (3)
- October 2022 (5)
- September 2022 (4)
- August 2022 (5)
- July 2022 (4)
- June 2022 (4)
- May 2022 (5)
- April 2022 (4)
- March 2022 (5)
- February 2022 (4)
- January 2022 (3)
- December 2021 (2)
- November 2021 (5)
- October 2021 (3)
- September 2021 (4)
- August 2021 (5)
- July 2021 (4)
- June 2021 (5)
- May 2021 (4)
- April 2021 (4)
- March 2021 (5)
- February 2021 (4)
- January 2021 (4)
- December 2020 (5)
- November 2020 (4)
- October 2020 (4)
- September 2020 (5)
- August 2020 (3)
- July 2020 (4)
- June 2020 (4)
- May 2020 (4)
- April 2020 (5)
- March 2020 (5)
- February 2020 (4)
- January 2020 (3)
- December 2019 (4)
- November 2019 (6)
- October 2019 (6)
- September 2019 (4)
- August 2019 (1)
- July 2019 (4)
- June 2019 (3)
- May 2019 (3)
- April 2019 (4)
- March 2019 (4)
- February 2019 (3)
- January 2019 (3)
- December 2018 (1)
- November 2018 (1)
- October 2018 (5)
- September 2018 (2)
- August 2018 (4)
- July 2018 (4)
- June 2018 (4)
- May 2018 (6)
- April 2018 (4)
- March 2018 (3)
- February 2018 (3)
- January 2018 (2)
- December 2017 (2)
- November 2017 (1)
- August 2017 (2)
- April 2017 (1)
- March 2017 (1)
- January 2017 (1)
- December 2016 (1)
- November 2016 (1)
- October 2016 (1)
- June 2016 (1)
- May 2016 (1)
- April 2016 (3)
- January 2016 (1)
- December 2015 (2)
- November 2015 (1)
- October 2015 (3)
- September 2015 (2)
- August 2015 (1)
- July 2015 (1)
- June 2015 (2)