WEBVTT
00:00:00.320 --> 00:00:02.720
Hello everybody, welcome to the Fire Science Show.
00:00:02.720 --> 00:00:12.800
In today's episode, we will be discussing how to pick the best tests to fulfill your needs, whatever the needs are in your fire safety engineering.
00:00:12.800 --> 00:00:15.439
And I've invited the nice guest, Dr.
00:00:15.439 --> 00:00:19.679
Andrea Franchini from University of Ghent to talk about this.
00:00:19.679 --> 00:00:27.120
And you may think we will be ranking test methods or giving you comparison between standards, and no, we're not gonna do that.
00:00:27.120 --> 00:00:40.640
Because what Andrea is doing is something quite more fundamental, which is gonna help those considerations about what test to use at pretty much any decision level you would be making.
00:00:40.640 --> 00:00:45.200
It is a part of a larger project, ATEST, by Dr.
00:00:45.200 --> 00:00:46.000
Ruben van Coile.
00:00:46.560 --> 00:00:50.640
Ruben is leading an ERC starting grant.
00:00:50.640 --> 00:00:53.840
We have discussed this in the podcast, I believe, two years ago.
00:00:53.840 --> 00:01:02.240
We had uh this uh podcast episode when Ruben was announcing his project and uh giving me an introduction about what he wants to do.
00:01:02.240 --> 00:01:18.719
Uh some years have passed, some work has been done, great work has been done, a massive paper was published, and here we are today able to discuss about the mathematical approach that is being used to decide upon the test methods.
00:01:18.719 --> 00:01:40.400
In this approach, instead of just thinking about uh the potential outcomes of a test instead of thinking about the criteria, pass fail criteria or other ways we would normally measure, this uses those, but uses those to inform what kind of information gain the test can give.
00:01:40.400 --> 00:01:52.319
This is a very specific term and Andrea will explain it later, but this kind of informs how much your expectations or centenities will move as you pursue the test.
00:01:52.319 --> 00:02:03.200
Because if you would like to do a test and it's not gonna move your knowledge by a bit, it's not gonna inform your decisions anymore, what's the point of running it?
00:02:03.200 --> 00:02:13.919
And if you have a limited budget, you would like to spend those money on the tests that will move your knowledge the most, so the reduced incentities the most or provide you some other metric.
00:02:13.919 --> 00:02:18.960
It's quite some complex mathematics and uh it's not an easy episode to be warned.
00:02:18.960 --> 00:02:32.879
But I think uh looking at the framework of what Ruben was proposing and looking today at the first outcomes and the proposals and the case study paper, I see this as a really, really good way.
00:02:32.879 --> 00:02:36.000
It's gonna be a challenge to introduce it into our paradigm.
00:02:36.000 --> 00:02:38.400
It doesn't work with our paradigm of fire testing.
00:02:38.400 --> 00:02:43.039
We need to change the paradigm to do this, but perhaps this is a good reason to change one.
00:02:43.039 --> 00:02:50.879
Um in the episode we go through two case studies, one related to concurmatry, one related to post-fire investigations.
00:02:50.879 --> 00:02:58.159
While you're uh listening to the episode, there's a Fire Safety Journal paper that covers the exact details of those two case studies.
00:02:58.159 --> 00:03:02.560
So if you have a chance uh to go through Fire Safety Journal paper, I would recommend doing that.
00:03:02.560 --> 00:03:08.400
And uh for now, please enjoy learning from Andrea about what has been done.
00:03:08.400 --> 00:03:11.520
Let's spin the intro and jump into the episode.
00:03:15.280 --> 00:03:17.360
Welcome to the Fire Science Show.
00:03:17.360 --> 00:03:21.039
My name is Wojciech Wegrzynski, and I will be your host.
00:03:36.319 --> 00:03:50.240
The Fire science Show is into its third year of continued support from its sponsor OFR consultants, who are an independent multi-award-winning fire engineering consultancy with a reputation for delivering innovative safety-driven solutions.
00:03:50.240 --> 00:04:04.000
As the UK leading independent fire consultancy, OFR's globally established team have developed a reputation for preeminent fire engineering expertise with colleagues working across the world to help protect people, property, and the planet.
00:04:04.000 --> 00:04:20.240
Established in the UK in 2016 as a startup business by two highly experienced fire engineering consultants, the business continues to grow at a phenomenal rate with offices across the country in eight locations, from Edinburgh to Bath, and plans for future expansions.
00:04:20.240 --> 00:04:28.959
If you're keen to find out more or join OFR consultants during this exciting period of growth, visit their website at OFRConsultants.com.
00:04:28.959 --> 00:04:31.040
And now back to the episode.
00:04:31.040 --> 00:04:32.160
Hello everybody.
00:04:32.160 --> 00:04:36.000
I am joined today by Andrea Franchini from Ghent University.
00:04:36.000 --> 00:04:37.519
Hey Andrea, good to have you in the show.
00:04:37.759 --> 00:04:38.000
Hi, Wojciech.
00:04:38.639 --> 00:04:39.519
Thanks for having me.
00:04:39.839 --> 00:04:46.720
Thanks for coming to discuss your interesting research carried out as a postdoc in the Ruben van Coile's uh ERC grant.
00:04:46.720 --> 00:04:48.480
How's working for ERC grant?
00:04:48.480 --> 00:04:51.680
We've advertised with Ruben uh the position in the podcast.
00:04:51.680 --> 00:04:53.519
I hope we have not overpromised.
00:04:54.000 --> 00:04:54.480
Yeah, yeah.
00:04:54.480 --> 00:04:57.680
I listened to the postcard before uh before applying for the position.
00:04:57.680 --> 00:04:59.600
And yeah, so far it has been great.
00:04:59.600 --> 00:05:03.199
It's a very mentally challenging and uh interesting project.
00:05:03.199 --> 00:05:10.160
And he gave us the opportunity to collaborate with a lot of people, and so it's very, very interesting project on my side.
00:05:10.480 --> 00:05:11.360
Yeah, super happy.
00:05:11.360 --> 00:05:21.519
And now seeing the first outcomes, or maybe not the first, but seeing the first big outcomes of the project that we're gonna discuss today, I'm I'm really excited for the next years of your research.
00:05:21.519 --> 00:05:26.399
So, Ruben had this presentation at ESFS conference in Ljubljana this year.
00:05:26.399 --> 00:05:31.120
Uh, he had a keynote on quantifying the utility of fire tests.
00:05:31.120 --> 00:05:38.000
I've talked with him and he told me that you've done all the hard jobs, so he wants you to uh speak on that, which I appreciate.
00:05:38.000 --> 00:05:48.079
Uh we know that the concept of Rubens ERC, the whole goal of A test was to like figure out how do we test in the future in such a way that it's the best.
00:05:48.079 --> 00:05:50.160
But now you have a framework.
00:05:50.160 --> 00:05:56.800
So let's introduce the listeners to the framework and perhaps let's start with the expected utility.
00:05:56.800 --> 00:05:58.000
I like that keyword.
00:05:58.000 --> 00:06:00.639
So so maybe let's start there and see where we get.
00:06:00.959 --> 00:06:01.920
Yeah, yeah, sure.
00:06:01.920 --> 00:06:07.120
So let me start with repeating the key problem we're trying to address.
00:06:07.120 --> 00:06:17.199
So we we want to understand which experimental protocol is best, meaning which experimental protocol uh experimenter should choose among all the options that he has.
00:06:17.199 --> 00:06:19.199
Should he go for uh a furnace test?
00:06:19.199 --> 00:06:22.000
Should you take a test in a concalorimeter?
00:06:22.000 --> 00:06:22.879
Which one?
00:06:23.279 --> 00:06:24.720
Assuming you have a freedom to choose.
00:06:24.959 --> 00:06:26.240
Assuming you have the freedom to choose.
00:06:26.240 --> 00:06:30.319
Yes, that's part of the FRPEST framework we we envision.
00:06:30.319 --> 00:06:42.399
So, assuming that you have this freedom, the choice among alternative experimental protocols is challenging because of several reasons, including that experimental outcomes are uncertain.
00:06:42.399 --> 00:06:45.600
Experiments are costly and time consuming.
00:06:45.600 --> 00:06:47.920
Probably you will know that better than me.
00:06:47.920 --> 00:06:54.959
And you can also have an environmental impact of your experiments, which you may want to account in your decision making.
00:06:54.959 --> 00:07:10.319
So the scope of our framework is to answer the question of which experimental protocol is best and we should choose by quantifying the expected utility of that experimental protocol before we actually conduct the experiment.
00:07:10.319 --> 00:07:41.279
So, in other words, we want to assess the potential benefit of collecting additional information through the experimental protocol we are planning, and we want to do that before doing the experiment, incorporating the available knowledge we have about the parameters we want to study and reflecting the experimental goals, which could be reducing uncertainty, reducing the economic impact of this uncertainty, or reducing, for example, the environmental impact of this uncertainty.
00:07:41.279 --> 00:07:48.240
So, in this sense, the expected utility you quantify captures the scopes of your experiment.
00:07:48.480 --> 00:07:52.399
And that utility is directly linked to your design goal.
00:07:52.399 --> 00:08:05.920
So, in what case you would be using that method, for example, you have to apply some very specific technology, and uh, I don't know, you want to know which product is the best to fit, and then you figure out which tests to do for that.
00:08:06.240 --> 00:08:08.800
Yeah, so this applies to both tests.
00:08:08.800 --> 00:08:15.199
So you want to, for example, demonstrate compliance or you want to classify a product, and it also applies to experiments.
00:08:15.199 --> 00:08:26.160
So for uh explorative experiments, you want to, for example, optimize your testing setup, you want to decide where you should put your sensors, and the utility definition aligns with your scope.
00:08:26.160 --> 00:08:30.879
So, for example, let's say you want to reduce uncertainty about some parameter.
00:08:30.879 --> 00:08:43.600
In that case, you can define a utility function that quantifies how much, in expectation, the outcomes of your experiment will reduce the uncertainty in those target parameters that you're tackling.
00:08:44.000 --> 00:08:54.240
Now, now that you said that I had the same feeling when Ruben was presenting in in Ljubljana that this is a very kind of universal framework which you can twist into different settings.
00:08:54.240 --> 00:09:00.320
You've uh twisted and fine-tuned it into uh fire testing and uh and and fire safety engineering.
00:09:00.320 --> 00:09:01.519
But indeed, you know what?
00:09:01.519 --> 00:09:07.919
When I when I heard Ruben's talk, I immediately thought about CFD and zone model uh dilemma.
00:09:07.919 --> 00:09:12.799
Like, is it better to run one CFD or uh or a thousand zone models in the same time?
00:09:12.799 --> 00:09:23.039
You know, because uh it's also like a kind of a challenge where you have like different centered, different levels of insight into the problem cost limitations.
00:09:23.039 --> 00:09:24.240
I I love it, I love it.
00:09:24.240 --> 00:09:28.559
But let's go back to the tests, to your applications that you've described.
00:09:28.559 --> 00:09:34.879
So in your paper, you have uh some practical examples how these have been implemented.
00:09:34.879 --> 00:09:38.240
So we'll probably discuss more of those in the discussion today.
00:09:38.240 --> 00:09:46.799
Perhaps let's introduce the listeners to a representative problem which we could then solve through the discussion, applying your your your thing.
00:09:46.799 --> 00:09:49.200
So uh let's let's go to the case study.
00:09:49.440 --> 00:10:07.039
Yeah, so for example, one of the case studies we presented pertaining the con calorimeter testing, and we wanted to understand how many tests we should run with the con calorimeter in order to reduce uncertainty in estimating the ignition time of a PMMA batch.
00:10:07.039 --> 00:10:08.480
That was the idea.
00:10:08.480 --> 00:10:26.080
And to answer this question, we apply the framework that we are discussing and uh we define a utility metric that captures the amount of uncertainty you have in different state of knowledge before you do the experiment and after you do the experiment.
00:10:26.080 --> 00:10:40.720
And then using this calculation that we can discuss more in detail uh later, you basically get an estimate of the number of tests that in expectation will minimize your uncertainty in the prediction of this ignition time.
00:10:40.960 --> 00:10:46.960
Okay, okay, so but but what are the things that you are playing with, for example, in the con calorimeter?
00:10:46.960 --> 00:10:57.679
Because, like, I mean, one way you could do it, the the classical way I would do it is I would just run the tests until I get some sort of convergence and say, okay, you know, this seems enough.
00:10:57.679 --> 00:11:03.759
But of course, uh, I do not know when the convergence will happen uh before I start doing that.
00:11:03.759 --> 00:11:04.720
So exactly.
00:11:04.720 --> 00:11:09.200
So uh this allows me to predict that the that that time when the convergence okay.
00:11:09.200 --> 00:11:10.159
So what do you say with it?
00:11:10.399 --> 00:11:13.840
Yeah, maybe I I can explain more what we mean by experimental protocol.
00:11:13.840 --> 00:11:18.799
Yes, so by experimental protocol, we mean a combination of experimental procedures.
00:11:18.799 --> 00:11:25.039
So you may have the concalorimeter, you may have the furnace test, and so on, and experimental design parameters.
00:11:25.039 --> 00:11:28.960
So, how do you tune your testing setup to run the experiment?
00:11:28.960 --> 00:11:33.759
For example, in the co-calorimeter, what is the heat flux exposure that you're going to use?
00:11:33.759 --> 00:11:39.679
Uh in in a different test, you may choose the temperature to which you're going to expose the sample, and so on.
00:11:39.679 --> 00:11:50.480
And in the case study of the ignition time that we have introduced, we say, okay, we already decided we want to use concalorimeter, so that's a fixed parameter.
00:11:50.480 --> 00:11:58.559
And the only experimental design parameter that we want to investigate is the number of con-calorimeter tests that we should run.
00:11:58.559 --> 00:12:03.519
So in this case, we just limit the analysis to one experimental design parameter.
00:12:03.519 --> 00:12:11.759
And we want to calculate how many tests we should do so that the uncertainty in predicting the ignition time is minimized.
00:12:11.759 --> 00:12:13.360
That's the that's the idea.
00:12:13.360 --> 00:12:17.360
But in in principle, you can include many more experimental design parameters.
00:12:17.360 --> 00:12:24.159
For example, the it flux and other variables that you can play with in with your experimental setup.
00:12:24.399 --> 00:12:31.120
And uh the procedure will uh look very very similar, no matter how many outcome variables you include.
00:12:31.120 --> 00:12:32.320
What changes?
00:12:32.639 --> 00:12:35.200
Yeah, so conceptually it works the same.
00:12:35.200 --> 00:12:48.559
The problem is that it becomes more challenging computationally and also conceptually, I think, if you have many design parameters, because you need to build models that are able to capture all your experimental design parameters.
00:12:48.559 --> 00:12:51.200
So conceptually, it works exactly the same.
00:12:51.200 --> 00:13:01.120
You can include as many experimental design parameters as you want, but you need to formulate the problem in a way such that it can account for all these experimental design parameters.
00:13:01.120 --> 00:13:04.159
You will need models that reproduce your experiment.
00:13:04.159 --> 00:13:08.799
So, whatever you want to optimize, you need a model that is able to capture that parameter.
00:13:09.200 --> 00:13:10.639
Yeah, that's what I wanted to ask.
00:13:10.639 --> 00:13:18.000
So, optimally, if it's a physical model, but do you need to also know the shape of the distribution of the outcome variable?
00:13:18.240 --> 00:13:26.240
Yeah, so maybe I can give you an overview of how the analysis works so that we know all the different uh pieces that we need.
00:13:26.799 --> 00:13:35.759
Maybe let's let's even step one step before because I feel it's gonna be interesting and and and intellectually engaging and challenging discussion.
00:13:35.759 --> 00:13:38.320
You use Bayesian framework for this.
00:13:38.320 --> 00:13:43.840
So maybe let's start with the framework itself and then build up into how this is applied.
00:13:43.840 --> 00:13:50.000
And let's try to keep it all in the the time to ignition of PMMA stage that we've set up.
00:13:50.000 --> 00:13:50.720
So that's great.
00:13:51.200 --> 00:13:57.919
Yes, Bayesian analysis is at the core of the methodology, and there are three main ideas in this methodology.
00:13:57.919 --> 00:14:05.360
One is the concept of state of knowledge, the second one is the concept of Bayesian analysis, and then the concept of utility.
00:14:05.360 --> 00:14:08.639
So let me link these three concepts for you.
00:14:08.639 --> 00:14:15.600
So we generally express our uncertain state of knowledge in terms of probability distributions.
00:14:15.600 --> 00:14:28.720
For example, back to the ignition time of PMMA, you may say, I have uncertainty about the ignition time, and to describe this uncertainty, I assign a distribution, let's say a normal distribution with a mean and a standard deviation.
00:14:29.039 --> 00:14:34.159
And then people do this all the time when they say like it's uh one minute plus minus 10 seconds.
00:14:34.159 --> 00:14:39.759
That's already a distribution you've given in this in this information, even if you don't think it's a distribution, you know.
00:14:39.759 --> 00:14:40.080
Okay.
00:14:40.480 --> 00:14:40.960
Absolutely.
00:14:40.960 --> 00:14:41.279
Yes.
00:14:41.279 --> 00:14:49.120
So you basically express your state of knowledge in terms of probability distributions that reflect uncertainty you have about some parameters.
00:14:49.120 --> 00:14:52.639
And second concept that introduced is Bayesian analysis.
00:14:52.639 --> 00:14:53.759
So, what's that?
00:14:53.759 --> 00:14:59.840
It's a method to compute and update probabilities after obtaining new data.
00:14:59.840 --> 00:15:05.759
Okay, so the key the all the Bayesian analysis is based on Bayes theorem.
00:15:05.759 --> 00:15:13.120
And the key idea of Bayes theorem is that evidence should not determine your state of knowledge, but it should update it.
00:15:13.120 --> 00:15:14.320
So, what does it mean?
00:15:14.320 --> 00:15:32.960
It means that if we, as we said, we assign a distribution to represent our current state of knowledge, we do the experiment, we get some data, and we use base theorem to get a second distribution that reflects your updated state of knowledge based on the observed experimental outcome.
00:15:32.960 --> 00:15:41.840
And for example, if you go back to the concrete method of testing, you have a distribution that reflects your knowledge of the ignition time before you do the experiment.
00:15:42.159 --> 00:15:45.200
So it's let's say 1210 seconds plus minus 30.
00:15:45.440 --> 00:15:46.159
Yes, exactly.
00:15:46.159 --> 00:15:52.559
You you can say it's a normal distribution with mean uh uh 210, as you said, and a standard deviation, as you said.
00:15:52.559 --> 00:16:06.159
So you do the experiment, you get some data points, and using Bayes' theorem, you can calculate a second probability distribution that somehow accounts for this updated knowledge that you get from your new data.
00:16:06.159 --> 00:16:12.480
So you get a second distribution that represents your updated state of knowledge after you do the experiment.
00:16:12.879 --> 00:16:16.399
So for example, now it's 207 plus minus 17.
00:16:16.399 --> 00:16:35.200
I do another one and I find it's plus minus 16, but eventually every time I do a new one, I just get the result within my standard deviation, it fits the pattern, and eventually I end up with the final standard uh normal distribution curve, which does not really change that much every time I run the experiment, potentially.
00:16:35.440 --> 00:16:39.360
Yeah, it may change or it may not, because if your experiment I don't know that, yes.
00:16:39.360 --> 00:16:40.639
Yeah, yeah, exactly.
00:16:40.639 --> 00:16:51.200
So if your experiments confirm your uh state of knowledge, meaning, for example, you get a result very close to your mean, to your estimating mean before you do the experiment, the distribution will remain essentially the same.
00:16:51.200 --> 00:17:01.440
But if you get, for example, an experimental data point very far from your initial state of knowledge, your distribution will somehow translate towards that value.
00:17:01.440 --> 00:17:11.279
So you will get an updated state of knowledge that tries to average your prior knowledge and the evidence that you get from the experiment.
00:17:11.519 --> 00:17:17.039
Yeah, the the that's probably a biggest challenge in here, like the outliers and how do you handle them.
00:17:17.039 --> 00:17:33.039
Because okay, going back to the world of fire science, it's not that you can afford to run cone colorimeter on PMMA 10,000 times, because there are also like economical and time restraints, and you probably would burn through three cone colorimeters running 10,000 samples anyway.
00:17:33.039 --> 00:17:37.359
So, yeah, there are physical limits in how much you can do.
00:17:37.359 --> 00:17:44.480
And if in general, we are usually working with very low sample sizes in fire science.
00:17:44.480 --> 00:17:46.240
Like if you have three, you're good.
00:17:46.240 --> 00:17:48.079
If you have five, you're plenty.
00:17:48.079 --> 00:17:54.480
If you have 30, you're David Morrison burning chairs because no one else does like 30 samples of one thing, right?
00:17:54.480 --> 00:18:03.200
Does this theorem also have a way to handle outliers, or you just you get one, you you you you face the consequences, they have to figure it out?
00:18:03.519 --> 00:18:14.799
Yeah, so in the theorem you have something called likelihood, which is the probability of observing that specific data point based on your prior state of knowledge.
00:18:14.799 --> 00:18:27.440
So if the data that you observe is very far from your current knowledge, it will essentially be assigned a very low probability so that it will uh tend to influence less your state of knowledge.
00:18:27.440 --> 00:18:29.200
So you you account for this.
00:18:29.599 --> 00:18:37.039
But in this approach, in no point of this kind of consideration, you try to understand why the deviation occurred, right?
00:18:37.039 --> 00:18:40.880
Because here it's just about the statistics of the outcome.
00:18:40.880 --> 00:18:56.720
While in reality, for example, uh someone was doing cone calorimeter tests on 25 kilowatts, the next day someone was doing on 50, and the third day a guy came back and didn't notice it's changed, and and maybe run a test in wrong conditions, so it could have been an error.
00:18:56.720 --> 00:19:02.720
So so but but the the theorem will not like it's a separate analysis to to clean out the data, I guess.
00:19:03.279 --> 00:19:08.799
Yes, so you you need to be critical about your data before you use the theorem.
00:19:08.799 --> 00:19:17.359
But one uh nice and elegant thing of Bayes theorem is that it's just centered on updating your belief.
00:19:17.359 --> 00:19:24.880
So even if you get your outlier, for example, once you can update your self-knowledge based on that outlier.
00:19:24.880 --> 00:19:28.880
So it's up to you to say, should I use the outlier and should I understand it?
00:19:28.880 --> 00:19:32.880
Yes, you definitely should, but that's not the the point of base theorem.
00:19:32.880 --> 00:19:47.119
You in other words, you don't need to take random data and throw them inside, although you can, you can do that, and you will still get a posterior distribution that reflects your updated state of knowledge based on the data that you give to the tier.
00:19:47.519 --> 00:20:01.519
Yeah, and I assume even if you did that and then you have uh hundreds of data points from your normal data, the the the theorem would result in a distribution that that's closer to the original and outlier will not impact it that much.
00:20:01.519 --> 00:20:03.200
I I that that's what I guess.
00:20:03.200 --> 00:20:03.759
Lovely.
00:20:03.759 --> 00:20:04.400
Yeah, yeah.
00:20:04.400 --> 00:20:09.680
And now how does uh okay, so so we know the the base theorem assumptions uh now.
00:20:09.680 --> 00:20:14.240
How do you apply this in in the in the technicalities of of the testing?
00:20:14.480 --> 00:20:20.559
Yeah, so the first component that's I mentioned for this framework is the concept of utility, right?
00:20:20.559 --> 00:20:24.240
So we define state of knowledge, we explain what Bayesian analysis does.
00:20:24.240 --> 00:20:31.279
So you update your state of knowledge, meaning you get a new distribution representing your state of knowledge after the experiment.
00:20:31.279 --> 00:20:41.599
And then what you can do is to assign to this state of knowledge a metric that describes how that state of knowledge is desirable to the user.
00:20:41.599 --> 00:20:43.200
How do we do that?
00:20:43.200 --> 00:20:52.480
We define utility function that quantifies the desirability of that state of knowledge based on your objectives.
00:20:52.480 --> 00:21:09.039
So, for example, if you're interested, as we are discussing for the case of the concolorimeter, in reducing uncertainty, you can define a utility metric that tells you how much uncertainty you have in your prior knowledge and how much uncertainty you have in your updated knowledge.
00:21:09.039 --> 00:21:19.839
And then you take the difference between the two and you assess whether your experimental data reduces the uncertainty with respect to your prior knowledge or increase it.
00:21:19.839 --> 00:21:20.319
Okay.
00:21:20.319 --> 00:21:24.640
So all this is if you have done the experiment, right?
00:21:24.640 --> 00:21:29.200
After you do it, after you have done the experiment, you can do all these calculations that I described.
00:21:29.200 --> 00:21:36.160
But as I mentioned in the beginning, the framework aims at calculate making this analysis before we do the experiment, right?
00:21:36.160 --> 00:21:37.359
So how do we do that?
00:21:37.359 --> 00:21:37.920
That's clear.
00:21:38.240 --> 00:21:41.759
So you want to budget how many cone calorimeters you want in your research grant.
00:21:41.759 --> 00:21:44.000
Do I need to go for five or fifty?
00:21:44.160 --> 00:21:44.480
Yeah.
00:21:44.799 --> 00:21:46.720
Your boss says, let's go for a hundred.
00:21:46.720 --> 00:21:48.880
We need uh exactly, exactly.
00:21:49.039 --> 00:21:51.200
And you want to do that before you run the experiment, right?
00:21:51.359 --> 00:21:51.759
Yes, yes.
00:21:52.079 --> 00:21:53.119
So, how do you do this?
00:21:53.119 --> 00:21:56.640
You need a model that simulates your experimental outcomes.
00:21:56.640 --> 00:21:59.599
Okay, based on your uncertain parameters.
00:21:59.599 --> 00:22:05.279
And you simulate multiple times the possible experimental outcomes of your experiment.
00:22:05.279 --> 00:22:18.160
You use the Bayesian theory that I described before for each of those outcomes, and for each of those analyses, you get an estimate of the utility of the experiment if the outcome was the one you simulate.
00:22:18.160 --> 00:22:25.599
And then you take the expected value of all these outcomes that you get, and that represents the expected utility of your experiment.
00:22:25.920 --> 00:22:54.559
So, as I understand you for the case of cone, a model of a cone would be some sort of uh, I don't know, let's say a machine learning of a cone based on a thousand samples or some formal previous statistical distribution, and then you expect the based on literature that the time of ignition of this material is like 300 seconds, and maybe the standard deviation would be 30 seconds, so you have no some expected outcome.
00:22:54.559 --> 00:23:02.720
Then, based on some model, you run the analysis and see that you potentially can reduce the uncertainties there.
00:23:02.720 --> 00:23:09.839
Well, in in this case, you probably can just use the model to predict your time to ignition if you have such a good model that predicts it.
00:23:09.839 --> 00:23:18.559
But I assume the wealth of this method comes when you have multiple tests to choose from and uh multiple uh utilities uh to balance.
00:23:18.559 --> 00:23:21.279
So that that that's when it it comes into play.
00:23:21.279 --> 00:23:28.960
The gain of information is this what was described in the papers as the information gain or the gain in utility?
00:23:28.960 --> 00:23:30.160
Yes, exactly.
00:23:30.160 --> 00:23:32.720
Let's introduce this concept to the listeners.
00:23:32.720 --> 00:23:37.920
So it's about how much more information you get per repeat of experiment.
00:23:37.920 --> 00:23:39.680
Is that I understand it correctly?
00:23:40.160 --> 00:23:40.480
Exactly.
00:23:40.480 --> 00:23:40.880
Yes.
00:23:40.880 --> 00:23:46.799
So what we did was calculating we we chose different number of tests in a concalorimeter.
00:23:46.799 --> 00:24:01.680
So we went from uh one all the way up to 30 tests, and we calculated for each of these possible number of tests, how much in expectation that number of tests will reduce uncertainty in the distribution of the ignition time.
00:24:01.680 --> 00:24:11.440
And then we, using the calculation that I described, we get that if you start at low trials, you see a very large increase in information gain.
00:24:11.440 --> 00:24:16.880
So any trials, like you have uh one, three, and so on, will give you a lot of information.
00:24:16.880 --> 00:24:19.839
But then you start seeing a plateau in the curve.
00:24:19.839 --> 00:24:30.400
So you see that the expected information gain that you get from running many tests reduces until at some point the marginal gain in this information is basically zero.
00:24:30.400 --> 00:24:40.240
Which means that if you want to understand an optimal number of tests that you should run, you should stop when this marginal expected information gain basically goes to zero.
00:24:40.240 --> 00:24:47.759
Because beyond that, based on your models, based on your assumption, the experiment will not give you more information.
00:24:48.079 --> 00:24:48.640
Brilliant.
00:24:48.640 --> 00:24:57.119
It's it's such a challenging concept because at some point like you may also be seeking those outliers, like to understand what happens in the outlier case.
00:24:57.119 --> 00:25:00.720
This is basically how the particle physics works, you know.