1 00:00:00,670 --> 00:00:02,720 - [Instructor] In the second part of the lecture, 2 00:00:02,720 --> 00:00:04,870 we'll look at the introduction to geostatistics 3 00:00:04,870 --> 00:00:07,900 and interpolation concepts. 4 00:00:07,900 --> 00:00:10,060 I won't lie, we're gonna get pretty close 5 00:00:10,060 --> 00:00:11,360 to some mathematics here 6 00:00:11,360 --> 00:00:13,490 but I think we make it out of the presentation 7 00:00:13,490 --> 00:00:16,200 without having to look at any formulas 8 00:00:16,200 --> 00:00:18,220 I'll be offering a relatively light treatment 9 00:00:18,220 --> 00:00:20,300 of both of these topics. 10 00:00:20,300 --> 00:00:22,940 So there's definitely plenty more to dig in on. 11 00:00:22,940 --> 00:00:24,820 If you're interested in the mathematics 12 00:00:24,820 --> 00:00:27,390 or just wanna know more in general, let me know, 13 00:00:27,390 --> 00:00:30,663 and I can find some additional resources for you to review. 14 00:00:33,206 --> 00:00:35,340 Geostatistics is a statistical approach 15 00:00:35,340 --> 00:00:40,340 for analyzing spatial and spacial temporal phenomenon. 16 00:00:40,410 --> 00:00:42,380 It considers both the location 17 00:00:42,380 --> 00:00:45,740 and the value of the observation at that location 18 00:00:45,740 --> 00:00:47,430 as part of the analysis. 19 00:00:47,430 --> 00:00:51,420 We can use geostatistics to describe patterns in our data 20 00:00:51,420 --> 00:00:54,410 to interpolate values or fill in the blanks 21 00:00:54,410 --> 00:00:57,310 between known point locations 22 00:00:57,310 --> 00:00:59,443 and even to estimate uncertainty. 23 00:01:01,949 --> 00:01:03,780 Geostatistics has evolved 24 00:01:04,990 --> 00:01:08,350 to be embraced by a broad range of disciplines 25 00:01:08,350 --> 00:01:10,603 for any number of applications. 26 00:01:11,720 --> 00:01:14,630 When we're talking about spatial interpolation, 27 00:01:14,630 --> 00:01:16,600 we consider the prediction of a variable 28 00:01:16,600 --> 00:01:20,260 at unmeasured location based on the value of variables 29 00:01:20,260 --> 00:01:21,720 at measured location. 30 00:01:21,720 --> 00:01:24,513 So out at sample points where you've collected data. 31 00:01:25,480 --> 00:01:28,700 Spatial prediction takes that one step further 32 00:01:28,700 --> 00:01:32,853 and includes other variables in the estimation process. 33 00:01:36,180 --> 00:01:39,500 Now, part of having data that you can use 34 00:01:39,500 --> 00:01:42,620 for geostatistical analysis includes sampling data 35 00:01:42,620 --> 00:01:45,020 or collecting data in the field, 36 00:01:45,020 --> 00:01:47,280 or potentially even just sampling data 37 00:01:47,280 --> 00:01:49,713 from a digital data set that you have. 38 00:01:50,850 --> 00:01:53,620 A couple of different kinds of sampling 39 00:01:53,620 --> 00:01:57,420 that we see on this page, systematic and random. 40 00:01:57,420 --> 00:02:00,030 Systematic samples are uniformly spaced 41 00:02:00,030 --> 00:02:02,170 over X and Y intervals. 42 00:02:02,170 --> 00:02:04,010 They're easy to plan and implement 43 00:02:04,010 --> 00:02:06,630 and it really removes any of the subjective judgment 44 00:02:06,630 --> 00:02:09,623 about where to collect sample data from. 45 00:02:10,700 --> 00:02:11,660 On the con side, 46 00:02:11,660 --> 00:02:13,810 it's not necessarily statistically efficient 47 00:02:13,810 --> 00:02:15,580 to collect data this way, 48 00:02:15,580 --> 00:02:18,730 and it certainly doesn't consider things like travel costs 49 00:02:18,730 --> 00:02:22,200 or inability to capture patterns at finer scales 50 00:02:23,310 --> 00:02:25,053 when using this type of program. 51 00:02:26,800 --> 00:02:29,585 Random samples on the other hand are just a random selection 52 00:02:29,585 --> 00:02:34,573 of X and Y coordinates usually selected separately. 53 00:02:35,830 --> 00:02:36,840 On the positive side, 54 00:02:36,840 --> 00:02:39,890 these are unlikely to match any pattern in the landscape 55 00:02:41,340 --> 00:02:44,970 meaning that it's good for capturing the diversity 56 00:02:44,970 --> 00:02:46,790 in the landscape itself, 57 00:02:46,790 --> 00:02:48,620 but on the con side 58 00:02:48,620 --> 00:02:49,964 it doesn't increase the number of samples 59 00:02:49,964 --> 00:02:52,403 in areas with high variation, 60 00:02:54,030 --> 00:02:56,650 so it's likely to miss some of those patterns 61 00:02:56,650 --> 00:02:59,163 and it's more challenging to plan and implement. 62 00:03:01,754 --> 00:03:04,410 We've also got an option for cluster sampling 63 00:03:04,410 --> 00:03:08,480 which kind of combines the random and systematic selection 64 00:03:09,380 --> 00:03:11,992 of the cluster centers and the clusters 65 00:03:11,992 --> 00:03:16,992 that actually surround those individual point locations. 66 00:03:17,000 --> 00:03:18,110 The US Forest Service 67 00:03:18,110 --> 00:03:21,840 uses a random cluster center point selection 68 00:03:21,840 --> 00:03:24,649 followed by systematic designation of the cluster 69 00:03:24,649 --> 00:03:26,540 of the surrounding points 70 00:03:26,540 --> 00:03:30,520 for some of their forest sampling and inventory activities. 71 00:03:30,520 --> 00:03:34,280 On the plus side, there's a real strong possibility 72 00:03:34,280 --> 00:03:39,280 to reduce your travel costs when sampling in this fashion. 73 00:03:39,410 --> 00:03:42,590 And then lastly, there's an adaptive sampling strategy 74 00:03:42,590 --> 00:03:44,635 where you collect more frequent samples 75 00:03:44,635 --> 00:03:49,635 in areas with variable information and less frequent samples 76 00:03:49,990 --> 00:03:52,770 in uniform areas of the landscape. 77 00:03:52,770 --> 00:03:55,810 This does require some assessment prior to field work 78 00:03:55,810 --> 00:03:58,040 and updates during field work 79 00:03:58,040 --> 00:04:00,863 which can present some logistical challenges. 80 00:04:04,830 --> 00:04:07,420 A few key bits of terminology I'd like to cover, 81 00:04:07,420 --> 00:04:11,760 interpolation as you may or may not know 82 00:04:11,760 --> 00:04:14,250 uses measured values to predict the values 83 00:04:14,250 --> 00:04:16,541 at unsampled locations. 84 00:04:16,541 --> 00:04:20,400 We'll look at two different kinds of interpolation, 85 00:04:20,400 --> 00:04:23,460 deterministic and geostatistical. 86 00:04:23,460 --> 00:04:27,640 So deterministic interpolation uses either a similarity 87 00:04:27,640 --> 00:04:29,930 or smoothing type function 88 00:04:29,930 --> 00:04:33,870 and it does not measure uncertainty or error 89 00:04:33,870 --> 00:04:35,857 in the output prediction. 90 00:04:35,857 --> 00:04:38,910 Geostatistical interpretation on the other hand, 91 00:04:38,910 --> 00:04:40,360 is a statistical model 92 00:04:40,360 --> 00:04:42,253 that include spatial autocorrelation. 93 00:04:43,170 --> 00:04:47,210 These can measure uncertainty in the output data, 94 00:04:47,210 --> 00:04:49,220 and are quite powerful 95 00:04:49,220 --> 00:04:52,100 for any number of different applications. 96 00:04:52,100 --> 00:04:54,160 We heard a little bit about the kernel already, 97 00:04:54,160 --> 00:04:55,820 but just by way of reminder, 98 00:04:55,820 --> 00:04:57,670 the kernel is a weighted function 99 00:04:57,670 --> 00:05:02,040 that assigns the weights to locations based on the proximity 100 00:05:02,040 --> 00:05:05,613 of that location to an observation or a sample point. 101 00:05:08,090 --> 00:05:12,510 Now, spatial autocorrelation is this assumption that 102 00:05:12,510 --> 00:05:15,470 the closer two observations are to one another, 103 00:05:15,470 --> 00:05:18,293 the more similar their values are likely to be. 104 00:05:19,200 --> 00:05:22,340 So depending on the type of analytical approach 105 00:05:22,340 --> 00:05:24,780 that we take to interpolation, 106 00:05:24,780 --> 00:05:27,780 spatial autocorrelation may be required, 107 00:05:27,780 --> 00:05:31,730 it might be assumed, or it might be ignored altogether. 108 00:05:31,730 --> 00:05:35,450 It's important to note that spatial autocorrelation 109 00:05:35,450 --> 00:05:38,110 violates this independence assumption 110 00:05:38,110 --> 00:05:39,520 that you would have otherwise 111 00:05:39,520 --> 00:05:41,463 in traditional statistical methods. 112 00:05:42,420 --> 00:05:45,830 For geostatistics, it's an important assumption 113 00:05:46,700 --> 00:05:48,650 that makes these calculations possible. 114 00:05:50,830 --> 00:05:52,900 We can look at a couple of different ways 115 00:05:52,900 --> 00:05:54,903 for analyzing data patterns, 116 00:05:56,280 --> 00:05:59,840 Moran's I and the Getis-Ord General G. 117 00:05:59,840 --> 00:06:01,841 Both of these are statistical measures 118 00:06:01,841 --> 00:06:04,900 that look at clusters of data. 119 00:06:04,900 --> 00:06:06,770 In the case of Moran's I, 120 00:06:06,770 --> 00:06:08,900 it's a measurement of spatial autocorrelation 121 00:06:08,900 --> 00:06:11,820 based on feature locations and their values 122 00:06:13,150 --> 00:06:15,900 and the algorithm determines if the data are clustered, 123 00:06:15,900 --> 00:06:17,543 dispersed or random. 124 00:06:18,910 --> 00:06:22,050 Like all statistical tests, this one has a null hypothesis 125 00:06:22,050 --> 00:06:24,730 that states that the attribute under consideration 126 00:06:24,730 --> 00:06:26,233 is randomly distributed. 127 00:06:27,400 --> 00:06:30,510 You can use Moran's I to help define both the size 128 00:06:30,510 --> 00:06:33,100 and the orientation of search neighborhoods 129 00:06:33,100 --> 00:06:36,073 that would be used for interpolation processes. 130 00:06:37,120 --> 00:06:39,990 Now, the Getis-Ord General G statistic 131 00:06:39,990 --> 00:06:43,180 measures clusters of high or low values. 132 00:06:43,180 --> 00:06:45,980 And in this case, it assumes a null hypothesis 133 00:06:45,980 --> 00:06:49,970 that there is no spatial clustering of the feature values. 134 00:06:49,970 --> 00:06:53,159 For both, depending on the P value and the Z value 135 00:06:53,159 --> 00:06:57,120 that you get from your estimation of these statistics, 136 00:06:57,120 --> 00:06:58,720 you can determine whether or not 137 00:06:59,570 --> 00:07:02,110 you can dismiss the null hypothesis 138 00:07:02,110 --> 00:07:04,870 and state something definitive about the pattern 139 00:07:04,870 --> 00:07:07,573 or lack thereof in the data that you're using. 140 00:07:09,936 --> 00:07:12,340 Now there's multiple mathematical approaches 141 00:07:12,340 --> 00:07:14,040 to spatial interpolation, 142 00:07:14,040 --> 00:07:16,900 but all of these use points with known values 143 00:07:16,900 --> 00:07:20,293 to estimate values at these unsamped locations. 144 00:07:21,540 --> 00:07:24,414 Location, both the distance between the sample location 145 00:07:24,414 --> 00:07:27,334 and the cell location and the attribute values 146 00:07:27,334 --> 00:07:31,040 are considered as part of the process. 147 00:07:31,040 --> 00:07:34,110 Now, once again, depending on the approach that you take 148 00:07:34,110 --> 00:07:35,346 varying weights would be assigned 149 00:07:35,346 --> 00:07:39,090 to neighboring observations and you do have some control 150 00:07:39,090 --> 00:07:41,040 over the way those weights are applied. 151 00:07:42,040 --> 00:07:43,050 Not surprisingly 152 00:07:43,050 --> 00:07:45,330 because of the multiple different approaches 153 00:07:45,330 --> 00:07:47,837 to performing spatial interpolation, 154 00:07:47,837 --> 00:07:50,743 the different approaches will produce different results. 155 00:07:52,310 --> 00:07:54,990 Deterministic and geostatistical approaches 156 00:07:55,830 --> 00:07:59,150 for interrelating continuous raster surfaces 157 00:07:59,150 --> 00:08:00,963 happen with control points. 158 00:08:03,540 --> 00:08:07,560 Now, even with a finite number of discrete objects in space, 159 00:08:07,560 --> 00:08:10,510 there are still likely too many to count, 160 00:08:10,510 --> 00:08:12,600 too many places to go survey 161 00:08:12,600 --> 00:08:16,280 to consider a completeness to that. 162 00:08:16,280 --> 00:08:20,324 And because of that, coupled with potentially costs, 163 00:08:20,324 --> 00:08:24,000 a lack of access for legal or practical reasons, 164 00:08:24,000 --> 00:08:26,470 site locations being dangerous, 165 00:08:26,470 --> 00:08:27,468 you might be missing samples, 166 00:08:27,468 --> 00:08:29,880 you might have erroneous samples, 167 00:08:29,880 --> 00:08:32,090 or you might be re sampling your data. 168 00:08:32,090 --> 00:08:36,170 These are all reasons why you might need to perform 169 00:08:36,170 --> 00:08:38,483 spatial interpolation operations. 170 00:08:39,410 --> 00:08:43,400 We have both global and local interpolators. 171 00:08:43,400 --> 00:08:44,320 On the global front, 172 00:08:44,320 --> 00:08:46,410 you're looking at trend surface analysis 173 00:08:46,410 --> 00:08:48,240 and regression models. 174 00:08:48,240 --> 00:08:51,860 We won't be doing much of that in this module. 175 00:08:51,860 --> 00:08:54,500 Instead, we'll focus on local operations 176 00:08:54,500 --> 00:08:57,230 that use a sample of the known points 177 00:08:57,230 --> 00:08:59,770 where the user has control over how many 178 00:08:59,770 --> 00:09:02,610 and how to search for those point locations. 179 00:09:02,610 --> 00:09:04,650 And we'll look at both geo statistical 180 00:09:04,650 --> 00:09:06,453 and deterministic approaches. 181 00:09:08,630 --> 00:09:11,040 Now control points as I mentioned before, 182 00:09:11,040 --> 00:09:13,900 are these locations with known values. 183 00:09:13,900 --> 00:09:16,890 There's a number of observations or values 184 00:09:16,890 --> 00:09:20,440 or sample locations and their distribution 185 00:09:20,440 --> 00:09:23,993 influence the accuracy of the estimated surface. 186 00:09:26,690 --> 00:09:29,024 In general, you wanna consider enough control points 187 00:09:29,024 --> 00:09:33,050 to yield adequate results but no more. 188 00:09:33,050 --> 00:09:35,920 Same fundamental principle of parsimony 189 00:09:35,920 --> 00:09:39,070 when constructing other statistical models. 190 00:09:39,070 --> 00:09:40,826 And it's up to you to analyze the properties 191 00:09:40,826 --> 00:09:44,790 of that surface to understand if or how 192 00:09:46,080 --> 00:09:50,610 the influence of the weight changes over space or distance. 193 00:09:50,610 --> 00:09:53,340 So again, looking for these patterns in the data 194 00:09:53,340 --> 00:09:56,693 that you're using to perform these interpolation functions. 195 00:09:58,230 --> 00:10:01,710 Now there's two types of interpolators, 196 00:10:01,710 --> 00:10:04,773 what's known as an exact or inexact predictor. 197 00:10:05,790 --> 00:10:07,780 For the exact, the predicted value 198 00:10:07,780 --> 00:10:09,800 is equal to the measured value. 199 00:10:09,800 --> 00:10:12,750 That means for each one of your sample locations, 200 00:10:12,750 --> 00:10:16,760 the curve that's fit through your data will perfectly align 201 00:10:16,760 --> 00:10:19,223 with the value at those sample points. 202 00:10:20,130 --> 00:10:21,690 In the case of the inexact, 203 00:10:21,690 --> 00:10:23,460 the predicted values might be different 204 00:10:23,460 --> 00:10:25,160 than the measured value. 205 00:10:25,160 --> 00:10:27,164 Oftentimes that means they will avoid 206 00:10:27,164 --> 00:10:30,890 or not be good at locating peaks and troughs 207 00:10:30,890 --> 00:10:31,923 in the landscape. 208 00:10:35,460 --> 00:10:39,150 So why you search neighborhoods? 209 00:10:39,150 --> 00:10:42,810 Again, it's this concept that the further away you are 210 00:10:42,810 --> 00:10:44,830 from a known point location, 211 00:10:44,830 --> 00:10:47,840 the less likely there is to be a relationship 212 00:10:47,840 --> 00:10:50,770 to the value in that known location. 213 00:10:50,770 --> 00:10:52,710 So distant points have a limited effect 214 00:10:52,710 --> 00:10:55,563 on the predictive value at some point, 215 00:10:56,410 --> 00:10:59,060 and they might even have a detrimental effect 216 00:10:59,060 --> 00:11:00,820 on the predicted value itself 217 00:11:00,820 --> 00:11:02,950 because of the disparity in the values 218 00:11:02,950 --> 00:11:04,563 from those two locations. 219 00:11:05,400 --> 00:11:07,530 Lastly, by applying search neighborhoods, 220 00:11:07,530 --> 00:11:09,690 you're reducing the number of points 221 00:11:09,690 --> 00:11:11,760 considered for each calculation 222 00:11:11,760 --> 00:11:14,103 which increases the computational speed. 223 00:11:16,420 --> 00:11:17,940 Now, in terms of neighborhoods, 224 00:11:17,940 --> 00:11:20,810 there's many different kinds that we can specify. 225 00:11:20,810 --> 00:11:24,480 You might simply say that you want the closest 10 points 226 00:11:24,480 --> 00:11:28,200 to the unknown value. 227 00:11:28,200 --> 00:11:31,810 You might want a specified number of points 228 00:11:31,810 --> 00:11:35,630 within some distance and you could specify both, 229 00:11:35,630 --> 00:11:37,080 the shape of that distance 230 00:11:37,080 --> 00:11:40,650 and the number of points you'd like to locate within there. 231 00:11:40,650 --> 00:11:44,010 And then lastly, you could look at something by quadrant 232 00:11:44,010 --> 00:11:46,870 and specify that you want at least a minimum number 233 00:11:46,870 --> 00:11:49,420 of points from each of the quadrants 234 00:11:49,420 --> 00:11:51,220 and you can rotate those quadrants 235 00:11:51,220 --> 00:11:52,435 and even combine the quadrants 236 00:11:52,435 --> 00:11:55,363 with a distance function as well. 237 00:11:58,150 --> 00:12:00,840 So if we look at one example here 238 00:12:00,840 --> 00:12:03,763 where I define a 200 meter radius, 239 00:12:04,880 --> 00:12:06,970 my neighborhood's going to be circular shape. 240 00:12:06,970 --> 00:12:10,600 It's going to have four sectors with a zero degree offset, 241 00:12:10,600 --> 00:12:13,110 and I want to identify at least two neighbors 242 00:12:13,110 --> 00:12:14,853 in each one of my sectors. 243 00:12:16,290 --> 00:12:21,140 We start by identifying that location of interest, 244 00:12:21,140 --> 00:12:24,433 plotting out our radius and the sectors, 245 00:12:25,350 --> 00:12:28,900 selecting those points within each one of the sectors 246 00:12:28,900 --> 00:12:31,630 and then using that information to make an estimate 247 00:12:31,630 --> 00:12:33,903 at that unknown value location. 248 00:12:34,780 --> 00:12:36,970 Now you can change the size and the shape 249 00:12:36,970 --> 00:12:38,270 of the search neighborhood 250 00:12:39,220 --> 00:12:42,010 if you're aware of directional trends 251 00:12:42,010 --> 00:12:44,190 in the location that you're analyzing. 252 00:12:44,190 --> 00:12:47,543 And we saw an example of that in the last lecture as well. 253 00:12:50,980 --> 00:12:52,880 In terms of directional influence, 254 00:12:52,880 --> 00:12:55,340 global trends affect all of the measurements 255 00:12:55,340 --> 00:12:56,900 in a deterministic manner 256 00:12:56,900 --> 00:13:00,670 and you can represent that using a mathematical formula. 257 00:13:00,670 --> 00:13:03,800 It's up to you to understand the directional signal 258 00:13:03,800 --> 00:13:07,360 whether one exists and what direction it's pointing. 259 00:13:07,360 --> 00:13:08,450 Some examples of this 260 00:13:08,450 --> 00:13:10,670 include something like a prevailing wind 261 00:13:10,670 --> 00:13:12,023 or groundwater flow. 262 00:13:13,360 --> 00:13:17,890 Another concept related to directional influences anistropy. 263 00:13:17,890 --> 00:13:20,200 So this is not deterministic. 264 00:13:20,200 --> 00:13:23,440 This is reflected more by a random process 265 00:13:23,440 --> 00:13:26,930 with higher auto correlation in one direction. 266 00:13:26,930 --> 00:13:30,160 So there is a potential for a deterministic work around 267 00:13:30,160 --> 00:13:34,140 to this somewhat funny issue by modeling the influence 268 00:13:35,800 --> 00:13:38,854 of that anistropy by changing the size 269 00:13:38,854 --> 00:13:41,190 and the shape of the search neighborhood 270 00:13:41,190 --> 00:13:45,853 to see if you can't account for that in those variations. 271 00:13:48,330 --> 00:13:50,690 In terms of output surfaces, 272 00:13:50,690 --> 00:13:52,360 there are a number of different surfaces 273 00:13:52,360 --> 00:13:55,173 that are produced by these interpolation methods. 274 00:13:56,170 --> 00:13:58,288 The prediction surface is the default output 275 00:13:58,288 --> 00:14:01,063 of all of the different interpolation methods. 276 00:14:02,460 --> 00:14:05,083 Predicted standard error is the standard deviation 277 00:14:05,083 --> 00:14:08,280 of the estimated value at each location. 278 00:14:08,280 --> 00:14:10,610 As the standard error increases, 279 00:14:10,610 --> 00:14:14,603 the precision of prediction at the location decreases. 280 00:14:15,630 --> 00:14:16,816 There's a probability surface, 281 00:14:16,816 --> 00:14:20,660 the probability that some threshold value will be exceeded 282 00:14:20,660 --> 00:14:21,993 at each location. 283 00:14:23,000 --> 00:14:26,840 And if you have a multinomial normal distribution 284 00:14:26,840 --> 00:14:29,600 for your data, you can assume that 285 00:14:29,600 --> 00:14:33,980 more than 99% of the values are reflected 286 00:14:33,980 --> 00:14:36,020 within three standard deviations 287 00:14:37,000 --> 00:14:37,833 of the data. 288 00:14:38,670 --> 00:14:41,790 You can also produce quantile output surfaces 289 00:14:41,790 --> 00:14:44,120 and these are really useful for monitoring 290 00:14:44,120 --> 00:14:46,833 or modeling best and worst case scenarios. 291 00:14:47,700 --> 00:14:50,310 Depending on the interpolation method 292 00:14:50,310 --> 00:14:53,710 that will determine the possible output surfaces. 293 00:14:53,710 --> 00:14:56,340 Again, the deterministic approaches 294 00:14:56,340 --> 00:14:59,340 will not produce any of the error probability 295 00:15:01,080 --> 00:15:03,083 surfaces as part of their output. 296 00:15:05,220 --> 00:15:08,290 And then lastly, we'll touch on accuracy. 297 00:15:08,290 --> 00:15:09,670 Now this is the measure difference 298 00:15:09,670 --> 00:15:12,723 between the interpolated value and the measured value. 299 00:15:13,770 --> 00:15:16,150 If you want to perform an accuracy assessment, 300 00:15:16,150 --> 00:15:18,580 you withhold some validation points 301 00:15:18,580 --> 00:15:21,120 not to be used in the estimation process, 302 00:15:21,120 --> 00:15:25,430 and then check against that once you've estimated a surface. 303 00:15:25,430 --> 00:15:28,420 You can do this for a subset of the data 304 00:15:28,420 --> 00:15:30,720 or for all of the data points in your dataset. 305 00:15:31,720 --> 00:15:35,550 Alternatively, it's somewhat more difficult and expensive. 306 00:15:35,550 --> 00:15:38,080 You could collect a validation dataset. 307 00:15:38,080 --> 00:15:40,160 So an entire dataset that's set aside 308 00:15:40,160 --> 00:15:44,010 from the estimation dataset, that's then used to compare 309 00:15:44,010 --> 00:15:47,793 against the output from your model estimation. 310 00:15:48,920 --> 00:15:50,180 The last thing I'll note here 311 00:15:50,180 --> 00:15:53,070 before we get into some examples of interpolation, 312 00:15:53,070 --> 00:15:55,103 is that no single method is the best solution 313 00:15:55,103 --> 00:15:57,300 for all instances. 314 00:15:57,300 --> 00:15:59,300 This will take some trial and error. 315 00:15:59,300 --> 00:16:02,770 You need to dig into your data and understand the contents 316 00:16:02,770 --> 00:16:06,460 particularly patterns that might exist in the data 317 00:16:06,460 --> 00:16:09,810 to be able to better parameterize your model 318 00:16:11,200 --> 00:16:13,763 and produce more realistic output surfaces. 319 00:16:14,750 --> 00:16:17,580 Well, that's it for key concepts and terminology. 320 00:16:17,580 --> 00:16:19,800 Let's take a look at some examples 321 00:16:19,800 --> 00:16:21,310 in both a deterministic 322 00:16:21,310 --> 00:16:24,083 and geostatistical interpolation approaches.