Palisades Fire Before and After

A Sample of Available Satellite Data for Post-Disaster Visual Assessment

Inspired by this post on LinkedIn from Brianna R. Pagán, Ph.D., in which she describes the lack of available real-time satellite data during the 2025 Palisades Fire, I wanted to see what types of satellite data were available now, several days into the disaster. By January 16th, 2025, nine days after the start of the fire, data from several providers had been made available, including Sentinel-2, Umbra, and Maxar.

Click on the visualization to interact with the map.

The map above shows two satellite imagery captures, January 2nd (left) and January 12th, 2025, from the Sentinel-2 satellite constellation. Comparing these two images, we can clearly see the Palisades fire rapid expansion across the Santa Monica Mountains and into the Pacific Palisades neighborhood (map - bottom-right).

Sentinel-2 passes over the earth every five days, so it’s not a rapid source of information but can be very useful, especially depending on the timing of the image capture and the relevant events. However, some of these contain extensive cloud cover, such as the image below. The images chosen for the map contain less than 20% cloud cover. When searching for images, you can filter by cloud cover percentage to find the best images for your analysis.

Sentinel-2 Cloud Cover

Sentinel-2 Capture from January 7, 2025. Smoke can be clearly seen near the Pacific Palisades neighborhood. Image Metadata

What’s missing in this visualization is the ability to quantify the impact across the urban infrastructure. While a general fire boundary could be generated - and the defined lines of the visible burn scar match up nicely with the published fire boundary - damage to individual buildings is hard to determine as the resolution of the Sentinel-2 data is only 10 meters per pixel. To fill this gap, higher-resolution satellite imagery, SAR data, aerial photography, or on-the-ground field work would be necessary. In this case, both Umbra and Maxar have included the disaster in their open data programs, offering samples of their data covering portions of the wildfire area for free. While this data is free to access, data licensing on derivatives could be problematic depending on the types of analysis performed and/or desired outputs.

Captures from Umbra (SAR, left) and Maxar covering portions of the areas affected by the Palisades Fire. Umbra Metadata

Using this higher-resolution data, we can easily make out damage to individual structures. Visual assessment of he Maxar data is straightforward, with damage to homes clearly visible, but the Umbra data requires a bit more interpretation. The SAR data is represented as a grayscale image where pixel intensity corresponds to the strength of the radar signal (backscatter) returned to the satellite. In the image above, the static or fuzziness in the image represents increased backscatter, which could represent damage to the structures, while intact buildings are represented by a more seamless return. However, this is not a direct correlation, and further analysis would be needed to determine the extent of the damage.

Additionally, with the SAR image, you can see that the projection is slightly off, potentially due to the slant effect of the SAR method. This is my first time working with SAR data, so it’s not clear if the shift can be corrected with information from the metadata or if this is a limitation of the data itself. More information on Umbra deliverables can be found in the Canopy Documentation.

Finall, while you can task these commercial satellites multiple times per day, and in the case of Umbra they can give usable results even through cloud-cover, to cover an area as large as the Palisades Fire the cost could would be significant - in the thousands to hundreds of thousands of dollars depending on the area and resolution.

How I Built It

To create the first map above, I used data from the Sentinel-2 satellite constellation pulled from Element 84’s Earth Search Console. You can find the same imagery via SkyFi, which includes additional providers such as Umbra. If you are brave you can search the raw Spatio Temporal Asset catalogs (STAC) for each of these providers, but the process can be tedious. Once you find an image from the Earth Console, you need to view the image metadata to get the download link for the visual GeoTIFF asset.

"visual": {
"href": "https://sentinel-cogs.s3.us-west-2.amazonaws.com/sentinel-s2-l2a-cogs/11/S/LT/2025/1/S2A_11SLT_20250112_0_L2A/TCI.tif",
"type": "image/tiff; application=geotiff; profile=cloud-optimized",
"title": "True color image"
}
Excerpt from the Sentinel-2 image metadata

Thanks to Geomatica and geotiff.js, we can add a cloud-optimized GeoTIFF directly into a MapLibre map. However, before we can use the Sentinel-2 imagery, we first need to convert it into a compatible compression format. The default compression used in the Sentinel-2 visual GeoTIFF imagery is DEFLATE, which is not supported by the browser. To convert the GeoTIFF into a compatible format, we can use gdal_translate from GDAL. For this visualization, I used a version of the script from my previous blog post on how to create cloud-optimized GeoTIFFs. After the conversion, I uploaded the GeoTIFFs to Cloudflare since the plugin allows for reading directly from object storage.

Once the images are converted, we can simply load them into the map and enable the swipe effect using the MapLibre Swipe Plugin.


Resources

Older Post
Deck GL and 3D Tiles

Related Posts