--- title: "How to download climate data using ColOpenData" output: rmarkdown::html_vignette: df_print: tibble vignette: > %\VignetteIndexEntry{How to download climate data using ColOpenData} %\VignetteEngine{knitr::rmarkdown} %\VignetteEncoding{UTF-8} --- ```{r, include = FALSE} knitr::opts_chunk$set( collapse = TRUE, comment = "#>" ) ``` ::: {style="text-align: justify;"} **ColOpenData** can be used to access open climate data from Colombia. This climate data is retrieved from the Institute of Hydrology, Meteorology and Environmental Studies (IDEAM). The climate module allows the user to consult climate data for any Region of Interest (ROI) inside the country and retrieve the information for each station contained inside. The available information from IDEAM can be accessed using specific internal tags as follows: ::: ```{r IDEAM table, echo = FALSE} tags <- c( "TSSM_CON", "THSM_CON", "TMN_CON", "TMX_CON", "TSTG_CON", "HR_CAL", "HRHG_CON", "TV_CAL", "TPR_CAL", "PTPM_CON", "PTPG_CON", "EVTE_CON", "FA_CON", "NB_CON", "RCAM_CON", "BSHG_CON", "VVAG_CON", "DVAG_CON", "VVMXAG_CON", "DVMXAG_CON" ) variable <- c( "Dry-bulb Temperature", "Wet-bulb Temperature", "Minimum Temperature", "Maximum Temperature", "Dry-bulb Temperature (Termograph)", "Relative Humidity", "Relative Humidity (Hydrograph)", "Vapour Pressure", "Dew Point", "Precipitation (Daily)", "Precipitation (Hourly)", "Evaporation", "Atmospheric Phenomenon", "Cloudiness", "Wind Trajectory", "Sunshine Duration", "Wind Speed", "Wind Direction", "Maximum Wind Speed", "Maximum Wind Direction" ) IDEAM_tags <- data.frame( Tags = tags, Variable = variable, stringsAsFactors = FALSE ) knitr::kable(IDEAM_tags) ``` ::: {style="text-align: justify;"} Each observation is subject to the availability of stations in the ROI and the stations' status (active, maintenance or suspended), as well as quality filters implemented by IDEAM. In this vignette you will learn: 1. How to download climate data using **ColOpenData**. 2. How to aggregate climate data by different frequencies 3. How to plot downloaded climate data For this example we will retrieve data for the municipality of Espinal in Colombia. We will download Dry-Bulb Temperature (TSSM_CON) from 2013 to 2016, to observe the increase in the average temperature during 2015 and 2016 due to the impact of El Nino (ENSO). **ColOpenData** offers three methods to do this, using different functions: - `download_climate_stations()` to download climate data from previously selected stations - `download_climate_geom()` to download climate data from a specified geometry (ROI) - `download_climate()` to download climate data from municipalities' or departments' already loaded geometries In this example, we will follow the three methods to get the same results, exploring the included functions. We will start by loading the needed libraries. ::: ```{r library imports, results = "hide", warning = FALSE, message = FALSE} library(ColOpenData) library(dplyr) library(sf) library(leaflet) library(ggplot2) ``` ::: {style="text-align: justify;"} **Disclaimer: all data is loaded to the environment in the user's R session, but is not downloaded to user's computer.** ::: ## Retrieving climate data for a ROI using stations' data ::: {style="text-align: justify;"} For this example, we will need to create a spatial polygon around the municipality of Espinal and use that as our ROI to retrieve the climate data. To create the spatial polygon we need to introduce the coordinates of the geometry. For simplicity, we will build a bounding box by introducing the 4 points which bound the municipality, and transform the created geometry into an `sf` object (see [**sf**](https://r-spatial.github.io/sf/articles/sf1.html) library for further details). ::: ```{r polygon creation} lat <- c(4.263744, 4.263744, 4.078156, 4.078156, 4.263744) lon <- c(-75.042067, -74.777022, -74.777022, -75.042067, -75.042067) polygon <- st_polygon(x = list(cbind(lon, lat))) %>% st_sfc() roi <- st_as_sf(polygon) ``` With our created ROI, we can make a simple visualization using `leaflet`. ```{r polygon plot} leaflet(roi) %>% addProviderTiles("OpenStreetMap") %>% addPolygons( stroke = TRUE, weight = 2, color = "#2e6930", fillColor = "#2e6930", opacity = 0.6 ) ``` ::: {style="text-align: justify;"} We can make a first exploration to check if there are any stations contained inside of it, using the function `stations_in_roi()`. ::: ```{r stations in roi} stations <- stations_in_roi(geometry = roi) head(stations) ``` ::: {style="text-align: justify;"} We can see that in the region there are 24 stations. Different categories are recorded by different stations, and can be checked at the column [categoria]{.underline}. Stations under the categories *Climática Principal* and *Climática Ordinaria* have records of temperature. Some stations are suspended, which means they are not taking measurements at the moment. This information is found at the column [estado]{.underline} where, if suspended, the observation would be *Suspendida* Also, at the column [fecha_suspension]{.underline} the observation would be different from `NA`, since suspended stations would have an associated suspension date. However, even if a station is suspended, the historical data (up to the suspension date) can be accessed. To filter the stations that recorded information during the desired period, we can delete the stations with suspension dates before 2013. ::: ```{r stations filtered} cw_stations <- stations %>% filter( as.Date(fecha_suspension) > as.Date("2013-01-01") | estado == "Activa", categoria %in% c("Climática Principal", "Climática Ordinaria") ) head(cw_stations) ``` ::: {style="text-align: justify;"} From the original 24 stations, only 1 was working for some or the whole period of interest and collected information for Dry-Bulb Temperature (TSSM_CON). It is important to consider that after data collection, some information might be lost due to quality attributes. With the stations, we can access TMX_CON from 2013 to 2016. To do so, we can use the function `download_climate_stations()`. This function has the following parameters: - `stations`: `data.frame` containing the stations' codes. This `data.frame` must be retrieved from the function `stations_in_roi`. - `start_date`: character with the first date to consult in the format `"YYYY-MM-DD"`. (First available date is `"1920-01-01"`). - `end_date`: character with the last date to consult in the format `"YYYY-MM_DD"`. (Last available date is `"2023-05-31"`). - `tag`: character containing climate tag to consult. ::: ```{r download climate stations} tssm_stations <- download_climate_stations( stations = cw_stations, start_date = "2013-01-01", end_date = "2016-12-31", tag = "TSSM_CON" ) head(tssm_stations) ``` ::: {style="text-align: justify;"} The returned tidy `data.frame` includes: individual and unique station code, longitude, latitude, date, hour, tag requested and value recorded at the specified time. The tidy structure reports a row for each observation, which makes the subset and plot easier. To plot a time series of the stations' data we can use `ggplot()` function from `ggplot2` package as follows: ::: ```{r plot temperatures stations} ggplot(data = tssm_stations) + geom_line(aes(x = date, y = value, group = station), color = "#106ba0") + ggtitle("Dry-bulb Temperature in Espinal by station") + xlab("Date") + ylab("Temperature [°C]") + facet_grid(rows = vars(station)) + theme_minimal() + theme( plot.background = element_rect(fill = "white", colour = "white"), panel.background = element_rect(fill = "white", colour = "white"), plot.title = element_text(hjust = 0.5) ) ``` ::: {style="text-align: justify;"} As we can see, only one station has data for the selected period. However, by having the data measured by hours, we cannot easily observe changes in the temperature patterns along time. To aid this issue, we will use the aggregation function `aggregate_climate()`, which aggregates climate data by time. This function takes by parameter the desired aggregation. ::: ```{r plot monthly} tssm_month <- tssm_stations %>% aggregate_climate(frequency = "month") ggplot(data = tssm_month) + geom_line(aes(x = date, y = value, group = station), color = "#106ba0") + ggtitle("Dry-bulb Temperature in Espinal by station") + xlab("Date") + ylab("Dry-bulb temperature [C]") + facet_grid(rows = vars(station)) + theme_minimal() + theme( plot.background = element_rect(fill = "white", colour = "white"), panel.background = element_rect(fill = "white", colour = "white"), plot.title = element_text(hjust = 0.5) ) ``` ## Other methods ::: {style="text-align: justify;"} To retrieve climate data for any ROI in the country, without manually extracting the stations' data, we can use the function `download_climate_geom()`. The function has the following parameters: - `geometry`: `sf` geometry containing the geometry for a given ROI. This geometry can be either a `POLYGON` or `MULTIPOLYGON`. - `start_date`: character with the first date to consult in the format `"YYYY-MM-DD"`. (First available date is `"1920-01-01"`). - `end_date`: character with the last date to consult in the format `"YYYY-MM_DD"`. (Last available date is `"2023-05-31"`). - `tag`: character containing climate tag to consult. To replicate the previous example, we can just use the previously created ROI and add the aggregation for month. We can add the aggregation function to the workflow using the pipe operator `%>%`. The following code should retrieve the same results as the previous one. ::: ```{r download climate data, eval = FALSE} tssm_roi <- download_climate_geom( geometry = roi, start_date = "2013-01-01", end_date = "2016-12-31", tag = "TSSM_CON" ) %>% aggregate_climate(frequency = "month") ``` To make the download process even easier, and avoid the creation of already known geometries like municipalities or departments, **ColOpenData** offers an extra function to download data using the areas' DIVIPOLA code. DIVIPOLA codification is standardized for the whole country, and contains departments' and municipalities' codes. For further details on DIVIPOLA codification and functions please refer to [Documentation and Dictionaries](https://epiverse-trace.github.io/ColOpenData/articles/documentation_and_dictionaries.html#DIVIPOLA). We will filter for the city of Espinal in the department Tolima. ::: ```{r municipality code} espinal_code <- name_to_code_mun("Tolima", "Espinal") espinal_code ``` ::: {style="text-align: justify;"} The function `download_climate()` will require almost the same arguments as `download_climate_geom()`, but instead of an `sf` object, it will take a character containing the DIVIPOLA code: - `code`: character with the DIVIPOLA code for the area. - `start_date`: character with the first date to consult in the format `"YYYY-MM-DD"`. (First available date is `"1920-01-01"`). - `end_date`: character with the last date to consult in the format `"YYYY-MM_DD"`. (Last available date is `"2023-05-31"`). - `tag`: character containing climate tag to consult. The code below can be used to get the same results as the previous two examples, without the need to create a whole geometry or filtering individual stations. ::: ```{r download climate mpio, eval = FALSE} tssm_mpio <- download_climate( code = espinal_code, start_date = "2013-01-01", end_date = "2016-12-31", tag = "TMX_CON" ) %>% aggregate_climate(frequency = "month") ``` ## Disclaimer ::: {style="text-align: justify;"} - Data availability is subdued to station's measurements and quality filters. In most cases, this leads to a lower amount of data, considering the extensive amount of climate stations. - Temporal aggregation is only available for some tags and is limited to the ones who have a specific methodology of aggregation reported by IDEAM. The daily, monthly and annual aggregation is available for: - `TSSM_CON`: Dry-bulb temperature - `TMX_CON`: Maximum temperature - `TMN_CON`: Minimum temperature - `PTPM_CON`: Precipitation - `BSHG_CON`: Sunshine duration - Temporal and spatial interpolation are not included in this version of **ColOpenData**. :::