Météo-France Open Data
[This article was first published on r.iresmi.net, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.
Great news! Météo-France has started to widen its open archive data. No API so far and a lot of files… What can we do?
library(tidyverse)
library(httr)
library(glue)
library(janitor)
library(jsonlite)
library(sf)
round_any <- function(x, accuracy, f = round) {
f(x / accuracy) * accuracy
}
We can start by selecting a few départements, here the 12 from Auvergne-Rhône-Alpes.
# which départements to download
sel_dep <- c("01", "03", "07", "15", "26", "38", "42", "43", "63", "69", "73", "74")
# for the map (code INSEE région)
sel_reg <- "84"
Scraping
Get some information about the available files for the daily data (Données climatologiques de base - quotidiennes).
# count files available
n_files <- GET("https://www.data.gouv.fr/api/2/datasets/6569b51ae64326786e4e8e1a/") |>
content() |>
pluck("resources", "total")
# get files informations
files_available <- GET(glue("https://www.data.gouv.fr/api/2/datasets/6569b51ae64326786e4e8e1a/resources/?page=1&page_size={n_files}&type=main")) |>
content(as = "text", encoding = "UTF-8") |>
fromJSON(flatten = TRUE) |>
pluck("data") |>
as_tibble(.name_repair = make_clean_names)
Download and open the files
if (!length(list.files("data"))) {
files_available |>
mutate(dep = str_extract(title, "(?<=departement_)[:alnum:]{2,3}(?=_)")) |>
filter(dep %in% sel_dep,
str_detect(title, "RR-T-Vent")) |>
pwalk(\(url, title, format, ...) {
GET(url,
write_disk(glue("data/{title}.{format}"),
overwrite = TRUE))
})
}
We get 36 files…Each département has 3 files representing the recent years, 1950-2021 and before 1950.
# parsing problems with readr::read_delim
# we use read.delim instead
meteo <- list.files("data", full.names = TRUE) |>
map(read.delim,
sep = ";",
colClasses = c("NUM_POSTE" = "character",
"AAAAMMJJ" = "character")) |>
list_rbind() |>
as_tibble(.name_repair = make_clean_names) |>
mutate(num_poste = str_trim(num_poste),
aaaammjj = ymd(aaaammjj))
# Map data
# See https://r.iresmi.net/posts/2021/simplifying_polygons_layers/
dep_sig <- read_sf("~/data/adminexpress/adminexpress_cog_simpl_000_2022.gpkg",
layer = "departement") |>
filter(insee_reg == sel_reg) |>
st_transform("EPSG:2154")
That’s a nice dataset of 19,222,848 rows!
Exploring
meteo |>
summarise(.by= c(nom_usuel, num_poste, lat, lon),
annee_max = max(year(aaaammjj))) |>
st_as_sf(coords = c("lon", "lat"), crs = "EPSG:4326") |>
st_transform("EPSG:2154") |>
ggplot() +
geom_sf(data = dep_sig) +
geom_sf(aes(color = annee_max == year(Sys.Date()),
shape = annee_max == year(Sys.Date())),
size = 2, alpha = 0.5) +
scale_color_manual(name = "statut",
values = c("FALSE" = "chocolate4",
"TRUE" = "chartreuse3"),
labels = c("FALSE" = "inactive",
"TRUE" = "active")) +
scale_shape_manual(name = "statut",
values = c("FALSE" = 16,
"TRUE" = 17),
labels = c("FALSE" = "inactive",
"TRUE" = "active")) +
labs(title = "Stations météo",
subtitle = "Auvergne-Rhône-Alpes",
caption = glue("données Météo-France (https://meteo.data.gouv.fr/)
carto : d'après IGN Admin Express 2022
https://r.iresmi.net/ {Sys.Date()}")) +
guides(color = guide_legend(reverse = TRUE),
shape = guide_legend(reverse = TRUE)) +
theme_minimal() +
theme(plot.caption = element_text(size = 6))

What would be an interesting station?
# longest temperature time series (possibly discontinuous)
meteo |>
filter(!is.na(tx)) |>
summarise(.by = c(nom_usuel, num_poste),
deb = min(year(aaaammjj)),
fin = max(year(aaaammjj)),
n_annee = n_distinct(year(aaaammjj))) |>
mutate(etendue = fin - deb) |>
arrange(desc(n_annee))
# A tibble: 922 × 6 nom_usuel num_poste deb fin n_annee etendue <chr> <chr> <dbl> <dbl> <int> <dbl> 1 ST-GENIS-LAVAL 69204002 1881 2023 129 142 2 CHAMONIX 74056001 1880 2023 110 143 3 MONESTIER 38242001 1916 2023 107 107 4 MONTELIMAR 26198001 1920 2023 104 103 5 LYON-BRON 69029001 1920 2023 104 103 6 BOURG EN BRESSE 01053001 1890 1992 103 102 7 CLERMONT-FD 63113001 1923 2023 101 100 8 PRIVAS 07186002 1865 1968 98 103 9 ANNECY 74010001 1876 1976 97 100 10 LE PUY-CHADRAC 43046001 1928 2023 96 95 # ℹ 912 more rows
Choosing the longest time series:
meteo |>
filter(num_poste == "69204002") |>
mutate(tm = if_else(is.na(tm), (tx + tn) / 2, tm)) |>
select(num_poste, nom_usuel, aaaammjj, tn, tm, tx) |>
pivot_longer(c(tn, tm, tx),
names_to = "mesure",
values_to = "temp") |>
mutate(mesure = factor(mesure,
levels = c("tx", "tm", "tn"),
labels = c("Tmax", "Tmoy", "Tmin"))) %>% {
ggplot(data = ., aes(aaaammjj, temp, color = mesure)) +
geom_point(size = 1, alpha = 0.01) +
geom_smooth(method = "gam",
formula = y ~ s(x, bs = "cs", k = 30)) +
scale_x_date(date_breaks = "10 years",
date_labels = "%Y",expand = expansion(),
limits = ymd(paste0(c(
round_any(min(year(.$aaaammjj), na.rm = TRUE), 10, floor),
round_any(max(year(.$aaaammjj), na.rm = TRUE), 10, ceiling)), "0101"))) +
scale_y_continuous(breaks = scales::breaks_pretty(4),
minor_breaks = scales::breaks_pretty(40)) +
scale_color_manual(values = c("Tmin" = "deepskyblue2",
"Tmoy" = "grey50",
"Tmax" = "firebrick2"),
labels = list("Tmax" = bquote(T[max]),
"Tmoy"= bquote(T[moy]),
"Tmin"= bquote(T[min]))) +
labs(title = glue("{.$nom_usuel[[1]]} ({str_sub(.$num_poste[[1]], 1, 2)})"),
subtitle = "Températures quotidiennes",
x = "année",
y = "℃",
caption = glue("données Météo-France (https://meteo.data.gouv.fr/)
https://r.iresmi.net/ {Sys.Date()}")) +
theme_minimal() +
theme(plot.caption = element_text(size = 6),
axis.title.y = element_text(angle = 0, vjust = 0.5))
}

To leave a comment for the author, please follow the link and comment on their blog: r.iresmi.net.
R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.