Filters entries in a map using the function specified (this is essentially a dplyr wrapper to the `map_filter(expr, func)` higher- order function, which is supported since Spark 3.0)

hof_map_filter(x, func, expr = NULL, dest_col = NULL, ...)



The Spark data frame to be processed


The filter function to apply (it should take (key, value) as arguments and return a boolean value, with FALSE indicating the key-value pair should be discarded and TRUE otherwise)


The map being filtered, could be any SQL expression evaluating to a map (default: the last column of the Spark data frame)


Column to store the filtered result (default: expr)


Additional params to dplyr::mutate


if (FALSE) { library(sparklyr) sc <- spark_connect(master = "local", version = "3.0.0") sdf <- sdf_len(sc, 1) %>% dplyr::mutate(m = map(1, 0, 2, 2, 3, -1)) filtered_sdf <- sdf %>% hof_map_filter(~ .x > .y) }