Filters entries in a map using the function specified (this is essentially a dplyr wrapper to the `map_filter(expr, func)` higher- order function, which is supported since Spark 3.0)

hof_map_filter(x, func, expr = NULL, dest_col = NULL, ...)

Arguments

x

The Spark data frame to be processed

func

The filter function to apply (it should take (key, value) as arguments and return a boolean value, with FALSE indicating the key-value pair should be discarded and TRUE otherwise)

expr

The map being filtered, could be any SQL expression evaluating to a map (default: the last column of the Spark data frame)

dest_col

Column to store the filtered result (default: expr)

...

Additional params to dplyr::mutate

Examples

if (FALSE) { library(sparklyr) sc <- spark_connect(master = "local", version = "3.0.0") sdf <- sdf_len(sc, 1) %>% dplyr::mutate(m = map(1, 0, 2, 2, 3, -1)) filtered_sdf <- sdf %>% hof_map_filter(~ .x > .y) }