I am using pyspark 1.5 getting my data from Hive tables and trying to use windowing functions.
According to this there exists an analytic function called firstValue
that will give me the first non-null value for a given window. I know this exists in Hive but I can not find this in pyspark anywhere.
Is there a way to implement this given that pyspark won't allow UserDefinedAggregateFunctions (UDAFs)?