You had no improvements essentially because you are not using Dask for the right use case. Dask, as Spark, is designed for data that do not fit your central memory, translating your ML code into a computation graph that can be distributed across several cores or machines, very much in the flavor of tensorflow.
While sklearn has improved over the years, Dask is still useful for massive big data tasks that can be converted into a graph.