Description
🚀 Feature
Extension of the methods in ot.da.*
for regression problems (by now only classification (?)).
Motivation
I already used ot.da.SinkhornLpl1Transport
for domain adaptation in (semi-)supervised classification problems (i.e. in ot.da.SinkhornLpl1Transport.fit(Xs, ys, Xt, yt)
, where yt
contains either the class label (a positive scalar) of a sample or -1
if the label is unknown). The only way I found in order to transfer this method to a (metric) regression problem is to convert the regression problem to a classification problem (e.g. by discretising the metric target value y
in e.g. 10 classes). Of course this conversion is not ideal as both the natural order of y
and distances between y
s get lost in a classification problem.
Pitch
Ideally yt
is capable of taking both class labels or metric target values. Samples without a label information are marked with e.g. numpy.nan
instead of -1
. The decision whether it is a regression or a classification problem is either clarified with an additional parameter, e.g. is_cls=True/False
or automatically (harder).
Alternatives
Maybe I am missing something and there is already a possibility for regression problems or it is impossible to implement as OT is not capable of working with yt
s of metric scale.
Additional context
Nothing to add here.