I wanted to get some opinions on using the original score feature. The original
score produced by Solr is intuitively a very important feature. In my data set
I'm seeing that the original score varies wildly between different queries.
This makes sense since the score generated by Solr is not normalized across all
queries. However, won't this mess with our training data? If this feature is
3269.4 for the top result for one query, and then 32.7 for the top result for
another query, it does not mean that the first document was 10x more relevant
to its query than the second document. I am using a normalize param within
Ranklib, but that only normalizes features between each other, not within one
feature, right? How are people handling this? Am I missing something?