-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to inspect/compare extracted features at scale? #13
Comments
quick status update on this. currently considering to:
|
In relation to this:
Sounds good! We still need to decide what features to focus on (or how to cluster them) and what criteria we want for the comparisons (are we doing any statistical testing for validation as you initially wrote as an idea?). |
one possible approach we discussed earlier would be to use multidimensional distance metrics for selection -- something along the lines of: a) doing PCA on the full feature set; b) then computing euclidean distances on rotated features (maybe with some weighting to account for different features capturing a different amount of variance), and selecting the prompt that yields the lowest average distance from the human completions. this could be a valid approach, though partly circular (or too adversarial) if we use the same features to build a classifier that discriminates between human- and AI-generated, so we may still consider alternative approaches based on hand-picking or a tiny bit human validation, or even using ROUGE-like scores for selection? if this is not operationally urgent, we can maybe defer this to a little meeting/sprint with the whole team involved |
I don't think it is urgent currently, so that sounds like a plan! |
an update for future selves, as it's been a long time since this has been opened. we have decided to go for minimal engineering of prompts -- they should be reasonable instructions, and generic enough to be reasonable prompts for multiple models. |
also closing because not needing action anymore |
Currently, even with TextDescriptives only (i.e., no additional feature sets for cognitive features) we have dozens of features and it is unfeasible to look at their distributions individually, both in the context of simple data exploration and for actual modeling purposes.
There are a few options to deal with exploding dimensionality here:
On top of this, additional output-driven dimensionality reduction (e.g., LASSO) could be applied to select for the feature set that is a) most affected by prompting; or b) most discriminative between human- and model-generated text
The text was updated successfully, but these errors were encountered: