Skip to content

Evolutionary search for survival analysis loss function for neural networks

Notifications You must be signed in to change notification settings

abdoush/SurvLossEvo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Improving Concordance Index in Regression-based Survival Analysis: Evolutionary Discovery of Loss Function for Neural Networks

The official repository for the paper "Improving Concordance Index in Regression-based Survival Analysis: Evolutionary Discovery of Loss Function for Neural Networks" accepted at GECCO 2024.

BibTeX Citation

Will be available soon

For each dataset (Nwtco, Flchain, Support), there are 6 notebooks. Run each of them to do the following:

  1. Search_[dataset_name]_LossRepeated:

    Experiment to optimize the full function f(x)+g(x), repeated 10 times.

  2. Search_[dataset_name]_LossRepeated_Fix_Left:

    Experiment to optimize the censored part g(x) and fixing the events part to f(x)=x^2, repeated 10 times..

  3. Search_[dataset_name]_LossRepeated_Fix_Right:

    Experiment to optimize the events part f(x) and fixing the censored part to g(x)=max(0,x)^2, repeated 10 times..

  4. Search_[dataset_name]_LossRepeated_Left_Right:

    Comparison between the results of the optimization of the Full function f(x)+g(x), fixing g(x), and fixing f(x). You need to copy the results from the first three notebooks.

  5. Softplus_Study_[dataset_name]:

    Comparison between MSCEsp and the truncated MSCEsp.

  6. Search_[dataset_name]_LossRepeated_Left_Right_LogSig:

    Comparison between optimization and MSCEsp fucntion. . You need to copy the results from the previous (1, 2, 3, and 5) notebooks.