Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the DeepRT+ calibration #13

Open
Chaohk opened this issue Jan 10, 2022 · 4 comments
Open

About the DeepRT+ calibration #13

Chaohk opened this issue Jan 10, 2022 · 4 comments

Comments

@Chaohk
Copy link

Chaohk commented Jan 10, 2022

Hi, could you please tell me that how the DeepRT+ do the calibration using a certain ratio of the test-group peptides after pretraining with the big data?

Thanks.

@horsepurve
Copy link
Owner

Hi! For the calibration from a pretrained model (run1) to another LC run (run2), please:

  1. Select a certain ratio of peptides from run2 as the calibration peptides (better with high identification confidence, and span across the whole LC run). Save as a txt file and assign its file path to "train_path" in "config.py".
  2. Set "pretrain_path" to the file path of your pretrained model, set "dict_path" and "max_length" to be the same with those of the pretrained model (see tutorial for details).
  3. Run by "python capsule_network_emb.py" and see if the prediction is improved and better than not using a model pretrained from bigger data.

@Chaohk
Copy link
Author

Chaohk commented Jan 11, 2022

Thank you for telling me the usage of the calibration function. Could I know what algorithm and methods are using in the RT calibration. Like piecewise linear regression?

@horsepurve
Copy link
Owner

The method is transfer learning. The model is first pre-trained on data1 from LC run1, then fine-tuned on data2 from LC run2.

The only difference between using/not using this pretraining & fine-tuning scheme is that: without it, we train a deep neural network (DNN) from scratch whose weights are randomly initialized; whereas pretraining provides a well-educated guess for the weight of DNN so that DNN can converge to a better minimum faster by stochastic gradient descent.

This pretraining & fine-tuning routine is extremely widely adopted in Computer Vision, Natural Language Processing, etc. In LCMS, our work is perhaps the first that uses this to integrate/calibrate multiple runs across different LCs/experiments/labs.

@Chaohk
Copy link
Author

Chaohk commented Jan 12, 2022

Fully understand. Thanks a lot!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants