Laboratory of Artificial Intelligence for Medicine and Science
- Computational biology & medicine - precision medicine, systems biology, & genetics
- Machine learning - explainable AI, interpretability, feature selection, & probabilistic graphical models
Modern machine learning (ML) models can accurately predict patient progress, an individual's phenotype, or molecular events such as transcription factor binding. However, they do not explain why selected features make sense or why a particular prediction was made. For example, a model may predict that a patient will get chronic kidney disease, which can lead to kidney failure. The lack of explanations about which features drove the prediction – e.g., high systolic blood pressure, high BMI, or others – hinders medical professionals in making diagnoses and decisions on appropriate clinical actions. The AIMS Lab seeks to develop approaches based on explainable artificial intelligence (AI) and machine learning (ML) for medicine and biology.
Some of her highlighted research include (i) finding therapeutic targets for Alzheimer’s disease (press articles), (ii) treating cancer based on a patient’s own molecular profile (Nature Communications 2018; Selected by F1000), (iii) ML theories on explainable AI (NeurIPS Dec 2017 ; Full Oral (top 1%); Cited 233 times as of June, 2019), and (iv) preventing complications during surgery (Nature BME 2018; Cover article), (v) predicting kidney diseases, (vi) enabling pre-hospital predictions for trauma patients, and (vii) improving our understanding of pan-cancer biology, (viii) human genome, and (ix) gene regulatory networks. Lee Lab is collaborating with biomedical researchers in UW School of Medicine, Allen Institutes, Harvard Medical School, University of British Columbia, etc.
- 2/21/2019: Nao's AIControl paper got accepted for publication in Nucleic Acids Research (IF: 11.56)! See Allen School News.
- 12/31/2018: Scott's SHAP paper (NeurIPS oral presentation) got cited 100 times as of today after about 1 year it was published.
- 11/1/2018: Our EMBARKER project on identifying therapeutic targets for Alzheimer's disease won the Madrona Prize at the Allen School 2018 Industry Affiliates Annual Research Day.
- Last year, Scott won the same prize on his model interpretation work, SHAP (NIPS oral presentation) and his Nature BME paper featured on the cover (see below).
- 10/10/2018: Scott's Prescience paper is published as a cover article of the October issue of Nature Biomedical Engineering.
- Nature BME Editorial - Towards trustable machine learning
- UW News - Prescience: Helping doctors predict the future
- GeekWire - Univ. of Washington researchers unveil Prescience, an AI system that predicts problems during surgery
- Allen School News - “Prescience” interpretable machine-learning system for predicting complications during surgery featured in Nature Biomedical Engineering
- 5/3/2018: Safiye Celik's and Su-In Lee's research is featured in GeekWire.
- 4/3/2018: Hugh Chen receives the NSF Graduate Research Fellowships Program (GRFP).
- 3/27/2018: Safiye Celik's MERGE paper (Nature Communications 2018) is recommended in F1000Prime as being of special significance in its field.
- 3/21/2018: Safiye Celik's research is featured in I Am CSE.
- Su-In's lecture was selected as the best talk at CGWI 2018 (Computational Genomics Winter Institute): Interpretable Machine Learning for Precision Medicine.
- 1/3/2018: Safiye's precision oncology paper got published in Nature Communications. See Allen School News, UW Huddle, and UW Medicine News.
- 9/4/2017: Scott Lundberg's SHAP paper is accepted to Neural Information Processing Systems (NIPS) 2017 for Full Oral Presentation.
- 8/7/2017: Gabriel Erion won the Best Poster Award: "Prediction and Prevention of Perioperative Adverse Events with Machine Learning Models", University of Washington MSTP (MD/PhD program) retreat, 2017.
- 2/7/2017: Scott Lundberg's ChromNet paper (Genome Biology 2016) is recommended in F1000Prime as being of special significance in its field.
- 11/20/2016: Scott Lundberg receives the Best Paper Award at the NIPS workshop "Interpretable Machine Learning for Complex Systems".
- 8/12/2016: Javad Hosseini's GRAB paper is accepted to Neural Information Processing Systems (NIPS) 2016.
- 7/15/2016: Nao Hiranuma's CloudControl paper is accepted to ACM Conference on Bioinformatics, Computational Biology (ACM-BCB) 2016.
- 7/6/2016: Safiye Celik's INSPIRE paper is featured in Casey Greene. The future is unsupervised. Science Translational Medicine July 2016.
- 6/10/2016: Safiye Celik's INSPIRE work is published in Genome Medicine June 2016.
- 5/4/2016: Maxim Grechkin's DISCERN work is published in PLOS Computational Biology May 2016.
- 4/30/2016: Scott Lundberg's ChromNet work is published in Genome Biology April 2016. See ChromNet browser.
- 2/8/2016: Javad Hosseini got ranked #1 in the DiMSUM competition. See Hosseini, Smith and Lee. NAACL Workshop SemEval 2016 Task10.
Prof. Su-In Lee's lab seeks to develop interpretable machine learning techniques to learn from big data: (1) how the human genome or protein works, (2) how to improve healthcare, and (3) how to treat challenging diseases such as cancer and Alzheimer's disease. Her research page lists her projects, including treating cancer based on a patient's own expression profile, finding therapeutic targets for Alzheimer's, predicting kidney disease, preventing complications during surgery, enabling pre-hospital predictions for trauma patients, analyzing medical images, and improving our understanding of pan-cancer biology and genome biology.