Quantitative digital image analysis and machine learning for staging of prostate cancer at diagnosis.

Cancer Research(2018)

引用 4|浏览24
暂无评分
摘要
Introduction and Objective: Prostate cancer (PC) with de novo bone metastases at diagnosis (M1) carries a 5-year survival rate of 28% and requires early, aggressive treatment. Clinical assays and the pathology of prostate needle biopsies (PNBX) cannot distinguish primary M1 tumors from high-grade localized (M0) cases. We hypothesized that digital image analysis can be applied to obtain morphologic biomarkers, not recognizable by pathologists. Here we demonstrate how novel software tools that involve deep learning frameworks can be used to systematically extract handcrafted and autoencoder features and to build models to predict M1 stage at the time of diagnosis. Methods: A study cohort, nested within a biorepository of 2150 PC patients at the Greater LA VA, consisted of 86 high-grade M0 and 85 M1 cases. Slides were digitized at 40X and 2 pathologists annotated all cancer foci. Approximately 30 image tiles were selected from each case. 62 handcrafted (HC) and 64 autoencoder (AE) features were extracted from nuclei. Feature values normalized. The normalized profile of each primary feature gave rise to 11 secondary features, representing the distribution of the feature within a case. We separated cases into training + testing versus validation groups at an 80:20 ratio. Using a bootstrapping method, we selected the best GLMNET models predicting M0 versus M1 status in the training + testing set and applied them to an independent validation set of cases. Results: After successful conversion of M0 and M1 image tiles to digital nuclear masks and color normalization, ~ 400,000 nuclei were isolated using parameters that enriched for nuclei from cancer cells. A denoising autoencoding neural network was used to generate AE biomarkers for each nucleus. HC features quantified nuclear shape, size, color, and texture. A systematic pipeline of preprocessing, normalization, and conversion to case-level secondary features was applied to AE and HC features. For both feature types, the average of 50,000 bootstrapping models resulted in an AUC of 0.8 for the training and an average accuracy of 0.7 for the test cohort. The best 38 AE models or 16 HC models were applied to the independent validation cohort of 24 cases and assigned each case to the M0 versus M1 groups by majority voting. At a threshold of 0.5, this resulted in an accuracy of 70% for M0 versus M1 distinction. Conclusion: We applied digital imaging technology and machine learning software to AE and HC features in order to predict M0 versus M1 stage from the tumor in PNBXs at diagnosis. Unexpectedly, hidden features in nuclei differed between M0 and M1 cases and succeeded in predicting metastatic disease with 70% accuracy. The ultimate goal is to combine the inexpensive risk prediction from quantitative imaging with clinical parameters and RNA sequencing data to develop accurate prediction models of occult metastases and risk of future metastatic progression at the time of diagnosis in all patients with high-grade PC. Funding: DOD PC131996, PCF-Movember GAP1 Unique TMAs Project, Prostate Cancer Foundation (PCF) Creativity Award, Jean Perkins Foundation, NIH/NCI P01 CA098912-09, NIH R01CA131255 and P50CA092131, Stephen Spielberg Team Science Award. Citation Format: Fangjin Huang, Nathan Ing, Miller Eric, Hootan Salemi, Michael Lewis, Isla Garraway, Arkadiusz Gertych, Beatrice Knudsen. Quantitative digital image analysis and machine learning for staging of prostate cancer at diagnosis [abstract]. In: Proceedings of the AACR Special Conference: Prostate Cancer: Advances in Basic, Translational, and Clinical Research; 2017 Dec 2-5; Orlando, Florida. Philadelphia (PA): AACR; Cancer Res 2018;78(16 Suppl):Abstract nr B094.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要