Deep-learning augmented RNA-seq analysis of transcript splicing | 用深度学习预测可变剪切

爱⌒轻易说出口 提交于 2020-04-24 07:22:16

可变剪切的预测已经很流行了,目前主要有两个流派:

  • 用DNA序列以及variant来预测可变剪切;GeneSplicer、MaxEntScan、dbscSNV、S-CAP、MMSplice、clinVar、spliceAI
  • 用RNA来预测可变剪切;MISO、rMATS、DARTS

前言废话

科研圈的热点扎堆现象是永远存在的,且一波接一波,大部分不屑于追热点且不出成果也基本都被圈子给淘汰了。

做纯方法开发的其实是很心累的,费时费力费脑,特别是自己的研究领域已经过时的时候,另外还得承受外行的歧视:“你们搞这个有什么用吗?文章也发不了,最后也没人用。”

最近这些年最大的一个热点就是“单细胞”,很多人都趁着这股东风捞了一些文章,最早一批开发方法的也发了不少nature method和NBT,bioinformatics和NAR更多。但大部分后面就销声匿迹了,因为门槛越来越低,进入者越来越多,经过几年的发展,现在已经成了三足鼎立之势,强者愈强,弱者退场。

写方法类的文章也有个潜规则,千万不要写得过于通俗易懂,大部分审稿人如果一眼就能看懂,就会自然认为你做的研究过于简单,没有发表的必要。最好要写得有理有据,且90%的审稿人没法一眼看懂,但细细琢磨后有那么点意思。哈哈,当笑话听就好。


跳到另外一篇用深度学习来预测可变剪切的。

Deep-learning augmented RNA-seq analysis of transcript splicing

PDF 

 

文章里面需要重点了解的基础知识:

Unlike methods that use cis sequence features to predict exon splicing patterns in specific samples7–10,看看前人是如何根据cis sequence特征来预测exon的剪切模式的

涉及到的文献:

The human splicing code reveals new insights into the genetic determinants of disease - 2015

Deciphering the splicing code - 2010

Deep learning of the tissue-regulated splicing code - 2014

BRIE: transcriptome-wide splicing quantification in single cells - 2017

哈哈,深度学习在可变剪切上的应用的风2014年就开始刮了,你不可能是第一个吃螃蟹的了。

想了解什么是AS,可以直接看现在开发的工具,里面肯定有图详细介绍,同时介绍其算法,一图胜千言

MISO (Mixture of Isoforms) software documentation 目前只支持python2版本,用conda的话还需要从文档中copy一下miso_settings.txt文件。

rMATS: Robust and flexible detection of differential alternative splicing from replicate RNA-Seq data

生物和信息之间存在一个巨大的gap,优秀的人能很快察觉到这个gap,并填补这个gap。

 

问题:

为什么AS的鉴定依赖测序深度?得了解现在主流的AS检测算法

怎么理解样本间的差异可变剪切这个概念?

如何理解cis sequence features,这个文件里都包含了哪些信息?

怎么predict exon-inclusion/skipping levels in bulk tissues or single cells

怎么理解we hypothesized that large-scale RNA-seq resources can be used to construct a deep-learning model of differential alternative splicing. 

 

两部分:

a deep neural network (DNN) model that predicts differential alternative splicing between two conditions on the basis of exon-specific sequence features and sample-specific regulatory features

a Bayesian hypothesis testing (BHT) statistical model that infers differential alternative splicing by integrating empirical evidence in a specific RNA-seq dataset with prior probability of differential alternative splicing

 

During training, large-scale RNA-seq data are analyzed by the DARTS BHT with an uninformative prior (DARTS BHT(flat), with only RNA-seq data used for the inference) to generate training labels of high-confidence differential or unchanged splicing events between conditions, which are then used to train the DARTS DNN. 

During application, the trained DARTS DNN is used to predict differential alternative splicing in a user-specific dataset.

This prediction is then incorporated as an informative prior with the observed RNA-seq read counts by the DARTS BHT (DARTS BHT(info)) for deeplearning-augmented splicing analysis.

差不多懂了,第一个BHT就是常规差异剪切分析工具(如MISO 和 rMATS)的升级版,用于制造有lable的训练数据。BHT的结果用于训练DNN模型;新的数据可以放进DNN模型里,得到的结果可以作为后面贝叶斯模型的先验,我们的RNA-seq数据则是用于更新先验形成后验,如果先验足够准确,则更新时对数据的依赖不搞,这也就是为什么该方法可以弥补RNA-seq测序深度不足的情形。

To generate training labels, we applied DARTS BHT(flat) to calculate the probability of an exon being differentially spliced or unchanged in each pairwise comparison. 

 

cis sequence features and messenger RNA (mRNA) levels of trans RNA-binding proteins (RBPs) in two conditions 

这个DNN把可变剪切转换成了一个regression的问题,特征就是上面两种,因为它们决定了最终的一个特征是否发生了可变剪切。

最终用到的特征:2,926 cis sequence features and 1,498 annotated RBPs

 

DNN用到的训练数据具体是什么?

large-scale RBP-depletion RNA-seq data in two human cell lines (K562 and HepG2) generated by the ENCODE consortium

We used RNA-seq data of 196 RBPs depleted by short-hairpin RNA (shRNA) in both cell lines, corresponding to 408 knockdown-versus-control pairwise comparisons

The remaining ENCODE data, corresponding to 58 RBPs depleted in only one cell line, were excluded from training and used as leave-out data for independent evaluation of the DARTS DNN 

From the high-confidence differentially spliced versus unchanged exons called by DARTS BHT(flat) (Supplementary Table 2), we used 90% of labeled events for training and fivefold cross-validation, and the remaining 10% of events for testing (Methods). 这样就把每个exon给的特征给提取出来了,lable也有了,就可以用于训练了。

 

比较了三个模型:

We used the leave-out data to compare the DARTS DNN with three alternative baseline methods: the identical DNN structure trained on individual leave-out datasets (DNN), logistic regression with L2 penalty (logistic), and random forest.

 

关于贝叶斯模型的部分:

incorporating the DARTS DNN predictions as the informative prior, and observed RNA-seq read counts as the likelihood (DARTS BHT(info)).

Simulation studies demonstrated that the informative prior improves the inference when the observed data are limited, for instance, because of low levels of gene expression or limited RNA-seq depth, but does not overwhelm the evidence in the observed data

 

如果文章看得迷迷糊糊的,就直接跑代码吧!

第一个功能BHT:

Darts_BHT bayes_infer --darts-count test_data/test_norep_data.txt --od test_data/

test_norep_data.txt文件是这样的:

ID	GeneID	geneSymbol	chr	strand	exonStart_0base	exonEnd	upstreamES	upstreamEE	downstreamES	downstreamEE	ID	IJC_SAMPLE_1	SJC_SAMPLE_1	IJC_SAMPLE_2	SJC_SAMPLE_2	IncFormLen	SkipFormLen
82439	ENSG00000169045.17_1	HNRNPH1	chr5	-	179046269	179046408	179045145	179045324	179047892	179048036	82439	15236	319	6774	834	180	90
21374	ENSG00000131876.16_3	SNRPA1	chr15	-	101826418	101826498	101825930	101826006	101827112	101827215	21374	4105	118	292	54	169	90
32815	ENSG00000141027.20_3	NCOR1	chr17	-	15990485	15990659	15989712	15989756	15995176	15995232	32815	624	564	549	1261	180	90
43143	ENSG00000133731.9_2	IMPA1	chr8	-	82597997	82598198	82593732	82593819	82598486	82598518	43143	155	332	22	341	180	90
111671	ENSG00000100320.22_3	RBFOX2	chr22	-	36232366	36232486	36205826	36206051	36236238	36236460	111671	93	193	35	534	180	90

每一行是一个基因,无冗余,然后就是一些属性.

跑出来的结果是这样的:

1 ID      I1      S1      I2      S2      inc_len skp_len psi1    psi2    delta.mle       post_pr
      2 1225    160     0       169     6       180     90      1       0.934   -0.0663 0.4367
      3 15829   52      58      12      41      180     90      0.31    0.128   -0.1819 0.8867
      4 20347   1084    930     371     615     180     90      0.368   0.232   -0.1365 1
      5 21374   4105    118     292     54      169     90      0.949   0.742   -0.2065 1
      6 24817   177     275     263     741     143     90      0.288   0.183   -0.1057 0.974
      7 32815   624     564     549     1261    180     90      0.356   0.179   -0.1774 1
      8 43143   155     332     22      341     180     90      0.189   0.031   -0.158  1
      9 46548   1685    4040    216     1752    180     90      0.173   0.058   -0.1145 1

每一行是对之前条目的预测。  

 

第二个功能DNN:

下载model

Darts_DNN get_data -d transFeature cisFeature trainedParam -t A5SS  

预测

Darts_DNN predict -i darts_bht.flat.txt -e RBP_tpm.txt -o pred.txt -t A5SS  

 其中的第一个文件是Input feature file (*.h5) or Darts_BHT output (*.txt)

ID      I1      S1      I2      S2      inc_len skp_len mu.mle  delta.mle       post_pr
chr1:-:10002681:10002840:10002738:10002840:9996576:9996685      581     0       462     0       155     99      1       0       0
chr1:-:100176361:100176505:100176389:100176505:100174753:100174815      28      0       49      2       126     99      1       -0.0493827160493827     0.248
chr1:-:109556441:109556547:109556462:109556547:109553537:109554340      2       37      0       81      119     99      0.0430341230167355
      -0.0430341230167355     0.188
chr1:-:11009680:11009871:11009758:11009871:11007699:11008901    11      2       49      4       176     99      0.755725190839695       0.117542135892979       0.329333333333333
chr1:-:11137386:11137500:11137421:11137500:11136898:11137005    80      750     64      738     133     99      0.0735580941766509      -0.0129207126090368     0

 第二个文件是Kallisto expression files

thymus  adipose
RPS11   2678.83013      2531.887535
ERAL1   14.350975       13.709394
DDX27   18.2573 14.02368
DEK     32.463558       14.520312
PSMA6   102.332592      77.089475
TRIM56  4.519675        6.14762566667
TRIM71  0.082009        0.0153936666667
UPF2    7.150812        5.23628033333
FARS2   6.332831        7.291382
ALKBH8  3.056208        1.27043633333
ZNF579  5.13265 8.248575

 结果文件,第一列是ID,第二列是真实的标签,第三列是预测的标签:

ID      Y_true  Y_pred
chr22:-:39136893:39137055:39137011:39137055:39136271:39136437   1.000000        0.318161
chr12:-:69326921:69326979:69326949:69326979:69326457:69326620   1.000000        0.073966
chr3:-:49053236:49053305:49053251:49053305:49052920:49053140    0.947333        0.295664
chr4:-:68358468:68358715:68358586:68358715:68357897:68357993    1.000000        0.304907
chr11:-:124972532:124972705:124972629:124972705:124972027:124972213     0.937333        0.365548
chr15:+:43695880:43696040:43695880:43695997:43696610:43696750   1.000000        0.450762

  

参考:

The Expanding Landscape of Alternative Splicing Variation in Human Populations. 

  

这篇是比较纯粹的DL应用:

Gene expression inference with deep learning | 基于深度学习的基因表达推测

案例文章:Gene expression inference with deep learning

uci-cbcl/D-GEX - github

深度学习的风已经过了几年了,目前在医疗影像处理上已经公认非常有效,所以后面想发文章必须数据足够大足够靓,方法上想创新太难。

LINCS L1000 data

核心的意思就是这个项目只测了不到一千个基因的表达,却要通过LR和DL来推测出其他全部的3万个基因的表达,所以称那978个基因叫landmark genes。

 

  

标签
易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!