
You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
Deep Learning for Regulatory Genomics - Regulator binding, Transcription Factors TFs - Lecture 08 (Spring 2021)
Deep Learning for Regulatory Genomics - Regulator binding, Transcription Factors TFs - Lecture 08 (Spring 2021)
The video discusses the use of deep learning for regulatory genomics and focuses on how DNA sequence can reveal different motifs present in enhancer and promoter regions, and their 3D looping. The video explains how Chromosome confirmation capture (3C) technology can probe chromosomal organization, and Hi-C technology can identify topologically associated domains (TADs), which interact with each other, and the compartment pattern in the genome. Convolutional filters are applied at every position of the DNA sequence to detect different features or motifs, and the deep learning framework can learn common properties, filters and motifs of the DNA sequence, which enable various prediction tasks to be carried out. The video also mentions how multitask learning is beneficial, and using additional layers in the deep learning network to recognize and combine multiple building block representations of transcription factor motifs could allow for more efficient recognition of complex motifs.
The speaker in this video discusses using deep learning for regulatory genomics with a focus on transcription factor binding and gene expression prediction. They explore the use of convolution structures and dilated convolutions to bring in large regions of DNA and make predictions in a multi-task framework for chromatin data and gene expression. The speaker also covers the use of residual connections to train deep neural nets and explains how the model can predict 3D contacts using IC data and models. Overall, deep learning can be a powerful tool for analyzing genomics data and making predictions based on DNA sequence with enough data and the right transformations.
Gene Expression Prediction - Lecture 09 - Deep Learning in Life Sciences (Spring 2021)
Gene Expression Prediction - Lecture 09 - Deep Learning in Life Sciences (Spring 2021)
The video discusses the use of deep learning in gene expression prediction and the challenges involved in analyzing biological data sets, including high dimensionality and noise. The lecture covers methodologies such as cluster analysis, low-rank approximations of matrices, and compressive sensing. The speaker also talks about the use of deep learning for gene expression prediction and chromatin, as well as weakly supervised learning for predicting enhancer activity sites. The lecture discusses several tools developed using primarily deep learning methodology, including danq, djgx, factory mat, and sc fin. The presenter also talks about the use of generative models for studying genomics data sets and introduces the idea of approximate inference methodology, particularly the popular one called variational inference.
In the second part of the lecture, the speaker discusses the application of deep learning in life sciences, specifically in gene expression prediction and genomic interpretation. The first topic focuses on the application of variation autoencoder models to RNA expression analysis for asthma datasets. The speaker proposes a framework to remove experimental artifacts using a conditional generative model. The second topic discusses Illumina's investment in deep learning networks to identify the sequence-to-function models for genomic interpretation, particularly for splicing. The company has developed SpliceAI, a deep convolutional neural network that predicts whether a nucleotide is a splice donor, acceptor, or neither. The third topic is about the speaker's research on predicting whether certain mutations will have cryptic splice function, which can lead to frameshifts and disease. The speaker also invites questions and applications for research positions, internships, and postdocs.
Single Cell Genomics - Lecture 10
Single Cell Genomics - Lecture 10 - Deep Learning in Life Sciences (Spring 2021)
In this lecture on single-cell genomics, the speaker discusses various methods and technologies used for profiling individual cells, including cell sorting and microfluidics. The focus is on three specific single-cell sequencing technologies - Smart-seq, drop-seq, and pooled approaches. The speaker also covers the process of analyzing single-cell transcriptomes, including preprocessing, visualization, clustering, and annotation, and the use of autoencoder architecture in community clustering. Deep learning methods are applied for domain adaptation and to reconstruct cell types in a stimulated fashion. The lecture also discusses the challenges involved in analyzing single-cell genomics data and proposes the use of a generative model to address these issues in a scalable and consistent way.
The second part of the video covers various topics related to single-cell genomics and deep learning. Topics discussed include variational inference, a generative process for single-cell RNA sequencing data, the SCVI model for mixing cell type datasets, CanVAE for propagating labels, and the implementation of various deep learning algorithms on a single code base called CVI tools. The speakers also address challenges in using posterior probabilities to calculate measures of gene expression and present methods for accurately calculating posterior expectations and controlling full discovery rates.
Dimensionality Reduction - Lecture 11
Dimensionality Reduction - Lecture 11 - Deep Learning in Life Sciences (Spring 2021)
The video lectures on deep learning in life sciences explore dimensionality reduction techniques for clustering and classification in single-cell data analysis. The lectures distinguish between supervised and unsupervised learning and explore the use of statistical hypothesis testing frameworks for evaluating differential expressions of genes. The lecture introduces the concept of manifold learning using principal component analysis, eigen decomposition, and singular value decomposition for linear dimensionality reduction and discusses the methods of t-distributed stochastic neighbor embedding and distributed stochastic neighbor embedding for clustering data preservation. The speaker also discusses the application of non-negative matrix factorization to genomic data and the integration of single-cell and multi-omic data sets. The ultimate goal of these techniques is to redefine cell types and identity in an unbiased and quantitative way.
The second part discusses several topics related to dimensionality reduction, specifically its application in life sciences. Integrative non-negative matrix factorization (iNMF) is used to link transcriptomic and epigenomic profiles to better understand cellular identity across various contexts. The lecture also discusses the benefits of using a mini-batch approach in deep learning, particularly for larger datasets, and how online algorithms can be leveraged to improve dimensionality reduction methods for analyzing large datasets. Additionally, the algorithm is introduced to integrate different types of data, such as RNA-seq and ATAC-seq data. Finally, the speaker expresses willingness to serve as a mentor for students interested in the field. Overall, the lecture was informative and well-received.
Disease Circuitry Dissection GWAS - Lecture 12
Disease Circuitry Dissection GWAS - Lecture 12 - Deep Learning in Life Science (Spring 2021)
This video on disease circuitry dissection GWAS covers the foundations of human genetics, the computational challenges for interpretation, and the various types of genetic variations examined in genome-wide association studies (GWAS). The video also explores methodologies such as Mendelian mapping, linkage analysis, and the identification of single nucleotide polymorphisms (SNPs) associated with diseases. Additionally, the speaker discusses the use of chi-square statistics, Manhattan plots, and QQ plots to visualize genomic regions significantly associated with disease phenotypes. The video also includes a case study on the FTO gene and how it was comprehensively dissected for its mechanistic implications in obesity. The challenges of understanding the genetic association with obesity and the steps to approach this issue are also discussed.
The lecture discusses the challenge of studying the impact of genomic variations on human health, and the importance of understanding how mutations affect different cell types. The speaker outlines their deep learning approach to predicting the effect of genomic sequence and variations, particularly in relation to predicting the binding of transcription factors and the organization of chromatin. They also describe their evaluation of these predictions using deeply sequenced genomic datasets to predict DNA sensitivity and histone mark QTLs, as well as their use of deep learning to predict the effect of mutations on gene expression and human diseases such as autism. Finally, they discuss their unbiased analysis of previously known gene sets and the use of a deep learning sequence model library.
GWAS mechanism - Lecture 13
GWAS mechanism - Lecture 13 - Deep Learning in Life Sciences (Spring 2021)
The lecture on GWAS mechanism in the Deep Learning in Life Sciences series looks at various methods to understand the function of non-coding genetic variants involved in complex traits. The lecture discusses the use of epigenomic annotations and deep learning models to identify global properties across genetically associated regions for a particular disease. It also covers enrichments across different tissues and enhancers and explains how these can be turned into empirical priors to predict the causal SNP within a locus. The lecture also discusses the use of intermediate molecular phenotypes like gene expression and methylation to study causality in genome-wide association studies and how to combine genotype and expression personal components to explain the phenotypic variable of expression. Lastly, the lecture examines the use of causal inference methods to determine the effect of changing a variable on outcome variables to identify causal versus anti-causal pathways.
The lecturer in this video discusses various techniques for inferring causal effects in genomics research. They cover the concept of d-separation and using natural randomization in genetics as a way to establish causal relationships. The lecturer also discusses Mendelian randomization and Rubin's Quasi-Inference Model, along with the potential outcome method for causal inference. They touch on the challenges of imputation and adjusting for biases in observational studies. The speaker also stresses the importance of using multiple orthogonal evidence to develop a robust causal algorithm. Additionally, they explain the use of genetics to perturb gene expressions and learn networks, and introduce the invariance condition as a way to identify causal structures in data. The lecture provides a comprehensive overview of various techniques and tools used in genomics research for causal inference.
Systems Genetics - Lecture 14
Systems Genetics - Lecture 14 - Deep Learning in Life Sciences (Spring 2021)
In this lecture on systems genetics and deep learning, the speaker covers several topics, including SNP heritability, partitioning heritability, stratified LD score regression, and deep learning in molecular phenotyping. They also explore the use of electronic health records, genomic association studies, and genomics to analyze a UK biobank dataset of around 500,000 individuals with thousands of phenotypes. The lecturer discusses how deep learning models can be used for sequence function prediction to understand the circuitry of disease loci and the use of linear mixed models for GWAS and EQTL calling. They also touch on the biases and violations of model assumptions in deep learning and highlight the importance of cell type-specific regulatory annotations in inferring disease-critical cell types. Lastly, the lecturer discusses the complexity of findings related to negative selection and causal effect sizes and introduces Professor Manuel Rivas from Stanford University to discuss the decomposition of genetic associations.
The lecture delves into the application of genetic data in various areas, including quantifying the composition and contribution components of traits, identifying genetic variants that contribute to adipogenesis or lipolysis, identifying mutations with strong effects on gene function and lower disease risk, and the development of risk prediction models using multivariate analysis. Additionally, the lecture discusses the application of polygenic risk score models in various biomarkers and stresses the need for data sharing across different populations to improve predictive accuracy, particularly in the case of non-European populations. The lecture concludes by expressing a willingness to supervise students interested in research projects related to UK Biobank polygenic scores and biotropic effects.
Graph Neural Networks - Lecture 15
Graph Neural Networks - Lecture 15 - Learning in Life Sciences (Spring 2021)
In this YouTube lecture on Graph Neural Networks, the speaker covers a wide range of topics, including the basics of graph networks, spectral representations, semi-supervised classification, and multi-relational data modeling. There is also a focus on the intersection of graph networks and natural language processing and how to generate graphs for drug discovery. The lecturer explains various methods to propagate information across graphs to obtain useful node embeddings that can be used for prediction tasks. The lecture also highlights the importance of contrastive learning for GNNs, the potential benefits of combining patch-based representations and attention-based methods, and the use of the transformer approach in NLP. The latter half of the lecture focuses on discussing papers that showcase the practical uses of GNNs in drug discovery and how to encode and decode the structure of molecules using a junction tree.
This video discusses multiple applications of graph neural networks (GNNs) in life sciences, including drug discovery and latent graph inference. The speaker highlights the issues and potential avenues in GNNs, such as the lack of spatial locality and fixed ordering, and the setup considered involves predicting the type of a given node, predicting a link between two nodes, measuring similarity between two nodes or two networks, and clustering nodes by performing community detection in the network. The lecturer also explains how GNNs can efficiently train and embed graphs, transform and aggregate information, and deal with polypharmacy side effects. Additionally, the lecture covers two methods for automatically learning representations in life sciences, with meta-learning models like MARS being leveraged to generalize to novel cell types. Lastly, the lecture discusses how GNNs can learn latent cell representations across multiple datasets to capture cell type heterogeneity.
AI for Drug Design - Lecture 16
AI for Drug Design - Lecture 16 - Deep Learning in the Life Sciences (Spring 2021)
This lecture discusses the use of deep learning for drug design. It explains how deep learning can be used to find novel compounds with antibiotic resistance. It also discusses how the deep learning models can be improved by incorporating biological knowledge.
This second part of the lecture provides an overview of how deep learning can be used in drug design, specifically for predicting the antiviral activity of drug combinations. The model was tested in vivo using cell-based assays and two novel synergistic drug combinations were identified.
Deep Learning for Protein Folding - Lecture 17
Deep Learning for Protein Folding - Lecture 17 - MIT Deep Learning in Life Sciences (Spring 2021)
This video discusses the use of deep learning in the field of protein folding, and specifically how geometric deep learning can be used to study protein structures and predict things such as ligand-binding sites and protein-protein interactions. The video also covers template-based vs. template-free modeling methods, various approaches for contact prediction in protein folding, and the use of residual neural networks for image modeling in protein structure prediction. Overall, the speaker emphasizes the promise of deep learning in advancing our understanding of protein structures and their functions, and provides detailed examples and results to back up this claim.
The video discusses various approaches to deep learning for protein folding, including the use of co-evolution predictions and templates for accurate modeling, the importance of finding better homologs, and the potential for deep learning to achieve comparable results without relying on traditional physics-based methods. The speakers also delve into the use of differentiable outputs and the importance of global accuracy, as well as the evolution of algorithm space and the potential for deep learning to predict protein confirmations based on factors such as genetic variation or small molecules. Overall, the video highlights the exciting potential for deep learning to revolutionize protein structure prediction and its many applications.