SlideShare a Scribd company logo
1 of 22
DEEP LEARNING FOR
FEATURE SELECTION IN
CASE AND CONTROL
PREDICTION
Ana Sanchez
Graduate Student Project Day
Spring 2015
Introduction
• Deep learning is a method that uses sets of
machine learning algorithms to model data.
• Learning representations of data algorithms are
used to obtain information that may otherwise not
be detected.
• In this project, deep learning is used for case and
control prediction in genome wide association
studies, specifically to construct new feature
vectors.
Methods
• SNP data is used to implement deep learning.
Currently experimenting with Upper GI Cancer
data.
• The data contains approximately 5000 subjects
(case and control) and was reduced to 10 case
subjects and 10 control subjects.
• There are 491,774 SNP features per subject. The
SNPs are condensed to patches of 100.
Methods - Patching
• The patches are generated by starting at index 0
to 99, then shifting by 50 SNPs and obtaining the
next 100 SNPs until the final SNP (ex: 1st patch 0-
99, 2nd patch 49-149, 3rd patch 100-199, etc.).
• Python program is written to accomplish patching.
Python Script for Patching
Output of Patching
A 196680 X 100 matrix. The matrix
represents 9834 patches per patient.
Methods – Deep Learning with K-Means
• To implement Deep learning, K-Means is used.
• The K-Means algorithm clusters data by
separating samples in N groups within cluster
sum of squares.
• Number of clusters must be specified.
Methods – Deep Learning with K-Means
• For the purpose of this project, Scikit-Learn was
used to implement the K-Means algorithm.
• Scikit-Learn is open source software that
provides libraries for machine learning in Python.
• Using imported libraries like numpy, sklearn
clusters, and sklearn metrics simple programs are
written to handle the complexities of the
algorithm.
• For K-Means, the number of K clusters were K=
1000 and K=10000.
Python Program for K-Means
Output for K-Means
One file contains a 196680 x 1 matrix.
This represents the label for each patch.
A second file contains a 1000 x 100 matrix. This is the
distance of each data point to the centroid. In other words,
the coordinates of the data points to each cluster.
Methods – Constructing a Distance Matrix
• At this point, three files have been created from
the original set, Patches, Centers, and Labels.
• The next step is to construct a pairwise distance
matrix. This will construct a file that contains the
distance of every data point, in each patch, to
every center cluster.
• Python program is written with patch and center
files as input.
Python Program for Distance Matrix
Output for Distance Matrix
A distance matrix 196680 x 1000 results.
Output for Distance Matrix
Screen Shot of word count for distance matrix output. Shows
196680 rows. When dividing 196680000 by the rows, the
columns are obtained. In this case there are 1000 columns.
Therefore, a 196680 x 1000 matrix resulted. This matrix
represents the number of patches by the number of K
centers.
Methods – Constructing New Feature Vector file
• Now that the Distance Matrix is obtained, the final
New Feature Vector file is created.
• Since the matrix has196680 x K, the sum of each
column for every 9834 rows needs to be
obtained. 9834 rows represent 1 patient and
should be represented in the final file.
• A python program was written that uses the
Distance Matrix file as input.
Python Program for New Features
Output of New Features
A New Vector matrix results, 20 x 1000.
Output of New Features
Screen Shot of word count for new feature vector output. Shows 20 rows.
When dividing 20000 by 20, the number of columns is obtained. In this
case, the number of columns is 1000. This represents the original 20
patients with their new features, 1000, based on the patching, kmeans,
pair wise, and summation methods that were applied.
Encountered Issues throughout Project
• Large files
• Need for extra programs to download
• When using Sci Kit, running normal K-means is extremely
time consuming. Instead MiniBatchKmeans is used.
• MBK makes subsets of input data, greatly reduces computation
time. Performs only slightly worse.
• Even with MBK, computation time for K=100000 is too
long and was not used.
• Still, long runtimes( k=1000 (~30 mins), k=10000
(~60mins), k=100000 (>16hours)).
• Long runtimes even for simple scripts due to large data.
Comparison of Kmeans and
MiniBatchKmeans
Image and comparison from SciKit-Learn website.
Conclusions
• The process was completed twice, for K= 1000, and K =
10000.
• Using deep learning ideas, a 20 x 491,774 data set was
properly processed into a 20x1000 file and a 20x10000
file.
• This significant reduction in file size saves memory.
• Classification algorithms can implemented on the new
files. Prediction Accuracy may be higher for the new files
than raw data.
References and Acknowledgments
• Scikit-learn: Machine Learning in Python, Pedregosa et al.,
JMLR 12, pp. 2825-2830, 2011.
• Michael K. K. Leung, Hui Yuan Xiong, Leo J. Lee, and
Brendan J. Frey Deep learning of the tissue-regulated
splicing code Bioinformatics 2014 30: i121-i129.
• Rasool Fakoor, Faisal Ladhak, Azade Nazi, Manfred Huber
Using Deep Learning to enhance cancer diagnosis and
classification Proceedings of the 30th International
Conference on Ma- chine Learning, Atlanta, Georgia, USA,
2013. JMLR: W&CP volume 28
Special Thanks to Mohammadreza Esfandiari, Payam, for
guidance and assistance throughout project.

More Related Content

What's hot

Introduction to Machine Learning with Python and scikit-learn
Introduction to Machine Learning with Python and scikit-learnIntroduction to Machine Learning with Python and scikit-learn
Introduction to Machine Learning with Python and scikit-learnMatt Hagy
 
One Algorithm to Rule Them All: How to Automate Statistical Computation
One Algorithm to Rule Them All: How to Automate Statistical ComputationOne Algorithm to Rule Them All: How to Automate Statistical Computation
One Algorithm to Rule Them All: How to Automate Statistical ComputationWork-Bench
 
Ml study notes id3
Ml study notes   id3Ml study notes   id3
Ml study notes id3Feri Handoyo
 
Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)Ha Phuong
 
Seminar_Presentation_ppt
Seminar_Presentation_pptSeminar_Presentation_ppt
Seminar_Presentation_pptAyushDixit52
 
A multilevel automatic thresholding method based on a genetic algorithm for a...
A multilevel automatic thresholding method based on a genetic algorithm for a...A multilevel automatic thresholding method based on a genetic algorithm for a...
A multilevel automatic thresholding method based on a genetic algorithm for a...Akshit Arora
 
Support Vector Machine (Classification) - Step by Step
Support Vector Machine (Classification) - Step by StepSupport Vector Machine (Classification) - Step by Step
Support Vector Machine (Classification) - Step by StepManish nath choudhary
 
Co-clustering of multi-view datasets: a parallelizable approach
Co-clustering of multi-view datasets: a parallelizable approachCo-clustering of multi-view datasets: a parallelizable approach
Co-clustering of multi-view datasets: a parallelizable approachAllen Wu
 
Clustbigfim frequent itemset mining of
Clustbigfim frequent itemset mining ofClustbigfim frequent itemset mining of
Clustbigfim frequent itemset mining ofijfcstjournal
 
A scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clusteringA scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clusteringAllenWu
 
Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...
Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...
Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...Kenko Nakamura
 
핵심 딥러닝 입문 4장 RNN
핵심 딥러닝 입문 4장 RNN핵심 딥러닝 입문 4장 RNN
핵심 딥러닝 입문 4장 RNNJaey Jeong
 
Dm part03 neural-networks-homework
Dm part03 neural-networks-homeworkDm part03 neural-networks-homework
Dm part03 neural-networks-homeworkokeee
 
Random forest r_packages_slides_02-02-2018_eng
Random forest r_packages_slides_02-02-2018_engRandom forest r_packages_slides_02-02-2018_eng
Random forest r_packages_slides_02-02-2018_engShuma Ishigami
 
Detection focal loss 딥러닝 논문읽기 모임 발표자료
Detection focal loss 딥러닝 논문읽기 모임 발표자료Detection focal loss 딥러닝 논문읽기 모임 발표자료
Detection focal loss 딥러닝 논문읽기 모임 발표자료taeseon ryu
 

What's hot (18)

Introduction to Machine Learning with Python and scikit-learn
Introduction to Machine Learning with Python and scikit-learnIntroduction to Machine Learning with Python and scikit-learn
Introduction to Machine Learning with Python and scikit-learn
 
One Algorithm to Rule Them All: How to Automate Statistical Computation
One Algorithm to Rule Them All: How to Automate Statistical ComputationOne Algorithm to Rule Them All: How to Automate Statistical Computation
One Algorithm to Rule Them All: How to Automate Statistical Computation
 
Ml study notes id3
Ml study notes   id3Ml study notes   id3
Ml study notes id3
 
Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)Tutorial of topological data analysis part 3(Mapper algorithm)
Tutorial of topological data analysis part 3(Mapper algorithm)
 
Seminar_Presentation_ppt
Seminar_Presentation_pptSeminar_Presentation_ppt
Seminar_Presentation_ppt
 
A multilevel automatic thresholding method based on a genetic algorithm for a...
A multilevel automatic thresholding method based on a genetic algorithm for a...A multilevel automatic thresholding method based on a genetic algorithm for a...
A multilevel automatic thresholding method based on a genetic algorithm for a...
 
Support Vector Machine (Classification) - Step by Step
Support Vector Machine (Classification) - Step by StepSupport Vector Machine (Classification) - Step by Step
Support Vector Machine (Classification) - Step by Step
 
Co-clustering of multi-view datasets: a parallelizable approach
Co-clustering of multi-view datasets: a parallelizable approachCo-clustering of multi-view datasets: a parallelizable approach
Co-clustering of multi-view datasets: a parallelizable approach
 
Clustbigfim frequent itemset mining of
Clustbigfim frequent itemset mining ofClustbigfim frequent itemset mining of
Clustbigfim frequent itemset mining of
 
A scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clusteringA scalable collaborative filtering framework based on co clustering
A scalable collaborative filtering framework based on co clustering
 
Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...
Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...
Fractality of Massive Graphs: Scalable Analysis with Sketch-Based Box-Coverin...
 
Improved k-means
Improved k-meansImproved k-means
Improved k-means
 
핵심 딥러닝 입문 4장 RNN
핵심 딥러닝 입문 4장 RNN핵심 딥러닝 입문 4장 RNN
핵심 딥러닝 입문 4장 RNN
 
Dm part03 neural-networks-homework
Dm part03 neural-networks-homeworkDm part03 neural-networks-homework
Dm part03 neural-networks-homework
 
Random forest r_packages_slides_02-02-2018_eng
Random forest r_packages_slides_02-02-2018_engRandom forest r_packages_slides_02-02-2018_eng
Random forest r_packages_slides_02-02-2018_eng
 
Ppt shuai
Ppt shuaiPpt shuai
Ppt shuai
 
Nonnegative Matrix Factorization with Side Information for Time Series Recove...
Nonnegative Matrix Factorization with Side Information for Time Series Recove...Nonnegative Matrix Factorization with Side Information for Time Series Recove...
Nonnegative Matrix Factorization with Side Information for Time Series Recove...
 
Detection focal loss 딥러닝 논문읽기 모임 발표자료
Detection focal loss 딥러닝 논문읽기 모임 발표자료Detection focal loss 딥러닝 논문읽기 모임 발표자료
Detection focal loss 딥러닝 논문읽기 모임 발표자료
 

Similar to DeepLearningProjV3

Enterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using SparkEnterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using SparkAlpine Data
 
Enterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using SparkEnterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using SparkSpark Summit
 
Hand Written Digit Classification
Hand Written Digit ClassificationHand Written Digit Classification
Hand Written Digit Classificationijtsrd
 
House price prediction
House price predictionHouse price prediction
House price predictionSabahBegum
 
Comprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIComprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIijtsrd
 
Power ai tensorflowworkloadtutorial-20171117
Power ai tensorflowworkloadtutorial-20171117Power ai tensorflowworkloadtutorial-20171117
Power ai tensorflowworkloadtutorial-20171117Ganesan Narayanasamy
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
 
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)
AI optimizing HPC simulations (presentation from  6th EULAG Workshop)AI optimizing HPC simulations (presentation from  6th EULAG Workshop)
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)byteLAKE
 
Waste Classification System using Convolutional Neural Networks.pptx
Waste Classification System using Convolutional Neural Networks.pptxWaste Classification System using Convolutional Neural Networks.pptx
Waste Classification System using Convolutional Neural Networks.pptxJohnPrasad14
 
Basic use of xcms
Basic use of xcmsBasic use of xcms
Basic use of xcmsXiuxia Du
 
Audio tagging system using densely connected convolutional networks (DCASE201...
Audio tagging system using densely connected convolutional networks (DCASE201...Audio tagging system using densely connected convolutional networks (DCASE201...
Audio tagging system using densely connected convolutional networks (DCASE201...Hyun-gui Lim
 
Large Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentLarge Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentShaleen Kumar Gupta
 
Block coordinate descent__in_computer_vision
Block coordinate descent__in_computer_visionBlock coordinate descent__in_computer_vision
Block coordinate descent__in_computer_visionYoussefKitane
 
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETS
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETSFAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETS
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETScsandit
 
Parallel Algorithms K – means Clustering
Parallel Algorithms K – means ClusteringParallel Algorithms K – means Clustering
Parallel Algorithms K – means ClusteringAndreina Uzcategui
 
Handwritten Digit Recognition and performance of various modelsation[autosaved]
Handwritten Digit Recognition and performance of various modelsation[autosaved]Handwritten Digit Recognition and performance of various modelsation[autosaved]
Handwritten Digit Recognition and performance of various modelsation[autosaved]SubhradeepMaji
 
Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...Ian Foster
 

Similar to DeepLearningProjV3 (20)

Enterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using SparkEnterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using Spark
 
Enterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using SparkEnterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using Spark
 
Hand Written Digit Classification
Hand Written Digit ClassificationHand Written Digit Classification
Hand Written Digit Classification
 
Mnist soln
Mnist solnMnist soln
Mnist soln
 
Neural networks
Neural networksNeural networks
Neural networks
 
House price prediction
House price predictionHouse price prediction
House price prediction
 
Deep learning
Deep learningDeep learning
Deep learning
 
Comprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPIComprehensive Performance Evaluation on Multiplication of Matrices using MPI
Comprehensive Performance Evaluation on Multiplication of Matrices using MPI
 
Power ai tensorflowworkloadtutorial-20171117
Power ai tensorflowworkloadtutorial-20171117Power ai tensorflowworkloadtutorial-20171117
Power ai tensorflowworkloadtutorial-20171117
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
 
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)
AI optimizing HPC simulations (presentation from  6th EULAG Workshop)AI optimizing HPC simulations (presentation from  6th EULAG Workshop)
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)
 
Waste Classification System using Convolutional Neural Networks.pptx
Waste Classification System using Convolutional Neural Networks.pptxWaste Classification System using Convolutional Neural Networks.pptx
Waste Classification System using Convolutional Neural Networks.pptx
 
Basic use of xcms
Basic use of xcmsBasic use of xcms
Basic use of xcms
 
Audio tagging system using densely connected convolutional networks (DCASE201...
Audio tagging system using densely connected convolutional networks (DCASE201...Audio tagging system using densely connected convolutional networks (DCASE201...
Audio tagging system using densely connected convolutional networks (DCASE201...
 
Large Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentLarge Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate Descent
 
Block coordinate descent__in_computer_vision
Block coordinate descent__in_computer_visionBlock coordinate descent__in_computer_vision
Block coordinate descent__in_computer_vision
 
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETS
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETSFAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETS
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETS
 
Parallel Algorithms K – means Clustering
Parallel Algorithms K – means ClusteringParallel Algorithms K – means Clustering
Parallel Algorithms K – means Clustering
 
Handwritten Digit Recognition and performance of various modelsation[autosaved]
Handwritten Digit Recognition and performance of various modelsation[autosaved]Handwritten Digit Recognition and performance of various modelsation[autosaved]
Handwritten Digit Recognition and performance of various modelsation[autosaved]
 
Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...
 

DeepLearningProjV3

  • 1. DEEP LEARNING FOR FEATURE SELECTION IN CASE AND CONTROL PREDICTION Ana Sanchez Graduate Student Project Day Spring 2015
  • 2. Introduction • Deep learning is a method that uses sets of machine learning algorithms to model data. • Learning representations of data algorithms are used to obtain information that may otherwise not be detected. • In this project, deep learning is used for case and control prediction in genome wide association studies, specifically to construct new feature vectors.
  • 3. Methods • SNP data is used to implement deep learning. Currently experimenting with Upper GI Cancer data. • The data contains approximately 5000 subjects (case and control) and was reduced to 10 case subjects and 10 control subjects. • There are 491,774 SNP features per subject. The SNPs are condensed to patches of 100.
  • 4. Methods - Patching • The patches are generated by starting at index 0 to 99, then shifting by 50 SNPs and obtaining the next 100 SNPs until the final SNP (ex: 1st patch 0- 99, 2nd patch 49-149, 3rd patch 100-199, etc.). • Python program is written to accomplish patching.
  • 5. Python Script for Patching
  • 6. Output of Patching A 196680 X 100 matrix. The matrix represents 9834 patches per patient.
  • 7. Methods – Deep Learning with K-Means • To implement Deep learning, K-Means is used. • The K-Means algorithm clusters data by separating samples in N groups within cluster sum of squares. • Number of clusters must be specified.
  • 8. Methods – Deep Learning with K-Means • For the purpose of this project, Scikit-Learn was used to implement the K-Means algorithm. • Scikit-Learn is open source software that provides libraries for machine learning in Python. • Using imported libraries like numpy, sklearn clusters, and sklearn metrics simple programs are written to handle the complexities of the algorithm. • For K-Means, the number of K clusters were K= 1000 and K=10000.
  • 10. Output for K-Means One file contains a 196680 x 1 matrix. This represents the label for each patch. A second file contains a 1000 x 100 matrix. This is the distance of each data point to the centroid. In other words, the coordinates of the data points to each cluster.
  • 11. Methods – Constructing a Distance Matrix • At this point, three files have been created from the original set, Patches, Centers, and Labels. • The next step is to construct a pairwise distance matrix. This will construct a file that contains the distance of every data point, in each patch, to every center cluster. • Python program is written with patch and center files as input.
  • 12. Python Program for Distance Matrix
  • 13. Output for Distance Matrix A distance matrix 196680 x 1000 results.
  • 14. Output for Distance Matrix Screen Shot of word count for distance matrix output. Shows 196680 rows. When dividing 196680000 by the rows, the columns are obtained. In this case there are 1000 columns. Therefore, a 196680 x 1000 matrix resulted. This matrix represents the number of patches by the number of K centers.
  • 15. Methods – Constructing New Feature Vector file • Now that the Distance Matrix is obtained, the final New Feature Vector file is created. • Since the matrix has196680 x K, the sum of each column for every 9834 rows needs to be obtained. 9834 rows represent 1 patient and should be represented in the final file. • A python program was written that uses the Distance Matrix file as input.
  • 16. Python Program for New Features
  • 17. Output of New Features A New Vector matrix results, 20 x 1000.
  • 18. Output of New Features Screen Shot of word count for new feature vector output. Shows 20 rows. When dividing 20000 by 20, the number of columns is obtained. In this case, the number of columns is 1000. This represents the original 20 patients with their new features, 1000, based on the patching, kmeans, pair wise, and summation methods that were applied.
  • 19. Encountered Issues throughout Project • Large files • Need for extra programs to download • When using Sci Kit, running normal K-means is extremely time consuming. Instead MiniBatchKmeans is used. • MBK makes subsets of input data, greatly reduces computation time. Performs only slightly worse. • Even with MBK, computation time for K=100000 is too long and was not used. • Still, long runtimes( k=1000 (~30 mins), k=10000 (~60mins), k=100000 (>16hours)). • Long runtimes even for simple scripts due to large data.
  • 20. Comparison of Kmeans and MiniBatchKmeans Image and comparison from SciKit-Learn website.
  • 21. Conclusions • The process was completed twice, for K= 1000, and K = 10000. • Using deep learning ideas, a 20 x 491,774 data set was properly processed into a 20x1000 file and a 20x10000 file. • This significant reduction in file size saves memory. • Classification algorithms can implemented on the new files. Prediction Accuracy may be higher for the new files than raw data.
  • 22. References and Acknowledgments • Scikit-learn: Machine Learning in Python, Pedregosa et al., JMLR 12, pp. 2825-2830, 2011. • Michael K. K. Leung, Hui Yuan Xiong, Leo J. Lee, and Brendan J. Frey Deep learning of the tissue-regulated splicing code Bioinformatics 2014 30: i121-i129. • Rasool Fakoor, Faisal Ladhak, Azade Nazi, Manfred Huber Using Deep Learning to enhance cancer diagnosis and classification Proceedings of the 30th International Conference on Ma- chine Learning, Atlanta, Georgia, USA, 2013. JMLR: W&CP volume 28 Special Thanks to Mohammadreza Esfandiari, Payam, for guidance and assistance throughout project.