The document presents a novel bio-computational model called "Sequence Miner" to analyze and classify dengue virus gene sequences. The model uses periodic association rules to identify co-occurrence patterns in dengue gene sequences and visualize classification results through an interactive tool. When tested on over 10,000 dengue virus sequences, the model accurately classified 96.74% of the sequences.
This document discusses the use of bioinformatics tools to analyze gene expression data and detect tumors and mutations in tissues. It summarizes the PhyloMap technique, which integrates principal coordinate analysis, vector quantization, and phylogenetic tree construction to provide improved visualization of large genomic data sets. PhyloMap allows researchers to better analyze and predict evolutionary relationships among influenza A virus genes. The document concludes that PhyloMap is an efficient algorithm for analyzing phylogenetic relationships in large genomic data compared to other techniques.
This document presents a computational method for estimating the population structure of viruses using pyrosequencing reads. The method involves four steps: 1) aligning reads to a reference genome, 2) correcting sequencing errors in the reads, 3) reconstructing haplotypes consistent with the reads, and 4) estimating the frequency of each haplotype in the population. The method is validated on pyrosequencing data from four HIV populations, with over 5000 reads each, by comparing the estimated populations to those obtained from clonal sequencing.
Microfluidics has shown promise for improving diagnostics and biology research by enabling rapid sample processing and precise control of fluids. However, adoption of microfluidic technologies in mainstream biology has not matched initial enthusiasm. While most microfluidic publications remain in engineering journals, their use in fields like cell biology, hematology and immunology is increasing. A case study of chemotaxis assays illustrates how traditional techniques have evolved over time to improve, whereas microfluidic solutions have yet to achieve widespread adoption in biology. Further development is needed to help microfluidics live up to its potential.
Application of Microarray Technology and softcomputing in cancer BiologyCSCJournals
DNA microarray technology has emerged as a boon to the scientific community in understanding the growth and development of life as well as in widening their knowledge in exploring the genetic causes of anomalies occurring in the working of the human body. microarray technology makes biologists be capable of monitoring expression of thousands of genes in a single experiment on a small chip. Extracting useful knowledge and info from these microarray has attracted the attention of many biologists and computer scientists. Knowledge engineering has revolutionalized the way in which the medical data is being looked at. Soft computing is a branch of computer science capable of analyzing complex medical data. Advances in the area of microarray –based expression analysis have led to the promise of cancer diagnosis using new molecular based approaches. Many studies and methodologies have come up which analyszes the gene espression data by using the techniques in data mining such as feature selection, classification, clustering etc. emboiding the soft computing methods for more accuracy. This review is an attempt to look at the recent advances in cancer research with DNA microarray technology , data mining and soft computing techniques.
COVID-19 : Targeting Cells For TreatmentAPRN World
COVID-19 is still affecting many people worldwide. The precautions to be taken are well known to the society by now. Across the globe, the governments are hard at work establishing the physical infrastructure to fight the andemic. At the same time, many laboratories across the world are working on clinical trials evaluating potential treatments. Researchers at MIT, the Ragon Institute of MGH, and Harvard along with colleagues from around the world have identified specific types of cells that appear to be the targets of the coronavirus, which is causing the Covid-19 pandemic.
A Numerical Study of the Spread of Malaria Disease with Self and Cross-Diffus...IOSRJM
: A study of the SIS model of malaria disease with a view to observing the effects of self and crossdiffusion on spatial dynamics is undertaken. Three different cases based on self-diffusion and cross-diffusion are chosen for the investigation. Two cases of cross-diffusion without self-diffusion are also considered in order to see the effects of diffusion on the transmission of malaria. Basic reproductive numbers and bifurcation values are calculated for each case. A series of numerical simulations based on self and cross-diffusion is performed. It is observed that with positive cross-diffusion and self-diffusion in the system, there is a significant increase in the proportion of both infected human and mosquito populations. The proportion of infected humans increases markedly with cross diffusion in the system. This also gives rise to some oscillations across the domain.
This document summarizes an update on DNA barcoding of human pathogenic fungi. It discusses that the ITS region has been proposed as the prime fungal barcode, but that other genetic loci like RPB1 may provide better resolution. It notes challenges with existing databases and the need for quality controlled reference databases. It outlines efforts to establish an international working group and reference database to standardize DNA barcoding for accurate identification of medically important fungi.
The document describes a one-year distance participation program in forensic science offered by the Bioinformatics Institute of India. The program aims to train aspiring and mid-career professionals in various forensic disciplines related to crime scene investigation and evidence analysis. The program covers topics such as forensic medicine, biology, toxicology, serology, criminology, and computer forensics. Completing the program qualifies participants to work as forensic scientists, DNA analysts, and other related roles.
This document discusses the use of bioinformatics tools to analyze gene expression data and detect tumors and mutations in tissues. It summarizes the PhyloMap technique, which integrates principal coordinate analysis, vector quantization, and phylogenetic tree construction to provide improved visualization of large genomic data sets. PhyloMap allows researchers to better analyze and predict evolutionary relationships among influenza A virus genes. The document concludes that PhyloMap is an efficient algorithm for analyzing phylogenetic relationships in large genomic data compared to other techniques.
This document presents a computational method for estimating the population structure of viruses using pyrosequencing reads. The method involves four steps: 1) aligning reads to a reference genome, 2) correcting sequencing errors in the reads, 3) reconstructing haplotypes consistent with the reads, and 4) estimating the frequency of each haplotype in the population. The method is validated on pyrosequencing data from four HIV populations, with over 5000 reads each, by comparing the estimated populations to those obtained from clonal sequencing.
Microfluidics has shown promise for improving diagnostics and biology research by enabling rapid sample processing and precise control of fluids. However, adoption of microfluidic technologies in mainstream biology has not matched initial enthusiasm. While most microfluidic publications remain in engineering journals, their use in fields like cell biology, hematology and immunology is increasing. A case study of chemotaxis assays illustrates how traditional techniques have evolved over time to improve, whereas microfluidic solutions have yet to achieve widespread adoption in biology. Further development is needed to help microfluidics live up to its potential.
Application of Microarray Technology and softcomputing in cancer BiologyCSCJournals
DNA microarray technology has emerged as a boon to the scientific community in understanding the growth and development of life as well as in widening their knowledge in exploring the genetic causes of anomalies occurring in the working of the human body. microarray technology makes biologists be capable of monitoring expression of thousands of genes in a single experiment on a small chip. Extracting useful knowledge and info from these microarray has attracted the attention of many biologists and computer scientists. Knowledge engineering has revolutionalized the way in which the medical data is being looked at. Soft computing is a branch of computer science capable of analyzing complex medical data. Advances in the area of microarray –based expression analysis have led to the promise of cancer diagnosis using new molecular based approaches. Many studies and methodologies have come up which analyszes the gene espression data by using the techniques in data mining such as feature selection, classification, clustering etc. emboiding the soft computing methods for more accuracy. This review is an attempt to look at the recent advances in cancer research with DNA microarray technology , data mining and soft computing techniques.
COVID-19 : Targeting Cells For TreatmentAPRN World
COVID-19 is still affecting many people worldwide. The precautions to be taken are well known to the society by now. Across the globe, the governments are hard at work establishing the physical infrastructure to fight the andemic. At the same time, many laboratories across the world are working on clinical trials evaluating potential treatments. Researchers at MIT, the Ragon Institute of MGH, and Harvard along with colleagues from around the world have identified specific types of cells that appear to be the targets of the coronavirus, which is causing the Covid-19 pandemic.
A Numerical Study of the Spread of Malaria Disease with Self and Cross-Diffus...IOSRJM
: A study of the SIS model of malaria disease with a view to observing the effects of self and crossdiffusion on spatial dynamics is undertaken. Three different cases based on self-diffusion and cross-diffusion are chosen for the investigation. Two cases of cross-diffusion without self-diffusion are also considered in order to see the effects of diffusion on the transmission of malaria. Basic reproductive numbers and bifurcation values are calculated for each case. A series of numerical simulations based on self and cross-diffusion is performed. It is observed that with positive cross-diffusion and self-diffusion in the system, there is a significant increase in the proportion of both infected human and mosquito populations. The proportion of infected humans increases markedly with cross diffusion in the system. This also gives rise to some oscillations across the domain.
This document summarizes an update on DNA barcoding of human pathogenic fungi. It discusses that the ITS region has been proposed as the prime fungal barcode, but that other genetic loci like RPB1 may provide better resolution. It notes challenges with existing databases and the need for quality controlled reference databases. It outlines efforts to establish an international working group and reference database to standardize DNA barcoding for accurate identification of medically important fungi.
The document describes a one-year distance participation program in forensic science offered by the Bioinformatics Institute of India. The program aims to train aspiring and mid-career professionals in various forensic disciplines related to crime scene investigation and evidence analysis. The program covers topics such as forensic medicine, biology, toxicology, serology, criminology, and computer forensics. Completing the program qualifies participants to work as forensic scientists, DNA analysts, and other related roles.
Uses of Artificial Intelligence in BioinformaticsPragya Pai
This presentation is about the usage of Artificial Intelligence in Bioinformatics. These slides give the basic knowledge about usage of Artificial Intelligence in Bioinformatics.
This document presents a mathematical model of the spread of two viral subtypes on a plant leaf. It describes a master's thesis that models the interaction between a wild viral strain and a mutant strain on a two-dimensional representation of a leaf. The model consists of two coupled reaction-diffusion equations, one for each strain, which include suppression of the other strain's reproduction. Numerical simulation of the model is used to test the hypothesis that co-infection of a cell by both subtypes is rare, as observed experimentally.
This document provides a summary of Peter WeiDer Li's background and experience. He has over 20 years of experience in informatics, software development, and scientific work. He has conceived and implemented various informatics solutions and databases across multiple roles at institutions including Mayo Clinic, Johns Hopkins University, and Celera Genomics. He has a broad range of technical skills and domain expertise in bioinformatics, genomics, and clinical informatics.
This chapter discusses molecular epidemiology and evolutionary genetics of infectious diseases. It argues that molecular epidemiology should incorporate concepts from population genetics and evolutionary biology. Characterizing pathogens should evaluate how genetic diversity impacts medical properties. Emerging technologies like massive sequencing, post-genomic studies, and bioinformatics are transforming this field. The chapter examines definitions of molecular epidemiology and how new technologies like PCR have revolutionized pathogen identification but have limitations. It also discusses how molecular epidemiology has helped clinical practice in some routine ways but is primarily a research tool rather than part of daily medical practice.
This candidate has 12 years of experience in drug discovery, primarily focused on oncology therapeutic targets involving cell biology assays and general laboratory skills. They currently serve as a lead scientist and lab group head presenting data to project teams. They are seeking a senior scientist role and have experience in areas like cell culture, proliferation assays, microscopy, and data analysis software. They have worked at companies like Novartis and Piramed Pharma.
Bioinformatics is defined as the field that studies biology using computers and information technology. It involves the collection, storage, and analysis of molecular biological data using techniques from computer science and statistics. Some key events in bioinformatics include Watson and Crick proposing the DNA double helix structure in 1953, and the development of sequence alignment and structure prediction algorithms in the 1970s. Bioinformatics aims to better understand living cells at the molecular level by analyzing raw molecular sequence and structure data. It provides globally accessible databases and analysis tools to enable sharing and study of biological data.
This document introduces bioinformatics and discusses some of its key concepts and applications. It defines bioinformatics as an interdisciplinary field that combines computer science, statistics and engineering to study and process biological data. It describes some basic cell components like DNA, RNA and proteins, and how genetics and the genetic code work. It also provides a brief history of bioinformatics, highlighting projects like the Human Genome Project. Finally, it outlines several applications of bioinformatics like phylogenetic analysis, drug design, microarray analysis and protein-protein interaction networks.
Nanoparticles as a novel and promising antiviral platform in veterinary medicineAhmed Hasham
Traditional veterinary virus vaccines, such as inactivated and live-attenuated vaccines, have achieved tremendous success
in controlling many viral diseases of livestock and chickens worldwide. However, many recent viral outbreaks caused by
diferent emerging and re-emerging viruses continue to be reported annually worldwide. It is therefore necessary to develop
new control regimens. Nanoparticle research has received considerable attention in the last two decades as a promising
platform with signifcant success in veterinary medicine, replacing traditional viral vector vaccines. However, the feld of
nanoparticle applications is still in its initial phase of growth. Here, we discuss various preparation methods, characteristics,
physical properties, antiviral efects, and pharmacokinetics of well-developed nanoparticles and the potential of nanoparticles
or nano-vaccines as a promising antiviral platform for veterinary medicine.
Brief presentation on the challenges and current state of play with regards to the bioinformatics of a pathogen, M. tuberculosis. Presented at the UWC/UCT Big Data workshop in January 2015
Determination of the Prevalence of Bluetongue Disease in Goats in Siirt Provi...AI Publications
Bluetongue disease is a disease caused by the bluetongue virus in the Orbivirus genus of the Reoviridae family. The disease can be transmitted by ticks, sperm and vertically but is mainly spread by the Culicoides type flies. While the disease is naturally observed in cattle and sheep, it is more rarely observed in goats. This study was carried out to determine the presence of bluetongue infection in goats raised in Siirt province serologically and to obtain information about its prevalence in the region. The material of the study consisted of a total of 372 goats over one year of age raised in Siirt province. For laboratory analyses, 5 ml blood samples were taken from the jugular vein of goats to tubes without anticoagulant. The samples were centrifuged at 3000 rpm for 10 minutes, and the sera were transferred to Eppendorf tubes and stored at -20 °C until analysis. The serum samples were analyzed by the ELISA (Thermo Scientific Multiskan Go) method using the commercial test kit (Bluetongue Virus VP7 Ab Test Kit, IDEXX). The results were evaluated as positive and negative. Of the 372 goats constituting the material of the study, 252 (67.74%) were seropositive and 120 (32.26%) were seronegative. As a result, in addition to measures for quarantine and vector control in the fight against the disease, it was concluded that vaccines specific to virus serotypes should be used and comprehensive and planned studies should be carried out in relation to the disease in the region and the country.
Bioinformatics can be applied to climate smart horticulture in several ways:
1) It allows for crop improvement through comparative genomics between crop plants and model species to identify important genes.
2) It facilitates plant breeding by providing tools for genome analysis, marker identification, and rational gene annotation.
3) Stress-tolerant varieties can be developed by using bioinformatics databases like KEGG to identify pathways and genes involved in drought resistance.
This document contains information about the International Conference and Exhibition on Nanomedicine and Drug Delivery that will take place from May 29-31, 2017 in Osaka, Japan. It includes the tentative program agenda, speaker biographies, organizing committee members, and registration information. The conference will feature keynote lectures, interactive sessions, workshops, and tracks on topics related to nanomedicine and drug delivery over the three-day event.
Bioinformatics involves the application of computer technology to manage biological information. Computers are used to gather, store, analyze, and integrate biological and genetic data, which can then be applied to areas like drug discovery. The need for bioinformatics arose from the large amount of genomic data generated by the Human Genome Project. It combines molecular biology and computer science to understand diseases and find new drug targets. Many universities, government agencies, and pharmaceutical companies have formed bioinformatics groups with computational biologists and computer scientists.
Free webinar-introduction to bioinformatics - biologist-1Elia Brodsky
The Omics Logic Introduction to Bioinformatics program is a one-month online training program that provides an introduction to the field of bioinformatics for beginners. The program consists of six sessions taught by an international team of experts, covering topics like genomics, transcriptomics, statistical analysis, machine learning, and a final bioinformatics project. Participants will learn data analysis skills in Python and R and how to extract insights from multi-omics datasets with applications in biomedicine. The goal is to prepare students for data-driven research in life sciences through interactive lessons, coding exercises, and independent projects.
This document provides an overview and syllabus for a course on bioinformatics. It discusses the goals of learning about available bioinformatics programs and tools, and interpreting their outputs. The course will cover topics like sequence alignment, phylogenetics, genome comparison and using databases. Assessment will include homework, exams, a report, and participation. The document contrasts the "old" and "new" biology, noting how the new biology generates large datasets that require computational analysis to make sense of the data. It emphasizes that bioinformatics uses algorithms and databases to organize, analyze and interpret biological data at large scales.
Introduction
Definition
History
Principle
Components of bioinformatics
Bioinformatics databases
Tools of bioinformatics
Applications of bioinformatics
Molecular medicine
Microbial genomics
Plant genomics
Animal genomics
Human genomics
Drug and vaccine designing
Proteomics
For studying biomolecular structures
In- silico testing
Conclusion
References
Presentation about how much bioinformatics involved in the medical field. This was presented at the University of Colombo in 2007 for an undergraduate seminar
This thesis examines genetic variation in early maize endosperm development, specifically endosperm cell number and endoreduplication level, across three mapping populations. Natural genetic variability was observed for both traits. Heritability estimates for endosperm cell number and mean ploidy level ranged from 0.23 to 0.43. A negative phenotypic correlation was found between the two traits. A composite trait of mean total chromatin content showed the strongest correlations with kernel weight. Quantitative trait loci (QTL) were identified for both endosperm cell number and mean ploidy level using composite interval mapping. Understanding the genetics controlling early endosperm development could aid in improving seed quality and yield.
Ceph Day Melbourne - Scale and performance: Servicing the Fabric and the Work...Ceph Community
The document discusses scale and performance challenges in providing storage infrastructure for research computing. It describes Monash University's implementation of the Ceph distributed storage system across multiple clusters to provide a "fabric" for researchers' storage needs in a flexible, scalable way. Key points include:
- Ceph provides software-defined storage that is scalable and can integrate with other systems like OpenStack.
- Multiple Ceph clusters have been implemented at Monash of varying sizes and purposes, including dedicated clusters for research data storage.
- The infrastructure provides different "tiers" of storage with varying performance and cost characteristics to meet different research needs.
- Ongoing work involves expanding capacity and upgrading hardware to improve performance
Uses of Artificial Intelligence in BioinformaticsPragya Pai
This presentation is about the usage of Artificial Intelligence in Bioinformatics. These slides give the basic knowledge about usage of Artificial Intelligence in Bioinformatics.
This document presents a mathematical model of the spread of two viral subtypes on a plant leaf. It describes a master's thesis that models the interaction between a wild viral strain and a mutant strain on a two-dimensional representation of a leaf. The model consists of two coupled reaction-diffusion equations, one for each strain, which include suppression of the other strain's reproduction. Numerical simulation of the model is used to test the hypothesis that co-infection of a cell by both subtypes is rare, as observed experimentally.
This document provides a summary of Peter WeiDer Li's background and experience. He has over 20 years of experience in informatics, software development, and scientific work. He has conceived and implemented various informatics solutions and databases across multiple roles at institutions including Mayo Clinic, Johns Hopkins University, and Celera Genomics. He has a broad range of technical skills and domain expertise in bioinformatics, genomics, and clinical informatics.
This chapter discusses molecular epidemiology and evolutionary genetics of infectious diseases. It argues that molecular epidemiology should incorporate concepts from population genetics and evolutionary biology. Characterizing pathogens should evaluate how genetic diversity impacts medical properties. Emerging technologies like massive sequencing, post-genomic studies, and bioinformatics are transforming this field. The chapter examines definitions of molecular epidemiology and how new technologies like PCR have revolutionized pathogen identification but have limitations. It also discusses how molecular epidemiology has helped clinical practice in some routine ways but is primarily a research tool rather than part of daily medical practice.
This candidate has 12 years of experience in drug discovery, primarily focused on oncology therapeutic targets involving cell biology assays and general laboratory skills. They currently serve as a lead scientist and lab group head presenting data to project teams. They are seeking a senior scientist role and have experience in areas like cell culture, proliferation assays, microscopy, and data analysis software. They have worked at companies like Novartis and Piramed Pharma.
Bioinformatics is defined as the field that studies biology using computers and information technology. It involves the collection, storage, and analysis of molecular biological data using techniques from computer science and statistics. Some key events in bioinformatics include Watson and Crick proposing the DNA double helix structure in 1953, and the development of sequence alignment and structure prediction algorithms in the 1970s. Bioinformatics aims to better understand living cells at the molecular level by analyzing raw molecular sequence and structure data. It provides globally accessible databases and analysis tools to enable sharing and study of biological data.
This document introduces bioinformatics and discusses some of its key concepts and applications. It defines bioinformatics as an interdisciplinary field that combines computer science, statistics and engineering to study and process biological data. It describes some basic cell components like DNA, RNA and proteins, and how genetics and the genetic code work. It also provides a brief history of bioinformatics, highlighting projects like the Human Genome Project. Finally, it outlines several applications of bioinformatics like phylogenetic analysis, drug design, microarray analysis and protein-protein interaction networks.
Nanoparticles as a novel and promising antiviral platform in veterinary medicineAhmed Hasham
Traditional veterinary virus vaccines, such as inactivated and live-attenuated vaccines, have achieved tremendous success
in controlling many viral diseases of livestock and chickens worldwide. However, many recent viral outbreaks caused by
diferent emerging and re-emerging viruses continue to be reported annually worldwide. It is therefore necessary to develop
new control regimens. Nanoparticle research has received considerable attention in the last two decades as a promising
platform with signifcant success in veterinary medicine, replacing traditional viral vector vaccines. However, the feld of
nanoparticle applications is still in its initial phase of growth. Here, we discuss various preparation methods, characteristics,
physical properties, antiviral efects, and pharmacokinetics of well-developed nanoparticles and the potential of nanoparticles
or nano-vaccines as a promising antiviral platform for veterinary medicine.
Brief presentation on the challenges and current state of play with regards to the bioinformatics of a pathogen, M. tuberculosis. Presented at the UWC/UCT Big Data workshop in January 2015
Determination of the Prevalence of Bluetongue Disease in Goats in Siirt Provi...AI Publications
Bluetongue disease is a disease caused by the bluetongue virus in the Orbivirus genus of the Reoviridae family. The disease can be transmitted by ticks, sperm and vertically but is mainly spread by the Culicoides type flies. While the disease is naturally observed in cattle and sheep, it is more rarely observed in goats. This study was carried out to determine the presence of bluetongue infection in goats raised in Siirt province serologically and to obtain information about its prevalence in the region. The material of the study consisted of a total of 372 goats over one year of age raised in Siirt province. For laboratory analyses, 5 ml blood samples were taken from the jugular vein of goats to tubes without anticoagulant. The samples were centrifuged at 3000 rpm for 10 minutes, and the sera were transferred to Eppendorf tubes and stored at -20 °C until analysis. The serum samples were analyzed by the ELISA (Thermo Scientific Multiskan Go) method using the commercial test kit (Bluetongue Virus VP7 Ab Test Kit, IDEXX). The results were evaluated as positive and negative. Of the 372 goats constituting the material of the study, 252 (67.74%) were seropositive and 120 (32.26%) were seronegative. As a result, in addition to measures for quarantine and vector control in the fight against the disease, it was concluded that vaccines specific to virus serotypes should be used and comprehensive and planned studies should be carried out in relation to the disease in the region and the country.
Bioinformatics can be applied to climate smart horticulture in several ways:
1) It allows for crop improvement through comparative genomics between crop plants and model species to identify important genes.
2) It facilitates plant breeding by providing tools for genome analysis, marker identification, and rational gene annotation.
3) Stress-tolerant varieties can be developed by using bioinformatics databases like KEGG to identify pathways and genes involved in drought resistance.
This document contains information about the International Conference and Exhibition on Nanomedicine and Drug Delivery that will take place from May 29-31, 2017 in Osaka, Japan. It includes the tentative program agenda, speaker biographies, organizing committee members, and registration information. The conference will feature keynote lectures, interactive sessions, workshops, and tracks on topics related to nanomedicine and drug delivery over the three-day event.
Bioinformatics involves the application of computer technology to manage biological information. Computers are used to gather, store, analyze, and integrate biological and genetic data, which can then be applied to areas like drug discovery. The need for bioinformatics arose from the large amount of genomic data generated by the Human Genome Project. It combines molecular biology and computer science to understand diseases and find new drug targets. Many universities, government agencies, and pharmaceutical companies have formed bioinformatics groups with computational biologists and computer scientists.
Free webinar-introduction to bioinformatics - biologist-1Elia Brodsky
The Omics Logic Introduction to Bioinformatics program is a one-month online training program that provides an introduction to the field of bioinformatics for beginners. The program consists of six sessions taught by an international team of experts, covering topics like genomics, transcriptomics, statistical analysis, machine learning, and a final bioinformatics project. Participants will learn data analysis skills in Python and R and how to extract insights from multi-omics datasets with applications in biomedicine. The goal is to prepare students for data-driven research in life sciences through interactive lessons, coding exercises, and independent projects.
This document provides an overview and syllabus for a course on bioinformatics. It discusses the goals of learning about available bioinformatics programs and tools, and interpreting their outputs. The course will cover topics like sequence alignment, phylogenetics, genome comparison and using databases. Assessment will include homework, exams, a report, and participation. The document contrasts the "old" and "new" biology, noting how the new biology generates large datasets that require computational analysis to make sense of the data. It emphasizes that bioinformatics uses algorithms and databases to organize, analyze and interpret biological data at large scales.
Introduction
Definition
History
Principle
Components of bioinformatics
Bioinformatics databases
Tools of bioinformatics
Applications of bioinformatics
Molecular medicine
Microbial genomics
Plant genomics
Animal genomics
Human genomics
Drug and vaccine designing
Proteomics
For studying biomolecular structures
In- silico testing
Conclusion
References
Presentation about how much bioinformatics involved in the medical field. This was presented at the University of Colombo in 2007 for an undergraduate seminar
This thesis examines genetic variation in early maize endosperm development, specifically endosperm cell number and endoreduplication level, across three mapping populations. Natural genetic variability was observed for both traits. Heritability estimates for endosperm cell number and mean ploidy level ranged from 0.23 to 0.43. A negative phenotypic correlation was found between the two traits. A composite trait of mean total chromatin content showed the strongest correlations with kernel weight. Quantitative trait loci (QTL) were identified for both endosperm cell number and mean ploidy level using composite interval mapping. Understanding the genetics controlling early endosperm development could aid in improving seed quality and yield.
Ceph Day Melbourne - Scale and performance: Servicing the Fabric and the Work...Ceph Community
The document discusses scale and performance challenges in providing storage infrastructure for research computing. It describes Monash University's implementation of the Ceph distributed storage system across multiple clusters to provide a "fabric" for researchers' storage needs in a flexible, scalable way. Key points include:
- Ceph provides software-defined storage that is scalable and can integrate with other systems like OpenStack.
- Multiple Ceph clusters have been implemented at Monash of varying sizes and purposes, including dedicated clusters for research data storage.
- The infrastructure provides different "tiers" of storage with varying performance and cost characteristics to meet different research needs.
- Ongoing work involves expanding capacity and upgrading hardware to improve performance
Modul ini membahas konsep sub program dalam bahasa Pascal yang terdiri dari prosedur dan fungsi. Prosedur adalah sub program yang berfungsi untuk memecahkan masalah menjadi modul-modul yang lebih kecil. Prosedur dapat menerima parameter baik secara nilai maupun acuan, yang memengaruhi perubahan nilai variabel antara prosedur dan program utama.
This document describes using a two-layer fuzzy logic controller with a unified power flow controller to dampen tie-line power oscillations in an interconnected thermal power system. Conventionally, automatic generation control is performed using primary governor control and a secondary proportional-integral controller. However, this can still result in frequency and tie-line power transients during load disturbances. The document proposes using a UPFC along with a two-layer fuzzy logic controller to further improve damping. Simulation results indicate this approach reduces transients and achieves faster settling times.
From Needs Analysis to Language Center: CALL for Change at Osaka UniversityParisa Mehran
A comprehensive needs analysis was conducted to assess the English language needs and difficulties of undergraduate Japanese EFL learners at Osaka University. The findings revealed that there is a pressing need for launching a CALL-based language center at this university which could contribute to the status of ELT.
This document discusses the practice of dynamic psychiatry in a small community. It is an article from 2005 written by Edward R. Shapiro from the Austen Riggs Center. The article has been read 8 times and Shapiro has 51 publications and 184 citations according to his ResearchGate profile.
Alex King, a student, is approached by a stranger who offers him a pill, telling him it will make his day amazing. Alex takes the pill but soon starts experiencing strange effects like losing time and having conversations he doesn't remember. He collapses at school and is taken to the hospital. There, he has flashbacks of taking the pill and confronts its dealer. The movie is about the negative effects of Alex taking the mysterious pill and how it sends his life into a tailspin.
The document discusses the past continuous passive voice. It provides the structure using "was/were being" or "was/were getting" followed by the past participle. Examples are given for each person and exercises provided for changing sentences from active to passive voice. Key points covered include questioning structures in the passive voice and changing negative sentences into the passive.
The document provides details of a lesson plan for an 8th standard physics class on the topic of mass. It includes the objectives to enable students to develop factual, conceptual, and procedural knowledge of mass through various activities and examples. The plan outlines pre-requisites, teaching-learning resources, introduction, discussion points, activities, review questions and a follow up activity to assess students' understanding of mass and its relationship to the amount of matter.
Dokumen ini berisi proyeksi dan cita-cita siswa kelas XII IPA1 SMA Negeri 1 Wonosari untuk lulusan tahun 2007 yang disusun oleh Sriyanta. Dokumen ini hanya menyebutkan judul proyeksi dan cita-cita siswa tanpa memberikan informasi lebih lanjut.
Dry-wit is a bash framework useful when writing new scripts. It provides a common layout and a toolbox of functions to help writing shell scripts from scratch.
Genetic algorithm to optimization mobility-based dengue mathematical modelIJECEIAES
Implementation of vaccines, mosquito repellents and several Wolbachia schemes have been proposed recently as strategies against dengue. Research showed that the use of vaccine and repellent is highly effective when implemented to individuals who are in area with high transmission rates, while the use of Wolbachia bacteria is strongly effective when implemented in area with low transmission rates. This research is to show a three-strategy combination to cope with the dengue using mathematical model. In dengue mathematical model construction, several parameters are not yet known, therefore a genetic algorithm method was used to estimate dengue model parameters. Numerical simulation results showed that the combination of three strategies were able to reduce the number of infected humans. The dynamic of the human population with the combination of three strategies on average was able to reduce the infected human population by 45.2% in immobility aspect. Furthermore, the mobility aspect in dengue model was presented by reviewing two areas; Yogyakarta and Semarang in Indonesia. The numerical solutions showed that the trend graph was almost similar between the two areas. With the maximum effort given, the combination control values decreased slowly until the 100th day.
This document reviews factors that can be used to predict dengue fever outbreaks using data mining techniques. It summarizes 10 research papers that used various data mining methods like k-means clustering, neural networks, and decision trees to predict dengue fever based on factors like temperature, rainfall, sunshine, and El Nino oscillation. The studies achieved prediction accuracies ranging from 71-96% and identified temperature, rainfall, and weather data as important influencing factors that are correlated with increased dengue fever incidence. The review concludes that data mining techniques can effectively use environmental and weather data to predict outbreaks of the mosquito-borne disease.
IRJET- Development of a Computer-Aided System for Swine-Flu Prediction using ...IRJET Journal
This document describes the development of a computer-aided system for predicting swine flu using computational intelligence techniques. Swine flu, also known as H1N1, is a contagious respiratory disease that affects pigs and can be transmitted to humans. The system aims to provide accurate diagnosis and prediction of swine flu at low cost using data mining techniques like support vector machines, random forests, naive Bayes classifiers, and k-means clustering. These techniques are compared based on metrics like accuracy, error rate, and execution time to determine the most effective approach for diagnostic and predictive modeling of swine flu infections.
An agent-based model to assess coronavirus disease 19 spread and health syst...IJECEIAES
The present pandemic has tremendously raised the health systems’ burden around the globe. It is important to understand the transmission dynamics of the infection and impose localized strategies across different geographies to curtail the spread of the infection. The present study was designed to assess the transmission dynamics and the health systems’ burden of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) using an agent-based modeling (ABM) approach. The study used a synthetic population with 31,738,240 agents representing 90.67 percent of the overall population of Telangana, India. The effects of imposing and lifting lockdowns, nonpharmaceutical interventions, and the role of immunity were analyzed. The distribution of people in different health states was measured separately for each district of Telangana. The spread dramatically increased and reached a peak soon after the lockdowns were relaxed. It was evident that is the protection offered is higher when a higher proportion of the population is exposed to the interventions. ABMs help to analyze grassroots details compared to compartmental models. Risk estimates provide insights on the proportion of the population protected by the adoption of one or more of the control measures, which is of practical significance for policymaking.
A CLOUD-BASED PROTOTYPE IMPLEMENTATION OF A DISEASE OUTBREAK NOTIFICATION SYS...IJCSEA Journal
This paper describes the design, prototype implementation and performance characteristics of a Disease Outbreak Notification System (DONS). The prototype was implemented in a hybrid cloud environment as an online/real-time system. It detects potential outbreaks of both listed and unknown diseases. It uses data mining techniques to choose the correct algorithm to detect outbreaks of unknown diseases. Our experiments showed that the proposed system has very high accuracy rate in choosing the correct detection algorithm. To our best knowledge, DONS is the first of its kind to detect outbreaks of unknown diseases using data mining techniques.
This document summarizes a research paper that aims to predict dengue cases in India using machine learning models. It finds that recurrent neural networks best model the seasonal variations in dengue data. The paper analyzes climate data from India and finds higher dengue cases correlate with increased urbanization and decreased vegetation. Time series models like SARIMA outperform ARIMA at forecasting dengue due to its seasonal components. Overall, recurrent neural networks and SARIMA models most accurately predict future dengue outbreaks in India based on historical climate and case data. This will help authorities prepare mitigation strategies to reduce dengue pandemics.
A Cloud-Based Prototype Implementation of a Disease Outbreak Notification SystemIJCSEA Journal
This paper describes the design, prototype implementation and performance characteristics of a Disease Outbreak Notification System (DONS). The prototype was implemented in a hybrid cloud environment as an online/real-time system. It detects potential outbreaks of both listed and unknown diseases. It uses data mining techniques to choose the correct algorithm to detect outbreaks of unknown diseases. Our experiments showed that the proposed system has very high accuracy rate in choosing the correct detection algorithm. To our best knowledge, DONS is the first of its kind to detect outbreaks of unknown diseases using data mining techniques.
Estimating the Statistical Significance of Classifiers used in the Predictio...IOSR Journals
This document summarizes a research paper that analyzes the statistical significance of different classifiers for predicting tuberculosis. The paper first compares the accuracy of classifiers like decision trees, support vector machines, k-nearest neighbor, and naive Bayes on tuberculosis data. It then evaluates the performance of these classifiers using a paired t-test to select the optimal model. The results showed that support vector machines and decision trees were not statistically significant, while support vector machines combined with naive Bayes and k-nearest neighbor were statistically significant.
RECOMMENDER SYSTEM FOR DETECTION OF DENGUE USING FUZZY LOGICIAEME Publication
The recommender System involved in health care is important since user can detect whether he has problem or not. A user will get whole information on the go. Today user doesn’t have much time and information about the dengue and it will be disclosed to the user at later stages. The dengue is deadly disease so its information should be disclosed at earlier stage. The proposed system works toward this aspect. The set of parameters including fever, TLC, blood pressure, severe headache etc. are analysed in proposed system. The filtering mechanism is also utilised in the proposed system which is integral part of recommender system. The content based filtering will be utilised in proposed system.
Epidemic Alert System: A Web-based Grassroots ModelIJECEIAES
This document summarizes research on web-based epidemic alert systems. It discusses how most current systems analyze large amounts of unstructured data from various online sources using complex algorithms, which can generate imprecise results given the lack of standards. The document then proposes a new grassroots web-based system that collects structured data directly from primary health centers, hospitals, and laboratories. This traditional approach uses threshold values based on percentiles to determine when an epidemic is triggered. If adopted, it could help standardize web-based disease surveillance.
This document summarizes a study that investigated the potential of using an individual's microbial fingerprint to link them to items they have touched. Fifteen individuals provided samples from their fingertips, personal laptops, chairs, a shared office doorknob, and photocopier. Bacterial DNA was extracted and 16S rRNA gene PCR and sequencing was used to analyze and compare the bacterial communities between samples. The results showed higher similarity between individuals' fingerprints and personal items than shared surfaces. Some individuals like regular users of the doorknob showed closer links between their fingerprints and the doorknob. Therefore, an individual's unique bacterial profile on their skin and possessions could potentially be used as a forensic tool to associate them with touched objects.
Disease Network is the science that has emerged to diagnose a disease from a network aspect
specifically. Networks are the group that interconnect to each others similarly disease networks are
the one that reveal concelled connection among apparently independent biomedical entities like
physiologic process, signaling receptors, in addition to genetic code, also they prove to exists
intitutive in addition to powerful way to learn/discover or diagnose a disease.Due to these networks,
we can now consume the elderly drugs and its method to learn/discover the new drug
accordingly.Example- Colchicine is used in gout but after repurposing it is also used in mediterranean
fever. This is because there are many factors that affect the body during mediterranean fever and
gout, we know that gout is a form of arthritis that causes pain in joints also mediterranean fever is the
one which is accompanied by pain in joints, therefore colchicine is used as a repurposed drug again.In
repurposing of medicines or drugs we first analyse the change in symptoms and identify the target
organ and accorgingly we produce a drug that is compatible with pharmacokinetics of the body. As
the availablity of transcriptomic,proteomic and metabolomic data sources are increasing day by day it helps in classification of disease .Also there are some networks reffered to as complex networks which can be called as collection of linked junctions/ nodes
Machine learning approaches in the diagnosis of infectious diseases-a review.pdfSmriti Mishra
This document reviews machine learning approaches for diagnosing infectious diseases. It discusses how machine learning algorithms like logistic regression, K-nearest neighbors, support vector machines, decision trees, naive Bayes, neural networks, and ensemble techniques have been applied to diagnose diseases like tuberculosis, influenza, HIV, dengue fever, COVID-19, cystitis, and nonspecific urethritis. Existing models have limitations related to data quality and availability. Ensemble approaches that combine multiple models may improve performance over single classifiers. Developing models requires diverse, high-quality data that accurately represents reality.
Binding site identification of COVID-19 main protease 3D structure by homolo...nooriasukmaningtyas
The influx of coronavirus in 2019 (COVID-19) has recorded millions of infection cases with several deaths worldwide. There is no effective treatment, but recent studies have shown that its enzymes maybe considered as potential drug target. The purpose of this work was to identify the binding site in-silico and present the 3D structure of COVID-19 main-protease (Mpro) by homology modeling through multiple alignment followed by optimization and validation. The modeling was done by Swiss-Model template library. The obtained homotrimer oligo-state model was verified for reliability using PROCHECK, Verify3D, MolProbity and QMEAN. HHBlits software was used to determine structures that matched the target sequence by evolution. Structure quality verification through Ramachandran plot showed an abundance of 99.3% of amino acid residues in allowed regions while 0.1% in disallowed region. The Verify3D rated the structure a 90.87% PASS of residues having an average 3D-1D score of at least 0.2, which validates a good environment profile for the Mpro model. The features of the secondary structure indicated that the structure contains 32.05% α-helix and 37.17% random coil with 25.92 extended strand. The result of this study suggests that blocking expression of this protein may constitute an efficient approach for infection transmission blockage.
Root cause analysis of COVID-19 cases by enhanced text mining processIJECEIAES
The main focus of this research is to find the reasons behind the fresh cases of COVID-19 from the public’s perception for data specific to India. The analysis is done using machine learning approaches and validating the inferences with medical professionals. The data processing and analysis is accomplished in three steps. First, the dimensionality of the vector space model (VSM) is reduced with improvised feature engineering (FE) process by using a weighted term frequency-inverse document frequency (TF-IDF) and forward scan trigrams (FST) followed by removal of weak features using feature hashing technique. In the second step, an enhanced K-means clustering algorithm is used for grouping, based on the public posts from Twitter®. In the last step, latent dirichlet allocation (LDA) is applied for discovering the trigram topics relevant to the reasons behind the increase of fresh COVID-19 cases. The enhanced K-means clustering improved Dunn index value by 18.11% when compared with the traditional K-means method. By incorporating improvised two-step FE process, LDA model improved by 14% in terms of coherence score and by 19% and 15% when compared with latent semantic analysis (LSA) and hierarchical dirichlet process (HDP) respectively thereby resulting in 14 root causes for spike in the disease.
ciclo autonomico-short paper - Witfor 2016 paper_42.. ..
This paper presents an ongoing project to develop a biocomputational platform to analyze genomic data from cancer patients and bacteria in Costa Rica. The platform will integrate genomic data processing, prediction of drug sensitivity, and identification of new therapeutic targets. It will use pattern recognition techniques and mathematical models on genomic and drug response data to predict personalized therapy. Preliminary results include databases to store cancer and bacteria genomic data, and tools for exploring relationships between genomic features and drug responses. The platform aims to help identify optimal personalized treatments to overcome drug resistance in cancer and bacterial infections.
potassium, chloride, bicarbonate, blood urea nitrogen (BUN), magnesium, creatinine, glucose, and sometimes calcium. Tests that focus on cholesterol levels can determine LDL and HDL cholesterol levels, as well as triglyceride levels.[6]
Big data approaches to healthcare systemsShubham Jain
The idea behind this presentation is to explore how big data will revolutionize existing healthcare system effectively by reducing healthcare concerns such as the selection of appropriate treatment paths, quality of healthcare systems and so on. Large amount of unstructured data is available in various organizations (payers, providers, pharmaceuticals). We will discuss all the intricacies involved in massive datasets of healthcare systems and how combination of VPH technologies and big data resulted into some mind-boggling consequences. Major opportunities in healthcare includes the integration of various data pools such as clinical data, pharmaceutical R&D data and patient behaviour and sentiment data. Finding potential insights from big data with the help of medical image processing techniques, predictive modelling etc. will eventually help us to leverage the ever-increasing costs of care, help providers practice more effective medicine, empower patients and caregivers, support fitness and preventive self-care, and to dream about more personalized medicine.
Semantic Similarity Measures between Terms in the Biomedical Domain within f...Editor IJCATR
The techniques and tests are tools used to define how measure the goodness of ontology or its resources. The similarity between biomedical classes/concepts is an important task for the biomedical information extraction and knowledge discovery. However, most of the semantic similarity techniques can be adopted to be used in the biomedical domain (UMLS). Many experiments have been conducted to check the applicability of these measures. In this paper, we investigate to measure semantic similarity between two terms within single ontology or multiple ontologies in ICD-10 “V1.0” as primary source, and compare my results to human experts score by correlation coefficient.
ABSTRACT
Genome-wide transcription profiling is a powerful technique in studying disease susceptible footprints. Moreover, when applied to disease tissue it may reveal quantitative and qualitative alterations in gene expression that give information on the context or underlying basis for the disease and may provide a new diagnostic approach. However, the data obtained from high-density microarrays is highly complex and poses considerable challenges in data mining. Past researches prove that neuro diseases damage the brain network interaction, protein- protein interaction and gene-gene interaction. A number of neurological research paper also analyze the relationship among damaged part. Analysis of gene-gene interaction network drawn by using state-of-the-art gene database of Alzheimer’s patient can conclude a lot of information. In this paper we used gene dataset affected with Alzheimer’s disease and normal patient’s dataset from NCBI databank. After proper processing the .CEL affymetrix data using RMA, we use the processed data to find gene interaction outputs. Then we filter the output files using probe set filtering attributes p-value and fold count and draw a gene-gene interaction network. Then we analyze the interaction network using GeneMania software.
Submission Deadline: 30th September 2022
Acceptance Notification: Within Three Days’ time period
Online Publication: Within 24 Hrs. time Period
Expected Date of Dispatch of Printed Journal: 5th October 2022
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...IAEME Publication
White layer thickness (WLT) formed and surface roughness in wire electric discharge turning (WEDT) of tungsten carbide composite has been made to model through response surface methodology (RSM). A Taguchi’s standard Design of experiments involving five input variables with three levels has been employed to establish a mathematical model between input parameters and responses. Percentage of cobalt content, spindle speed, Pulse on-time, wire feed and pulse off-time were changed during the experimental tests based on the Taguchi’s orthogonal array L27 (3^13). Analysis of variance (ANOVA) revealed that the mathematical models obtained can adequately describe performance within the parameters of the factors considered. There was a good agreement between the experimental and predicted values in this study.
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSIAEME Publication
The study explores the reasons for a transgender to become entrepreneurs. In this study transgender entrepreneur was taken as independent variable and reasons to become as dependent variable. Data were collected through a structured questionnaire containing a five point Likert Scale. The study examined the data of 30 transgender entrepreneurs in Salem Municipal Corporation of Tamil Nadu State, India. Simple Random sampling technique was used. Garrett Ranking Technique (Percentile Position, Mean Scores) was used as the analysis for the present study to identify the top 13 stimulus factors for establishment of trans entrepreneurial venture. Economic advancement of a nation is governed upon the upshot of a resolute entrepreneurial doings. The conception of entrepreneurship has stretched and materialized to the socially deflated uncharted sections of transgender community. Presently transgenders have smashed their stereotypes and are making recent headlines of achievements in various fields of our Indian society. The trans-community is gradually being observed in a new light and has been trying to achieve prospective growth in entrepreneurship. The findings of the research revealed that the optimistic changes are taking place to change affirmative societal outlook of the transgender for entrepreneurial ventureship. It also laid emphasis on other transgenders to renovate their traditional living. The paper also highlights that legislators, supervisory body should endorse an impartial canons and reforms in Tamil Nadu Transgender Welfare Board Association.
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSIAEME Publication
Since ages gender difference is always a debatable theme whether caused by nature, evolution or environment. The birth of a transgender is dreadful not only for the child but also for their parents. The pain of living in the wrong physique and treated as second class victimized citizen is outrageous and fully harboured with vicious baseless negative scruples. For so long, social exclusion had perpetuated inequality and deprivation experiencing ingrained malign stigma and besieged victims of crime or violence across their life spans. They are pushed into the murky way of life with a source of eternal disgust, bereft sexual potency and perennial fear. Although they are highly visible but very little is known about them. The common public needs to comprehend the ravaged arrogance on these insensitive souls and assist in integrating them into the mainstream by offering equal opportunity, treat with humanity and respect their dignity. Entrepreneurship in the current age is endorsing the gender fairness movement. Unstable careers and economic inadequacy had inclined one of the gender variant people called Transgender to become entrepreneurs. These tiny budding entrepreneurs resulted in economic transition by means of employment, free from the clutches of stereotype jobs, raised standard of living and handful of financial empowerment. Besides all these inhibitions, they were able to witness a platform for skill set development that ignited them to enter into entrepreneurial domain. This paper epitomizes skill sets involved in trans-entrepreneurs of Thoothukudi Municipal Corporation of Tamil Nadu State and is a groundbreaking determination to sightsee various skills incorporated and the impact on entrepreneurship.
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSIAEME Publication
The banking and financial services industries are experiencing increased technology penetration. Among them, the banking industry has made technological advancements to better serve the general populace. The economy focused on transforming the banking sector's system into a cashless, paperless, and faceless one. The researcher wants to evaluate the user's intention for utilising a mobile banking application. The study also examines the variables affecting the user's behaviour intention when selecting specific applications for financial transactions. The researcher employed a well-structured questionnaire and a descriptive study methodology to gather the respondents' primary data utilising the snowball sampling technique. The study includes variables like performance expectations, effort expectations, social impact, enabling circumstances, and perceived risk. Each of the aforementioned variables has a major impact on how users utilise mobile banking applications. The outcome will assist the service provider in comprehending the user's history with mobile banking applications.
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSIAEME Publication
Technology upgradation in banking sector took the economy to view that payment mode towards online transactions using mobile applications. This system enabled connectivity between banks, Merchant and user in a convenient mode. there are various applications used for online transactions such as Google pay, Paytm, freecharge, mobikiwi, oxygen, phonepe and so on and it also includes mobile banking applications. The study aimed at evaluating the predilection of the user in adopting digital transaction. The study is descriptive in nature. The researcher used random sample techniques to collect the data. The findings reveal that mobile applications differ with the quality of service rendered by Gpay and Phonepe. The researcher suggest the Phonepe application should focus on implementing the application should be user friendly interface and Gpay on motivating the users to feel the importance of request for money and modes of payments in the application.
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOIAEME Publication
The prototype of a voice-based ATM for visually impaired using Arduino is to help people who are blind. This uses RFID cards which contain users fingerprint encrypted on it and interacts with the users through voice commands. ATM operates when sensor detects the presence of one person in the cabin. After scanning the RFID card, it will ask to select the mode like –normal or blind. User can select the respective mode through voice input, if blind mode is selected the balance check or cash withdraw can be done through voice input. Normal mode procedure is same as the existing ATM.
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IAEME Publication
There is increasing acceptability of emotional intelligence as a major factor in personality assessment and effective human resource management. Emotional intelligence as the ability to build capacity, empathize, co-operate, motivate and develop others cannot be divorced from both effective performance and human resource management systems. The human person is crucial in defining organizational leadership and fortunes in terms of challenges and opportunities and walking across both multinational and bilateral relationships. The growing complexity of the business world requires a great deal of self-confidence, integrity, communication, conflict and diversity management to keep the global enterprise within the paths of productivity and sustainability. Using the exploratory research design and 255 participants the result of this original study indicates strong positive correlation between emotional intelligence and effective human resource management. The paper offers suggestions on further studies between emotional intelligence and human capital development and recommends for conflict management as an integral part of effective human resource management.
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYIAEME Publication
Our life journey, in general, is closely defined by the way we understand the meaning of why we coexist and deal with its challenges. As we develop the "inspiration economy", we could say that nearly all of the challenges we have faced are opportunities that help us to discover the rest of our journey. In this note paper, we explore how being faced with the opportunity of being a close carer for an aging parent with dementia brought intangible discoveries that changed our insight of the meaning of the rest of our life journey.
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...IAEME Publication
The main objective of this study is to analyze the impact of aspects of Organizational Culture on the Effectiveness of the Performance Management System (PMS) in the Health Care Organization at Thanjavur. Organizational Culture and PMS play a crucial role in present-day organizations in achieving their objectives. PMS needs employees’ cooperation to achieve its intended objectives. Employees' cooperation depends upon the organization’s culture. The present study uses exploratory research to examine the relationship between the Organization's culture and the Effectiveness of the Performance Management System. The study uses a Structured Questionnaire to collect the primary data. For this study, Thirty-six non-clinical employees were selected from twelve randomly selected Health Care organizations at Thanjavur. Thirty-two fully completed questionnaires were received.
Living in 21st century in itself reminds all of us the necessity of police and its administration. As more and more we are entering into the modern society and culture, the more we require the services of the so called ‘Khaki Worthy’ men i.e., the police personnel. Whether we talk of Indian police or the other nation’s police, they all have the same recognition as they have in India. But as already mentioned, their services and requirements are different after the like 26th November, 2008 incidents, where they without saving their own lives has sacrificed themselves without any hitch and without caring about their respective family members and wards. In other words, they are like our heroes and mentors who can guide us from the darkness of fear, militancy, corruption and other dark sides of life and so on. Now the question arises, if Gandhi would have been alive today, what would have been his reaction/opinion to the police and its functioning? Would he have some thing different in his mind now what he had been in his mind before the partition or would he be going to start some Satyagraha in the form of some improvement in the functioning of the police administration? Really these questions or rather night mares can come to any one’s mind, when there is too much confusion is prevailing in our minds, when there is too much corruption in the society and when the polices working is also in the questioning because of one or the other case throughout the India. It is matter of great concern that we have to thing over our administration and our practical approach because the police personals are also like us, they are part and parcel of our society and among one of us, so why we all are pin pointing towards them.
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...IAEME Publication
The goal of this study was to see how talent management affected employee retention in the selected IT organizations in Chennai. The fundamental issue was the difficulty to attract, hire, and retain talented personnel who perform well and the gap between supply and demand of talent acquisition and retaining them within the firms. The study's main goals were to determine the impact of talent management on employee retention in IT companies in Chennai, investigate talent management strategies that IT companies could use to improve talent acquisition, performance management, career planning and formulate retention strategies that the IT firms could use. The respondents were given a structured close-ended questionnaire with the 5 Point Likert Scale as part of the study's quantitative research design. The target population consisted of 289 IT professionals. The questionnaires were distributed and collected by the researcher directly. The Statistical Package for Social Sciences (SPSS) was used to collect and analyse the questionnaire responses. Hypotheses that were formulated for the various areas of the study were tested using a variety of statistical tests. The key findings of the study suggested that talent management had an impact on employee retention. The studies also found that there is a clear link between the implementation of talent management and retention measures. Management should provide enough training and development for employees, clarify job responsibilities, provide adequate remuneration packages, and recognise employees for exceptional performance.
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...IAEME Publication
Globally, Millions of dollars were spent by the organizations for employing skilled Information Technology (IT) professionals. It is costly to replace unskilled employees with IT professionals possessing technical skills and competencies that aid in interconnecting the business processes. The organization’s employment tactics were forced to alter by globalization along with technological innovations as they consistently diminish to remain lean, outsource to concentrate on core competencies along with restructuring/reallocate personnel to gather efficiency. As other jobs, organizations or professions have become reasonably more appropriate in a shifting employment landscape, the above alterations trigger both involuntary as well as voluntary turnover. The employee view on jobs is also afflicted by the COVID-19 pandemic along with the employee-driven labour market. So, having effective strategies is necessary to tackle the withdrawal rate of employees. By associating Emotional Intelligence (EI) along with Talent Management (TM) in the IT industry, the rise in attrition rate was analyzed in this study. Only 303 respondents were collected out of 350 participants to whom questionnaires were distributed. From the employees of IT organizations located in Bangalore (India), the data were congregated. A simple random sampling methodology was employed to congregate data as of the respondents. Generating the hypothesis along with testing is eventuated. The effect of EI and TM along with regression analysis between TM and EI was analyzed. The outcomes indicated that employee and Organizational Performance (OP) were elevated by effective EI along with TM.
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...IAEME Publication
By implementing talent management strategy, organizations would have the option to retain their skilled professionals while additionally working on their overall performance. It is the course of appropriately utilizing the ideal individuals, setting them up for future top positions, exploring and dealing with their performance, and holding them back from leaving the organization. It is employee performance that determines the success of every organization. The firm quickly obtains an upper hand over its rivals in the event that its employees having particular skills that cannot be duplicated by the competitors. Thus, firms are centred on creating successful talent management practices and processes to deal with the unique human resources. Firms are additionally endeavouring to keep their top/key staff since on the off chance that they leave; the whole store of information leaves the firm's hands. The study's objective was to determine the impact of talent management on organizational performance among the selected IT organizations in Chennai. The study recommends that talent management limitedly affects performance. On the off chance that this talent is appropriately management and implemented properly, organizations might benefit as much as possible from their maintained assets to support development and productivity, both monetarily and non-monetarily.
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...IAEME Publication
Banking regulations act of India, 1949 defines banking as “acceptance of deposits for the purpose of lending or investment from the public, repayment on demand or otherwise and withdrawable through cheques, drafts order or otherwise”, the major participants of the Indian financial system are commercial banks, the financial institution encompassing term lending institutions. Investments institutions, specialized financial institution and the state level development banks, non banking financial companies (NBFC) and other market intermediaries such has the stock brokers and money lenders are among the oldest of the certain variants of NBFC and the oldest market participants. The asset quality of banks is one of the most important indicators of their financial health. The Indian banking sector has been facing severe problems of increasing Non- Performing Assets (NPAs). The NPAs growth directly and indirectly affects the quality of assets and profitability of banks. It also shows the efficiency of banks credit risk management and the recovery effectiveness. NPA do not generate any income, whereas, the bank is required to make provisions for such as assets that why is a double edge weapon. This paper outlines the concept of quality of bank loans of different types like Housing, Agriculture and MSME loans in state Haryana of selected public and private sector banks. This study is highlighting problems associated with the role of commercial bank in financing Small and Medium Scale Enterprises (SME). The overall objective of the research was to assess the effect of the financing provisions existing for the setting up and operations of MSMEs in the country and to generate recommendations for more robust financing mechanisms for successful operation of the MSMEs, in turn understanding the impact of MSME loans on financial institutions due to NPA. There are many research conducted on the topic of Non- Performing Assets (NPA) Management, concerning particular bank, comparative study of public and private banks etc. In this paper the researcher is considering the aggregate data of selected public sector and private sector banks and attempts to compare the NPA of Housing, Agriculture and MSME loans in state Haryana of public and private sector banks. The tools used in the study are average and Anova test and variance. The findings reveal that NPA is common problem for both public and private sector banks and is associated with all types of loans either that is housing loans, agriculture loans and loans to SMES. NPAs of both public and private sector banks show the increasing trend. In 2010-11 GNPA of public and private sector were at same level it was 2% but after 2010-11 it increased in many fold and at present there is GNPA in some more than 15%. It shows the dark area of Indian banking sector.
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...IAEME Publication
An experiment conducted in this study found that BaSO4 changed Nylon 6's mechanical properties. By changing the weight ratios, BaSO4 was used to make Nylon 6. This Researcher looked into how hard Nylon-6/BaSO4 composites are and how well they wear. Experiments were done based on Taguchi design L9. Nylon-6/BaSO4 composites can be tested for their hardness number using a Rockwell hardness testing apparatus. On Nylon/BaSO4, the wear behavior was measured by a wear monitor, pinon-disc friction by varying reinforcement, sliding speed, and sliding distance, and the microstructure of the crack surfaces was observed by SEM. This study provides significant contributions to ultimate strength by increasing BaSO4 content up to 16% in the composites, and sliding speed contributes 72.45% to the wear rate
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...IAEME Publication
The majority of the population in India lives in villages. The village is the back bone of the country. Village or rural industries play an important role in the national economy, particularly in the rural development. Developing the rural economy is one of the key indicators towards a country’s success. Whether it be the need to look after the welfare of the farmers or invest in rural infrastructure, Governments have to ensure that rural development isn’t compromised. The economic development of our country largely depends on the progress of rural areas and the standard of living of rural masses. Village or rural industries play an important role in the national economy, particularly in the rural development. Rural entrepreneurship is based on stimulating local entrepreneurial talent and the subsequent growth of indigenous enterprises. It recognizes opportunity in the rural areas and accelerates a unique blend of resources either inside or outside of agriculture. Rural entrepreneurship brings an economic value to the rural sector by creating new methods of production, new markets, new products and generate employment opportunities thereby ensuring continuous rural development. Social Entrepreneurship has the direct and primary objective of serving the society along with the earning profits. So, social entrepreneurship is different from the economic entrepreneurship as its basic objective is not to earn profits but for providing innovative solutions to meet the society needs which are not taken care by majority of the entrepreneurs as they are in the business for profit making as a sole objective. So, the Social Entrepreneurs have the huge growth potential particularly in the developing countries like India where we have huge societal disparities in terms of the financial positions of the population. Still 22 percent of the Indian population is below the poverty line and also there is disparity among the rural & urban population in terms of families living under BPL. 25.7 percent of the rural population & 13.7 percent of the urban population is under BPL which clearly shows the disparity of the poor people in the rural and urban areas. The need to develop social entrepreneurship in agriculture is dictated by a large number of social problems. Such problems include low living standards, unemployment, and social tension. The reasons that led to the emergence of the practice of social entrepreneurship are the above factors. The research problem lays upon disclosing the importance of role of social entrepreneurship in rural development of India. The paper the tendencies of social entrepreneurship in India, to present successful examples of such business for providing recommendations how to improve situation in rural areas in terms of social entrepreneurship development. Indian government has made some steps towards development of social enterprises, social entrepreneurship, and social in- novation, but a lot remains to be improved.
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...IAEME Publication
Distribution system is a critical link between the electric power distributor and the consumers. Most of the distribution networks commonly used by the electric utility is the radial distribution network. However in this type of network, it has technical issues such as enormous power losses which affect the quality of the supply. Nowadays, the introduction of Distributed Generation (DG) units in the system help improve and support the voltage profile of the network as well as the performance of the system components through power loss mitigation. In this study network reconfiguration was done using two meta-heuristic algorithms Particle Swarm Optimization and Gravitational Search Algorithm (PSO-GSA) to enhance power quality and voltage profile in the system when simultaneously applied with the DG units. Backward/Forward Sweep Method was used in the load flow analysis and simulated using the MATLAB program. Five cases were considered in the Reconfiguration based on the contribution of DG units. The proposed method was tested using IEEE 33 bus system. Based on the results, there was a voltage profile improvement in the system from 0.9038 p.u. to 0.9594 p.u.. The integration of DG in the network also reduced power losses from 210.98 kW to 69.3963 kW. Simulated results are drawn to show the performance of each case.
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...IAEME Publication
Manufacturing industries have witnessed an outburst in productivity. For productivity improvement manufacturing industries are taking various initiatives by using lean tools and techniques. However, in different manufacturing industries, frugal approach is applied in product design and services as a tool for improvement. Frugal approach contributed to prove less is more and seems indirectly contributing to improve productivity. Hence, there is need to understand status of frugal approach application in manufacturing industries. All manufacturing industries are trying hard and putting continuous efforts for competitive existence. For productivity improvements, manufacturing industries are coming up with different effective and efficient solutions in manufacturing processes and operations. To overcome current challenges, manufacturing industries have started using frugal approach in product design and services. For this study, methodology adopted with both primary and secondary sources of data. For primary source interview and observation technique is used and for secondary source review has done based on available literatures in website, printed magazines, manual etc. An attempt has made for understanding application of frugal approach with the study of manufacturing industry project. Manufacturing industry selected for this project study is Mahindra and Mahindra Ltd. This paper will help researcher to find the connections between the two concepts productivity improvement and frugal approach. This paper will help to understand significance of frugal approach for productivity improvement in manufacturing industry. This will also help to understand current scenario of frugal approach in manufacturing industry. In manufacturing industries various process are involved to deliver the final product. In the process of converting input in to output through manufacturing process productivity plays very critical role. Hence this study will help to evolve status of frugal approach in productivity improvement programme. The notion of frugal can be viewed as an approach towards productivity improvement in manufacturing industries.
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTIAEME Publication
In this paper, we investigated a queuing model of fuzzy environment-based a multiple channel queuing model (M/M/C) ( /FCFS) and study its performance under realistic conditions. It applies a nonagonal fuzzy number to analyse the relevant performance of a multiple channel queuing model (M/M/C) ( /FCFS). Based on the sub interval average ranking method for nonagonal fuzzy number, we convert fuzzy number to crisp one. Numerical results reveal that the efficiency of this method. Intuitively, the fuzzy environment adapts well to a multiple channel queuing models (M/M/C) ( /FCFS) are very well.
Applications of artificial Intelligence in Mechanical Engineering.pdfAtif Razi
Historically, mechanical engineering has relied heavily on human expertise and empirical methods to solve complex problems. With the introduction of computer-aided design (CAD) and finite element analysis (FEA), the field took its first steps towards digitization. These tools allowed engineers to simulate and analyze mechanical systems with greater accuracy and efficiency. However, the sheer volume of data generated by modern engineering systems and the increasing complexity of these systems have necessitated more advanced analytical tools, paving the way for AI.
AI offers the capability to process vast amounts of data, identify patterns, and make predictions with a level of speed and accuracy unattainable by traditional methods. This has profound implications for mechanical engineering, enabling more efficient design processes, predictive maintenance strategies, and optimized manufacturing operations. AI-driven tools can learn from historical data, adapt to new information, and continuously improve their performance, making them invaluable in tackling the multifaceted challenges of modern mechanical engineering.
Null Bangalore | Pentesters Approach to AWS IAMDivyanshu
#Abstract:
- Learn more about the real-world methods for auditing AWS IAM (Identity and Access Management) as a pentester. So let us proceed with a brief discussion of IAM as well as some typical misconfigurations and their potential exploits in order to reinforce the understanding of IAM security best practices.
- Gain actionable insights into AWS IAM policies and roles, using hands on approach.
#Prerequisites:
- Basic understanding of AWS services and architecture
- Familiarity with cloud security concepts
- Experience using the AWS Management Console or AWS CLI.
- For hands on lab create account on [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
# Scenario Covered:
- Basics of IAM in AWS
- Implementing IAM Policies with Least Privilege to Manage S3 Bucket
- Objective: Create an S3 bucket with least privilege IAM policy and validate access.
- Steps:
- Create S3 bucket.
- Attach least privilege policy to IAM user.
- Validate access.
- Exploiting IAM PassRole Misconfiguration
-Allows a user to pass a specific IAM role to an AWS service (ec2), typically used for service access delegation. Then exploit PassRole Misconfiguration granting unauthorized access to sensitive resources.
- Objective: Demonstrate how a PassRole misconfiguration can grant unauthorized access.
- Steps:
- Allow user to pass IAM role to EC2.
- Exploit misconfiguration for unauthorized access.
- Access sensitive resources.
- Exploiting IAM AssumeRole Misconfiguration with Overly Permissive Role
- An overly permissive IAM role configuration can lead to privilege escalation by creating a role with administrative privileges and allow a user to assume this role.
- Objective: Show how overly permissive IAM roles can lead to privilege escalation.
- Steps:
- Create role with administrative privileges.
- Allow user to assume the role.
- Perform administrative actions.
- Differentiation between PassRole vs AssumeRole
Try at [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
Gas agency management system project report.pdfKamal Acharya
The project entitled "Gas Agency" is done to make the manual process easier by making it a computerized system for billing and maintaining stock. The Gas Agencies get the order request through phone calls or by personal from their customers and deliver the gas cylinders to their address based on their demand and previous delivery date. This process is made computerized and the customer's name, address and stock details are stored in a database. Based on this the billing for a customer is made simple and easier, since a customer order for gas can be accepted only after completing a certain period from the previous delivery. This can be calculated and billed easily through this. There are two types of delivery like domestic purpose use delivery and commercial purpose use delivery. The bill rate and capacity differs for both. This can be easily maintained and charged accordingly.
Generative AI Use cases applications solutions and implementation.pdfmahaffeycheryld
Generative AI solutions encompass a range of capabilities from content creation to complex problem-solving across industries. Implementing generative AI involves identifying specific business needs, developing tailored AI models using techniques like GANs and VAEs, and integrating these models into existing workflows. Data quality and continuous model refinement are crucial for effective implementation. Businesses must also consider ethical implications and ensure transparency in AI decision-making. Generative AI's implementation aims to enhance efficiency, creativity, and innovation by leveraging autonomous generation and sophisticated learning algorithms to meet diverse business challenges.
https://www.leewayhertz.com/generative-ai-use-cases-and-applications/
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
Software Engineering and Project Management - Introduction, Modeling Concepts...Prakhyath Rai
Introduction, Modeling Concepts and Class Modeling: What is Object orientation? What is OO development? OO Themes; Evidence for usefulness of OO development; OO modeling history. Modeling
as Design technique: Modeling, abstraction, The Three models. Class Modeling: Object and Class Concept, Link and associations concepts, Generalization and Inheritance, A sample class model, Navigation of class models, and UML diagrams
Building the Analysis Models: Requirement Analysis, Analysis Model Approaches, Data modeling Concepts, Object Oriented Analysis, Scenario-Based Modeling, Flow-Oriented Modeling, class Based Modeling, Creating a Behavioral Model.
2. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 18 editor@iaeme.com
1. INTRODUCTION
Bioinformatics has evolved and expanded continuously over the past four decades and
has grown into a very important bridging discipline in life science research. With the
advent of high-throughput biotechnologies, biological data like DNA, RNA, and
protein data are generated faster than ever. Huge amounts of data are being produced
and collected. The biologist needs computational models to help manage and analyze
such large and complex data sets. Database and Web technologies are used to build
plenty of online data banks for data storage and sharing. Most of the data collected
have been put on the World Wide Web and can be shared and accessed online. To
know the updated numbers of complete genomes, nucleotides, and protein coding
sequences, the reader can check the Genome Reviews of EMBL-EBI
(http://www.ebi.ac.uk/GenomeReviews/stats/). The researcher is also referred to
Protein Data Bank for the number of known protein structures. As for the analysis of
the data, data mining technologies can be utilized. The mystery of life hidden in the
biological data might be decoded much faster and more accurately with the data
mining technologies. To follow the scientific output produced regarding a single
disease, such as dengue, a scientist would have to scan more than a hundred different
journals and read a few dozen papers per day. Currently different biological types of
data, such as sequences, protein structures and families, proteomics data, gene
ontologies, gene expression and other experimental data are stored in distinct
databases [1]. Existing databases or data collection can be very specialized and often
they store the information using specific data formats [11].
The challenge lies in the analysis of a huge amount of data to extract meaningful
information and use them to answer some of the fundamental biological questions. So,
there is the need to develop an interactive tool to visualize the representation of
information together with data analysis techniques to simplify the interpretation of data.
The incidence of dengue has grown dramatically around the world in recent decades.
Over 2.5 billion people, 40% of the world's population, are now at risk on account of
Dengue. World Health Organization (WHO) currently estimates that there may be 50–
100 million Dengue infections worldwide every year. As per the medical record of
Government of TamilNadu, India, 15,535 persons were affected and 96 were expired in
the year 2009. The outbreak of dengue in India in the year 2012 was the worst in the
previous six years. In the months of December 2014 and January 2015 alone, nearly 20
persons including children, expired in the Virudhunagar District of TamilNadu [9].
Under these circumstances, the work on the genome sequence of dengue virus
plays a vital role in the diagnosis of the disease. Therefore, it is necessary to predict
the presence of co-occurrence patterns which are the similar elements present in
dengue gene sequences. The dengue virus belongs to Flavi viridae family that is
transmitted to people through the bite of the mosquitoes named Aedes aegypti or
Aedes albopictus. Serotype refers to the subdivisions of a virus that are classified
based on their cell surface. They are listed in the Table-1.
Table 1 Types of Dengue Virus Serotypes
Virus type Name of the Virus
DEN-1 Strain Hawaii
DEN-2 Strain New Guinea C
DEN-3 Strain H87
DEN-4 Strain H241
3. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 19 editor@iaeme.com
There are three main types of dengue infection viz. Classic Dengue Fever (CD),
Dengue Hemorrhagic (DH) fever and Dengue Shock Syndrome (DSS). All the types
of dengue fever begin with noticeable symptoms within four to seven days after the
Aedes aegypti mosquito’s bite. The symptoms of CD include headache, pain behind
the eyes, joints and muscles, vomiting and body rash. It also reduces the count of
White Blood Cells (WBC). DH fever includes all the classic symptoms with higher
fever and sharp decrease in the number of platelets in the blood. Platelets are small,
disk shaped fragments that are the natural source of growth factors. They are
circulated in the blood and involved in the formation of blood clots. As a result of
this, victims bleed from the nose, gums and skin. DSS is the most severe form of the
disease that causes massive bleeding and fall in the blood pressure [14]. Each virus
type has its own characteristics.
In Section 2, the work related to bio-computational tools is outlined. Section 3
demonstrates the methodologies related to the proposed sequence miner tool. Section
4 exhibits the experimental results that were obtained using dengue virus serotype
dataset. Finally, Section 5 describes conclusion.
2. RELATED WORK
Basic biological research includes a wide range of studies focused on learning how
the dengue virus is transmitted, how it infects cells and causes disease. Further many
research works investigate several aspects of dengue viral biology that includes
exploration of the interactions between the virus and humans as well as the repetition
of dengue virus serotypes. Researchers have also been studying the dengue viruses to
understand the factors that are responsible for transmitting the virus to humans. They
found that specific viral sequences are associated with other severe dengue symptoms
[5]. Therefore, the literature on dengue fever viewed as three aspects viz. biological,
computational and bio-computational. For this paper, the related work focused on bio-
computational aspect.
There are several computational biology tools that have been developed over the
last two decades. These tools are selected to cover the range of different
functionalities and features for data analysis and visualization [2]. Some of the tools
are reviewed here.
Medusa [8] is a Java application oriented and is available as an applet. It is an
open source product under the General Public License (GPL). The visualization is
based on the Fruchterman-Reingold [6] algorithm and it provides two dimensional
representations of networks of medium size for up to a few hundred nodes and edges.
It is less suited for the visualization of big datasets. Medusa uses non-directed, multi-
edge connections, which allow the simultaneous representation of more than one
connection between two bio-entities. Additional nodes can be fixed in order to
facilitate pattern recognition and the spring embedded layout algorithms help the
relaxation of the network. It supports weighted graphs and represents the significance
and importance of a connection by varying line thickness. The compatibility of
Medusa has its own text file format that is not compatible with other visualization
tools or integrated with other data sources. The input file format allows the user to
annotate each node or connection. It allows the selection and analysis of subsets of
nodes. A text search, which supports regular expressions, can be applied to find
nodes. The status of a network can be saved and reloaded at any time when Medusa is
not currently connected to any data source.
4. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 20 editor@iaeme.com
Cytoscape [13] is a standalone Java application. It is an open source project under
LesserGPL (LGPL) license. It mainly provides two dimensional representations and is
suitable for large-scale network analysis with hundredth thousands of nodes and
edges. It can support directed, undirected and weighted graphs and comes with
powerful visual styles that allow the user to change the properties of nodes or edges.
The tool provides a variety of layout algorithms including cyclic and spring-
embedded layouts. Furthermore, expression data can be mapped as node color, label,
border thickness, or border color. Cytoscape comes with various data parsers or filters
that make it compatible with other tools. The file formats that are supported to save or
load the graphs are Setup InFormation (SIF), Geography Markup Language (GML),
eXtensible Graph Markup and Modeling Language (XGMML) and Biology
PAthaway eXchange (BioPAX). It also allows the user to import messengerRNA
(mRNA) expression profiles and gene functional annotations from the Gene Ontology
(GO). Users can also directly import GO Terms and annotations from gene
association files. It is highly interactive and the user can zoom in or out and browse
the network. The status of the network as well as the edge or node properties can be
saved and reloaded. In addition, Cytoscape comes with a network manager to easily
organize multiple networks. The user can have many different panels that hold the
status of the network at different time points which makes it an efficient tool to
compare networks between each other. It also comes with efficient network filtering
capabilities. Users can select subsets of nodes and/or interactions and search for active
sub networks or pathway modules. It incorporates statistical analysis of the network
and makes it easy to cluster or detect highly interconnected regions. The main purpose
of this tool is the visualization of molecular interaction networks and their integration
with gene expression profiles and other data. It also allows the user to manipulate and
compare multiple networks. Many plug-ins created by users are available and allow
more specialized analysis of networks and molecular profiles.
Osprey [3] is a standalone application running under a wide range of platforms. It
can be licensed for non-commercial use and the source code is currently not available.
Osprey provides two dimensional representations of directed, undirected and
weighted networks. It is not efficient for large scale network analysis, various layout
options and ways to arrange nodes in various geometric distributions. The layouts
range from the relax algorithm over a simple circular layout to a more advanced dual
spiked ring layout that displays up to 1500 – 2000 nodes in a easily manageable
format. The user can change the size and the colors of most Osprey objects such as
edges, nodes, labels, and arrow heads. Data can be loaded into the tool either using
different text formats or by connecting directly to several databases, such as the
General Repository of Interaction Datasets (GRID) or BioGRID [15] database. In
addition to its own Osprey file format, the tool can also load custom gene network and
gene list formats, making Osprey compatible with other tools relying on the same file
formats. Osprey networks can be saved in Scalable Vector Graphics (SVG), Portable
Network Graphics (PNG) and Joint Photographic Experts Group (JPEG) format. The
tool provides several features for functional assessment and comparative analysis of
different networks together with network and connectivity filters and dataset
superimposing. Osprey also has the ability to cluster genes by GO Processes. Network
filters can extract biological information that is supplied to Osprey either by the user
or by instructions inside the GRID dataset. Connectivity filters identify nodes based
on their connectivity levels. Finally, Osprey includes basic functions such as selecting
and moving individual nodes or groups of nodes or removing nodes and edges. With
its various filtering capabilities, Osprey is a powerful tool for network manipulation.
5. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 21 editor@iaeme.com
The ability to incorporate new interactions into an already existing network might be
considered the tool's biggest asset.
ProViz [10] is a standalone open source application under the GPL license. It
comes with both two dimensional and pseudo- three dimensional display support to
render data. It can manipulate single graphs in large-scale dataset with millions of
nodes or connections. It generates appealing 3-Dimensional (3D) visualizations. In
addition, the tool also offers a circular and a hierarchical layout, which improve the
detection of metabolic pathways or gene regulation networks in large datasets. ProViz
is ideal to gain a first overview of networks because it allows fast navigation through
graphs. Graphs are saved and loaded in Tulip format which is a drawing package.
Networks can also be exported in PNG format. Subgraphs are produced by selection,
filtering or clustering methods and can be automatically organized into views. With
ProViz it is possible to annotate each node and each edge with comments or merge
different datasets into a single graph. Users can also enrich the networks by querying
available online databases. ProViz uses a controlled vocabulary on bio-molecules and
interactions, described in eXtensible Markup Language (XML) format. It has its
strength in the area of protein – protein interaction networks and their analysis using
arbitrary properties and taxonomic identifier. Its plug-in architecture allows a
diversification of function according to the user's needs.
Ondex [7] is a standalone freely available open source application. It provides two
dimensional representations of directed, undirected and weighted networks. It can
handle large scale networks of hundred thousands of nodes and edges. It also supports
bidirectional connections, which are represented as curves. Moreover, different types
of data are separated by placing them in different disk-circles interconnected between
each other. Data may be imported through a number of 'parsers' for public-domain
and other databases, such as TRANScription FACtor (TRANSFAC), TRANScription
PATH (TRANSPATH), CHEmical Entities of Biological Interest (CHEBI), GO, Kyto
Encyclopedia of Genes and Genomes (KEGG), Drastic, Enzyme Nomenclature,
Expert Protein Analysis System (ExPASy), Pathway Tools, Pathway Genome
DataBases (PGDBs), Plant Ontology and Medical Subject Headings Vocabulary
(MeSH). Graph objects can be exported to Cell Illustrator and XML formats. To
reload or feed into other applications graph objects may be saved as ONDEX XML or
an XGMML form. Ondex integrates various filters that selectively add or remove
connected nodes from the display according to user selectable rules of connectivity
type like distance, level or equivalence [12].
Pathway Analysis Tools for Integration and Knowledge Acquisition (PATIKA)
[4] is a web based non-open source application publicly available for non-commercial
use. It has its own license. It provides 2D representations of single or directed graphs.
There are no limitations regarding the size of the graphs. It offers a very intuitive and
widely accepted representation for cellular processes using directed graphs where
nodes correspond to molecules and edges correspond to interactions between them.
Even though the implemented variety of layout algorithms is rather limited, PATIKA
is able to support bipartite graph of states and transitions. It represents different types
of edges: product edges, where the source and target nodes of a product edge define
the transition and a product of this transition, activator edges, where the source and
target nodes of an activator edge define the activating state and the transition that is
activated by this state, inhibitor edges where the source and target nodes of an
activator edge define the inhibiting state and the transition that is inhibited by this
state and substrate edges where the target and source nodes of a substrate edge define
6. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 22 editor@iaeme.com
the transition and a substrate of this transition respectively. It integrates data from
several sources, including Entrez Gene, Universal Protein Resource (UniProt), GO,
Human Protein Reference Database (HPRD) and Reactome pathway databases. Users
can query and access data using PATIKA's web query interface, and save their results
in XML format or export them as common picture formats. BioPAX and Systems
Biology Markup Language (SBML) exporters can be used as part of Patikas Web
service. The user can connect to the server and query the database to construct the
desired pathway. Pathways are created on the fly, and drawn automatically. The user
can manipulate a pathway through operations such as add new state or remove an
existing transition, edit its contents such as the description of a state or transition or
change the graphical view of a pathway component. PATIKA is a tool for data
integration and pathway analysis. It is an integrated software environment designed to
provide researchers a complete solution for modeling and analyzing cellular
processes. It is one of the few tools that allow visualizing transitions efficiently.
Though, there are various tools available to perform the sequence analysis, the
works related to find all the periodicities are very limited. Further the existing works
concentrate mainly on sequence alignment. Therefore there is a need for holistic
approach that computes all kinds of periodicities and their associations. In the current
work, we propose a tool called ‘Sequence Miner’ to classify the given sequence and
visualize the structure of the protein.
3. METHODOLOGY
The following steps are involved in the development of the bio-computational model
named ‘Sequence Miner’ for classifying the evolution of dengue virus serotypes. The
work flow of the proposed work is illustrated in Figure.1.
Figure 1 Work flow of the Sequence Miner
3.1. Data Collection
The primary step is to acquire the knowledge about dengue virus such as its serotypes,
replication cycle, symptoms caused and diagnosis methods available to detect dengue
and also identify the toxic protein in dengue virus. Each virus has toxic proteins that
cause diseases in human. Dengue virus has toxic proteins E and M. Then the online
composite database National Center for Biotechnology Information (NCBI) is used to
collect the gene sequences of all four dengue virus serotypes.
3.2. Data Preprocessing
The next step is to preprocess the collected data using sequence alignment algorithms
such as local alignment and global alignment [5]. The length variation and
inconsistent of the sequence are eliminated through the preprocessing. After
preprocessing, the aligned sequences are transferred to the next process.
7. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 23 editor@iaeme.com
3.3. Periodic Association Rules
A further step in this direction is the prediction of co-occurrence patterns among the
dengue gene sequences. This can be done by evaluating the rules that can reveal the
occurrence of an element or subsequence. Such rules are called Periodic Association
Rules, and the corresponding technique is called Periodic Association Rule Mining
(PARM). The PARM is similar to market basket analysis. In PARM terminology, the
nucleic or amino acids may be considered as items and the gene subsequences as the
baskets that contain the items. In the traditional association rules, only the number of
frequent items is calculated whereas PARM calculates the occurrence order of
frequent item sets along with its periodic position.
To obtain periodic association rule, the frequencies of nucleic or amino-acids are
computed in each dengue gene sequence. The rule can be expressed as A! C, where
A and C are the associated items. The rules state that if a nucleic acid A is present in a
given sequence with f1 periodicity then there will be another nucleic acid C that will
have similar periodicity with respect to their respective initial positions. The PARM
procedure enables to find the periodicity f1 along with their starting positions.
Let I = {i1, ...., ik} be a set of k elements, called items. Let Is = {b1, ...., bn} be a set
of n subsets of I. We call each bi as a set of transaction. In the market basket
application [6], the set I denotes the items stocked by a retail outlet and each basket bi
is the set of items of a transaction. Similarly, in case of gene sequence the set I denote
the elements of nucleic or amino acid and the basket bi is the orderly subsequences.
The order and frequency of the elements can be evaluated using the suffix tree. The
PAR is intended to capture the orderly dependence among the elements of dengue
virus dataset and the rule can be represented as i1 ! i2 along with the period and
starting position of i1 and i2, provided the following conditions hold good:
1. i1 and i2 occur at regular intervals in the sequence for at least s% of the n baskets
where s is the support and n is the number of subsequences.
2. For all the subsequences containing i1, at least c% of subsequences contains i2 where
c is the confidence.
The above definition can be extended to form multidimensional periodic
association rule such as AC ! GT, where AC and GT are element of nucleic acid
with periodic dependence. The association rules are considered interesting if they
satisfy both minimum support and confidence thresholds. The threshold values are set
by users based on their domain expertise.
To evaluate the PAR we propose the RECurrence FINder (RECFIN) algorithm.
The following steps are involved in the RECFIN algorithm:
1. Based on the occurrence positions the elements are mapped into integers.
2. Based on the support threshold the element periodicity is found. The set of elements
that satisfies the minimum support threshold is called the frequent item set.
3. The frequent item sets are used to generate association rules. For example, consider
the item set {A, C, G}. The following rules can be evaluated using the given
item set:
Rule 1: A ^ C ! G
Rule 2: C ^ G ! A
Rule 3: A ^ G ! C
8. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 24 editor@iaeme.com
Rule 4: G ^ A ! C
Rule 5: C ^ A ! G
Rule 6: G ^ C ! A
In the above rules the element that appears in left hand side is called antecedent and
that of the right hand side is called consequent. The confidence is computed using the
conditional probability of antecedent. For example, the confidence of the rule 1 is
computed as follows:
Confidence = support {A, C, G}/support{A,C}]
If the confidence is equal to or greater than a given confidence threshold, the rule is
considered as interesting rule.
4. Based on the support and confidence the PAR is generated.
3. 4. Amino Acid Component based Classification (AACC)
The AACC algorithm is based on the ID3 classifier. Based on PAR the given
sequence may be classified into six components such as Sulfur, Aromatic, Alphatic,
Acidic, Basic and Neutral. The classifier model has two phases viz. i) model
construction ii) model usage as illustrated in Figure.2. (a) and (b)
Neutral component produces the Asparagine, Serine, Threonine and Glutamine.
Sulfur component produces the Cytosine and Methoine. Alphatic component produces
the Leucine, Isoleucine, Glycine, Valine and Alanine. The Basic component produces
the amino acids Arginine and Lysine. Acidic component produces Glutamic and
Aspartic acids. Aromatic component produces the Phenylalanine, Tryptophan and
Tyrosine.
Figure 2(a) Model Construction
9. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 25 editor@iaeme.com
Figure 2(b) Model Usage
The classification model was trained by 10,735 sequences and the testing phase
was conducted by 10,198 sequences through this model.
Apart from these, a new dengue virus serotype was recently found by the
scientists that also creates the DH fever. DEN5 is the Non-Structural (NS) Protein
which indicates the new type of serotypes emerged from the existing serotypes.
Therefore, the proposed bio-computational model aims to predict the future evolution
of dengue virus serotype by analyzing the existing sequence of dengue virus
serotypes. Also this model is helpful to analyze the different kinds of gene sequences
like Nucleic acid (DNA, RNA) and Amino acid (Protein) sequences and visualize
them.
In fact, dengue provides to the drug designer significant difficulties that may not
be found for other virus infections like malaria, yellow fever, bird flu fever, etc. The
proposed tool aims to detect all other infections which are caused by various virus
serotypes.
3.5. Visualization / Graphical Representation
Bio-computational tools are the software programs for analyzing the biological data
and extracting the patterns from them. In addition to that tools must be user friendly,
even beginners can also be benefited by using them. The tool “Sequence Miner” will
be suitable for both the experts and the beginners to get knowledge about different
organisms via sequence analysis.
All the processes are graphically visualized in the sequence miner tool. It has the
interactive feature to classify the given sequence along with its structure. The
following are the effective features of the proposed tool: i) data collection through the
online ii) sequence alignment iii) generation of periodic association rules iv)
classification based on the amino acids and v) visualize the structure of the protein.
The layout of the sequence miner tool is illustrated in Figure.3.
The input sequences are collected from various online data repositories such as
NCBI, GenBank and related web sites as illustrated in Figure.4. The file format of the
input sequence may be text file or access number of the whole sequence.
10. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 26 editor@iaeme.com
Figure 3 Sequence Miner - Layout
Figure 4 Online Database
3.5.1. Sequence Comparison
The compare menu performs two tasks on dengue Serotypes: (i) Hit Rate (ii) Longest
Common Subsequence (LCS). Hit Rate compares two DNA or Protein sequences and
displays the number of matches between the two sequences and its matching
percentage as shown in Figure.5. LCS compares two DNA or Protein sequences then
predicts and displays the all common subsequence and longest common subsequences
using edit distance method. It also displays the execution time of the algorithm to
produce the result as shown in Figure.6. The comparison ratio also differs from
DEN1, DEN2, DEN3 and DEN4. However, DEN3 and DEN4 shows the similar hit
rate ranging from 70% to 86 %.
3.5.2. Sequence Alignment
The aim of the sequence alignment is to match the most similar elements of two
sequences. In comparing sequences, one should account for the influence of
11. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 27 editor@iaeme.com
molecular evolution. The probability of acceptably replacing an amino acid with a
similar amino acid is greater than replacement by a very different one. Substitution
matrices evaluate potential replacements for protein and nucleic acid sequences.
Figure 5 Comparison of Two Sequences
Figure 6 Finding the LCS in the given sequences
An optimal pairwise alignment is an alignment which has the maximum amount
of similarity with the minimum number of residue 'substitutions'. There are two types
of substitution matrix available:
1. Point Accepted Mutation (PAM)
2. BLOck SUbstitution Matrix (BLOSUM)
PAM is constructed by examining the kind of mutation that occurs in closely
related protein sequences i.e. mutation of one residue accepted by evolution.
BLOSUM is derived based on direct observation for every possible amino acid
substitution in multiple sequence alignment and it depends only on the identity of
protein sequences. BLOSUM matrices effectively represent more distant sequence
relationships, and BLOSUM62 has become a standard matrix. So in the pairwise
alignment module, BLOSUM 62 is used to find out both local and
12. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 28 editor@iaeme.com
global alignments between sequences.
Figure 7 Needleman – Wunsch Global Alignment Algorithm
The pairwise alignment module performs two tasks on dengue serotypes: (i)
Needleman- Wunsch (NW) Align with Blosum62 (ii) Smith- Waterman (SW) Align
with Blosum62. NW Align with Blosum62 performs global alignment. Global
alignment optimizes the alignment over the full-length of the sequences to find out the
similarity between two closely related sequences. In order to carry out the global
alignments on DNA or Protein sequences, Sequence Miner implemented the dynamic
algorithms NW algorithm and uses the Blosum62 substitution matrix as shown in
Figure.7. SW algorithm with Blosum62 performs local alignment. Local alignment is
for determining similar regions between two distinctly related DNA and Protein
sequences. In order to achieve the local alignments on DNA or Protein sequences,
Sequence Miner implemented the dynamic SW algorithm and uses the Blosum62
substitution matrix. It also displays the execution time of the algorithm for the given
input as shown in Figure.8.
Figure 8 Smith-Waterman Local Alignment Algorithm
3.5.3. Pattern Matching
Pattern matching is an important task in bioinformatics algorithms that try to find a
place where one or several patterns are found within a larger sequence or text.
13. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 29 editor@iaeme.com
Sequence Miner implements the Boyer Moore and Suffix tree algorithm in order to
highlight the user specified protein or nucleotide pattern in the inputted dengue
sequence. It also displays the periodical patterns along with periodic association rules
and execution time of the algorithm needed to find and highlight the searching
patterns as shown in Figure.9. The periodic association rules are mined with latent
periodicity that also exhibits the evolutionary relationship among DEN3 and DEN4.
3.5.4. Periodic Association Rules
The periodic patterns are extracted from the aligned sequences using novel RECFIN
algorithm. The RECFIN algorithm finds element, subsequence and latent periodicities
using suffix tree. The sample input sequence is given as in Figure.10. The suffix tree
finds the subsequence periodicities present in the given sequence. The sample result
displayed in Figure.11. The periodic patterns including latent periodicities are
identified along with their position as shown in Figure.12. The resultant PAR mined
with the help of minimum support and confident thresholds as shown in Figure.13.
Figure 9 Pattern Matching Algorithms
Figure 10 Sample Input Sequence
14. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 30 editor@iaeme.com
Figure 11 Suffix Tree for the partial subsequence
Figure 12 Periodic patterns along with position
Figure 13 Mining Periodic Association Rules
15. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 31 editor@iaeme.com
3.5.5. Amino Acid Component based Classification
This classification works based on the novel AACC algorithm. The mined amino
acids are classified based on the amino acid components and the classification results
illustrated in Figure.13.
3.5.6. Visualization
The visualization module exhibits the 3-Dimensional structure of the proteins as
shown in Figure.15. The variation of the protein structure of all dengue serotypes are
visualized in this module. Some additional requirements are needed for the
visualization such as Rasmol (Graphics Visualization), Swiss Protein DataBase
Viewer (SPDBV).
Figure 14 Classification Results
4. EXPERIMENTAL RESULTS
The total number of dengue gene sequences available in the NCBI is 21, 026. From
this huge number of sequences 10, 735 sequences have been taken for the training set.
The proposed model classifies 10, 198 correctly. The accuracy of classification result
is 96.74%. The resultant sequence can be calculated by subtracting the non classified
sequence from the total sequences.
Result = 10,735 – 537 = 10, 198
The same set of data given into other bioinformatics tools. The comparison of the
classification results is illustrated in Figure 16.
The features of the existing tools are compared with our proposed sequence miner
tool. Table.1. shows the salient features of the proposed tool that are compared with
other existing tools.
Figure 15 Visualization of the Protein Structure
16. T. Marimuthu and V. Balamurugan
http://www.iaeme.com/IJCIET/index.asp 32 editor@iaeme.com
Figure 16 Accuracy of the Classification Results
Table.1. Comparative Analysis
Name of the
Tool
Input Type Compatibility Visualization Functionality
Medusa text file Not compatible
with other
visualization
tools
2- Dimensional
Representation
Text search
through regular
expression
Cytoscape text file Load graphs 2- Dimensional
Representation
Zoom in and
Zoom out
Osprey text file Different text
formats, grid
2- Dimensional
Representation
Gene Ontology
ProViz text file, Image Graphs 2- Dimensional
Representation
Sub graphs
Ondex text file Data supports
with other
formats
2- Dimensional
Representation
Filter
PATIKA XML Format Data supports
with other
formats
2- Dimensional
Representation
Data integration
Sequence
Miner
Text file,
Sequence data,
XML
Compatible with
other file formats
2- Dimensional
and 3-
Dimensional
Representation
Classification,
Protein structure
prediction, Gene
Ontology
5. CONCLUSION
The proposed tool “Sequence Miner” is a novel approach designed to perform
sequence analysis through the traditional methods such as LCS, Pairwise alignment
(Local, Global and Multiple), Pattern matching algorithms, and Phylogenetic tree
construction. The classification results of this work clearly exhibit the evolutionary
relationship of dengue virus serotypes from the existing serotypes. Therefore, there is
the chance to the presence of DEN3 or DEN4 in the recently discovered DEN5
serotype. There is no evidence to find the structure of DEN5, however the E and M
proteins of DEN5 may be associated with the existing serotypes. The proposed bio-
computational model will be helpful to make the confirmation of the toxic proteins
presence in the recently discovered virus serotype. On the whole, the relationship
between dengue serotypes predicted via the proposed tool will definitely help the
17. A Novel Bio-Computational Model for Mining the Dengue Gene Sequences
http://www.iaeme.com/IJCIET/index.asp 33 editor@iaeme.com
biotechnologists and drug designers to move one step forward in discovering an
effective vaccine for dengue.
REFERENCES
[1] Ahdesmaki, M., Lahdesmaki, H., Yli-Harja, O., “Robust Fisher’s Test for
Periodicity Detection in Noisy Biological Time Series”, in the proc. of IEEE Int.
Workshop on Genomic Signal Processing and Statistics, Tuusula, FINLAND,
Vol. No 6(3), pp. 175-181, 2007.
[2] Bioinformatics Educational Resources Documentation (online), European
Bioinformatics Institute United Kingdom. Available:
http://www.ebi.ac.uk/2can/tutorials/protein/align.html
[3] Breitkreutz : “Osprey: a network visualization system”, Int. Journal of Genome
Biology, Vol. 4(3), 2003.
[4] Demir, OB., Dogrusoz, U., Gursoy, A., Nisanci, G., Cetin-Atalay, R., Ozturk, M.,
“PATIKA: An Integrated Visual Environment for Collaborative Construction and
Analysis of Cellular Pathways”, Int. Journal of Bioinformatics, Vol.18, pp.996-
1003, 2002.
[5] FASTA Format Description (online), NGFN-BLAST. Available at:
http://ngfnblast.gbf.de/docs/fasta.html
[6] Fruchterman, TMJ., Reingold, EM., “Graph Drawing by Force-Directed
Placement, Software, Practice and Experience”, First Edition, John Wiley Ltd.,
pp. 1129-1164, 1991.
[7] Hermjakob, H., Montecchi-Palazzi, L., Lewington, C., Mudali, S., Kerrien, S.,
“IntAct: an open source molecular interaction database”, Int. Journal of Nucleic
Acids Research, Vol. 21(2), pp.452-455, 2004.
[8] Hooper, SD., Bork, P., “Medusa: a simple tool for interaction graph analysis”,
Int. Journal of Bioinformatics, Vol.21(24), pp. 4432-4433, 2005.
[9] http://www.thehindu.com/antimosquitos/
[10] Iragne, F., Nikolski, M., Mathieu, B., Auber, D., Sherman, D., “ProViz: protein
interaction visualization and exploration”, Int. Journal of Bioinformatics, Vol.
21(2), pp.272-274, 2005.
[11] Lenzerini, M., “Data Integration: A Theoretical Perspective”, Pipeline Open Data
Standards, John Wiley Ltd., pp.243-246, 2002.
[12] Maglott, D., Ostell, J., Pruitt Kim, D., Tatusova. T., “Entrez Gene: gene-centered
information at NCBI”, Int. Journal of Nucleic Acids Research, Vol. 33, pp.54-
D58, 2005.
[13] Shannon, P., Markiel, A,, Ozier, O., Baliga, NS., Wang, JT., Ramage, D.,
“Cytoscape: a software environment for integrated models of biomolecular
interaction networks”, Int. Journal of Genome Research, Vol.13(11), pp. 2498-
2504, 2005.
[14] Sukmal, F., Benediktus,Y., Hidayat, A., “Molecular Surveillance of Dengue
VirusSerotype-1”, Int. Journal on Molecular Biology, Vol. No. 2(1) pp. 345-349,
2013.
[15] www.thebiogrid.org.