Kuan-ming Lin is interested in data mining, particularly mining biological databases, web documents, and the semantic web. He has skills in data mining techniques including machine learning, feature selection, and support vector machines. He has published papers on data integration of microarray data and structure prediction of HIV coreceptors. He hopes to continue a career in data mining and cloud computing.
Course developed by Dr. Joan E. Hughes at The University of Texas at Austin
The purpose of this class is to introduce you to the theories, assumptions, and practices underlying the use of qualitative research in education. In the tradition of survey courses, this class examines the broad history, concepts, and themes that distinguish multiple methods of qualitative research, specifically as they relate to education research. Students will study, practice, and reflect on different qualitative research methodologies and consider the components and challenges faced when engaging in qualitative research methods. Each student will design and conduct his/her own qualitative study. Issues related to data collection, negotiating access to the field, ethics, and representation will be of particular importance. While it is not assumed that you will gain a comprehensive, rich understanding of any one particular qualitative research tradition over the trajectory of the course, it is expected that upon completion you will acquire the foundational knowledge and experience to begin evaluating, selecting, and defending appropriate qualitative methods for use in your own future research projects.
Goals:
1. Understand historical background and fundamental tenets of qualitative research.
2. Understand ethical issues within qualitative research.
3. Develop a researchable question.
4. Identify the limits and affordances of qualitative research designs.
5. Develop a beginning awareness of qualitative inquiry approaches, including ethnography, case studies, narrative, postmodern, critical, and basic interpretive.
6. Engage in qualitative research activities, including: field observations, interview, coding, analysis, and report writing.
Course developed by Dr. Joan E. Hughes at The University of Texas at Austin
The purpose of this class is to introduce you to the theories, assumptions, and practices underlying the use of qualitative research in education. In the tradition of survey courses, this class examines the broad history, concepts, and themes that distinguish multiple methods of qualitative research, specifically as they relate to education research. Students will study, practice, and reflect on different qualitative research methodologies and consider the components and challenges faced when engaging in qualitative research methods. Each student will design and conduct his/her own qualitative study. Issues related to data collection, negotiating access to the field, ethics, and representation will be of particular importance. While it is not assumed that you will gain a comprehensive, rich understanding of any one particular qualitative research tradition over the trajectory of the course, it is expected that upon completion you will acquire the foundational knowledge and experience to begin evaluating, selecting, and defending appropriate qualitative methods for use in your own future research projects.
Goals:
1. Understand historical background and fundamental tenets of qualitative research.
2. Understand ethical issues within qualitative research.
3. Develop a researchable question.
4. Identify the limits and affordances of qualitative research designs.
5. Develop a beginning awareness of qualitative inquiry approaches, including ethnography, case studies, narrative, postmodern, critical, and basic interpretive.
6. Engage in qualitative research activities, including: field observations, interview, coding, analysis, and report writing.
steps in research proposal or aspect of research proposalZaryabQureshi3
this slide is all about the research methodology or research proposal in this slide you will found the steps of doing research or present your proposal.
This presentation gives effcient information as for writing a Scientific Research Paper. There is also an article which has more details regarding this topic https://essay-academy.com/account/blog/writing-a-scientific-research-paper
Web Mining Based Framework for Ontology Learningcsandit
Today, the notion of Semantic Web has emerged as a prominent solution to the problem oforganizing the immense information provided by World Wide Web, and its focus on supporting
a better co-operation between humans and machines is noteworthy. Ontology forms the major
component of Semantic Web in its realization. However, manual method of ontology
construction is time-consuming, costly, error-prone and inflexible to change and in addition, it requires a complete participation of knowledge engineer or domain expert. To address this
issue, researchers hoped that a semi-automatic or automatic process would result in faster and
better ontology construction and enrichment. Ontology learning has become recently a major
area of research, whose goal is to facilitate construction of ontologies, which reduces the effort
in developing ontology for a new domain. However, there are few research studies that attempt
to construct ontology from semi-structured Web pages. In this paper, we present a complete
framework for ontology learning that facilitates the semi-automation of constructing and
enriching web site ontology from semi structured Web pages. The proposed framework employs
Web Content Mining and Web Usage mining in extracting conceptual relationship from Web.
The main idea behind this concept was to incorporate the web author's ideas as well as web
users’ intentions in the ontology development and its evolution.
steps in research proposal or aspect of research proposalZaryabQureshi3
this slide is all about the research methodology or research proposal in this slide you will found the steps of doing research or present your proposal.
This presentation gives effcient information as for writing a Scientific Research Paper. There is also an article which has more details regarding this topic https://essay-academy.com/account/blog/writing-a-scientific-research-paper
Web Mining Based Framework for Ontology Learningcsandit
Today, the notion of Semantic Web has emerged as a prominent solution to the problem oforganizing the immense information provided by World Wide Web, and its focus on supporting
a better co-operation between humans and machines is noteworthy. Ontology forms the major
component of Semantic Web in its realization. However, manual method of ontology
construction is time-consuming, costly, error-prone and inflexible to change and in addition, it requires a complete participation of knowledge engineer or domain expert. To address this
issue, researchers hoped that a semi-automatic or automatic process would result in faster and
better ontology construction and enrichment. Ontology learning has become recently a major
area of research, whose goal is to facilitate construction of ontologies, which reduces the effort
in developing ontology for a new domain. However, there are few research studies that attempt
to construct ontology from semi-structured Web pages. In this paper, we present a complete
framework for ontology learning that facilitates the semi-automation of constructing and
enriching web site ontology from semi structured Web pages. The proposed framework employs
Web Content Mining and Web Usage mining in extracting conceptual relationship from Web.
The main idea behind this concept was to incorporate the web author's ideas as well as web
users’ intentions in the ontology development and its evolution.
The present society is considered an information society. A society where the creation, distribution, use, integration, and manipulation of digital information have become the most significant activity in all aspects. Information is producing from every sector of any society, which has resulted in an information explosion. Modern technologies are also having a huge impact. So managing this voluminous information is really a tough job. Again WWW has opened the door to connect anyone or anything within a fraction of a second. This study discussed the Semantic Web and linked data technologies and their effect and application to libraries for the handling of various types of resources.
Researcher Reliance on Digital Libraries: A Descriptive AnalysisIJAEMSJORNAL
The digital library is an information technology that is structured as a digital knowledge resource, or can be alluded to a medium that stores information for a huge scope and is teamed up with the information the board gadget equipped for showing the information or information required by the client. Digital libraries can be extensively characterized as an information stockpiling and recovery frameworks that control digital information in the media (text, pictures, sound, static or dynamic) on the web. The main aim of this study is to study the awareness and using pattern of digital library by the researchers, to analyse the influence of digital library on researchers’ efficiency, analyse the purpose of using Digital Library Consortium, decide the effect of problems and motivational components of the digital library on the users, evaluate the satisfaction level of users with coverage of journals and perspectives on training and awareness programs and propose the available resources for effective utilization of the Digital Library.
An efficient educational data mining approach to support e-learningVenu Madhav
The e-learning is a recent development that has
emerged in the educational system due to the growth of the
information technology. The common challenges involved
in The e-learning platform include the collection and
annotation of the learning materials, organization of the
knowledge in a useful way, the retrieval and discovery of
the useful learning materials from the knowledge space in a
more significant way, and the delivery of the adaptive and
personalized learning materials. In order to handle these
challenges, the proposed system is developed using five
different steps of knowledge input such as the annotation of
the learning materials, creation of knowledge space,
indexing of learning materials using the multi-dimensional
knowledge and XML structure to generate a knowledge
grid and the retrieval of learning materials performed by
matching the user query with the indexed database and
ontology. The process is carried out in two modules such as
the server module and client module. The proposed
approach is evaluated using various parameters such as the
precision, recall and F-measure. Comprehensive results are
achieved by varying the keywords, number of documents
and the K-size. The proposed approach has yielded
excellent results by obtaining the higher evaluation metric,
together with an average precision of 0.81, average
WEB MINING: PATTERN DISCOVERY ON THE WORLD WIDE WEB - 2011Mustafa TURAN
Web Mining: Pattern Discovery on the World Wide Web
The study covers auto Turkish text content gathering from most popular social media platforms (twitter, facebook, blogs, news, etc...) and auto classifying sentimentally those contents.
Data mining refers to the process of analysing the data from different perspectives and summarizing it into useful information.
Data mining software is one of the number of tools used for analysing data. It allows users to analyse from many different dimensions and angles, categorize it, and summarize the relationship identified.
Data mining is about technique for finding and describing Structural Patterns in data.
Data mining is the process of finding correlation or patterns among fields in large relational databases.
The process of extracting valid, previously unknown, comprehensible , and actionable information from large databases and using it to make crucial business decisions.
Topic Modeling : Clustering of Deep Webpagescsandit
The internet is comprised of massive amount of info
rmation in the form of zillions of web
pages.This information can be categorized into the
surface web and the deep web. The existing
search engines can effectively make use of surface
web information.But the deep web remains
unexploited yet. Machine learning techniques have b
een commonly employed to access deep
web content.
Under Machine Learning, topic models provide a simp
le way to analyze large volumes of
unlabeled text. A "topic" consists of a cluster of
words that frequently occur together. Using
contextual clues, topic models can connect words wi
th similar meanings and distinguish
between words with multiple meanings. Clustering is
one of the key solutions to organize the
deep web databases.In this paper, we cluster deep w
eb databases based on the relevance found
among deep web forms by employing a generative prob
abilistic model called Latent Dirichlet
Allocation(LDA) for modeling content representative
of deep web databases. This is
implemented after preprocessing the set of web page
s to extract page contents and form
contents.Further, we contrive the distribution of “
topics per document” and “words per topic”
using the technique of Gibbs sampling. Experimental
results show that the proposed method
clearly outperforms the existing clustering methods
.
Topic Modeling : Clustering of Deep Webpagescsandit
The internet is comprised of massive amount of information in the form of zillions of web pages.This information can be categorized into the surface web and the deep web. The existing search engines can effectively make use of surface web information.But the deep web remains unexploited yet. Machine learning techniques have been commonly employed to access deep web content.
Under Machine Learning, topic models provide a simple way to analyze large volumes of unlabeled text. A "topic" consists of a cluster of words that frequently occur together. Using
contextual clues, topic models can connect words with similar meanings and distinguish between words with multiple meanings. Clustering is one of the key solutions to organize the deep web databases.In this paper, we cluster deep web databases based on the relevance found among deep web forms by employing a generative probabilistic model called Latent Dirichlet
Allocation(LDA) for modeling content representative of deep web databases. This is implemented after preprocessing the set of web pages to extract page contents and form
contents.Further, we contrive the distribution of “topics per document” and “words per topic”
using the technique of Gibbs sampling. Experimental results show that the proposed method clearly outperforms the existing clustering methods.
1. Research Statement
Kuan-ming Lin
January 2007
Since I took the Database course by Prof. Jun Yang at Duke Computer Science
Department, I have been attracted by the excitement of data mining. Broadly
speaking, I am interested in finding useful patterns in large stores of data which
can be handled only by computers. In earlier years, such data of large size were
collected by business or organizations and kept in proprietary data warehouses.
Therefore, many data mining techniques were specialized to tackle individual
cases, thus their generalizability remained questionable. Nowadays, as the
Internet has become the largest worldwide information storage, it is possible for
the public to perform data analysis on public databases, which also motivates my
career choice. In particular, I am curious in mining three categories of public data,
namely biological databases[3][5], Web documents[1], and the Semantic Web.
A number of repositories of biological data, particularly via high throughput
methods representing different aspects of biological processes, have been
publicized in the past decade. The data collectively are believed to provide more
information than what they were explained individually. For example, we could
find functions to unknown genes by integrative analysis on multiple microarrays,
genome sequences, NMR and X-ray data, etc. Then, how to extract
complementary information across heterogeneous datasets becomes appealing
for identifying new biological relations. Since there is little known underlying
theoretical model available for biological datasets, this is very challenging and
worth solving.
As the rapid growing of information on the Web, retrieving interesting
information from Web documents, such as what Google and Facebook have done,
has high commercial potential value. Among various multimedia mining
challenges, I believe Web text mining research, benefited from traditional
information retrieval, should provides new opportunities, in that Web documents
are by nature highly unstructured yet interconnected through hyperlinks.
Designing new techniques for mining Web document could lead to great
applications, so I will be happy to be involved in related groups.
The Semantic Web is a solution to extend the current Web content to provide
intelligent mining and analysis on the Web. Unlike the currently prevalent HTML
2. and XML, Semantic Web languages explicitly define associations between
semantics, or meaning, with the Web document content. The semantics are
formally specified in well-designed ontology; therefore, they can be either shared
via the Internet or refined for local needs. The current standard for the Semantic
Web exists: OWL, a W3C Recommendation, for example, yet integrations from
heterogeneous Semantic Web groups and ontology are still in its infantry. Besides,
the current semantic reasoning and ontology updating tools are still far from
satisfactory. Therefore, I would like to contribute to this new field with my
already cumulated knowledge on data mining and database integration.
Throughout the undergraduate and master studies, I have developed skills in the
data mining subfields, ranging from theoretical studies like approximation
algorithms for integrative data mining, to machine learning issues like feature
selection and discovery and SVM[2][4], to applications in bioinformatics and Web
mining. A few of my studies focused on bioinformatics and statistical learning,
and I have gained experiences on Web text mining through my research
internship and a summer school lecture series in Taiwan. Meanwhile, I have
published two papers at Duke[3][5].
To sum up, I hope to continue my career in data mining related software fields, in
Taiwan or other places where cloud computing and analysis are encouraged.
References
[1] C.-C. Huang, K.-M. Lin, and L.-F. Chien. Automatic Training Corpora
Acquisition through Web Mining. In Proceedings of The 2005 IEEE/WIC/ACM
International Conference on Web Intelligence
[2] K.-M. Lin. Reduction Techniques for Support Vector Machines. Master’s thesis,
National Taiwan University, Taipei, Taiwan, 2002. Adviser: Chih-Jen Lin.
[3] K.-M. Lin and J. Kang. Exploiting Inter-gene Information for Microarray Data
Integration. In Proceedings of The ACM Symposium on Applied Computing (SAC),
March 2007.
[4] K.-M. Lin and C.-J. Lin. A Study on Reduced Support Vector Machines. IEEE
Transactions on Neural Networks, 14(6): 1449–1559, 2003.
[5] K.-M. Lin, J. Zheng, and J. Zhang. Automatic Structure Prediction of HIV
Coreceptor CCR5. In Proceedings of Bioinformatics in Taiwan Symposium (BIT),
Taichung, Taiwan, September 2006.