1) The document outlines the tasks, tools, and topics explored by Vipul Divyanshu during a summer internship at India Innovation Labs, including data analytics on a medium-sized database and building a recommender engine.
2) Key tools explored include Mahout for machine learning algorithms, Hadoop for distributed processing, and Rush Analyzer (with KNIME) for data visualization and analytics.
3) Vipul implemented recommendation engines including user-based, item-based, and SlopeOne recommenders and evaluated performance using recommender evaluators.
Talk 1: Google App Engine Development: Java, Data Models, and other things you should know
(Navin Kumar, CTO of Socialwok.com) http://www.linkedin.com/pub/navin-kumar/3/100/a08
Description: Google AppEngine is a cloud architecture designed to run and scale your own web applications. It makes it easy to develop applications, and with the introduction of Java language support, allows you to develop using the standard servlet development model and along with GWT allows end-to-end Java development of powerful Ajax-based web applications. Here we will describe the tips and tricks that are used to develop Socialwok, a rich,on-demand enterprise microblogging platform, which is deployed on Google AppEngine using the Java language support. Finally we will introduce a neat example that illustrates some of the tricks that can be used on AppEngine to develop your own applications.
So, you need a Content Management System on the .NET framework. While your business might spend wheelbarrows of money on a platform that is powerful and extensible, your personal site would abandon extensibility for a free, open-source solution. But what if we had an option that was free and powerful and extensible? We do: Orchard CMS. Since we already know that Orchard is free, in this session we will discuss the power of Orchard’s CMS engine. You will learn how to build new modules for the Orchard platform, allowing you to extend functionality as you see fit to meet the needs of your site, your business, and customers.
Professional Services Analytics Experience Summer InternshipAxcess Financial
Discover a career as an analyst with Nielsen! As an Analyst within our Professional Services Analytics Experience (PSAE), you’ll be working with our several hundred clients, like Coca-Cola, General Mills, SC Johnson, Google, Amazon, Target, Unilever, Johnson & Johnson, and P&G, to name a few.
Talk 1: Google App Engine Development: Java, Data Models, and other things you should know
(Navin Kumar, CTO of Socialwok.com) http://www.linkedin.com/pub/navin-kumar/3/100/a08
Description: Google AppEngine is a cloud architecture designed to run and scale your own web applications. It makes it easy to develop applications, and with the introduction of Java language support, allows you to develop using the standard servlet development model and along with GWT allows end-to-end Java development of powerful Ajax-based web applications. Here we will describe the tips and tricks that are used to develop Socialwok, a rich,on-demand enterprise microblogging platform, which is deployed on Google AppEngine using the Java language support. Finally we will introduce a neat example that illustrates some of the tricks that can be used on AppEngine to develop your own applications.
So, you need a Content Management System on the .NET framework. While your business might spend wheelbarrows of money on a platform that is powerful and extensible, your personal site would abandon extensibility for a free, open-source solution. But what if we had an option that was free and powerful and extensible? We do: Orchard CMS. Since we already know that Orchard is free, in this session we will discuss the power of Orchard’s CMS engine. You will learn how to build new modules for the Orchard platform, allowing you to extend functionality as you see fit to meet the needs of your site, your business, and customers.
Professional Services Analytics Experience Summer InternshipAxcess Financial
Discover a career as an analyst with Nielsen! As an Analyst within our Professional Services Analytics Experience (PSAE), you’ll be working with our several hundred clients, like Coca-Cola, General Mills, SC Johnson, Google, Amazon, Target, Unilever, Johnson & Johnson, and P&G, to name a few.
Jumpstart your career with the world’s most in-demand technology: Hadoop. Hadooptrainingacademy provides best Hadoop online training with quality videos, comprehensive
online live training and detailed study material. Join today!
For more info, visit: http://www.hadooptrainingacademy.com/
Contact Us:
8121660088
732-419-2619
http://www.hadooptrainingacademy.com/
Hadoop Training, Enhance your Big data subject knowledge with Online Training without wasting your time. Register for Free LIVE DEMO Class.
For more info: http://www.hadooponlinetutor.com
Contact Us:
8121660044
732-419-2619
http://www.hadooponlinetutor.com
Hadoop is one of the booming and innovative data analytics technology which can effectively handle Big Data problems and achieve the data security. It is an open source and trending technology which involves in data collection, data processing and data analytics using HDFS (Hadoop Distributed File System) and MapReduce algorithms.
Michael Choi's process for designing web application(s), including which programming language to use, when to use Node.js, when to use a light-weight framework vs a heavy MVC framework, how to set up git for collaboration based on complexity of the project, how a tool like Jenkins can be used for continuous integration, continuous delivery, and continuous deployment, where to host the data, what services to use for orchestrating containers or servers.
Maven: Managing Software Projects for Repeatable ResultsSteve Keener
This presentation explores the features of Maven 2, a software build tool that has achieved prominence in the Java development arena. Find out how Maven can be leveraged to automate key project tasks.
Hadoop is one of the booming and innovative data analytics technology which can effectively handle Big Data problems and achieve the data security. It is an open source and trending technology which involves in data collection, data processing and data analytics using HDFS (Hadoop Distributed File System) and MapReduce algorithms.
This is a presentation I did for the Cedar Rapids .NET User Group (CRineta.org). I also presented it at work (Fiserv Insurance Solutions - now StoneRiver) for fellow developers.
this is a colloquium report. check it and do changes according to your needs. this report prepared with the help of websites, reading books, searching articles on web.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Jumpstart your career with the world’s most in-demand technology: Hadoop. Hadooptrainingacademy provides best Hadoop online training with quality videos, comprehensive
online live training and detailed study material. Join today!
For more info, visit: http://www.hadooptrainingacademy.com/
Contact Us:
8121660088
732-419-2619
http://www.hadooptrainingacademy.com/
Hadoop Training, Enhance your Big data subject knowledge with Online Training without wasting your time. Register for Free LIVE DEMO Class.
For more info: http://www.hadooponlinetutor.com
Contact Us:
8121660044
732-419-2619
http://www.hadooponlinetutor.com
Hadoop is one of the booming and innovative data analytics technology which can effectively handle Big Data problems and achieve the data security. It is an open source and trending technology which involves in data collection, data processing and data analytics using HDFS (Hadoop Distributed File System) and MapReduce algorithms.
Michael Choi's process for designing web application(s), including which programming language to use, when to use Node.js, when to use a light-weight framework vs a heavy MVC framework, how to set up git for collaboration based on complexity of the project, how a tool like Jenkins can be used for continuous integration, continuous delivery, and continuous deployment, where to host the data, what services to use for orchestrating containers or servers.
Maven: Managing Software Projects for Repeatable ResultsSteve Keener
This presentation explores the features of Maven 2, a software build tool that has achieved prominence in the Java development arena. Find out how Maven can be leveraged to automate key project tasks.
Hadoop is one of the booming and innovative data analytics technology which can effectively handle Big Data problems and achieve the data security. It is an open source and trending technology which involves in data collection, data processing and data analytics using HDFS (Hadoop Distributed File System) and MapReduce algorithms.
This is a presentation I did for the Cedar Rapids .NET User Group (CRineta.org). I also presented it at work (Fiserv Insurance Solutions - now StoneRiver) for fellow developers.
this is a colloquium report. check it and do changes according to your needs. this report prepared with the help of websites, reading books, searching articles on web.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
Vipul divyanshu mahout_documentation
1. Data Analytics
Project Documentation
Vipul Divyanshu
IIL/2012/14
Summer Internship
Mentor: Saish Kamat
India Innovation Labs
Tasks at hand:
*Data Analytics on a Medium Size Data Base
*Building an Recommender Engine for products
Tools and topics Explored:
Mahout
Root
Hadoop
Data Rush
Rush Analyser (with KNIME)
Google Analytics engine
Analysis of the tools and what was explored:
Mahout:Mahout is an open source machine learning library from Apache. The
algorithmsit implements fall under the broad umbrella of machine learning or collective
Intelligence.
Mahout currently has:
Collaborative Filtering
User and Item based recommenders
K-Means, Fuzzy K-Means clustering
Mean Shift clustering
Dirichlet process clustering
Latent Dirichlet Allocation
Singular value decomposition
2. Parallel Frequent Pattern mining
Complementary Naive Bayes classifier
Random forest decision tree based classifier
High performance java collections (previously colt collections)
The fact that mahout has this many features and sub tools and libraries to work with,
it is the best suited tool for the self-designed data analytics programs.
And mahout also has core libraries are highly optimized to allow for good
performance also for non-distributed algorithms.
NOTE: For a well understanding of Mahout, the book ‗Mahout In action‘ is suggested.
ROOT:It is an object-oriented framework aimed at solving the
dataanalysis challenges of high-energy physics.
Below, you can find a quick overview of the ROOT framework:
Save data. You can save your data (and any C++ object) in a compressed
binaryform in a ROOT file. The object format is also saved in the same file.
ROOT
provides a data structure that is extremely powerful for fast access of huge
amounts of data - orders of magnitude faster than any database.
Access data. Data saved into one or several ROOT files can be accessed
from your PC, from the web and from large-scale file delivery systems used
e.g. in the GRID. ROOT trees spread over several files can be chained and
accessed as a unique object, allowing for loops over huge amounts of data.
Process data. Powerful mathematical and statistical tools are provided to
operate on your data. The full power of a C++ application and of parallel
processing is available for any kind of data manipulation. Data can also
be generated following any statistical distribution, making it possible to
simulate complex systems.
Show results. Results are best shown with histograms, scatter plots,
fitting functions, etc. ROOT graphics may be adjusted real-time by few
mouse clicks. High-quality plots can be saved in PDF or other format.
Interactive or built application. You can use the CINT C++ interpreter or
Python for your interactive sessions and to write macros, or compile your
program to run at full speed. In both cases, you can also create a GUI.
Link to know more about root: http://root.cern.ch/drupal/
Link for ROOT user‘s guide: http://root.cern.ch/download/doc/ROOTUsersGuide.pdf
Constrains of ROOT:
What was found was that it is concentrates more on displaying and the graphical
presentation of the collected data and on the representation of computed
(processed) result in the form of canvas, histograms, TGraphs. This can be used in later
point of time to present the processed data in a well-defined and interactive manner.
3. Screenshot:
HADOOP:
Hadoop is an open source framework for writing and running distributed
applications that process large amounts of data on different networks.
Key distinctions of Hadoop are:
Accessible—Hadoop runs on large clusters of commodity machines or on cloud
computingservices such as Amazon‘s Elastic Compute Cloud (EC2).
Robust—because it is intended to run on commodity hardware, Hadoop is architected
withthe assumption of frequent hardware malfunctions. It can gracefully handle most
suchfailures.
Scalable—Hadoop scales linearly to handle larger data by adding more nodes to the
cluster.
Simple—Hadoop allows users to quickly write efficient parallel code.
Link to explore more in Hadoop: http://hadoop.apache.org/
NOTE: For a well understanding of hadoop, the book ‗Hadoop In action‘ is suggested.
4. Setting Up Mahout development environment in Eclipse:
NOTE: The following explanation is for Ubuntu (Linux) OS .we can even implement
it on any other OS such as windows.
PREREQUIREMENTS:
1. Java SDK 6u23 x64
2. Maven 3.0.2
3. ANY UPDATED MAHOUT LIBRARY
4. IDE(I had used eclipse)
5. CYGWIN (in case of windows OS)
Running your first sample code:
Once all the above requirements are met we are ready to execute our
first sample code.
Step 1:
At first, start Eclipse and create a workspace. We takeit ―UsersVipulworkspace‖ for
the present.
Extract the source of Mahout below the workspace. Itis
―UsersVipulworkspacemahout-distribution-0.4″ for the present.
Convert Maven project of Mahout into Eclipse project with the belowcommand.
cd UsersVipulworkspacemahout-distribution-0.4
mvn eclipse: eclipse
Now set the classpath variable M2_REPO of Eclipse to Maven2 localrepository.
mvn -Declipse.workspace= eclipse: add-maven-repo
But ―Maven – Guide to using Eclipse with Maven 2.x‖ says ―Issue: The
command does not work‖. So set it in Eclipse directly.
Open Window > Preferences > Java > Build Path > Classpath Valirables
from Eclipse‘s menu.
Press ―New‖ and Add Name as ―M2_REPO‖ and Path as Maven 2
repository path (its default is .m2/repository at your user directory).
Finally import the converted Eclipse project of Mahout.
Open File > Import > General > Existing Projects into Workspace from
Eclipse menu.
Select the project directory UsersVipulworkspacemahoutdistribution-0.6 and all projects.
NOTE: Now you need to have your first code to be implemented
ready. If so proceed to Step 2.
Step 2:
At first, generate a Maven project for sample codes on the Eclipse
workspace directory.
$ cd Users/Vipul/workspace
$ mvn archetype: create -DgroupId=mia.recommender -
DartifactId=recommender
Do the following.
Delete a generated Skelton code src/main/App.java and copy the
code into src/main/java/mia/recommender of the ‗recommender‘
project.
Convert the Maven project into Eclipse project.
$ cd Users/Vipul/workspace/recommender
$ mvn eclipse: eclipse
Import the project into Eclipse.
Open File > Import > General > Existing Projects into Workspace
from Eclipse menu and select the ‗recommender‘ project.
Then the ‗recommender‘ project is available on Eclipse workspace,
but all classes have errors because of no Mahout Library reference.
5. Right click the „recommender‟ project, select Properties > Java Build Path >
Projects from pop-up menu and click „Add‟ and select the below Mahout projects.
mahout-core
mahout-examples
mahout-taste-webapp
mahout-math
mahout-utils
Then only 4 errors remain.
6. Hence they are conflicts with updated APIs, these error correction need to modify codes.
For example, open mia.recommender.ch03.IREvaluatorBooleanPrefIntro2 and press ctrl+1
at error line in it.
This error says that the code does not catch or declare a exception of
TasteException which NearestNUserNeighborhood‟s constructor throws. So you
can choise whichever you like a solution in the pop up menu. Others as well.
The classes which has main() function can be executed on Eclipse.
For example, select mia.recommender.ch02.RecommenderIntro and click Run >
Run in Eclipse‟s menu (or may press ctrl+F11 insted). Then It throws an
exception as „Exception in thread “main” java.io.FileNotFoundException:
intro.csv‟.
To make it read a sample data file „intro.csv‟ in src/mia/recommender/ch02, click
Run > Run Configurations in Eclipse‟s menu and select the configuration of
7. RecommenderIntro which is created by the above execution. Then set
mia/recommender/ch02 to Working directory in Arguments tab(see the below
figure). Click “Workspace…” button and select the directory.
Then it outputs a result like “RecommendedItem[item:104, value:4.257081]“.
If you want to make a project, repeat from Maven project creation.
8. RECOMMENDATION ENGINE:
Recommendation isall about predicting patterns of taste, and using them to discover
new and desirable things you didn‘t already know about.We have many types of
recommender like:
GenericUserBasedRecommender
GenericItemBasedRecommender
SlopeOneRecommender
SVDRecommender
KnnItemBasedRecommender
Well I had implemented the code for the first three but with time in hand the
other two and some more can be implemented.
NOTE: For every recommender to feed the data to it we need a file normally of
type .csv and don‘t forget to place it in the same folder in which we have our
pom file of the current project being build.
THE USER BASED RECOMMENDATION ENGINE
All the required details of the user based recommender engine are given in detail in the
book which I had mentioned before. The output of my recommender is shown below:
The output if the above code can be observed in the ellipse.
9. THE ITEM BASED RECOMMENDATION ENGINE:
It is similar to that of the user based recommendation engine the only difference is that
it finds the similarity between the item instead of users.
Note: Due to the above reason it is more suited in the case when we is a fast growing list
of users and a slower growing product or item list.
The output of the Item based recommender code is:
THE SLOPE-ONE RECOMMENDATION ENGINE:
It is similar to that to of item based recommendation engine but has a pre-processing
state and the output is on the basis of the relation between the different items.
The output of my code is:
10. THE EVLUATOR FOR RECOMMENDATION ENGINE:
There are many possible ways to evaluate the performance of an the recommender
engine, I have explored the following:
RecommenderIRStatsEvaluator
AverageAbsoluteDifferenceRecommenderEvaluator
RMSRecommenderEvaluator.
Well I had implemented the first two of them
AVERAGEABSOLUTEDIFFERENCERECOMMENDEREVALUATOR
It takes the a part of data as test data and rest as training data and recommends items
for our test data and latter is matched with the real values of the test data. The output
for my code is:
12. Note: To test the above codes on a larger scale we can download the
Input files for them from: http://www.grouplens.org/node/12
Mahout is still in development stage and still many fields can be explored
like clustering, network pattern learning and classification.
The Hadoop could be used with mahout to implement a cluster and map-
reduce to receive data.
Rush Analyser (with Knime):
This tool is also in Java and eclipse is needed. It was downloaded from the link :
http://bigdata.pervasive.com/Products/Download-Center.aspx
Is the graphical version of Data rush and is very handy in the terms of data analytics and
visualisation.
Here is a snapshot of my work where I have loaded the 10K movie rating data
downloaded from the test data download link given.
13. In the image different nodes can be seen used to perform different operations on the
data set.
This is the parallel plot of the data set.
14. This is the scatter plot generated for the same 10K data value scattered on the 2-D plan.
By use of clustering blocks in the rush analyser the data was analysed.
Few of the blocks explored by me are:
Regression
Classifiers
Recommender
Clustering
Filters
Data from different Databases can be directly imported by the use of Data Base reader
block.
These are few of the topics explored in Rush Analyser(a interactive Datarush tool)
15. And it only the tip of the ice berg as Rush Analyser has a lot more in store to be
explored.For more info go to The given link could be referred for exploring data rush:
http://bigdata.pervasive.com/Products/Analytic-Engine-Pervasive-DataRush.aspx.
The potential of the DATA RUSHis still to be explored for the project.
Thank You IIL:
Vipul Divyanshu
IIL/2012/14