Sunteți pe pagina 1din 10

ANALYSIS AND COMPARISON OF MULTI - AGENT DATA MINING TOOLS

*1
Mr. Sivabalan M, *2 Dr. Arulanandam K.,
*1
Guest Lecturer, PG and Research Department of Computer Science, Government Thirumagal Mills
College, Gudiyattam, Tamilnadu, India.
*2
Assistant Professor & Head of Department, PG and Research Department of Computer Science,
Government Thirumagal Mills College, Gudiyattam, Tamilnadu, India
---------------------------------------------------------------------***---------------------------------------------------------
Abstract: INTRODUCTION

The main objectives of this research is to Multi Agent Systems (MAS) often deal with
conduct a comparative study of the architectural complex Applications that require distributed
designs of Multi-Agent development toolkits, problem solving. In many applications the
studying the individual methodology used in the individual and collective behavior of the agents
designing the architectures within Multi-Agent depends on the observed data from distributed
development environments. The study aims to sources. In a typical distributed environment
research widely into the various design analyzing distributed data is a non-trivial problem
methodology used in implementing various Multi- because of many constraints such as limited
Agent design environments, analysis different bandwidth (e.g. wireless networks), privacy-
aspects relating to the components of the sensitive data, distributed compute nodes, only to
development tools.The FIPA (Foundation for mention a few. The field of Distributed Data Mining
Intelligent Physical Agents) a part of the analysis (DDM) deals with these challenges in analyzing
and evaluation of the specification providing the distributed data and offers many algorithmic
necessary requirements for agent systems. solutions to perform different data analysis and
mining operations in a fundamentally distributed
This analysis is to compare and contrast
manner that pays careful attention to the resource
each agent against and other properties. To make an
constraints. Since multi-agent systems are also
effective comparison as well as evaluate individual
distributed systems, combining DDM with MAS
agent tools using Multi-Agent development tool to
for data intensive applications is appealing.
compare their performance and a high level
understanding of a particular agent–oriented The aim is to understand the paradigm of
development methodologies put it use. The main agent oriented development tools, studying the
objective can allow in the analysis agent methodology used in Multi-Agent design models
architecture entirely. The analysis of Multi-Agent and the applications developed. This investigation
development tools aims to give insight into the will be carried out in the form of a comparative
many theoretical aspects of software engineering study that will include agent development tools to
and the drive for greater understanding of software understand each agent by carrying out a detailed
engineer methodology as well as the concept analysis of each Multi-Agent System (MAS)
behind agent design for an agent development building tools.
environment such as Multi-Agent systems.
The objective of this study is widely to fill in the
Keywords: Distributed Data Mining, Agent gap in knowledge as to agent-oriented
mining, Multi Agent System (MAS),Weka, developments tools and the methodology each has
Knime, Tanagra. chosen to use individually then carrying out a
comparison on many levels of its abstraction,
making an evaluation from a comparison
collectively reviewing the agent development depending on the environment of the system it is
toolkit thoroughly as well as effectiveness. visiting. The agents can interact with other agents
or with the user, as needed [15]. But FTP agents do
Existing Agent Simulation Tools not require any user interaction—based on push
technology, they can move from system to system,
In this section we would like to discuss the popular respond to events, and perform tasks according to
agent simulation tools like Aglets, JATLite, FTP criteria predefined by the user. An Agent Manager
Software Agent and Voyager. Aglets are Java-based is responsible for launching the agent. Voyager,
autonomous agents developed by IBM, which from Object Space, Inc., is an agent-enhanced
provide the basic capabilities required for mobility Object Request Broker (ORB) coded in Java. An
and has a globally unique name. A travel itinerary is ORB provides the capability to create objects on a
used to specify the destinations to which the agent remote system and invoke methods on those objects
must travel and what actions it must take at each [16]. Voyager augments the traditional ORB with
location. In order for an aglet to run on a particular agent capabilities. Voyager agents have mobility
system, the target system must be running an aglet and autonomy which is provided in the base class,
host application which provides a platform-neutral Agent. An Agent can move itself from one location
execution environment for the aglet. The aglet to another and can leave behind a forwarding
workbench includes a configurable Java security address with a secretary so that future message scan
manager. Aglets can communicate using a be forwarded to its new location. Specialized
whiteboard that allows agents to collaborate and agents, called Messengers, are used to deliver
share information asynchronously. Synchronous messages.
and asynchronous message passing is also Need of Agents:
supported for aglet communication. Aglets are
streamed using standard Java serialization or In Data mining perspective, an agent can be defined
externalization. A network agent class loader is as an object whose behavior is described by a
supplied which allows an aglet’s byte code stream "script", with its own calculation ways, and can
and state to travel across a network [13]. Java Agent move from place to place to communicate with
Template Lite (JATLite) is a set of light-weight other agents. With its "script", the agent is able to
Java packages being developed at Stanford follow a life behavior that will be instilled at the
University that can be used to build multiagent time of implementation and that will allow him to
systems. It is a layered architecture which provides have as main feature to be fully autonomous. One
a different communication protocol at each layer. of the discriminate characteristics of the agents is
The JATLite framework is intended for developing the representation and reasoning on the
typed-message, autonomous agents that environment (the external world and other agents),
communicate using a peer-to-peer protocol. Both based on this feature, we find two different classes,
synchronous and asynchronous message passing are which are Cognitive agents, Reactive agents.
supported. Messages can be delivered through
polling or message queuing. The framework A cognitive agent is an agent that has an explicit
provides additional security which checks the agent representation of its purpose and its environment.
name and password for a more secure connection The actions it performs to achieve its goal are the
[14]. FTP Software Agent Technology is Java-based result from a reasoning on the state of the
software designed to manage heterogeneous environment. Usually a cognitive system includes a
networks across the Internet using agent small number of agents; each is similar to a more or
technology. The agents are autonomous and mobile, less complex expert system. In this case we speak
and can move to any system in the network which of high granularity agent.
has an Agent Responder installed. As the agent
moves from system to system, its tasks may change,
Challenges in Distributed Data mining and 3.1.1Weka
Mining Agents:
Weka is the tool most commonly used due to its
Agents can enhance data mining through involving vast functionality and supported features. This java
agent intelligence in data mining systems, while an based data mining tool provides user with both GUI
agent system can benefit from data mining via and simple CLI for performing and managing tasks
extending agents’ knowledge discovery capability to be performed. It supports all data mining tasks
[12]. Nevertheless, the agent mining interaction from preprocessing, classification, and clustering to
symbiosis cannot be established if mutual issues are visualization and feature selection.[9]
not solved [13]. These mutual issues involve
fundamental challenges hidden on both sides and
particularly within the interaction and integration.
Issues in agent-mining interaction highlighting the
existence of mutual issues. Mutual issues
constraining agent-mining interaction and
integration consist of many aspects such as
architecture and infrastructure, constraint and
environment, domain intelligence, human
intelligence, knowledge engineering and
management, and nonfunctional requirements.
Figure 1: Weka GUI
Architecture and infrastructure Data mining always
3.1.2 Knime
faces a problem in how to implement a system that
can support those brilliant functions and algorithms KNIME (Konstanz Information Miner) is an open
studied in academia. API workflow based data mining tool that provides
easy accessibility to new nodes to be added into the
Nonfunctional requirements Nonfunctional requests
workflow. It provides its user with the GUI which
are essential in real-world mining and agent
aid with the simplification of workflow generation
systems. The agent-mining simians may more or
by the user. It also provides with features to modify
less address nonfunctional requirements such as
a particular node accordingly a nd execution of
efficiency, effectiveness, action ability, and user and
partial data flow[10].
business friendliness.
3.1.3 Tanagra
Constraint and environment both agent and mining
systems need to interact with the environment, and This extension of SIPINA provides the users an
tackle the constraints surrounding a system [15]. In easy to use interface for the analysis of either real
agent communities, environment could present or artificial data. It allows the researchers to easily
characters such as openness, accessibility, add their own data mining research methodology or
uncertainty, diversity, temporality, spatiality, and/or any newly identified data mining processing
evolutionary and dynamic processes. These factors technique and also supports byproviding them with
form varying constraints on agents and agent architecture and a means to compare their
systems. methodology performances. It provides the
beginners or naives with a platform where they can
Human intelligence both agent and mining need to
carry out their experimental procedures[6].
consider the roles and components of human
intelligence. Many roles may be better played by
humans in agent-mining interaction.

Data Mining Tools


nearest neighbor it is 58.44% to 94.11%, for C4.5 it
is 86.12% to 92.13%.The KNIME does not provide
any implementation for the ZeroR and OneR
algorithms which proved to be a shortcoming in the
accuracy of the tool. K-nearest neighbor is not
compatible with the data type for Audiology but
produced an accuracy of 41.54% for instances in
Zoo data set. C4.5 achieved the accuracy between
62.55% and 89.36%.Tanagra, similar to KNIME,
does not provide with the implementation of ZeroR
Figure 2: Tanagra explorer for result
and OneR algorithm. K-nearest neighbor algorithm,
visualization
like the case with other tools, does not conclude
3.2 Experimental Analysis over the discrete values but runs well over the Zoo
data set values. K nearest neighbor algorithm
The performance of these tools has been analyzed attains an accuracy percentage of 75.32% with
by first running them with several datasets available percentage split. C4.5 obtains result between
on UCI repository. Several different algorithms for 74.38%and 80.51%.
classification and clustering were implemented in
this analysis and performance of these algorithms Table 1: Accuracy for classification over
was observed. In this section a sample of the audiology data set.
experiment performed during the research is
presented and conclusion of the results of different
tools is discussed

Data Set:

Data set of (1)Audiology isused with data type


multivariate, attribute type categorical, number of
attributes 69, number of instances number of Table 2: Accuracy achieved over Zoo data set.
attributes, number of instances 226.(2) Zoo: data
type: multivariate, attribute type: categorical,
integer, number of attributes: 18, number of
instances: 101[13].

Preliminaries:

The classification was carried out of data set with


percentage split methodology of 60% for training it has been observed that Weka has successfully run
data and remaining 40% for the test data. The and implemented all the algorithms and produced
obtained measure of accuracy is used as the appropriate results for the algorithms but with
criterion for the performance analysis of the tools. lowest accuracy that of ZeroR. Though ZeroR and
OneR did not provide result, KNIME produced an
Experiment: accuracy of 89.36% with C4.5 over the zoo data.set
The k nearest neighbor algorithm did not run on the which is very close then 92.13% of that confirmed
dataset, the possible reason being the incompatible by C4.5 in Weka. Beside the non availability of
data type as the data was of discrete values. The ZeroR and OneR algorithm implementation,
accuracy measure of Weka with OneR is 42.85%, to Tanagra’s accuracy with C4.5 is satisfactory with
38.23%, for ZeroR it is 27.27% to 38.23%, for k- result between 74.38% and 80.51% which is more
stable then compared to 62.55% to 89.36% of what validation or independent validation, Weka lacks to
it is with KNIME. allow saving model in order to avoid rebuilding for
other datasets[1].
3.3Features and Functionalities:
3.3.2Knime:
3.3.1Weka:
Database system support: KNIME provides with
Database system support: Weka supports reading a great strength being able to establish database
of files from several different data bases. It also connections with any number of databases that
provide feature to import the data over internet, provides JDBC. This tool also provides with a
from web pages or from a remotely located SQL unique functionality of ports to different data
database server by entering merely the URL of sources and databases. With this there is no need to
resource. This hence allows Weka to support variety modify the SQL query. Users can use these ports in
of different data formats[9]. order to integrate from several databases and
Graphical representation:Weka provides limited modify the dimension of the database accordingly
support to the visualization of concluded data. [4].
Although the Weka API provides with various data Graphical representation:Workbench in KNIME
mining and processing methods, it lacks in the provides user with an easy way to handle different
representation of the result of processing. The functions and data flow of the process by merely
representation of the results, graphs and plots, lacks dragging and dropping new nodes which can then
in detailed representation. But that really is not of be customized as required. It also provides with its
much concern as Weka provides appreciable up gradation to improve the workbench
support for other functionalities and also the functionality.
visualization provided is sufficient to represent the
view of data on which the analyses has been KNIME can also be implemented deploying further
performed and the results of the data analyses and improvement with individual efforts over a
preprocessing. Weka also allow for available add- particular function and over the data so as to be able
on functions to be included in order to be able to to represent variety of data types and their
interface with R statistical package with improved properties [4].Its applicability varies. from
statistical analysis and representation of result[4]. execution of basic processes to the integration with
other software for enhanced visualization and
Analysis and Processing Capabilities: It allows performance. However, the manual effort required
users to use R application as can directly interact with KNIME makes it a less suitable option for
with R package. Weka also provides with separate large complex workflows.
GUI for knowledge flow, for Experimenter, used to
compare various results and explorer to analyze Analysis and Processing Capabilities: KNIME
different data sets. It also provides for creating own provides all the supported operation for data mining
filters for filtering out instances[4]. with a graphical interface where the naïve user can
import, filter and generate workflow without any
Issues:However, Weka is not better suitable option need to formulate SQL queries and code [10].On
for the large data sets as theyare roughly handled. the other hand KNIME allows experts to write their
The databases with large unstructured data are not own programming script for any new node or to
suitable as it hinders the pre-processing and download extensions from other users to attain
computing time of Weka. It can perform well and more specific functionality.
provide more accurate results with smaller
databases. This tool has limited ability to partition
dataset to training and test sets. It does facilitate to
save parameters for future application. For cross
Issues: KNIME also has limited partitioning ability databases. It is required to integrate different
like Weka but does provide the functionality to save datasets outside Tanagra and then import before
parameters and validation model for cross performing any operations. Out of the available
validation and independent validation[1]. operator, not all are applicable to every node. Some
of these often fail when applied to inappropriate
3.3.3Tanagra: location and that too without proper display of any
Database system support: Importing of different error message. Tanagra has limited partitioning
data sources is not supported in Tanagra because, as ability and does not provide to save parameter
it provides only tree representation, so it would values. It also does not support for saving
disturb the tree nature of the stream diagram. Even independent validation model [12].
it can also not read from any other data source 4.EVALUATION OF COMPARATIVE STUDY
format or integrated database directly[12].
There are conclusions drawn from the study of
Graphical representation: The visualization of these tools. The analysis of these tools has provided
various models is although not like other tools but it us with idea for the betterment of the whole data
includes several statistical measures. Tanagra is mining procedure. Even though these tools have
based on stream diagram paradigm where a user proved to be appropriate for the specific data
defines data sources, data operation and a linking domains and specific data mining tasks such as
path through the sources and operations. This path classification, clustering, etc, the shortcomings,
describes the flow similar to workflow. But Tanagra flaws or specificity of these tools have acted as a
allows the graph to be represented only as a tree so pullback from the implementation of a general
as to maintain the simplicity. However, this creates framework for data mining process. The major
a bottleneck with the performance of Tanagra as common drawback with theses tools is that their
now there can be only one data source to any processing of data, classification, clustering,
operation [4].Tanagra provides with different ways prediction and inferring of rules all is based on the
of visualization of data as in scatter plots, graphs selection of the algorithm for data mining over a
and tables. However, it is difficult to develop particular type of data set.If the selection of
graphs by specifying several parameter values as algorithm is not appropriate regarding the domain
some may not be applied to a given data set but of data then the produced patterns or predictions
should be applied to the sub-node of the data cannot be completely relied. For example, the
indirectly. This leads to the missing values for some classification over Audiology data set results with
views. Instead the visualization of data result is 84.41% correctly classified instances with Simple
more useful as high dimensional data can be Logistic while with ZeroR it results in only 27.27%.
represented in 2D.The representation of clustering If an inappropriate algorithm is used for future data
and classification result is in text[12]. value prediction then it would produce incorrect
Analysis capabilities: Tanagra can import text files results. Another issue with these tools is that the
with whitespace delimited fields. It also provides current state of art does not provide an automated
with a conversion tool for files it cannot read from mining technique. All the tasks such as
arff. This tool provides robust statistical analysis classification and clustering are performed
functionality by complementing with a range of consecutively and are specific to an application[5].
uni-and multivariate parametric and nonparametric A theoretical framework is required for
tests. It also includes correspondence analysis, implementation of unified theory where these data
principal component analysis, and the partial least mining tasks can be unified and overcome with the
squares methods.[11] shortcomings of these tools.
Issues: Tanagra cannot implement association rule
mining for large datasets. It cannot integrate
Multi agent systems are those systems which break This knowledge is then analyzed by the user if is
a domain in components and each component required by him according to his business rules[5].
performs its part of processing to achieve a There is no need to manually implement algorithms
common goal[5]. These agents are adoptive to new for classification and clustering in the whole
environment, autonomous, flexible, and easily process, however it is though required to specify
modifiable and pertains some characteristics them at the beginning of the process.
relating to artificial intelligence. These intelligent
agents handle several independent tasks. Such multi Comparison of Data Mining Tools
agent system is modular and provides robustness to
the system. A MAS based data mining tool would
provide with:

 Most suitable processing technique

 Preprocessing of new data and updating in


the database.

 A dataset of patterns obtained every time


after any new data addition and suggest with
possible knowledge that can be discovered.
Table: Comparison of Data Mining Tools
 Reduced processing time for pattern
Multi-agent systems and their applications
recognition and representation of analyzed
data. Agents and Multi-Agent Systems (MAS) have
emerged as a powerful technology to cope with the
increasing complexity of a variety of Information
Technology scenarios. We are not going to explain
the full details of the agents because these are
covered in other chapters. We are only going to
provide a basic overview.

The most widely accepted definition for the “agent”


term is that “an agent acts on behalf of someone
else, after having been authorized”. This definition
can be applied to software agents, which are
instantiated and act instead of a user or a software
Fig: Data Mining process.
program that controls them. The difficulty in
In the above depicted process, the initial stages defining an agent arises from the fact that the
(data gathering, data preprocessing, data cleansing, various aspects of agency are weighted differently,
and dataset preparation) are same as the data with respect to the application domain at hand.
mining process. However, the next stage in this Wooldridge & Jennings have succeeded in
process, the mining tasks, classification, clustering combining general agent features into the following
and visualization are unified and the process is generic abstract definition integrating all the
called Unified Data Mining Process UDMP. characteristics into the notion of an agent: “An
agent is an autonomous software entity that
In this technique of data mining, the only required -functioning continuously - carries out a set of goal-
input is a dataset in which the proceeding tasks are oriented tasks on behalf of another entity, either
performed automatically and knowledge is mined. human or software system. This software entity is
able to perceive its environment through sensors IND(D) i.e. they represent different decision
and act upon it through effectors, and in doing so, concepts. M(C) is a diagonal symmetric matrix.
employ some knowledge or representation of the
user's preferences” (Wooldridge, 1999). mij = 0

An agent may have, depending on the domain it is fx i, xj ∈ same IND(D) = {c ∈ C: f(c, xi) ≠ f(c, xj)}
situated in, some or all of the properties listed if xi, xj ∈ different IND(D)
below (Symeonidis & Mitkas, 2005): 2.2. Find the CORE from discernibility matrix:
• Autonomy (considered a must-have feature by For any c ∈ C, c ∈ CORE(C) if and only if there
many researchers in the field of agents) exists i, j, 1 ≤ j < i ≤ N such that mij = {c}. Note
that a core may be empty
• Interactivity: Reactivity or Pro-activeness
2.3. Determine the attribute set UA which user
• Adaptability prefers to emphasize. If UA is empty that means
that the user does not have preference for any
• Sociability
attribute.
• Cooperativity
2.4. Let RED = CORE ∪ UA
• Competitiveness
2.5. AR = AR – RED
• Mobility
2.6. Find attribute a in AR which has the
• Learning maximum SGF(a, RED, D)

The Rough Sets algorithm implemented in 2.7 RED = RED ∪ {ai}, AR = AR – {ai} (i =1, 2,
MASCE can be summarized as follows: …m)

This algorithm takes as input a decision table S = 2.8 If k(RED, D) = 1, then stop, otherwise go to
(U, C∪ D, V, f) and produces as output the set of step
decision rules {τij} (Attia at al., 2004).
Step 3: Generate the reduced relation by removing
Step 1: Vertical reduction: those attributes which are not in the best reduct
RED.
The vote value is calculated for all the tuples
(similar tuples are collapsed into one and their Step 4: Combine similar tuples in the reduced
number is added to the vote). relation.

Then tuples, with vote values less than the noise Step 5(a): Transform tuples in the reduced relation
filter threshold, are removed from the database into decision rules for each class in D.
table.
Step 5(b): For the same class in the reduced table,
Step 2: Horizontal reduction: Attributes reduction two tuples can be combined if the values of the
is made by calculating the best reduct RED as condition attributes differ in only one attribute, thus
follows, let all attributes be called AR and the user obtaining a more general set of decision rules. Or
preferred attributes if any be UA. Begin instead of steps 5(a) and 5(b) we can use the
following alternative method for generation of
2.1. Construct the modified discernibility matrix decision rules:
M(C): Each entry mij contains the condition
attributes whose values are not identical on both xi Step 6(a): Extract the decision rule which is called
and xj where xi, xj belong to different classes of τij as follows: τij = DesC(Xi) ⇒ DesD(Yj) such that
Xi ∩ Yj ≠ Φ for Xi ∈ C* and Yj ∈ D*.
Step 6(b): Call A rule (deterministic) if Xi ⊆ Y j i.e. give a partition of the set of values of the given
Xi ∩ Y j = Xi, otherwise a rule is nondeterministic. characteristic.
The set of all decision rules {τij} for all classes Yj
Conclusion:
∈ D * is called the decision algorithm of the
information system S. The above study was conducted by using four
algorithms over a data set: Zero Rule (ZeroR), One
/* End of the algorithm*/ The significance of an
Rule (OneR), decision tree (C4.5), and k-nearest
individual attribute {a} added to the set A with
neighbor (KNN). Tools were run over the data set
respect to the dependency between A and D
and results were observed for each algorithm.
(Decision set) is represented by significant factor
Accuracy percentage served as performance
SGF, given by:
measure. Weka was indentified a better performer
SGF (a, A, D) = k(A+{a}, D) – k(A, D) where k(A, with the specified algorithms, followed by KNIME
D) = card(POSA(D)) / card(U) . and Tanagra. This performance ranking based on
the type of data set used and how the classifier is
Decision trees mining in MASCE implemented within the tool, as task of
Decision tree induction is a well-known discipline classification is affected by so. But Weka still
in Machine Learning presented by Quinlan in 1986 proved to be better as it provided with the
(Quinlan, 1986). The basic algorithm for decision implementation of ZeroR and OneR over data types
tree induction is a greedy algorithm that constructs where other tools did not. The functionality offered
decision trees in a top-down recursive divide-and- by these tools, like API support and graphical
conquer manner. In the process of constructing a presentation along with other features aid with the
tree, the criteria of selecting test attributes selection of tool best suitable according to the usage
influences the classification accuracy of the tree. by different users. The methodologies discussed for
Presently, there are many criteria for choosing the UDMT formulation offer better approach towards
test attribute in building decision tree, such as ID3 data mining but still are left with some issues due to
(Quinlan, 1986) and C4.5 (Quinlan, 1993) which an incomplete theory for their correct formulation.
use an entropy-based measure known as The MAS suffer from the problem of selection of
information gain as a heuristic for selecting the appropriate set of algorithms for classification,
attribute. Decision trees represent a supervised clustering and visualization. The correct application
approach to classification. A decision tree is a of an algorithm as a function is also an issues
simple structure where non terminal nodes The Inductive Databases theory addresses the need
(internal) represent tests on one or more attributes for a language to design inductive queries and
and terminal (leaf) nodes reflect decision outcomes. generation of databases including both data and
The ordinary tree consists of one root, branches, patterns. All the defined patterns are then required
nodes (places where branches are divided) and to satisfy the constraints imposed upon. This results
leaves. In the same way the decision tree consists of in need of better algorithms for constraint based
nodes which stand for circles, the branches stand data mining.
for segments connecting the nodes. A decision tree
is usually drawn from left to right or beginning
from the root downwards, so it is easier to draw it.
Future Work:
The first node is a root. The end of the chain “root -
branch - node-...- node” is called “leaf”. From each Based on our result we concluded that there
internal node (i.e. not a leaf) may grow out two or is enough claim for Agent – oriented methodology
more branches. Each node corresponds with a varies regarding the different methodology applied;
certain characteristic and the branches correspond three out of the five architectures studied has
with a range of values. These ranges of values must applied the BDI theory into practices. Other method
such as O-MaSE used in Agent Tool3 and the
Cognitive method used in Courage has been applied
only within limited architectures. We found
significant performance boost that proofs our agent-
oriented methodology. Enhance comparison of
agent-oriented tools.

References:

[1]“A comparative analysis of data mining tools in agent


based systems”, Sharon Christa, K. Laxmi Madhuri,
V.Suma, research and industry incubation centre,
Dhyanchand Sagar institutions.

[2]“A Comparisonstudy between data mining tools with


some classification methods”, Abdullah H. Wahbeh,
Qasem A. Al-Radaideh, Mohammed N. Al-Kabi, and
Emad M. Al-Shawakfa.

[3]“Evaluation of Fourteen Desktop Data Mining


Tools”, Michel A. King and John F. Elder IV, Ph.D.,
Department of Systems Engineering-summary.

[4]“Suitability analysis of data mining toolsand


methods” Samuel Kováč, Masaryk university faculty of
informatics.

[5]“Towards the Formulation of a Unified DataMining


Theory, Implemented by Means ofMultiagent Systems
(MASs)”, Dost Muhammad Khan, Nawaz Mohamudally
and D. K. R. Babajee.

[6]“A Study of Data Mining Toolsin Knowledge


Discovery Process”, Y. Ramamohan, K. Vasantharao, C.
Kalyana Chakravarti, A.S.K.Ratnam.

[7]“Towards a General Framework for Data


Mining”,Saˇso DˇzeroskiJoˇzef Stefan Institute, Jamova
39, 1000 Ljubljana, Slovenia

[8]Principles of Data Mining, by David Hand,


HeikkiMannila and Padhraic Smyth

[9]WEKA, the University of Waikato, Available


at:http://www.cs.waikato.ac.nz/ml/weka/downloading.ht
ml

[10]KNIMEAvailableat:http://www.knime.org/downloa
d-desktop[11]“Open-Source Tools for Data Mining”
Blaz Zupan, PhD, Janez Demsar, PhD.[12]“Tanagra: An
Evaluation” Jess

S-ar putea să vă placă și