weka-3-4-4
所属分类:人工智能/神经网络/深度学习
开发工具:Java
文件大小:8347KB
下载次数:366
上传日期:2005-07-30 10:46:00
上 传 者:
Simenxiao
说明: 由java开发的软件包,里面有人工智能所用的很多东东,包括神经网络,支持向量机,决策树等分类和回归分析方法,集成化软件哦!
(by java development package, which has artificial intelligence used by many of the Eastern, including neural networks, support vector machines, such as decision tree classification and regression analysis, integrated software Oh!)
文件列表:
weka-3-4-4 (0, 2005-03-07)
weka-3-4-4\changelogs (0, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-1-4 (13106, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-1-5 (6938, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-1-6 (8222, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-1-7 (29256, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-1-8 (33326, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-1-9 (22553, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-2 (35645, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-3 (6282, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-3-1 (35879, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-3-2 (11029, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-3-3 (22535, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-3-4 (1389, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-3-5 (26175, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-3-6 (19581, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-4 (29462, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-4-1 (26737, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-4-2 (29308, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-4-3 (19564, 2005-03-07)
weka-3-4-4\changelogs\CHANGELOG-3-4-4.html (56660, 2005-03-07)
weka-3-4-4\COPYING (17977, 2005-03-07)
weka-3-4-4\data (0, 2005-03-07)
weka-3-4-4\data\contact-lenses.arff (2890, 2005-03-07)
weka-3-4-4\data\cpu.arff (6938, 2005-03-07)
weka-3-4-4\data\iris.arff (7486, 2005-03-07)
weka-3-4-4\data\labor.arff (8255, 2005-03-07)
weka-3-4-4\data\segment-challenge.arff (200410, 2005-03-07)
weka-3-4-4\data\segment-test.arff (109984, 2005-03-07)
weka-3-4-4\data\soybean.arff (202935, 2005-03-07)
weka-3-4-4\data\weather.arff (489, 2005-03-07)
weka-3-4-4\data\weather.nominal.arff (587, 2005-03-07)
weka-3-4-4\doc (0, 2005-03-07)
weka-3-4-4\doc\allclasses-frame.html (92367, 2005-03-07)
weka-3-4-4\doc\allclasses-noframe.html (78967, 2005-03-07)
weka-3-4-4\doc\constant-values.html (108747, 2005-03-07)
weka-3-4-4\doc\deprecated-list.html (5783, 2005-03-07)
weka-3-4-4\doc\help-doc.html (8121, 2005-03-07)
weka-3-4-4\doc\index-all.html (2801315, 2005-03-07)
weka-3-4-4\doc\index.html (897, 2005-03-07)
... ...
=====================================================================
======
README
======
WEKA 3.4.4
7 March 2005
Java Programs for Machine Learning
Copyright (C) 19***-2005 University of Waikato
web: http://www.cs.waikato.ac.nz/~ml
=====================================================================
Contents:
---------
1. Using one of the graphical user interfaces in Weka
2. The Weka data format (ARFF)
3. Using Weka from the command line
- Classifiers
- Association rules
- Filters
4. Database access
5. The Experiment package
6. Tutorial
7. Source code
8. Credits
9. Submission of code and bug reports
10. Copyright
----------------------------------------------------------------------
1. Using one of the graphical user interfaces in Weka:
------------------------------------------------------
This assumes that the Weka archive that you have downloaded has been
extracted into a directory containing this README and that you haven't
used an automatic installer (e.g. the one provided for Windows).
Weka 3.4 requires Java 1.4 or higher. Depending on your platform you
may be able to just double-click on the weka.jar icon to run the
graphical user interfaces for Weka. Otherwise, from a command-line
(assuming you are in the directory containing weka.jar), type
java -jar weka.jar
or if you are using Windows use
javaw -jar weka.jar
This will start a small graphical user interface (GUIChooser) from
which you can select the SimpleCLI interface or the more sophisticated
Explorer, Experimenter, and Knowledge Flow interfaces. SimpleCLI just
acts like a simple command shell. The Explorer is currently the main
interface for data analysis using Weka. The Experimenter can be used
to compare the performance of different learning algorithms across
various datasets. The Knowledge Flow provides a component-based
alternative to the Explorer interface.
Example datasets that can be used with Weka are in the sub-directory
called "data", which should be located in the same directory as this
README file.
The Weka user interfaces provide extensive built-in help facilities
(tool tips, etc.). Documentation for the Explorer can be found in
ExplorerGuide.pdf (also in the same directory as this
README).
You can also start the GUIChooser from within weka.jar:
java -classpath weka.jar:$CLASSPATH weka.gui.GUIChooser
or if you are using Windows use
javaw -classpath weka.jar;$CLASSPATH weka.gui.GUIChooser
----------------------------------------------------------------------
2. The Weka data format (ARFF):
-------------------------------
Datasets for WEKA should be formatted according to the ARFF
format. (However, there are several converters included in WEKA that
can convert other file formats to ARFF. The Weka Explorer will use
these automatically if it doesn't recognize a given file as an ARFF
file.) Examples of ARFF files can be found in the "data" subdirectory.
What follows is a short description of the file format. A more
complete description is available from the Weka web page.
A dataset has to start with a declaration of its name:
@relation name
followed by a list of all the attributes in the dataset (including
the class attribute). These declarations have the form
@attribute attribute_name specification
If an attribute is nominal, specification contains a list of the
possible attribute values in curly brackets:
@attribute nominal_attribute {first_value, second_value, third_value}
If an attribute is numeric, specification is replaced by the keyword
numeric: (Integer values are treated as real numbers in WEKA.)
@attribute numeric_attribute numeric
In addition to these two types of attributes, there also exists a
string attribute type. This attribute provides the possibility to
store a comment or ID field for each of the instances in a dataset:
@attribute string_attribute string
After the attribute declarations, the actual data is introduced by a
@data
tag, which is followed by a list of all the instances. The instances
are listed in comma-separated format, with a question mark
representing a missing value.
Comments are lines starting with % and are ignored by Weka.
----------------------------------------------------------------------
3. Using Weka from the command line:
------------------------------------
If you want to use Weka from your standard command-line interface
(e.g. bash under Linux):
a) Set WEKAHOME to be the directory which contains this README.
b) Add $WEKAHOME/weka.jar to your CLASSPATH environment variable.
c) Bookmark $WEKAHOME/doc/packages.html in your web browser.
Alternatively you can try using the SimpleCLI user interface available
from the GUI chooser discussed above.
In the following, the names of files assume use of a unix command-line
with environment variables. For other command-lines (including
SimpleCLI) you should substitute the name of the directory where
weka.jar lives for $WEKAHOME. If your platform uses something other
character than / as the path separator, also make the appropriate
substitutions.
===========
Classifiers
===========
Try:
java weka.classifiers.trees.J48 -t $WEKAHOME/data/iris.arff
This prints out a decision tree classifier for the iris dataset
and ten-fold cross-validation estimates of its performance. If you
don't pass any options to the classifier, WEKA will list all the
available options. Try:
java weka.classifiers.trees.J48
The options are divided into "general" options that apply to most
classification schemes in WEKA, and scheme-specific options that only
apply to the current scheme---in this case J48. WEKA has a common
interface to all classification methods. Any class that implements a
classifier can be used in the same way as J48 is used above. WEKA
knows that a class implements a classifier if it extends the
Classifier class in weka.classifiers. Almost all classes in
weka.classifiers fall into this category. Try, for example:
java weka.classifiers.bayes.NaiveBayes -t $WEKAHOME/data/labor.arff
Here is a list of some of the classifiers currently implemented in
weka.classifiers:
a) Classifiers for categorical prediction:
weka.classifiers.lazy.IBk: k-nearest neighbour learner
weka.classifiers.trees.J48: C4.5 decision trees
weka.classifiers.rules.PART: rule learner
weka.classifiers.bayes.NaiveBayes: naive Bayes with/without kernels
weka.classifiers.rules.OneR: Holte's OneR
weka.classifiers.functions.SMO: support vector machines
weka.classifiers.functions.Logistic: logistic regression
weka.classifiers.meta.AdaBoostM1: AdaBoost
weka.classifiers.meta.LogitBoost: logit boost
weka.classifiers.trees.DecisionStump: decision stumps (for boosting)
etc.
b) Classifiers for numeric prediction:
weka.classifiers.functions.LinearRegression: linear regression
weka.classifiers.trees.M5P: model trees
weka.classifiers.rules.M5Rules: model rules
weka.classifiers.lazy.IBk: k-nearest neighbour learner
weka.classifiers.lazy.LWR: locally weighted regression
=================
Association rules
=================
Next to classification schemes, there is some other useful stuff in
WEKA. Association rules, for example, can be extracted using the
Apriori algorithm. Try
java weka.associations.Apriori -t $WEKAHOME/data/weather.nominal.arff
=======
Filters
=======
There are also a number of tools that allow you to manipulate a
dataset. These tools are called filters in WEKA and can be found
in weka.filters.
weka.filters.unsupervised.attribute.Discretize: discretizes numeric data
weka.filters.unsupervised.attribute.Remove: deletes/selects attributes
etc.
Try:
java weka.filters.supervised.attribute.Discretize -i
$WEKAHOME/data/iris.arff -c last
----------------------------------------------------------------------
4. Database access:
-------------------
In terms of database connectivity, you should be able to use any
database with a Java JDBC driver. When using classes that access a
database (e.g. the Explorer), you will probably want to create a
properties file that specifies which JDBC drivers to use, where to
find the database, and specify a mapping for the data types. This file
should reside in your home directory or the current directory and be
called "DatabaseUtils.props". An example is provided in
weka/experiment (you need to expand wek.jar to be able to look a this
file). Note that the settings in this file are used unless they are
overidden by settings in the DatabaseUtils.props file in your home
directory or the current directory (in that order).
----------------------------------------------------------------------
5. The Experiment package:
--------------------------
There is support for running experiments that involve evaluating
classifiers on repeated randomizations of datasets, over multiple
datasets (you can do much more than this, besides). The classes for
this reside in the weka.experiment package. The basic architecture is
that a ResultProducer (which generates results on some randomization
of a dataset) sends results to a ResultListener (which is responsible
for stating whether it already has the result, and otherwise storing
results).
Example ResultListeners include:
weka.experiment.CSVResultListener: outputs results as
comma-separated-value files.
weka.experiment.InstancesResultListener: converts results into a set
of Instances.
weka.experiment.DatabaseResultListener: sends results to a database
via JDBC.
Example ResultProducers include:
weka.experiment.RandomSplitResultProducer: train/test on a % split
weka.experiment.CrossValidationResultProducer: n-fold cross-validation
weka.experiment.AveragingResultProducer: averages results from another
ResultPoducer
weka.experiment.DatabaseResultProducer: acts as a cache for results,
storing them in a database.
The RandomSplitResultProducer and CrossValidatioResultProducer make
use of a SplitEvaluator to obtain actual results for a particular
split, provided are ClassifierSplitEvaluator (for nominal
classification) and RegressionSplitEvaluator (for numeric
classification). Each of these uses a Classifier for actual results
generation.
So, you might have a DatabaseResultListener, that is sent results from
an AveragingResultProducer, which produces averages over the n results
produced for each run of an n-fold CrossValidationResultProducer,
which in turn is doing nominal classification through a
ClassifierSplitEvaluator, which uses OneR for prediction. Whew. But
you can combine these things together to do pretty much whatever you
want. You might want to write a LearningRateResultProducer that splits
a dataset into increasing numbers of training instances.
To run a simple experiment from the command line, try:
java weka.experiment.Experiment -r -T datasets/UCI/iris.arff \
-D weka.experiment.InstancesResultListener \
-P weka.experiment.RandomSplitResultProducer -- \
-W weka.experiment.ClassifierSplitEvaluator -- \
-W weka.classifiers.rules.OneR
(Try "java weka.experiment.Experiment -h" to find out what these
options mean)
If you have your results as a set of instances, you can perform paired
t-tests using weka.experiment.PairedTTester (use the -h option to find
out what options it needs).
However, all this is much easier if you use the Experimenter GUI.
----------------------------------------------------------------------
6. Tutorial:
------------
A tutorial on how to use WEKA is in $WEKAHOME/Tutorial.pdf. However,
not everything in WEKA is covered in the Tutorial, and the package
structure has changed quite a bit. For a complete list you have to
look at the online documentation $WEKAHOME/doc/packages.html. In
particular, Tutorial.pdf is a draft from the "Data Mining" book (see
our web page), and so only describes features in the stable 3.0
release.
----------------------------------------------------------------------
7. Source code:
---------------
The source code for WEKA is in $WEKAHOME/weka-src.jar. To expand it,
use the jar utility that's in every Java distribution.
----------------------------------------------------------------------
8. Credits:
-----------
Refer to the web page for a list of contributors:
http://www.cs.waikato.ac.nz/~ml/weka/
----------------------------------------------------------------------
9. Call for code and bug reports:
---------------------------------
If you have implemented a learning scheme, filter, application,
visualization tool, etc., using the WEKA classes, and you think it
should be included in WEKA, send us the code, and we can potentially
put it in the next WEKA distribution.
If you find any bugs, send a fix to mlcontrib@cs.waikato.ac.nz.
If that's too hard, just send a bug report to the wekalist mailing
list.
-----------------------------------------------------------------------
10. Copyright:
--------------
WEKA is distributed under the GNU public license. Please read
the file COPYING.
-----------------------------------------------------------------------
近期下载者:
相关文件:
收藏者: