daniel blanchard michael heilman nitin madnanifilesthat are similar in format to windows ini...

54
SciKit-Learn Laboratory Documentation Release 0.28.0 Daniel Blanchard Michael Heilman Nitin Madnani November 01, 2014

Upload: others

Post on 07-Jun-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

SciKit-Learn LaboratoryDocumentation

Release 0.28.0

Daniel Blanchard Michael HeilmanNitin Madnani

November 01, 2014

Contents

1 Documentation 31.1 Step 1: Installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31.2 License . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31.3 Step 2: Tutorial! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31.4 Running Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51.5 Utility Scripts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161.6 Python API . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

2 Indices and tables 45

Python Module Index 47

i

ii

SciKit-Learn Laboratory Documentation, Release 0.28.0

SKLL (pronounced “skull”) provides a number of utilities to make it simpler to run common scikit-learn experiments with pre-generated features.

There are two primary means of using SKLL: the run_experiment script and the Python API.

Contents 1

SciKit-Learn Laboratory Documentation, Release 0.28.0

2 Contents

CHAPTER 1

Documentation

1.1 Step 1: Installation

SKLL can easily be installed via pip:

pip install skll

or via conda (for Linux and OS X):

conda install -c https://conda.binstar.org/dan_blanchard skll

It can also be downloaded directly from Github.

1.2 License

SKLL is distributed under the 3-clause BSD License.

1.3 Step 2: Tutorial!

For this tutorial, we’re going to make use of the examples provided in the examples/ directoryin your copy of SKLL. Of course, the provided examples are already perfect and ready to use. Ifthis weren’t the case, you would need to...

1.3.1 1. Get your data into the correct format

SKLL can work with several common data formats, each of which are described here.

3

SciKit-Learn Laboratory Documentation, Release 0.28.0

Should you need your data to be in any number of these formats, or should you like to have say,one ARFF file for use with SKLL and Weka, the skll_convert script can help you out. This is aseasy as:

$ skll_convert examples/iris/train/example_iris_features.jsonlines examples/iris/train/example_iris_features.arff

skll_convert can also create sparse data files, very useful for speeding up large experimentswith lots of zero-valued features.

At minimum you will probably want to work with a training set and a testing set. You will probablywant to place these in clearly-labeled directories, since both files need to have the same name, andbe in the same format, unless they are listed explicitly in your config file. This will be discussed inthe next section.

At this time, utilities necessary to split data into training and testing are not part of SKLL.

1.3.2 2. Create a configuration file for the experiment you’d like to run

For this tutorial, we will refer to an “experiment” as having a single data set split into (at minimum)training and testing portions. As part of each experiment, we can train and test several models,either simultaneously or sequentially, depending on the availability of a grid engine. This will bedescribed in more detail later on, when we are ready to run our experiment.

You can consult the full list of learners currently available in SKLL to get an idea for the thingsyou can do. As part of this tutorial, we will use the following learners:

• Random Forest (RandomForestClassifier), C-Support Vector Classification (SVC),Linear Support Vector Classification (LinearSVC), Logistic Regression, MultinomialNaïve Bays (MultinomialNB) with the Iris example;

• Random Forest, Decision Tree, C-SVC, and Multinomial Naïve Bayes with the Titanicexample;

• Random Forest Regression (RandomForestRegressor), Support Vector Regression(SVR), and Linear Regression with the Boston example.

Optionally, you can also tune your models in such a way that they maximize a certain evaluationmetric, which we refer to as an “objective function”. The full list of available objective functionscan be found here.

Configuration File for the Iris Example

Let’s take a look at the options specified in iris/evaluate.cfg. Here, we are only goingto train a model and evaluate its performance, because in the General section, task is set to“evaluate”. We will explore the other options for task later.

In the Input section, you may want to adjust train_location and test_locationto point to the directories containing the Iris training and testing data (most likely

4 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

skll/examples/iris/train and skll/examples/iris/test respectively, relativeto your installation of SKLL). featuresets indicates the name of both the training and testingfiles. suffix is the suffix, indicating the type of file that both the training and testing sets are in.learners must always be specified in between [ ] brackets, even if you only want to use onelearner. This is similar to the featuresets option, which requires two sets of brackets, sincemultiple sets of different-yet-similar features can be provided. We will keep our examples simple,however, and only use one set of features per experiment.

The Tuning section defines how we want our model to be tuned. Setting grid_search to“True” here employs scikit-learn’s GridSearchCV class, which is an implementation of the stan-dard, brute-force approach to hyperparameter optimization. objective refers to the desiredobjective function; here, “f1_score_micro” will optimize for micro-averaged F1.

In the Output section, the arguments to each of these are directories you’d like all of the outputof your experiment to go. results refers to the results of the experiment, and log refers tolog files containing any status, warning, or error messages generated during model training andevaluation.

This experiment is not configured to serialize the constructed models.

Configuration File for the Iris Cross-Validation Example

The file iris/cross_val.cfg almost identical to iris/evaluate.cfg, except that it isconfigured to do three-fold cross-validation on the training set. This is indicated by setting taskto “cross_validate” in the General section.

Configuration File for the Titanic Example

1.4 Running Experiments

The simplest way to use SKLL is to create configuration files that describe experiments you wouldlike to run on pre-generated features. This document describes the supported feature file formats,how to create configuration files (and layout your directories), and how to use run_experimentto get things going.

1.4.1 Quick Example

If you don’t want to read the whole document, and just want an example of how things work, dothe following from the command prompt:

$ cd examples$ python make_example_iris_data.py # download a simple dataset

1.4. Running Experiments 5

SciKit-Learn Laboratory Documentation, Release 0.28.0

$ cd iris$ run_experiment --local evaluate.cfg # run an experiment

1.4.2 Feature file formats

The following feature file formats are supported:

arff The same file format used by Weka with the following added restrictions:

• Only simple numeric, string, and nomimal values are supported.

• Nominal values are converted to strings.

• If there is an attribute called “id” present, this will be treated as the ID foreach row.

• If the data is labelled, there must be an attribute with the name specified bylabel_col in the Input section of the configuartion file you create for yourexperiment. This defaults to ‘y’. This must also be the final attribute listed(like in Weka).

csv/tsv A simple comma or tab-delimited format with the following restrictions:

• If the data is labelled, there must be a column with the name specified bylabel_col in the Input section of the configuartion file you create for yourexperiment. This defaults to ‘y’.

• If there is a column called “id” present, this will be treated as the ID for eachrow.

• All other columns contain feature values, and every feature value must bespecified (making this a poor choice for sparse data).

jsonlines/ndj (Recommended) A twist on the JSON format where every line is a ei-ther JSON dictionary (the entire contents of a normal JSON file), or a commentline starting with //. Each dictionary is expected to contain the following keys:

• y: The class label.

• x: A dictionary of feature values.

• id: An optional instance ID.

This is the preferred file format for SKLL, as it is sparse and can be slightly fasterto load than other formats.

libsvm While we can process the standard input file format supported by LibSVM,LibLinear, and SVMLight, we also support specifying extra metadata usuallymissing from the format in comments at the of each line. The comments are notmandatory, but without them, your classes and features will not have names. Thecomment is structured as follows:

6 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

ID | 1=ClassX | 1=FeatureA 2=FeatureB

The entire format would like this:

2 1:2.0 3:8.1 # Example1 | 2=ClassY | 1=FeatureA 3=FeatureC1 5:7.0 6:19.1 # Example2 | 1=ClassX | 5=FeatureE 6=FeatureF

Note: IDs, classes, and feature names cannot contain the following characters:| # =

megam An expanded form of the input format for the MegaM classification packagewith the -fvals switch.

The basic format is:

# Instance1CLASS1 F0 2.5 F1 3 FEATURE_2 -152000# Instance2CLASS2 F1 7.524

where the optional comments before each instance specify the ID for the fol-lowing line, class names are separated from feature-value pairs with a tab, andfeature-value pairs are separated by spaces. Any omitted features for a given in-stance are assumed to be zero, so this format is handy when dealing with sparsedata. We also include several utility scripts for converting to/from this MegaMformat and for adding/removing features from the files.

1.4.3 Creating configuration files

The experiment configuration files that run_experiment accepts are standard Python configurationfiles that are similar in format to Windows INI files. 1 There are four expected sections in aconfiguration file: General, Input, Tuning, and Output. A detailed description of eachpossible settings for each section is provided below, but to summarize:

• If you want to do cross-validation, specify a path to training feature files, and set task tocross_validate. Please note that the cross-validation currently uses StratifiedKFold.You also can optionally use predetermined folds with the cv_folds_location setting.

• If you want to train a model and evaluate it on some data, specify a training location, a testlocation, and a directory to store to store results, and set task to evaluate.

• If you want to just train a model and generate predictions, specify a training location, atest location, and set task to predict.

1 We are considering adding support for JSON configuration files in the future, but we have not added this func-tionality yet.

1.4. Running Experiments 7

SciKit-Learn Laboratory Documentation, Release 0.28.0

• If you want to just train a model, specify a training location, and set task to train.

• A list of classifiers/regressors to try on your feature files is required.

Example configuration files are available here.

General

experiment_name A string used to identify this particular experiment configuration.When generating result summary files, this name helps prevent overwriting pre-vious summaries.

task What types of experiment we’re trying to run. Valid options are:cross_validate, evaluate, predict, and train. See above for de-scriptions.

Input

train_location Path to directory containing training data files. There must be a filefor each featureset.

test_location (Optional) Path to directory containing test data files. There must be afile for each featureset.

label_col (Optional) If you’re using ARFF, CSV, or TSV files, the class labels foreach instance are assumed to be in a column with this name. If no column withthis name is found, the data is assumed to be unlabelled. Defaults to ‘y’. ForARFF files only, this must also be the final column to count as the label (forcompatibility with Weka).

ids_to_floats (Optional) If you have a dataset with lots of examples, and your inputfiles have IDs that look like numbers (can be converted by float()), then settingthis to True will save you some memory by storing IDs as floats. Note that thiswill cause IDs to be printed as floats in prediction files (e.g., “4.0” instead of “4”or “0004” or “4.000”).

shuffle (Optional) If True, shuffle the examples in the training data before usingthem for learning. This happens automatically when doing a grid search butit might be useful in other scenarios as well, e.g., online learning. Defaults toFalse.

class_map (Optional) If you would like to collapse several classes into one, or other-wise modify your labels (without modifying your original feature files), you canspecify a dictionary mapping from new class labels to lists of original class la-bels. For example, if you wanted to collapse the classes “beagle” and “dachsund”into a “dog” class, you would specify the following for class_map:

8 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

{’dog’: [’beagle’, ’dachsund’]}

Any classes not included in the dictionary will be left untouched.

cv_folds_location (Optional) Path to a csv file (with a header that is ignored) specify-ingfolds for cross-validation. The first column should consist of training set IDsand the second should be a string for the fold ID (e.g., 1 through 5, A through D,etc.). If specified, the CV and grid search will leave one fold ID out at a time. 2

featuresets List of lists of prefixes for the files containing the features you would liketo train/test on. Each list will end up being a job. IDs are required to be the samein all of the feature files, and a ValueErrorwill be raised if this is not the case.

sampler (Optional) It performs a non-linear transformations of the input, whichcan serve as a basis for linear classification or other algorithms. Validoptions are: Nystroem, RBFSampler, SkewedChi2Sampler andAdditiveChi2Sampler. For addition information see the scikit-learn docu-mentation.

sampler_parameters (Optional) dict containing parameters you want to have fixedfor the sampler. Any empty ones will be ignored (and the defaults will beused).

The default fixed parameters (beyond those that scikit-learn sets) are:

Nystroem

{’random_state’: 123456789}

RBFSampler

{’random_state’: 123456789}

SkewedChi2Sampler

{’random_state’: 123456789}

feature_hasher (Optional) If “true”, this enables a high-speed, low-memory vector-izer that uses feature hashing for converting feature dictionaries into NumPy ar-rays instead of using a DictVectorizer. This flag will drastically reducememory consumption for data sets with a large number of features. If enabled,the user should also specify the number of features in the hasher_featuresfield. For addition information see the scikit-learn documentation.

hasher_features (Optional) The number of features used by the FeatureHasherif the feature_hasher flag is enabled. It is suggested to use the power oftwo larger than the number of features in the data set. For example, if you had17 features, you would want to set the flag to 32.

2 K-1 folds will be used for grid search within CV, so there should be at least 3 fold IDs.

1.4. Running Experiments 9

SciKit-Learn Laboratory Documentation, Release 0.28.0

suffix (Optional) The file format the training/test files are in. Valid option are .arff,.csv, .jsonlines, .libsvm, .megam,, .ndj, and .tsv”.

If you omit this field, it is assumed that the “prefixes” listed in featuresetsare actually complete filenames. This can be useful if you have feature files thatare all in different formats that you would like to combine.

featureset_names (Optional) Optional list of names for the feature sets. If omitted,then the prefixes will be munged together to make names.

learners 3 List of scikit-learn models to try using. A separate job will be run foreach combination of classifier and feature-set. Acceptable values are describedbelow. Names in parentheses are aliases that can also be used inconfigurationfiles. Custom learners can also be specified. See custom_learner_path.

Classifiers:

• AdaBoostClassifier: AdaBoost Classifier

• DecisionTreeClassifier (dtree): Decision Tree Classifier

• GradientBoostingClassifier (gradient): Gradient Boosting Classifier

• KNeighborsClassifier: K-Nearest Neighbors Classifier

• LinearSVC (svm_linear): SVM using LibLinear

• LogisticRegression (logistic): Logistic regression using LibLinear

• MultinomialNB (naivebayes): Multinomial Naive Bayes

• RandomForestClassifier (rforest): Random Forest Classifier

• SGDClassifier: Stochastic Gradient Descent Classifier

• SVC (svm_radial): SVM using LibSVM

Regressors:

• AdaBoostRegressor: AdaBoost Regressor

• DecisionTreeRegressor: Decision Tree Regressor

• ElasticNet: ElasticNet Regression

• GradientBoostingRegressor (gb_regressor): Gradient Boosting Re-gressor

• KNeighborsRegressor: K-Nearest Neighbors Regressor

• Lasso: Lasso Regression

• LinearRegression: Linear Regression

• RandomForestRegressor: Random Forest Regressor

• Ridge (ridge): Ridge Regression

10 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

• SGDRegressor: Stochastic Gradient Descent Regressor

• SVR (svr_linear): Support Vector Regression with a linear kernel.Can use other kernels by specifying a ‘kernel’ fixed parameter inthe fixed_parameters list.

For all regressors you can also prepend Rescaled to the beginningof the full name (e.g., RescaledSVR) to get a version of the regres-sor where predictions are rescaled and constrained to better match thetraining set.

custom_learner_path (Optional) Path to a .py file that defines a customlearner. This file will be imported dynamically. This is only re-quired if a custom learner in specified in the list of learners. Cus-tom learners must implement the fit and predict methods and inheritsklearn.base.BaseEstimator. Custom regression learners must alsoinherit sklearn.base.RegressorMixin. Models that require dense ma-trices should implement a method requires_dense that returns True.

fixed_parameters (Optional) List of dicts containing parameters you want to havefixed for each classifier in learners list. Any empty ones will be ignored (andthe defaults will be used).

The default fixed parameters (beyond those that scikit-learn sets) are:

LogisticRegression

{’random_state’: 123456789}

LinearSVC

{’random_state’: 123456789}

SVC

{’cache_size’: 1000}

DecisionTreeClassifier and DecisionTreeRegressor

{’random_state’: 123456789}

RandomForestClassifier and RandomForestRegressor

{’n_estimators’: 500, ’random_state’: 123456789}

GradientBoostingClassifier and GradientBoostingRegressor

{’n_estimators’: 500, ’random_state’: 123456789}

SVR

1.4. Running Experiments 11

SciKit-Learn Laboratory Documentation, Release 0.28.0

{’cache_size’: 1000, ’kernel’: b’linear’}

Note: This option helps us to manage the case of imbalancedata sets with the parameter class_weight for the classifiers: SVC,LogisticRegression, LinearSVC and SGDClassifier.

Two possible options are available. The first one with an ‘auto’ option that usesthe values of y (the classes) to automatically adjust weights inversely propor-tional to class frequencies, as shown in the following code:

{’class_weight’: ’auto’}

And the second option allows you to assign an specific weight per each class.The default weight per class is 1. An example could be:

{’class_weight’: {1: 10}}

Additional examples and information can be seen here.

Tuning

feature_scaling (Optional) Whether to scale features by their mean and/or their stan-dard deviation. This defaults to none, which does no scaling of any kind. If youscale by mean, your data will automatically be converted to dense, so use cautionwhen you have a very large dataset. Valid options are:

• none: perform no feature scaling at all.

• with_std: Scale feature values by their standard deviation.

• with_mean: Center features by subtracting their mean.

• both: perform both centering and scaling.

Defaults to none.

grid_search (Optional) Whether or not to perform grid search to find optimal param-eters for classifier. Defaults to False.

grid_search_jobs (Optional) Number of folds to run in parallel when using gridsearch. Defaults to number of grid search folds.

min_feature_count (Optional) The minimum number of examples for a which eachfeature must be nonzero to be included in the model. Defaults to 1.

objective (Optional) The objective function to use for tuning. Valid options are:

Classification:

• accuracy: Overall accuracy

12 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

• precision: Precision

• recall: Recall

• f1: The default scikit-learn F1 score (F1 of the positive class for binaryclassification, or the weighted average F1 for multiclass classification)

• f1_score_micro: Micro-averaged F1 score

• f1_score_macro: Macro-averaged F1 score

• f1_score_weighted: Weighted average F1 score

• f1_score_least_frequent: F1 score of the least frequent class. The least fre-quent class may vary from fold to fold for certain data distributions.

• average_precision: Area under PR curve (for binary classification)

• roc_auc: Area under ROC curve (for binary classification)

Regression or classification with integer classes:

• unweighted_kappa: Unweighted Cohen’s kappa (any floating point valuesare rounded to ints)

• linear_weighted_kappa: Linear weighted kappa (any floating point valuesare rounded to ints)

• quadratic_weighted_kappa: Quadratic weighted kappa (any floating pointvalues are rounded to ints)

• uwk_off_by_one: Same as unweighted_kappa, but all ranking differ-ences are discounted by one. In other words, a ranking of 1 and a ranking of2 would be considered equal.

• lwk_off_by_one: Same as linear_weighted_kappa, but all rankingdifferences are discounted by one.

• qwk_off_by_one: Same as quadratic_weighted_kappa, but all rank-ing differences are discounted by one.

Regression or classification with binary classes:

• kendall_tau: Kendall’s tau

• pearson: Pearson correlation

• spearman: Spearman rank-correlation

Regression:

• r2: R2

• mean_squared_error: Mean squared error regression loss

Defaults to f1_score_micro.

1.4. Running Experiments 13

SciKit-Learn Laboratory Documentation, Release 0.28.0

param_grids (Optional) List of parameter grids to search for each classifier. Eachparameter grid should be a list of dictionaries mapping from strings to lists ofparameter values. When you specify an empty list for a classifier, the defaultparameter grid for that classifier will be searched.

The default parameter grids for each classifier are:

LogisticRegression

[{’C’: [0.01, 0.1, 1.0, 10.0, 100.0]}]

LinearSVC

[{’C’: [0.01, 0.1, 1.0, 10.0, 100.0]}]

SVC

[{’C’: [0.01, 0.1, 1.0, 10.0, 100.0]}]

MultinomialNB

[{’alpha’: [0.1, 0.25, 0.5, 0.75, 1.0]}]

DecisionTreeClassifier and DecisionTreeRegressor

[{’max_features’: ["auto", None]}]

RandomForestClassifier and RandomForestRegressor

[{’max_depth’: [1, 5, 10, None]}]

GradientBoostingClassifier and GradientBoostingRegressor

[{’max_depth’: [1, 3, 5], ’n_estimators’: [500]}]

ElasticNet, Lasso, and Ridge

[{’alpha’: [0.01, 0.1, 1.0, 10.0, 100.0]}]

SVR

[{’C’: [0.01, 0.1, 1.0, 10.0, 100.0]}]

pos_label_str (Optional) The string label for the positive class in the binary classifi-cation setting. If unspecified, an arbitrary class is picked.

Output

probability (Optional) Whether or not to output probabilities for each class instead ofthe most probable class for each instance. Only really makes a difference whenstoring predictions. Defaults to False.

14 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

results (Optional) Directory to store result files in. If omitted, the current workingdirectory is used.

log (Optional) Directory to store result files in. If omitted, the current working direc-tory is used.

models (Optional) Directory to store trained models in. Can be omitted to not storemodels.

predictions (Optional) Directory to store prediction files in. Can be omitted to notstore predictions.

Note: you can use the same directory for results, log, models, and predictions.

1.4.4 Using run_experiment

Once you have create the configuration file for your experiment, you can usually just getyour experiment started by running run_experiment CONFIGFILE. That said, there are afew options that are specified via command-line arguments instead of in the configuration file:--ablation, --keep-models, and --resume.

--ablation NUM_FEATURES Runs an ablation study where repeated experi-ments are conducted with the specified number of feature files in each featuresetin the configuration file held out. For example, if you have three feature files(A, B, and C) in your featureset and you specifiy --ablation 1, there will bethree three experiments conducted with the following featuresets: [[A, B],[B, C], [A, C]].

If you would like to try all possible combinations of feature files, you can use the--ablation_all option instead.

--keep-models If trained models already exist for any of the learner/featuresetcombinations in your configuration file, just load those models and do not re-train/overwrite them.

--resume If result files already exist for an experiment, do not overwrite them. Thisis very useful when doing a large ablation experiment and part of it crashes.

If you have Grid Map installed, run_experiment will automatically schedule jobs on yourDRMAA-compatible cluster. However, if you would just like to run things locally, you can specifythe --local option. 4 You can also customize the queue and machines that are used for runningyour jobs via the --queue and --machines arguments. For complete details on how to specifythese options, just run run_experiment --help.

The result, log, model, and prediction files generated by run_experiment will all share the followingautomatically generated prefix EXPERIMENT_FEATURESET_LEARNER, where the followingdefinitions hold:

4 This will happen automatically if Grid Map cannot be imported.

1.4. Running Experiments 15

SciKit-Learn Laboratory Documentation, Release 0.28.0

EXPERIMENT The name specified as experiment_name in the configuration file.

FEATURESET The feature set we’re training on joined with “+”.

LEARNER The learner the current results/model/etc. was generated using.

For every experiment you run, there will also be a result summary file generated that is a tab-delimited file summarizing the results for each learner-featureset combination you have in yourconfiguration file. It is named EXPERIMENT_summary.tsv.

1.5 Utility Scripts

In addition to the main script, run_experiment, SKLL comes with a number of helpful utility scriptsthat can be used to prepare feature files and perform other routine tasks. Each is described brieflybelow. For full details of how to use each of these scripts, just run them with the --help option.

1.5.1 compute_eval_from_predictions

Compute evaluation metrics from prediction files after you have run an experiment.

1.5.2 filter_megam

Filter MegaM file to remove features with names in stop word list (or non alphabetic characters).Also has side-effect of removing TEST, TRAIN, and DEV lines if they are present.

1.5.3 generate_predictions

Loads a trained model and outputs predictions based on input feature files. Useful if you want toreuse a trained model as part of a larger system without creating configuration files.

1.5.4 join_megam

Combine MegaM files that contain features for the same examples.

1.5.5 print_model_weights

Prints out the weights of a given trained model.

16 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

1.5.6 skll_convert

Convert between .arff, .csv., .jsonlines, .libsvm, .megam, and .tsv formats.

1.5.7 summarize_results

Creates an experiment summary TSV file from a list of JSON files generated byrun_experiment.

1.6 Python API

The complete Python API is documented below, but here is a quick run-down of how you accom-plish common tasks.

Load some examples:

from skll import load_examples

train_examples = load_examples(’myexamples.megam’)

Train a linear svm (assuming we have train_examples):

from skll import Learner

learner = Learner(’LinearSVC’)learner.train(train_examples)

Evaluate a trained model:

test_examples = load_examples(’test.tsv’)conf_matrix, accuracy, prf_dict, model_params, obj_score = learner.evaluate(test_examples)

Perform ten-fold cross-validation with a radial SVM:

learner = Learner(’SVC’)fold_result_list, grid_search_scores = learner.cross-validate(train_examples)

fold_result_list in this case is a list of the results returned by learner.evaluate foreach fold, and grid_search_scores is the highest objective function value achieved whentuning the model.

Generate predictions from a trained model:

predictions = learner.predict(test_examples)

1.6. Python API 17

SciKit-Learn Laboratory Documentation, Release 0.28.0

1.6.1 skll Package

The most useful parts of our API are available at the package level in addition to the module level.They are documented in both places for convenience.

From data Module

skll.convert_examples(example_dicts, sparse=True, ids_to_floats=False)This function is to facilitate programmatic use of Learner.predict() and other functions thattake FeatureSet objects as input. It converts a .jsonlines/.ndj-style list of dictionaries into anFeatureSet.

Parameters example_dicts (iterable of dicts) – An list of dictionaries followingthe .jsonlines/.ndj format (i.e., features ‘x’, label ‘y’, and ‘id’).

Returns an FeatureSet representing the examples in example_dicts.

skll.load_examples(path, quiet=False, sparse=True, label_col=u’y’,ids_to_floats=False, class_map=None, feature_hasher=False,num_features=None)

Loads examples in the .arff, .csv, .jsonlines, .libsvm, .megam, .ndj, or.tsv formats.

If you would like to include example/instance IDs in your files, they must be specified in thefollowing ways:

•MegaM: As a comment line directly preceding the line with feature values.

•LibSVM: As the first item in the three-part comment described below.

•CSV/TSV/ARFF: An “id” column.

•JSONLINES: An “id” key in each JSON dictionary.

Also, for ARFF, CSV, and TSV files, there must be a column with the name specified bylabel_col if the data is labelled. For ARFF files, this column must also be the final one (as itis in Weka).

For LibSVM files, we use a specially formatted comment for storing example IDs, classnames, and feature names, which are normally not supported by the format. The comment isnot mandatory, but without it, your classes and features will not have names. The commentis structured as follows:

ExampleID | 1=FirstClass | 1=FirstFeature 2=SecondFeature

Parameters

• path (str or dict) – The path to the file to load the examples from, or alist of example dictionaries (like you would pass to convert_examples).

18 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

• quiet (bool) – Do not print “Loading...” status message to stderr.

• sparse (bool) – Whether or not to store the features in a numpy CSRmatrix.

• label_col (str) – Name of the column which contains the class labels forARFF/CSV/TSV files. If no column with that name exists, or None isspecified, the data is considered to be unlabelled.

• ids_to_floats (bool) – Convert IDs to float to save memory. Will raiseerror if we encounter an a non-numeric ID.

• class_map (dict from str to str) – Mapping from original class labelsto new ones. This is mainly used for collapsing multiple classes into asingle class. Anything not in the mapping will be kept the same.

Returns 4-tuple of an array example ids, an array of class labels, a scipy CSR ma-trix of features, and a DictVectorizer containing the mapping between featurenames and the column indices in the feature matrix.

skll.write_feature_file(path, ids, classes, features, feat_vectorizer=None,id_prefix=u’EXAMPLE_’, label_col=u’y’,arff_regression=False, arff_relation=u’skll_relation’,subsets=None, label_map=None)

Writes a feature file in either .arff, .csv, .jsonlines, .megam, .ndj, or .tsvformats with the given a list of IDs, classes, and features.

Parameters

• path (str) – A path to the feature file we would like to create. The suf-fix to this filename must be .arff, .csv, .jsonlines, .libsvm,.megam, .ndj, or .tsv. If subsets is not None, this is assumed tobe a string containing the path to the directory to write the feature fileswith an additional file extension specifying the file type. For example/foo/.csv.

• ids (list of str) – The IDs for each instance in the feature list/array. IfNone, IDs will be automatically generated with the prefix specified byid_prefix followed by the row number. If id_prefix is also None, no IDswill be written to the file.

• classes (list of str) – The class labels for each instance in the featurelist/array. If None, no class labels will be added to output file.

• features (list of dict or array-like) – The features for each instance rep-resented as either a list of dictionaries or an array-like (if feat_vectorizeris also specified).

• feat_vectorizer (DictVectorizer) – A DictVectorizer to map to/from fea-ture columns indices and names.

1.6. Python API 19

SciKit-Learn Laboratory Documentation, Release 0.28.0

• label_col (str) – Name of the column which contains the class labels forCSV/TSV files. If no column with that name exists, or None is specified,the data is considered to be unlabelled.

• arff_regression (bool) – A boolean value indicating whether the ARFFfiles that are written should be written for regression rather than classifi-cation, i.e., the class variable y is numerical rather than an enumerationof classes and all non-numeric attributes are removed.

• arff_relation (str) – Relation name for ARFF file.

• subsets (dict (str to list of str)) – A mapping from subset names tolists of feature names that are included in those sets. If given, a fea-ture file will be written for every subset (with the name containing thesubset name as suffix to path). Note, since string- valued features areautomatically converted into boolean features with names of the formFEATURE_NAME=STRING_VALUE, when doing the filtering, the por-tion before the = is all that’s used for matching. Therefore, you do notneed to enumerate all of these boolean feature names in your mapping.

• label_map (dict (str -> int)) – Mapping from class name to numbers touse for writing LibSVM files.

From experiments Module

skll.run_ablation(config_path, local=False, overwrite=True, queue=u’all.q’,hosts=None, quiet=False, all_combos=False)

Takes a configuration file and runs repeated experiments where each feature set has beenremoved from the configuration.

Parameters

• config_path (str) – Path to the configuration file we would like to use.

• local (bool) – Should this be run locally instead of on the cluster?

• overwrite (bool) – If the model files already exist, should we overwritethem instead of re-using them?

• queue (str) – The DRMAA queue to use if we’re running on the cluster.

• hosts (list of str) – If running on the cluster, these are the machines weshould use.

• quiet (bool) – Suppress printing of “Loading...” messages.

• all_combos (bool) – By default we only exclude one feature from set,but if all_combos is True, we do a true ablation study with all featurecombinations.

20 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

Deprecated since version 0.20.0: Use run_configuration() with the ablation argu-ment instead.

skll.run_configuration(config_file, local=False, overwrite=True,queue=u’all.q’, hosts=None, write_summary=True,quiet=False, ablation=0, resume=False)

Takes a configuration file and runs the specified jobs on the grid.

Parameters

• config_path (str) – Path to the configuration file we would like to use.

• local (bool) – Should this be run locally instead of on the cluster?

• overwrite (bool) – If the model files already exist, should we overwritethem instead of re-using them?

• queue (str) – The DRMAA queue to use if we’re running on the cluster.

• hosts (list of str) – If running on the cluster, these are the machines weshould use.

• write_summary (bool) – Write a tsv file with a summary of the results.

• quiet (bool) – Suppress printing of “Loading...” messages.

• ablation (int or None) – Number of features to remove when doing anablation experiment. If positive, we will perform repeated ablation runsfor all combinations of features removing the specified number at a time.If None, we will use all combinations of all lengths. If 0, the default, noablation is performed. If negative, a ValueError is raised.

• resume (bool) – If result files already exist for an experiment, do notoverwrite them. This is very useful when doing a large ablation experi-ment and part of it crashes.

Returns A list of paths to .json results files for each variation in the experiment.

Return type list of str

From learner Module

class skll.Learner(model_type, probability=False, feature_scaling=u’none’,model_kwargs=None, pos_label_str=None,min_feature_count=1, sampler=None, sampler_kwargs=None,custom_learner_path=None)

Bases: object

A simpler learner interface around many scikit-learn classification and regression functions.

Parameters

1.6. Python API 21

SciKit-Learn Laboratory Documentation, Release 0.28.0

• model_type (str) – Type of estimator to create (e.g., LogisticRegres-sion). See the skll package documentation for valid options.

• probability (bool) – Should learner return probabilities of all classes(instead of just class with highest probability)?

• model_kwargs (dict) – A dictionary of keyword arguments to pass tothe initializer for the specified model.

• pos_label_str (str) – The string for the positive class in the binary clas-sification setting. Otherwise, an arbitrary class is picked.

• feature_scaling (str) – how to scale the features, if at all. Options are:‘with_std’: scale features using the standard deviation, ‘with_mean’:center features using the mean, ‘both’: do both scaling as well as center-ing, ‘none’: do neither scaling nor centering

• min_feature_count (int) – The minimum number of examples a featuremust have a nonzero value in to be included.

• custom_learner_path (str) – Path to code where a custom classifier isdefined.

cross_validate(examples, stratified=True, cv_folds=10,grid_search=False, grid_search_folds=3, grid_jobs=None,grid_objective=u’f1_score_micro’, prediction_prefix=None,param_grid=None, shuffle=False, feature_hasher=False)

Cross-validates a given model on the training examples.

Parameters

• examples (FeatureSet) – The data to cross-validate learner perfor-mance on.

• stratified (bool) – Should we stratify the folds to ensure an even distri-bution of classes for each fold?

• cv_folds (int or dict) – The number of folds to use for cross-validation,or a mapping from example IDs to folds.

• grid_search (bool) – Should we do grid search when training eachfold? Note: This will make this take much longer.

• grid_search_folds (int) – The number of folds to use when doing thegrid search (ignored if cv_folds is set to a dictionary mapping examplesto folds).

• grid_jobs (int) – The number of jobs to run in parallel when doing thegrid search. If unspecified or 0, the number of grid search folds will beused.

22 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

• grid_objective (function) – The objective function to use when doingthe grid search.

• param_grid (list of dicts mapping from strs to lists of parameter val-ues) – The parameter grid to search through for grid search. If unspec-ified, a default parameter grid will be used.

• prediction_prefix (str) – If saving the predictions, this is the prefix thatwill be used for the filename. It will be followed by ”.predictions”

• shuffle (bool) – Shuffle examples before splitting into folds for CV.

• feature_hasher (bool) – Are we using feature_hasher?

Returns The confusion matrix, overall accuracy, per-class PRFs, and modelparameters for each fold.

Return type list of 4-tuples

evaluate(examples, prediction_prefix=None, append=False,grid_objective=None, feature_hasher=False)

Evaluates a given model on a given dev or test example set.

Parameters

• examples (FeatureSet) – The examples to evaluate the performance ofthe model on.

• prediction_prefix (str) – If saving the predictions, this is the prefix thatwill be used for the filename. It will be followed by ”.predictions”

• append (bool) – Should we append the current predictions to the file ifit exists?

• grid_objective (function) – The objective function that was used whendoing the grid search.

• feature_hasher (bool) – are we using a feature_hasher?

Returns The confusion matrix, the overall accuracy, the per-class PRFs, themodel parameters, and the grid search objective function score.

Return type 5-tuple

classmethod from_file(learner_path)

Returns New instance of Learner from the pickle at the specified path.

load(learner_path)Replace the current learner instance with a saved learner.

Parameters learner_path (str) – The path to the file to load.

modelThe underlying scikit-learn model

1.6. Python API 23

SciKit-Learn Laboratory Documentation, Release 0.28.0

model_kwargsA dictionary of the underlying scikit-learn model’s keyword arguments

model_paramsModel parameters (i.e., weights) for LinearModel (e.g., Ridge) regression andliblinear models.

Returns Labeled weights and (labeled if more than one) intercept value(s)

Return type tuple of (weights, intercepts), where weights is a dictand intercepts is a dictionary

model_typeThe model type (i.e., the class)

predict(examples, prediction_prefix=None, append=False, class_labels=False,feature_hasher=False)

Uses a given model to generate predictions on a given data set

Parameters

• examples (FeatureSet) – The examples to predict the classes for.

• prediction_prefix (str) – If saving the predictions, this is the prefix thatwill be used for the filename. It will be followed by ”.predictions”

• append (bool) – Should we append the current predictions to the file ifit exists?

• class_labels (bool) – For classifier, should we convert class indices totheir (str) labels?

• feature_hasher (bool) – For classifier, are we using a feature_hasher?

Returns The predictions returned by the learner.

Return type array

probabilityShould learner return probabilities of all classes (instead of just class with highest prob-ability)?

save(learner_path)Save the learner to a file.

Parameters learner_path (str) – The path to where you want to save thelearner.

train(examples, param_grid=None, grid_search_folds=3, grid_search=True,grid_objective=u’f1_score_micro’, grid_jobs=None, shuffle=False, fea-ture_hasher=False, run_create_label_dict=True)

Train a classification model and return the model, score, feature vectorizer, scaler, labeldictionary, and inverse label dictionary.

24 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

Parameters

• examples (FeatureSet) – The examples to train the model on.

• param_grid (list of dicts mapping from strs to lists of parameter val-ues) – The parameter grid to search through for grid search. If unspec-ified, a default parameter grid will be used.

• grid_search_folds (int or dict) – The number of folds to use whendoing the grid search, or a mapping from example IDs to folds.

• grid_search (bool) – Should we do grid search?

• grid_objective (function) – The objective function to use when doingthe grid search.

• grid_jobs (int) – The number of jobs to run in parallel when doing thegrid search. If unspecified or 0, the number of grid search folds will beused.

• shuffle (bool) – Shuffle examples (e.g., for grid search CV.)

Returns The best grid search objective function score, or 0 if we’re not doinggrid search.

Return type float

From metrics Module

skll.f1_score_least_frequent(y_true, y_pred)Optimize the hyperparameter values during the grid search based on the F1 measure of theleast frequent class.

This is mostly intended for use when you’re doing binary classification and your data ishighly skewed. You should probably use f1_score_macro if your data is skewed and you’redoing multi-class classification.

skll.kappa(y_true, y_pred, weights=None, allow_off_by_one=False)Calculates the kappa inter-rater agreement between two the gold standard and the predictedratings. Potential values range from -1 (representing complete disagreement) to 1 (repre-senting complete agreement). A kappa value of 0 is expected if all agreement is due tochance.

In the course of calculating kappa, all items in y_true and y_pred will first be converted tofloats and then rounded to integers.

It is assumed that y_true and y_pred contain the complete range of possible ratings.

This function contains a combination of code from yorchopolis’s kappa-stats and Ben Ham-ner’s Metrics projects on Github.

1.6. Python API 25

SciKit-Learn Laboratory Documentation, Release 0.28.0

Parameters

• weights (str or numpy array) – Specifies the weight matrix for the cal-culation. Options are:

– None = unweighted-kappa

– ‘quadratic’ = quadratic-weighted kappa

– ‘linear’ = linear-weighted kappa

– two-dimensional numpy array = a custom matrix of weights. Eachweight corresponds to the 𝑤𝑖𝑗 values in the wikipedia description ofhow to calculate weighted Cohen’s kappa.

• allow_off_by_one (bool) – If true, ratings that are off by one are countedas equal, and all other differences are reduced by one. For example, 1 and2 will be considered to be equal, whereas 1 and 3 will have a differenceof 1 for when building the weights matrix.

skll.kendall_tau(y_true, y_pred)Optimize the hyperparameter values during the grid search based on Kendall’s tau.

This is useful in cases where you want to use the actual probabilities of the different classesafter the fact, and not just the optimize based on the classification accuracy.

skll.spearman(y_true, y_pred)Optimize the hyperparameter values during the grid search based on Spearman rank correla-tion.

This is useful in cases where you want to use the actual probabilities of the different classesafter the fact, and not just the optimize based on the classification accuracy.

skll.pearson(y_true, y_pred)Optimize the hyperparameter values during the grid search based on Pearson correlation.

1.6.2 data Package

data.featureset Module

Classes related to storing/merging feature sets.

author Dan Blanchard ([email protected])

organization ETS

class skll.data.featureset.ExamplesTuple(ids=None, classes=None,features=None,feat_vectorizer=None)

Bases: skll.data.featureset.FeatureSet

26 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

Deprecated class only here to help people transition to Featureset.

Parameters

• name (str) – The name of this feature set.

• ids (np.array) – Example IDs for this set.

• classes (np.array) – Classes for this set.

• features (np.array) – Feature matrix as created by the given feature vec-torizer.

• feat_vectorizer (DictVectorizer or FeatureHasher) – Vectorizer that cre-ated feature matrix.

class skll.data.featureset.FeatureSet(name, ids, classes=None, fea-tures=None, vectorizer=None)

Bases: object

Encapsulation of all of the features, values, and metadata about a given set of data.

This replaces ExamplesTuple in older versions.

Parameters

• name (str) – The name of this feature set.

• ids (np.array) – Example IDs for this set.

• classes (np.array) – Classes for this set.

• features (list of dict or array-like) – The features for each instance rep-resented as either a list of dictionaries or an array-like (if feat_vectorizeris also specified).

• vectorizer (DictVectorizer or FeatureHasher) – Vectorizer that createdfeature matrix.

Note: If ids, classes, and/or features are not None, the number of rows in each array mustbe equal.

feat_vectorizerBackward compatible name for vectorizer

filter(ids=None, classes=None, features=None, inverse=False)Removes or keeps features and/or examples from the Featureset depending on thepassed in parameters.

Parameters

• ids (list of str/float) – Examples to keep in the FeatureSet. If None, noID filtering takes place.

1.6. Python API 27

SciKit-Learn Laboratory Documentation, Release 0.28.0

• classes (list of str/float) – Classes that we want to retain examples for.If None, no class filtering takes place.

• features (list of str) – Features to keep in the FeatureSet. To helpwith filtering string-valued features that were converted to sequencesof boolean features when read in, any features in the FeatureSet thatcontain a = will be split on the first occurrence and the prefix will bechecked to see if it is in features. If None, no feature filtering takesplace. Cannot be used if FeatureSet uses a FeatureHasher for vector-ization.

• inverse (bool) – Instead of keeping features and/or examples in lists,remove them.

filtered_iter(ids=None, classes=None, features=None, inverse=False)A version of __iter__ that retains only the specified features and/or examples fromthe output.

Parameters

• ids (list of str/float) – Examples in the FeatureSet to keep. If None, noID filtering takes place.

• classes (list of str/float) – Classes that we want to retain examples for.If None, no class filtering takes place.

• features (list of str) – Features in the FeatureSet to keep. To helpwith filtering string-valued features that were converted to sequencesof boolean features when read in, any features in the FeatureSet thatcontain a = will be split on the first occurrence and the prefix will bechecked to see if it is in features. If None, no feature filtering takesplace. Cannot be used if FeatureSet uses a FeatureHasher for vector-ization.

• inverse (bool) – Instead of keeping features and/or examples in lists,remove them.

has_classesWhether or not this FeatureSet has any finite classes.

data.readers Module

Handles loading data from various types of data files.

author Dan Blanchard ([email protected])

author Michael Heilman ([email protected])

author Nitin Madnani ([email protected])

28 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

organization ETS

class skll.data.readers.ARFFReader(path_or_list, **kwargs)Bases: skll.data.readers.DelimitedReader

Reader for creating a FeatureSet out of an ARFF file.

If you would like to include example/instance IDs in your files, they must be specified as an“id” column.

Also, there must be a column with the name specified by label_col if the data is labelled, andthis column must be the final one (as it is in Weka).

static split_with_quotes(s, delimiter=u’ ‘, quote_char=u””’, es-cape_char=u’\’)

A replacement for string.split that won’t split delimiters enclosed in quotes.

class skll.data.readers.CSVReader(path_or_list, **kwargs)Bases: skll.data.readers.DelimitedReader

Reader for creating a FeatureSet out of a CSV file.

If you would like to include example/instance IDs in your files, they must be specified as an“id” column.

Also, there must be a column with the name specified by label_col if the data is labelled.

class skll.data.readers.DelimitedReader(path_or_list, **kwargs)Bases: skll.data.readers.Reader

Reader for creating a FeatureSet out of a delimited (CSV/TSV) file.

If you would like to include example/instance IDs in your files, they must be specified as an“id” column.

Also, for ARFF, CSV, and TSV files, there must be a column with the name specified bylabel_col if the data is labelled. For ARFF files, this column must also be the final one (as itis in Weka).

Parameters dialect (str) – The dialect of to pass on to the underlying CSVreader. Default: ‘excel-tab’

class skll.data.readers.DictListReader(path_or_list, quiet=True,ids_to_floats=False, label_col=u’y’,class_map=None, sparse=True,feature_hasher=False,num_features=None)

Bases: skll.data.readers.Reader

This class is to facilitate programmatic use of Learner.predict() and other functionsthat take FeatureSet objects as input. It iterates over examples in the same way as other‘‘Reader‘‘s, but uses a list of example dictionaries instead of a path to a file.

1.6. Python API 29

SciKit-Learn Laboratory Documentation, Release 0.28.0

Parameters

• path_or_list (Iterable of dict) – List of example dictionaries.

• quiet (bool) – Do not print “Loading...” status message to stderr.

• ids_to_floats (bool) – Convert IDs to float to save memory. Will raiseerror if we encounter an a non-numeric ID.

class skll.data.readers.LibSVMReader(path_or_list, quiet=True,ids_to_floats=False, la-bel_col=u’y’, class_map=None,sparse=True, feature_hasher=False,num_features=None)

Bases: skll.data.readers.Reader

Reader to create a FeatureSet out ouf a LibSVM/LibLinear/SVMLight file.

We use a specially formatted comment for storing example IDs, class names, and featurenames, which are normally not supported by the format. The comment is not mandatory,but without it, your classes and features will not have names. The comment is structured asfollows:

ExampleID | 1=FirstClass | 1=FirstFeature 2=SecondFeature

class skll.data.readers.MegaMReader(path_or_list, quiet=True,ids_to_floats=False, la-bel_col=u’y’, class_map=None,sparse=True, feature_hasher=False,num_features=None)

Bases: skll.data.readers.Reader

Reader to create a FeatureSet out ouf a MegaM -fvals file.

If you would like to include example/instance IDs in your files, they must be specified as acomment line directly preceding the line with feature values.

class skll.data.readers.NDJReader(path_or_list, quiet=True,ids_to_floats=False, label_col=u’y’,class_map=None, sparse=True, fea-ture_hasher=False, num_features=None)

Bases: skll.data.readers.Reader

Reader to create a FeatureSet out of a .jsonlines/.ndj file

If you would like to include example/instance IDs in your files, they must be specified in thefollowing ways as an “id” key in each JSON dictionary.

class skll.data.readers.Reader(path_or_list, quiet=True, ids_to_floats=False,label_col=u’y’, class_map=None, sparse=True,feature_hasher=False, num_features=None)

Bases: object

30 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

A little helper class to make picklable iterators out of example dictionary generators

Parameters

• path_or_list (str or list of dict) – Path or a list of example dictionaries.

• quiet (bool) – Do not print “Loading...” status message to stderr.

• ids_to_floats (bool) – Convert IDs to float to save memory. Will raiseerror if we encounter an a non-numeric ID.

• label_col (str) – Name of the column which contains the class labels forARFF/CSV/TSV files. If no column with that name exists, or None isspecified, the data is considered to be unlabelled.

• class_map (dict from str to str) – Mapping from original class labelsto new ones. This is mainly used for collapsing multiple classes into asingle class. Anything not in the mapping will be kept the same.

• sparse (bool) – Whether or not to store the features in a numpy CSRmatrix when using a DictVectorizer to vectorize the features.

• feature_hasher (bool) – Whether or not a FeatureHasher should be usedto vectorize the features.

• num_features (int) – If using a FeatureHasher, how many featuresshould the resulting matrix have? You should set this to at least twicethe number of features you have to avoid collisions.

read()Loads examples in the .arff, .csv, .jsonlines, .libsvm, .megam, .ndj, or.tsv formats.

Returns FeatureSet representing the file we read in.

class skll.data.readers.TSVReader(path_or_list, **kwargs)Bases: skll.data.readers.DelimitedReader

Reader for creating a FeatureSet out of a TSV file.

If you would like to include example/instance IDs in your files, they must be specified as an“id” column.

Also there must be a column with the name specified by label_col if the data is labelled.

skll.data.readers.convert_examples(example_dicts, sparse=True,ids_to_floats=False)

This function is to facilitate programmatic use of Learner.predict() and other functions thattake FeatureSet objects as input. It converts a .jsonlines/.ndj-style list of dictionaries into anFeatureSet.

Parameters example_dicts (iterable of dicts) – An list of dictionaries followingthe .jsonlines/.ndj format (i.e., features ‘x’, label ‘y’, and ‘id’).

1.6. Python API 31

SciKit-Learn Laboratory Documentation, Release 0.28.0

Returns an FeatureSet representing the examples in example_dicts.

skll.data.readers.load_examples(path, quiet=False, sparse=True, la-bel_col=u’y’, ids_to_floats=False,class_map=None, feature_hasher=False,num_features=None)

Loads examples in the .arff, .csv, .jsonlines, .libsvm, .megam, .ndj, or.tsv formats.

If you would like to include example/instance IDs in your files, they must be specified in thefollowing ways:

•MegaM: As a comment line directly preceding the line with feature values.

•LibSVM: As the first item in the three-part comment described below.

•CSV/TSV/ARFF: An “id” column.

•JSONLINES: An “id” key in each JSON dictionary.

Also, for ARFF, CSV, and TSV files, there must be a column with the name specified bylabel_col if the data is labelled. For ARFF files, this column must also be the final one (as itis in Weka).

For LibSVM files, we use a specially formatted comment for storing example IDs, classnames, and feature names, which are normally not supported by the format. The comment isnot mandatory, but without it, your classes and features will not have names. The commentis structured as follows:

ExampleID | 1=FirstClass | 1=FirstFeature 2=SecondFeature

Parameters

• path (str or dict) – The path to the file to load the examples from, or alist of example dictionaries (like you would pass to convert_examples).

• quiet (bool) – Do not print “Loading...” status message to stderr.

• sparse (bool) – Whether or not to store the features in a numpy CSRmatrix.

• label_col (str) – Name of the column which contains the class labels forARFF/CSV/TSV files. If no column with that name exists, or None isspecified, the data is considered to be unlabelled.

• ids_to_floats (bool) – Convert IDs to float to save memory. Will raiseerror if we encounter an a non-numeric ID.

• class_map (dict from str to str) – Mapping from original class labelsto new ones. This is mainly used for collapsing multiple classes into asingle class. Anything not in the mapping will be kept the same.

32 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

Returns 4-tuple of an array example ids, an array of class labels, a scipy CSR ma-trix of features, and a DictVectorizer containing the mapping between featurenames and the column indices in the feature matrix.

skll.data.readers.safe_float(text, replace_dict=None)Attempts to convert a string to a float, but if that’s not possible, just returns the original value.

Parameters

• text (str) – The text to convert.

• replace_dict (dict from str to str) – Mapping from text to replacementtext values. This is mainly used for collapsing multiple classes into asingle class. Replacing happens before conversion to floats. Anythingnot in the mapping will be kept the same.

data.writers Module

Handles loading data from various types of data files.

author Dan Blanchard ([email protected])

author Michael Heilman ([email protected])

author Nitin Madnani ([email protected])

organization ETS

class skll.data.writers.ARFFWriter(path, feature_set, **kwargs)Bases: skll.data.writers.DelimitedFileWriter

FeatureSetWriter for writing out FeatureSets as ARFF files.

Parameters

• path (str) – A path to the feature file we would like to create. Ifsubsets is not None, this is assumed to be a string containing the pathto the directory to write the feature files with an additional file extensionspecifying the file type. For example /foo/.csv.

• feature_set (FeatureSet) – The FeatureSet to dump a file.

• requires_binary (bool) – Whether or not the FeatureSetWriter mustopen the file in binary mode for writing with Python 2.

• quiet (bool) – Do not print “Writing...” status message to stderr.

• relation (str) – The name of the relation in the ARFF file. Default:’skll_relation’

• regression (bool) – Is this an ARFF file to be used for regression? De-fault: False

1.6. Python API 33

SciKit-Learn Laboratory Documentation, Release 0.28.0

class skll.data.writers.CSVWriter(path, feature_set, **kwargs)Bases: skll.data.writers.DelimitedFileWriter

FeatureSetWriter for writing out FeatureSets as TSV files.

Parameters

• path (str) – A path to the feature file we would like to create. Ifsubsets is not None, this is assumed to be a string containing the pathto the directory to write the feature files with an additional file extensionspecifying the file type. For example /foo/.csv.

• feature_set (FeatureSet) – The FeatureSet to dump a file.

• quiet (bool) – Do not print “Writing...” status message to stderr.

class skll.data.writers.DelimitedFileWriter(path, feature_set, **kwargs)Bases: skll.data.writers.FeatureSetWriter

FeatureSetWriter for writing out FeatureSets as TSV/CSV files.

Parameters

• path (str) – A path to the feature file we would like to create. Ifsubsets is not None, this is assumed to be a string containing the pathto the directory to write the feature files with an additional file extensionspecifying the file type. For example /foo/.csv.

• feature_set (FeatureSet) – The FeatureSet to dump to a file.

• quiet (bool) – Do not print “Writing...” status message to stderr.

• label_col (str) – Name of the column which contains the class labels forCSV/TSV files. If no column with that name exists, or None is specified,the data is considered to be unlabelled.

• dialect (str) – The dialect to use for the underlying csv.DictWriterDefault: ‘excel-tab’

class skll.data.writers.FeatureSetWriter(path, feature_set, **kwargs)Bases: object

Helper class for writing out FeatureSets to files.

Parameters

• path (str) – A path to the feature file we would like to create. The suf-fix to this filename must be .arff, .csv, .jsonlines, .libsvm,.megam, .ndj, or .tsv. If subsets is not None, path is assumed tobe a string containing the path to the directory to write the feature fileswith an additional file extension specifying the file type. For example/foo/.csv.

• feature_set (FeatureSet) – The FeatureSet to dump a file.

34 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

• requires_binary (bool) – Whether or not the FeatureSetWriter mustopen the file in binary mode for writing with Python 2.

• quiet (bool) – Do not print “Writing...” status message to stderr.

write(subsets=None)Writes out this FeatureSetWriter’s FeatureSet to a file in its format.

Parameters subsets (dict (str to list of str)) – A mapping from subset namesto lists of feature names that are included in those sets. If given, a fea-ture file will be written for every subset (with the name containing thesubset name as suffix to path). Note, since string- valued features areautomatically converted into boolean features with names of the formFEATURE_NAME=STRING_VALUE, when doing the filtering, the por-tion before the = is all that’s used for matching. Therefore, you do notneed to enumerate all of these boolean feature names in your mapping.

class skll.data.writers.LibSVMWriter(path, feature_set, **kwargs)Bases: skll.data.writers.FeatureSetWriter

FeatureSetWriter for writing out FeatureSets as LibSVM/SVMLight files.

class skll.data.writers.MegaMWriter(path, feature_set, **kwargs)Bases: skll.data.writers.FeatureSetWriter

FeatureSetWriter for writing out FeatureSets as MegaM files.

class skll.data.writers.NDJWriter(path, feature_set, **kwargs)Bases: skll.data.writers.FeatureSetWriter

FeatureSetWriter for writing out FeatureSets as .jsonlines/.ndj files.

class skll.data.writers.TSVWriter(path, feature_set, **kwargs)Bases: skll.data.writers.DelimitedFileWriter

FeatureSetWriter for writing out FeatureSets as TSV files.

Parameters

• path (str) – A path to the feature file we would like to create. Ifsubsets is not None, this is assumed to be a string containing the pathto the directory to write the feature files with an additional file extensionspecifying the file type. For example /foo/.csv.

• feature_set (FeatureSet) – The FeatureSet to dump a file.

• quiet (bool) – Do not print “Writing...” status message to stderr.

1.6. Python API 35

SciKit-Learn Laboratory Documentation, Release 0.28.0

skll.data.writers.write_feature_file(path, ids, classes, fea-tures, feat_vectorizer=None,id_prefix=u’EXAMPLE_’,label_col=u’y’,arff_regression=False,arff_relation=u’skll_relation’,subsets=None, label_map=None)

Writes a feature file in either .arff, .csv, .jsonlines, .megam, .ndj, or .tsvformats with the given a list of IDs, classes, and features.

Parameters

• path (str) – A path to the feature file we would like to create. The suf-fix to this filename must be .arff, .csv, .jsonlines, .libsvm,.megam, .ndj, or .tsv. If subsets is not None, this is assumed tobe a string containing the path to the directory to write the feature fileswith an additional file extension specifying the file type. For example/foo/.csv.

• ids (list of str) – The IDs for each instance in the feature list/array. IfNone, IDs will be automatically generated with the prefix specified byid_prefix followed by the row number. If id_prefix is also None, no IDswill be written to the file.

• classes (list of str) – The class labels for each instance in the featurelist/array. If None, no class labels will be added to output file.

• features (list of dict or array-like) – The features for each instance rep-resented as either a list of dictionaries or an array-like (if feat_vectorizeris also specified).

• feat_vectorizer (DictVectorizer) – A DictVectorizer to map to/from fea-ture columns indices and names.

• label_col (str) – Name of the column which contains the class labels forCSV/TSV files. If no column with that name exists, or None is specified,the data is considered to be unlabelled.

• arff_regression (bool) – A boolean value indicating whether the ARFFfiles that are written should be written for regression rather than classifi-cation, i.e., the class variable y is numerical rather than an enumerationof classes and all non-numeric attributes are removed.

• arff_relation (str) – Relation name for ARFF file.

• subsets (dict (str to list of str)) – A mapping from subset names tolists of feature names that are included in those sets. If given, a fea-ture file will be written for every subset (with the name containing thesubset name as suffix to path). Note, since string- valued features areautomatically converted into boolean features with names of the form

36 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

FEATURE_NAME=STRING_VALUE, when doing the filtering, the por-tion before the = is all that’s used for matching. Therefore, you do notneed to enumerate all of these boolean feature names in your mapping.

• label_map (dict (str -> int)) – Mapping from class name to numbers touse for writing LibSVM files.

1.6.3 experiments Module

Functions related to running experiments and parsing configuration files.

author Dan Blanchard ([email protected])

author Michael Heilman ([email protected])

author Nitin Madnani ([email protected])

author Chee Wee Leong ([email protected])

skll.experiments.run_ablation(config_path, local=False, overwrite=True,queue=u’all.q’, hosts=None, quiet=False,all_combos=False)

Takes a configuration file and runs repeated experiments where each feature set has beenremoved from the configuration.

Parameters

• config_path (str) – Path to the configuration file we would like to use.

• local (bool) – Should this be run locally instead of on the cluster?

• overwrite (bool) – If the model files already exist, should we overwritethem instead of re-using them?

• queue (str) – The DRMAA queue to use if we’re running on the cluster.

• hosts (list of str) – If running on the cluster, these are the machines weshould use.

• quiet (bool) – Suppress printing of “Loading...” messages.

• all_combos (bool) – By default we only exclude one feature from set,but if all_combos is True, we do a true ablation study with all featurecombinations.

Deprecated since version 0.20.0: Use run_configuration() with the ablation argu-ment instead.

1.6. Python API 37

SciKit-Learn Laboratory Documentation, Release 0.28.0

skll.experiments.run_configuration(config_file, local=False, over-write=True, queue=u’all.q’,hosts=None, write_summary=True,quiet=False, ablation=0, re-sume=False)

Takes a configuration file and runs the specified jobs on the grid.

Parameters

• config_path (str) – Path to the configuration file we would like to use.

• local (bool) – Should this be run locally instead of on the cluster?

• overwrite (bool) – If the model files already exist, should we overwritethem instead of re-using them?

• queue (str) – The DRMAA queue to use if we’re running on the cluster.

• hosts (list of str) – If running on the cluster, these are the machines weshould use.

• write_summary (bool) – Write a tsv file with a summary of the results.

• quiet (bool) – Suppress printing of “Loading...” messages.

• ablation (int or None) – Number of features to remove when doing anablation experiment. If positive, we will perform repeated ablation runsfor all combinations of features removing the specified number at a time.If None, we will use all combinations of all lengths. If 0, the default, noablation is performed. If negative, a ValueError is raised.

• resume (bool) – If result files already exist for an experiment, do notoverwrite them. This is very useful when doing a large ablation experi-ment and part of it crashes.

Returns A list of paths to .json results files for each variation in the experiment.

Return type list of str

1.6.4 learner Module

Provides easy-to-use wrapper around scikit-learn.

author Michael Heilman ([email protected])

author Nitin Madnani ([email protected])

author Dan Blanchard ([email protected])

author Aoife Cahill ([email protected])

organization ETS

38 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

class skll.learner.FilteredLeaveOneLabelOut(labels, keep, examples)Bases: sklearn.cross_validation.LeaveOneLabelOut

Version of LeaveOneLabelOut cross-validation iterator that only outputs indices of instanceswith IDs in a prespecified set.

class skll.learner.Learner(model_type, probability=False, fea-ture_scaling=u’none’, model_kwargs=None,pos_label_str=None, min_feature_count=1,sampler=None, sampler_kwargs=None, cus-tom_learner_path=None)

Bases: object

A simpler learner interface around many scikit-learn classification and regression functions.

Parameters

• model_type (str) – Type of estimator to create (e.g., LogisticRegres-sion). See the skll package documentation for valid options.

• probability (bool) – Should learner return probabilities of all classes(instead of just class with highest probability)?

• model_kwargs (dict) – A dictionary of keyword arguments to pass tothe initializer for the specified model.

• pos_label_str (str) – The string for the positive class in the binary clas-sification setting. Otherwise, an arbitrary class is picked.

• feature_scaling (str) – how to scale the features, if at all. Options are:‘with_std’: scale features using the standard deviation, ‘with_mean’:center features using the mean, ‘both’: do both scaling as well as center-ing, ‘none’: do neither scaling nor centering

• min_feature_count (int) – The minimum number of examples a featuremust have a nonzero value in to be included.

• custom_learner_path (str) – Path to code where a custom classifier isdefined.

cross_validate(examples, stratified=True, cv_folds=10,grid_search=False, grid_search_folds=3, grid_jobs=None,grid_objective=u’f1_score_micro’, prediction_prefix=None,param_grid=None, shuffle=False, feature_hasher=False)

Cross-validates a given model on the training examples.

Parameters

• examples (FeatureSet) – The data to cross-validate learner perfor-mance on.

1.6. Python API 39

SciKit-Learn Laboratory Documentation, Release 0.28.0

• stratified (bool) – Should we stratify the folds to ensure an even distri-bution of classes for each fold?

• cv_folds (int or dict) – The number of folds to use for cross-validation,or a mapping from example IDs to folds.

• grid_search (bool) – Should we do grid search when training eachfold? Note: This will make this take much longer.

• grid_search_folds (int) – The number of folds to use when doing thegrid search (ignored if cv_folds is set to a dictionary mapping examplesto folds).

• grid_jobs (int) – The number of jobs to run in parallel when doing thegrid search. If unspecified or 0, the number of grid search folds will beused.

• grid_objective (function) – The objective function to use when doingthe grid search.

• param_grid (list of dicts mapping from strs to lists of parameter val-ues) – The parameter grid to search through for grid search. If unspec-ified, a default parameter grid will be used.

• prediction_prefix (str) – If saving the predictions, this is the prefix thatwill be used for the filename. It will be followed by ”.predictions”

• shuffle (bool) – Shuffle examples before splitting into folds for CV.

• feature_hasher (bool) – Are we using feature_hasher?

Returns The confusion matrix, overall accuracy, per-class PRFs, and modelparameters for each fold.

Return type list of 4-tuples

evaluate(examples, prediction_prefix=None, append=False,grid_objective=None, feature_hasher=False)

Evaluates a given model on a given dev or test example set.

Parameters

• examples (FeatureSet) – The examples to evaluate the performance ofthe model on.

• prediction_prefix (str) – If saving the predictions, this is the prefix thatwill be used for the filename. It will be followed by ”.predictions”

• append (bool) – Should we append the current predictions to the file ifit exists?

• grid_objective (function) – The objective function that was used whendoing the grid search.

40 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

• feature_hasher (bool) – are we using a feature_hasher?

Returns The confusion matrix, the overall accuracy, the per-class PRFs, themodel parameters, and the grid search objective function score.

Return type 5-tuple

classmethod from_file(learner_path)

Returns New instance of Learner from the pickle at the specified path.

load(learner_path)Replace the current learner instance with a saved learner.

Parameters learner_path (str) – The path to the file to load.

modelThe underlying scikit-learn model

model_kwargsA dictionary of the underlying scikit-learn model’s keyword arguments

model_paramsModel parameters (i.e., weights) for LinearModel (e.g., Ridge) regression andliblinear models.

Returns Labeled weights and (labeled if more than one) intercept value(s)

Return type tuple of (weights, intercepts), where weights is a dictand intercepts is a dictionary

model_typeThe model type (i.e., the class)

predict(examples, prediction_prefix=None, append=False, class_labels=False,feature_hasher=False)

Uses a given model to generate predictions on a given data set

Parameters

• examples (FeatureSet) – The examples to predict the classes for.

• prediction_prefix (str) – If saving the predictions, this is the prefix thatwill be used for the filename. It will be followed by ”.predictions”

• append (bool) – Should we append the current predictions to the file ifit exists?

• class_labels (bool) – For classifier, should we convert class indices totheir (str) labels?

• feature_hasher (bool) – For classifier, are we using a feature_hasher?

Returns The predictions returned by the learner.

1.6. Python API 41

SciKit-Learn Laboratory Documentation, Release 0.28.0

Return type array

probabilityShould learner return probabilities of all classes (instead of just class with highest prob-ability)?

save(learner_path)Save the learner to a file.

Parameters learner_path (str) – The path to where you want to save thelearner.

train(examples, param_grid=None, grid_search_folds=3, grid_search=True,grid_objective=u’f1_score_micro’, grid_jobs=None, shuffle=False, fea-ture_hasher=False, run_create_label_dict=True)

Train a classification model and return the model, score, feature vectorizer, scaler, labeldictionary, and inverse label dictionary.

Parameters

• examples (FeatureSet) – The examples to train the model on.

• param_grid (list of dicts mapping from strs to lists of parameter val-ues) – The parameter grid to search through for grid search. If unspec-ified, a default parameter grid will be used.

• grid_search_folds (int or dict) – The number of folds to use whendoing the grid search, or a mapping from example IDs to folds.

• grid_search (bool) – Should we do grid search?

• grid_objective (function) – The objective function to use when doingthe grid search.

• grid_jobs (int) – The number of jobs to run in parallel when doing thegrid search. If unspecified or 0, the number of grid search folds will beused.

• shuffle (bool) – Shuffle examples (e.g., for grid search CV.)

Returns The best grid search objective function score, or 0 if we’re not doinggrid search.

Return type float

class skll.learner.SelectByMinCount(min_count=1)Bases: sklearn.feature_selection.univariate_selection.SelectKBest

Select features ocurring in more (and/or fewer than) than a specified number of examples inthe training data (or a CV training fold).

skll.learner.rescaled(cls)Decorator to create regressors that store a min and a max for the training data and make sure

42 Chapter 1. Documentation

SciKit-Learn Laboratory Documentation, Release 0.28.0

that predictions fall within that range. It also stores the means and SDs of the gold standardand the predictions on the training set to rescale the predictions (e.g., as in e-rater).

Parameters cls (BaseEstimator) – A regressor to add rescaling to.

Returns Modified version of class with rescaled functions added.

1.6.5 metrics Module

This module contains a bunch of evaluation metrics that can be used to evaluate the performanceof learners.

author Michael Heilman ([email protected])

author Nitin Madnani ([email protected])

author Dan Blanchard ([email protected])

organization ETS

skll.metrics.f1_score_least_frequent(y_true, y_pred)Optimize the hyperparameter values during the grid search based on the F1 measure of theleast frequent class.

This is mostly intended for use when you’re doing binary classification and your data ishighly skewed. You should probably use f1_score_macro if your data is skewed and you’redoing multi-class classification.

skll.metrics.kappa(y_true, y_pred, weights=None, allow_off_by_one=False)Calculates the kappa inter-rater agreement between two the gold standard and the predictedratings. Potential values range from -1 (representing complete disagreement) to 1 (repre-senting complete agreement). A kappa value of 0 is expected if all agreement is due tochance.

In the course of calculating kappa, all items in y_true and y_pred will first be converted tofloats and then rounded to integers.

It is assumed that y_true and y_pred contain the complete range of possible ratings.

This function contains a combination of code from yorchopolis’s kappa-stats and Ben Ham-ner’s Metrics projects on Github.

Parameters

• weights (str or numpy array) – Specifies the weight matrix for the cal-culation. Options are:

– None = unweighted-kappa

– ‘quadratic’ = quadratic-weighted kappa

– ‘linear’ = linear-weighted kappa

1.6. Python API 43

SciKit-Learn Laboratory Documentation, Release 0.28.0

– two-dimensional numpy array = a custom matrix of weights. Eachweight corresponds to the 𝑤𝑖𝑗 values in the wikipedia description ofhow to calculate weighted Cohen’s kappa.

• allow_off_by_one (bool) – If true, ratings that are off by one are countedas equal, and all other differences are reduced by one. For example, 1 and2 will be considered to be equal, whereas 1 and 3 will have a differenceof 1 for when building the weights matrix.

skll.metrics.kendall_tau(y_true, y_pred)Optimize the hyperparameter values during the grid search based on Kendall’s tau.

This is useful in cases where you want to use the actual probabilities of the different classesafter the fact, and not just the optimize based on the classification accuracy.

skll.metrics.pearson(y_true, y_pred)Optimize the hyperparameter values during the grid search based on Pearson correlation.

skll.metrics.spearman(y_true, y_pred)Optimize the hyperparameter values during the grid search based on Spearman rank correla-tion.

This is useful in cases where you want to use the actual probabilities of the different classesafter the fact, and not just the optimize based on the classification accuracy.

skll.metrics.use_score_func(func_name, y_true, y_pred)Call the scoring function in sklearn.metrics.SCORERS with the given name. This takescare of handling keyword arguments that were pre-specified when creating the scorer. Thisapplies any sign-flipping that was specified by make_scorer when the scorer was created.

44 Chapter 1. Documentation

CHAPTER 2

Indices and tables

• genindex

• modindex

• search

45

SciKit-Learn Laboratory Documentation, Release 0.28.0

46 Chapter 2. Indices and tables

Python Module Index

sskll.data.featureset, 26skll.data.readers, 28skll.data.writers, 33skll.experiments, 37skll.learner, 38skll.metrics, 43

47

SciKit-Learn Laboratory Documentation, Release 0.28.0

48 Python Module Index

Index

AARFFReader (class in skll.data.readers), 29ARFFWriter (class in skll.data.writers), 33

Cconvert_examples() (in module skll), 18convert_examples() (in module

skll.data.readers), 31cross_validate() (skll.Learner method), 22cross_validate() (skll.learner.Learner method),

39CSVReader (class in skll.data.readers), 29CSVWriter (class in skll.data.writers), 33

DDelimitedFileWriter (class in skll.data.writers),

34DelimitedReader (class in skll.data.readers), 29DictListReader (class in skll.data.readers), 29

Eevaluate() (skll.Learner method), 23evaluate() (skll.learner.Learner method), 40ExamplesTuple (class in skll.data.featureset), 26

Ff1_score_least_frequent() (in module skll), 25f1_score_least_frequent() (in module

skll.metrics), 43feat_vectorizer (skll.data.featureset.FeatureSet

attribute), 27FeatureSet (class in skll.data.featureset), 27FeatureSetWriter (class in skll.data.writers), 34

filter() (skll.data.featureset.FeatureSet method),27

filtered_iter() (skll.data.featureset.FeatureSetmethod), 28

FilteredLeaveOneLabelOut (class inskll.learner), 38

from_file() (skll.Learner class method), 23from_file() (skll.learner.Learner class method),

41

Hhas_classes (skll.data.featureset.FeatureSet at-

tribute), 28

Kkappa() (in module skll), 25kappa() (in module skll.metrics), 43kendall_tau() (in module skll), 26kendall_tau() (in module skll.metrics), 44

LLearner (class in skll), 21Learner (class in skll.learner), 39LibSVMReader (class in skll.data.readers), 30LibSVMWriter (class in skll.data.writers), 35load() (skll.Learner method), 23load() (skll.learner.Learner method), 41load_examples() (in module skll), 18load_examples() (in module skll.data.readers),

32

MMegaMReader (class in skll.data.readers), 30MegaMWriter (class in skll.data.writers), 35

49

SciKit-Learn Laboratory Documentation, Release 0.28.0

model (skll.Learner attribute), 23model (skll.learner.Learner attribute), 41model_kwargs (skll.Learner attribute), 23model_kwargs (skll.learner.Learner attribute),

41model_params (skll.Learner attribute), 24model_params (skll.learner.Learner attribute),

41model_type (skll.Learner attribute), 24model_type (skll.learner.Learner attribute), 41

NNDJReader (class in skll.data.readers), 30NDJWriter (class in skll.data.writers), 35

Ppearson() (in module skll), 26pearson() (in module skll.metrics), 44predict() (skll.Learner method), 24predict() (skll.learner.Learner method), 41probability (skll.Learner attribute), 24probability (skll.learner.Learner attribute), 42

Rread() (skll.data.readers.Reader method), 31Reader (class in skll.data.readers), 30rescaled() (in module skll.learner), 42run_ablation() (in module skll), 20run_ablation() (in module skll.experiments), 37run_configuration() (in module skll), 21run_configuration() (in module

skll.experiments), 37

Ssafe_float() (in module skll.data.readers), 33save() (skll.Learner method), 24save() (skll.learner.Learner method), 42SelectByMinCount (class in skll.learner), 42skll.data.featureset (module), 26skll.data.readers (module), 28skll.data.writers (module), 33skll.experiments (module), 37skll.learner (module), 38skll.metrics (module), 43spearman() (in module skll), 26

spearman() (in module skll.metrics), 44split_with_quotes()

(skll.data.readers.ARFFReader staticmethod), 29

Ttrain() (skll.Learner method), 24train() (skll.learner.Learner method), 42TSVReader (class in skll.data.readers), 31TSVWriter (class in skll.data.writers), 35

Uuse_score_func() (in module skll.metrics), 44

Wwrite() (skll.data.writers.FeatureSetWriter

method), 35write_feature_file() (in module skll), 19write_feature_file() (in module

skll.data.writers), 35

50 Index