OpenML
Filter results by:
Imputation transformer for completing missing values.
0 runs0 likes0 downloads0 reach0 impact
DummyClassifier is a classifier that makes predictions using simple rules. This classifier is useful as a simple baseline to compare with other (real) classifiers. Do not use it for real problems.
0 runs0 likes0 downloads0 reach0 impact
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
1 runs0 likes0 downloads0 reach0 impact
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
1 runs0 likes0 downloads0 reach0 impact
Imputation transformer for completing missing values.
0 runs0 likes0 downloads0 reach0 impact
DummyClassifier is a classifier that makes predictions using simple rules. This classifier is useful as a simple baseline to compare with other (real) classifiers. Do not use it for real problems.
0 runs0 likes0 downloads0 reach0 impact
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
1 runs0 likes0 downloads0 reach0 impact
0 likes - 0 downloads - 0 reach - mean_absolute_error: 0.1486, mean_prior_absolute_error: 0.1491, number_of_instances: 2178, relative_absolute_error: 0.9962, root_mean_prior_squared_error: 0.1894, root_mean_squared_error: 0.189, root_relative_squared_error: 0.9981,
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
1 runs0 likes0 downloads0 reach0 impact
Imputation transformer for completing missing values.
0 runs0 likes0 downloads0 reach0 impact
Ordinary least squares Linear Regression. LinearRegression fits a linear model with coefficients w = (w1, ..., wp) to minimize the residual sum of squares between the observed targets in the dataset,…
0 runs0 likes0 downloads0 reach0 impact
Encode categorical features as an integer array. The input to this transformer should be an array-like of integers or strings, denoting the values taken on by categorical (discrete) features. The…
0 runs0 likes0 downloads0 reach0 impact
0 likes - 0 downloads - 0 reach - area_under_roc_curve: 0.5, kb_relative_information_score: 0.0048, mean_absolute_error: 0.4568, mean_prior_absolute_error: 0.4589, weighted_recall: 0.6364, number_of_instances: 253, predictive_accuracy: 0.6364, prior_entropy: 0.9463, relative_absolute_error: 0.9955, root_mean_prior_squared_error: 0.4813, root_mean_squared_error: 0.4815, root_relative_squared_error: 1.0006, unweighted_recall: 0.5,
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
1 runs0 likes0 downloads0 reach0 impact
Imputation transformer for completing missing values.
0 runs0 likes0 downloads0 reach0 impact
DummyClassifier is a classifier that makes predictions using simple rules. This classifier is useful as a simple baseline to compare with other (real) classifiers. Do not use it for real problems.
0 runs0 likes0 downloads0 reach0 impact
Encode categorical features as an integer array. The input to this transformer should be an array-like of integers or strings, denoting the values taken on by categorical (discrete) features. The…
0 runs0 likes0 downloads0 reach0 impact
Encode categorical features as a one-hot numeric array. The input to this transformer should be an array-like of integers or strings, denoting the values taken on by categorical (discrete) features.…
0 runs0 likes0 downloads0 reach0 impact
Standardize features by removing the mean and scaling to unit variance The standard score of a sample `x` is calculated as: z = (x - u) / s where `u` is the mean of the training samples or zero if…
0 runs0 likes0 downloads0 reach0 impact
Concatenates results of multiple transformer objects. This estimator applies a list of transformer objects in parallel to the input data, then concatenates the results. This is useful to combine…
0 runs0 likes0 downloads0 reach0 impact
Dimensionality reduction using truncated SVD (aka LSA). This transformer performs linear dimensionality reduction by means of truncated singular value decomposition (SVD). Contrary to PCA, this…
0 runs0 likes0 downloads0 reach0 impact
Select features according to a percentile of the highest scores.
0 runs0 likes0 downloads0 reach0 impact
An AdaBoost classifier. An AdaBoost [1] classifier is a meta-estimator that begins by fitting a classifier on the original dataset and then fits additional copies of the classifier on the same dataset…
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A Bagging classifier. A Bagging classifier is an ensemble meta-estimator that fits base classifiers each on random subsets of the original dataset and then aggregate their individual predictions…
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A Bagging classifier. A Bagging classifier is an ensemble meta-estimator that fits base classifiers each on random subsets of the original dataset and then aggregate their individual predictions…
0 runs0 likes0 downloads0 reach0 impact
An AdaBoost classifier. An AdaBoost [1] classifier is a meta-estimator that begins by fitting a classifier on the original dataset and then fits additional copies of the classifier on the same dataset…
0 runs0 likes0 downloads0 reach0 impact
Test study for the Python tutorial on studies
3 datasets, 3 tasks, 1 flows, 3 runs
Soft Voting/Majority Rule classifier for unfitted estimators. .. versionadded:: 0.17
0 runs0 likes0 downloads0 reach0 impact
0 likes - 0 downloads - 0 reach - area_under_roc_curve: 0.8083, f_measure: 0.8231, kappa: 0.4358, kb_relative_information_score: 0.2535, mean_absolute_error: 0.2187, mean_prior_absolute_error: 0.3266, weighted_recall: 0.8314, number_of_instances: 522, precision: 0.8196, predictive_accuracy: 0.8314, prior_entropy: 0.7318, relative_absolute_error: 0.6697, root_mean_prior_squared_error: 0.4037, root_mean_squared_error: 0.3467, root_relative_squared_error: 0.8588, unweighted_recall: 0.6998,
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
Logistic Regression (aka logit, MaxEnt) classifier. In the multiclass case, the training algorithm uses the one-vs-rest (OvR) scheme if the 'multi_class' option is set to 'ovr', and uses the…
0 runs0 likes0 downloads0 reach0 impact
Soft Voting/Majority Rule classifier for unfitted estimators. .. versionadded:: 0.17
0 runs0 likes0 downloads0 reach0 impact
test description
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
0 likes - 0 downloads - 0 reach - area_under_roc_curve: 0.9945, f_measure: 0.8956, kappa: 0.8916, kb_relative_information_score: 0.6688, mean_absolute_error: 0.0694, mean_prior_absolute_error: 0.1212, weighted_recall: 0.902, number_of_instances: 500, precision: 0.9064, predictive_accuracy: 0.902, prior_entropy: 3.6489, relative_absolute_error: 0.5727, root_mean_prior_squared_error: 0.246, root_mean_squared_error: 0.1571, root_relative_squared_error: 0.6383, unweighted_recall: 0.804,
Test study for the Python tutorial on studies
3 datasets, 3 tasks, 1 flows, 3 runs
0 likes - 0 downloads - 0 reach - area_under_roc_curve: 0.8122, f_measure: 0.8279, kappa: 0.4446, kb_relative_information_score: 0.2504, mean_absolute_error: 0.2185, mean_prior_absolute_error: 0.3266, weighted_recall: 0.8391, number_of_instances: 522, precision: 0.8258, predictive_accuracy: 0.8391, prior_entropy: 0.7318, relative_absolute_error: 0.6691, root_mean_prior_squared_error: 0.4037, root_mean_squared_error: 0.346, root_relative_squared_error: 0.857, unweighted_recall: 0.6976,
0 likes - 0 downloads - 0 reach - area_under_roc_curve: 0.9936, f_measure: 0.8764, kappa: 0.874, kb_relative_information_score: 0.6635, mean_absolute_error: 0.0701, mean_prior_absolute_error: 0.1212, weighted_recall: 0.886, number_of_instances: 500, precision: 0.8828, predictive_accuracy: 0.886, prior_entropy: 3.6489, relative_absolute_error: 0.5784, root_mean_prior_squared_error: 0.246, root_mean_squared_error: 0.1586, root_relative_squared_error: 0.6445, unweighted_recall: 0.7701,
0 likes - 0 downloads - 0 reach - area_under_roc_curve: 0.8297, f_measure: 0.7721, kappa: 0.4922, kb_relative_information_score: 0.3136, mean_absolute_error: 0.3177, mean_prior_absolute_error: 0.4545, weighted_recall: 0.776, number_of_instances: 768, precision: 0.7715, predictive_accuracy: 0.776, prior_entropy: 0.9331, relative_absolute_error: 0.6991, root_mean_prior_squared_error: 0.4766, root_mean_squared_error: 0.3976, root_relative_squared_error: 0.8342, unweighted_recall: 0.7388,
0 likes - 0 downloads - 0 reach - area_under_roc_curve: 0.8256, f_measure: 0.7609, kappa: 0.468, kb_relative_information_score: 0.3124, mean_absolute_error: 0.3177, mean_prior_absolute_error: 0.4545, weighted_recall: 0.7643, number_of_instances: 768, precision: 0.7598, predictive_accuracy: 0.7643, prior_entropy: 0.9331, relative_absolute_error: 0.699, root_mean_prior_squared_error: 0.4766, root_mean_squared_error: 0.4, root_relative_squared_error: 0.8392, unweighted_recall: 0.7281,
Imputation transformer for completing missing values.
0 runs0 likes0 downloads0 reach0 impact
Encode categorical features as a one-hot numeric array. The input to this transformer should be an array-like of integers or strings, denoting the values taken on by categorical (discrete) features.…
0 runs0 likes0 downloads0 reach0 impact
Feature selector that removes all low-variance features. This feature selection algorithm looks only at the features (X), not the desired outputs (y), and can thus be used for unsupervised learning.
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Test suite for the Python tutorial on benchmark suites
20 datasets, 20 tasks, 0 flows, 0 runs
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Ordinary least squares Linear Regression. LinearRegression fits a linear model with coefficients w = (w1, ..., wp) to minimize the residual sum of squares between the observed targets in the dataset,…
0 runs0 likes0 downloads0 reach0 impact
Ordinary least squares Linear Regression. LinearRegression fits a linear model with coefficients w = (w1, ..., wp) to minimize the residual sum of squares between the observed targets in the dataset,…
0 runs0 likes0 downloads0 reach0 impact
A decision tree classifier.
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
A random forest classifier. A random forest is a meta estimator that fits a number of decision tree classifiers on various sub-samples of the dataset and uses averaging to improve the predictive…
0 runs0 likes0 downloads0 reach0 impact
Randomized search on hyper parameters. RandomizedSearchCV implements a "fit" and a "score" method. It also implements "score_samples", "predict", "predict_proba", "decision_function", "transform" and…
1 runs0 likes0 downloads0 reach0 impact
Ordinary least squares Linear Regression. LinearRegression fits a linear model with coefficients w = (w1, ..., wp) to minimize the residual sum of squares between the observed targets in the dataset,…
1 runs0 likes0 downloads0 reach0 impact
Standardize features by removing the mean and scaling to unit variance The standard score of a sample `x` is calculated as: z = (x - u) / s where `u` is the mean of the training samples or zero if…
0 runs0 likes0 downloads0 reach0 impact
DummyClassifier is a classifier that makes predictions using simple rules. This classifier is useful as a simple baseline to compare with other (real) classifiers. Do not use it for real problems.
0 runs0 likes0 downloads0 reach0 impact
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
0 runs0 likes0 downloads0 reach0 impact
Gaussian Naive Bayes (GaussianNB) Can perform online updates to model parameters via :meth:`partial_fit`. For details on algorithm used to update feature means and variance online, see Stanford CS…
0 runs0 likes0 downloads0 reach0 impact
Test suite for the Python tutorial on benchmark suites
20 datasets, 20 tasks, 0 flows, 0 runs
A random forest classifier. A random forest is a meta estimator that fits a number of decision tree classifiers on various sub-samples of the dataset and uses averaging to improve the predictive…
0 runs0 likes0 downloads0 reach0 impact
Exhaustive search over specified parameter values for an estimator. Important members are fit, predict. GridSearchCV implements a "fit" and a "score" method. It also implements "score_samples",…
1 runs0 likes0 downloads0 reach0 impact
Standardize features by removing the mean and scaling to unit variance The standard score of a sample `x` is calculated as: z = (x - u) / s where `u` is the mean of the training samples or zero if…
0 runs0 likes0 downloads0 reach0 impact
DummyClassifier is a classifier that makes predictions using simple rules. This classifier is useful as a simple baseline to compare with other (real) classifiers. Do not use it for real problems.
0 runs0 likes0 downloads0 reach0 impact
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
0 runs0 likes0 downloads0 reach0 impact
A random forest classifier. A random forest is a meta estimator that fits a number of decision tree classifiers on various sub-samples of the dataset and uses averaging to improve the predictive…
0 runs0 likes0 downloads0 reach0 impact
Exhaustive search over specified parameter values for an estimator. Important members are fit, predict. GridSearchCV implements a "fit" and a "score" method. It also implements "score_samples",…
1 runs0 likes0 downloads0 reach0 impact
Standardize features by removing the mean and scaling to unit variance The standard score of a sample `x` is calculated as: z = (x - u) / s where `u` is the mean of the training samples or zero if…
0 runs0 likes0 downloads0 reach0 impact
DummyClassifier is a classifier that makes predictions using simple rules. This classifier is useful as a simple baseline to compare with other (real) classifiers. Do not use it for real problems.
0 runs0 likes0 downloads0 reach0 impact
Pipeline of transforms with a final estimator. Sequentially apply a list of transforms and a final estimator. Intermediate steps of the pipeline must be 'transforms', that is, they must implement fit…
1 runs0 likes0 downloads0 reach0 impact
Logistic Regression (aka logit, MaxEnt) classifier. In the multiclass case, the training algorithm uses the one-vs-rest (OvR) scheme if the 'multi_class' option is set to 'ovr', and uses the…
0 runs0 likes0 downloads0 reach0 impact
Encode categorical features as a one-hot numeric array. The input to this transformer should be an array-like of integers or strings, denoting the values taken on by categorical (discrete) features.…
0 runs0 likes0 downloads0 reach0 impact
Standardize features by removing the mean and scaling to unit variance The standard score of a sample `x` is calculated as: z = (x - u) / s where `u` is the mean of the training samples or zero if…
0 runs0 likes0 downloads0 reach0 impact
Concatenates results of multiple transformer objects. This estimator applies a list of transformer objects in parallel to the input data, then concatenates the results. This is useful to combine…
0 runs0 likes0 downloads0 reach0 impact
Dimensionality reduction using truncated SVD (aka LSA). This transformer performs linear dimensionality reduction by means of truncated singular value decomposition (SVD). Contrary to PCA, this…
0 runs0 likes0 downloads0 reach0 impact
Select features according to a percentile of the highest scores.
0 runs0 likes0 downloads0 reach0 impact
An AdaBoost classifier. An AdaBoost [1] classifier is a meta-estimator that begins by fitting a classifier on the original dataset and then fits additional copies of the classifier on the same dataset…
0 runs0 likes0 downloads0 reach0 impact