Download JAR files tagged by quadratic with all dependencies
josqp from group com.quantego (version 0.6.5)
A Java port of the OSQP quadratic programming solver
Artifact josqp
Group com.quantego
Version 0.6.5
Last update 14. November 2023
Organization not specified
URL https://github.com/loehndorf/josqp
License MIT License
Dependencies amount 1
Dependencies junit-jupiter-engine,
There are maybe transitive dependencies!
Group com.quantego
Version 0.6.5
Last update 14. November 2023
Organization not specified
URL https://github.com/loehndorf/josqp
License MIT License
Dependencies amount 1
Dependencies junit-jupiter-engine,
There are maybe transitive dependencies!
Diversity from group com.github.sergejzr.lib (version 0.0.1)
We propose two efficient algorithms for exploring topic diversity in large document corpora such as user generated content on the social web, bibliographic data, or other web repositories. Analyzing diversity is useful for obtaining insights into knowledge evolution, trends, periodicities, and topic heterogeneity of such collections. Calculating diversity statistics requires averaging over the similarity of all object pairs, which, for large corpora, is prohibitive from a computational point of view. Our proposed algorithms overcome the quadratic complexity of the average pair-wise similarity computation, and allow for constant time (depending on dataset properties) or linear time approximation with probabilistic guarantees.
Artifact Diversity
Group com.github.sergejzr.lib
Version 0.0.1
Last update 09. September 2016
Organization not specified
URL http://dl.acm.org/citation.cfm?id=2398445
License The Apache License, Version 2.0
Dependencies amount 0
Dependencies No dependencies
There are maybe transitive dependencies!
Group com.github.sergejzr.lib
Version 0.0.1
Last update 09. September 2016
Organization not specified
URL http://dl.acm.org/citation.cfm?id=2398445
License The Apache License, Version 2.0
Dependencies amount 0
Dependencies No dependencies
There are maybe transitive dependencies!
dagging from group nz.ac.waikato.cms.weka (version 1.0.3)
This meta classifier creates a number of disjoint, stratified folds out of the data and feeds each chunk of data to a copy of the supplied base classifier. Predictions are made via majority vote, since all the generated base classifiers are put into the Vote meta classifier.
Useful for base classifiers that are quadratic or worse in time behavior, regarding number of instances in the training data.
For more information, see:
Ting, K. M., Witten, I. H.: Stacking Bagged and Dagged Models. In: Fourteenth international Conference on Machine Learning, San Francisco, CA, 367-375, 1997.
2 downloads
Artifact dagging
Group nz.ac.waikato.cms.weka
Version 1.0.3
Last update 29. April 2014
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/dagging
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!
Group nz.ac.waikato.cms.weka
Version 1.0.3
Last update 29. April 2014
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/dagging
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!
kernelLogisticRegression from group nz.ac.waikato.cms.weka (version 1.0.0)
This package contains a classifier that can be used to train a two-class kernel logistic regression model with the kernel functions that are available in WEKA. It optimises the negative log-likelihood with a quadratic penalty. Both, BFGS and conjugate gradient descent, are available as optimisation methods, but the former is normally faster. It is possible to use multiple threads, but the speed-up is generally very marginal when used with BFGS optimisation. With conjugate gradient descent optimisation, greater speed-ups can be achieved when using multiple threads. With the default kernel, the dot product kernel, this method produces results that are close to identical to those obtained using standard logistic regression in WEKA, provided a sufficiently large value for the parameter determining the size of the quadratic penalty is used in both cases.
0 downloads
Artifact kernelLogisticRegression
Group nz.ac.waikato.cms.weka
Version 1.0.0
Last update 26. June 2013
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/kernelLogisticRegression
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!
Group nz.ac.waikato.cms.weka
Version 1.0.0
Last update 26. June 2013
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/kernelLogisticRegression
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!
multiLayerPerceptrons from group nz.ac.waikato.cms.weka (version 1.0.10)
This package currently contains classes for training multilayer perceptrons with one hidden layer, where the number of hidden units is user specified. MLPClassifier can be used for classification problems and MLPRegressor is the corresponding class for numeric prediction tasks. The former has as many output units as there are classes, the latter only one output unit. Both minimise a penalised squared error with a quadratic penalty on the (non-bias) weights, i.e., they implement "weight decay", where this penalised error is averaged over all training instances. The size of the penalty can be determined by the user by modifying the "ridge" parameter to control overfitting. The sum of squared weights is multiplied by this parameter before added to the squared error. Both classes use BFGS optimisation by default to find parameters that correspond to a local minimum of the error function. but optionally conjugated gradient descent is available, which can be faster for problems with many parameters. Logistic functions are used as the activation functions for all units apart from the output unit in MLPRegressor, which employs the identity function. Input attributes are standardised to zero mean and unit variance. MLPRegressor also rescales the target attribute (i.e., "class") using standardisation. All network parameters are initialised with small normally distributed random values.
Group: nz.ac.waikato.cms.weka Artifact: multiLayerPerceptrons
Show all versions Show documentation Show source
Show all versions Show documentation Show source
10 downloads
Artifact multiLayerPerceptrons
Group nz.ac.waikato.cms.weka
Version 1.0.10
Last update 31. October 2016
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/multiLayerPerceptrons
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!
Group nz.ac.waikato.cms.weka
Version 1.0.10
Last update 31. October 2016
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/multiLayerPerceptrons
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!
Page 1 from 1 (items total 5)