All Downloads are FREE. Search and download functionalities are using the official Maven repository.

Download JAR files tagged by faster with all dependencies

Search JAR files by class name

annotation from group com.groupdocs (version 1.3.0)

GroupDocs.Annotation lets you add notes to PDF and Word documents, as well as to image files – all directly from a web browser. It is a convenient web-based tool that doesn’t require any software installation and allows you and your colleagues to annotate documents online. Moreover, with GroupDocs.Annotation, you can add your notes to a document and then send it for approval or review, or share the document with others for online collaborative review in real-time. This way you get feedback faster and can keep everyone’s notes and comments in a single file.

Group: com.groupdocs Artifact: annotation
Show all versions Show documentation Show source 
 

0 downloads
Artifact annotation
Group com.groupdocs
Version 1.3.0
Last update 09. June 2014
Organization not specified
URL http://maven.apache.org
License GroupDocs License, Version 1.0
Dependencies amount 4
Dependencies viewer, sqlite-jdbc, gson, atmosphere-runtime,
There are maybe transitive dependencies!

low-latency-primitive-concurrent-queues from group uk.co.boundedbuffer (version 1.0.0)

An low latency, lock free, primitive bounded blocking queue backed by an int[]. This class mimics the interface of {@linkplain java.util.concurrent.BlockingQueue BlockingQueue}, however works with primitive ints rather than objects, so is unable to actually implement the BlockingQueue. This class takes advantage of the Unsafe.putOrderedObject, which allows us to create non-blocking code with guaranteed writes. These writes will not be re-orderd by instruction reordering. Under the covers it uses the faster store-store barrier, rather than the the slower store-load barrier, which is used when doing a volatile write. One of the trade off with this improved performance is we are limited to a single producer, single consumer.

Group: uk.co.boundedbuffer Artifact: low-latency-primitive-concurrent-queues
Show documentation Show source 
 

0 downloads
Artifact low-latency-primitive-concurrent-queues
Group uk.co.boundedbuffer
Version 1.0.0
Last update 24. February 2014
Organization not specified
URL http://www.boundedbuffer.co.uk
License The Apache Software License, Version 2.0
Dependencies amount 2
Dependencies mockito-core, japex-maven-plugin,
There are maybe transitive dependencies!

kernelLogisticRegression from group nz.ac.waikato.cms.weka (version 1.0.0)

This package contains a classifier that can be used to train a two-class kernel logistic regression model with the kernel functions that are available in WEKA. It optimises the negative log-likelihood with a quadratic penalty. Both, BFGS and conjugate gradient descent, are available as optimisation methods, but the former is normally faster. It is possible to use multiple threads, but the speed-up is generally very marginal when used with BFGS optimisation. With conjugate gradient descent optimisation, greater speed-ups can be achieved when using multiple threads. With the default kernel, the dot product kernel, this method produces results that are close to identical to those obtained using standard logistic regression in WEKA, provided a sufficiently large value for the parameter determining the size of the quadratic penalty is used in both cases.

Group: nz.ac.waikato.cms.weka Artifact: kernelLogisticRegression
Show documentation Show source 
 

0 downloads
Artifact kernelLogisticRegression
Group nz.ac.waikato.cms.weka
Version 1.0.0
Last update 26. June 2013
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/kernelLogisticRegression
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!

algorithms from group de.cit-ec.tcs.alignment (version 3.1.1)

This module defines the interface for AlignmentAlgorithms as well as some helper classes. An AlignmentAlgorithm computes an Alignment of two given input sequences, given a Comparator that works in these sequences. More details on the AlignmentAlgorithm can be found in the respective interface. More information on Comparators can be found in the comparators module. The resulting 'Alignment' may be just a real-valued dissimilarity between the input sequence or may incorporate additional information, such as a full Alignment, a PathList, a PathMap or a CooptimalModel. If those results support the calculation of a Gradient, they implement the DerivableAlignmentDistance interface. In more detail, the Alignment class represents the result of a backtracing scheme, listing all Operations that have been applied in one co-optimal Alignment. A classic AlignmentAlgorithm does not result in a differentiable dissimilarity, because the minimum function is not differentiable. Therefore, this package also contains utility functions for a soft approximation of the minimum function, namely Softmin. For faster (parallel) computation of many different alignments or gradients we also provide the ParallelProcessingEngine, the SquareParallelProcessingEngine and the ParallelGradientEngine.

Group: de.cit-ec.tcs.alignment Artifact: algorithms
Show all versions Show documentation Show source 
 

0 downloads
Artifact algorithms
Group de.cit-ec.tcs.alignment
Version 3.1.1
Last update 26. October 2018
Organization not specified
URL http://openresearch.cit-ec.de/projects/tcs
License The GNU Affero General Public License, Version 3
Dependencies amount 3
Dependencies comparators, parallel, lombok,
There are maybe transitive dependencies!

multiLayerPerceptrons from group nz.ac.waikato.cms.weka (version 1.0.10)

This package currently contains classes for training multilayer perceptrons with one hidden layer, where the number of hidden units is user specified. MLPClassifier can be used for classification problems and MLPRegressor is the corresponding class for numeric prediction tasks. The former has as many output units as there are classes, the latter only one output unit. Both minimise a penalised squared error with a quadratic penalty on the (non-bias) weights, i.e., they implement "weight decay", where this penalised error is averaged over all training instances. The size of the penalty can be determined by the user by modifying the "ridge" parameter to control overfitting. The sum of squared weights is multiplied by this parameter before added to the squared error. Both classes use BFGS optimisation by default to find parameters that correspond to a local minimum of the error function. but optionally conjugated gradient descent is available, which can be faster for problems with many parameters. Logistic functions are used as the activation functions for all units apart from the output unit in MLPRegressor, which employs the identity function. Input attributes are standardised to zero mean and unit variance. MLPRegressor also rescales the target attribute (i.e., "class") using standardisation. All network parameters are initialised with small normally distributed random values.

Group: nz.ac.waikato.cms.weka Artifact: multiLayerPerceptrons
Show all versions Show documentation Show source 
 

10 downloads
Artifact multiLayerPerceptrons
Group nz.ac.waikato.cms.weka
Version 1.0.10
Last update 31. October 2016
Organization University of Waikato, Hamilton, NZ
URL http://weka.sourceforge.net/doc.packages/multiLayerPerceptrons
License GNU General Public License 3
Dependencies amount 1
Dependencies weka-dev,
There are maybe transitive dependencies!



Page 29 from 29 (items total 285)


© 2015 - 2024 Weber Informatics LLC | Privacy Policy