Jubatus

From Wikipedia, the free encyclopedia
Jubatus
Developer(s)Nippon Telegraph and Telephone &
Stable release
0.4.3 / April 19, 2013 (2013-04-19)
Written inC++
Operating systemLinux
Typemachine learning
LicenseGNU Lesser General Public License 2.1
Websitejubat.us/en/

Jubatus is an open-source online machine learning and distributed computing framework developed at Nippon Telegraph and Telephone and . Its features include classification, recommendation, regression, anomaly detection and graph mining. It supports many client languages, including C++, Java, Ruby and Python. It uses Iterative Parameter Mixture[1][2] for distributed machine learning.

Notable Features[]

Jubatus supports:

References[]

  1. ^ Ryan McDonald, K. Hall and G. Mann, Distributed Training Strategies for the Structured Perceptron, North American Association for Computational Linguistics (NAACL), 2010.
  2. ^ Gideon Mann, R. McDonald, M. Mohri, N. Silberman, and D. Walker, Efficient Large-Scale Distributed Training of Conditional Maximum Entropy Models, Neural Information Processing Systems (NIPS), 2009.
  3. ^ Crammer, Koby; Dekel, Ofer; Shalev-Shwartz, Shai; Singer, Yoram (2003). Online Passive-Aggressive Algorithms. Proceedings of the Sixteenth Annual Conference on Neural Information Processing Systems (NIPS).
  4. ^ Koby Crammer and Yoram Singer. Ultraconservative online algorithms for multiclass problems. Journal of Machine Learning Research, 2003.
  5. ^ Koby Crammer, Ofer Dekel, Joseph Keshet, Shai Shalev-Shwartz, Yoram Singer, Online Passive-Aggressive Algorithms. Journal of Machine Learning Research, 2006.
  6. ^ Mark Dredze, Koby Crammer and Fernando Pereira, Confidence-Weighted Linear Classification, Proceedings of the 25th International Conference on Machine Learning (ICML), 2008
  7. ^ Koby Crammer, Mark Dredze and Fernando Pereira, Exact Convex Confidence-Weighted Learning, Proceedings of the Twenty Second Annual Conference on Neural Information Processing Systems (NIPS), 2008
  8. ^ Koby Crammer, Mark Dredze and Alex Kulesza, Multi-Class Confidence Weighted Algorithms, Empirical Methods in Natural Language Processing (EMNLP), 2009
  9. ^ Koby Crammer, Alex Kulesza and Mark Dredze, Adaptive Regularization Of Weight Vectors, Advances in Neural Information Processing Systems, 2009
  10. ^ Koby Crammer and Daniel D. Lee, Learning via Gaussian Herding, Neural Information Processing Systems (NIPS), 2010.


Retrieved from ""