Jensen: An Easily-Extensible C++ Toolkit for Production-Level Machine Learning and Convex Optimization
2018-07-17Code Available0· sign in to hype
Rishabh Iyer, John T. Halloran, Kai Wei
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/rishabhk108/jensenOfficialIn papernone★ 0
- github.com/johnhalloran321/cvxpy_classifiersnone★ 0
Abstract
This paper introduces Jensen, an easily extensible and scalable toolkit for production-level machine learning and convex optimization. Jensen implements a framework of convex (or loss) functions, convex optimization algorithms (including Gradient Descent, L-BFGS, Stochastic Gradient Descent, Conjugate Gradient, etc.), and a family of machine learning classifiers and regressors (Logistic Regression, SVMs, Least Square Regression, etc.). This framework makes it possible to deploy and train models with a few lines of code, and also extend and build upon this by integrating new loss functions and optimization algorithms.