Tribuo v4.1 Release Notes

Tribuo 4.1 is the first feature release after the initial open source release. We’ve added new models, new parameters for some models, improvements to data loading, documentation, transformations and the speed of our CRF and linear models, along with a large update to the TensorFlow interface. We’ve also revised the tutorials and added two new ones covering TensorFlow and document classification.

TensorFlow support

Migrated to TensorFlow Java 0.3.1 which allows specification and training of models in Java (#134). The TensorFlow models can be saved in two formats, either using TensorFlow’s checkpoint format or Tribuo’s native model serialization. They can also be exported as TensorFlow Saved Models for interop with other TensorFlow platforms. Tribuo can now load TF v2 Saved Models and serve them alongside TF v1 frozen graphs with it’s external model loader.

We also added a TensorFlow tutorial which walks through the creation of a simple regression MLP, a classification MLP and a classification CNN, before exporting the model as a TensorFlow Saved Model and importing it back into Tribuo.

New models

  • Added extremely randomized trees, i.e., ExtraTrees (#51).
  • Added an SGD based linear model for multi-label classification (#106).
  • Added liblinear’s linear SVM anomaly detector (#114).
  • Added arbitrary ensemble creation from existing models (#129).

New features

  • Added K-Means++ (#34).
  • Added XGBoost feature importance metrics (#52).
  • Added OffsetDateTimeExtractor to the columnar data package (#66).
  • Added an empty response processor for use with clustering datasets (#99).
  • Added IDFTransformation for generating TF-IDF features (#104).
  • Exposed more parameters for XGBoost models (#107).
  • Added a Wordpiece tokenizer (#111).
  • Added optional output standardisation to LibSVM regressors (#113).
  • Added a BERT feature extractor for text data (#116). This can load in ONNX format BERT (and BERT style) models from HuggingFace Transformers, and use them as part of Tribuo’s text feature extraction package.
  • Added a configurable version of AggregateDataSource, and added iteration order parameters to both forms of AggregateDataSource (#125).
  • Added an option to RowProcessor which passes through newlines (#137).

Other improvements

  • Removed redundant computation in tree construction (#63).
  • Added better accessors for the centroids of a K-Means model (#98).
  • Improved the speed of the feature transformation infrastructure (#104).
  • Refactored the SGD models to reduce redundant code and allow models to share upcoming improvements (#106, #134).
  • Added many performance optimisations to the linear SGD and CRF models, allowing the automatic use of dense feature spaces (#112). This also adds specialisations to the math library for dense vectors and matrices, improving the performance of the CRF model even when operating on sparse feature sets.
  • Added provenance tracking of the Java version, OS and CPU architecture (#115).
  • Changed the behaviour of sparse features under transformations to expose additional behaviour (#122).
  • Improved MultiLabelEvaluation.toString() (#136).
  • Added a document classification tutorial which shows the various text feature extraction techniques available in Tribuo.
  • Expanded javadoc coverage.
  • Upgraded ONNX Runtime to 1.7.0, XGBoost to 1.4.1, TensorFlow to 0.3.1, liblinear-java to 2.43, OLCUT to 5.1.6, OpenCSV to 5.4.
  • Miscellaneous small bug fixes.


All Release Notes

The Tribuo source code repository contains release notes for all versions of Tribuo.