Release Notes - SuperML Java v3.1.2
Release Date: April 30, 2026
Version: 3.1.2
Maven Central: org.superml:superml-core:3.1.2
Breaking Changes: None (fully backward compatible with 3.0.x and 2.x.x)
SuperML Java 3.1.2 is an incremental release that builds upon the solid foundation of 3.0.1, focusing on performance optimizations, bug fixes, and enhanced stability across all modules. This release ensures the framework remains production-ready with improved reliability and efficiency.
π Key Improvements
1. Performance Enhancements β‘
Core Algorithm Optimizations
- 15% faster training in Linear Models (LinearRegression, LogisticRegression, Ridge, Lasso)
- 10% improved memory usage in Decision Tree and Random Forest implementations
- Enhanced vectorization in mathematical operations across all modules
- Optimized matrix operations using more efficient BLAS-level computations
Transformer Performance Improvements
- 20% faster attention computation through optimized matrix multiplication
- Reduced memory footprint for large sequence processing
- Improved batch processing efficiency in transformer training
- Better GPU utilization when available (preparation for future CUDA support)
PMML Export Optimizations
- Faster XML generation for large models (50% speed improvement)
- Reduced memory overhead during PMML conversion
- Improved validation performance for complex model structures
2. Bug Fixes and Stability π§
Security Fixes
- Fixed: CVE-2025-27820 β upgraded
httpclient55.4.1 β 5.4.3 - Fixed: CVE-2023-6378 β upgraded
logback-classic1.2.12 β 1.5.12,slf4j1.7.36 β 2.0.16
Serialization
- Added:
DecisionTreeandTreeNodenow implementjava.io.Serializablefor model persistence
Core Module Fixes
- Fixed: Thread safety issues in concurrent model training scenarios
- Fixed: Memory leak in model persistence when handling large datasets
- Fixed: Numerical stability improvements in gradient descent algorithms
- Fixed: Edge cases in cross-validation when dealing with small datasets
Transformer Module Fixes
- Fixed: Attention mask handling for variable sequence lengths
- Fixed: Position encoding overflow for very long sequences (>10K tokens)
- Fixed: Gradient clipping edge cases during training
- Fixed: Token padding inconsistencies in batch processing
PMML Module Fixes
- Fixed: XML encoding issues with special characters in feature names
- Fixed: Schema validation errors for certain Random Forest configurations
- Fixed: Precision loss in coefficient export for very small values
- Fixed: Memory usage spikes during large model PMML conversion
3. Enhanced Documentation and Examples π
Updated Documentation
- Comprehensive API documentation with detailed parameter explanations
- Performance benchmarking guides with real-world scenarios
- Advanced usage patterns and best practices documentation
- Troubleshooting guides for common integration issues
New Examples and Tutorials
- Production deployment examples showing real-world integration patterns
- Performance tuning tutorials for different use cases
- Memory optimization guides for large-scale applications
- Advanced transformer fine-tuning examples
π Technical Specifications
Performance Benchmarks (vs 3.0.1)
- Training Speed: +15% average improvement across all algorithms
- Memory Usage: -10% reduction in peak memory consumption
- Prediction Speed: +8% faster inference across the board
- PMML Export: +50% faster conversion for large models
Compatibility Matrix
- Java Versions: 8, 11, 17, 21 (fully tested)
- Maven: 3.6.0+ (recommended 3.8.0+)
- Operating Systems: Windows 10+, macOS 10.15+, Linux (Ubuntu 18.04+)
- Memory Requirements: Minimum 2GB, Recommended 8GB+ for large models
Module Status (21/21 Modules)
β
superml-core - Enhanced performance, bug fixes
β
superml-linear-models - 15% faster training, stability improvements
β
superml-tree-models - Serializable DecisionTree/TreeNode, memory optimizations
β
superml-transformers - Attention optimizations, JUnit 5 migration
β
superml-pmml - 50% faster export, validation improvements
β
superml-clustering - Numerical stability improvements
β
superml-preprocessing - Enhanced scaling algorithms
β
superml-model-selection - Cross-validation edge case fixes
β
superml-pipeline - Thread safety improvements
β
superml-datasets - Memory efficient data loading
β
superml-metrics - Enhanced computation accuracy, condition number fix
β
superml-visualization - Improved chart rendering
β
superml-persistence - Memory leak fixes
β
superml-inference - Performance optimizations
β
superml-autotrainer - Algorithm selection improvements
β
superml-drift - Real-time monitoring optimizations
β
superml-utils - Utility improvements
β
superml-bundle-all - All-in-one bundle
β
superml-examples - Updated with latest patterns
β
superml-testcases - 172+ tests passing (full coverage)
β
superml-pmml - PMML 4.4 export support
π Migration Guide
From v3.0.1 to v3.1.2
No breaking changes - this is a drop-in replacement:
- Update Maven dependency: ```xml
Or use the all-in-one bundle:
```xml
<dependency>
<groupId>org.superml</groupId>
<artifactId>superml-bundle-all</artifactId>
<version>3.1.2</version>
</dependency>
- No code changes required - all existing code continues to work
- Optional: Take advantage of new performance optimizations automatically
- Optional: Update to newer documentation and examples
From v2.x.x to v3.1.2
Follow the v3.0.1 migration guide first, then upgrade to v3.1.2:
- See Migration Guide 2.x β 3.0.1
- All v3.0.1 migration steps apply
- Additional v3.1.2 benefits are automatic
π Usage Examples
Enhanced Performance Example
import org.superml.linear_model.LinearRegression;
import org.superml.preprocessing.StandardScaler;
import org.superml.pipeline.Pipeline;
// Performance improvements are automatic
LinearRegression model = new LinearRegression();
StandardScaler scaler = new StandardScaler();
Pipeline pipeline = new Pipeline()
.addStep("scaler", scaler)
.addStep("model", model);
// 15% faster training in v3.1.2
long startTime = System.currentTimeMillis();
pipeline.fit(X_train, y_train);
long trainingTime = System.currentTimeMillis() - startTime;
System.out.println("Training completed in: " + trainingTime + "ms");
// 8% faster predictions
double[] predictions = pipeline.predict(X_test);
Optimized Transformer Usage
import org.superml.transformers.TransformerEncoder;
import org.superml.transformers.MultiHeadAttention;
// Memory and speed optimizations are automatic
TransformerEncoder encoder = new TransformerEncoder.Builder()
.modelDimension(512)
.numLayers(6)
.numAttentionHeads(8)
.build();
// 20% faster attention computation in v3.1.2
encoder.train(trainingData);
Faster PMML Export
import org.superml.pmml.PMMLConverter;
import org.superml.tree_models.RandomForest;
RandomForest model = new RandomForest();
model.fit(X, y);
PMMLConverter converter = new PMMLConverter();
// 50% faster PMML generation in v3.1.2
long startTime = System.currentTimeMillis();
String pmml = converter.convertToXML(model);
long exportTime = System.currentTimeMillis() - startTime;
System.out.println("PMML export completed in: " + exportTime + "ms");
π Performance Comparison
Training Performance (1M samples, 100 features)
| Algorithm | v3.0.1 | v3.1.2 | Improvement | |ββββ|βββ|βββ|ββββ-| | LinearRegression | 2.3s | 2.0s | +15% | | LogisticRegression | 3.1s | 2.7s | +13% | | RandomForest | 12.4s | 11.2s | +10% | | TransformerEncoder | 45.2s | 36.1s | +20% |
Memory Usage (Peak during training)
| Algorithm | v3.0.1 | v3.1.2 | Reduction | |ββββ|βββ|βββ|ββββ| | DecisionTree | 1.2GB | 1.1GB | -8% | | RandomForest | 3.4GB | 3.1GB | -9% | | TransformerEncoder | 2.8GB | 2.5GB | -11% | | PMML Export | 800MB | 650MB | -19% |
Prediction Speed (10K predictions)
| Algorithm | v3.0.1 | v3.1.2 | Improvement | |ββββ|βββ|βββ|ββββ-| | LinearRegression | 15ms | 14ms | +7% | | LogisticRegression | 18ms | 16ms | +11% | | RandomForest | 95ms | 87ms | +8% | | TransformerEncoder | 340ms | 315ms | +7% |
π Related Resources
Documentation Updates
- Performance Guide - New optimization techniques
- Troubleshooting Guide - Common issue solutions
- Advanced Examples - Production patterns
Previous Releases
- Release Notes v3.0.1 - Major transformer and PMML release
- Whatβs New in v3.0.1 - Feature overview and migration
- Release Notes v2.1.0 - Previous stable release
Community and Support
- GitHub Issues - Bug reports and feature requests
- Community Forum - Discussions and support
- SuperML.dev - Developer resources and tutorials
π Acknowledgments
Special thanks to the SuperML community for:
- Performance testing across different environments
- Bug reports that helped identify edge cases
- Feature requests that guided optimization priorities
- Documentation improvements and examples
Contributors to v3.1.2
- Performance optimization team
- QA and testing contributors
- Documentation and example contributors
- Community feedback providers
π Known Issues and Limitations
Minor Known Issues
- Transformer training: Very long sequences (>16K tokens) may require manual memory management
- PMML export: Some edge cases with custom feature transformations may need manual validation
- Visualization: ASCII charts may not render perfectly in all terminal environments
Future Improvements (v3.2.0)
- GPU acceleration for transformer training
- Distributed training support for large datasets
- Enhanced ONNX export with more model types
- Real-time inference streaming capabilities
β Upgrade Recommendation
SuperML Java v3.1.2 is highly recommended for all users:
- β Production environments: Enhanced stability and performance
- β Development: Better debugging and faster iteration cycles
- β Research: Improved transformer performance for experiments
- β Enterprise: Better memory management and reliability
Upgrade is safe and straightforward with no breaking changes and immediate performance benefits.
For complete documentation and examples, visit SuperML Java Documentation