RkGrit opened a new pull request, #16794:
URL: https://github.com/apache/iotdb/pull/16794
This PR introduces significant improvements in the model storage, loading,
and inference pipeline management for better extensibility, efficiency, and
ease of use. The changes include the refactoring of model storage to support a
wider range of models, streamlining the model loading process, and the
introduction of a unified inference pipeline. These improvements aim to
optimize model management, reduce memory usage, and enhance the overall
inference workflow.
- [ ] Model Storage Refactoring
- [ ] Extended Support for Models: The system now supports not only
built-in models like TimerXL and Sundial but also allows the integration of
fine-tuned and user-defined models.
- [ ] Unified Model Management: A new model management system enables
model registration, deletion, and loading from both local paths and Hugging
Face.
- [ ] Code Optimization: Redundant code from previous versions has been
removed, and hard-coded model management has been replaced by a more flexible
approach that integrates seamlessly with the Hugging Face Transformers
ecosystem.
- [ ] Model Loading Refactoring
- [ ] Simplified Model Loading: The previous custom loading logic with
complex if...else... conditions has been replaced by a unified model loading
interface, simplifying the process.
- [ ] Automatic Model Type Detection: The system now automatically
detects the model type and selects the appropriate loading method, supporting
models from Transformers, sktime, and PyTorch.
- [ ] Lazy Loading: The PR introduces lazy loading for Python modules,
eliminating the need to load multiple modules at startup, reducing
initialization time and memory consumption.
- [ ] Inference Pipeline Addition
- [ ] Unified Inference Workflow: The introduction of the Inference
Pipeline encapsulates the entire model inference process, offering a
standardized interface for preprocessing, inference, and post-processing.
- [ ] Support for Multiple Tasks: The pipeline is versatile, supporting
various inference tasks such as prediction, classification, and dialogue-based
tasks.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]