RkGrit opened a new pull request, #16794:
URL: https://github.com/apache/iotdb/pull/16794

   This PR introduces significant improvements in the model storage, loading, 
and inference pipeline management for better extensibility, efficiency, and 
ease of use. The changes include the refactoring of model storage to support a 
wider range of models, streamlining the model loading process, and the 
introduction of a unified inference pipeline. These improvements aim to 
optimize model management, reduce memory usage, and enhance the overall 
inference workflow.
   
   - [ ] Model Storage Refactoring
       - [ ] Extended Support for Models: The system now supports not only 
built-in models like TimerXL and Sundial but also allows the integration of 
fine-tuned and user-defined models.
       - [ ] Unified Model Management: A new model management system enables 
model registration, deletion, and loading from both local paths and Hugging 
Face.
       - [ ] Code Optimization: Redundant code from previous versions has been 
removed, and hard-coded model management has been replaced by a more flexible 
approach that integrates seamlessly with the Hugging Face Transformers 
ecosystem.
   
   - [ ]  Model Loading Refactoring
       - [ ] Simplified Model Loading: The previous custom loading logic with 
complex if...else... conditions has been replaced by a unified model loading 
interface, simplifying the process.
       - [ ] Automatic Model Type Detection: The system now automatically 
detects the model type and selects the appropriate loading method, supporting 
models from Transformers, sktime, and PyTorch.
       - [ ] Lazy Loading: The PR introduces lazy loading for Python modules, 
eliminating the need to load multiple modules at startup, reducing 
initialization time and memory consumption.
   
   - [ ] Inference Pipeline Addition
       - [ ] Unified Inference Workflow: The introduction of the Inference 
Pipeline encapsulates the entire model inference process, offering a 
standardized interface for preprocessing, inference, and post-processing.
       - [ ] Support for Multiple Tasks: The pipeline is versatile, supporting 
various inference tasks such as prediction, classification, and dialogue-based 
tasks.
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to