-
Notifications
You must be signed in to change notification settings - Fork 1.1k
[AINode] Refactoring of Model Storage, Loading, and Inference Pipeline #16819
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
Codecov Report❌ Patch coverage is Additional details and impacted files@@ Coverage Diff @@
## master #16819 +/- ##
============================================
+ Coverage 38.87% 38.98% +0.11%
Complexity 207 207
============================================
Files 5022 5009 -13
Lines 333113 332063 -1050
Branches 42390 42260 -130
============================================
- Hits 129488 129449 -39
+ Misses 203625 202614 -1011 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
9a726eb to
3335e1e
Compare
CRZbulabula
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
PTAL.
iotdb-core/ainode/iotdb/ainode/core/inference/inference_request_pool.py
Outdated
Show resolved
Hide resolved
iotdb-core/ainode/iotdb/ainode/core/model/sundial/pipeline_sundial.py
Outdated
Show resolved
Hide resolved
iotdb-core/ainode/iotdb/ainode/core/model/timer_xl/pipeline_timer.py
Outdated
Show resolved
Hide resolved
remove useless codes in IoTDB
fix ci
Update AINodeInstanceManagementIT.java Fix. CI
* stash * Support loading inference pipelines for user-defined models * Support loading inference pipelines for different models
…ient model loading (#16865)
09f98a0 to
ef84301
Compare
|



This PR introduces significant improvements in the model storage, loading, and inference pipeline management for better extensibility, efficiency, and ease of use. The changes include the refactoring of model storage to support a wider range of models, streamlining the model loading process, and the introduction of a unified inference pipeline. These improvements aim to optimize model management, reduce memory usage, and enhance the overall inference workflow.
Model Storage Refactoring
Model Loading Refactoring
Inference Pipeline Addition