-
Notifications
You must be signed in to change notification settings - Fork 1.1k
[AINode] Refactoring of Model Storage, Loading, and Inference Pipeline #16819
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
Codecov Report❌ Patch coverage is Additional details and impacted files@@ Coverage Diff @@
## master #16819 +/- ##
============================================
+ Coverage 38.76% 38.92% +0.15%
Complexity 207 207
============================================
Files 5006 5012 +6
Lines 332106 332171 +65
Branches 42232 42286 +54
============================================
+ Hits 128756 129287 +531
+ Misses 203350 202884 -466 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
21ad4ac to
12418ac
Compare
remove useless codes in IoTDB
12418ac to
ca94b06
Compare
a59af47 to
687ae9f
Compare
Update AINodeInstanceManagementIT.java
ee343b2 to
9a726eb
Compare
|



This PR introduces significant improvements in the model storage, loading, and inference pipeline management for better extensibility, efficiency, and ease of use. The changes include the refactoring of model storage to support a wider range of models, streamlining the model loading process, and the introduction of a unified inference pipeline. These improvements aim to optimize model management, reduce memory usage, and enhance the overall inference workflow.
Model Storage Refactoring
Model Loading Refactoring
Inference Pipeline Addition