[AI_IMS-MED] Adaptive Model Delivery
This contribution revises previous documents (S4-251799, S4aR250211) on adaptive model delivery, incorporating the agreed call flow for device inferencing from S4aR260014 (agreed in SA4#134). The work builds upon TR 26.927 which documented AI/ML model delivery procedures.
The document addresses the critical challenge of timely model delivery for UE-centric inference in IMS DC-based AI/ML applications. Key points:
Based on TR 26.927 clause 5.2.2.2:
The document references two agreed high-level call flows:
Key steps include:
1. MMTel service establishment
2. BDC establishment between UE and MF
3. DCSF creates DC application list based on subscription filter and UE static capabilities
4. Application list includes AI service information
5. User selects app based on AI service
6. App download via BDC
7. Task selection and model variant selection
8. ADC establishment
9. Three inferencing modes: Local, Remote, or Split
Detailed 15-step procedure including:
- Application discovery with AI_APPLICATION_DISCOVERY_REQUEST/RESPONSE messages
- Application metadata including AI feature tags and task descriptions
- Task manifest delivery
- Model selection and delivery via BDC or ADC
- Support for task reselection during session
Figure X.X-1: Basic Model Delivery over IMS DC
14-step procedure:
0. UE1 registers to IMS with AI/ML capability indication
1. MMTEL session establishment
2. IMS AS allocates DC resources
3. Session established between UE1 and UE2
4. Bootstrap Data Channel (bDC) establishment
5. DCSF creates subscriber-specific application list
6. Application list delivery over bDC
7. App selection and download with app manifest (includes inference tasks and model lists)
8. UE2 side DC procedures
9-10. Application data channel establishment with DC AS
11-12. Model selection and delivery (from DC AS or DCAR via DCSF)
13. Media exchange over MMTEL session
14. Inference execution on local or remote media
Figure X.Y-2: Adaptive Model Delivery over IMS DC
Enhanced procedure building on basic delivery:
Steps 1-10: Same as basic model delivery, with lower precision model selection in step 10
Step 11: Request for updatable model via MF
Steps 12a/12b: Model delivery from either:
- Option a: DCAR via DCSF
- Option b: DC-AS
Step 13: Model download to UE
Step 14: Inference loop starts and continues
Step 15: UE requests model update via MF
Steps 16a/16b: Model update delivery from either:
- Option a: DCAR via DCSF
- Option b: DC-AS
Step 17: Model update download via MF
Step 18: UE applies model update to initial model
Step 19: Inference continues with potential for further updates
The referenced S4aR260014 document contains an Editor's Note indicating:
- Whether MF needs to understand AI task semantics requires clarification (FFS)
- Application type handling needs clarification
- Large model handling procedures need clarification