S4-260129 - AI Summary

[AIML_IMS-MED] Call flow for split inferencing

Back to Agenda Download Summary
AI-Generated Summary AI

Summary of S4-260129: Call Flow for Split Inferencing

Document Information

  • Source: Samsung Electronics Co., Ltd.
  • Meeting: TSG-SA WG4 Meeting #135 (February 2026, Goa, India)
  • Work Item: AIML_IMS-MED
  • Purpose: Approval of call flow for split inferencing

Main Technical Contribution

This document proposes a detailed call flow for split inferencing in IMS-based AI/ML services, where AI model execution is distributed between the UE and network elements (MF - Media Function). The contribution is intended for inclusion in clause AC.4.3 of the base Change Request.

Split Inferencing Call Flow

Session Establishment and Bootstrap (Steps 1-2)

  • MMTel service establishment
  • Bootstrap Data Channel (BDC) establishment between UE and MF per TS 23.228, clause AC.7.1

Application Discovery and Selection (Steps 3-6)

  • Application List Request: UE requests available DC applications from MF via HTTP over BDC
  • MF Routing: MF replaces root URL with replacement URL and forwards to DCSF
  • Application Metadata Creation: DCSF creates user-specific DC application list based on:
  • User subscription information
  • Application metadata including:
    • Generic app information (description, app ID, URL)
    • AI-specific information (AI feature tag indicating requirements, AI task descriptions)
  • Metadata Delivery: DCSF provides application list URL and metadata to UE via MF
  • User Selection: User selects application based on AI service description and AI task annotations

Application Download (Steps 7-9)

  • UE requests selected application from MF
  • MF fetches AI application from DCSF
  • Application downloaded to UE via BDC along with AI task metadata (expressed as task manifest per clause AC.7)

AI Task Selection and Configuration (Steps 10-13)

  • Task Presentation: User presented with list of AI tasks supported by application, including:
  • Annotations from AI task metadata
  • Task description information
  • Information on execution endpoints supported by each task/subtask
  • User Task Selection: User selects desired AI task(s)
  • Application Data Channel: Established between UE and DC AS per TS 23.228, clause AC.7.2
  • Split Configuration Decision: UE identifies which tasks/AI models to execute locally vs. in network based on:
  • User-selected AI tasks
  • AI task metadata
  • UE capabilities
  • Configuration Request: UE requests split inference configuration from network, identifying AI models for UE and network execution

Model Distribution and Configuration Response (Steps 14-16)

  • Requirements Check: MF verifies requirements for network-side AI tasks/models; MF reallocation if requirements not met
  • Model Fetching: MF obtains AI models for both UE and network execution from either:
  • DCAR via DCSF (step 15a), or
  • DC AS (step 15b - alternative)
  • Configuration Response: MF sends response to UE including AI models for UE execution

Inference Execution (Steps 17-22)

  • SDP Re-negotiation: Associates media/data/intermediate data flows between UE and MF with corresponding tasks
  • UE Inference: Tasks designated for UE execution are performed
  • Data Transfer to Network: Output (media/data/intermediate data) from UE tasks sent to MF
  • Network Inference: MF executes tasks designated for network execution
  • Result Delivery: MF sends output (results or intermediate data for further UE processing) to UE
  • Optional Further UE Processing: UE may execute additional tasks as part of selected AI task(s)

Dynamic Task Reselection (Step 23)

  • User/UE may reselect AI tasks during session using AI task metadata from step 9
  • On reselection, flow returns to step 12 (split configuration decision)

Key Technical Features

Metadata Framework

  • Application metadata includes both generic and AI-specific information
  • AI task metadata (task manifest) provides detailed information on:
  • Task descriptions
  • Execution endpoint options
  • Requirements for split execution

Flexibility in Execution Distribution

  • UE determines split configuration based on capabilities and metadata
  • Network validates requirements and may reallocate MF resources
  • Dynamic task reselection supported during active session

Model Distribution Options

  • Multiple sources for AI model retrieval (DCAR via DCSF or DC AS)
  • Models distributed to both UE and network as needed for split execution

Media/Data Flow Management

  • SDP re-negotiation ensures proper association of data flows with tasks
  • Support for intermediate data exchange between UE and network for multi-stage inference pipelines
Document Information
Source:
Samsung Electronics Iberia SA
Type:
discussion
For:
Agreement
Original Document:
View on 3GPP
Title: [AIML_IMS-MED] Call flow for split inferencing
Agenda item: 10.5
Agenda item description: AI_IMS-MED (Media aspects for AI/ML in IMS services)
Doc type: discussion
For action: Agreement
Contact: Eric Yip
Uploaded: 2026-02-03T15:43:27.897000
Contact ID: 86783
TDoc Status: merged
Reservation date: 03/02/2026 07:36:37
Agenda item sort order: 52