[go: up one dir, main page]

WO2024210402A1 - Procédé et appareil pour une émission de données ia/ml dans un système de communication sans fil - Google Patents

Procédé et appareil pour une émission de données ia/ml dans un système de communication sans fil Download PDF

Info

Publication number
WO2024210402A1
WO2024210402A1 PCT/KR2024/003988 KR2024003988W WO2024210402A1 WO 2024210402 A1 WO2024210402 A1 WO 2024210402A1 KR 2024003988 W KR2024003988 W KR 2024003988W WO 2024210402 A1 WO2024210402 A1 WO 2024210402A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
network
network entity
data transfer
transfer function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
PCT/KR2024/003988
Other languages
English (en)
Inventor
Morteza KHEIRKHAH
David GUTIERREZ ESTEVEZ
Chadi KHIRALLAH
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Publication of WO2024210402A1 publication Critical patent/WO2024210402A1/fr
Anticipated expiration legal-status Critical
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/16Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks using machine learning or artificial intelligence

Definitions

  • the present disclosure relate to methods, apparatus and/or systems for transferring (e.g., delivering and/or collecting) AI(artificial intelligence)/ML(machine learning) related data in a network.
  • Various examples provide methods, apparatus and/or systems wherein a new network function is provided in one or more network entities, with this network function being responsible for delivering AI/ML models and/or related data, and optionally for collecting data for AI/ML purposes also.
  • the new network function is a AI data delivery function provided in a UE and in a user plane function (and, optionally, in a RAN) to provide AI/ML related data to the UE over the UP.
  • 6G communication systems which are expected to be commercialized around 2030, will have a peak data rate of tera (1,000 giga)-level bit per second (bps) and a radio latency less than 100 ⁇ sec, and thus will be 50 times as fast as 5G communication systems and have the 1/10 radio latency thereof.
  • a terahertz (THz) band for example, 95 gigahertz (GHz) to 3THz bands. It is expected that, due to severer path loss and atmospheric absorption in the terahertz bands than those in mmWave bands introduced in 5G, technologies capable of securing the signal transmission distance (that is, coverage) will become more crucial.
  • Radio Frequency (RF) elements it is necessary to develop, as major technologies for securing the coverage, Radio Frequency (RF) elements, antennas, novel waveforms having a better coverage than Orthogonal Frequency Division Multiplexing (OFDM), beamforming and massive Multiple-input Multiple-Output (MIMO), Full Dimensional MIMO (FD-MIMO), array antennas, and multiantenna transmission technologies such as large-scale antennas.
  • OFDM Orthogonal Frequency Division Multiplexing
  • MIMO massive Multiple-input Multiple-Output
  • FD-MIMO Full Dimensional MIMO
  • array antennas and multiantenna transmission technologies such as large-scale antennas.
  • OFDM Orthogonal Frequency Division Multiplexing
  • MIMO massive Multiple-input Multiple-Output
  • FD-MIMO Full Dimensional MIMO
  • array antennas and multiantenna transmission technologies such as large-scale antennas.
  • OFDM Orthogonal Frequency Division Multiplexing
  • MIMO massive Multiple-input Multiple-Out
  • a full-duplex technology for enabling an uplink transmission and a downlink transmission to simultaneously use the same frequency resource at the same time
  • a network technology for utilizing satellites, High-Altitude Platform Stations (HAPS), and the like in an integrated manner
  • HAPS High-Altitude Platform Stations
  • an improved network structure for supporting mobile base stations and the like and enabling network operation optimization and automation and the like
  • a dynamic spectrum sharing technology via collision avoidance based on a prediction of spectrum usage an use of Artificial Intelligence (AI) in wireless communication for improvement of overall network operation by utilizing AI from a designing phase for developing 6G and internalizing end-to-end AI support functions
  • a next-generation distributed computing technology for overcoming the limit of UE computing ability through reachable super-high-performance communication and computing resources (such as Mobile Edge Computing (MEC), clouds, and the like) over the network.
  • MEC Mobile Edge Computing
  • 6G communication systems in hyper-connectivity, including person to machine (P2M) as well as machine to machine (M2M), will allow the next hyper-connected experience.
  • services such as truly immersive eXtended Reality (XR), high-fidelity mobile hologram, and digital replica could be provided through 6G communication systems.
  • services such as remote surgery for security and reliability enhancement, industrial automation, and emergency response will be provided through the 6G communication system such that the technologies could be applied in various fields such as industry, medical care, automobiles, and home appliances.
  • the present disclosure relate to methods, apparatus and/or systems for transferring (e.g., delivering and/or collecting) AI(artificial intelligence)/ML(machine learning) related data in a network.
  • Various examples provide methods, apparatus and/or systems wherein a new network function is provided in one or more network entities, with this network function being responsible for delivering AI/ML models and/or related data, and optionally for collecting data for AI/ML purposes also.
  • the new network function is a AI data delivery function provided in a UE and in a user plane function (and, optionally, in a RAN) to provide AI/ML related data to the UE over the UP.
  • a first network entity comprising a first artificial intelligence (AI)/machine learning (ML) data transfer function
  • AI artificial intelligence
  • ML machine learning
  • the first network entity is included in a network and comprises: a transmitter; a receiver; and at least one processor configured to: establish a first connection between the first AI/ML data transfer function and at least one second AI/ML data transfer function included in the network; and control the first AI/ML data transfer function to coordinate communicating AI/ML data with the network over user plane (UP) or control plane (CP) based on at least one rule or policy, wherein the AI/ML data relates to an AI/ML operation; wherein each of the at least one second AI/ML data transfer function is included in a second network entity.
  • AI artificial intelligence
  • ML machine learning
  • the first AI/ML data transfer function is configured to control transfer of the AI/ML data over the UP or the CP based on data characteristics of the AI/ML data.
  • the first AI/ML data transfer function is connected to the at least one second AI/ML data transfer function via at least one existing or new protocol data unit (PDU) session.
  • PDU protocol data unit
  • the first connection is between the first AI/ML data transfer function and a plurality of the second AI/ML data transfer functions; and wherein the at least one processor is configured to: determine at least one transport protocol for use in establishing the first connection based on at least one of: a type of the AI/ML data, priority of the AI/ML, reliability and security of the transport protocol or available access in the case of multi-access scenarios; or receive, from the network, an indication of the at least one transport protocol.
  • one instance is used in the first connection for connecting the first AI/ML data transfer function to the plurality of second AI/ML data transfer functions or to a subset of the plurality of second AI/ML data transfer functions, over the at least one transport protocol.
  • the at least one transport protocol comprises a plurality of transport protocols; and wherein the first AI/ML data transfer function is configured to simultaneously use multiple transport sessions, using the plurality of transport protocols, for coordinating communication of the AI/ML data.
  • different transport protocols among the plurality of transport protocols are used for communicating different types of AI/ML data.
  • the first network entity is a user equipment (UE).
  • UE user equipment
  • the at least one processor is configured to: control the first AI/ML data transfer function to coordinate collection of at least a first portion of the AI/ML data; and transmit the first portion of the AI/ML data to the at least one second network entity based on the first connection.
  • the at least one processor is configured to: transmit, to a core network (CN) during initial UE registration in the network, an indication that the first network entity includes the first AI/ML data transfer function or supports the first AI/ML data transfer function; and/or receive, from the CN, a PDU related message indicating that a PDU session corresponding to the PDU related message supports the first AI/ML data transfer function, wherein the PDU related message comprises an IP address of the at least one second AI/ML data transfer function; and wherein the at least one existing or new PDU session is the indicated PDU session.
  • CN core network
  • the at least one rule or policy are received from the network and indicates: how measurements are to be configured at the first network entity, data to be collected by the first network entity and the frequency of the collection of the data, how at least a second portion of the AI/ML data is to be communicated by the first network entity, and/or a PDU session to be used for the AI/ML data, wherein the at least one existing or new PDU session is the indicated PDU session.
  • the at least one rule or policy indicates how to transmit the second portion of the AI/ML data based on one or more of a size of the AI/ML data, a type of the AI/ML data, a priority of the AI/ML data, or privacy of the AI/ML data.
  • the at least one rule or policy indicates: how the AI/ML data is to be communicated, a priority for communicating the AI/ML data, and/or, if the at least one rule or policy comprises a plurality of different rules or policies, a relative priority between the plurality of different rules or policies.
  • the at least one rule or policy is received from the network; and/or wherein the at least one rule or policy indicates, for each of a plurality of different network conditions, how the AI/ML data is to be communicated and/or the priority for communicating the AI/ML data.
  • the at least one rule or policy comprises the plurality of different rules or policies, and the plurality of different rules or policies originate from a plurality of different entities within the network; and wherein the relative priority between the plurality of different rules or policies indicates whether a rule or policy, among the plurality of different rules or policies, originating from one of the plurality of different entities can be prioritised over another rule or policy, among the plurality of different rules or policies, originating from a different one of the plurality of different entities.
  • the at least one processor is configured to: receive at least a third portion of the AI/ML data over the UP from the at least one second network entity; and wherein the third portion of the AI/ML data includes one or more of an AI/ML trained model, AI/ML model construction, AI/ML model topology, neural network weights, datasets for training, or measurements and statistics for model training.
  • the at least one processor is configured to: establish a second connection between the first AI/ML data transfer function and at least one third AI/ML data transfer function included in the network; and the at least one third AI/ML data transfer function is included at a third network entity; and wherein the first network entity is a user equipment (UE), the at least one second network entity comprises at least one user plane function (UPF), and the third network entity is a next generation node B (gNB) or next generation radio access network (NG-RAN).
  • UE user equipment
  • UPF user plane function
  • gNB next generation node B
  • NG-RAN next generation radio access network
  • the first network entity is a user plane function (UPF).
  • UPF user plane function
  • the first network entity is a next generation node B (gNB) or next generation radio access network (NG-RAN).
  • gNB next generation node B
  • NG-RAN next generation radio access network
  • the embodiment herein is to provide a first network entity comprising a first artificial intelligence (AI)/machine learning (ML) data transfer function.
  • the first network entity includes a transmitter, a receiver, and at least one processor.
  • the processor is configured to establish a first connection between the first AI/ML data transfer function and at least one second AI/ML data transfer function included in the network, and control the first AI/ML data transfer function to coordinate communicating AI/ML data with the network over user plane (UP) or control plane (CP) based on at least one rule or policy.
  • UP user plane
  • CP control plane
  • the AI/ML data relates to an AI/ML operation, each of the at least one second AI/ML data transfer function is included in a second network entity.
  • the first AI/ML data transfer function is configured to control transfer of the AI/ML data over the UP or the CP based on data characteristics of the AI/ML data.
  • the first AI/ML data transfer function is connected to the at least one second AI/ML data transfer function via at least one existing or new protocol data unit (PDU) session.
  • PDU protocol data unit
  • the first connection is between the first AI/ML data transfer function and a plurality of the second AI/ML data transfer functions.
  • the at least one processor is configured to determine at least one transport protocol for use in establishing the first connection based on at least one of: a type of the AI/ML data, priority of the AI/ML, reliability and security of the transport protocol or available access in the case of multi-access scenarios, or receive, from the network, an indication of the at least one transport protocol.
  • one instance is used in the first connection for connecting the first AI/ML data transfer function to the plurality of second AI/ML data transfer functions or to a subset of the plurality of second AI/ML data transfer functions, over the at least one transport protocol.
  • the at least one transport protocol comprises a plurality of transport protocols.
  • the first AI/ML data transfer function is configured to simultaneously use multiple transport sessions, using the plurality of transport protocols, for coordinating communication of the AI/ML data.
  • different transport protocols among the plurality of transport protocols are used for communicating different types of AI/ML data.
  • the AI/ML data comprises a first AI/ML model and a second AI/ML model, wherein the first AI/ML model is a different type of AI/ML data to the second AI/ML model.
  • the first AI/ML data transfer function is configured to use a first transport protocol, among the plurality of transport protocols, for transmitting or receiving the first AI/ML model and use a second transport protocol, among the plurality of transport protocols.
  • the second transport protocol is different to the first transport protocol.
  • the first network entity is a user equipment (UE).
  • UE user equipment
  • the at least one processor is configured to control the first AI/ML data transfer function to coordinate collection of at least a first portion of the AI/ML data; and transmit the first portion of the AI/ML data to the at least one second network entity based on the first connection.
  • the at least one processor is configured to transmit, to a core network (CN) during initial UE registration in the network, an indication that the first network entity includes the first AI/ML data transfer function or supports the first AI/ML data transfer function; and/or receive, from the CN, a PDU related message indicating that a PDU session corresponding to the PDU related message supports the first AI/ML data transfer function, wherein the PDU related message comprises an IP address of the at least one second AI/ML data transfer function, and the at least one existing or new PDU session is the indicated PDU session.
  • CN core network
  • the at least one rule or policy are received from the network and indicates how measurements are to be configured at the first network entity, data to be collected by the first network entity and the frequency of the collection of the data, how at least a second portion of the AI/ML data is to be communicated by the first network entity, and/or a PDU session to be used for the AI/ML data, wherein the at least one existing or new PDU session is the indicated PDU session.
  • the at least one rule or policy indicates how to transmit the second portion of the AI/ML data based on one or more of a size of the AI/ML data, a type of the AI/ML data, a priority of the AI/ML data, or privacy of the AI/ML data.
  • the at least one rule or policy indicates how the AI/ML data is to be communicated, a priority for communicating the AI/ML data, and/or, if the at least one rule or policy comprises a plurality of different rules or policies, a relative priority between the plurality of different rules or policies.
  • the at least one rule or policy is received from the network; and/or the at least one rule or policy indicates, for each of a plurality of different network conditions, how the AI/ML data is to be communicated and/or the priority for communicating the AI/ML data.
  • the at least one rule or policy comprises the plurality of different rules or policies, and the plurality of different rules or policies originate from a plurality of different entities within the network and the relative priority between the plurality of different rules or policies indicates whether a rule or policy, among the plurality of different rules or policies, originating from one of the plurality of different entities can be prioritised over another rule or policy, among the plurality of different rules or policies, originating from a different one of the plurality of different entities.
  • the at least one processor is configured to receive at least a third portion of the AI/ML data over the UP from the at least one second network entity and the third portion of the AI/ML data includes one or more of an AI/ML trained model, AI/ML model construction, AI/ML model topology, neural network weights, datasets for training, or measurements and statistics for model training.
  • the at least one processor is configured to establish a second connection between the first AI/ML data transfer function and at least one third AI/ML data transfer function included in the network and the at least one third AI/ML data transfer function is included at a third network entity.
  • the first network entity is a user equipment (UE)
  • the at least one second network entity comprises at least one user plane function (UPF)
  • the third network entity is a next generation node B (gNB) or next generation radio access network (NG-RAN).
  • gNB next generation node B
  • NG-RAN next generation radio access network
  • the at least one processor is configured to receive, from a session management function (SMF) or a policy control function (PCF) in the network, the at least one rule or policy and configure the first AI/ML data transfer function based on the received at least one rule or policy.
  • SMS session management function
  • PCF policy control function
  • the at least one processor is configured to allocate an IP address for the first AI/ML data transfer function and transmit information on the IP address to a session management function (SMF) included in the network.
  • SMF session management function
  • the at least one second network entity is a user equipment (UE), and the first network entity is configured to interact with the UE over the UP via a data radio bearer. Further, the at least one second network entity comprises at least one user plane function (UPF), and the first network entity is configured to interact with the at least one UPF over the UP via a new transport session over N3 interface.
  • UE user equipment
  • UPF user plane function
  • the embodiment herein is to provide a method of a first network entity comprising a first artificial intelligence (AI)/machine learning (ML) data transfer function and being included in a network.
  • the method includes establishing a first connection between the first AI/ML data transfer function and at least one second AI/ML data transfer function included in the network, and controlling the first AI/ML data transfer function to coordinate communicating AI/ML data with the network over user plane (UP) or control plane (CP) based on at least one rule or policy, wherein the AI/ML data relates to an AI/ML operation.
  • UP user plane
  • CP control plane
  • each of the at least one second AI/ML data transfer function is included in a second network entity.
  • the expression “at least one of A, B and/or C” (or the like), the expression “and/or”, and the expression “one or more of A, B and/or C” (or the like) should be seen to separately include all possible combinations, for example: A, B, C, A and B, A and C, A and B and C.
  • X for Y (where Y is some action, process, operation, function, activity or step and X is some means for carrying out that action, process, operation, function, activity or step) encompasses means X adapted, configured or arranged specifically, but not necessarily exclusively, to do Y.
  • Certain examples of the present disclosure provide methods, apparatus and/or systems for transferring (e.g., delivering and/or collecting) AI/ML related data in a network.
  • Various examples provide methods, apparatus and/or systems wherein a new network function is provided in one or more network entities, with this network function being responsible for delivering AI/ML models and/or related data, and optionally for collecting data for AI/ML purposes also.
  • the new network function is a AI data delivery function provided in a UE and in a user plane function (and, optionally, in a RAN) to provide AI/ML related data to the UE over the UP.
  • 3GPP 5G 3rd Generation Partnership Project
  • the techniques disclosed herein are not limited to these examples or to 3GPP 5G, and may be applied in any suitable system or standard, for example one or more existing and/or future generation wireless communication systems or standards.
  • the techniques disclosed herein may be applied in any existing or future releases of 3GPP 5G NR or any other relevant standard.
  • the functionality of the various network entities and other features disclosed herein may be applied to corresponding or equivalent entities or features in other communication systems or standards.
  • Corresponding or equivalent entities or features may be regarded as entities or features that perform the same or similar role, function, operation or purpose within the network.
  • the following disclosure should be considered in relation to 6G also, which is expected to use at least part of the 5G architecture, or equivalent, and to which the present disclosure also relates.
  • a particular network entity may be implemented as a network element on a dedicated hardware, as a software instance running on a dedicated hardware, and/or as a virtualised function instantiated on an appropriate platform, e.g. on a cloud infrastructure.
  • One or more of the messages in the examples disclosed herein may be replaced with one or more alternative messages, signals or other type of information carriers that communicate equivalent or corresponding information.
  • One or more non-essential elements, entities and/or messages may be omitted in certain examples.
  • ⁇ Information carried by a particular message in one example may be carried by two or more separate messages in an alternative example.
  • ⁇ Information carried by two or more separate messages in one example may be carried by a single message in an alternative example.
  • the transmission of information between network entities is not limited to the specific form, type and/or order of messages described in relation to the examples disclosed herein.
  • an apparatus/device/network entity configured to perform one or more defined network functions and/or a method therefor.
  • Such an apparatus/device/network entity may comprise one or more elements, for example one or more of receivers, transmitters, transceivers, processors, controllers, modules, units, and the like, each element configured to perform one or more corresponding processes, operations and/or method steps for implementing the techniques described herein.
  • an operation/function of X may be performed by a module configured to perform X (or an X-module).
  • Certain examples of the present disclosure may be provided in the form of a system (e.g., a network) comprising one or more such apparatuses/devices/network entities, and/or a method therefor.
  • examples of the present disclosure may be realized in the form of hardware, software or a combination of hardware and software.
  • Certain examples of the present disclosure may provide a computer program comprising instructions or code which, when executed, implement a method, system and/or apparatus in accordance with any aspect, example and/or embodiment disclosed herein.
  • Certain embodiments of the present disclosure provide a machine-readable storage storing such a program.
  • a network may include one or more of a Network Data Analytics Function (NWDAF) entity, an Access and Mobility Management Function (AMF) entity, a Session Management Function (SMF) entity, a Network Slice Selection Function (NSSF) entity, a Network Repository Function (NRF) entity, Application Function (AF) entity, and an Operation and Maintenance (OAM) entity.
  • NWDAF Network Data Analytics Function
  • AMF Access and Mobility Management Function
  • SMF Session Management Function
  • NSSF Network Slice Selection Function
  • NRF Network Repository Function
  • AF Application Function
  • OFAM Operation and Maintenance
  • the network may include one or more Service Consumers (including one or more of the entities mentioned above and/or one or more other entities) that receive analytics from NWDAF.
  • Service Consumers including one or more of the entities mentioned above and/or one or more other entities
  • a network may omit one or more of the entities mentioned above and/or may comprise one or more additional entities
  • Sync-FL may be computationally intensive for a network entity or participant (e.g., a UE).
  • a network entity or participant e.g., a UE
  • Sync-FL is an example which highlights the importance of minimizing partial or total disturbance of data collection and/or data transfer for AI/ML operations.
  • Another important scenario could be related to multi-agent, multi-device ML operations where typically a big data processing task (e.g., distributed training) is split among a set of devices (UEs) by ML agents.
  • Wireless or mobile (cellular) communications networks in which a mobile terminal (e.g., user equipment (UE), such as a mobile handset) communicates via a radio link with a network of base stations, or other wireless access points or nodes, have undergone rapid development through a number of generations.
  • a mobile terminal e.g., user equipment (UE), such as a mobile handset
  • 3GPP 3rd Generation Partnership Project
  • 4G and 5G systems are now widely deployed, while beyond 5G (B5G) and 6G systems are being considered.
  • 3GPP standards for 4G systems include an Evolved Packet Core (EPC) and an Enhanced-UTRAN (E-UTRAN: an Enhanced Universal Terrestrial Radio Access Network).
  • EPC Evolved Packet Core
  • E-UTRAN Enhanced-UTRAN
  • LTE Long Term Evolution
  • LTE is commonly used to refer to the whole system including both the EPC and the E-UTRAN, and LTE is used in this sense in the remainder of this document.
  • LTE should also be taken to include LTE enhancements such as LTE Advanced and LTE Pro, which offer enhanced data rates compared to LTE.
  • 5G New Radio 5G New Radio
  • 5G NR 5G New Radio
  • B5G systems, such as 6G B5G systems, such as 6G, are currently being considered and developed, and are expected to at least partly build on 5G systems.
  • SBA Service-Based Architecture
  • every network function (NF) in 5G Core (5GC) such as Access and Mobility Management Function (AMF), Session Management Function (SMF), Unified Data Manager (UDM), Policy Control Function (PCF), etc.
  • AMF Access and Mobility Management Function
  • SMF Session Management Function
  • UDM Unified Data Manager
  • PCF Policy Control Function
  • a transport protocol e.g., TCP, UDP, SCTP, etc.
  • TCP Transmission Control Protocol
  • UDP User Data Manager
  • SCTP Policy Control Function
  • the SBA simplifies interactions between NFs given that interfaces can be defined at the application layer and they can be easily expandable. For example, if AMF wishes to interact with a newly proposed NF in 5GC, then this can be performed by a simple software update.
  • each network function expresses its functionalities through Service-Based Interfaces (SBIs), which include a set of services.
  • SBIs Service-Based Interfaces
  • Each service includes a set of service operations (e.g., a set of RESTful APIs).
  • RESTful APIs e.g., RESTful APIs
  • FIG 1 from TS 23.501 [2] (Section 4.2.3), illustrates a high level architecture of 5GC (5G System Architecture) where several NFs are connected to one another via a bus.
  • 5GC 5G System Architecture
  • SCP Service Communication Proxy
  • SCP provides several benefits for 5GC networks such as load balancing, routing, message periodization, overload control, etc.
  • the interfaces N1, N2, N3, N4, and N6 do not support SBI while any interface name starts with “Nxx” support SBI (e.g., Npcf, Naf, Namf, etc.).
  • FIG 2 from TS 23.501 [2] (Section 4.2.3), illustrates a non-roaming 5G system (5GS) architecture (Non-Roaming 5G System Architecture in reference point representation) with some key NFs interacting with one another.
  • 5GS Non-Roaming 5G System Architecture in reference point representation
  • AMF is well connected to other NFs, given that it is an anchor point for relaying messages from NG-RAN and UE over N2 and N1 reference points, respectively.
  • SMF is also connected to several other NFs because it manages PDU sessions.
  • New frameworks and architectures are being developed as part of 5G network (and beyond, such as 6G networks) in order to increase the range of functionality and use cases available through 5G networks.
  • One such new framework is the use of artificial intelligence(AI)/machine learning(ML), which may be used for the optimisation of the operation of 5G networks.
  • AI artificial intelligence
  • ML machine learning
  • AI/ML models and/or data might be transferred across the AI/ML applications (e.g., application functions (AFs)), 5GC (5G core), UEs (user equipments) etc.).
  • AI/ML works could be divided into two main phases: model training and inference. During model training and inference, multiple rounds of interaction may be required.
  • the AI/ML operation/model is split into multiple parts according to the current task and environment.
  • the intention is to offload the computation-intensive, energy-intensive parts to network endpoints, whereas leave the privacy-sensitive and delay-sensitive parts at the end device.
  • the device executes the operation/model up to a specific part/layer and then sends the intermediate data to the network endpoint.
  • the network endpoint executes the remaining parts/layers and feeds the inference results back to the device.
  • Multi-functional mobile terminals might need to switch the AI/ML model in response to task and environment variations.
  • the condition of adaptive model selection is that the models to be selected are available for the mobile device.
  • it can be determined to not pre-load all candidate AI/ML models on-board.
  • Online model distribution i.e. new model downloading
  • NW network
  • the model performance at the UE needs to be monitored constantly.
  • the cloud server trains a global model by aggregating local models partially-trained by each end devices.
  • a UE performs the training based on the model downloaded from the AI server using the local training data. Then the UE reports the interim training results to the cloud server via 5G UL channels.
  • the server aggregates the interim training results from the UEs and updates the global model. The updated global model is then distributed back to the UEs and the UEs can perform the training for the next iteration.
  • model training In general, the AI/ML works can be divided into three main phases: model training, model transfer and inference. More specifically, with the introduction of federated learning, model transfer has become a crucial phase to successfully perform some AI/ML operations. Time spent for model training, inference and transmission of the AI/ML models and for output of the inference depend on computation and/or communication capabilities of participating nodes/components; hence, the time varies among different nodes/components.
  • Federated Learning in more detail, is an important machine learning technique that allows a set of participants (e.g. UEs) to engage in distributed model training without exposing their own parameters (data) other than a set of weights to outside entities. It is predicted that FL will be significantly used in several 5G and/or 6G use cases and thus, it generates a significant amount of traffic on 5G/6G networks. Therefore, it is crucial to be handled gracefully within 3GPP networks.
  • FIG 3 illustrates key interactions between the main entities involved in a distributed FL(federated learning) model training session.
  • Participants can include a car, robot, smartphone, and/or drone and the central FL server may be located in a 5G or 6G cloud within or outside the 3GPP network.
  • a central FL server may be located in other locations such as RAN, 5GC and/or UE, especially in a hierarchical FL model.
  • the FL server selects a set of participants to be part of the next training cycle based on a set of criteria that may be dynamically changed between training cycles.
  • the FL server then distributed the latest trained global model to selected participants.
  • Each participant then starts its local model training when it receives the global model and its related configurations.
  • the local training may use a particular or locally available dataset or real environment parameters.
  • Each participant then sends the trained model to the FL server once its local training is completed (i.e., when the local model is converged and is stable).
  • the FL server then aggregates all locally trained models from participants, creating a new global model which can be further distributed between participants in the next round of training sessions.
  • Synchronous Federated Learning (Sync-FL)
  • Synchronous Federated Learning is a form of FL where participants have a strict deadline for their local model training completion and also uploading the results to the central FL server. If a participant can’t meet the deadline, its results (i.e., trained models) may be unused by the central FL server, wasting 3GPP resources across UE, RAN, and Core Network. On that basis, typically, the central FL server indicates these time constraints to participants so that we can adjust their compute and network resources accordingly. That said, completing local model training at UE is also computationally intensive and thus requires a significant amount of power which is an important resource, especially for mobile devices with limited battery capacity.
  • Table 1 (Latency and user experienced UL/DL data rates for uncompressed FL), from TR 22.874 [3] (Section 7.1.6.1), shows an example of resource usage:
  • Figure 4 Performance gap vs. experience of learning for a given task: (1) with disturbance (e.g. delay) of data collection/transfer (dashed line 11) (2) without disturbance (e.g. delay) of data collection/transfer (solid line 13)), from TR 22.874 [3] (Section 7.3.1), illustrates the achievable learning performance (i.e., Experience of Learning) towards a given task when data collection/transfer is distributed (dashed line 11) and not (solid line 13).
  • AI/ML data training results
  • the vertical line 17 and the horizontal line 15 between dashed line 11 and solid line 13 shows this performance gap).
  • AI/ML data training results, etc.
  • these reasons may include one or more of the following:
  • Figure 5 from TR 22.874 [3] (Section 7.3.1), illustrates a scenario where a UE attempts to prevent data disturbance by intelligently scheduling the UL data transmission for a particular set of bits so that the required UL data transmission deadline of one second is met.
  • the UE delivered the relevant data with a size of 3 bits in two seconds as represented by data transfer 21, which caused the UL transmission deadline of one second to be missed.
  • the UE used a higher data rate of 3bps, delivering the concerning data in one second (meeting the UL transmission deadline of one second) as represented by data transfer 23.
  • the base station e.g. gNB
  • the base station has more resources to grant to other UEs for their UL transmission.
  • a UE may support two steering functionalities in this architecture: (1) MPTCP (MultiPath-TCP) and/or (2) ATSSS-LL.
  • MPTCP MultiPath-TCP
  • ATSSS-LL ATSSS-LL.
  • Each steering functionality mentioned above enables traffic steering, switching and splitting across 3GPP access and non-3GPP access following the ATSSS rules provided by the network, specifically PCF (Policy Control Function)/SMF (Session Management Function).
  • PCF Policy Control Function
  • SMS Session Management Function
  • the MPTCP Proxy functionality may be supported by UPF.
  • the UPF can then communicate with the MPTCP functionality in the UE by using MultiPath-TCP (MPTCP) (IETF RFC 8684). Additionally, the UPF may support Performance Measurement Functionality (PMF), which the UE can use to obtain access performance measurements over the UP of 3GPP and non-3GPP accesses.
  • MPTCP MultiPath-TCP
  • PMF Performance Measurement Functionality
  • the 3GPP system is moving towards a fully automated system where AI/ML techniques play a crucial role.
  • 3GPP is discussing potential use cases where AI/ML techniques can improve the 3GPP system performance significantly.
  • RAN1 and RAN2 working groups are currently discussing a few use cases in which AI/ML approaches could be an asset for improving the air interface performance, e.g., in CSI feedback, beam management, and location services.
  • these few use cases are just the beginning of using AI/ML for the air interface.
  • More 3GPP system functionality will be dependent on AI/ML techniques in the near future (not only for the air interface but generally for the entire 3GPP system, including UE, RAN, and CN).
  • the 5G control plane architecture is not currently designed to handle control messages which are very large in size, that is the case with AI/ML-related data, including neural network (NN) models, datasets, and large data collections.
  • the current 5G control plane is designed (both in CN and RAN) for handling small and high-priority messages.
  • RRC signaling cannot handle (or has difficulties in handling) large data transfers, which also impacts NAS-based solutions.
  • CP control plane
  • CP control plane
  • large transfers may impact small high-priority messages if these two types of traffic compete for the shared resources.
  • AI/ML techniques are intended to be an integral part of the 3GPP system, an architectural modification to the current 5G system is required to support such data transfers between the 3GPP system’s entities, including UE, RAN, and Core Network.
  • UP user plane
  • N3 tunnel to exchange non-user data (e.g., AI/ML-related data) between UE and CN, and also between RAN and CN.
  • an AI/ML trained model by the network (e.g. gNB) can be delivered to UE via CN over UP, and which entity in CN should receive it and relay to UE.
  • gNB intends to deliver AI/ML data (including model, training, dataset, and others) to UE, which 3GPP entities should decide on the priority of this transmission.
  • AI/ML models could be transferred to UE in multiple ways simultaneously who should decide which one to choose.
  • AI/ML data or related data could be related to AI/ML model construction, AI/ML model topology, neural network weights, datasets for training, measurements and statistics for model training, and/or any other data which can be used by AI/ML components and/or other components handling AI/ML aspects (or traffic). Additionally, it should be noted that the examples, embodiments etc. disclosed herein could also be used for non-AI/ML use cases.
  • a method of AI/ML model delivery over the user plane between CN and UE is provided.
  • AI-DDF AI Data Delivery Function
  • the AI-DDF may not only can take responsibility for delivering AI/ML models and/or related data over UP from CN to UE (and vice-versa), but the AI-DDF may also be configured to take responsibility for data collection for AI/ML purposes.
  • Figure 7 illustrates a representation of network architecture in which AI-DDF is provided/integrated.
  • AI-DDF provided in the network (for example, in accordance with the below)
  • any AI/ML data with potentially different sizes can be delivered between UE and CN with no issue.
  • the system architecture at least partly reflects 5G system architecture, it will be appreciated that it may be extended to a B5G or 6G system.
  • Figure 7 illustrates a UE 100, UPF 200, RAN 300, DN 400, 5GC 500 (including a number of NFs) and a PDU session 600 (e.g., established between UE 100 and UPF 200).
  • UE 100 includes AI-DDF 110 and UPF 200 includes AI-DDF 210, thereby implementing the new NF mentioned above.
  • Fig. 7 Also shown in Fig. 7 are the various interfaces between different entities, including N1 710, N2 720, N3 730, N4 740, Uu 750, N6 760.
  • each entity indicated in Figure. 7 may comprise one or more such entity.
  • UPF 200 may represent a single UPF or a plurality of UPFs.
  • reference to an AI-DDF may refer to an AI-DDF at one network entity (e.g., either AI-DDF 110, or AI-DDF 210 in the example of Fig. 7), to a plurality of AI-DDFs at the same network entity (e.g., in a case where multiple AI-DDFs are implemented at UE 100, UPF 200 or elsewhere), or to AI-DDFs across different network entities (e.g., both AI-DDF 110 and AI-DDF 210, each of which may also comprise, or refer to, multiple AI-DDFs).
  • Figure. 7 illustrates 5G architecture with the inclusion of the new AI-DDF component at both the UPF 200 and the UE 100.
  • the AI-DDF is provided in a UE 100 and in the CN (e.g., in UPF 200).
  • the AI-DDF components 110, 220 connect to each other via an existing PDU Session (e.g., as may be represented by PDU session 600) or a new PDU dedicated PDU Session (e.g., as may be represented by PDU session 600).
  • multiple instances of AI-DDF exist in the network (e.g., in UPF 200) connecting to the single instance at UE.
  • the AI/ML data is delivered between the UE 100 and CN entity (or entities), without the need for segmentation (an issue that exists with CP-based solutions), given that it is transferred over UP.
  • the AI-DDF is configured to assign a different priority to data transmission of different AI/ML models (e.g., if an AI/ML model has a large size and it is not time-sensitive, then it can be delivered over a path with low priority, i.e., a low priority QFI can be assigned).
  • the AI-DDF (e.g. AI-DDF 110 and/or AI-DDF 210) is implemented per PDU session (this way, e.g., multiple instances of AI-DDF may need to be established within a single UPF 200).
  • the AI-DDF is used over multiple PDU sessions terminated at the same UPF 200 where this function is activated.
  • one AI-DDF (e.g., AI-DDF 110, 210 implemented at UE 100 and UPF 200) is implemented for a plurality of PDU sessions terminated at UPF 200.
  • the AI-DDF may operate over an existing PDU session or, alternatively, a dedicated PDU session may be established with an existing UPF.
  • a network slice could additionally (or alternatively) be allocated to the AI-DDF in a PDU session (e.g., PUD session 600).
  • the network e.g., CN
  • the network should determine which UPF(s), and thus which AI-DDF, should be activated to undertake responsibility of AI/ML-related data delivery and collection.
  • the UE may use only one instance connecting to the plurality of instances (or a subset thereof) over a transport protocol such as TCP, UDP, SCTP, or MPTCP. Accordingly, the UE may determine a transport protocol to use for the instance (e.g., for the AI-DDF at the UE connecting to AI-DDFs elsewhere, such as at UPF(s)). Alternatively, another network entity (e.g., the network) may make this determination and inform the UE of the selected transport protocol.
  • a transport protocol such as TCP, UDP, SCTP, or MPTCP.
  • the choice of transport protocol depends on at least one of: the type of data to be transported (e.g., neural network models, training datasets, collected measurements and statistics, data analytics), data priority (e.g., whether data to be transferred is high priority or low priority, as may be defined in the network), reliability and security (e.g., of each transport protocol), available access in the case of multi-access scenarios, etc.
  • An AI-DDF e.g., AI-DDF 110, 210) may use multiple transport sessions simultaneously using different transport protocols. The decision of which transport protocol may be used for which AI/ML data transport could be configured by the network.
  • the AI-DDF may need to coordinate data collection across different requests (e.g. across or from different NFs).
  • an LMF and an NWDAF simultaneously train an offline AI/ML model to be used within an air interface (e.g. a gNB) and/or a UE.
  • data collection and measurement configurations may need to be configured (e.g. at the gNB and/or UE) in an optimal manner (e.g., not collecting the same information twice due to a different collection interval and deliver it over UP connections).
  • coordination with the network e.g. gNB
  • normal air interface measurements are configured (e.g., MDT, SON, RRM, RRC measurement reports, CSI, etc.).
  • an AI-DDF (at the UE and/or gNB) receives, from the network, one or more measurement configurations and/or data collection instructions, to be used or implemented by the AI-DDF.
  • the AI-DDF (e.g., AI-DDF 110 and/or AI-DDF 210) is enabled during PDU Session Establishment or Modification procedures.
  • this can be signaled within the 5G-SM capability information element (IE) of PDU Session Establishment/Modification Request message (i.e., a PDU setup message) where a UE (e.g., UE 100) can express its supported capabilities and/or desired PDU session, for example, Reflective QoS, MH6-PDU, MA-PDU Session, etc.
  • IE 5G-SM capability information element
  • the UE may need to indicate its AI/ML data delivery capability (e.g., whether it supports AI-DDF) during initial UE registration to the network so that such functionality can be activated later on over PDU Session(s) if the UE has required permission to do use it.
  • the NAS signaling may need to be extended.
  • the 5GMM capability IE may be extended with a new AIML IE as part of the UE Registration Request message.
  • the UE may indicate its support of AIML data delivery capability to the network based on a request from the network.
  • the UE is configured to transmit, to the network, an indication that the UE supports AI-DDF (or an indication that the UE does not support AI-DDF). In further examples, this may be in response to a request (of whether the UE does or does not support AI-DDF) from the network.
  • the network stores (or saves) the UE’s support of AI/ML data delivery capability in the UE context.
  • the network e.g., a network entity in the network
  • the PCF may compile and send a set of rules to the UPF (e.g., UPF 200) (directly if UPF supports SBI (Service-Based Interface) or indirectly, e.g., via SMF if not) to configure the AI-DDF (e.g., AI-DDF 210). For example, this could be done during the PDU Session Establishment/Modification procedure (e.g., PDU session setup) or via DL NAS transport (over PS-Signaling).
  • the UPF e.g., UPF 200
  • SBI Service-Based Interface
  • SMF Service-Based Interface
  • the PCF sends the set of rules to SMF first (e.g. via API(s), such as Npcf_SMPolicyControl APIs).
  • SMF then creates rules, e.g. N4 rules (including rules related to AI-DDF and, optionally, other rules such as PDR, FAR, URR, and/or QER, etc.), and pushes them to UPF.
  • N4 rules including rules related to AI-DDF and, optionally, other rules such as PDR, FAR, URR, and/or QER, etc.
  • PCF can send these rules to the UPF directly.
  • the sending of the rules to the UPF may still be preferred to be done via SMF because the SMF, in general (as a session manager), may need to be aware of such rules/policies. This may be important given that SMF needs to select the UPF at this stage and thus, selecting a UPF which supports the AI-DDF is required.
  • SMF may already know (e.g. from or during PDU Session Establishment/Modification procedure) whether the UE has requested to activate the AI-DDF and also whether the UE’s subscription is allowed such functionality (due to prior interactions with UDM).
  • the UE may have transmitted a request to activate the AI-DDF, and the SMF may have identified that such a request has been transmitted.
  • the SMF may identify whether the UE’s subscription (e.g., in the network) allows for this functionality (e.g., activating the AI-DFF).
  • the UPF may allocate an IP address for the AI-DDF (e.g., AI-DFF 210), and send this information to SMF (e.g., during N4 session establishment).
  • the SMF may send this information to AMF (e.g., via Namf_Communication_N1N2MessageTransfer message).
  • the AMF may send a PDU related message (e.g., the PDU Session Accept message) to the UE (e.g. via the NAS signaling) indicating to UE that this PDU session supports the AI-DDF.
  • the IP address of the AI-DDF e.g. AI-DFF 210 at the UPF (e.g., UPF 200) side is also provided in this message to UE.
  • this message may include rules related to AI-DDF for delivering AI/ML models and also data collection, etc.
  • the rules and policies may indicate how different AI/ML models should be delivered and/or with what transmission priority under different network conditions (or in general), and/or how rules indicated by different network entities at different times should be prioritized over one another (e.g., whether NWDAF or LMF may overwrite one or more rules originally populated by PCF, in some conditions).
  • SMF or PCF can compile and send a set of rules to UE (e.g., UE 100) for configuring AI-DDF (e.g., AI-DFF 110) at the UE side.
  • AI-DDF rules may include: how measurements should be configured at UE and/or which data should be collected in what frequency, and/or how AI/ML data should delivered at the UL direction depending on one or more of the AI-ML data size, type, priority, privacy, etc.
  • these rules may be delivered to UE during PDU Session Establishment/Modification procedure (e.g., while setting up a PDU session), e.g., within a new IE in N1 SM Container.
  • a set of USRP rules may also needed at UE so that when a particular AI/ML application traffic (e.g., related to model transfer) arrives at lower layer, the URSP rules determine which PDU session (i.e., which AI-DDF) should be used. Accordingly, the UE may receive a set of USRP rules and, based on the USRP rules, determine a PDU session or AI-DDF to be used for received AI/ML traffic (e.g. related to model transfer).
  • methods and apparatus support model delivery (or data therefor) over user plane between RAN and CN/UE. That is, the concepts and features described in the examples, embodiments etc., above are also viable at another network entity, such as NG-RAN.
  • Figure. 8 illustrates a system architecture where, in addition to the features shown in Figure. 7, RAN 300 comprises AI-DDF 310.
  • the system architecture at least partly reflects 5G system architecture, it will be appreciated that it may be extended to a B5G or 6G system.
  • DRB 800 is established, or exists, between UE 100 and RAN 300, and an N3 tunnel 900 is established, or exists, between RAN 300 and UPF 200.
  • each entity indicated in Figure. 8 may comprise one or more such entity.
  • UPF 200 may represent a single UPF or a plurality of UPFs.
  • an AI-DDF may refer to an AI-DDF at one network entity (e.g., either AI-DDF 110, AI-DDF 210 or AI-DDF 310, in the example of Figure. 8), to a plurality of AI-DDFs at the same network entity (e.g., in a case where multiple AI-DDFs are implemented at UE 100, UPF 200, RAN 300 or elsewhere), or to AI-DDFs across different network entities (e.g., two or more of AI-DDF 110, AI-DDF 210 and AI-DDF 310, each of which may also comprise, or refer to, multiple AI-DDFs).
  • AI-DDF at one network entity
  • AI-DDF 110 e.g., AI-DDF 110, AI-DDF 210 or AI-DDF 310, in the example of Figure. 8
  • a plurality of AI-DDFs at the same network entity e.g., in a case where multiple AI-DDFs are implemented at UE 100, UPF
  • Figure. 8 illustrates the AI-DDF 110, 210, 310 at the UE 100, RAN 300 (e.g., gNB) and CN (i.e., UPF 200).
  • the AI-DDF 310 at RAN 300 e.g. NG-RAN 300, or gNB
  • the AI-DDF 110, 210, 310 can deliver AI/ML-related data (including large models) over user plane.
  • AI-DDF can be activated per gNB (e.g., corresponding to RAN 300) where a new connection between gNB and UPF (e.g., UPF 200) can be established (e.g., over an existing N3 tunnel such as N3 tunnel 900).
  • gNB e.g., corresponding to RAN 300
  • UPF e.g., UPF 200
  • the connection between gNB (or RAN 300) and UPF (e.g., UPF 200) is similar to the case of UE and UPF, where a new transport protocol can be used over the existing N3 tunnel.
  • the choice of transport protocol e.g., UDP, TCP, MPTCP
  • This approach may work even in the case of the UE is in RRC_INACTIVE state, because, in that state, the N3 resources are still available between the gNB and UPF.
  • a similar session is established between the UE (e.g., UE 100) and gNB (or RAN 300), e.g., after the completion of a PDU Session Establishment procedure or after the UE has registered in a PLMN.
  • the AI-DDF e.g., AI-DDF 310
  • the gNB or NG-RAN 300
  • This new session between the UE and gNB can be established during PDU Session Established/Modification procedure (e.g.
  • DRB 800 may be set up by RRC procedure, such as RRC Reconfiguration procedure), reusing the same security and integrity credentials of the related PDU session.
  • the information regarding the IP address of the AI-DDF at gNB (e.g., AI-DDF 310) may be sent to UE, for example, via the NAS PDU Session Accept Message. In some cases, this approach only works when UE is in RRC_CONNECTED and thus all UP resources (e.g., DRBs) should be allocated between the UE and gNB.
  • the UE 100 can interact with the gNB (or NG-RAN 300) over the data plane (via a DRB 800), and the gNB (or NG-RAN 300) can interact with UPF 200 over a new transport session over N3 (e.g., via N3 tunnel 900).
  • the gNB or NG-RAN 300
  • N3 e.g., via N3 tunnel 900
  • an offline model is trained in CN (e.g., by NWDAF or LMF or in another newly added or existing network entity or function) and it should be delivered to the gNB, but the CP route cannot be used due to the large size of AI/ML data, which may cause congestion on the CP path to gNB, delaying important CP messages.
  • the model can be delivered to UPF either indirectly, such as through SMF (e.g. via N4), or directly if UPF supports SBI.
  • the AI-DDF at UPF may then deliver this to the AI-DDF component at the gNB (e.g., over N3 interface), where it will be delivered further to UE if needed (i.e. optionally delivered to AI-DDF at the UE), such as shown in Figure 8.
  • the gNB trains an offline AI/ML model that should be used at a UE.
  • the gNB could deliver it to UPF via a new AI-DDF session (between gNB and UPF over N3, for instance), and then the AI/ML model will be delivered to the UE via non-3GPP access, such as Wi-Fi, in the case of an MA-PDU session.
  • the gNB may directly deliver this AI/ML related data to the UE over the UP connection between AI-DDF instances at UE and gNB.
  • the UE trains an offline AI/ML model, that should be used at a gNB.
  • the UE may transport the trained model to UPF via WiFi access, such as in the case of the MA-PDU session (i.e., UP over non-3GPP access).
  • the UPF may then deliver it to gNB over N3, as discussed earlier.
  • This scenario may be advantageous when the UE does not have a good connection over 3GPP access (e.g., over Uu interface) but has a good connection over non-3GPP access (e.g., over WiFi), and the AI/ML model should be transported to gNB as soon as possible.
  • the UE trains an AI/ML model, which should be used at CN (e.g., LMF or NWDAF).
  • the UE may initially transport the trained model to a gNB via UP over a low-priority DRB. From the gNB, the model can be delivered to UPF over N3 (via UP). From UPF (e.g., by AI-DDF) to its destination (e.g., LMF), the model may be delivered either directly over SBI or via N4. This way, the 5GC control plane is used minimally for transferring AI/ML data, which are typically very large and time-insensitive.
  • the UE may transport the trained model to UPF (e.g. AI-DDF at UPF) via WiFi access in the case of the MA-PDU session. The UPF (AI-DDF) may then deliver it to LMF via N4 or directly if UPF supports SBI.
  • the network pushes a set of one or more rules to gNB (e.g. AI-DDF at the gNB).
  • the PCF initially pushes policy and charging control (PCC) rule(s) to SMF.
  • PCC policy and charging control
  • the SMF may then create AI-DDF configurations or rules/policies configurations accordingly (e.g., based on the PCC rule(s)) and send them to AMF, e.g. via Namp_Communication API over non-ue-n2-messages transfer service operation.
  • the AMF may then relay this to the gNB, e.g. over NGAP (NG application protocol).
  • NGAP NG application protocol
  • the SMF may also transfer rules to gNB via Namf_Communication_N1N2MessageTransfer message, indicating some information needed to be parsed by gNB (the N2 part).
  • the gNB then pushes those rules to the AI-DDF component (e.g., AI-DDF at the gNB) accordingly.
  • the SMF directly pushes rules to gNB over HTTP. Otherwise, AMF will be used to relay messages between SMF and gNB (i..e AI-DDF at gNB). Direct interaction with AI-DDF may not be necessary at gNB and UPF by SMF, provided that the gNB and UPF act as a relay.
  • an AI-DDF e.g., in or across UE, RAN or UPF
  • the AI/ML data is time-insensitive, e.g., when delivering an offline trained model, then the AI-DDF may determine to use UP paths.
  • every AI-DDF instance may determine how to deliver AI/ML data to the next node (this may be according to a policy provided by the network and/or one or more analytics that AI-DDF may collect from the network and UE).
  • the AI-DDF at the UE may initially determine to deliver AI/ML data to gNB via CP.
  • the AI-DDF at the gNB may determine to deliver (the) AI/ML data to an NF in 5GC via UPF (i.e., partially via UP) or directly via CP.
  • the UE and gNB may exchange AI/ML-related models and/or other AI/ML-related data via RRC signaling over an SRB (e.g., SRB4 or a new SRB with a particular priority depending on the AI/ML data size and latency requirements, assuming that the problem of RRC segmentation is resolved).
  • SRB e.g., SRB4 or a new SRB with a particular priority depending on the AI/ML data size and latency requirements, assuming that the problem of RRC segmentation is resolved.
  • SRB e.g., SRB4 or a new SRB with a particular priority depending on the AI/ML data size and latency requirements, assuming that the problem of RRC segmentation is resolved.
  • SRB e.g., SRB4 or a new SRB with a particular priority depending on the AI/ML data size and latency requirements, assuming that the problem of RRC segmentation is resolved.
  • NWDAF NWDAF
  • other NFs in 5GC e.g., 5
  • the UE may initially use RRC signaling to carry the model to gNB over the Uu interface (this may also be a valid case when the UE is in RRC_INACTIVE and wishes to be in that state due to a lack of battery resource, for example).
  • the model can be delivered to UPF (e.g., AI-DDF at UPF) over UP (passing through the N3 tunnel) and from UPF to NWDAF/LMF, directly if UPF supports SBI or via SMF (N4) otherwise.
  • UPF e.g., AI-DDF at UPF
  • UP passing through the N3 tunnel
  • N4 SMF
  • the AMF, and CP may not be a bottleneck by the large data size of AI/ML traffic.
  • the CP of 5GC is partially being used by AI/ML traffic (e.g., between UPF and NWDAF/LMF if the UPF support SBI, which may be the case).
  • using NGAP to transfer AI/ML-related data to CN from gNB may not be an optimal approach in several scenarios, especially when the data size is very large; transferring AI/ML-related data via NAS signaling between UE and CN may not be an optimal approach in several scenarios, especially when the data size is very large and not time-sensitive; and, in such scenarios, UP paths may be preferred for transporting AI/ML-related data.
  • a network comprising a first network entity (e.g., a UE) including a first data transfer component (e.g., at least one first AI-DDF) and a second network entity (e.g., a UPF) including a second data transfer component (e.g., at least one second AI-DDF), wherein: the first data transfer component is configured to receive AI/ML related data (or other data) from, or transmit AI/ML related data (or other data) to, the second data transfer component (e.g., via one or more PDU sessions or a network slice allocated to the first and second data transfer components in a PDU session).
  • a first network entity e.g., a UE
  • first data transfer component e.g., at least one first AI-DDF
  • a second network entity e.g., a UPF
  • second data transfer component e.g., at least one second AI-DDF
  • the first and/or second data transfer component is configured to obtain, from one or more network functions included in the network, the AI/ML related data (e.g., data for training a AI/ML model, information for configuring measurement of data for training an AI/ML model etc.).
  • the AI/ML related data e.g., data for training a AI/ML model, information for configuring measurement of data for training an AI/ML model etc.
  • the first and/or second data transfer component is configured to receive, from the network, configuration information for obtaining the AI/ML related data; and, optionally, to obtain the AI/ML related data based on the configuration information.
  • the first network entity is configured to indicate, to the network (e.g., during initial registration in the network), that the first network entity comprises or supports (or does not comprise or support) the first data transfer component.
  • this indication may be included in an IE transmitted to the network by the first network entity.
  • the indication is transmitted in response to a request, relating to whether the first network entity includes or supports the data transfer component, received form the network.
  • a third network entity in the network may be configured to store the indication from the first network entity.
  • the second data transfer component is configured or activated in the second network entity in response to a message (e.g., instruction) received from the network.
  • the first data transfer component and/or the second data transfer component determines whether to use a first communication method (e.g., communicating via control plane) or a second communication method (e.g., communicating via user plane) for transmitting specific AI/ML related data.
  • the determination may be based on a characteristic of the specific AI/ML related data (e.g., if the data is to be used for AI/ML inference, if the data is time-insensitive, a priority of the data etc.).
  • the first data transfer component and/or the second data transfer component may re-determine a communication method (e.g. the first communication method, the second communication method or another communication method) for transmitting specific AI/ML related data.
  • a first network entity in accordance with any one or more of the above examples.
  • a second network entity in accordance with any one or more of the above examples.
  • the first data transfer component and/or the second data transfer are implemented in a single PDU session, or in a plurality of PDU sessions terminated at the second network entity.
  • the first data transfer component and/or the second data transfer are configured to operate over an existing PDU session or in a dedicated PDU session.
  • the second network entity is configured to receive, from a fourth network entity (e.g., PCF or SMF), one or more rules for configuring the second data transfer component.
  • the one or more rules are compiled by a fifth network entity (e.g., PCF) and transmitted to the second network entity via the fourth network entity (e.g., SMF).
  • the fourth network entity e.g., SMF
  • the one or more rules compiled by the fifth network entity are received via the fourth network entity (e.g., SMF).
  • the fourth network entity e.g., SMF
  • the network further comprises a sixth network entity (e.g., RAN, or gNB) comprising a third data transfer component (e.g., third AI-DDL).
  • the third data transfer component is configured to cooperate with at least one of the first data transfer component and the second data transfer component to exchange the AI/ML related data.
  • Figure 9 is a block diagram of an exemplary apparatus, or network entity, that may be used in examples of the present disclosure.
  • entity may be implemented, for example, as a network element on a dedicated hardware, as a software instance running on a dedicated hardware, and/or as a virtualised function instantiated on an appropriate platform, e.g. on a cloud infrastructure.
  • the entity 1000 comprises a processor (or controller) 1001, a transmitter 1003 and a receiver 1005.
  • the receiver 1005 is configured for receiving one or more messages from one or more other network entities, for example as described above.
  • the transmitter 1003 is configured for transmitting one or more messages to one or more other network entities, for example as described above.
  • the processor 1001 is configured for performing one or more operations, for example according to the operations as described above.
  • Figure 10 is a flow diagram illustrating a method according to various examples of the present disclosure.
  • the method is performed by a first network entity comprising a first artificial intelligence (AI)/machine learning (ML) data transfer function.
  • the first network entity is included in a network.
  • the first network entity is a UE, a UPF or a gNB or NG-RAN.
  • the first network entity establishes a first connection between the first AI/ML data transfer function and at least one second AI/ML data transfer function included in the network.
  • the first network entity controls the first AI/ML data transfer function to coordinate communicating AI/ML data with the network over user plane (UP) or control plane (CP) based on at least one rule or policy, wherein the AI/ML data relates to an AI/ML operation.
  • UP user plane
  • CP control plane
  • Each of the at least one second AI/ML data transfer function is included in a second network entity.
  • communicating AI/ML data e.g. an AI/ML model
  • communicating AI/ML data e.g. an AI/ML model
  • communicating AI/ML data e.g. an AI/ML model
  • communicating AI/ML data e.g. an AI/ML model
  • communicating AI/ML data to a UE as described herein may also be regarded as communicating the AI/ML data to an AI-DDF at the UE, and vice versa (and similar being the case for a UPF too).
  • one or more features or operations may be omitted, modified or moved (e.g., to change the order of the features or the operations), if desired and appropriate.
  • one or more features or operations from any example/embodiment may be combined with features or operations from any other example/embodiment.
  • the present disclosure should be considered to include all combinations of two or more of the embodiments, examples etc. disclosed herein, and all combinations of two or more of the features disclosed herein.
  • Such an apparatus and/or system may be configured to perform a method according to any aspect, embodiment or example disclosed herein.
  • Such an apparatus may comprise one or more elements, for example one or more of receivers, transmitters, transceivers, processors, controllers, modules, units, and the like, each element configured to perform one or more corresponding processes, operations and/or method steps for implementing the techniques described herein.
  • an operation/function of X may be performed by a module configured to perform X (or an X-module).
  • the one or more elements may be implemented in the form of hardware, software, or any combination of hardware and software.
  • examples of the present disclosure may be implemented in the form of hardware, software or any combination of hardware and software. Any such software may be stored in the form of volatile or non-volatile storage, for example a storage device like a ROM, whether erasable or rewritable or not, or in the form of memory such as, for example, RAM, memory chips, device or integrated circuits or on an optically or magnetically readable medium such as, for example, a CD, DVD, magnetic disk or magnetic tape or the like.
  • volatile or non-volatile storage for example a storage device like a ROM, whether erasable or rewritable or not
  • memory such as, for example, RAM, memory chips, device or integrated circuits or on an optically or magnetically readable medium such as, for example, a CD, DVD, magnetic disk or magnetic tape or the like.
  • the storage devices and storage media are embodiments of machine-readable storage that are suitable for storing a program or programs comprising instructions that, when executed, implement certain examples of the present disclosure. Accordingly, certain examples provide a program comprising code for implementing a method, apparatus or system according to any example, embodiment and/or aspect disclosed herein, and/or a machine-readable storage storing such a program. Still further, such programs may be conveyed electronically via any medium, for example a communication signal carried over a wired or wireless connection.
  • Figure. 12 illustrates a block diagram of a terminal (or a user equipment (UE)), according to embodiments of the present disclosure.
  • Figure. 12 corresponds to the example of the UE of Figure. 1.
  • the UE may include a transceiver 1210, a memory 1220, and a processor 1230.
  • the transceiver 1210, the memory 1220, and the processor 1230 of the UE may operate according to a communication method of the UE described above.
  • the components of the UE are not limited thereto.
  • the UE may include more or fewer components than those described above.
  • the processor 1230, the transceiver 1210, and the memory 1220 may be implemented as a single chip.
  • the processor 1230 may include at least one processor.
  • the transceiver 1210 collectively refers to a UE receiver and a UE transmitter, and may transmit/receive a signal to/from a base station or a network entity.
  • the signal transmitted or received to or from the base station or a network entity may include control information and data.
  • the transceiver 1210 may include a RF transmitter for up-converting and amplifying a frequency of a transmitted signal, and a RF receiver for amplifying low-noise and down-converting a frequency of a received signal.
  • the transceiver 1210 may receive and output, to the processor 1230, a signal through a wireless channel, and transmit a signal output from the processor 1230 through the wireless channel.
  • the memory 1220 may store a program and data required for operations of the UE. Also, the memory 1220 may store control information or data included in a signal obtained by the UE.
  • the memory 1220 may be a storage medium, such as read-only memory (ROM), random access memory (RAM), a hard disk, a CD-ROM, and a DVD, or a combination of storage media.
  • the processor 1230 may control a series of processes such that the UE operates as described above.
  • the transceiver 1210 may receive a data signal including a control signal transmitted by the base station or the network entity, and the processor 1230 may determine a result of receiving the control signal and the data signal transmitted by the base station or the network entity.
  • Figure. 13 illustrates a block diagram of a base station, according to embodiments of the present disclosure.
  • Figure. 13 corresponds to the example of the RAN of Figure. 1.
  • the base station may include a transceiver 1310, a memory 1320, and a processor 1330.
  • the transceiver 1310, the memory 1320, and the processor 1330 of the base station may operate according to a communication method of the base station described above.
  • the components of the base station are not limited thereto.
  • the base station may include more or fewer components than those described above.
  • the processor 1330, the transceiver 1310, and the memory 1320 may be implemented as a single chip.
  • the processor 1330 may include at least one processor.
  • the transceiver 1310 collectively refers to a base station receiver and a base station transmitter, and may transmit/receive a signal to/from a terminal or a network entity.
  • the signal transmitted or received to or from the terminal or a network entity may include control information and data.
  • the transceiver 1310 may include a RF transmitter for up-converting and amplifying a frequency of a transmitted signal, and a RF receiver for amplifying low-noise and down-converting a frequency of a received signal.
  • the transceiver 1310 may receive and output, to the processor 1330, a signal through a wireless channel, and transmit a signal output from the processor 1330 through the wireless channel.
  • the memory 1320 may store a program and data required for operations of the base station. Also, the memory 1320 may store control information or data included in a signal obtained by the base station.
  • the memory 1320 may be a storage medium, such as read-only memory (ROM), random access memory (RAM), a hard disk, a CD-ROM, and a DVD, or a combination of storage media.
  • the processor 1330 may control a series of processes such that the base station operates as described above.
  • the transceiver v10 may receive a data signal including a control signal transmitted by the terminal, and the processor 1330 may determine a result of receiving the control signal and the data signal transmitted by the terminal.

Landscapes

  • Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

La présente divulgation porte sur un système de communication 5G ou 6G permettant de prendre en charge un débit supérieur d'émission de données. Selon un exemple, la présente divulgation concerne une première entité de réseau comprenant une première fonction de transfert de données d'intelligence artificielle (IA)/apprentissage automatique (ML). La première entité de réseau est incluse dans un réseau et comprend : un émetteur; un récepteur; et au moins un processeur configuré pour : établir une première connexion entre la première fonction de transfert de données IA/ML et au moins une seconde fonction de transfert de données IA/ML incluse dans le réseau; et commander la première fonction de transfert de données IA/ML pour coordonner la communication de données IA/ML avec le réseau sur le plan utilisateur (UP) ou le plan de contrôle (CP) sur la base d'au moins une règle ou politique, les données IA/ML se rapportant à une opération IA/ML; chacune de la ou des secondes fonctions de transfert de données IA/ML étant incluse dans une seconde entité de réseau. L'invention divulgue également des exemples de procédés d'une première entité de réseau.
PCT/KR2024/003988 2023-04-06 2024-03-28 Procédé et appareil pour une émission de données ia/ml dans un système de communication sans fil Pending WO2024210402A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
GBGB2305221.0A GB202305221D0 (en) 2023-04-06 2023-04-06 Methods and apparatus for ai'ml data transfer
GB2305221.0 2023-04-06
GB2403214.6 2024-03-05
GB2403214.6A GB2632725A (en) 2023-04-06 2024-03-05 Methods and apparatus for AI/ML data transfer

Publications (1)

Publication Number Publication Date
WO2024210402A1 true WO2024210402A1 (fr) 2024-10-10

Family

ID=86378833

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2024/003988 Pending WO2024210402A1 (fr) 2023-04-06 2024-03-28 Procédé et appareil pour une émission de données ia/ml dans un système de communication sans fil

Country Status (2)

Country Link
GB (2) GB202305221D0 (fr)
WO (1) WO2024210402A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190208379A1 (en) * 2016-05-04 2019-07-04 Nokia Solutions And Networks Oy Control plane user plane correlation function
US20200374863A1 (en) * 2019-05-24 2020-11-26 Huawei Technologies Co., Ltd. Location-based beam prediction using machine learning
US20220108214A1 (en) * 2020-08-13 2022-04-07 Electronics And Telecommunications Research Institute Management method of machine learning model for network data analytics function device
WO2023015431A1 (fr) * 2021-08-10 2023-02-16 Qualcomm Incorporated Indication basée sur des dci pour déclencher le modèle ml combiné
WO2023015430A1 (fr) * 2021-08-10 2023-02-16 Qualcomm Incorporated Configuration de paramètres de structure ml combinée

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190208379A1 (en) * 2016-05-04 2019-07-04 Nokia Solutions And Networks Oy Control plane user plane correlation function
US20200374863A1 (en) * 2019-05-24 2020-11-26 Huawei Technologies Co., Ltd. Location-based beam prediction using machine learning
US20220108214A1 (en) * 2020-08-13 2022-04-07 Electronics And Telecommunications Research Institute Management method of machine learning model for network data analytics function device
WO2023015431A1 (fr) * 2021-08-10 2023-02-16 Qualcomm Incorporated Indication basée sur des dci pour déclencher le modèle ml combiné
WO2023015430A1 (fr) * 2021-08-10 2023-02-16 Qualcomm Incorporated Configuration de paramètres de structure ml combinée

Also Published As

Publication number Publication date
GB2632725A (en) 2025-02-19
GB202403214D0 (en) 2024-04-17
GB202305221D0 (en) 2023-05-24

Similar Documents

Publication Publication Date Title
WO2018110952A2 (fr) Appareil et procédé de commande de flux de données dans un système de communication sans fil
EP4458050A1 (fr) Procédé et dispositif de communication pour service xr dans un système de communication sans fil
EP3738339A1 (fr) Procédé de délestage de données, procédé d'attribution d'identifiant de drb, procédé de libération de ressource, et dispositif
WO2021172874A1 (fr) Procédé et appareil d'exécution de fonction de réseau virtualisé
WO2021091321A1 (fr) Dispositif et procédé permettant de prendre en charge la qualité de service dans un système de communication sans fil
WO2023068854A1 (fr) Procédé et appareil de gestion de qualité de service de trafic sur liaison montante dans un système de communication sans fil
WO2024096613A1 (fr) Procédé et appareil pour connecter un terminal basé sur un flux qos dans un système de communication sans fil
WO2023191502A1 (fr) Procédé et dispositif de fourniture d'un trajet d'accès dans un système de communication sans fil
WO2023146310A1 (fr) Procédé et appareil pour la prise en charge de changement de tranche de réseau dans un système de communication sans fil
WO2019160298A1 (fr) Procédé de délestage de données, procédé d'attribution d'identifiant de drb, procédé de libération de ressource, et dispositif
WO2024035033A1 (fr) Procédé et appareil de service utilisant la fonction modem dans un système de communication sans fil
WO2024071762A1 (fr) Dispositif électronique et procédé d'audit d'abonnement
WO2024096710A1 (fr) Entraînement fl à multiples fonctionnalités de modèle d'un modèle d'apprentissage ia/ml pour de multiples fonctionnalités de modèle
WO2024210402A1 (fr) Procédé et appareil pour une émission de données ia/ml dans un système de communication sans fil
WO2024005589A1 (fr) Procédé et dispositif de notification d'événement de régulation de congestion dans un système de communication sans fil
WO2024167340A1 (fr) Procédé et appareil de sélection d'un type de support radio dans des déploiements à double connectivité prenant en charge l'ia/ml
WO2024210369A2 (fr) Procédé et appareil de gestion d'une session pdu dans un système de communication sans fil
WO2024232572A1 (fr) Procédé et appareil de gestion de données d'ia et d'am dans un système de communication sans fil
WO2025058409A1 (fr) Procédé et appareil permettant d'établir une connexion dans un système de communication sans fil
WO2025174060A1 (fr) Procédé et appareil de direction de trafic dans un système de communication sans fil
WO2024096638A1 (fr) Procédés et appareil relatifs à la gestion de faisceaux
WO2025034089A1 (fr) Procédé et appareil de synchronisation de clés dans un groupe de cellule secondaire sélectif
WO2024048871A1 (fr) Procédé et appareil de délestage de trafic dans une communication sans fil
WO2024237527A1 (fr) Gestion d'id de modèle dans une opération ia/ml dans un réseau
WO2024096620A1 (fr) Procédés exécutés par un nœud de réseau et un nœud, et dispositif de nœud

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 24785109

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE