目录
Meeting Agenda
Meeting Agenda for #108 FS_NR_AIML_air
8.21 Study on Artificial Intelligence (AI)/Machine Learning (ML) for NR air interface
8.21.1 General and work plan
8.21.2 Specific issues related to use case for AI/ML
8.21.3 Interoperability and testability aspect
8.21.4 Moderator summary and conclusions
Meeting schedule for #108 FS_NR_AIML_air
RAN4 Main session, 11:00-13:00, August 23
RAN4 Main Ad hoc session, 18:20-19:20, August 25
General and work plan sub-topics summary
-
For Requirements for data collection
•Option 1: RAN4 to study requirements for data collection (e.g. accuracy) especially for training data
•Option 2: RAN4 to study requirements for data collection depending on outcome of other groups.
•Option 3: RAN4 should not study requirements for data collection(in particular for training)
-
For Handling of generalization - robustness, whether should RAN4 requirements/tests ensure that performance is maintained under different scenarios (AI/ML model maintains performance level under “unseen” inputs in training)?
-
For Handling of generalization - dynamically changing environment, whether should RAN4 study the requirements/tests?
-
For AI/ML model complexity,
•Option 1: KPIs related to model computation complexity should be considered (actual KPIs can be further discussed: FLOPS, # of parameters, etc.).
•Option 2: Model complexity should be set as a side condition for requirements.
-
For Requirements for LCM,
•Option 1: Wait for progress in other working groups before further discussing any LCM related topics
•Option 2: Study multi-sample / multi-user involved performance evaluation.
•Option 3: Study requirements definition for dynamically changing scenarios(accuracy and latency of monitoring)
•Option 4: No need to study anything else
-
What is RAN4 testing goals?
•Option 1: The testing goal is to verify whether a specific AI/ML model can be conducted in a proper way.
•Option 2: The testing goal is to verify whether the performance gain of AI/ML model can be achieved for a static scenario/configuration.
-
Whether is there need to study a framework to enable post deployment tests for model updates and/or drift validation(and possible other use cases)?
-
Whether should Overhead be considered when formulating performance requirements and comparing with legacy performance?
-
For Encoder/decoder terminology for two sided model, whether is there need for reference encoder/decoder definition ?
Specific issues related to use case for AI/ML sub-topics summary
-
For Metrics/KPIs for CSI requirements/tests,
•Option 1: Only use throughput (absolute or relative)
•Option 2: Use throughput and other intermediate metrics/KPIs(SGCS, NMSE, etc).
•Option 3: use throughput and overhead
-
For Metrics/KPIs for Beam prediction requirements/tests,
•Option 1: RSRP accuracy
•Option 2: beam prediction accuracy :Top-1(%), Top-K(%)
•Option 3: maximum RSRP among top-K predicted beams is larger than the RSRP of the strongest beam – x dB, x>relative measurement accuracy requirement
•Option 4: overhead/latency reduction
•Option 5: combinations of above options
-
For Metrics/KPIs for positioning requirements/tests,
•Option 1: direct positioning accuracy (ground truth vs. reported)
•Option 2: RSTD/UE Rx-Tx accuracy
•Option 3: CIR/PDP/RSRP accuracy
•Option 4: LOS/NLOS
-
Whether is there need to develop requirements for model delivery/update/transfer?
-
To study the Feasibility of Intermediate KPIs for CSI requirements or LCM
•How such metrics(SGCS, NMSE, etc) can be accessed and how to set a requirement on them or compare them to the ground truth
•How can the ground truth be established in a testing environment
-
Whether is there need to study the possibility of defining accuracy requirements for measurement data or labelled data?
-
Whether is there need to study a framework for beam prediction requirements on the network side?
Interoperability and testability aspect Sub-topics summary
-
For Encoder/decoder for 2-sided model, Pros/cons are listed below.
•Recommended WF: Down-select option 6
Option | content | pros | cons |
---|---|---|---|
1 | reference decoder is provided by the UE vendor of the encoder under test so that the encoder and decoder are jointly designed and trained | • alleviate the impact of model mismatch. | • Encoder may not work for the decoders of structure mismatch or not being jointly trained. |
2 | reference decoder is provided by the vendor of the decoder(infra-vendors) so that the encoder and decoder are jointly designed and trained | • Can test with real infra vendor decoder, no additional RAN4 decoder used in practice and the test can reflect the performance in the field. | • different network vendors provide different models. |
3 | The reference decoder(s) are fully specified and captured in RAN4 spec to ensure identical implementation across equipment vendors without additional training procedure needed | • Simpler testing procedure since TE can directly implement the decoder | • Possible lengthy RAN4 discussion to agree on one (or more) fully specified reference decoder |
4 | The reference decoder(s) are partially specified and captured in RAN4 spec | • TBD | • the unspecified part is left to TE vendor implementation, TEs may have different reference decoders |
6 | Test decoder is specified and captured in RAN4 and is provided by test environment vendor. The encoder and decoder can be jointly trained | • TBD | • TBD |
-
For Test encoder/decoder further discussion, pros, cons, TE implementation issues, high level test procedure, RAN4 testing issues and more relevant details will be gathered and discussed.
-
For Reference block diagram for 1-sided model,
-
For Reference block diagram for 2-sided model,
-
For Interoperability aspects,
Model Training | Model monitoring and Model selection/(de)activation/switching/fallback | Model Inference | |
---|---|---|---|
N/W-UE Collaboration Level-x | N/A (training in non-3GPP entities or offline training as baseline, model training perf. guaranteed by model inference perf.) | N/A | Interoperability guaranteed by - Use case KPI |
N/W-UE Collaboration Level-y | N/A (training in non-3GPP entities or offline training as baseline, model training perf. guaranteed by model inference perf.) | Interoperability guaranteed by - Model monitoring perf. - Model selection/(de)activation/switching/fallback perf. | Interoperability guaranteed by - Use case KPI |
N/W-UE Collaboration Level-z | N/A for one-sided model training (training in non-3GPP entities or offline training as baseline, model training perf. guaranteed by model inference perf.) | Interoperability guaranteed by - Model monitoring perf. - Model selection/(de)activation/switching/fallback perf. | Interoperability guaranteed by - Use case KPI |
-
For Channel Models for testing,
•Option 1: RAN4 should start discussing/developing CDL models
•Option 2: TDL models are enough.
•Option 3: Postpone this discussion for now
-
Whether should RAN4 send an LS to RAN1 for now to ask about how graceful the degradation of an AI model is when scenarios are changing?