-
RAN3 Release 17: Full Breakdown of 3GPP Enhancements, Outcomes, and Technical Priorities

RAN3 Release 17: In-Depth Look at 5G Upgrades, Timelines, and Focus Areas
3GPP Release 17 marks a significant milestone in the 5G network’s development. While RAN1 is busy improving the physical layer and RAN2 is focused on L2 and RRC protocols, RAN3 plays a vital role in shaping RAN architecture, interfaces, and coordinating higher layers.
The image included gives a detailed quarterly breakdown of RAN3’s activities for Release 17, showcasing the various work items, their timelines, and the allocations for Technical Units (TUs). This blog post takes that visual guide and turns it into a straightforward narrative that’s optimized for search engines and accurate for telecom professionals and advanced learners alike.
What RAN3 Does in the 3GPP Ecosystem
RAN3 is in charge of the Radio Access Network’s architecture, which covers:
The X2/Xn interfaces
NG-RAN architecture
Core-RAN interactions
Coordinating mobile handovers
Enhancing Quality of Experience (QoE) and mobility
Harmonizing Multi-RAT architecture
With Release 17, 5G architecture gets even stronger—this is especially important as networks move toward NTN, enterprise private 5G, integrated access/backhaul (IAB), and improved positioning.
Timeline Overview: RAN3 Release 17 Work (2020 Q2 – 2021 Q2)
The timeline lays out various work items that are being progressed through multiple phases. Here’s a summary of the features illustrated in the image:
| Feature | Quarter | 2020 Q2 | 2020 Q3 | 2020 Q4 | 2021 Q1 | 2021 Q2 |
| SON/MDT Enhancements | 6 | 2 | 4 | 2 | 3 |
| NR QoE Study | 2 | 1 | 2 | 1 | 2 |
| NR Multicast | 2 | 2 | 2 | 1 | 1 |
| NR Non-Public Network Enhancements | 0 | 1 | 1 | 5 | 1 |
| IAB Enhancement | 2 | 1 | 2 | 2 | 3 |
| NR over NTN | 1 | 5 | 1 | 2 | 10 |
| NB-IoT over NTN | – | – | – | – | 1 |
| NR Positioning Enhancements | 0 | 0 | 2 | 0 | 5 |
| LTE UP/CP Split | 0 | 1 | 2 | 0 | 0 |
| Corrections / TElx | 5 | 1 | 3 | 0 | 1 |
| Misc. Impacts (RAN1/2/SA/CT) | 4.5 (-1) | 2 (0) | 3.5 (+0.5) | 4 (+0.5) | 9 (0) |
Total available TUs per meeting: 12
Overbudget units are noted in brackets.
This allocation gives insight into how RAN3 manages architectural complexity across various projects happening simultaneously.
Key Enhancements in RAN3 Release 17
Let’s take a closer look at the features based on the timeline.
- SON/MDT Enhancements (Self-Organizing Networks & Minimization of Drive Tests)
RAN3 provides architectural backing for:
Improved MDT reporting for better coverage mapping
Enhanced analytics for SON algorithms
Fewer physical drive tests needed
Increased signaling efficiency between RAN and core
These improvements are key, especially as networks incorporate more small cells, private networks, and adaptable spectrum systems.
- NR QoE Study & Architecture Enhancements
RAN3 tackles Quality of Experience (QoE) by looking at:
Reporting mechanisms for per-service QoE
Linking KPIs like latency, jitter, and throughput to user experience classifications
Signaling improvements at the interface level for detailed QoE monitoring
With Release 17, this study translates into tangible architectural changes, enabling:
Mobility decisions that reflect QoE
Optimizations at the session level for XR, gaming, and real-time applications.
- NR Multicast (Enhanced Broadcast/Multicast Architecture)
RAN3 bolsters the architecture needed for multicast delivery by enhancing:
RRC signaling across various cells
Coordination for inter-node multicast
Resource pooling for group broadcast sessions
This is vital for areas like:
Public safety messaging
Stadium broadcasts
Connected vehicle platooning
- Non-Public Network (NPN) Enhancements
With enterprise 5G booming, RAN3 strengthens the NPN framework by offering:
Better separation between public and private slices
Clear boundaries between RAN and core
Smooth mobility between public and private networks
Enhanced access control and authentication mechanisms
These advancements are directly aimed at supporting Industry 4.0 implementations.
- IAB (Integrated Access and Backhaul) Enhancements
RAN3 fine-tunes the architecture for operating multi-hop IAB. Upgrades include:
Handling dynamic topologies
Efficient signaling between IAB donor nodes and downstream relays
Optimizing control-plane tunnels
Advanced routing for multi-hop backhaul
IAB is crucial in mmWave settings where a wired backhaul isn’t a practical option.
- NR over NTN (Non-Terrestrial Networks)
Release 17 facilitates 5G satellite integration, with RAN3 focusing on:
NG-RAN interface adaptations for NTN
Managing satellite mobility and delays
RRC improvements for tracking in orbit
Support for hybrid terrestrial-satellite mobility
RAN3 also backs NB-IoT over NTN, which connects low-power IoT devices through LEO/GEO satellites.
- NR Positioning Enhancements
Positioning is a standout feature for 5G. RAN3 supports:
New signaling methods for precise multi-cell positioning
Integration between location servers and RAN nodes
Architecture designed for uplink TDOA, AoA, and PRS-based positioning
Potential applications are:
Autonomous vehicles
Indoor navigation
Robotics
XR alignment systems
- LTE UP/CP Split Enhancements
Even though LTE is well-established, Release 17 updates help it coexist with 5G. Enhancements cover:
Improved handling of splits in Cloud RAN architectures
Better signaling processes for mid-session split adjustments
More efficient bearer anchoring across LTE and NR nodes
This supports scenarios where LTE continues to play a role in multiple RAT setups.
- Corrections & TElx Work
This section encompasses:
Fixes to specifications
Protocol improvements
Enhancements for interoperability
Technical editorial work (TElx)
While these tasks may not seem exciting, they’re crucial for ensuring stable implementations across various vendors.
- Miscellaneous Impacts from RAN1, RAN2, SA, and CT
There are cross-group implications from enhancements led by:
RAN1 (Physical Layer)
RAN2 (L2/RRC Protocols)
SA (System Architecture)
CT (Core Network & Terminals)
These effects necessitate adjustments from RAN3, often resulting in significant TU usage, especially as the complexity across layers increases.
Why RAN3 Release 17 Is Critical for 5G Evolution
RAN3 Release 17 may not add eye-catching PHY features, but it plays an essential role in keeping network architecture:
Scalable
Interoperable
Multi-RAT capable
Prepared for NTN and private enterprise setups
Ready for next-gen services like XR and V2X
This kind of coordination helps ensure that developments in RAN1 and RAN2 can be effectively integrated into live networks.
Conclusion
RAN3 Release 17 is key to enhancing the robustness of 5G architecture. With significant improvements across QoE, NTN, IAB, multicast, NPN, and positioning, these updates set the stage for more capable, adaptable, and interconnected 5G networks.
By syncing with RAN1, RAN2, SA, and CT efforts, RAN3 guarantees a cohesive end-to-end framework that can handle the growing needs of enterprise networks, consumer broadband, IoT, satellite connectivity, and advanced industrial applications.
-
RAN2 Release 17: Key Enhancements, Timeline, and Technical Priorities Explained

RAN2 Release 17: A Comprehensive Guide to 5G Upgrades and Timeline
With 5G networks constantly evolving, the 3GPP Release 17 standards mark a significant step forward in boosting performance, broadening capabilities, and supporting advanced use cases like XR, Industrial IoT, NR Multicast, and Non-Terrestrial Networks (NTN).
The image attached gives you a timeline view of RAN2 Release 17 work items from Q2 2020 to Q2 2021, highlighting how Technical Units (TUs) are distributed across different features. In this blog, we’ll go through these upgrades and showcase how RAN2’s efforts enhance the overall 5G NR ecosystem.
Getting to Know RAN2 in 3GPP
RAN2 mainly zeroes in on Radio Interface Layer 2 (L2) and Radio Resource Control (RRC) protocol specs. This includes:
PDCP (Packet Data Convergence Protocol)
RLC (Radio Link Control)
MAC (Medium Access Control)
RRC signaling
Release 17 builds on the solid groundwork laid by Release 15 (the initial version of 5G NR) and Release 16 (which focused on industrial improvements) by introducing even more flexibility, reliability, efficiency, and global reach.
Key Features of the RAN2 Release 17 Timeline
The timeline featured in the image covers:
2020 Q2 → 2021 Q2
Color-coded work items represent their progression
Technical Units (TUs) illustrate how RAN2 resources are allocated
Here’s a simplified look at how TUs break down, as shown in the image:
Feature / Quarter2020 Q22020 Q32020 Q42020 Q12021 Q2NR Multicast12424NPN (Non-Public Network) Enhancements00.5112Multiradio DC/CA Enhancements0.51212Multi-SIM01212NR for NTN122.513NR Positioning Enhancement14424NR IAB Enhancement01212NR IIoT / URLLC01212Small Data Enhancements01.531.53NR Sidelink Relay024/02/14/1NR XR Study00.5111UE Power Saving Enhancements11212SON/MDT Enhancements01212RAN1/3/SA Impacts149613
Major Enhancements in RAN2 Release 17
Let’s break down the key areas highlighted in the timeline.
NR Multicast: Boosting Broadcast Capabilities
NR Multicast offers:
Efficient delivery of content to multiple users at once
Support for live events, public safety, and V2X
Reduced latency compared to LTE eMBMS
Improved resource utilization
RAN2 emphasizes enhancements to PDCP/RLC for delivering content to groups while ensuring the sessions remain continuous.
Enhancements for Non-Public Networks (NPN)
Release 17 ramps up NPN capabilities for:
Private 5G networks for enterprises
Factory automation
Dedicated campus networks
Upgrades include:
Stronger RRC security updates
Enhanced mobility management for isolated networks
Improvements in Multiradio Dual Connectivity (DC) and Carrier Aggregation (CA)
This aspect boosts:
Multi-radio access combining LTE, NR, and Wi-Fi
CA that enhances uplink reliability and increases throughput
Better mobility management when using dual connectivity
This is vital for devices needing seamless connectivity across various radio technologies.
Optimizations for Multi-SIM functionality
Release 17 refines:
Paging monitoring
DRX operation across several SIMs
Lower power consumption for dual-SIM 5G devices
Support for Non-Terrestrial Networks (NTN)
One of the major highlights of Release 17 is the NTN support for:
LEO/GEO satellites
Coverage in remote areas
Reliable connectivity during emergencies
RAN2’s contributions involve:
NTN-aware behavior for RLC and PDCP
Managing large delays and Doppler shifts
Supporting IoT over NTN, targeting narrowband use cases
Enhancements in NR Positioning
Accurate 5G positioning becomes essential for:
Autonomous vehicles
Industrial robotics
AR/VR applications
Release 17 enhances:
Positioning reference signals
RRC signaling improvements
Support for multi-cell measurements
Integrated Access and Backhaul (IAB) Enhancements
IAB steps up coverage utilizing relay-like nodes. Release 17 reinforces:
Mobility of IAB nodes
Adaptation of backhaul links
Multi-hop capabilities
Advancements in NR IIoT & URLLC
Industrial IoT demands ultra-reliable signaling. RAN2 introduces:
Quicker retransmissions
Improved packet duplication management
More consistent latency handling
Improvements for Small Data and Massive IoT
To enhance energy and bandwidth efficiency for lightweight devices, Release 17 presents:
Reduced signaling overhead
Optimizations in RRC state transitions
Early data transmission for quicker uplink bursts
NR Sidelink Relay
This is a significant advancement for V2X and public safety. It allows devices to relay information over:
L2 (MAC/RLC)
L3 (RRC signaling)
This supports better connectivity even in areas with poor signal or no coverage.
NR XR Study & RAN Slicing
Extended Reality (XR) needs include:
Latency under 10 ms
High data throughput
Control over jitter
RAN2 is looking into protocol changes to ensure quality of service through slicing, which will dedicate resources for XR streams.
Enhancements for UE Power Saving
Managing power consumption is crucial for 5G devices. Release 17 introduces:
Improved DRX cycles
Enhanced wake-up signaling
Better coordination across layers for sleep states
These updates benefit smartphones, wearables, and industrial sensors alike.
SON/MDT Enhancements
Self-Organizing Networks (SON) and Minimization of Drive Tests (MDT) receive:
Improved mobility tracking
Enhanced validation of coverage
More accurate quality of experience indicators
Impacts from RAN1/3 and SA Work
RAN2 adjusts its protocols to be in sync with:
RAN1 improvements at the physical layer
RAN3 interface updates
Enhancements in the SA/CT core network
These collaborative impacts account for a significant segment of TUs.
Why Release 17 is Important for 5G Evolution
Release 17 broadens 5G’s impact, extending beyond high-performance mobile broadband into:
Satellite connectivity
Advanced IoT applications
Real-time XR experiences
Critical industrial automation
Multicast services for wider areas
It fine-tunes capability, efficiency, reliability, and global accessibility.
Conclusion
RAN2 Release 17 is crucial for paving the way for next-gen 5G NR features, showcasing significant strides in NR multicast, positioning, sidelink, IIoT, NTN, and XR support. The provided timeline and TU distribution reveal a well-organized roadmap aimed at boosting 5G’s reach, versatility, and reliability.
As rollout efforts intensify around the globe, Release 17 is set to be an essential stepping stone towards 5G-Advanced and what lies ahead, ensuring that networks are ready for demanding new applications—from industrial automation to satellite-based worldwide coverage.
-
RAN2 Release 17 Explained: Key 5G NR Enhancements, Timeline, and RAN2 Feature Evolution

RAN2 Release 17: An In-Depth Look at Signaling, Mobility, and Architectural Enhancements for 5G NR
3GPP Release 17 marks a major step forward in the evolution of 5G, with RAN2 playing a crucial role in developing control-plane procedures, managing mobility, refining RRC functionalities, and enhancing protocols that shape the performance of NR networks.
The image included outlines RAN2’s work on a quarterly basis starting from 2020 Q2 to 2021 Q2, highlighting key features along with their corresponding TUs (Textual Units) that represent the workload involved.
This article takes that visual data and turns it into a detailed, technically accurate, and easy-to-understand guide. The aim is to help Telecom professionals and tech enthusiasts grasp the significance and implications of RAN2 in Release 17.
What’s the Role of RAN2 in 3GPP?
RAN2 handles:
NR RRC (Radio Resource Control)
Mobility processes
Dual connectivity and CA signaling
Sidelink control
QoS and slicing mechanisms
IAB (Integrated Access and Backhaul) signaling
Non-Access Stratum interactions impacting RAN
With Release 17, we see enhancements in efficiency, flexibility, and capabilities, all geared toward facilitating new 5G use cases and emerging sectors.
Key RAN2 Features in Release 17
Here’s a thorough look at the major features highlighted in the timeline.
- NR Multicast Enhancements
Timeline: 2020 Q2 → 2021 Q2
Multicast gets some significant upgrades to bolster advanced broadcasting for:
Public safety
Live content distribution
Effective group communications
Improvements focus on:
Greater reliability
Synchronized multicast operations
Improved RRC state handling
- Non-Public Network Enhancements (Led by SA2)
Timeline: Q2 2020 → Q1 2021
These updates enhance private 5G networks for companies.
Key features include:
Better mobility between NPN and PLMN
Improved security alignment
RRC optimizations for standalone NPN setups
- Multiradio DC/CA Enhancements
Timeline: 2020 Q2 → 2021 Q2
This supports better:
Dual connectivity (NR-NR, NR-LTE)
Carrier aggregation signaling
Quick and smooth transitions between RATs
This is crucial for multi-band setups and varied networks.
- Multi-SIM Support
Timeline: Q2 2020 → Q1 2021
This feature enhances how devices perform with dual-SIM or multi-SIM setups. Enhancements include:
Coordinated paging
RRC state management
Less interference and missed calls
- NR for NTN (Non-Terrestrial Networks)
Timeline: Q2 2020 → Q1 2021, followed by IoT for NTN in 2021 Q2.
Satellite-based NR extends 5G coverage to remote areas. Target improvements include:
Mobility under high Doppler shifts
Timing and synchronization challenges
Satellite-specific RRC procedures
- NR Positioning Enhancements
Timeline: 2020 Q2 → 2021 Q1
Crucial for applications like:
Industrial robotics
Autonomous vehicles
Indoor navigation
RAN2 enhances positioning through:
Better measurement reporting
RRC assistance info
Multi-TRP positioning processes
- NR IAB Enhancements
Timeline: 2020 Q3 → 2021 Q1
Integrated Access and Backhaul (IAB) sees improvements through:
Enhanced mobility for parent/child nodes
Sturdier backhaul links
Streamlined RRC configurations
This is vital for dense urban mmWave setups.
- NR IIoT/URLLC Enhancements
Timeline: 2020 Q3 → 2021 Q2
Strengthening NR performance for industrial IoT and ultra-reliable low latency applications involves:
Quicker control-plane procedures
Reliable performance
Consistent QoS enforcement
- Enhancements for Small Data Transmission
Timeline: 2020 Q3 → 2021 Q2
This is optimized for devices that only need to send small, sporadic packets. Improvements include:
RRC reductions
Lighter signaling
Faster access and release
- NR Sidelink Relay (Based on SI outcome)
Timeline:
Study: 2020 Q2–Q3
Specification: 2020 Q4 → 2021 Q2
Supports extending network coverage with UE relays for:
Disaster recovery
Rural connectivity
V2X situations
- NR XR Study and RAN Slicing (Based on SI Outcome)
Timeline: 2020 Q2 → 2021 Q2
XR enhancements provide high throughput and low jitter for AR/VR/MR applications. RAN slicing allows for flexible virtual networks tailored to specific use cases.
Key improvements include:
Refined QoS flows
Slice-aware mobility processes
XR-focused QoE mechanisms
- UE Power Saving Enhancements
Timeline: 2020 Q3 → 2021 Q2
Focusing on extending battery life through:
Improved DRX patterns
Lightweight signaling
Better paging cycles
This is critical for both IoT devices and smartphones.
- SON/MDT Enhancements (Led by RAN3)
Timeline: 2020 Q3 → 2021 Q2
Smart network optimization and Minimization of Drive Tests (MDT) see improvements through:
More accurate reporting
Efficient setups
Support for advanced AI-driven optimization
- Cross-RAN Impacts (Led by RAN1/3/4, etc.)
Includes changes linked to:
Physical layer constraints
Spectrum/antenna specifications
Core network dependencies
These alterations ensure alignment across the entire architecture.
RAN2 Release 17 TUs (Workload Allocation)
The image includes a TU table showing quarterly workloads:
QuarterTotal TUsNotes2020 Q29Initial studies start2020 Q324Growing feature set2020 Q452Peak workload as specs develop2021 Q130Stabilization and refining2021 Q260Finalization, high activity
The sharp spike by 2021 Q2 highlights the finalization efforts across various parallel features.
RAN2 Work Structure in Release 17 (Based on Image)
Major Mobility & Control Enhancements
Multi-SIM
NR positioning
IAB improvements
RAN slicing
Coverage & Connectivity Extensions
NR sidelink relay
NTN satellite systems
Advancements in private networks
Signaling Efficiency Improvements
Small data transmission
Power-saving optimizations
Strengthening IIoT/URLLC control-plane
Core Support for New Use Cases
XR signaling enhancements
Enhanced multicast
Industrial IoT advancements
How RAN2 Release 17 Boosts 5G
Release 17’s RAN2 improvements significantly enhance 5G capabilities in several areas:
- Scalability
RAN2 introduces optimizations that help 5G networks scale effectively across private networks, satellite connections, and dense multi-band setups.
- Reliability for Key Industries
Enhancements in URLLC, positioning, and IAB allow for more advanced industrial automation, robotics, and transport systems.
- Extended Coverage
Sidelink relay and NTN take 5G’s reach far beyond traditional limits.
- User Experience
Better XR support, power savings, and small data optimizations lead to smoother and more efficient device operation.
- Network Automation
SON/MDT enhancements decrease manual work and set the stage for AI-driven RAN.
Conclusion
RAN2 Release 17 brings a wealth of enhancements that boost the 5G control plane, enhance coverage, improve industrial capabilities, and get networks ready for sophisticated future applications. The work noted in the timeline illustrates how RAN2 gradually built support for XR, NTN, private networks, sidelink relay, advanced mobility, and optimized signaling from mid-2020 to mid-2021.
For those in telecom, getting a handle on RAN2’s progress is key for planning deployments, designing devices, and aligning future architecture strategies. Release 17 doesn’t just push current 5G operations forward; it lays the groundwork for 5G-Advanced (Release 18 and beyond), where intelligence, flexibility, and global reach will shape the next generation of networks.
-
RAN1 Release 17 Explained: Key 5G Enhancements, Timeline, and Technical Upgrades

RAN1 Release 17: A Detailed Look at 5G NR Updates and Timeline
5G is evolving quickly, and 3GPP Release 17 is one of the most significant updates for the physical layer, particularly for RAN1, the group that handles 5G NR Layer-1 specifications.
The image included gives a clear view of how RAN1’s Release 17 features developed from Q1 2020 to Q1 2021 and the Textual Units (TUs) assigned each quarter. These features include performance improvements, coverage upgrades, new use cases like XR, and extending 5G NR into non-terrestrial networks (NTN).
In this blog post, we’ll break down each feature category and explain the timeline in a way that’s easy to understand for both Telecom pros and tech fans.
What Is RAN1 Release 17?
RAN1 in 3GPP focuses on the physical layer of 5G NR, dealing with things like modulation, coding, MIMO, beams, sidelink, channel models, and more. Release 17 builds on what was done in Rel-15 and Rel-16, enhancing performance, expanding coverage, and opening up new areas like industrial IoT, XR, and satellite NR (NTN).
The timeline shows when every feature was analyzed, designed, and finalized, along with the TUs assigned each quarter.
Key Features of RAN1 Release 17
Here’s a rundown of the main RAN1 Release 17 work items shown in the timeline.
- NR MIMO Enhancements
Timeline: Q1 2020 → Q1 2021
These enhancements focus on improving advanced beamforming, massive MIMO efficiency, mobility robustness, and CSI feedback.
Why it matters:
Better spectral efficiency
Stronger performance at cell edges
Optimized multi-user scenarios
- NR Sidelink Enhancements
Timeline: Q2 2020 → Q1 2021
Release 17 boosts sidelink capabilities for:
V2X improvements
Device-to-device communication
Public safety messaging
Improvements include:
Higher reliability
Better synchronization
Low-latency transmission modes
- Extending NR Operation to 71 GHz
Timeline: Q2–Q3 2020
This focuses on extending FR2 from 52.6 GHz to 71 GHz, studying how they can coexist and the feasibility of waveforms.
Importance:
Unlocks more mmWave spectrum
Supports ultra-high capacity hotspots
Enables next-gen indoor/outdoor setups
- Dynamic Spectrum Sharing (DSS) Enhancements
Timeline: Q3–Q4 2020
Improvements aim at better coexistence between LTE and NR in shared spectrum.
Key upgrades:
Enhanced scheduling efficiency
Reduced overhead
Lower latency for NR users in DSS bands
- NR Industrial IoT / URLLC Enhancements
Timeline: Q2 2020 → Q1 2021
These refinements target ultra-low latency and reliability.
Focus areas:
Smart factories
Robotics
Precision automation
- IoT over NTN (Non-Terrestrial Networks)
Timeline: Starts Q3 2020 → Q1 2021
NR NTN introduces 5G for satellite connectivity.
Applications include:
Maritime
Aviation
Rural areas
Emergency coverage in remote locations
- NR Positioning Enhancements
Timeline: Q1 2020 → Q1 2021
Accurate positioning is crucial for autonomous vehicles and industrial automation.
Enhancements include:
Greater accuracy (sub-meter level)
Faster time-to-first-fix
Better mobility tracking
- Low-Complexity NR Devices
Timeline: Q1 2020 → Q1 2021
These are designed to be cost-effective and use less power.
Good for:
Wearables
Low-end IoT sensors
Battery-constrained devices
Release 17 brings reduced bandwidth and lighter processing requirements.
- Power Saving Improvements
Timeline: Q2–Q3 2020
This includes new DRX cycles, optimized wake-up signals, and energy-efficient scheduling.
Benefits:
Longer battery life
More efficient IoT operations
Lower overall network power use
- NR Coverage Enhancements
Timeline: Q1 2020 → Q1 2021
This targets areas like:
Indoor 5G
Rural macro deployments
Deep indoor coverage
Techniques used include:
Repetitions
Increased beamforming gain
Improved channel estimation
- XR Study (Extended Reality)
Timeline: Q3 2020–Q1 2021
Release 17 adapts the physical layer for XR applications.
Goals:
Stable and low latency (under 20 ms)
High throughput
Consistent QoS for AR/VR/MR
- NR IoT / eMTC Enhancements
Timeline: Q1 2020 → Q1 2021
This focuses on the coexistence of LTE-M / NB-IoT with NR and aims to boost performance for massive IoT.
- Impacts from RAN2/3/4
RAN1 also takes into account dependencies from higher layers, including:
Integrated Access and Backhaul (IAB)
Multicast
RAN4 spectrum/antenna constraints
- Reserved for TEI17
This is a late-stage buffer for additional study/work items.
RAN1 Release 17 TUs (Workload Allocation)
The right side of the image shows the quarterly TUs assigned for Release 17 work. These TUs represent the workload for each item.
QuarterTotal TUsNotes2020 Q115Initial setup of studies2020 Q242Peak load as studies mature2020 Q325Transition from study to specification work2020 Q452Heavy standardization work2021 Q127Finalization and freeze preparation
The workload demonstrates the complexity and range of Release 17 features.
Structured Overview of Release 17 Features
Major Physical Layer Enhancements
NR MIMO improvements
DSS refinement
Better FR2 spectrum utilization
Coverage and positioning boosts
Device & IoT-Level Enhancements
Low complexity NR devices
Power saving
Industrial IoT & URLLC upgrades
Evolution of eMTC & NR IoT
New Use Case Support
XR optimization
NTN-based IoT
Advanced sidelink
Why Release 17 Matters for the Future of 5G
Release 17 lays the groundwork for future 5G-Advanced developments (Release 18 and beyond). It allows for:
A broader 5G device ecosystem
Better reliability for mission-critical services
Worldwide satellite-integrated 5G coverage
Immersive XR experiences over mobile networks
Efficient IoT scaling with optimized designs
This isn’t just a small update—it’s a strategic growth of 5G’s capabilities.
Conclusion
RAN1 Release 17 is a significant advancement for 5G NR, enhancing performance and pushing the limits of cellular systems. From MIMO and sidelink improvements to innovative NTN support and XR optimization, Release 17 paves the way for 5G-Advanced.
For Telecom experts and tech enthusiasts alike, this release marks the start of a more versatile, efficient, and widely accessible 5G ecosystem. Grasping the timeline and range of these upgrades will help prepare for future deployments, product innovations, and network developments in the years to come.
-
3GPP RAN Timeline Explained: Release 16, Release 17, and the Road to Release 18

3GPP RAN Timeline: Tracking the Growth of Release 16, Release 17, and Release 18
The development of mobile communication standards is shaped by detailed timelines established by 3GPP through its Radio Access Network (RAN) working groups. The image included here shows an Overall RAN Timeline that runs from late 2019 to 2022, marking important freeze points, package approvals, and performance completions for 3GPP Releases 16, 17, and the initial phase of Release 18.
For telecom professionals, these timelines are crucial for planning deployments, developing chipsets, optimizing networks, and creating long-term technology strategies. This article breaks down each key milestone and presents the information in a straightforward, accessible way.
What the RAN Timeline Represents
The RAN timeline depicted in the image details activities from:
RAN1: Focused on physical layer specifications
RAN2: Concerned with the radio interface layer 2 and RRC
RAN3: Concentrating on architecture and X2/F1 interfaces
RAN4: Dealing with RF and performance requirements
Each release moves through distinct technical phases:
RAN1 Freeze
RAN2/3/4 Stage-3 Freeze (Core specifications)
ASN.1 Freeze
Performance Completion
Package Approval
The timeline visually illustrates how these phases overlap and affect one another across the releases.
Release 16 Timeline Breakdown
Release 16 is the second big step in 5G since Release 15, refining 5G NR, enhancing URLLC, improving industrial IoT, and expanding V2X capabilities.
- RAN1 Freeze – Q1 2020
The image highlights the “Rel-16 RAN1 Freeze” occurring early in 2020. This freeze finalizes all physical layer features like:
Slot structures
MIMO schemes
NR enhancements
Latency reduction features
Once RAN1 freezes, the PHY layer features are set, allowing chipset makers to kick off silicon development.
- Stage-3 Freeze – Q2 2020
This stage marks the completion of:
RRC protocols (RAN2)
Xn/NG interface specifications (RAN3)
Radio resource control behaviors
With the Stage-3 freeze, the protocol behaviors stabilize enough for vendors to start working on compatible software.
- ASN.1 Freeze – Mid-2020
The ASN.1 freeze locks in message structures for:
RRC
NGAP
XnAP
This ensures the interface schemas remain unchanged, allowing for:
Vendor interoperability
Conformance testing
Stable deployments
- RAN4 Performance Completion – Q3 2020
RAN4 sets the groundwork for:
RF requirements
Sensitivity thresholds
Adjacent channel leakage ratios
Power class definitions
The performance completion stage guarantees hardware vendors have the final RF parameters they need for device certification.
- Release 16 Completion
The image wraps up with the “Release 16 Completion” section, indicating that Rel-16 officially concluded around Q3/Q4 2020.
Release 17 Timeline Breakdown
Release 17 pushes 5G further, adding improvements for:
RedCap (reduced capability devices)
Satellite-to-NR (NTN)
Enhanced positioning accuracy
QoE improvements
Better network slicing capabilities
Its timeline extends from early 2020 into 2022, reflecting delays due to pandemic-driven remote meetings.
Package Approvals in 2020
Two major package approvals are noted in the image:
Rel-17 RAN1/2/3 Package Approval – Early 2020
Rel-17 RAN4 Package Approval – Q3 2020
These approvals allow the working groups to start drafting technical features officially.
15-Month Release Length
A long horizontal bracket in the image shows a 15-month release duration, emphasizing that delays pushed Rel-17 past its initial schedule.
RAN1 Freeze – Q1 2021
RAN1 is the first to freeze since PHY features need to be stable early in the process. This includes:
NR-Light / RedCap
MIMO improvements
Energy-efficiency features
Sidelink enhancements
Stage-3 Freeze – Q2 2021
At this point, specifications for:
RAN2
RAN3
RAN4 and the core network freeze at the same time.
This synchronicity helps ensure consistent cross-layer behavior in NR features.
ASN.1 Freeze for RAN2/3 – Q3/Q4 2021
The image points to the “Rel-17 ASN.1 Freeze (RAN 2/3)” in late 2021.
This step finalizes all signaling message structures and wraps up the 5G NR protocol schemas.
RAN4 Performance Completion – Early 2022
This milestone confirms that RF parameters for all Release-17 features are now stable, which is vital for:
Device manufacturers
Small cell and base station vendors
Certification bodies
After this performance completion, vendors can confidently start the commercialization process.
Release 18 – Early Activities in 2021
Release 18 kicks off 5G-Advanced, focusing on:
AI-driven network optimization
Sidelink upgrades for XR
Spectrum efficiency tweaks
Improved NTN-NR integration
The image also notes the:
Rel-18 RAN1/2/3 Package Approval (TBC) – Q2 2021
The “TBC” label indicates this approval was on the docket but awaiting confirmation when the timeline was created.
Work on Release 18 begins even while Release 17 is being wrapped up—this overlap is typical and necessary for ongoing industry advancement.
Summary Table of Timelines
Release Milestone Timeline Meaning Rel-16RAN1 Freeze Q1 2020PHY features locked Stage-3 Freeze Q2 2020 Protocol behavior setASN.1 Freeze Mid-2020 Messaging locked RAN4 Completion Q3 2020RF finalized Rel-17RAN1 Freeze Q1 2021NR-Light, enhancements fixed Stage-3 Freeze Q2 2021RAN2/3/4 + Core freeze ASN.1 Freeze Q3/Q4 2021 Schemas finalized RAN4 Completion Q1 2022 RF finalized Rel-18RAN1/2/3 Package Approval 2021 (TBC)5G-Advanced kickoff
Why These Timelines Matter
For Network Vendors
These timelines guide chipset development, help in choosing PHY layer features, and influence hardware design cycles.
For Operators
They signal when specific features (like NR-Light and improved slicing) can be expected to roll out in commercial networks.
For Standardization Experts
They clarify the interdependencies between working groups and illustrate how releases evolve side by side.
For Researchers and Academics
Getting a grip on these freeze points is useful for aligning research efforts with future standards.
Conclusion
The uploaded Overall RAN Timeline gives a detailed overview of how 3GPP structures and delivers 5G advancements across Releases 16, 17, and 18. Every milestone—RAN1 freezes, Stage-3 freezes, ASN.1 finalizations, and performance completions—marks vital technical progress shaping the global 5G landscape.
Release 16 laid the groundwork, Release 17 expanded 5G’s capabilities into IoT, NTN, and positioning, and Release 18 heralds the start of the 5G-Advanced era. Grasping this timeline is key for professionals to anticipate when technologies will be ready, plan deployments effectively, and align innovations with global standards.
-
3GPP Release-17 Timeline Explained: Key Milestones, Freeze Dates, and 5G-Advanced Evolution

3GPP Release-17 Timeline: Key Milestones Shaping the Future of 5G and Beyond
As 5G continues to mature, the 3GPP’s Release-17 (Rel-17) stands out as a key milestone on the road to 5G-Advanced and eventually 6G. The timeline shared here outlines how the 3GPP mapped out the multi-year development cycle for Release-17 from 2020 to 2023. It points out freeze points, numbers for technical specification group (TSG) meetings, and approval events that help clarify how the global mobile industry agrees on architecture, features, interfaces, and protocol progress.
This blog aims to break down each milestone in a clear, tech-friendly way, making it easier for engineers, planners, and enthusiasts to see where Rel-17 fits within the larger 5G picture.
Understanding 3GPP and Release-17
3GPP (3rd Generation Partnership Project) brings together global bodies focused on telecom standards to create specifications for 3G, 4G, 5G, 5G-Advanced, and the future 6G. Each “Release” represents a bundle of standards packed with features, enhancements, and architectural upgrades.
Release-17 is especially significant because it:
Expands the abilities of 5G NR (New Radio)
Boosts support for Massive IoT and RedCap devices
Improves accuracy for positioning
Enhances network slicing and private network capabilities
Sets the groundwork for 5G-Advanced (Release-18)
The timeline image captures the organized journey toward completing this release.
Timeline Overview: 2020 to 2023
The timeline stretches from Q4 2020 (TSG#90-e) to Q1 2023 (TSG#99). The “e” next to the meeting numbers means electronic (virtual) meetings, reflecting the global changes during the pandemic.
There are three major freeze milestones worth noting:
Stage-2 Freeze
Stage-3 Freeze
Protocol Coding Freeze (ASN.1, OpenAPI)
Also, the Release-18 package approval comes up mid-timeline, marking the start of work on 5G-Advanced.
Release-17 Timeline Breakdown
Here’s a breakdown of the timeline based on the image shown.
2020 – Foundation and Initial Study Phase (TSG#90-e)
In Q4 2020, during meeting TSG#90-e, the focus was on discussing preliminary study items and work for Release-17. This phase involved things like:
Defining the project’s scope
Spotting industry needs
Divvying up contributions
Estimating what’s needed for features
Switching to virtual meetings caused some scheduling hiccups, leading 3GPP to tweak target dates.
2021 – Active Specification Work and Stage-2 Freeze
Q1 2021 – TSG#91-e
Work carried on in virtual settings, centering on feature definitions for NR, system architecture (SA), and RAN improvements.
Q2 2021 – TSG#92-e
Specifications started to take shape. This is when technical reports began evolving into complete specs.
Key Milestones Table from the Image
Milestone Timing (from image) – Meaning
Stage-2 Freeze: Mid-2021 – Architecture and logical procedures locked in.
Stage-3 Freeze: Early 2022 – Protocol specs finalized.
ASN.1 & OpenAPI Coding Freeze: Q2 2022 – Interface schemas finalized for smooth interoperability.
Release-18 Package Approval: Mid-2022 – Official kickoff for 5G-Advanced.
Finalization Meetings: Late 2022 to Early 2023 – Editorial tweaks and wrap-up.
Why Release-17 Matters for 5G Evolution
Release-17 really boosts what 5G networks can do. Here are some key areas it covers:
- Enhanced mMTC and RedCap Devices
Lightweight IoT devices will see benefits from lower complexity and better battery life.
- Advanced Positioning
With improved accuracy, this will help industrial robotics, AR/VR, and autonomous systems.
- Extended NR Operation
This includes support for:
Spectrum above 52.6 GHz
More robust mmWave enhancements
- Strengthened Private Networks
Improvements in slicing, resource partitioning, and network capability exposure.
- Expanded Non-Terrestrial Networks
Connecting satellites to devices gets a boost in standardization.
- URLLC and Industrial Enhancements
You’ll get more reliable connections for mission-critical automation tasks.
Conclusion
The uploaded 3GPP Release-17 timeline gives us a great overview of how the global telecom community has worked together to shape the next phase of 5G. From the Stage-2 and Stage-3 freezes to the finalization of coding and the approval of Release-18, each milestone shows the dedication of countless engineers and organizations over the years. Grasping this timeline helps professionals understand the complexities of standardization and gears up the industry for rolling out and optimizing the complete set of Release-17 features—setting us up for 5G-Advanced and ultimately, 6G.
-
PDCP Duplication in 5G: How Carrier Aggregation and Dual Connectivity Boost Reliability

Understanding PDCP Duplication: Carrier Aggregation and Dual Connectivity Explained
As 5G networks expand to support critical applications like self-driving cars, industrial automation, remote surgeries, and widespread sensor networks, ultra-reliable performance is just as key as high data rates. One crucial technique that ensures this kind of reliability is PDCP duplication, which is a key feature in both Carrier Aggregation (CA) and Dual Connectivity (DC).
The diagram we’ve included shows how PDCP packets are duplicated and sent over various radio connections, helping to cut down latency, enhance reliability, and protect users during radio link failures.
In this post, we’ll dive into the architecture, the mechanisms behind it, scheduling aspects, and how PDCP duplication is applied in both CA and DC scenarios.
What Is PDCP Duplication?
PDCP (Packet Data Convergence Protocol) duplication is when the same PDCP Protocol Data Unit (PDU) is sent through two or more different radio links. The user equipment (UE) picks up these duplicates and uses PDCP sequence numbering to filter out the repeats.
Why Duplication Is So Effective
The chance that both links fail at once is really low.
Sending duplicate data helps reduce packet loss and boosts reliability.
Upload (UL) and download (DL) performance is sturdier in poor radio conditions.
It keeps latency stable, which is crucial for time-sensitive traffic.
That’s why PDCP duplication is vital for URLLC (Ultra-Reliable Low Latency Communication) and critical IoT applications.
PDCP Duplication with Carrier Aggregation (CA)
Carrier Aggregation lets the UE connect to multiple component carriers from one gNB. In the illustration (on the left), the CA setup includes:
One PDCP-NR entity that creates two PDCP PDUs: * One original packet * One duplicate packet
Two RLC entities (Entity 1 & Entity 2)
A MAC entity that manages scheduling and multiplexing
Two PHY layers and two component carriers (CC1 and CC2)
How It Works
PDCP-NR generates a PDU.
It also creates a duplicate PDCP PDU (highlighted in red).
The two PDUs are forwarded to RLC Entity 1 and RLC Entity 2 independently.
MAC schedules each PDU on separate carriers—CC1 and CC2.
The UE gets both PDUs over 5G NR.
PDCP at the UE filters out the duplicate using sequence number matching.
Benefits of PDCP Duplication via CA
Very low variation in latency (perfect for real-time industrial uses)
Stronger resistance to issues if one carrier’s link degrades
Better coverage reliability in CA-wide setups
Shorter retransmission delays since at least one PDU will probably be successful
CA-based duplication works best when the UE stays connected to the same gNB and can utilize a wide frequency range.
PDCP Duplication with Dual Connectivity (DC)
Dual Connectivity takes things a step further, allowing the UE to connect to two different base stations:
A Master Node (MN)—usually LTE or NR
A Secondary Node (SN)—usually NR
In the image (on the right), the DC architecture consists of:
A PDCP-NR instance producing duplicate packets
One bearer through the LTE MN
One bearer through the NR SN
An X2-U interface that carries the duplicate copy
An LTE stack (RLC-LTE, MAC-LTE, PHY-LTE)
An NR stack (RLC-NR, MAC-NR, PHY-NR)
Flow of DC-Based PDCP Duplication
The PDCP-NR entity gets packets from the S-GW.
It produces both the original and duplicate PDUs.
One copy goes through the LTE Master Node.
The other is routed through the NR Secondary Node.
Both LTE and NR deliver the packets separately to the UE.
PDCP at the UE merges and removes duplicates.
Unique Advantages of Dual Connectivity
The redundancy covers different RATs (LTE + NR), meaning: * Different frequencies * Different base stations * Different schedulers
This significantly lowers the chances of linked failures occurring together.
It’s perfect for mobility situations where one node might experience temporary issues.
It’s crucial for the initial phases of 5G deployments, where LTE still provides essential coverage.
DC-based duplication is favored for high-mobility, coverage-challenged, or mission-critical scenarios.
PDCP Duplication vs. RLC Retransmissions
So, why not just use HARQ & RLC?
Well, relying on those mechanisms brings about:
Retransmission delays
Increased latency variability
Higher signaling overhead when things get congested
PDCP duplication sidesteps these problems by allowing immediate parallel transmission—no need for retransmissions.
When Is PDCP Duplication Activated?
3GPP outlines different situations where PDCP duplication kicks in:
Activated by:
The necessity for high reliability
Congestion in the RLC buffer
Poor SINR conditions
Activation of the URLLC profile
Handover preparation or when on the move
Deactivated When:
There’s only one radio link available
The UE’s battery condition is low
The type of bearer doesn’t call for high reliability
Typically, PDCP duplication is dynamic and policy-controlled by the network.
Which Use Cases Gain the Most?
Critical Industrial IoT
Robot control
Motion control loops
Safety mechanisms
Autonomous Vehicles
V2X safety messages
Cooperative maneuvers
Collision prevention
Healthcare
Remote surgeries
Remote haptics
Public Safety Networks
Mission-critical push-to-talk
Real-time video feeds
Emergency drone data
Reliable Consumer Services
Cloud gaming
Augmented/Virtual Reality
Managing crowd situations at stadiums
Comparison Table: CA vs DC in PDCP Duplication
FeatureCarrier AggregationDual ConnectivityNumber of Base Stations1 (same gNB)2 (MN + SN)Redundancy TypeFrequency-levelNode-level + RAT-levelLatency ImpactLow and consistentSlight variation between RATsReliabilityHighExtremely highIdeal ForStationary / low mobilityHigh mobility, coverage gaps
Conclusion
PDCP duplication stands out as one of the key features in the 5G toolkit for achieving ultra-reliable, low-latency communication. By sending duplicate packets across multiple carriers (CA) or multiple nodes and RATs (DC), networks can drastically minimize packet loss, stabilize latency, and provide the resilience needed for critical services.
As we move forward into the next phases of 5G and eventually 6G systems, PDCP duplication will still be a fundamental feature for things like industrial automation, autonomous technologies, public safety, and any application where we can’t afford to lose even a single packet.
-
Multi-Radio Dual Connectivity (MR-DC) Explained: How LTE & 5G Work Together for Higher Performance

MAC Carrier Aggregation: A Comprehensive Technical Guide for 5G Networks
Carrier Aggregation (CA) stands out as one of the most valuable assets in today’s 5G and LTE-Advanced networks. It enables operators to merge several frequency carriers—whether they’re in the same band or different ones—leading to much higher data speeds, better coverage, and more efficient use of spectrum.
The image included gives a clear visual of MAC Carrier Aggregation, showcasing how the protocol layers work together, how the MAC layer manages multiple carriers, and how devices connect to various cells at the same time. In this blog post, we’ll take a deep dive into the architecture and break down how MAC-level carrier aggregation operates within 5G NR.
Understanding the Architecture in the Image
The diagram highlights several key components working in harmony within a 5G RAN:
AMF and RRC (control plane)
UPF and SDAP (user plane)
PDCP, RLC, and MAC (radio protocol stack)
PHY layer managing several aggregated carriers
Multiple cells (Cell 1, Cell 2, Cell 3, Cell 4…)
User devices accessing multiple carriers at once
This setup is crucial for how 5G achieves faster data rates and lower latency by aggregating radio resources.
What Is MAC Carrier Aggregation?
MAC Carrier Aggregation involves the Medium Access Control (MAC) layer’s capability to orchestrate and manage data transmission across various component carriers (CCs).
These carriers can be:
Intra-band contiguous
Intra-band non-contiguous
Inter-band carriers from different frequency bands
With CA, you can combine up to 16 component carriers in 5G, allowing for multi-gigabit downlink speeds.
The MAC layer serves as the central hub, managing:
Resource scheduling
HARQ processes
Logical channel prioritization
Multi-carrier buffer management
Meanwhile, the PHY layer takes care of physical modulation, coding, and transmission across each carrier.
Protocol Stack Overview in the Diagram
- Control Plane: AMF → RRC
The Access and Mobility Management Function (AMF) interacts with the RRC (Radio Resource Control) layer. RRC is in charge of:
Setting up connections
Configuring carriers
Managing mobility and handovers
Activating/deactivating CA
- User Plane: UPF → SDAP
The User Plane Function (UPF) connects with the SDAP (Service Data Adaptation Protocol), which aligns QoS flows with radio bearers. CA enables SDAP to utilize multiple carriers to fulfill QoS needs (e.g., URLLC, eMBB).
- PDCP → RLC → MAC Stack
These layers handle standard 5G NR functions:
PDCP: compression, ciphering
RLC: segmentation, retransmission
MAC: multiplexing, scheduling, and carrier selection
PHY: modulation, coding, and transmission for each cell
The MAC layer is crucial to Carrier Aggregation, directing how data moves across the combined carriers.
How MAC Carrier Aggregation Works
- MAC Takes Data From RLC
The MAC layer gets RLC PDUs and determines:
Which carrier to utilize
The appropriate scheduling grant size
How to manage HARQ processes for each carrier
- Scheduling Across Multiple Component Carriers
Every carrier has its own:
Bandwidth
Frequency
Propagation conditions
Cell ID
The MAC scheduler has to optimize resource allocation across all of these at once.
- PHY Sends the Data Through Multiple Cells
The PHY layer is responsible for physical transmissions. The image illustrates:
Cell 1
Cell 2
Cell 3
Cell 4
These cells can serve the same device using various carriers. Devices receive multiple downlink signals, which you can see represented by different colored arrows.
- The UE Combines the Data
The device’s PHY and MAC layers work together to merge the aggregated data streams into one high-rate data flow.
Why Carrier Aggregation Matters in 5G
- Significantly Higher Throughput
Combining carriers allows operators to offer a broader effective bandwidth:
Number of CCsTotal BandwidthThroughput Impact1 CC20 MHzBaseline3 CCs60 MHz~3× data rates16 CCs (5G NR)640 MHzMulti-Gbps speeds
- Better Spectrum Utilization
Operators often lack large blocks of contiguous spectrum. CA addresses this by combining fragmented pieces from:
Mid-band
Low-band
mmWave
- Improved Coverage and Reliability
Low-band carriers enhance coverage, while high-band carriers offer higher speeds. CA allows for the simultaneous use of both.
- Enhanced User Experience
Users enjoy:
Faster downloads
Higher streaming quality
More stable connections
Better performance at the edges of cells
CA Types Explained
- Intra-band Contiguous CA
Carriers are adjacent in frequency.
Easiest to put into practice.
- Intra-band Non-Contiguous CA
Carriers are in the same band but have gaps between them.
Requires more complicated RF filtering.
- Inter-band CA
Carriers come from different frequency bands.
This is the most common configuration in 5G (e.g., 700 MHz + 3.5 GHz + 28 GHz).
The diagram hints at a multi-cell inter-band scenario.
Multi-Cell Carrier Aggregation (as shown in the image)
The image goes beyond typical CA. It showcases multi-cell Carrier Aggregation, where carriers come from not just different frequencies but also from different cells.
This supports:
Greater network capacity
Flexible load balancing
Multi-cell connectivity before full dual connectivity (DC) is necessary
Users maintain connections to several cells at once, which enhances:
Resilience
Handover performance
Throughput aggregation across sectors or gNodeBs
MAC Layer Challenges in Multi-Cell CA
Managing multiple carriers across cells calls for:
- Advanced Scheduling Algorithms
MAC needs to take into account:
Load distribution
Carrier quality
Interference
Hardware constraints
- Separate HARQ Processes Per Carrier
Each carrier requires its own count and timing for HARQ processes.
- Cross-carrier Scheduling
Scheduling grants for one carrier might affect another.
- Increased UE Complexity
Devices need to support:
Multiple RF chains
Various antenna paths
Multi-carrier demodulation capability
Benefits of MAC Carrier Aggregation in 5G Networks
✔ Enhanced spectral efficiency
✔ Higher data rates (multi-Gbps)
✔ Flexible spectrum use across fragmented bands
✔ More stable connectivity for users at cell edges
✔ Improved QoS support for eMBB, URLLC, and mMTC
✔ Better network load balancing
✔ Superior mobility performance
Conclusion
MAC Carrier Aggregation is a fundamental technology propelling the high-performance features of 5G networks. By effectively managing multiple carriers at the MAC layer and transmitting them through the PHY layer across different cells, CA significantly boosts network throughput, spectrum efficiency, and user satisfaction. The image illustrates how various layers—RRC, SDAP, PDCP, RLC, MAC, and PHY—collaborate to enable seamless multi-carrier communication across multiple cells.
As we progress towards 5G-Advanced and eventually 6G, carrier aggregation will keep evolving, accommodating more component carriers, wider bandwidths, and increasingly varied spectrum bands. For those in telecom and technology, getting a handle on MAC Carrier Aggregation is crucial for mastering the future of wireless communication.
-
MAC Carrier Aggregation Explained: How 5G Combines Multiple Cells for Higher Throughput

MAC Carrier Aggregation: A Comprehensive Technical Guide for 5G Networks
Carrier Aggregation (CA) really stands out as a key feature in today’s 5G and LTE-Advanced networks. It allows network operators to merge multiple frequency carriers—whether they’re in the same band or different ones—to boost throughput, enhance coverage, and make better use of available spectrum.
The image uploaded gives a clear visual of MAC Carrier Aggregation, illustrating how different protocol layers interact, how the MAC layer manages multiple carriers, and how devices connect to various cells at the same time. This blog post delves into the architecture and breaks down how MAC-level carrier aggregation operates within 5G NR.
Grasping the Architecture in the Image
The diagram showcases several essential components working in tandem within a 5G Radio Access Network (RAN):
AMF and RRC for the control plane
UPF and SDAP for the user plane
PDCP, RLC, and MAC making up the radio protocol stack
The PHY layer that handles multiple combined carriers
Multiple cells (Cell 1, Cell 2, Cell 3, Cell 4, etc.)
User devices connecting to various carriers all at once
This setup is crucial for how 5G achieves faster data rates and lower latency through aggregated radio resources.
So, What Is MAC Carrier Aggregation?
MAC Carrier Aggregation is all about the Medium Access Control (MAC) layer’s ability to manage and schedule data transmission across multiple component carriers (CCs). These carriers can be:
Intra-band contiguous
Intra-band non-contiguous
Inter-band carriers coming from different frequency bands
In 5G, CA can merge up to 16 component carriers, enabling multi-gigabit downlink speeds.
The MAC layer serves as the main hub for handling:
Resource scheduling
HARQ processes
Prioritizing logical channels
Managing buffers for multiple carriers
The PHY layer’s job is to provide physical modulation, coding, and transmission through each carrier.
A Look at the Protocol Stack in the Diagram
- Control Plane: AMF → RRC
The Access and Mobility Management Function (AMF) connects with the RRC (Radio Resource Control) layer, which oversees:
Setting up connections
Configuring carriers
Mobility and handovers
Activating or deactivating CA
- User Plane: UPF → SDAP
The User Plane Function (UPF) links to SDAP (Service Data Adaptation Protocol), matching QoS flows to radio bearers. CA enables SDAP to utilize several carriers to fulfill QoS needs (like URLLC, eMBB).
- PDCP → RLC → MAC Stack
These layers cover standard 5G NR functionalities:
PDCP: handles compression and ciphering
RLC: manages segmentation and retransmissions
MAC: is in charge of multiplexing, scheduling, and selecting carriers
PHY: deals with modulation, coding, and data transmission over cells
The MAC layer is essentially the backbone of Carrier Aggregation, directing how data moves across the combined carriers.
How MAC Carrier Aggregation Operates
- MAC Receives Data From RLC
The MAC layer gets the RLC Protocol Data Units (PDUs) and determines:
Which carrier to utilize
The size of the scheduling grant to assign
How to handle HARQ processes for each carrier
- Scheduling Across Multiple Component Carriers
Every carrier comes with its own:
Bandwidth
Frequency
Propagation conditions
Cell ID
The MAC scheduler needs to optimize resource distribution across all these aspects simultaneously.
- PHY Sends Data Through Multiple Cells
The PHY layer takes care of physical transmissions. The image outlines:
Cell 1
Cell 2
Cell 3
Cell 4
Each of these cells can cater to the same device using different carriers, with multiple downlink signals illustrated by various colored arrows.
- The UE Combines the Data
The device’s PHY and MAC layers piece together the aggregated data streams into a single, high-speed data flow.
The Importance of Carrier Aggregation in 5G
- Much Higher Throughput
By merging carriers, operators can offer broader effective bandwidth:
Number of CCsTotal BandwidthThroughput Impact1 CC20 MHzBaseline3 CCs60 MHz~3× data rates16 CCs (5G NR)640 MHzMulti-Gbps speeds
- Better Spectrum Utilization
Operators often don’t have large contiguous spectrum blocks. CA fixes this issue by bringing together fragmented spectrum across:
Mid-band
Low-band
mmWave
- Enhanced Coverage and Reliability
Low-band carriers boost coverage, while high-band carriers offer fast speeds. With CA, both can be used at once.
- Improved User Experience
Users enjoy:
Quicker downloads
Higher streaming quality
More stable connections
Better performance at cell edges
Explaining the Different Types of CA
- Intra-band Contiguous CA
Carriers are adjacent in frequency.
This is the simplest to implement.
- Intra-band Non-Contiguous CA
Carriers belong to the same band but have gaps in between.
This requires advanced RF filtering techniques.
- Inter-band CA
Carriers are from different frequency bands.
Most prevalent in 5G, for example, combining 700 MHz + 3.5 GHz + 28 GHz.
The diagram hints at a multi-cell inter-band situation.
Multi-Cell Carrier Aggregation (as depicted in the image)
The image goes beyond traditional CA. It showcases multi-cell Carrier Aggregation, where carriers are drawn not just from various frequencies, but also across different cells.
This setup supports:
Greater network capacity
Flexible load balancing
Multi-cell connectivity before fully implementing dual connectivity (DC)
Users can stay connected to multiple cells at once, boosting:
Resilience
Handover performance
Throughput aggregation across sectors or gNodeBs.
Challenges for the MAC Layer in Multi-Cell CA
Handling multiple carriers across cells brings about:
- Advanced Scheduling Algorithms
The MAC must factor in:
Load distribution
Carrier quality
Interference issues
Hardware limits
- Separate HARQ Processes for Each Carrier
Each carrier requires its own count and timing for HARQ processes.
- Cross-carrier Scheduling
Scheduling grants for one carrier might come through another.
- Increased Complexity for User Equipment (UE)
Devices need to support:
Multiple RF chains
Various antenna paths
Capability for multi-carrier demodulation
Advantages of MAC Carrier Aggregation in 5G Networks
✔ Greater spectral efficiency
✔ Faster data rates (multi-Gbps)
✔ Flexible use of spectrum across fragmented bands
✔ Enhanced connectivity stability for users at cell edges
✔ Improved QoS support for eMBB, URLLC, and mMTC
✔ More effective load balancing of the network
✔ Boosted mobility performance
Conclusion
MAC Carrier Aggregation is essential for achieving the high-performance features of 5G networks. By effectively coordinating multiple carriers at the MAC layer and transmitting them through the PHY layer across different cells, CA significantly enhances network throughput, spectrum efficiency, and the overall user experience. The image nicely illustrates how the RRC, SDAP, PDCP, RLC, MAC, and PHY layers work together to enable smooth multi-carrier communication across several cells.
As we advance toward 5G-Advanced and eventually 6G, we can expect CA to keep evolving, supporting even more component carriers, broader bandwidths, and a wider array of spectrum bands. For those in the telecom field, getting a grip on MAC Carrier Aggregation is crucial for navigating the future of wireless communication systems.
-
MAC Resource Allocation and Scheduling in 5G/6G: How Networks Balance eMBB, URLLC, mMTC, V2X & Broadcast Traffic

MAC Resource Allocation and Scheduling in 5G and 6G Networks
Efficient radio resource allocation is really the backbone of today’s wireless networks. The Medium Access Control (MAC) layer is key in making sure that different types of traffic—like eMBB, URLLC, mMTC, V2X, and broadcast—can operate together with minimal interference, top-notch quality of service (QoS), and ultra-high efficiency.
The diagram included gives a visual look at how frequency and time resources are shared dynamically among various service types. It illustrates MAC-level scheduling across both dimensions, showing the complexity and adaptability of scheduling as we move into 5G and the upcoming 6G networks.
This article breaks down the image, explaining how MAC resource scheduling operates, why it’s important, and how networks juggle different traffic needs at the same time.
Understanding the MAC Layer in Modern Wireless Systems
The MAC layer handles:
Resource allocation (who gets which time-frequency blocks)
Scheduling (when and how resources are assigned)
QoS differentiation (meeting targets for latency, reliability, and throughput)
Collision avoidance and coordination
HARQ mechanisms to ensure reliability
In 5G and 6G, the MAC layer has to cater to a wide range of services that come with vastly different needs—from the high throughput of eMBB to the ultra-low latency demanded by URLLC, and even the massive device density for mMTC.
Breaking Down the Image: Frequency–Time Resource Grid
The diagram outlines resource allocation along two axes:
Vertical axis = Frequency resources
Horizontal axis = Time resources
Various service types take up different areas, reflecting their unique traits and scheduling priorities.
The main traffic types shown include:
eMBB (Enhanced Mobile Broadband)
URLLC (Ultra-Reliable Low-Latency Communications)
mMTC / NB-IoT (Massive Machine-Type Communications)
eMTC (Extended Machine-Type Communications)
V2X (Vehicle-to-Everything Communication)
Broadcast Services
Blank Resources (unused or reserved blocks)
This illustrates how the MAC scheduler organizes traffic flows into time-frequency slots while honoring the specific performance needs of each service.
eMBB Allocation: High Throughput in Large Bands
The large teal areas in the diagram point out eMBB, which requires:
High bandwidth
Moderate latency
Flexible scheduling
High spectral efficiency
eMBB often gets contiguous blocks of spectrum for data-heavy tasks like:
UHD/VR streaming
Cloud gaming
High-capacity mobile broadband
The scheduler assigns sizable continuous time-frequency blocks to crank up throughput.
URLLC: Priority Scheduling for Low Latency and High Reliability
The slim gray vertical segments denote URLLC, showing how its resources fit into the grid even when eMBB is in the mix.
URLLC has strict needs:
1 ms → 0.1 ms latency
99.9999% reliability (10⁻⁶ to 10⁻⁹ BLER)
Instant scheduling
Key points about URLLC resource allocation include:
Scheduled preemptively over eMBB when necessary
Takes up small but frequent resource blocks
Utilizes robust coding and mini-slots
This guarantees that mission-critical traffic—like industrial automation, remote robotics, and autonomous systems—meets strict QoS standards.
V2X Allocation: High Reliability & Mid-Latency Resources
The block labeled V2X (Vehicle-to-Everything) is for a service class that needs:
Mid-tier latency
High reliability
Moderate bandwidth
V2X services include:
Vehicle-to-vehicle communication
Cooperative awareness messages
Collision avoidance
Connected traffic infrastructure
V2X occupies a specific area on the grid, ensuring predictable scheduling.
Broadcast Scheduling: Efficient Use of Shared Resources
The orange block showing broadcast illustrates how networks plan:
Firmware updates
Multimedia broadcast (eMBMS)
Public warning systems
IoT multicast delivery
Broadcast transmissions are advantageous because of:
One-to-many efficiency
Wide-area coverage
Fixed time-frequency reservation
In multi-service settings, designating broadcast in distinct blocks boosts predictability and cuts down on interference.
mMTC, eMTC, and NB-IoT: Dense IoT Traffic Requires Narrowband Allocation
The lower right section displays:
mMTC–eMTC (Machine-type Communication)
mMTC–NB-IoT
These services focus on:
Massive device density (1M+ devices/km²)
Sporadic traffic
Low data rates
Long battery life
Narrowband and block-scheduled allocations work best for these IoT devices because:
They cut down on energy consumption
Narrowband lessens noise and boosts link budget
Scheduling is periodic and predictable
NB-IoT often operates within LTE/5G carriers using a compact narrowband slot.
Blank Regions: Reserved or DTX Periods
The blank sections represent:
Unused spectrum
Reserved guard periods
Muted subframes for interference control
Slots for dynamic TDD switching
Space for future traffic spikes
These blank resource blocks are crucial for:
Avoiding collisions
Managing inter-cell interference
Preparing for sudden URLLC traffic
Flexible slicing for new service classes
These unallocated blocks give schedulers the flexibility to adapt traffic on the fly.
How the MAC Scheduler Makes Decisions
Modern 5G/6G schedulers leverage advanced algorithms, including machine learning, to decide how to allocate resources.
Key factors for scheduling include:
- Channel Conditions
CQI, SINR, Doppler, and fading behavior.
- Traffic Class
eMBB, URLLC, mMTC, V2X, etc.
- QoS Requirements
Latency, reliability, throughput, jitter.
- Network Slicing Rules
SLAs for specific sectors (like smart factories).
- Priority Levels
Mission-critical → highest
eMBB → moderate
Non-time-critical → lowest
- Spectrum Availability
Carrier aggregation, TDD patterns, and bandwidth parts.
Schedulers work at sub-millisecond precision to meet these goals.
MAC Scheduling Techniques Used in 5G/6G
Common algorithms comprise:
Round Robin (RR)
Proportional Fair (PF)
Maximum Throughput (MT)
Latency-Optimized Scheduling
URLLC Preemption Scheduling
Machine Learning–Assisted Scheduling
New 6G MAC designs are incorporating AI-driven scheduling, semantic-aware allocation, and a combined approach to communication and sensing resource management.
Conclusion
The diagram gives a clear view of how the MAC layer handles resource allocation and scheduling across various service categories in 5G and the evolving 6G networks. By dynamically dividing time-frequency resources among eMBB, URLLC, V2X, mMTC, NB-IoT, and broadcast services, the MAC layer ensures that each traffic type gets the appropriate mix of throughput, latency, and reliability.
As networks gear up for 6G, MAC scheduling is set to become smarter—with AI-driven optimizations, ultra-flexible resource blocks, and new air-interface designs. Grasping MAC resource allocation today is key to crafting the networks of tomorrow.
[awsmjobs]