Wcdma
-
Upload
degdag-abdulrahman -
Category
Documents
-
view
2.189 -
download
18
Transcript of Wcdma
1) Soft Handover Overhead is high Soft Handover Overhead is higher than 45% in RNC, the value can’t meet KPI request, customer ask to
optimize SHO overhead.
Check cell coverage for improving overshooting and reducing SHO overhead with iNastar, we find some cells coverage to larger, and then ask to customer to down antenna tilt of those cells.
some value of parameters are different HW’s recommend value, particularly TrigTime1A (1A Time to trigger) still using NSN’s setting, after swap NSN network 2 years.
After changing TrigTime1A = D320 on Oct. 9th
2) PS CDR reduced due to inactivity timer opt. PS DCR was improved after 10/11 due to change PS inactive timer (10sec -> 5sec))
SET UPSINACTTIMER
PsInactTmrForCon
PsInactTmrForStr
PsInactTmrForInt
PsInactTmrForBac
Meaning: When detecting that the Ps' User had no data to transfer for a long time which longer than this timer, the PDCP layer would request the RRC layer to release this Radio Access Bear. So the number of normal release will increase which will result in decreasing the PS CDR = Abnormal Release / Abnormal Release + Normal Release
1) External Interference We found the KPI for Our site is not good, and the RTWP for all cells are very High.
We check the RTWP for Site New Sites GHB968:
We make a trace for RF Frequency Scaning by which we are confermed that there is some External
Interference
After This we conferm that there is some External Interference in our Network, so we just inform to our
coustomer to make it clear.
Always check the results for surrounding sites , if you are suspecting Interference.
1) Optimize PS RB Setup timer PS Drops are very high at RNC
After investigations we found a lot of Ps Drops due to coverage, SRB, TRB Resets and UU No Reply
RbSetupRspTmr
Wait RB setup response timer
Meaning: A timer to RNC wait for the RB setup response from UE in the RB procedure. Refer to the Note. The RB reconfiguration message may retransmit three times when the timer expires.The parameter modification has no impact on the equipment. GUI Value Range: 300~300000 Unit: ms Actual Value Range: 300~300000 MML Default Value: None Recommended Value: 5000 Parameter Relationship: None Service Interrupted After Modification : No (No impact on the UE in idle mode) Impact on Network Performance: None
So what's recommended is as below:
3) High RTWP Due to Micro Wave Interference New 3G NodeB has completed integration, RTWP was very high. This site was 2G and 3G
collocation site,before GSM is 1800M band, now UMTS is 2100M
From M2000 we got the RTWP value, the top sector 2 RTWP value was -80, sector 1 and
sector 3 were more than -100, it was serious problem. We did some work for this site below:
1. We exchanged the feeder and jumper, the RTWP didn't change with jumper and feeder ;
2. We replaced the all WRFU and WBBP board, the high RTWP not disappeared;
3. We blocked GSM all TRX in the morning during idle hour, but no any improvement.
4. After we monitor several days KPI,we found that the RTWP can reach the normal level on
sometime , we doubted that it was interference cause RTWP.so we check the installation, we
saw one antenna very near the Huawei antenna.
Negotiated with the other customer regarding reducing their MW power, after they reduce the
power ,the RTWP can reach normal value.
1) DL power congestion solved by admission control and CPICH power optimization
Cells suffer from high DL power congestion affecting accessibility KPIs (RRC, CS RAB & PS RAB %)
We took two actions:
Optimize CPICH power by decreasing it in both carrier’s
MOD UCELL:CellId=40483, PCPICHPower=340;
MOD UCELL:CellId=40488, PCPICHPower=340;
optimize the DL load threshold by controlling the admission control (CAC) of conversational
AMR service, conversational non-AMR service, and handover scenarios thresholds, where they
decide when to accept the call only if the load after admitting it is less than above four
thresholds depending on type (default values: 80, 80, 85, 75%)
MML Commands
MODUCELLCAC:CellId=40483,DlConvAMRThd=92,DlConvNonAMRThd=92,DlOtherThd=90,
DlHOThd=93, DlCellTotalThd=95;
MODUCELLCAC:CellId=40488,DlConvAMRThd=92,DlConvNonAMRThd=92,DlOtherThd=90,
DlHOThd=93, DlCellTotalThd=95;
40483: DL power congestion released and accessibility KPIs improved
40488: DL power congestion released and accessibility KPIs improved
1) PS Data traffic Increases drastically & HSDPA
traffic Decreases Simultaneously due to changing thresholds
Suddenly There is an Increases in PS data traffic & decreases in HSDPA traffic
First we need to check there is increases or decreases in RAB attemts
• If we look to HS RAB Attempts then there is an 50 % Decreases hence the HS traffic
decreases. Analysis
We checked the codes assigned for HS services. But before & after codes assigned is
same there is no change in PS & HS assigned codes. Means for HS it is 7 and
remaining codes is for R99
Then we found a change in parameter below that has been changed from D768 to D64
Parameter Name DL BE traffic threshold on HSDPA
Meaning Rate threshold for decision to use HS-DSCH to carry DL PS background/interactive services. When the maximum DL service rate is higher than or equal to this threshold, the service will be carried on HS-DSCH. Otherwise, it will be carried on DCH.
GUI Value Range D8, D16, D32, D64, D128, D144, D256, D384, D512, D768, D1024, D1536, D1800, D2048, D3600, D7200, D8640, D10100, D13900
Recommended Value
D64
After returning it back to its original
4) Relief High UL CE congestion by LDR action site 4092 suffers from high UL CE congestion affected PS RAB SR (Success Rate)%
Load Reshuffling (LDR) is required to reduce the cell load and increase the access success rate.
We enable Cell Credit(CELL_CREDIT_LDR) LDR, NodeB credit(NODEB_CREDIT_LDR ) LDR, Cell
Group Credit (LCG_CREDIT_LDR)
MODUCELLALGOSWITCH:CellId=40926, CELL_CREDIT_LDR-1;
MODUCELLALGOSWITCH:CellId=40927, CELL_CREDIT_LDR-1;
MODUNODEBALGOPARA:NodeBName="C1_0_DEL4092P1(DSK_TE)",NodeBLdcAlgoSwitch=NOD
EB_CREDIT_LDR-1&LCG_CREDIT_LDR-1; as both cells under same node-b
Then I define the 1st, 2nd , 3rd actions of the LDR to ones that can solve the UL CE problem, as not
all actions in LDR can solve UL CE as inter-freq HO as example
4092 high CE Usage and after LDR action the CE usage decreased
CE Congestion released & PS RAB SR improved
5) Poor PS CSSR due to UL Power congestion For lot of cells had this problem we took on each cell one or more of below actions:
1) increase UlTotalEquseNum from 160 to 200
As in CAC, UL is admitted if algorithm 2 is applied which is the case if
{{{{{(ENUtotal + ENUnew)/ UlTotalEqUserNum}}}}} <
{{{{{UlNonCtrlThdForHo/AMR/NonAMR/Other}}}}}
2) Activated UL LDR CE/Power and modified UL LDR actions to correspond to UL CE
We enable Cell Credit(CELL_CREDIT_LDR) LDR, NodeB credit(NODEB_CREDIT_LDR ) LDR, Cell
Group Credit (LCG_CREDIT_LDR) and UL_UU_LDR-1;
3) lower UL LDR trigger threshold from 65 to 55
To make LDR work faster
UlLdrTrigThd=55, UlLdrRelThd=45;
Conclusion: Top 3 worst cells UL power Cong recover:
6) IRAT Performance Improvement Actions
Cause Analysis:
CS IRAT and PS IRAT bad bec high physical channel failure at worst cells (which refers to failure due to RF problems) + failure due to congestion (found only in CS as PS has no preparation) After finding out 2 major reasons for CS and PS IRAT failures we investigate further and found bellow mentioned conclusions –
Handling Process:
Now we know that route cause of poor IRAT performance was congestion at target 2G cells and poor 2G coverage at time of IRAT handovers. Capacity augmentation done by 2G team on request for congested 2G cells on and PS IRAT performance improved after this. We also done bellow mentioned parameter optimization to further improve IRAT performance as it was still bellow baseline –
1) 3A event:
The estimated quality of the currently used UTRAN frequency is below a certain
threshold and the estimated quality of the other system is above a certain threshold
QOtherRAT + CIOOtherRAT ≥ TOtherRAT + H3a/2 QUsed ≤ TUsed - H3a/2
Recommended values of TOtherRAT:
Parameter Recommended Value
TargetRatCsThd 16, namely -95dBm
TargetRatR99PsThd 16, namely -95dBm
TargetRatHThd 16, namely -95dBm
We changed TargetRatHThd=16 to 26
2)
Also PenaltyTimeForPhyChFail=30 to 60 at worst cells
Parameter ID PenaltyTimeForPhyChFail
Parameter Name Inter-RAT HO Physical Channel Failure Penalty Timer
Meaning Duration of the penalty for inter-RAT handover failure due to physical channel failure. The UE is not allowed to make inter-RAT handover attempts within the penalty time. For details about the physical channel failure, see 3GPP TS 25.331.
Unit s
Actual Value Range
0~65535
Default Value 30
3)
In 3A:
QOtherRAT + CIOOtherRAT ≥ TOtherRAT + H3a/2 CIO is composite of CIO(2G) + CIOoffset(3G2G), so we decreased the CIOoffset to give less priority to
2G to HO to it
4) Increase timer T309
Parameter ID T309
Parameter Name Timer 309
Meaning T309 is started after the UE is reselected to a cell belonging to another radio access system in connected mode, or the CELL CHANGE ORDER FROM UTRAN message is received. It is stopped after the UE is successfully connected in the new cell. The UE will continue the connection to UTRAN upon expiry. Protocol default value is 5.
Unit s
Actual Value Range
1~8
Default Value 5
7) different RTWP between F1 and F2 of the same sector
during normal audits of the network we found that for some sectors there is a diffrence in the RTWP between F1
and F2 cell of the same sector,
To check we have to verify the following three parts:
1. we had to make sure that the equipment is not faulty
to check the equipment we swapped the sectors between sector1 and sector3 (connected the antenna of sector3 to the RRU and the feaders of sector1 and antenna of sector1 to the RRU and the feaders of sector3) and when we did that we found that the RTWP is the same and didnt move from sector3 to sector
2. we have to make sure that it is no external interference
checked using spectrum annalizer and we found that there is no external interference
3. we have to confirm it is traffic load or not
was the problem, basically the second carrier is used for data traffic, and it was
noticed that the HSDPA traffic on this cell is relatively high compared with the trend of the first carrier, Such traffic difference especially in HSDPA and HSUPA can be the reason of the difference between RTWP of the first and second carrier cells. It is so clear from the below hourly snap that the RTWP is increasing and decreasing with the change of the HSDPA and HSUPA number of users
here is F2 G31377
here is for F1 G31373 and F2 G31377
1) HSDPA low throughput analysis DT of a cluster we found that the throughput is not high in special areas as per the below snap
Radio conditions was good, CQI of that road was very good (average more than 23) which we verified as per the
below snap
the IUB utlization is normal and there is no congestion as well as the power, below snaps of the IUB utilization at
the test time:
we went deeper to check the number of codes assigned to the UE during the test we found that the number of
codes was very low as per the snap
Reason we found that the NodeB lisence for the number of codes was normal and the feature of the dynamic
codes allocation is activated on the nodeB, but when we checked the average number of users ber hour in a day
we found that the cell is covering alot of users of HSDPA services below snap to show the number of users hourly
8) HSDPA Rate was LOW due to 16QAM not activated
was swapping vendor and after we swapped the first cluster we found the HSDPA rate is Low comparing to
the value we have before Swap
1- We sent a DT Engineer and started to make a test. 2- Also we checked the IUB BW and the number of HSDPA users configured on the sites and the
number of codes configured for each site. 3- From point 2 we found everything is OK. 4- But from the DT log files we found the following:
5- the DT log files we found the following, We found all the samples under the QPSK and zero samples at 16 QAM.
we checked the NodeB configuration found the 16QAM switch enabled on all the sites from LST MACHSPAR we found one parameter was not exist in our NodeB License: HSDPA RRM license, after activating it 16-
QAM worked and throughput for the same HSDPA traffic increased
1) Idle Mode 2G-3G optimization to stay more on 3G
To offload traffic over 2G and to make user under 3G coverage more, Change parameter
FDDQMIN from -10 dB to -14 dB on 2G side
SSEARCHRAT from -8 to -9 on 3G side
Inter-RAT measurement:
Squal ≤ SsearchRATm
↓
Qqualmeas − Qqualmin ≤ SsearchRATm
↓
Qqualmeas ≤ Qqualmin + SsearchRATm
3G Coverage and traffic increase which can be seen from increase in HSDPA throughput ( more user in 3G for
longer time duration) also face power and CE blocking due to increase in 3G users on those sites which was fixed.
Huawei Confidential
HSDPA UE Mean Cell (increased after change, but reduced again since 20-Oct, probably due
to increased of power blocking)
1) Low PS traffic on F2 cells due to missing Blind HO neighbors.
-The problem was that After F3 Expansion on one site and during KPIs check for the period before expansion We found that site had very low PS traffic (very low PS RAB ATT) On F2 cells and it have very high traffic (very High PS RAB ATT) on F1 cells while the network strategy don’t permit for this scenario to be happened .
We found that the blind HO was not defined from F1F2 ADD
UINTERFREQNCELL:RNCID=1,CELLID=5022,NCELLRNCID=1,NCELLID=5025,BLINDHOFLAG=TRUE, NPRIOFLAG=FALSE,INTERNCELLQUALREQFLAG=FALSE;
9) PS RAB Succ Rate Degraded due to DRD Parameter and Blind HO
PS RAB degraded below baseline on 1st Sept 2012. From statistic, it is cause by top worst 2
nd
cells and not related to all cells in RNC level.
Start Time Period NE Name BSC6900UCell Carrier
RRC succ
rate
(RAN12)
(%)
RRC att
(RAN12)
(times)
RRC
succ(RAN
12)
(times)
AMR
RAB SR
(none)
AMR
RAB
Attempt
(none)
No.of
AMR
RAB
failure
(none)
AMR
RAB
Success
(none)
PS RAB
SR (none)
PS RAB
Setup
Attempt
(none)
PS RAB
Setup
Success
(none)
PS
traffic(R99
+H) (MB)
09/15/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.936 15830 15820 100 352 0 352 99.907 16295 16280 5217.114
09/15/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.908 35884 35851 99.845 646 1 645 99.961 36233 36219 6400.582
09/15/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.923 31532 31508 100 662 0 662 99.966 32585 32574 9391.246
09/15/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 10 0 10 100 4 4 870.541
09/15/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 - 0 0 100 19 0 19 100 20 20 641.662
09/15/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 - 0 0 100 6 0 6 100 5 5 1309.272
09/16/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.956 15938 15931 100 300 0 300 99.933 16489 16478 2901.555
09/16/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.931 34830 34806 100 572 0 572 99.98 35060 35053 5361.683
09/16/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.918 32950 32923 99.881 847 1 846 99.982 33808 33802 11972.92
09/16/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 5 0 5 100 13 13 295.1
09/16/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 - 0 0 100 6 0 6 50 2 1 793.231
09/16/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 - 0 0 100 9 0 9 100 10 10 732.952
09/17/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.952 16991 16983 100 375 0 375 99.919 17419 17405 2831.682
09/17/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.911 30405 30378 100 620 0 620 99.97 30656 30647 9296.874
09/17/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.894 34031 33995 100 626 0 626 99.962 34727 34714 9351.929
09/17/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 8 0 8 100 8 8 273.691
09/17/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 - 0 0 100 14 0 14 100 8 8 1023.958
09/17/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 - 0 0 100 9 0 9 100 6 6 3540.679
09/18/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.916 15504 15491 100 318 0 318 99.93 15929 15918 2747.42
09/18/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.885 34989 34949 99.843 640 1 639 99.966 35298 35286 7015.812
09/18/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.915 29601 29576 100 705 0 705 99.98 30358 30352 9678.257
09/18/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 6 0 6 100 8 8 629.362
09/18/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 - 0 0 100 12 0 12 100 10 10 688.146
09/18/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 - 0 0 100 15 0 15 100 13 13 512.788
09/19/2012 00:00:00 Label=GNH089C, CellID=8949 99.92 16372 16359 100 312 0 312 99.983 6025 6024 677.259
09/19/2012 00:00:00 Label=GNH089B, CellID=8948 99.897 31101 31069 100 406 0 406 99.971 14128 14124 1970.258
09/19/2012 00:00:00 Label=GNH089A, CellID=8947 99.941 29261 29244 99.78 455 1 454 99.99 11010 11009 3445.6
09/19/2012 00:00:00 Label=GNH089F, CellID=8952 100 1 1 100 71 0 71 99.932 10375 10368 4510.544
09/19/2012 00:00:00 Label=GNH089E, CellID=8951 100 1 1 100 131 0 131 99.935 16997 16986 3590.282
09/19/2012 00:00:00 Label=GNH089D, CellID=8950 - 0 0 100 115 0 115 99.913 18526 18510 6874.892
09/20/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.907 19448 19430 100 336 0 336 99.876 1615 1613 657.522
09/20/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.951 27057 27044 100 561 0 561 99.963 5477 5475 1068.383
09/20/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.565 28324 28201 100 609 0 609 100 3450 3450 4601.665
09/20/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 96 0 96 99.878 16484 16464 3402.389
09/20/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 100 1 1 100 123 0 123 99.984 19116 19113 10759.17
09/20/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 100 1 1 100 131 0 131 99.949 21689 21678 9006.012
09/21/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.96 17602 17595 99.519 208 1 207 99.973 3754 3753 743.707
09/21/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.94 38351 38328 99.703 675 2 673 99.917 4830 4826 1850.86
09/21/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.932 32648 32626 100 574 0 574 99.838 4940 4932 774.914
09/21/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 38 0 38 99.796 5417 5406 2419.261
09/21/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 - 0 0 98.496 133 2 131 99.909 16602 16587 7385.943
09/21/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 100 1 1 98.888 90 1 89 99.908 10978 10968 8592.546
09/22/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.934 18324 18312 100 371 0 371 100 3460 3460 1950.093
09/22/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.947 26804 26790 100 543 0 543 99.955 4491 4489 1672.599
09/22/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.924 30626 30603 100 679 0 679 99.979 4951 4950 912.409
09/22/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 58 0 58 99.789 7618 7602 2638.071
09/22/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 - 0 0 99.152 118 1 117 99.933 12118 12110 13333.32
09/22/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 - 0 0 98.611 72 1 71 99.899 11923 11911 19733.02
09/23/2012 00:00:00 1440 TUBRNC1 Label=GNH089C, CellID=8949 99.942 17461 17451 100 337 0 337 99.974 3933 3932 1000.152
09/23/2012 00:00:00 1440 TUBRNC1 Label=GNH089B, CellID=8948 99.928 25335 25317 99.584 481 2 479 99.977 4348 4347 337.365
09/23/2012 00:00:00 1440 TUBRNC1 Label=GNH089A, CellID=8947 99.942 27838 27822 100 447 0 447 100 1903 1903 808.139
09/23/2012 00:00:00 1440 TUBRNC1 Label=GNH089F, CellID=8952 - 0 0 100 55 0 55 99.852 6792 6782 2361.191
09/23/2012 00:00:00 1440 TUBRNC1 Label=GNH089E, CellID=8951 - 0 0 100 56 0 56 99.989 9998 9997 5304.287
09/23/2012 00:00:00 1440 TUBRNC1 Label=GNH089D, CellID=8950 - 0 0 100 86 0 86 99.915 11779 11769 5175.125
Modification
date
F1
F2
F1
F2
F1
F2
F1
F2
F1
F2
F1
F2
F1
F2
F1
F2
F1
F2
was due to is PS RAB UUFail with its sub counter PS RAB PhyChFail and PS RAB UuNoReply.
The reason for this degrade was following two reasons that after setting them right the things
returned normal as seen in above 2 figures
1. Blind HO Flag for Multi carrier cells inter-frequency relation was wrong setting
10) CSSR PS Degraded due to high PS Code Congestion after swap
PS CSSR was low because after investigating founf Failed due to Code Congestion
Later For soution We decided to change the Algorithm and Open the LDR in Cell Level at 2 sectors which
had code congestion.
The Parameters are
MOD UCELLALGOSWITCH: CellId=10051, NBMLdcAlgoSwitch=CELL_CODE_LDR-1; MOD UCELLLDR: CellId=10051, DlLdrFirstAction=CodeAdj, DlLdrSecondAction=Berated, DlLdrBERateReductionRabNum=1, GoldUserLoadControlSwitch=ON; PS CSSR Improved after Opening the LDR parameters
11) Low CS IRAT Handover Success Rate due to miss configuration in GSM band
The requested CS IRAT Handover Success Rate target is 95% but these 2 sites (3 sectors each)
could only achieve around 60% during busy hour as shown in picture below
main reason for the CS IRAT HO failure is due to IRATHO.FailOutCS.PhyChFail.
Note that blue counter is sum of the other 2
Next, checking on cell_gcell counter, found that almost all of the failures happened to the co-
site GSM as highlighted below
Checking from the Ios trace, it is found that after the RNC sends the
RRC_HO_FROM_UTRAN_CMD_GSM to UE, the UE replied an
RRC_HO_FROM_UTRAN_FAIL, and the reason is physicalChannelFailure as shown
below.
Row Labels Sum of VS.IRATHO.AttOutCS.GCell Sum of VS.IRATHO.SuccOutCS.GCell Number of Failure
UCELL_GCELL:43017:740:01:14149:19093 1 1 0
UCELL_GCELL:43017:740:01:14149:19383 3 2 1
UCELL_GCELL:43017:740:01:14149:2812 29 27 2
UCELL_GCELL:43017:740:01:14149:40492 2 2 0
UCELL_GCELL:43017:740:01:14149:41001 17 2 15
UCELL_GCELL:43017:740:01:14149:41002 6 0 6
UCELL_GCELL:43017:740:01:14149:41003 5 0 5
UCELL_GCELL:43018:740:01:14149:19093 1 1 0
UCELL_GCELL:43018:740:01:14149:19383 2 2 0
UCELL_GCELL:43018:740:01:14149:19643 2 2 0
UCELL_GCELL:43018:740:01:14149:2812 11 11 0
UCELL_GCELL:43018:740:01:14149:40492 2 2 0
UCELL_GCELL:43018:740:01:14149:41001 10 2 8
UCELL_GCELL:43018:740:01:14149:41002 24 4 20
UCELL_GCELL:43018:740:01:14149:41003 3 0 3
UCELL_GCELL:43019:740:01:14149:19093 3 3 0
UCELL_GCELL:43019:740:01:14149:19383 2 2 0
UCELL_GCELL:43019:740:01:14149:2811 0 0 0
UCELL_GCELL:43019:740:01:14149:2812 103 101 2
UCELL_GCELL:43019:740:01:14149:2813 0 0 0
UCELL_GCELL:43019:740:01:14149:40492 4 4 0
UCELL_GCELL:43019:740:01:14149:41001 1 0 1
UCELL_GCELL:43019:740:01:14149:41002 1 0 1
UCELL_GCELL:43019:740:01:14149:41003 19 3 16
UCELL_GCELL:43027:740:01:14150:41011 59 6 53
UCELL_GCELL:43027:740:01:14150:41012 6 3 3
UCELL_GCELL:43027:740:01:14150:41013 23 0 23
UCELL_GCELL:43027:740:01:16202:19082 85 83 2
UCELL_GCELL:43027:740:01:16202:19083 9 9 0
UCELL_GCELL:43027:740:01:16202:19261 8 8 0
UCELL_GCELL:43027:740:01:16202:19262 7 6 1
UCELL_GCELL:43028:740:01:14150:41011 3 0 3
UCELL_GCELL:43028:740:01:14150:41012 40 11 29
UCELL_GCELL:43028:740:01:14150:41013 16 0 16
UCELL_GCELL:43028:740:01:16202:19261 6 5 1
UCELL_GCELL:43028:740:01:16202:19262 17 16 1
UCELL_GCELL:43028:740:01:16202:40071 7 5 2
UCELL_GCELL:43028:740:01:16202:40073 1 1 0
UCELL_GCELL:43029:740:01:14150:41011 16 3 13
UCELL_GCELL:43029:740:01:14150:41012 8 1 7
UCELL_GCELL:43029:740:01:14150:41013 78 16 62
UCELL_GCELL:43029:740:01:16202:19082 7 7 0
UCELL_GCELL:43029:740:01:16202:19083 4 4 0
UCELL_GCELL:43029:740:01:16202:19261 33 31 2
UCELL_GCELL:43029:740:01:16202:19262 16 15 1
UCELL_GCELL:43029:740:01:16202:40071 6 6 0
UCELL_GCELL:43029:740:01:16202:40073 6 6 0
The problem was that the GSM cell when created and configured to be in co-BCCH mode which
the main BCCH is in 850MHz while 1900MHz as below from ADD GCELL
But when GSM is defined as external neighbor to the UMTS, it was defined in a band different
from the actual one
TYPE Freq. Band
Meaning: This parameter specifies the frequency band of new cells. Each new cell can be allocated frequencies of only one frequency band. Once the frequency band is selected, it cannot be changed. GSM900: The cell supports GSM900 frequency band. DCS1800: The cell supports DCS1800 frequency band. GSM900_DCS1800: The cell supports GSM900 and DCS1800 frequency bands. GSM850: The cell supports GSM850 frequency band. GSM850_DCS1800: The cell supports GSM850 and DCS1800 frequency bands. PCS1900: The cell supports PCS1900 frequency band. GSM850_PCS1900: The cell supports GSM850 and PCS1900 frequency bands. TGSM810: The cell supports TGSM810 frequency band. GUI Value Range: GSM900, DCS1800, GSM900_DCS1800, GSM850, PCS1900, GSM850_1800, GSM850_1900, TGSM810 Unit: None Actual Value Range: GSM900, DCS1800, GSM900_DCS1800, GSM850, PCS1900, GSM850_1800, GSM850_1900, TGSM810
MML Default Value: None Recommended Value: None Parameter Relationship: None Service Interrupted After Modification : Not involved Impact on Network Performance: None
ADD UEXT2GCELL):
BandInd Inter-RAT Cell Frequency Band Indicator
Meaning: When the inter-RAT cell frequency number is within the range 512-810, the parameter indicates whether this frequency number belongs to the DSC1800 or PCS1900 frequency band. GUI Value Range: GSM900_DCS1800_BAND_USED(Use GSM900M or 1800M frequency band), PCS1900_BAND_USED(Use GSM1900M frequency band) Unit: None Actual Value Range: GSM900_DCS1800_BAND_USED, PCS1900_BAND_USED MML Default Value: GSM900_DCS1800_BAND_USED Recommended Value: GSM900_DCS1800_BAND_USED Parameter Relationship: None Service Interrupted After Modification : No (No impact on the UE in idle mode) Impact on Network Performance: None
So when the UE try to make the handover to GSM PCS1900MHz band, the RNC had instructed
the UE to search for DCS1800 band which caused the failure.
After the implementation, the CS IRAT Handover Success Rate has improved obviously as below:
12) Abnormal high RTWP due to improper setting on NodeB
During cluster acceptance O operator swap project, it was found cell W6374B3 and W6229B3 always be
the top worst cells in AMR drops.
AMR drops for the 7days.
PS DCR was also having relatively poor KPIs, which was 5~30% in these 2 cells.
Scanning through for possible reason of drops, it was found both cells having abnormal high RTWP
We checked hardware problems related to parameters as following:
It was found there is improper setting in desensitization intensity (DSP DESENS) in both problem
cells as shown below.
1. After revert, RTWP of both cells back to normal, on level of -105dBm as shown below.
2. PS DCR of these 2 cells (W6229B3 & W6374B3) showed significant improvement to level of 1% as shown below.
13) Poor PS IRAT Handover SSR due to congestion issue on adjacent 2G sites
Symptom:
PS IRAT handover SSR of sector B and C degraded significantly at busy time.
Cause
Analysis:
1. Missing neighbouring 2G cells;
2. Poor coverage;
3. IRAT configuration (3G or 2G side);
4. Congestion on adjacent 2G sites;
5. PS - CN Topology and configurations ( Intra-SGSN or Inter-SGSN handover, Routing Area Update failures )
6. Others
Handling
Process:
1. Checked the CS IRAT HO SSR of the site, which is much better than PS IRAT HO SSR and acceptable. So neighbors
and coverage should not be the issue; (most probably is congestion as CS prepare channel while PS don’t)
2. PS IRAT HO SSR degraded only at busy time, which is most probably caused by congestion issue on adjacent 2G
sites. Checked TBF, GPRS and Edge congestion situation of adjacent 2G sites, and there are serious congestion issue
found.
T591B:
T591C:
T6425B:
T6574A:
T5565C:
3. After expansion on adjacent 2G sites, PS IRAT HO SSR was improved significantly.
14) Analysis Report for bad RTWP on one NodeB caused by External Interference
bad RTWP on one NodeB.
Action Plan: 1st Action: Request FLM team to perform below actions:
Check connectors/combiner.
Replace combiner,
Check WMPT,
And if still issue not clear, then re-commission the site.
After performing all above actions the RTWP issue still exist on this site (3 sectors), suspected
internal/external interference.
2nd Action: Request to change UARFCN from Freq1 band 1 (UL 9613 DL 10563) to Freq Band 6
(UL 9738 DL 10688) which is 25M apart from 1st freq on site “120031_A_Dahlan_3G” for trial
purpose,
After change frequency RTWP normal
So now we know that there is interference on the 1
st freq, so we will continue using this 2
nd trial freq until
interference is solved in first one, but the problem with 2nd freq is that the KPI’s where not good as seen
below:
CSSR decrease: RRC.FailConnEstab.NoReply bad.
DCR Increase: VS.RAB.AbnormRel.PS.RF.SRBReset/VS.RAB.AbnormRel.PS.RF.ULSync
/VS.RAB.AbnormRel.PS.RF.UuNoReply bad.
Traffic increased.
So we want to find what is the problem
3rd Action: the first thing found wrong on 2nd freq from Audit Parameters is that there is no inter-freq
HO activated as in 1st freq from below parameter,
we found the HOSWITCH_HO_INTER_FREQ_HARD_HO_SWITCH=FALSE which states that there
is no IFHO performed Note that there is another switch HO_ALGO_LDR_ALLOW_SHO_SWITCH: this switch is to activate the inter-freq HO triggered by LDR
and LDR only, it means whether LDR action “inter-freq” can trigger inter-freq HO or not, while the previous one is whether inter-freq is
activated or not which is a must as if not activated this parameter will not have any meaning
so before in 1st freq some UE’s performed inter-freq as there was no good intra-freq cell, so if no inter-
freq the UE will keep work on the current freq that will increase traffic on current freq and also this will
result in more CDR probability
After fix switch: IFHO comes normal, here below KPI of IFHO success rate
there is improvement in all KPIs but still not good, so we need to improve more
4th
Action: we wanted to enhance the KPI’s for the 2nd freq even more, Check propagation delay
distribution for site 120031_A_Dahlan_3G before and after changing the freq: Found site
overshooting after change frequency:
ID Counter Description
73423486 VS.TP.UE.0 Number of RRC Connection Establishment Requests with Propagation Delay of 0
73423488 VS.TP.UE.1 Number of RRC Connection Establishment Requests with Propagation Delay of 1
73423490 VS.TP.UE.2 Number of RRC Connection Establishment Requests with Propagation Delay of 2
73423492 VS.TP.UE.3 Number of RRC Connection Establishment Requests with Propagation Delay of 3
73423494 VS.TP.UE.4 Number of RRC Connection Establishment Requests with Propagation Delay of 4
73423496 VS.TP.UE.5 Number of RRC Connection Establishment Requests with Propagation Delay of 5
73423498 VS.TP.UE.6.9 Number of RRC Connection Establishment Requests with Propagation Delay of 6~9
73423510 VS.TP.UE.10.15 Number of RRC Connection Establishment Requests with Propagation Delay of 10~15
73423502 VS.TP.UE.16.25 Number of RRC Connection Establishment Requests with Propagation Delay of 16~25
73423504 VS.TP.UE.26.35 Number of RRC Connection Establishment Requests with
ID Counter Description
Propagation Delay of 26~35
73423506 VS.TP.UE.36.55 Number of RRC Connection Establishment Requests with Propagation Delay of 36~55
73423508 VS.TP.UE.More55 Number of RRC Connection Establishment Requests with Propagation Delay Greater than 55
Each propagation delay represents three chips. The propagation distance of one chip is 78 m. Therefore, one propagation delay corresponds to 234 m. When the propagation delay is 0, it indicates that the UE is 0-234 m away from the base station. When the propagation delay is 1, it indicates that the UE is 234-468 m away from the base station. When the propagation delay is 2, it indicates that the UE is 468-702 m away from the base station. ...... When the propagation delay is 55, it indicates that the UE is 12870-13104 m away from the base station
Here is before changing freq for 3 sectors
Here is after changing and RTWP was fixed
So as u can see the 2nd freq has more coverage, this comes from the fact that 2nd freq has no
continues coverage as 1st freq, as not commonly used freq by other neighbor sited, so this
resulted in less HO that made coverage is more
1) Bad Quality (ECIO) for due to high Users/RTWP
There was bad Ec/No as seen below in DT
This is not a permanent issue as found mainly in busy hour as seen below
The problem mainly was due to high traffic as seen below when number of users increase the RTWP increase up
to -92dB which degrade the quality (Ec/No) in UL which is the same in DL
So the problem was due to not external interference but high traffic
So there are number of solutions to solve high traffic
1) SHO failure due to Iur congestion The main problem in this swap was IuR congestion
Counter Description
VS.SHO.FailRLRecfgIur.OM.Tx
Number of failed radio link synchronous reconfigurations by DRNC on Iur interface because of OM intervention (cause value: OM Intervention)
VS.SHO.FailRLRecfgIur.CongTx
Number of failed radio link synchronous reconfigurations by DRNC on Iur interface because of insufficient RNC capability (cause value: Cell not Available, UL Scrambling Code Already in Use, DL Radio Resources not Available, UL Radio Resources not Available, Combining Resources not Available, Measurement Temporarily not Available, Cell Reserved for Operator Use, Control Processing Overload, or Not enough User Plane Processing Resources)
VS.SHO.FailRLRecfgIur.CfgUTx
Number of failed radio link synchronous reconfigurations by DRNC on Iur interface because of improper configurations (cause value: UL SF not supported, DL SF not supported, Downlink Shared Channel Type not supported, Uplink Shared Channel Type not supported, CM not supported, Number of DL codes not supported, or Number of UL codes not supported)
VS.SHO.FailRLRecfgIur.HW.Tx
Number of failed radio link synchronous reconfigurations by DRNC on Iur interface because of hardware failure (cause value: Hardware Failure)
VS.SHO.FailRLRecfgIur.TransCongRx
Number of failed radio link synchronous reconfigurations by DRNC on Iur interface because of insufficient RNC transmission capability (cause value: Transport Resource Unavailable)
Note that if the counter is Tx it refers to DRNC while Rx refers to SRNC
According to the RNC statistics, the DRNC (ZTE) shows a big amount of failures
(VS.SHO.FailRLRecfgIur.CongTx, VS.SHO.FailRLAddIur.Cong.Tx and
VS.SHO.FailRLSetupIur.CongTx) than the SRNC(Huawei). Please find below the respective
pictures.
After investigation of the traces was detected the next problems which is there is big
congestion in code at ZTE RNC, here below is counters for some cells in ZTE RNC
So ZTE activated some algorithms on its side and changed some parameters to solve the
problem, which was actually solved as seen below
RNCId CellId CellName Time(As
day) VS.RAC.DCCC.Fail.Code.Cong VS.RAB.SFOccupy.Ratio VS.RAB.SFOccupy.MAX VS.RAB.SFOccupy
79 25656 256C5_6 2012-07-18 3.0000 0.9136 251.0000 233.8861
79 25652 256C5_2 2012-07-18 754.0000 0.9121 256.0000 233.5064
79 25655 256C5_5 2012-07-18 0 0.9107 246.0000 233.1368
79 14242 142U4_2 2012-07-21 822.0000 0.9097 255.0000 232.8829
79 28095 280C9_5 2012-07-18 0 0.9085 240.0000 232.5664
79 28891 288C9_1 2012-07-18 77.0000 0.9080 248.0000 232.4595
79 28896 288C9_6 2012-07-18 0 0.9080 243.0000 232.4520
79 45053 450C5_3 2012-07-18 85.0000 0.9080 253.0000 232.4490
79 27894 278C9_4 2012-07-22 63.0000 0.9072 255.0000 232.2551
79 62342 623U4_2 2012-07-25 808.0000 0.9068 254.0000 232.1405
79 24351 243C5_1 2012-07-18 89.0000 0.9067 255.0000 232.1035
79 62341 623U4_1 2012-07-18 223.0000 0.9066 254.0000 232.1025
79 14245 142U4_5 2012-07-18 0 0.9062 254.0000 231.9770
79 62343 623U4_3 2012-07-25 173.0000 0.9060 255.0000 231.9387
79 25651 256C5_1 2012-07-26 1562.0000 0.9059 256.0000 231.9010
79 53245 532U4_5 2012-07-18 0 0.9056 240.0000 231.8272
79 3754 037C5_4 2012-07-18 0 0.9051 255.0000 231.7155
79 25656 256C5_6 2012-07-20 0 0.9051 247.0000 231.6953
79 43752 437C5_2 2012-07-31 1025.0000 0.9051 255.0000 231.6940
79 3855 038C5_5 2012-07-18 34.0000 0.9049 256.0000 231.6653
79 25652 256C5_2 2012-07-27 109.0000 0.9049 256.0000 231.6500
79 28094 280C9_4 2012-07-18 18.0000 0.9049 256.0000 231.6447
79 28092 280C9_2 2012-07-18 874.0000 0.9049 248.0000 231.6443
79 43752 437C5_2 2012-07-29 906.0000 0.9048 256.0000 231.6314
79 24352 243C5_2 2012-07-18 30.0000 0.9047 248.0000 231.6035
79 17993 179C9_3 2012-07-23 585.0000 0.9047 255.0000 231.5929
79 43752 437C5_2 2012-07-30 871.0000 0.9045 256.0000 231.5526
79 25656 256C5_6 2012-07-19 1.0000 0.9045 246.0000 231.5394
79 25652 256C5_2 2012-07-23 31.0000 0.9044 255.0000 231.5190
79 25652 256C5_2 2012-07-19 200.0000 0.9043 253.0000 231.4931
79 62342 623U4_2 2012-07-26 219.0000 0.9041 256.0000 231.4475
79 62343 623U4_3 2012-07-26 1157.0000 0.9041 256.0000 231.4468
79 3653 036C5_3 2012-07-31 560.0000 0.9040 256.0000 231.4336
79 27896 278C9_6 2012-07-29 1247.0000 0.9040 256.0000 231.4212
15) DCR KPI degraded after NodeB rehoming from one RNC to another
Phenomen
on
Description
:
rehoming of 29 NodeBs to a new RNCon 24May. The following showed the abnormal release (DCR
nom) increased significantly after 24Jul while normal release (DCR denom) remained almost same
level.
Cause
Analysis:
1) Missing ncells 2) RNC parameters or switches 3) RNC license
Handling
Process:
This is a case of post rehoming KPI degradation, thus we, first of all, checked the ncells script for the rehoming operation. Found to have few missing ncells for Inter RNC neighboring relations. Complete ncells added on 25Jul night. DCR improved around 60%. Still it was suspected there is another reason behind the degradation.
After checking all the KPI again, it was found there is abnormal increase in CS traffic after rehoming. Thus we started to suspect these increase are related to the DCR degradation.
Then we went into details to check raw counters of every KPIs, and found that the CS IRAT HO attempts decreased till almost zero value, same went to PS attempts as well. This explained the reason why DCR increased and CS traffic increased abnormally as the CS calls have been kept and dragged in 3G till call drops.
3. Based on this assumption, we tried to compare the configuration of RNC Depok and RNC Depok2. No different in term of parameters and switches configuration.
4. Then we continued the verification on RNC license, found there was missing item called “Coverage Based Inter-RAT Handover Between UMTS and GSM/GPRS=ON” in RNC Depok2.
16) External Interference Interference Found in below cells.
• Amar_Taru (2286) – 3rd Sector.
• Panneri (2149) – 1st Sector.
Interference Test Analysis of Amar_Taru – 3rd Sector / Panneri 1st Sector
Field test observation – we had changed Azimuth of Panneri 1st Sector from 40* to 160* on
that time RTWP suddenly decreased that mean some Unknown frequency generating by
unknown source which is available near to Andheri Station which is same or very close to
RCOM UL Centre Frequency (1961.5MHz) .
17) AMR Call Drop Resolution By 2D 2F Parameter
change
Phenome
non
Descripti
on
RNC having high AMR call drop rate
Analysis
1. It is found that AMR call drop is happening after the compress mode is triggered from NASTAR.
Solution
Change the 2D 2F parameter setting of issued cells from:
After Parameter change:
there is improvement in AMR call drop rate after the changes done in IRAT 2D 2F
parameter settings.
18) Low PS CSSR due to Uplink Power Congestion Low PS CSSR on sector B of the site at busy time.
Cause Analysis
1. Resource Congestion;
2. Improper configuration;
3. RF issue;
4. CN issue;
5. Others
Handling Process:
1. Checked the traffic of the sector B, and the site has high traffic;
RTWP is very high at busy time
2. Check the PS RAB establish congestion on M2000, and the site has significantly high uplink
power congestion;
The HSUPA user number always hits the limit (20);
3. Analyze the coverage on Naster. The analysis result shows that the site can reach a distant area
(TP=20, Distance=4.6km).
4. With the Nastar result, we then check the site on Google earth. It is clear that the site has
overshooting and overlapping issue. Adjusting azimuth or downtilt is suggested.
Adjust the downtilt and azimuth as the red arrow shows, the issue was recovered with the
reduced traffic.
19) WCDMA DL Power Congestion Troubleshooting
we have found DL power congestion instatntly
If TCP ratio is very high, it means downlink power congestion. Then we can:
1. For single carrier cells, we can use downlink LDR: MOD CELLALGOSWITCH: CellId=0, NBMLdcAlgoSwitch=DL_UU_LDR-1; MOD CELLLDR: CellId=0, DlLdrFirstAction=BERateRed, DlLdrBERateReductionRabNum=1;
GoldUserLoadControlSwitch=ON;
2. For F1 cell, Setting LDR as follows: MOD CELLALGOSWITCH: CellId=0, NBMLdcAlgoSwitch=DL_UU_LDR-1; MOD CELLLDR: CellId=0, DlLdrFirstAction=BERateRed, DlLdrSecondAction=InterFreqLDHO,
DlLdrBERateReductionRabNum=1, GoldUserLoadControlSwitch=ON;
Then we can monitor the counters as follows to check the effect of LDR action:
VS.LCC.LDR.InterFreq
VS.LCC.LDR.BERateDL
VS.LCC.LDR.BERateUL
Note: usually power congestion will not happen in dual carrier cell. For single carrier
site, if power congestion is serious, expand carrier is recommended.
1) PSC planning to enhance CSSR
RNC having normal CSSR but to improve more PSC audit and change should be done
After the PSC change, CSSR improved.
Below is the cells that had PSC planning on
1) Uplink Power Congestion
HUAWEI Confidential
Uplink Power Congestion
Main Root Problem:
• High RAB failures on site 102373_SEKELOA_3G due to uplink power congestion.
Analysis :
• Uplink power congestion was found on site 102373_SEKELOA_3G although parameter
ULTOTALEQUSERNUM has been set to 200 (=maximum value)
HUAWEI Confidential
Counter Description for LDR State:
Uplink Power Congestion
Action :
Disable UL power CAC for cell with high UL power congestion. For any cell with UL power
congestion still appear although ULTOTALEQUSERNUM has been set to 200 (=maximum value),
we decide to disable UL power CAC by setting NBMUlCacAlgoSelSwitch in UCELLALGOSWITCH
to ALGORITHM_OFF.
HUAWEI Confidential
Result :
• After changing NBMUlCacAlgoSelSwitch setting improvement in uplink power congestion.
Uplink Power Congestion
1) RF Coverage problem Solved Later by Modifying parameters related to cell radius
From DT
Found EC/IO and RSCP (little) Poor Near the Cell 080086 which was causing main problems
Solution According to Coverage Prediction Plot from Atoll we found that there is coverage shrink in the area due
to bad cell environment and so planned to change the cpich power
Increase Power CPICH from 330 to 390
RlMaxDlPwr from 0 to 10 for CS Services and 20 to 40 For PS Services for RAB 384 and
256 Kpbs
Service RL Max Downlink Transmit Power (dB) RL Min Downlink Transmit Power (dB)
Downlink SF
CS Domain
12.2 kbps AMR -3 -18 128
28 kbps -2 -17 64
32 kbps -2 -17 64
56 kbps 0 -15 32
64 kbps (VP) 0 -15 32
PS Domain
0 kbps -2 -17 256
8 kbps -8 -23 128
32 kbps -4 -19 64
64 kbps -2 -17 32
144 kbps 0 -15 16
256 kbps 2 -13 8
384 kbps 4 -11 8
Also
1) RRC Rej and RAB Fail and reason are RRC Rej and RAB Fail due to Code Congetion in WCDMA
KPI Analysis:
Solution:
If HS-PDSCH Reserved Code’s value is excessively high, the HSDPA code resource is wasted and the
admission rejection rate of R99 services increases due to code resource.
so we have change this parameter from 12 to 5 .
As I checked the site parameter config. And found Code number for HS-PDSCH is 12. So change it to 5 as per baseline.
After reduce the HS-PDSCH Code problem is solved.
20) CS IRAT HO Problem due to LAC miss-configuration [HO]
When we implemented the work order of RNC in one region we got the IRAT HO Success Rate
of 24%.
After we executed one work order on 69 sites of One RNC in one region we got so many IRAT
failures.
BSC6900UCell IRATHO.FailRelocPrepOutCS.UKnowRNC
Label=UBEN077_S1, CellID=20771 3350
Label=UBEN077_S3, CellID=20773 1998
Label=UBEN007_S2, CellID=20072 1796
Label=UBEN077_S2, CellID=20772 940
Label=UBEN017_S1, CellID=20171 874
Label=UBEN038_S3, CellID=20383 844
Label=UBEN070_S1, CellID=20701 631
Label=UBEN901_S2, CellID=29012 507
Label=UBEN039_S1, CellID=20391 482
Label=UBEN901_S3, CellID=29013 388
Label=UBEN901_S1, CellID=29011 327
Label=UBEN017_S2, CellID=20172 314
Label=UBEN070_S2, CellID=20702 308
Label=UBEN028_S2, CellID=20282 255
Label=UBEN025_S1, CellID=20251 252
Label=UBEN032_S2, CellID=20322 218
1. Checked neighbor data from 3G to GSM Handover in RNC, checks each NGSM cell
information, there is no problem in that.
2. Traced singling in RNC using LMT and found many prepare handover failed, the
reason is “unknown target RNC”. What backed it out is that the counters from
M2000 that counts are
IRATHO.FailOutCS.PhyChFail
IRATHO.FailRelocPrepOutCS.UKnowRNC
3. Based on that we have checked the configured LAC in MSC, checked MSC data and
find LAI is wrong.
After the LAI modifications in the RNC & MSC we have got The IRAT HO success Rate of
97%
21) How to improve PS IRAT Success rate
3G to 3G and 3G to 2G neighbor list review and optimization
3G-to-2G Handover Measurement Events - 2D QUsed ≤ TUsed2d - H2d/2 TUsed2d :
Parameter Recommended Value
InterRATCSThd2DEcN0 -14, namely -14dB
InterRATR99PsThd2DEcN0 -15, namely -15dB
InterRATHThd2DEcN0 -15, namely-15dB
InterRATCSThd2DRSCP -100, namely -100dBm
InterRATR99PsThd2DRSCP -110, namely -110dBm
InterRATHThd2DRSCP -110, namely -110dBm
HystFor2D 4, namely 2dB
TimeToTrig2D D320, namely 320ms
- Speed up handover to avoid failure due to poor RF by increased INTERRATR99PSTHD2DRSCP from -110 to -100dBm and INTERRATHTHD2DRSCP from -110 to -105dBm.
- Increase the penalty time PENALTYTIMEFORPHYCHFAIL from 30s to 60s to alleviate 2G congestion and control the number of 3G to 2G handovers ( avoid handover to high congestion 2G cell). - Adjust parameter INTERRATPHYCHFAILNUM from 3 to 1 to speed up the penalty period after first time physical channel failure.
Parameter ID InterRatPhyChFailNum
Parameter Name Inter-RAT HO Physical Channel Failure THD
Meaning Maximum number of inter-RAT handover failures allowed due to physical channel failure. When the number of inter-RAT handover failures due to physical channel failure exceeds the threshold, a penalty is given to the UE. During the time specified by "PenaltyTimeForInterRatPhyChFail", the UE is not allowed to make inter-RAT handover attempts. For details about the physical channel failure, see 3GPP TS 25.331.
3G-to-2G Handover Measurement Events - 3A QOtherRAT + CIOOtherRAT ≥ TOtherRAT + H3a/2
QUsed ≤ TUsed - H3a/2 TOtherRAT is the absolute inter-RAT handover threshold. Based on different
service types (CS , PS domain R99 service, or PS domain HSPA service), this threshold can be configured through the following parameters: TargetRatCsThd TargetRatR99PsThd TargetRatHThd
Parameters Optimization (SET 2) Adjust parameter TARGETRATR99PSTHD and TARGETRATHTHD from -95 to -90 dBm.
- GSM cells that contribute with high failure that affect IRAT success rate, you can decrease its priority by adjusting target cell (NPRIOFLAG, NPRIO, RATCELLTYPE).
Conclusion & Recommendations: >After implemented the actions according to KPI Improvement plan (page 3) , the target KPI : PS IRAT HO Success Rate significant improve from about 85.6% to 94.8 %.
1) R99 PS drop ratio increase after action the
64QAM due to CM on HSPA+ not activated
R99 PS Drop increase after activation of the 64QAM in March 5:
Firstly, activation time is confirmed by RNC operation log:
From counter analysis, we found per RNC that there are nearly 300 drops on PS R99 drop:
And TOP cell has nearly 30 drops R99 PS drop, other cell has several times R99 PS drop:
At the same time, H2D time begins to increase when activation of 64QAM is made:
Analyzing the RNC configuration, find that HSPA+ service is not allowed to start CM:
This configuration will cause 64QAM user in the bad coverage must turn to DCH from
HSDPA, then the user starts CM. This is more possible to drop.
In the IOS, some user drop after 64QAM UE return to DCH for bad coverage:
Solution
According to the above analysis, HSPA+ service can’t support CM, so HSPA+ user in bad
coverage return to DCH that causes R99 PS drop ratio increase.
SET UCMCF: EHSPACMPermissionInd=TRUE
2) SHO OVERHEAD PROBLEM solved by optimizing event 1B
During working on B project i found problem of SHO Overhead in RNC's is high
In Trial Optimisation : I present 2 batches for the optimisation
1st Batch 1.Select Cells where SHO Overhead is high and have high traffic/congestion. 2.Adjust antenna e-tilt to control coverage. If antenna e-tilt is already at maximum then go to (3). 3.Adjust SHO parameters IntraRelThdFor1BCSNVP and IntraRelThdFor1BPS from 12
(means 6dB) to 10 (means 5dB) to increase probabilities of triggering event 1B and improve SHO Overhead If SHO overhead is not Improved then we have to apply 2nd Batch 2nd Batch 1.Select Cells where SHO Overhead is still high Change TRIGTIME1B from 640 to 320 (ms) to further improvement.
After applying above, significant improvement occurred
22) FACH Congestion Reduction by increasing Air + Iub B.W
FACH congestion can be thought to be due to one of the below 3
reasons:
a. Air Interface congestion (SF64 where SCCPCH is configured at
is the bottleneck)
b. Iub Interface congestion (FACH BW which is configured to
4500byte/sec is the bottleneck)
c. Both Air and Iub interfaces are bottlenecks
Trial proposed area
ID Counter Description
67199740 VS.CRNC.IUB.FACH.Bandwidth FACH Bandwidth of CRNC for Cell
This counter provide the bandwidth of common channels for the CRNC on the Iub interface in the unit of bytes per second.
ID Counter Description
73439970 VS.FACH.DCCH.CONG.TIME Congestion Duration of DCCHs Carried over FACHs for Cell
73439971 VS.FACH.DTCH.CONG.TIME Congestion Duration of DTCHs Carried over FACHs for Cell
These counters provide the duration for which the DCCHs/DTCHs carried over the FACHs in a cell are congested. Unit:sec
Step1: Increasing the SF of 2nd SCCPCH from
SF64 to be SF32
It got result but not acceptable result
Step2: Increasing the Iub of the FACH to be
9000B/s instead of 4500 B/s (on top of SF32)
This solved the problem
23) Soft handover Overhead Reduction using event 1A
It is found that the main contributor to the SPU load is the soft handover. Most of the N
odeB are six sector NodeB, therefore, there will be more RL established per UE
From network audit analysis, 27% of the SPU load is caused by softhandover
Solution: Event 1A triggering threshold is reduced to make the event less likely to occur. Below is the command:
MOD UCELLINTRAFREQHO: RNCId=XX, CellId=XXXX, IntraRelThdFor1ACSVP=5,
IntraRelThdFor1ACSNVP=5, IntraRelThdFor1APS=5;
İt was changed from default value 6. Below is the result after change:
The soft handover overhead and SPU Load reduced after the change. The SPU
load usage reduction more than 10%
In addition, the call drop rate have not changed after the changes
Degrade in Paging Success Rate after IU-FLEX implementation
Customer in Country M, at office M , reported that there are degradations in Paging Success Rate for 1 RNC, IPRN5. The Paging
Success Rate (PSR) for idle UE on RNC IPRN5 was degraded since 14th
Sep 2012.
CAUSE ANALYSIS
o The problem is shown in Figure 1, where the IU Paging Success Ratio is degraded.
Figure 1 PSR for idle UE
o As shown in Figure 2, the RRC successful connection rate stayed almost the same. This indicated that there is nothing wrong with the common part which RRC connection and paging share together, including UU interface, NODEB, IUB, some internal modul
es of RNC.
Figure 2 RRC successful connection rate
o Besides, there’s no flow control/ discarded detected, as shown in Figure 3.
Figure 3 CPUSALLVS.PAGING.FC.Disc.Num.CPUs
o In addition, from the performance file, there is no PCH congestion found at all, as shown in figure 4, and there is no paging discarded too.
o It shows that, the paging message should successfully be delivered from IU interface to UU interface. This conclusion together with point 1 indicates the PSR deterioration is not caused by UTRAN.
Figure 4 UCELLALLVS.RRC.Paging1.Loss.PCHCong.Cell
ROOT CAUSE ANALYSIS:
PSR for the idle UE on RNC is calculated by the formula:PSR=VS.RANAP.Paging.Att.IdleUE/VS.RANAP.Paging.Succ.IdleUE. The denominator and the numerator are shown in Figure 5.
Figure 5 The denominator and the numerator for PSR
From an hour IU Trace, there is 286 location update failure out of 4042 location update requests in total with the reason shown as Figure 7.All the
failure was received from CN.
Figure 6 Location updating failure with different cause
FINDINGS:
From the analysis, we could say that after IU-FLEX, repeated paging mechanism could be altered, which could bring in more useless paging attempts. As a result, PSR on RNC is degraded.
Uplink power Congestion analysis and solution
I country M project, as the new construction developing, the network environment ,the type of service and number of users have also changed ,some cells of new UMTS sites uplink power congests,a great impact to the cell KPIs.
None
From M2000,Extract the top issue cell 050076_3G-3 counter“VS.RAB.FailEstabPS.ULPower.Cong” as below:
Time BSC6900UCell VS.RAB.FailEstabPS.ULPower.Cong
2012-10-15 Label=050076_3G-3, CellID=37836 220
2012-10-16 Label=050076_3G-3, CellID=37836 453
2012-10-17 Label=050076_3G-3, CellID=37836 124
We found when the UL power congest, the traffic is a little high,so we reduce CPICH power 1DB to decreases the coverage ,but we found the UL power still congestion after the revision, we doubt that lack of resources is not the root cause. We check the current network parameters, found uplink CAC algorithm switch of the issue cell is set to ALGORITHM_SECOND(The equivalent user number algorithm).
Algorithm Content
ALGORITHM_OFF Uplink power admission control algorithm disabled.
ALGORITHM_FIRST Power-based increment prediction algorithm for uplink admission control.
ALGORITHM_SECOND ENU-based admission algorithm for uplink admission control.
ALGORITHM_THIRD Power-based non-increment prediction algorithm for uplink admission control.
If we use ALGORITHM_SECOND,the network performs admission control based on the uplink equivalent number of users (ENU) of the cell and the predicted ENU caused by admitting new users. It means according to the different service types, equivalent to different number of users. When the cell equivalent nu
mber of users exceeds the set value(Here is 95), the cell will deny user access.
According the algorithm principle,we use “ALGORITHM_OFF “to disable uplink call admission control algorithm. After we monitor several days KPI,we found that the KPI can reach the normal level,and there are no abnormal fluct
uations with other KPIs:
For the uplink power congestion,we could analyze from the following two aspects 1.Lack of resources. a:Check CE adequacy of resources; b:Adjust the coverage.by modifying the pilot power and the maximum transmission power or by RF optimal adjustment. 2. Lead to the issue of parameter settings. Adjust cell parameters:as the access control algorithm
Page 1
Cells Location with LAC Borders
Below mentioned plot shows cells location with less than 98% RRC Registration
success rate with LAC borders, most of cells are located on LAC borders / covering in
open areas.
FACH Power was changed on B-A LAC
border cells from 1 to 1.8dB. Changes were
implemented on 20th July night .
RRC Registration has shown slight
improvement when compared to last Monday
hourly trend.
FACH Power & IdleQhyst2s Trial
RRC Registration attempts reduced as
expected after changing IDLEQHYST2S from 2
to 4dBm, but there was no change in RRC
success rate for reg. Changes were Reverted
on 20th July before FACH trial changes.
Page 3
Shown below is the Overall TP distribution for X area Cells. As shown in Map these cells are
facing in open area with no 3G coverage overlap.
Nearly 20.0% of samples lies in >1.5 Km
CS Traffic has increased after swap hence there is no loss of coverage after swap from legacy
Cluster
Name Pre-Swap KPI Post-Swap KPI
Cell Traffic Volume, CS / Week 33,795 40,479
Cause of the problem was attenuation not set and TMA not configured but were physically present on the Site .
On investigation we found that the cell having High RSSI were having TMA before Swap . But were not configured in the Huawei System afterwards . Also attenuation needs to be set accordingly . Here is the process and commands to check .
When there is no TMA, the attenuation value is set to 0. When the 12 dB TMA is used, the
attenuation value is set between 4 dB to 11 dB. When the 24 dB TMA is used, the attenuation
value is set between 11 dB and 22 dB. When the 32 dB TMA is used, the attenuation value is set
between 22 dB and 30 dB.
This command takes effects immediately after the execution.
ATTEN Attenuation of RX Channel(dB)
Meaning: It is the value of WRFU/RRU Rx attenuation. GUI Value Range: 0, 4~30 Actual Value Range: 0, 4~30 Unit: dB Default value: - Recommended value: None
Post Correction of the Antennuation .Here is RSSI post implementation. So after Swap we should check these to avoid the RTWP issue .
Report for PS RAB Success/UL Power Congestion analysis and Improved by changing CELL Loading Reshuffling –
CELLULDR Parameters
Detail Analysis:
In Moran RNC on Mosaic project >> PS RAB Success/UL Power congestion noticed
and due to which PS RAB get affected. To improve it >> Cell Loading Reshuffling
parameters UCELL_UU_LDR changed and due to change PS RAB get OK.
Failures Reason Analysis
Analyse the counter related to PS RABs and it is found that many call are failing on
counter : >> Number of Failed PS RAB Establishments for Cell (UL Power
Congestion) (none) <<<< On 1 particular Cell >> Eircom Wicklow_1
This counter means that there is UL power congestion in the uplink.
Pre KPI attached attached for refernece of Eircom Wiclow_1
Cluster Name Start time RAB Setup
Success Rate(PS)
Call Setup
Success Rate(PS)
Number of Failed PS RAB Establishments
for Cell (UL Power Congestion) (none)
Eircom Wicklow_1 11/23/2012 15:00 96.27% 96.12% 95
Eircom Wicklow_1 11/23/2012 16:00 93.45% 92.33% 208
Eircom Wicklow_1 11/23/201217:00 54.76% 54.33% 297
Eircom Wicklow_1 11/23/2012 18:00 74.66% 71.56% 592
Eircom Wicklow_1 11/23/2012 19:00 48.18% 47.55% 653
Eircom Wicklow_1 11/23/2012 20:00 69.51% 69.10% 432
Eircom Wicklow_1 11/23/2012 21:00 89.42% 89.03% 210
Eircom Wicklow_1 11/23/2012 22:00 94.21% 94.11% 117
Action Taken to Improve:
To Improve UL Power congestion>> 1 parameter related to CAC Algorithm is changed:
Make MOD UCELLCAC >>>> UlTotalEqUserNum (UL Total no. of User)>>>from 95 user>>> 200 user
But still after changing this parameter, UL Power Congestion problem did not resolve, there was
some improvement but Congestion was there.
So we change CELL LOADING RESHUFFLING PARAMETER
STEPS:
A) First SWITCH ON the UL LDR Switch by command:
MOD UCELL ALGOSWITCH: CELLID=65361; NBMLdcAlgoSwitch=UL_UU_LDR-1;
B) Change LDR Parameter:
MOD UCELLLDR: CELLID=65361; ULLdrFirstAction= BERateRed,
ULLdrBERatReductionRabNum=1; GoldUserLoadControlSwitch=ON:
DESCRIPTION OF PARAMETER:
KPI Analysed : KPI analysed after cahnge and Improvement found : KPI ATTACHED for refernece:
Cluster Name Start time
RAB Setup
Success Rate(PS)
Call Setup
Success Rate(PS)
Number of Failed
PS RAB Establishments
for Cell (UL Power
Congestion) (none)
Eircom Wicklow_1 12/08/2012 15:00 99.85% 99.73% 3
Eircom Wicklow_1 12/08/2012 16:00 99.79% 99.56% 0
Eircom Wicklow_1 12/08/2012 17:00 100.00% 99.64% 0
Eircom Wicklow_1 12/08/2012 18:00 99.89% 99.84% 0
Eircom Wicklow_1 12/08/2012 19:00 99.92% 99.88% 0
Eircom Wicklow_1 12/08/2012 20:00 99.88% 99.67% 0
Eircom Wicklow_1 12/08/2012 21:00 100.00% 100.00% 0
Eircom Wicklow_1 12/08/2012 22:00 99.97% 99.89% 0
1.4 UL Power congestion problem get resolved after changing this parameter
Report for PS RAB Failure due to UL Power Congestion and Improved by changing UCELLCAC UL User equivalent
number Parameter
Detail Analysis:
In Meteor RNC on Mosaic project PS RAB get degraded on 1 Site,So to improve
UCELL CAC UL UE equivalent parameter changed and due to change PS RAB get
OK.
Failures Reason Analysis
Analyse the counter related to CS/PS RABs and it is found that many call are failing on
counter : >> Number of Failed PS RAB Establishments for Cell (UL Power
Congestion) (none) <<<< On 1 particular Cell >> Ashford_MMC_2
This counter means that there is UL power congestion in the uplink.
Pre KPI attached attached for refernece of Ashford _2
Cluster Name Start time
RAB Setup
Success Rate(CS)
Call Setup
Success Rate(CS)
RAB Setup
Success Rate(PS)
Call Setup
Success Rate(PS)
Number of Failed PS RAB
Establishments for
Cell (UL Power Congestion)
(none)
Ashford_MMC_2 12/5/2012 19:00 79.31% 79.31% 92.23% 92.11% 47
Ashford_MMC_2 12/5/2012 19:00 100.00% 100.00% 92.22% 91.79% 100
Ashford_MMC_2 12/5/2012 20:00 86.21% 86.21% 54.76% 54.33% 306
Ashford_MMC_2 12/5/2012 20:00 82.98% 82.98% 54.23% 53.96% 590
Ashford_MMC_2 12/5/2012 21:00 82.61% 82.61% 43.78% 43.68% 574
Ashford_MMC_2 12/5/2012 21:00 87.50% 85.00% 46.31% 46.20% 422
Ashford_MMC_2 12/5/2012 22:00 86.67% 86.67% 86.73% 86.63% 102
Ashford_MMC_2 12/5/2012 22:00 100.00% 100.00% 87.31% 87.31% 168
Action Taken to Improve:
To Improve UL Power congestion>> 1 parameter related to CAC Algorithm is
changed:
Make MOD UCELLCAC >>>> UlTotalEqUserNum (UL Total no. of
User)>>>from 95 user>>> 200 user
DESCRIPTION OF PARAMETER:
Impact on Network Performance: If the value is too high, the system load after
admission may be over large, which impacts system stability and leads to system
congestion. If the value is too low, the possibility of user rejects may increase, resulting
in waste in idle resources.
KPI Analysed : KPI analysed after cahnge and Improvement found : KPI ATTACHED for refernece:
Cluster Name Start time RAB Setup
Success Rate(CS)
Call Setup
Success Rate(CS)
RAB Setup
Success Rate(PS)
Call Setup
Success Rate(PS)
Number of Failed
PS RAB Establishments
for Cell (UL Power
Congestion) (none)
Ashford_MMC_2 12/15/2012 19:00 96.77% 96.77% 99.75% 99.62% 0
Ashford_MMC_2 12/15/2012 19:00 100.00% 100.00% 99.83% 99.28% 0
Ashford_MMC_2 12/15/2012 20:00 100.00% 100.00% 100.00% 99.75% 0
Ashford_MMC_2 12/15/2012 20:00 100.00% 100.00% 99.92% 99.82% 0
Ashford_MMC_2 12/15/2012 21:00 100.00% 93.33% 99.89% 99.78% 0
Ashford_MMC_2 12/15/2012 21:00 100.00% 100.00% 100.00% 99.58% 0
Ashford_MMC_2 12/15/2012 22:00 100.00% 90.00% 100.00% 100.00% 0
Ashford_MMC_2 12/15/2012 22:00 100.00% 100.00% 99.89% 99.55% 0
1.4 UL Power congestion problem get resolved after changing this parameter
Report for CS RAB Failure due to DL Power Congestion and Improved by changing DLALGOSWITCH OFF Parameter
Detail Analysis:
In Meteor RNC on Mosaic project CS RAB get bad of 1 Site, So to improve
DLALGOWSITCH OFF parameter changed and due to change>> CS RAB get OK.
Failures Reason Analysis
Analyse the counter related to CS RABs and it is found that many call are failing on
counter : >> Number of Failed CS RAB Establishments for Cell (DL Power
Congestion) (none) <<<< On 1 particular Cell >> Balliguille Hill_1
This counter means that there is DL power congestion in the downlink.
Pre KPI attached attached for refernece of Balligullie Hill _1
Cluster Name Start time
Call Setup
Success Rate(CS)
Number of Failed CS RAB Establishments
for Cell (DL Power
Congestion) (none)
BallyguileHill_MMC_F1_1 11/28/2012 18:00 98.02% 44
BallyguileHill_MMC_F2_1 11/28/2012 19:00 97.17% 3
BallyguileHill_MMC_F1_1 11/28/2012 19:00 96.37% 17
BallyguileHill_MMC_F2_1 11/28/2012 20:00 97.00% 2
BallyguileHill_MMC_F1_1 11/28/2012 20:00 96.87% 14
BallyguileHill_MMC_F2_1 11/28/2012 21:00 96.99% 3
BallyguileHill_MMC_F1_1 11/28/2012 21:00 96.37% 13
BallyguileHill_MMC_F1_1 11/28/2012 22:00 98.00% 1
BallyguileHill_MMC_F1_1 11/28/2012 23:00 99.48% 2
Action Taken to Improve:
To Improve DL Power congestion>> 1 parameter related to CAC Algorithm is
changed:
Make DL CAC Algorithm Switch >>>> OFF >>>from>>> Algorithm First state
DESCRIPTION OF PARAMETER:
1. In OFF condition : DL CAC algorithm is disable.
In Algorithm First condition: The load factor prediction is ON.
If Algorithm first applied than after reaching load factor, new calls are rejected. While if
we disable it than it can take new call. We make it OFF most of the time at the time of
more load on site, while Algorithm First is used when we have more sites nearby and
reaching certain load/threshold, it can transfer calls to near by BTS
1.3 KPI Analysed : KPI analysed after cahnge and Improvement found : KPI ATTACHED for refernece:
Cluster Name Start time Call Setup Success Rate(CS)
Number of Failed CS RAB
Establishments for Cell (DL
Power Congestion)
(none)
BallyguileHill_MMC_F1_1 11/30/2012
18:00 99.42% 0
BallyguileHill_MMC_F1_2 11/30/2012
18:00 100.00% 0
BallyguileHill_MMC_F1_1 11/30/2012
19:00 99.65% 0
BallyguileHill_MMC_F1_2 11/30/2012
19:00 100.00% 0
BallyguileHill_MMC_F1_1 11/30/2012
20:00 99.75% 0
BallyguileHill_MMC_F1_2 11/30/2012
20:00 100.00% 0
BallyguileHill_MMC_F1_1 11/30/2012
21:00 99.04% 0
BallyguileHill_MMC_F1_2 11/30/2012
21:00 100.00% 0
BallyguileHill_MMC_F1_1 11/30/2012
22:00 99.60% 0
BallyguileHill_MMC_F1_2 11/30/2012
22:00 100.00% 0
BallyguileHill_MMC_F1_1 11/30/2012
23:00 99.58% 0
BallyguileHill_MMC_F1_2 11/30/2012
23:00 100.00% 0
1.4 DL Power congestion problem get resolved after changing this parameter
Phenomenon Description
Hsupa call drop increase after hsupa cm is permitted:
Cm permission ind on hsupa is changed from “limited” to “permit”
list rnc-oriented cmcf algorithm parameters
-------------------------------------------
cm permission ind on hsdpa = permit
cm permission ind on hsupa = permit
cm permission ind on hspa+ = permit
Alarm Information
none
Cause Analysis
Check behavior of all counters in hsupa call drop formula
Check expected behavior of the system when cm hsupa is permitted
Time
(As hour)
VS.HSUPA.
RAB.
Release
VS.HSUPA.
RAB
.AbnormRel.
Rate
VS.HSUPA.
RAB.
AbnormRel
VS.HSUPA.
RAB.
NormRel
VS.HSUPA.
E2D.
Succ
VS.HSUPA.
HHO.E2D.
SuccOut
IntraFreq
VS.HSUP
A.HHO.
E2D.
SuccOut
InterFreq
VS.HSUPA
.E2F.
Succ
2011-04-03 00:00:00 51682 1,02% 526 47033 4123 0 0 0
2011-04-03 01:00:00 47012 1,06% 498 43439 3075 0 0 0
2011-04-03 02:00:00 42068 0,54% 228 39714 2126 0 0 0
2011-04-03 03:00:00 39811 0,62% 246 37729 1836 0 0 0
2011-04-03 04:00:00 37147 0,48% 179 35489 1479 0 0 0
2011-04-03 05:00:00 35628 0,46% 165 34323 1140 0 0 0
2011-04-03 06:00:00 35007 0,47% 165 33860 982 0 0 0
2011-04-03 07:00:00 33478 0,48% 161 32371 946 0 0 0
2011-04-03 08:00:00 33488 0,47% 158 32243 1087 0 0 0
2011-04-03 09:00:00 36558 0,59% 216 34963 1379 0 0 0
2011-04-03 10:00:00 43005 0,78% 337 40493 2175 0 0 0
2011-04-03 11:00:00 46745 1,01% 472 43090 3183 0 0 0
2011-04-03 12:00:00 50449 0,92% 466 46371 3612 0 0 0
2011-04-03 13:00:00 53865 1,20% 646 48550 4669 0 0 0
2011-04-03 14:00:00 53655 1,21% 649 48341 4665 0 0 0
2011-04-03 15:00:00 53326 1,20% 640 48418 4266 2 0 0
2011-04-03 16:00:00 53662 1,25% 669 48145 4848 0 0 0
2011-04-03 17:00:00 56492 1,21% 685 50469 5338 0 0 0
2011-04-03 18:00:00 56744 1,32% 749 50089 5905 1 0 0
2011-04-03 19:00:00 59140 1,16% 688 52716 5736 0 0 0
2011-04-03 20:00:00 61355 1,30% 800 54158 6397 0 0 0
2011-04-03 21:00:00 60632 1,27% 771 53661 6198 2 0 0
2011-04-03 22:00:00 61460 1,19% 730 54917 5813 0 0 0
2011-04-03 23:00:00 57151 1,01% 575 51068 5508 0 0 0
2011-04-04 00:00:00 49978 0,83% 413 46046 3519 0 0 0
2011-04-04 01:00:00 45064 0,50% 227 42762 2075 0 0 0
2011-04-04 02:00:00 41767 0,57% 240 40027 1500 0 0 0
2011-04-04 03:00:00 38890 0,41% 161 37628 1101 0 0 0
2011-04-04 04:00:00 38198 0,33% 125 37153 920 0 0 0
2011-04-04 05:00:00 37880 0,30% 115 36891 874 0 0 0
2011-04-04 06:00:00 39438 0,31% 124 38379 935 0 0 0
2011-04-04 07:00:00 49245 0,52% 258 47385 1602 0 0 0
2011-04-04 08:00:00 76818 0,90% 694 72340 3784 0 0 0
2011-04-04 09:00:00 97637 0,81% 790 90664 6183 0 0 0
2011-04-04 10:00:00 101384 1,16% 1172 97995 2217 0 0 0
2011-04-04 11:00:00 102138 1,03% 1054 100975 107 1 1 0
2011-04-04 12:00:00 106681 1,09% 1165 105377 138 0 1 0
2011-04-04 13:00:00 107342 1,08% 1156 106054 130 0 2 0
2011-04-04 14:00:00 103931 1,15% 1194 102660 75 1 1 0
2011-04-04 15:00:00 100534 1,27% 1275 99181 77 0 1 0
2011-04-04 16:00:00 102318 1,22% 1249 100988 79 0 2 0
2011-04-04 17:00:00 103256 1,22% 1257 101918 79 1 1 0
2011-04-04 18:00:00 98919 1,46% 1443 97404 71 1 0 0
2011-04-04 19:00:00 89741 1,48% 1325 88373 43 0 0 0
2011-04-04 20:00:00 75692 1,50% 1138 74528 25 1 0 0
2011-04-04 21:00:00 70472 1,49% 1049 69387 36 0 0 0
2011-04-04 22:00:00 66384 1,50% 997 65327 59 1 0 0
2011-04-04 23:00:00 60195 1,61% 971 59160 62 2 0 0
Suggestions and Summary
it is important to analyze system behavior after one feature is activated in the network, so we can
explain the root cause of abnormal kpi behavior
Case name: Abnormal distribution of VS.RRC.AttConnEstab.Reg
Phenomenon Description: In country R during WCDMA optimization project, at the step of RRC CSSR optimization
RNO team found abnormal distribution of RRC attempts for registration reason. It takes around 50% of total RRC Attempts. Hardware version is BSC6810V200R011C00SPC100.
Symptoms: 1. High RRC attempts quantity.
2. Abnormal distribution of RRC attempts for registration reason
3. No any hardware alarms.
Analyze sequence: 1. Localize the problem.
2. Analyze possible reasons.
3. Perform Drive Test.
4. Check RNC level parameters.
Analyse Procedure: From statistic for RNC 4016 VS.RRC.AttConnEstab.Reg teaks around 50% of
total RRC Attempts Connection Establishment. Attempts are normally distributed among cells.
RNCName Time(As day) VS.RRC.AttConnEstab VS.RRC.AttConnEstab.Reg
RNC:4016 2011-08-10 791541 414010
RNC:4016 2011-08-11 811675 462559
RNC:4016 2011-08-12 796428 424042
RNC:4016 2011-08-13 815134 446783
RNC:4016 2011-08-14 835164 450958
02000004000006000008000001000000
2011
-08
-10
2011
-08
-11
2011
-08
-12
2011
-08
-13
2011
-08-
14
RNC:4016RNC:4016RNC:4016RNC:4016RNC:4016
VS.RRC.AttConnEstab
VS.RRC.AttConnEstab.Reg
At the same time for other 2 RNC's no such situation, RRC Attempts with Registration reason are
no more than 15%.
Such results exclude problem of CN because all 3 RNC’s of this region share same CN.
So possible reasons of such situation are:
1. Wrong RNC/Cell Level parameter settings. 2. Bad coverage and frequent reselection of 2G <-> 3G networks. For first reason we use Nastar Configuration Analysis Function to check difference in
parameters setting. No any difference.
For second reason RNO team decide to perform Drive Test to check coverage and UE
behavior. As result found that UE repeat to perform Combined RA/LA update and Location
Update every time failed with reason “MSC temporarily not reachable“. RA Update is
performed successfully.
This is root reason why registration quantity is so high.
About combined RA/LA: If the optional Gs-interface is implemented and the UE has entered a new LA as well as a new RA, a combined
RA/LA update will be performed. From the MS point of view, all signalling exchange takes place towards the SGSN. The
SGSN then updates the MSC/VLR. A combined RA/LA update takes place in network operation mode I when the UE enters a new RA or when a
GPRS-attached UE performs IMSI attach. The UE sends a Routing Area Update Request indicating that a LA update may
also need to be performed, in which case the SGSN forwards the LA update to the VLR. This concerns only CS idle mode, since no combined RA/LA updates are performed during a CS connection.
For our network Gs interface is not configured, so we checked Network Operation Mode for PS CNDomain. It
was set to NMO=Mode1. ADD CNDOMAIN:CNDOMAINID=PS_DOMAIN, DRXCYCLELENCOEF=6, NMO=MODE1; For other RNC’s it was set to NMO=Mode2. Nastar didn’t found configuration difference because it’s related to CN
configuration. After modification of NMO=Mode2 problem was solved and RRC attempts with registration reason decreased to 5% level.
Suggestion: For RAN performance optimization needs to pay attention at whole network
structure including Transmission and Core Network. Wrong setting of such global parameter like
NMO brings additional UE power, radio resource consumption, additional RNC SPU and CN
signalling loading.
0200000400000600000800000
1000000
2011
-08
-…
2011
-08
-…
2011
-08-
…
2011
-08
-…
2011
-08
-…
2011
-08
-…
2011
-08-
…
2011
-08
-…
2011
-08
-…
VS.RRC.AttConnEstab
VS.RRC.AttConnEstab.Reg