Saturday, 16 November 2013
Agile IMS for the All-IP Era
Labels:
AIPN,
IMS,
LTE & 5G World Series,
Mobile Cloud,
OTT,
SDN / NFV
Tuesday, 12 November 2013
Mobile Video Offload using Wi-Fi is the only solution in the coming years
A very interesting infographic from Skyfire some months back highlighted some very valid issues about Video on mobiles.
Personally, I do watch quite a bit of video on my phone and tablet but only when connected using Wi-Fi. Occasionally when I am out, if someone sends me video clip on Whatsapp or some link to watch Video on youtube, I do try and see it. Most of the time the quality is too disappointing. It could be because my operator has been rated as the worst operator in UK. Anyway, as the infographic above suggests, there needs to be some kind of an optimisation done to make sure that end users are happy. OR, the users cn offload to Wi-Fi when possible to get a better experience.
This is one of the main reasons why operators are actively considering offloading to Wi-Fi and have carrier WiFi solutions in place. The standards are actively working in the same direction. Two of my recent posts on the topic of 'roaming using ANDSF' and 'challenges with seamless cellular/Wi-Fi handover' have been quite popular.
Recently I attended a webinar on the topic of 'Video Offload'. While the webinar reinforced my beliefs about why offload should be done, it did teach me a thing or two (like when is a Hotspot called a Homespot - see here). The presentation and the Video is embedded below. Before that, I want to show the result of a poll conducted during the webinar where the people present (and I would imagine there were quite a few people) were asked about how they think MNO will approach the WiFi solution in their network. Result as follows:
Here is the presentation:
Here is the video of the event:
Personally, I do watch quite a bit of video on my phone and tablet but only when connected using Wi-Fi. Occasionally when I am out, if someone sends me video clip on Whatsapp or some link to watch Video on youtube, I do try and see it. Most of the time the quality is too disappointing. It could be because my operator has been rated as the worst operator in UK. Anyway, as the infographic above suggests, there needs to be some kind of an optimisation done to make sure that end users are happy. OR, the users cn offload to Wi-Fi when possible to get a better experience.
This is one of the main reasons why operators are actively considering offloading to Wi-Fi and have carrier WiFi solutions in place. The standards are actively working in the same direction. Two of my recent posts on the topic of 'roaming using ANDSF' and 'challenges with seamless cellular/Wi-Fi handover' have been quite popular.
Recently I attended a webinar on the topic of 'Video Offload'. While the webinar reinforced my beliefs about why offload should be done, it did teach me a thing or two (like when is a Hotspot called a Homespot - see here). The presentation and the Video is embedded below. Before that, I want to show the result of a poll conducted during the webinar where the people present (and I would imagine there were quite a few people) were asked about how they think MNO will approach the WiFi solution in their network. Result as follows:
Here is the presentation:
Here is the video of the event:
Labels:
Apps,
Apps Video Streaming,
Data Offload,
Stats,
Videos,
Wi-Fi
Sunday, 10 November 2013
SIPTO Evolution
Couple of years back I did a post on SIPTO (Selected IP Traffic Offload) and related technologies coming as part of Rel-10. I also put up a comparison for SIPTO, LIPA and IFOM here. Having left it for couple of years, I found that there have been some enhancements to the architecture from the basic one described here.
I have embedded the NEC paper below for someone wanting to investigate further the different options shown in the picture above. I think that even though the operator may offload certain type of traffic locally, they would still consider that data as part of the bundle and would like to charge for it. At the same time there would be a requirement on the operator for lawful interception, so not sure how this will be managed for different architectures. Anyway, feel free to leave comments if you have any additional info.
Labels:
Data Offload,
Data Traffic Management,
NEC,
Release 10,
Release 12,
SIPTO
Wednesday, 6 November 2013
The Relentless Rise of Mobile Technology
Mobiles have been rising and rising. Couple of weeks back I read 'Mobile is considered the first and most important screen by nearly half of the 18- to 34-year-old demographic, according to research commissioned by Weve.'
The finding placed mobile ahead of laptops or PCs (chosen by 30.6 per cent) and way ahead of TV (12.4 per cent) as the first and most important screen in the lives of people between the ages of 18 and 34.
Just 5.8 per cent of those surveyed in the age group chose a tablet as their "first screen".
The research also found that 45 per cent of 18- to 34-year-olds consider their mobile their first choice of device when interacting with online content, placing the platform just ahead of laptops and PCs, which scored 43 per cent.
Among the wider 18 to 55 age group surveyed, a PC or laptop was seen as the "first screen" with 39.8 per cent naming either computer as their most important screen, while smartphones came second on 28 per cent.
TV was in third place with 27 per cent of people naming it as their most important screen. Five per cent of the total group said they considered a tablet their "first screen".
Only a quarter of the 18 to 55 age group said mobile would be their first choice platform if they wanted to access the internet, while nearly two thirds preferred to use a PC or laptop.Tomi Ahonen has always been referring to Mobile as the 7th Mass Media.
So when I saw this above picture (and there are more of them) in Ben Evaans slide deck (embedded below), it just reiterated my belief that Mobile will take over the world sooner or later. Anyway, the slides are interesting to go through.
Labels:
Apps,
Mobile Phones and Devices,
Smartphones,
Stats,
Trends
Monday, 4 November 2013
Key challenges with automatic Wi-Fi / Cellular handover
Recently in a conference I mentioned that the 3GPP standards are working on standards that will allow automatic and seamless handovers between Cellular and Wi-Fi. At the same time operators may want to have a control where they can automatically switch on a users Wi-Fi radio (if switched off) and offload to Wi-Fi whenever possible. It upset quite a few people who were reasoning against the problems this could cause and the issues that need to be solved.
I have been meaning to list the possible issues which could be present in this scenario of automatically handing over between Wi-Fi and cellular, luckily I found that they have been listed very well in the recent 4G Americas whitepaper. The whitepaper is embedded below but here are the issues I had been wanting to discuss:
I have been meaning to list the possible issues which could be present in this scenario of automatically handing over between Wi-Fi and cellular, luckily I found that they have been listed very well in the recent 4G Americas whitepaper. The whitepaper is embedded below but here are the issues I had been wanting to discuss:
In particular, many of the challenges facing Wi-Fi/Cellular integration have to do with realizing a complete intelligent network selection solution that allows operators to steer traffic in a manner that maximizes user experience and addresses some of the challenges at the boundaries between RATs (2G, 3G, LTE and Wi-Fi).Here is the paper:
Figure 1 (see above) below illustrates four of the key challenges at the Wi-Fi/Cellular boundary.
1) Premature Wi-Fi Selection: As devices with Wi-Fi enabled move into Wi-Fi coverage, they reselect to Wi-Fi without comparative evaluation of existing cellular and incoming Wi-Fi capabilities. This can result in degradation of end user experience due to premature reselection to Wi-Fi. Real time throughput based traffic steering can be used to mitigate this.
2) Unhealthy choices: In a mixed wireless network of LTE, HSPA and Wi-Fi, reselection may occur to a strong Wi-Fi network, which is under heavy load. The resulting ‘unhealthy’ choice results in a degradation of end user experience as performance on the cell edge of a lightly loaded cellular network may be superior to performance close to a heavily loaded Wi-Fi AP. Real time load based traffic steering can be used to mitigate this.
3) Lower capabilities: In some cases, reselection to a strong Wi-Fi AP may result in reduced performance (e.g. if the Wi-Fi AP is served by lower bandwidth in the backhaul than the cellular base station presently serving the device). Evaluation of criteria beyond wireless capabilities prior to access selection can be used to mitigate this.
4) Ping-Pong: This is an example of reduced end user experience due to ping-ponging between Wi-Fi and cellular accesses. This could be a result of premature Wi-Fi selection and mobility in a cellular environment with signal strengths very similar in both access types. Hysteresis concepts used in access selection similar to cellular IRAT, applied between Wi-Fi and cellular accesses can be used to mitigate this.
Labels:
4G,
Data Offload,
Handovers,
White Papers and Reports,
Wi-Fi
Tuesday, 29 October 2013
ANDSF: Evolution and Roaming with Hotspot 2.0
Access Network Discovery and Selection Function (ANDSF) is still evolving and with the introduction of Hotspot 2.0 (HS 2), there is a good possibility to provide seamless roaming from Cellular to Wi-Fi, Wi-Fi to Wi-Fi and Wi-Fi to Cellular.
There is a good paper (not very recent) by Alcatel-Lucent and BT that explains these roaming scenarios and other ANDSF policies related information very well. Its embedded below:
Labels:
Alcatel-Lucent,
ANDSF,
LTE,
Network Architecture,
Release 10,
Release 11,
Release 12,
Roaming,
Wi-Fi
Sunday, 27 October 2013
TDD-FDD Joint CA
From a recent NTT Docomo presentation (embedded below). Whereas right now 3GPP has only been working on FDD or TDD scenarios, this proposal is a combination of FDD as P-Cell and TDD as S-Cell. Inter-Technology carrier aggregation is another possible option. Anyway, the complete presentation is below.
LTE-Advanced Enhancements and Future Radio Access Toward 2020 and Beyond from Zahid Ghadialy
Updated on 29/10/2013
3GPP has already started working on this work item. See RP-131399 for details.
Updated on 29/10/2013
3GPP has already started working on this work item. See RP-131399 for details.
Labels:
Carrier Aggregation,
LTE,
LTE-Advanced,
NTT DoCoMo,
Release 12,
TD-LTE,
TDD
Tuesday, 22 October 2013
Korea Telecom ‘Route Decision System’ for midnight buses
Labels:
Big Data,
KT,
LTE & 5G World Series,
South Korea
Tuesday, 15 October 2013
What is Network Function Virtualisation (NFV)?
Software Defined Networking (SDN) and Network Function Virtualization (NFV) are the two recent buzzwords taking the telecoms market by storm. Every network vendor now has some kind of strategy to use this NFV and SDN to help operators save money. So what exactly is NFV? I found a good simple video by Spirent that explains this well. Here it is:
To add a description to this, I would borrow an explanation and a very good example from Wendy Zajack, Director Product Communications, Alcatel-Lucent in ALU blog:
Let’s take this virtualization concept to a network environment. For me cloud means I can get my stuff where ever I am and on any device – meaning I can pull out my smart phone, my iPad, my computer – and show my mom the latest pictures of her grand kids. I am not limited to only having one type of photo album I put my photos in – and only that. I can also show her both photos and videos together – and am not just limited to showing her the kids in one format and on one device.
Today in a telecom network is a lot of equipment that can only do one thing. These machines are focused on what they are do and they do it really well – this is why telecom providers are considered so ‘trusted.’ Back in the days of landline phones even when the power was out you could always make a call. These machines run alone with dedicated resources. These machines are made by various different vendors and speak various languages or ‘protocols’ to exchange information with each other when necessary. Some don’t even talk at all – they are just set-up and then left to run. So, every day your operator is running a mini United Nations and corralling that to get you to access all of your stuff. But it is a United Nations with a fixed number of seats, and with only a specific nation allowed to occupy a specific seat, with the seat left unused if there was a no-show. That is a lot of underutilized equipment that is tough and expensive to manage. It also has a shelf life of 15 years… while your average store-bought computer is doubling in speed every 18 months.
Virtualizing the network means the ability to run a variety of applications (or functions) on a standard piece of computing equipment, rather than on dedicated, specialized processors and equipment, to drive lower costs (more value), more re-use of the equipment between applications (more sharing), and a greater ability to change what is using the equipment to meet the changing user needs (more responsiveness). This has already started in enterprises as a way to control IT costs and improve the performance and of course way greener.
To give this a sports analogy – imagine if in American football instead of having specialists in all the different positions (QB, LB, RB, etc), you had a bunch of generalists who could play any position – you might only need a 22 or 33 man squad (2 or 3 players for every position) rather than the normal squad of 53. The management of your team would be much simpler as ‘one player fits all’ positions. It is easy to see how this would benefit a service provider – simplifying the procurement and management of the network elements (team) and giving them the ability to do more, with less.
Dimitris Mavrakis from Informa wrote an excellent summary from the IIR SDN and NFV conference in Informa blog here. Its worth reading his article but I want to highlight one section that shows how the operators think deployment would be done:
The speaker from BT provided a good roadmap for implementing SDN and NFV:
- Start with a small part of the network, which may not be critical for the operation of the whole. Perhaps introduce incremental capacity upgrades or improvements in specific and isolated parts of the network.
- Integrate with existing OSS/BSS and other parts of the network.
- Plan a larger-scale rollout so that it fits with the longer-term network strategy.
Deutsche Telecom is now considered to be deploying in the first phase, with a small trial in Hrvatski Telecom, its Croatian subsidiary, called Project Terrastream. BT, Telefonica, NTT Communications and other operators are at a similar stage, although DT is considered the first to deploy SDN and NFV for commercial network services beyond the data center.
Stage 2 in the roadmap is a far more complicated task. Integrating with existing components that may perform the same function but are not virtualized requires east-west APIs that are not clearly defined, especially when a network is multivendor. This is a very active point of discussion, but it remains to be seen whether Tier-1 vendors will be willing to openly integrate with their peers and even smaller, specialist vendors. OSS/BSS is also a major challenge, where multivendor networks are controlled by multiple systems and introducing a new service may require risking several parameters in many of these OSS/BSS consoles. This is another area that is not likely to change rapidly but rather in small, incremental steps.
The final stage is perhaps the biggest barrier due to the financial commitment and resources required. Long-term strategy may translate to five or even 10 years ahead – when networks are fully virtualized – and the economic environment may not allow such bold investments. Moreover, it is not clear if SDN and NFV guarantee new services and revenues outside the data center or operator cloud. If they do not, both technologies – and similar IT concepts – are likely to be deployed incrementally and replace equipment that reaches end-of-life. Cost savings in the network currently do not justify forklift upgrades or the replacement of adequately functional network components.
There is also a growing realization that bare-metal platforms (i.e., the proprietary hardware-based platforms that power today’s networks) are here to stay for several years. This hardware has been customized and adapted for use in telecom networks, allowing high performance for radio, core, transport, fixed and optical networks. Replacing these high-capacity components with virtualized ones is likely to affect performance significantly and operators are certainly not willing to take the risk of disrupting the operation of their network.
A major theme at the conference was that proprietary platforms (particularly ATCA) will be replaced by common off-the-shelf (COTS) hardware. ATCA is a hardware platform designed specifically for telecoms, but several vendors have adapted the platform to their own cause, creating fragmentation, incompatibility and vendor lock-in. Although ATCA is in theory telecoms-specific COTS, proprietary extensions have forced operators to turn to COTS, which is now driven by IT vendors, including Intel, HP, IBM, Dell and others.
ETSI has just published first specifications on NFV. Their press release here says:
ETSI has published the first five specifications on Network Functions Virtualisation (NFV). This is a major milestone towards the use of NFV to simplify the roll-out of new network services, reduce deployment and operational costs and encourage innovation.
These documents clearly identify an agreed framework and terminology for NFV which will help the industry to channel its efforts towards fully interoperable NFV solutions. This in turn will make it easier for network operators and NFV solutions providers to work together and will facilitate global economies of scale.
The IT and Network industries are collaborating in ETSI's Industry Specification Group for Network Functions Virtualisation (NFV ISG) to achieve a consistent approach and common architecture for the hardware and software infrastructure needed to support virtualised network functions. Early NFV deployments are already underway and are expected to accelerate during 2014-15. These new specifications have been produced in less than 10 months to satisfy the high industry demand – NFV ISG only began work in January 2013.
NFV ISG was initiated by the world's leading telecoms network operators. The work has attracted broad industry support and participation has risen rapidly to over 150 companies of all sizes from all over the world, including network operators, telecommunication equipment vendors, IT vendors and technology providers. Like all ETSI standards, these NFV specifications have been agreed by a consensus of all those involved.
The five published documents (which are publicly available via www.etsi.org/nfv) include four ETSI Group Specifications (GSs) designed to align understanding about NFV across the industry. They cover NFV use cases, requirements, the architectural framework, and terminology. The fifth GS defines a framework for co-ordinating and promoting public demonstrations of Proof of Concept (PoC) platforms illustrating key aspects of NFV. Its objective is to encourage the development of an open ecosystem by integrating components from different players.
Work is continuing in NFV ISG to develop further guidance to industry, and more detailed specifications are scheduled for 2014. In addition, to avoid the duplication of effort and to minimise fragmentation amongst multiple standards development organisations, NFV ISG is undertaking a gap analysis to identify what additional work needs to be done, and which bodies are best placed to do it.
The ETSI specifications are available at: http://www.etsi.org/technologies-clusters/technologies/nfvThe first document that shows various use cases is embedded below:
Sunday, 13 October 2013
Handset Antenna Design
Came across this presentation on Handset Antenna design from a recent Cambridge Wireless event here. Its interesting to see how the antenna technology has evolved and is still evolving. Another recent whitepaper from 4G Americas on meeting the 1000x challenge (here) showed how the different wavelengths are affecting the antenna design.
Maybe its better to move to higher frequencies from the handset design point of view. Anyway, the Cambridge Wireless presentation is embedded below:
Subscribe to:
Posts (Atom)