Click here to close now.

Welcome!

Virtualization Authors: Pat Romanski, Carmen Gonzalez, Clinton Wolfe, John M. Hawkins, Leo Reiter

News Feed Item

Fujitsu Develops New Data Transfer Protocol Enabling Improved Transmissions Speeds

- Software-only approach enables over 30 times improvement in file transfer speeds between Japan and the US
- Reduces virtual desktop operating latency to less than 1/6 of previous levels

Kawasaki, Japan, Jan 29, 2013 - (JCN Newswire) - Fujitsu Laboratories Limited today announced the development of a new data transfer protocol that, by taking a software-only approach, can significantly improve the performance of file transfers, virtual desktops and other various communications applications.

Conventionally, when using transmission control protocol (TCP)(1) - the standard protocol employed in communications applications - in a low-quality communications environment, such as when connected to a wireless network or during times of line congestion, data loss (packet loss) can occur, leading to significant drops in transmission performance due to increased latency from having to retransmit data.

To address this problem, Fujitsu Laboratories has succeeded at a software-only approach, developing: 1) A new protocol that incorporates an efficient proprietarily developed retransmission method based on user datagram protocol (UDP)(2), an optimized way to deliver streaming media able to reduce latency resulting from data retransmission when packet loss occurs; 2) Control technology that addresses the problem of UDP transmissions consuming excess bandwidth by performing a real-time measurement of available network bandwidth and securing an optimal amount of communications bandwidth without overwhelming TCP's share of the bandwidth; and 3) Technology that, by employing the new protocol, makes it possible to easily speed up existing TCP applications without having to modify them.

Through a simple software installation, the new technology will make it possible to speed up TCP applications that previously required costly specialized hardware, and it can also be easily incorporated into mobile devices and other kinds of equipment. Moreover, compared with TCP, the technology enables a greater than 30 times improvement in file transfer speeds between Japan and the US, in addition to reducing virtual desktop operating latency to less than 1/6 of previous levels. This, in turn, is expected to make it easier to take advantage of various applications employing international communication lines and wireless networks which are anticipated to become increasingly widespread.

Background

With the increased popularity of mobile devices and cloud services in recent years, a wide range of applications have begun to utilize communications capabilities. In many applications, such as file transfer, virtual desktop, and other communications applications, TCP is employed as a standard communications protocol. One issue with TCP is that data loss (packet loss) can occur in low-quality communications environments, resulting in significant drops in transmission performance (reduced throughput and higher latency) due to increased latency from having to retransmit data. In the future, it is expected that there will be greater opportunities to take advantage of international communications lines and wireless networks, making it necessary to ensure that transmission performance does not drop even when connected to a low-quality communications environment.

Technological Challenges

Currently, one well-known method of speeding up application transmission speeds in low-quality communications environments is to employ specialized acceleration hardware. This kind of specialized equipment, however, is expensive and bulky, making it difficult to incorporate into mobile devices. High-speed transmission methods for transferring files using software-based acceleration also exist, but to support a variety of existing TCP applications using these methods, it has been necessary to make modifications to the traffic processing components of each application.

Newly Developed Technology

By developing a proprietary software-based transfer protocol, Fujitsu Laboratories has succeeded in significantly improving the throughput and operating latency of existing TCP applications.

Key features of the new technology are as follows:

1) New protocol improves throughput and latency in low-quality communications environments

Fujitsu has developed a new protocol that incorporates a proprietarily developed and efficient retransmission method based on UDP, a protocol optimized for delivering streaming media. As a result, the new protocol is able to reduce latency resulting from data retransmission when packet loss occurs. The protocol can quickly distinguish between lost packets and packets that have not yet arrived at their destination, thereby preventing unnecessary retransmissions and latency from occurring. By incorporating the new protocol as a software add-on to UDP, it is possible to maintain the high speeds typical of UDP while avoiding packet loss and packets being sent in reverse order, UDP's main weaknesses. This, in turn, has enabled improvements in packet delivery and latency. In a comparison with standard TCP, the new protocol achieved a throughput increase of over 30 times during a simulated file transfer between Japan and the US, and operating packet delivery latency was reduced to less than 1/6 of previous levels.

2) Communications bandwidth control technology using real-time measurement of available network bandwidth

Fujitsu Laboratories developed a control technology that, by performing real-time measurement of available network bandwidth, can secure an optimal amount of communications bandwidth without overwhelming the share of bandwidth used by other TCP communications in a mixed TCP environment. For example, when other TCP communications are using relatively little bandwidth, the bandwidth share for the new protocol will increase, and when other TCP communications are taking up a higher percentage of bandwidth, the new protocol will use a smaller share.

3) Technology for accelerating existing TCP applications without any modifications

Fujitsu Laboratories has developed a technology that automatically converts TCP traffic standard for a wide variety of applications into the new protocol described in (1) above. This makes it possible to significantly improve the speed of a host of existing applications, including file transfer applications, virtual desktop applications, and web browsing applications, all without the need for any modifications.

Results

The use of the new technology is expected to speed up the performance of a wide range of communications applications employing international communication lines and wireless networks which are anticipated to become widely used more and more. For instance, the technology can help speed up web browsing and file download speeds in mobile communications environments where there is deterioration due to building obstructions or movement. In addition, the technology can improve data transfer speeds between datacenters in Japan and the US. It is also expected to help improve the usability of virtual desktops when accessing a virtual desktop located on a remote server using a low-quality communications environment.

Future Development

During fiscal 2013, Fujitsu Laboratories aims to commercialize the new technology as a communications middleware solution for improving communications speeds without having to modify existing TCP applications.

(1) Transmission Control Protocol (TCP): An Internet protocol that guarantees data delivery through a retransmission mechanism.
(2) User Datagram Protocol (UDP): An Internet protocol that does not guarantee data delivery.

About Fujitsu Limited

Fujitsu is the leading Japanese information and communication technology (ICT) company offering a full range of technology products, solutions and services. Over 170,000 Fujitsu people support customers in more than 100 countries. We use our experience and the power of ICT to shape the future of society with our customers. Fujitsu Limited (TSE:6702) reported consolidated revenues of 4.5 trillion yen (US$54 billion) for the fiscal year ended March 31, 2012. For more information, please see www.fujitsu.com.



Source: Fujitsu Limited

Contact:
Fujitsu Limited
Public and Investor Relations
www.fujitsu.com/global/news/contacts/
+81-3-3215-5259


Copyright 2013 JCN Newswire. All rights reserved. www.japancorp.net

More Stories By JCN Newswire

Copyright 2008 JCN Newswire. All rights reserved. Republication or redistribution of JCN Newswire content is expressly prohibited without the prior written consent of JCN Newswire. JCN Newswire shall not be liable for any errors or delays in the content, or for any actions taken in reliance thereon.

@ThingsExpo Stories
The Internet of Things (IoT) promises to evolve the way the world does business; however, understanding how to apply it to your company can be a mystery. Most people struggle with understanding the potential business uses or tend to get caught up in the technology, resulting in solutions that fail to meet even minimum business goals. In his session at @ThingsExpo, Jesse Shiah, CEO / President / Co-Founder of AgilePoint Inc., showed what is needed to leverage the IoT to transform your business. He discussed opportunities and challenges ahead for the IoT from a market and technical point of vie...
IoT is still a vague buzzword for many people. In his session at @ThingsExpo, Mike Kavis, Vice President & Principal Cloud Architect at Cloud Technology Partners, discussed the business value of IoT that goes far beyond the general public's perception that IoT is all about wearables and home consumer services. He also discussed how IoT is perceived by investors and how venture capitalist access this space. Other topics discussed were barriers to success, what is new, what is old, and what the future may hold. Mike Kavis is Vice President & Principal Cloud Architect at Cloud Technology Pa...
Hadoop as a Service (as offered by handful of niche vendors now) is a cloud computing solution that makes medium and large-scale data processing accessible, easy, fast and inexpensive. In his session at Big Data Expo, Kumar Ramamurthy, Vice President and Chief Technologist, EIM & Big Data, at Virtusa, will discuss how this is achieved by eliminating the operational challenges of running Hadoop, so one can focus on business growth. The fragmented Hadoop distribution world and various PaaS solutions that provide a Hadoop flavor either make choices for customers very flexible in the name of opti...
The true value of the Internet of Things (IoT) lies not just in the data, but through the services that protect the data, perform the analysis and present findings in a usable way. With many IoT elements rooted in traditional IT components, Big Data and IoT isn’t just a play for enterprise. In fact, the IoT presents SMBs with the prospect of launching entirely new activities and exploring innovative areas. CompTIA research identifies several areas where IoT is expected to have the greatest impact.
Advanced Persistent Threats (APTs) are increasing at an unprecedented rate. The threat landscape of today is drastically different than just a few years ago. Attacks are much more organized and sophisticated. They are harder to detect and even harder to anticipate. In the foreseeable future it's going to get a whole lot harder. Everything you know today will change. Keeping up with this changing landscape is already a daunting task. Your organization needs to use the latest tools, methods and expertise to guard against those threats. But will that be enough? In the foreseeable future attacks w...
Disruptive macro trends in technology are impacting and dramatically changing the "art of the possible" relative to supply chain management practices through the innovative use of IoT, cloud, machine learning and Big Data to enable connected ecosystems of engagement. Enterprise informatics can now move beyond point solutions that merely monitor the past and implement integrated enterprise fabrics that enable end-to-end supply chain visibility to improve customer service delivery and optimize supplier management. Learn about enterprise architecture strategies for designing connected systems tha...
Dale Kim is the Director of Industry Solutions at MapR. His background includes a variety of technical and management roles at information technology companies. While his experience includes work with relational databases, much of his career pertains to non-relational data in the areas of search, content management, and NoSQL, and includes senior roles in technical marketing, sales engineering, and support engineering. Dale holds an MBA from Santa Clara University, and a BA in Computer Science from the University of California, Berkeley.
Wearable devices have come of age. The primary applications of wearables so far have been "the Quantified Self" or the tracking of one's fitness and health status. We propose the evolution of wearables into social and emotional communication devices. Our BE(tm) sensor uses light to visualize the skin conductance response. Our sensors are very inexpensive and can be massively distributed to audiences or groups of any size, in order to gauge reactions to performances, video, or any kind of presentation. In her session at @ThingsExpo, Jocelyn Scheirer, CEO & Founder of Bionolux, will discuss ho...
The cloud is now a fact of life but generating recurring revenues that are driven by solutions and services on a consumption model have been hard to implement, until now. In their session at 16th Cloud Expo, Ermanno Bonifazi, CEO & Founder of Solgenia, and Ian Khan, Global Strategic Positioning & Brand Manager at Solgenia, will discuss how a top European telco has leveraged the innovative recurring revenue generating capability of the consumption cloud to enable a unique cloud monetization model to drive results.
As organizations shift toward IT-as-a-service models, the need for managing and protecting data residing across physical, virtual, and now cloud environments grows with it. CommVault can ensure protection &E-Discovery of your data – whether in a private cloud, a Service Provider delivered public cloud, or a hybrid cloud environment – across the heterogeneous enterprise. In his session at 16th Cloud Expo, Randy De Meno, Chief Technologist - Windows Products and Microsoft Partnerships, will discuss how to cut costs, scale easily, and unleash insight with CommVault Simpana software, the only si...
Analytics is the foundation of smart data and now, with the ability to run Hadoop directly on smart storage systems like Cloudian HyperStore, enterprises will gain huge business advantages in terms of scalability, efficiency and cost savings as they move closer to realizing the potential of the Internet of Things. In his session at 16th Cloud Expo, Paul Turner, technology evangelist and CMO at Cloudian, Inc., will discuss the revolutionary notion that the storage world is transitioning from mere Big Data to smart data. He will argue that today’s hybrid cloud storage solutions, with commodity...
Every innovation or invention was originally a daydream. You like to imagine a “what-if” scenario. And with all the attention being paid to the so-called Internet of Things (IoT) you don’t have to stretch the imagination too much to see how this may impact commercial and homeowners insurance. We’re beyond the point of accepting this as a leap of faith. The groundwork is laid. Now it’s just a matter of time. We can thank the inventors of smart thermostats for developing a practical business application that everyone can relate to. Gone are the salad days of smart home apps, the early chalkb...
Cloud data governance was previously an avoided function when cloud deployments were relatively small. With the rapid adoption in public cloud – both rogue and sanctioned, it’s not uncommon to find regulated data dumped into public cloud and unprotected. This is why enterprises and cloud providers alike need to embrace a cloud data governance function and map policies, processes and technology controls accordingly. In her session at 15th Cloud Expo, Evelyn de Souza, Data Privacy and Compliance Strategy Leader at Cisco Systems, will focus on how to set up a cloud data governance program and s...
Roberto Medrano, Executive Vice President at SOA Software, had reached 30,000 page views on his home page - http://RobertoMedrano.SYS-CON.com/ - on the SYS-CON family of online magazines, which includes Cloud Computing Journal, Internet of Things Journal, Big Data Journal, and SOA World Magazine. He is a recognized executive in the information technology fields of SOA, internet security, governance, and compliance. He has extensive experience with both start-ups and large companies, having been involved at the beginning of four IT industries: EDA, Open Systems, Computer Security and now SOA.
The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information. In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focused on understanding how industrial data can create intelligence for industrial operations. Imagine ...
We certainly live in interesting technological times. And no more interesting than the current competing IoT standards for connectivity. Various standards bodies, approaches, and ecosystems are vying for mindshare and positioning for a competitive edge. It is clear that when the dust settles, we will have new protocols, evolved protocols, that will change the way we interact with devices and infrastructure. We will also have evolved web protocols, like HTTP/2, that will be changing the very core of our infrastructures. At the same time, we have old approaches made new again like micro-services...
Operational Hadoop and the Lambda Architecture for Streaming Data Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing and analyzing streaming data is the Lambda Architecture, representing a model of how to analyze rea...
SYS-CON Events announced today that Vitria Technology, Inc. will exhibit at SYS-CON’s @ThingsExpo, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Vitria will showcase the company’s new IoT Analytics Platform through live demonstrations at booth #330. Vitria’s IoT Analytics Platform, fully integrated and powered by an operational intelligence engine, enables customers to rapidly build and operationalize advanced analytics to deliver timely business outcomes for use cases across the industrial, enterprise, and consumer segments.
Today’s enterprise is being driven by disruptive competitive and human capital requirements to provide enterprise application access through not only desktops, but also mobile devices. To retrofit existing programs across all these devices using traditional programming methods is very costly and time consuming – often prohibitively so. In his session at @ThingsExpo, Jesse Shiah, CEO, President, and Co-Founder of AgilePoint Inc., discussed how you can create applications that run on all mobile devices as well as laptops and desktops using a visual drag-and-drop application – and eForms-buildi...
Containers and microservices have become topics of intense interest throughout the cloud developer and enterprise IT communities. Accordingly, attendees at the upcoming 16th Cloud Expo at the Javits Center in New York June 9-11 will find fresh new content in a new track called PaaS | Containers & Microservices Containers are not being considered for the first time by the cloud community, but a current era of re-consideration has pushed them to the top of the cloud agenda. With the launch of Docker's initial release in March of 2013, interest was revved up several notches. Then late last...