From the Wires
SCinet Research Sandbox Shows Off Groundbreaking Network Research @ SC12
By: Business Wire
Nov. 5, 2012 03:20 PM
This year, SC12 will not only showcase the next generation of high performance computing applications but it will also be home to seven innovative network research projects through a special program called the SCinet Research Sandbox (SRS).
SCinet is the primary high performance network infrastructure built each year for SC exhibitors to highlight their cutting edge computing applications and collaborations. As a key component of SCinet, the SRS is designed to enable researchers to experiment and demonstrate their ideas on innovative network architectures, applications and protocols in the unique live environment of the SCinet network. This year, the SRS will provide researchers with dedicated access to multiple 100 Gigabits per second wide area network links as well as a 10 Gigabit per second (Gbps) OpenFlow network testbed.
“In addition to supporting the extreme demands of the HPC-based demonstrations that have become the trademark of the conference, SCinet also seeks to foster and highlight developments in network research that will be necessary to support the next-generation of science applications,” said Brian Tierney, SRS co-chair for SC12 and head of ESnet’s Advanced Network Technologies Group. “Both 100 Gbps networking and OpenFlow have become some of the most influential networking technologies of this decade. SRS allows the community to showcase innovations on these platforms while in their infancy to demonstrate the impact they may have on the entire HPC community in the future.”
“Openflow has the potential to greatly improve applications that are necessary for advancements in HPC such as GridFTP and others. By allowing these applications to access the network in a highly intelligent and programmable manner, OpenFlow can significantly improve end-to-end network performance especially for bulk data transfers, which will be an increasingly difficult challenge in the era of data intensive science,” said Andrew Lee, SRS co-chair for SC12 and Principal Network Systems Engineer for the Global Research Network Operations Center at Indiana University. “Demonstrations like those being supported by the SCinet Research Sandbox are laying the groundwork for these critical advancements as well as demonstrating to the community - in a tangible way - the possibilities that Openflow provides."
Seven projects have been selected as part of the SRS program, all of which will be showcased as part of the Technical Program and will be demonstrating their research in several exhibit booths during the conference. For detailed information on the projects and their presentations visit: http://sc12.supercomputing.org/content/scinet-research-sandbox
2012 SRS projects include:
Efficient LHC Data Distribution across 100Gbps Networks
The analysis of data leading to the recent discoveries at the Large Hadron Collider (LHC) produces data flows of more than 100 Petabytes per year, and increasingly relies on the efficient movement of data sets between the globally distributed computing sites. The team will demonstrate the state-of-the-art data movement tools, as enabling technology for high-throughput data distribution over 100Gbps WAN circuits. The demo will interconnect three major LHC Tier-2 computing sites and the SC12 show floor (booth 809) using 100Gbps technology.
Collaborating organizations: University of Victoria,
University of Michigan, California Institute of Technology, Vanderbilt
University, Internet2, ESnet, CENIC, Starlight, PacWave/GLORIAD,
KNU/KISTI as well as vendors: Alcatel Lucent, Ciena, Cisco, Juniper
Networks, Mellanox. Dell-Force10, SuperMicro, ASA Micro Systems, Data
Direct Network, Fusion-IO, PADTECH
Exploiting Network Parallelism for Improving Data Transfer Performance
The task of scientific bulk data movement, e.g. migrating collected results from the instrumentation to the processing and storage facilities, is hampered by a lack of available network resources. Traditional R&E connectivity can be congested on portions of an end-to-end path causing degradation of overall performance. This SRS project will explore dynamic network control to facilitate efficient bulk data movement, combining opportunistic use of "traditional" networks with dedicated reservations over virtual circuits and OpenFlow enabled resources. The GridFTP application has been instrumented with the eXtensible Session Protocol (XSP), an intelligent system capable of controlling programmable networks. The project intends to show end-to-end performance improvement between the SC12 conference and campuses involved in the DYNES project, through a combination of regular connectivity, dynamic bandwidth allocations, TCP acceleration, and operations using multiple paths.
Collaborating organizations: Indiana University, Lawrence
Berkeley National Laboratory, Argonne National Laboratory and Internet2
Multipathing with MPTCP and OpenFlow
This demo shows several emerging network technologies and how these can be used in big data transfers between data centers. In this demonstration traffic is sent simultaneously across multiple OpenFlow controlled paths between Geneva and Salt Lake City. The congestion control mechanism of Multipath TCP (MPTCP) favours the least congested paths and ensures that the load balancing across the paths is always optimal.
Collaborating organizations: SURFnet, SARA, iCAIR and the
California Institute of Technology
The emerging era of “Big Science” demands the highest possible network performance. End-to-end circuit automation and workflow-driven customization are two essential capabilities needed for networks to scale to meet this challenge. This demonstration showcases how combining software-defined networking techniques with virtual circuits capabilities can transform the network into a dynamic, customer-configurable virtual switch. In doing so, users are able to rapidly customize network capabilities to meet their unique workflows with little to no configuration effort. The demo also highlights how the network can be automated to support multiple collaborations in parallel.
Collaborating organizations: ESnet, Ciena
OpenFlow Enabled Hadoop over Local and Wide Area Clusters
The Hadoop Distributed File Systems and Hadoop's implementation of MapReduce is one of the most widely used platforms for data intensive computing. The shuffle and sort phases of a MapReduce computation often saturate network links to nodes and the reduce phase of the computation must wait for data. This study explores the use of OpenFlow to the control network configuration for different flows to thereby provide different network characteristics for different categories of Hadoop traffic.
Collaborating organizations: Infoblox, Laboratory for
Advanced Computing, University of Chicago, Open Cloud Consortium,
Livermore Valley Open Campus
Large-scale data intensive science requires global collaboration and sophisticated high capacity data management. The emergence of more flexible networking, for example, using techniques based on OpenFlow, provides opportunities to address these issues because these techniques enable a high degree of network customization and dynamic provisioning. These techniques enable large-scale facilities to be created that can be used to prototype new architecture, services, protocols, and technologies. A number of research organizations from several countries have designed and implemented a persistent international experimental research facility that can be used to prototype, investigate, and test network innovations for large-scale global science. For SC12, this international experimental network facility will be extended to from sites across the world to the conference showfloor, and it will be used to support several testbeds and to showcase a series of complementary demonstrations.
Collaborating organizations: iCAIR,
Northwestern University; National Center for High-Performance Computing,
Taiwan; University of Applied Sciences, Taiwan; National Cheng-Kung
University, Taiwan; SARA, The Netherlands, California Institute of
Technology/CERN; SURFnet, The Netherlands.
Scalable Cyber-Security for Terabit Cloud Computing
Reservoir Labs will demonstrate R-Scope®, a scalable, high-performance network packet inspection technology that forms the core of a new generation of Intrusion Detection Systems enabling the construction and deployment of cyber security infrastructures scaling to terabit per second ingest bandwidths. This scalability is enabled by the use of low- power and high-performance manycore network processors combined with Reservoir’s enhancements to Bro. The innovative R-Scope PACE-T appliance, implemented on a 1U Tilera TILExtreme-Gx platform, will demonstrate the capacity to perform cyber-security analysis at 80Gbps, by combining cyber-security aware front-end network traffic load balancing tightly coupled with the full back-end analytic power of Bro. This fully-programmable platform incorporates the full Bro semantics into the appliance’s load-balancing front-end and the back-end analytic nodes.
Collaborating organizations: Reservoir Labs, SCinet
This press release is available online at http://sc12.supercomputing.org/content/scinet-research-sandbox-shows-groundbreaking-network-research.
SC12, sponsored by the ACM (Association for Computing Machinery) and the IEEE Computer Society, offers a complete technical education program and exhibition to showcase the many ways high performance computing, networking, storage and analysis lead to advances in scientific discovery, research, education and commerce. This premier international conference includes a globally attended technical program, workshops, tutorials, a world class exhibit area, demonstrations and opportunities for hands-on learning.
Photos/Multimedia Gallery Available: http://www.businesswire.com/cgi-bin/mmg.cgi?eid=50466050&lang=en
Latest Cloud Developer Stories
Subscribe to the World's Most Powerful Newsletters
Subscribe to Our Rss Feeds & Get Your SYS-CON News Live!
SYS-CON Featured Whitepapers
Most Read This Week
Breaking Cloud Computing News