IDIO: Orchestrating inbound network data on server processors

Mohammad Alian, Jongmin Shin, Ki Dong Kang, Ren Wang, Alexandros Daglis, Daehoon Kim, Nam Sung Kim

Research output: Contribution to journalArticlepeer-review

Abstract

Network bandwidth demand in datacenters is doubling every 12 to 15 months. In response to this demand, high-bandwidth network interface cards, each capable of transferring 100s of Gigabits of data per second, are making inroads into the servers of next-generation datacenters. Such unprecedented data delivery rates on server endpoints raise new challenges, as inbound network traffic placement decisions within the memory hierarchy have a direct impact on end-to-end performance. Modern server-class Intel processors leverage DDIO technology to steer all inbound network data into the last-level cache (LLC), regardless of the network traffic's nature. This static data placement policy is suboptimal, both from a performance and an energy efficiency standpoint. In this work, we design IDIO, a framework that - unlike DDIO - dynamically decides where to place inbound network traffic within a server's multi-level memory hierarchy. IDIO dynamically monitors system behavior and distinguishes between different traffic classes to determine and periodically re-evaluate the best placement location for each flow: LLC, mid-level (L2) cache or DRAM. Our results show that IDIO increases a server's maximum sustainable load by up to ∼ 33.3% across various network functions.

Original languageEnglish (US)
Article number9294125
Pages (from-to)30-33
Number of pages4
JournalIEEE Computer Architecture Letters
Volume20
Issue number1
DOIs
StatePublished - Jan 1 2021

Keywords

  • Cache
  • data direct I/O
  • datacenters
  • network

ASJC Scopus subject areas

  • Hardware and Architecture

Fingerprint Dive into the research topics of 'IDIO: Orchestrating inbound network data on server processors'. Together they form a unique fingerprint.

Cite this