Application-transparent near-memory processing architecture with memory channel network

The physical memory capacity of servers is expected to increase drastically with the deployment of the forthcoming non-volatile memory technologies. This is a welcomed improvement for the emerging data-intensive applications. For such servers to be cost-effective, nonetheless, we must cost-effective...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:2018 51st Annual IEEE ACM International Symposium on Microarchitecture (MICRO) S. 802 - 814
Hauptverfasser: Alian, Mohammad, Min, Seung Won, Asgharimoghaddam, Hadi, Dhar, Ashutosh, Wang, Dong Kai, Roewer, Thomas, McPadden, Adam, O'Halloran, Oliver, Chen, Deming, Xiong, Jinjun, Kim, Daehoon, Hwu, Wen-mei, Kim, Nam Sung
Format: Tagungsbericht
Sprache:Englisch
Veröffentlicht: Piscataway, NJ, USA IEEE Press 20.10.2018
IEEE
Schriftenreihe:ACM Conferences
Schlagworte:
ISBN:9781538662403, 153866240X
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Abstract The physical memory capacity of servers is expected to increase drastically with the deployment of the forthcoming non-volatile memory technologies. This is a welcomed improvement for the emerging data-intensive applications. For such servers to be cost-effective, nonetheless, we must cost-effectively increase compute throughput and memory bandwidth commensurate with the increase in memory capacity without compromising the application readiness. Tackling this challenge, we present Memory Channel Network (MCN) architecture in this paper. Specifically, first, we propose an MCN DIMM, an extension of a buffered DIMM where a small but capable processor called MCN processor is integrated with a buffer device on the DIMM for near-memory processing. Second, we implement device drivers to give the host and MCN processors in a server an illusion that they are independent heterogeneous nodes connected through an Ethernet link. These allow the host and MCN processors in a server to run a given data-intensive application together based on popular distributed computing frameworks such as MPI and Spark without any change in the host processor hardware and its application software, while offering the benefits of high-bandwidth and low-latency communication between the host and MCN processors over the memory channels. As such, MCN can serve as an application-transparent framework which can seamlessly unify the near-memory processing within a server and the distributed computing across such servers for data-intensive applications. Our simulation running the full software stack shows that a server with 8 MCN DIMMs offers 4.56 x higher throughput and consume 47.5% less energy than a cluster with 9 conventional nodes connected through Ethernet links, as it facilitates up to 8.17 x higher aggregate DRAM bandwidth utilization. Lastly, we demonstrate the feasibility of MCN with an IBM POWER8 system and an experimental buffered DIMM.
AbstractList The physical memory capacity of servers is expected to increase drastically with deployment of the forthcoming non-volatile memory technologies. This is a welcomed improvement for emerging data-intensive applications. For such servers to be cost-effective, nonetheless, we must cost-effectively increase compute throughput and memory bandwidth commensurate with the increase in memory capacity without compromising application readiness. Tackling this challenge, we present Memory Channel Network (MCN) architecture in this paper. Specifically, first, we propose an MCN DIMM, an extension of a buffered DIMM where a small but capable processor called MCN processor is integrated with a buffer device on the DIMM for near-memory processing. Second, we implement device drivers to give the host and MCN processors in a server an illusion that they are independent heterogeneous nodes connected through an Ethernet link. These allow the host and MCN processors in a server to run a given data-intensive application together based on popular distributed computing frameworks such as MPI and Spark without any change in the host processor hardware and its application software, while offering the benefits of high-bandwidth and low-latency communications between the host and the MCN processors over memory channels. As such, MCN can serve as an application-transparent framework which can seamlessly unify near-memory processing within a server and distributed computing across such servers for data-intensive applications. Our simulation running the full software stack shows that a server with 8 MCN DIMMs offers 4.56X higher throughput and consume 47.5% less energy than a cluster with 9 conventional nodes connected through Ethernet links, as it facilitates up to 8.17X higher aggregate DRAM bandwidth utilization. Lastly, we demonstrate the feasibility of MCN with an IBM POWER8 system and an experimental buffered DIMM.
The physical memory capacity of servers is expected to increase drastically with the deployment of the forthcoming non-volatile memory technologies. This is a welcomed improvement for the emerging data-intensive applications. For such servers to be cost-effective, nonetheless, we must cost-effectively increase compute throughput and memory bandwidth commensurate with the increase in memory capacity without compromising the application readiness. Tackling this challenge, we present Memory Channel Network (MCN) architecture in this paper. Specifically, first, we propose an MCN DIMM, an extension of a buffered DIMM where a small but capable processor called MCN processor is integrated with a buffer device on the DIMM for near-memory processing. Second, we implement device drivers to give the host and MCN processors in a server an illusion that they are independent heterogeneous nodes connected through an Ethernet link. These allow the host and MCN processors in a server to run a given data-intensive application together based on popular distributed computing frameworks such as MPI and Spark without any change in the host processor hardware and its application software, while offering the benefits of high-bandwidth and low-latency communication between the host and MCN processors over the memory channels. As such, MCN can serve as an application-transparent framework which can seamlessly unify the near-memory processing within a server and the distributed computing across such servers for data-intensive applications. Our simulation running the full software stack shows that a server with 8 MCN DIMMs offers 4.56 x higher throughput and consume 47.5% less energy than a cluster with 9 conventional nodes connected through Ethernet links, as it facilitates up to 8.17 x higher aggregate DRAM bandwidth utilization. Lastly, we demonstrate the feasibility of MCN with an IBM POWER8 system and an experimental buffered DIMM.
Author McPadden, Adam
O'Halloran, Oliver
Roewer, Thomas
Asgharimoghaddam, Hadi
Dhar, Ashutosh
Alian, Mohammad
Kim, Nam Sung
Hwu, Wen-mei
Wang, Dong Kai
Chen, Deming
Min, Seung Won
Xiong, Jinjun
Kim, Daehoon
Author_xml – sequence: 1
  givenname: Mohammad
  surname: Alian
  fullname: Alian, Mohammad
  organization: University of Illinois, Urbana-Champaign
– sequence: 2
  givenname: Seung Won
  surname: Min
  fullname: Min, Seung Won
  organization: University of Illinois, Urbana-Champaign
– sequence: 3
  givenname: Hadi
  surname: Asgharimoghaddam
  fullname: Asgharimoghaddam, Hadi
  organization: University of Illinois, Urbana-Champaign
– sequence: 4
  givenname: Ashutosh
  surname: Dhar
  fullname: Dhar, Ashutosh
  organization: University of Illinois, Urbana-Champaign
– sequence: 5
  givenname: Dong Kai
  surname: Wang
  fullname: Wang, Dong Kai
  organization: University of Illinois, Urbana-Champaign
– sequence: 6
  givenname: Thomas
  surname: Roewer
  fullname: Roewer, Thomas
  organization: IBM Research
– sequence: 7
  givenname: Adam
  surname: McPadden
  fullname: McPadden, Adam
  organization: IBM Systems
– sequence: 8
  givenname: Oliver
  surname: O'Halloran
  fullname: O'Halloran, Oliver
  organization: IBM Systems
– sequence: 9
  givenname: Deming
  surname: Chen
  fullname: Chen, Deming
  organization: University of Illinois, Urbana-Champaign
– sequence: 10
  givenname: Jinjun
  surname: Xiong
  fullname: Xiong, Jinjun
  organization: IBM Research
– sequence: 11
  givenname: Daehoon
  surname: Kim
  fullname: Kim, Daehoon
  organization: DGIST
– sequence: 12
  givenname: Wen-mei
  surname: Hwu
  fullname: Hwu, Wen-mei
  organization: University of Illinois, Urbana-Champaign
– sequence: 13
  givenname: Nam Sung
  surname: Kim
  fullname: Kim, Nam Sung
  organization: University of Illinois, Urbana-Champaign
BookMark eNqNkDtPwzAUhY0ACSjdkViyMqS9jhPbGauKR6WiSgjEaNnODTVtncgxqvrvSR8TE9PV0TnfHb4bcuEbj4TcURhRCuX4dTZ9W4wyoHIEAALOyLAUkhZMcp7lwM7_5Csy7LrvfppxySTwa_I5adu1szq6xqcxaN-1OqCPiUcd0g1umrBL2tBY7DrnvxId7NJFtPEnYLJ1cZmcNnapvcd1z8VtE1a35LLW6w6HpzsgH0-P79OXdL54nk0n81RTnsc0Bw2G8UxAQWvMpGFMVzVURgCnRlgjSioNSM1AFBLyrMx5LViJpqoqW3I2IPfHvw4RVRvcRoedkoXICyn6dnxstd0o0zSrTlFQe3XqoE7t1amDOmWCw7onHv5LsF9KznCj
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/MICRO.2018.00070
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE/IET Electronic Library (IEL) (UW System Shared)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList

Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Computer Science
EISBN 9781538662403
153866240X
EndPage 814
ExternalDocumentID 8574587
Genre orig-research
GroupedDBID 6IE
6IF
6IL
6IN
AAJGR
ABLEC
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
IEGSK
OCL
RIB
RIC
RIE
RIL
AAWTH
ID FETCH-LOGICAL-a164t-40a0b3627051fe28b33adf0db7061b7cb7918b08a30758042946f739ebdddc963
IEDL.DBID RIE
ISBN 9781538662403
153866240X
IngestDate Wed Aug 27 02:46:37 EDT 2025
Sat Jun 15 16:36:41 EDT 2024
Wed Jan 31 06:41:02 EST 2024
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MeetingName MICRO-51: The 51st Annual IEEE/ACM International Symposium on Microarchitecture
MergedId FETCHMERGED-LOGICAL-a164t-40a0b3627051fe28b33adf0db7061b7cb7918b08a30758042946f739ebdddc963
PageCount 13
ParticipantIDs ieee_primary_8574587
acm_books_10_1109_MICRO_2018_00070_brief
acm_books_10_1109_MICRO_2018_00070
PublicationCentury 2000
PublicationDate 20181020
2018-Oct
PublicationDateYYYYMMDD 2018-10-20
2018-10-01
PublicationDate_xml – month: 10
  year: 2018
  text: 20181020
  day: 20
PublicationDecade 2010
PublicationPlace Piscataway, NJ, USA
PublicationPlace_xml – name: Piscataway, NJ, USA
PublicationSeriesTitle ACM Conferences
PublicationTitle 2018 51st Annual IEEE ACM International Symposium on Microarchitecture (MICRO)
PublicationTitleAbbrev MICRO
PublicationYear 2018
Publisher IEEE Press
IEEE
Publisher_xml – name: IEEE Press
– name: IEEE
SSID ssj0002683806
Score 2.2941432
Snippet The physical memory capacity of servers is expected to increase drastically with the deployment of the forthcoming non-volatile memory technologies. This is a...
The physical memory capacity of servers is expected to increase drastically with deployment of the forthcoming non-volatile memory technologies. This is a...
SourceID ieee
acm
SourceType Publisher
StartPage 802
SubjectTerms Application Transparent
Bandwidth
Buffer Device
Computer architecture
Distributed Systems
DRAM
Ethernet
General and reference
General and reference -- Cross-computing tools and techniques
General and reference -- Cross-computing tools and techniques -- Performance
Hardware
Memory Channel
Mobile Processors
Near Memory Processing
Processing In Memory
Random access memory
Servers
TCP IP
Title Application-transparent near-memory processing architecture with memory channel network
URI https://ieeexplore.ieee.org/document/8574587
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV3fS8MwED624YNPUzdx_iKIDz5Yl60_kjwOcfjiHKK4t9CmKQy0k60T_O-9S2unIIhvpSQp_cjx5e6-ywGcR3jq8C03XjDgvod8bTwVpBTIVxk6XEhCceyaTYjJRM5matqAy7oWxlrrxGf2ih5dLj9dmDWFyvoyFEEoRROaQkRlrVYdTxlG0pe8zkRy1cffeLgn8RapJTl1I27G5vVHFxVHIuP2_z6_A91NNR6b1jyzCw2b70H7qx0Dq6yzA8-jTTLaKy8tp0qvgk1wN3t3JKn9YFVhAC7ERt9yCIzisawaQxUHuX3BeU4i3oWn8c3j9a1X9U3wYnR-CnQJY54gMQk0uMwOZeL7cZrxNBFI3okwiVADmXAZo32HkhgpiDLhK5ukaWrQIvehlS9yewAsSq3xfRXawKqA2pIToQfcRFlo8MAb9uAMMdXkEKy08ye40g54TcBrB3wPLv4epJPl3GY96BDu-q28akNXkB_-_voItmmBUlt3DK1iubYnsGXei_lqeer2xyc4SLer
linkProvider IEEE
linkToHtml http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwlV3fS8MwED62KejT1E2cP4v44IN12Zq2yeMQx8RtDpm4t9CmKQx0k60T_O-9y-qmIIhvpSSh_cjx5e6-ywFcBHjq8AzTLm8wz0W-1q7kCQXyZYoOF5JQFNlmE2G_L0YjOSjA1aoWxhhjxWfmmh5tLj-Z6gWFyurCD7kvwiJs-Jw32bJaaxVRaQbCE2yVi2Syjj_y-EDyLdJLMupHXIz0648-KpZG2uX_fcAOVNf1eM5gxTS7UDCTPSh_NWRwcvuswHNrnY52l9eWU61X5vRxP7s9EtV-OHlpAC7ktL5lERyKyDr5GKo5mJgXnGdF4lV4at8Obzpu3jnBjdD9ydApjFiM1BSiyaWmKWLPi5KUJXGI9B2HOg5lQ8RMRGjhviBO4kEaetLESZJotMl9KE2mE3MATpAY7XnSN9xITo3JidI500Hqazzy-jU4R0wVuQRzZT0KJpUFXhHwygJfg8u_B6l4NjZpDSqEu3pbXrahcsgPf399BludYa-runf9-yPYpsWWSrtjKGWzhTmBTf2ejeezU7tXPgHy-7ry
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=Proceedings+of+the+51st+Annual+IEEE%2FACM+International+Symposium+on+Microarchitecture&rft.atitle=Application-transparent+near-memory+processing+architecture+with+memory+channel+network&rft.au=Alian%2C+Mohammad&rft.au=Min%2C+Seung+Won&rft.au=Asgharimoghaddam%2C+Hadi&rft.au=Dhar%2C+Ashutosh&rft.series=ACM+Conferences&rft.date=2018-10-20&rft.pub=IEEE+Press&rft.isbn=9781538662403&rft.spage=802&rft.epage=814&rft_id=info:doi/10.1109%2FMICRO.2018.00070
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9781538662403/lc.gif&client=summon&freeimage=true
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9781538662403/mc.gif&client=summon&freeimage=true
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9781538662403/sc.gif&client=summon&freeimage=true