Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards

Cited 11 time in webofscience Cited 0 time in scopus
  • Hit : 230
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKwon, Youngeunko
dc.contributor.authorRhu, Minsooko
dc.date.accessioned2022-11-24T10:01:37Z-
dc.date.available2022-11-24T10:01:37Z-
dc.date.created2022-11-20-
dc.date.issued2022-06-
dc.identifier.citation49th IEEE/ACM International Symposium on Computer Architecture, ISCA 2022, pp.860 - 873-
dc.identifier.issn1063-6897-
dc.identifier.urihttp://hdl.handle.net/10203/300888-
dc.description.abstractPersonalized recommendation models (RecSys) are one of the most popular machine learning workload serviced by hyperscalers. A critical challenge of training RecSys is its high memory capacity requirements, reaching hundreds of GBs to TBs of model size. In RecSys, the so-called embedding layers account for the majority of memory usage so current systems employ a hybrid CPU-GPU design to have the large CPU memory store the memory hungry embedding layers. Unfortunately, training embeddings involve several memory bandwidth intensive operations which is at odds with the slow CPU memory, causing performance overheads. Prior work proposed to cache frequently accessed embeddings inside GPU memory as means to filter down the embedding layer traffic to CPU memory, but this paper observes several limitations with such cache design. In this work, we present a fundamentally different approach in designing embedding caches for RecSys. Our proposed ScratchPipe architecture utilizes unique properties of RecSys training to develop an embedding cache that not only sees the past but also the "future"cache accesses. ScratchPipe exploits such property to guarantee that the active working set of embedding layers can "always"be captured inside our proposed cache design, enabling embedding layer training to be conducted at GPU memory speed.-
dc.languageEnglish-
dc.publisherIEEE/ACM-
dc.titleTraining Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards-
dc.typeConference-
dc.identifier.wosid000852702500060-
dc.identifier.scopusid2-s2.0-85132797165-
dc.type.rimsCONF-
dc.citation.beginningpage860-
dc.citation.endingpage873-
dc.citation.publicationname49th IEEE/ACM International Symposium on Computer Architecture, ISCA 2022-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationNew York-
dc.identifier.doi10.1145/3470496.3527386-
dc.contributor.localauthorRhu, Minsoo-
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 11 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0