WebDynamic Tensor Rematerialization (DTR) Marisa Kirisame, Steven Lyubomirsky, Altan Haan, Jennifer Brennan, Mike He, Jared Roesch, Tianqi Chen, Zachary Tatlock. Save memory for NN by dynamically discarding and recomputing intermediate results at runtime. By being smart about what to keep and what to discard, train larger models under a tight … WebJun 16, 2024 · Checkmate: Breaking the memory wall with optimal tensor rematerialization. In Proceedings of Machine Learning and Systems 2024, pages 497 …
Venues OpenReview
Web2024) identifies the optimal rematerialization schedule for arbitrary static graphs. Shah et al. (2024) extends Check-mate with operator implementation selection, but this is orthogonal to our work’s scheduling problem. Dynamic Tensor Rematerialization (DTR) (Kirisame et al., 2024) finds an approximation of Checkmate that is near-optimal WebWe demonstrate that a simple online algorithm can achieve comparable performance by introducing Dynamic Tensor Rematerialization (DTR), a greedy online algorithm for … birds that build nest with moss
[2006.09616] Dynamic Tensor Rematerialization - arXiv.org
Web2 Dynamic Tensor Rematerialization DTR is designed as a thin runtime layer that intercepts tensor allocations, accesses, and deallocations, eliminating the need for ahead-of-time program (e.g., DL model) analysis. Figure 1 sketches DTR’s high-level approach. When a tensor allocation occurs, DTR first checks if sufficient memory is available. http://sampl.cs.washington.edu/research.html WebThe dashed and dotted lines represent the last ratio before thrashing and out-of-memory errors, respectively. - "Dynamic Tensor Rematerialization" Figure 2: Simulated results comparing different heuristics on various models, comparing rate of computational slowdown for different budgets (fractions of the original peak memory usage). ... dance a chachucha fandango bolero