Figure 1 - uploaded by Philip Machanick
Content may be subject to copyright.
The growing CPU-DRAM speed gap, expressed as relative speed over time (log scale). The vertical lines represent intervals at which the speed gap has doubled.

The growing CPU-DRAM speed gap, expressed as relative speed over time (log scale). The vertical lines represent intervals at which the speed gap has doubled.

Source publication
Article
Full-text available
The memory wall is the predicted situation where improvements to processor speed will be masked by the much slower improvement in dynamic random access (DRAM) memory speed. Since the prediction was made in 1995, considerable progress has been made in addressing the memory wall. There have been advances in DRAM organization, improved approaches to m...

Context in source publication

Context 1
... ) t ≈ 1.4 t , resulting in the gap doubling every 2.1 years. Figure 1 illustrates the trends, assuming a conservative CPU speed improvement of 50% per year. Consequently, even if the number of memory references to DRAM is a relatively small fraction of the total, the fraction of overall run time spend waiting for DRAM -if the problem is not addressed -will grow over time, and eventually dominate the total time spent executing a program [Wulf and McKee 1995]. ...

Similar publications

Article
Full-text available
This paper presents the Multilevel A-Diakoptics methodology (Diakoptics based on Actors) for the dynamic load flow simulation of hybrid distribution systems, which are power systems working at different base frequencies. In the development of the smart grid several challenges have been identified, such as the connection of non-conventional loads, d...
Conference Paper
Full-text available
Multicore architectures in computer hardware bring an unprecedented need for parallel programming. In the work considered in this presentation, we are especially interested in multithreaded programs with shared data. In this widely- used programming paradigm, "lock" primitives are employed to control access to the shared data within the program thr...
Article
Full-text available
We present an implementation of the ATLAS High Level Trigger, HLT, that provides parallel execution of trigger algorithms within the ATLAS multithreaded software framework, AthenaMT. This development will enable the ATLAS HLT to meet future challenges due to the evolution of computing hardware and upgrades of the Large Hadron Collider, LHC, and ATL...
Article
Full-text available
Although modern computer hardware offers an increasing number of processing elements organized in nonuniform memory access (NUMA) architectures, prevailing middleware engines for executing business processes, workflows, and Web service compositions have not been optimized for properly exploiting the abundant processing resources of such machines. A...
Article
Full-text available
As the energy consumption has been surging in an unsustainable way, it is important to understand the impact of existing architecture designs from energy efficiency perspective, which is especially valuable for High Performance Computing (HPC) and datacenter environment hosting tens of thousands of servers. One obstacle hindering the advance of com...

Citations

... Meanwhile, it provides direct and efficient support in hardware for the compiler and runtime system to exploit thread level parallelism in the program, in which it was believed to have much more potential for performance improvement [44]. Furthermore, the multithreaded execution model provides an easy way to hide the ever increasing memory access latency [45] and thus improves the program execution efficiency. In short, multicore processor is a very promising computer architecture that can effectively translate the billion number of transistors on the chip to program performance without violating power consumption limits. ...