Samsung announces high bandwidth memory, processing-in-memory architecture

You are interested in Samsung announces high bandwidth memory, processing-in-memory architecture right? So let's go together Zliu.info look forward to seeing this article right here!

Samsung announces high bandwidth memory, processing-in-memory architecture
Credit: Samsung

Samsung Electronics has announced on its Newsroom webpage the development of a new kind of memory chip architecture called high-bandwidth memory, processing-in-memory—HBM-PIM. The architecture adds artificial intelligence processing to high-bandwidth memory chips. The new chips will be marketed as a way to speed up data centers, boost speed in high performance computers and to further enable AI applications.

Computer engineers have long been working to remove the bottlenecks that arise in computers due to the need to shuffle data back and forth between a computer’s CPU and its memory chips. Most efforts to do so have involved adding small amounts of fast memory caches to the CPUs—unfortunately, doing so increases energy consumption, leading to more heat production. In this new effort, the team at Samsung has taken the opposite approach—giving memory chips the ability to take on some of the processing. With the new HBM-PIM, Samsung has placed what it describes as “a DRAM optimized AI engine inside of a memory bank.” This reduces the processing load on the CPU by offloading some of its work to remote databanks. Not only is the workload reduced, but the speed of processing is increased due to a reduction in data movement.

Specs for the HBM-PIM include a PCU running at 300MHz controlled by the host CPU using conventional memory commands. With this approach, the PCU can be instructed to carry out FP16 calculations directly inside of the DRAM unit. Notably, the HBM-PIM can operate as normal RAM when a system is running applications that have not been written for it.

See also  Sensor research helps fight wildfires

Samsung notes that when they tested the new technology with their existing HBM2 Aquabolt systems, system performance doubled and energy consumption was reduced by 70%. They also noted that installing HBM-PIMs in existing systems would not require any other changes to in-place hardware or software. Their HBM-PIM technology is currently being tested with previously developed AI accelerators created by AI solution partners—they expect to see the results in the first half of this year.

Representatives of the company will be presenting a paper they have written describing the new technology at this year’s International Solid-State Circuits Virtual Conference.

Conclusion: So above is the Samsung announces high bandwidth memory, processing-in-memory architecture article. Hopefully with this article you can help you in life, always follow and read our good articles on the website: Zliu.info


Hi, I'm Wenda, currently working on Zliu.info. This is my personal Blog, where I will share the tips and knowledge that I have learned. If you have any questions, please contact me at Email: [email protected]! Thank you !

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button