Samsung develops industry’s largest capacity HBM3E chips for faster AI learning

이재림 2024. 2. 27. 12:05
음성재생 설정 이동 통신망에서 음성 재생 시 데이터 요금이 발생할 수 있습니다. 글자 수 10,000자 초과 시 일부만 음성으로 제공합니다.
글자크기 설정 파란원을 좌우로 움직이시면 글자크기가 변경 됩니다.

이 글자크기로 변경됩니다.

(예시) 가장 빠른 뉴스가 있고 다양한 정보, 쌍방향 소통이 숨쉬는 다음뉴스를 만나보세요. 다음뉴스는 국내외 주요이슈와 실시간 속보, 문화생활 및 다양한 분야의 뉴스를 입체적으로 전달하고 있습니다.

Samsung Electronic succeeded in developing its fifth-generation high bandwidth memory (HBM) chips HBM3E that has the industry’s largest capacity of 36 gigabytes, the Korean chipmaker said on Tuesday.
Samsung Electronics' fifth-generation HBM3E chips [SAMSUNG ELECTRONICS]

Samsung Electronics succeeded in developing its fifth-generation high bandwidth memory (HBM) chips HBM3E which has the industry’s largest capacity of 36 gigabytes, the Korean chipmaker said on Tuesday.

The Suwon, Gyeonggi-based manufacturer has already started sending out product samples to its clients and the chips are slated for mass production by the first half of this year.

HBM3E chips can stack twelve 24-gigabit dynamic random access memory (DRAM) chips, making it 50 percent more efficient than its fourth-generation predecessor, HBM3, which can stack eight layers.

The chipmaker believes that its latest product arrives on the global market at an opportune time, coinciding with the rapid evolution of AI-driven services and heightened demands for data processing.

For instance, when applying 12-layer HBM3E chips, the speed of AI machine learning will improve by 34 percent compared to 8-layer HBM3 chips, while up to 11.5 times more users will be able to utilize AI services due to increased processing capacity and efficiency.

BY LEE JAE-LIM [lee.jaelim@joongang.co.kr]

Copyright © 코리아중앙데일리. 무단전재 및 재배포 금지.