Go Sugizaki
Fujitsu
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Go Sugizaki.
international solid-state circuits conference | 2013
Ryuji Kan; Tomohiro Tanaka; Go Sugizaki; Kinya Ishizaka; Ryuichi Nishiyama; Sota Sakabayashi; Yoichi Koyanagi; Ryuji Iwatsuki; Kazumi Hayasaka; Taiki Uemura; Gaku Ito; Yoshitomo Ozeki; Hiroyuki Adachi; Kazuhiro Furuya; Tsuyoshi Motokurumada
The 10th generation SPARC64™ processor named SPARC64 X contains 3-billion transistors on a 588mm2 die fabricated in an enhanced 28nm high-κ metal-gate (HKMG) CMOS process, with 13 layers of copper interconnect with low-κ dielectrics. More stress control, SiGe improvement and S/D optimization achieve about 10% higher performance than the standard 28nm high performance (28HP) process. SPARC64 X runs at 3.0GHz and consists of 16 cores, shared 24MB level 2 (L2) cache, four channels of 1.6GHz DDR3 controller, two ports of PCIe Gen3 controller, and five ports of system interface controller. ccNUMA is adopted as its memory system, and a cache coherence control unit for multi-chip systems with up to 64 processors is integrated into L2 cache control circuitry for lower latency and reduced area and power consumption.
IEICE Transactions on Electronics | 2007
Mariko Sakamoto; Akira Katsuno; Go Sugizaki; Toshio Yoshida; Aiichiro Inoue; Koji Inoue; Kazuaki Murakami
Broadcast and synchronization techniques are used for cache coherence control in conventional larger scale snoop-based SMP systems. The penalty for synchronization is directly proportional to system size. Meanwhile, advances in LSI technology now enable placing a memory controller on a CPU die. The latency to access directly linked memory is drastically reduced by an on-die controller. Developing an enterprise server system with these CPUs allows us an opportunity to achieve higher performance. Though the penalty of synchronization is counted whenever a cache miss occurs, it is necessary to improve the coherence method to receive the full benefit of this effect. In this paper, we demonstrate a coherence directory organization that fits into DSM enterprise server systems. Originally, a directory-based method was adopted in high performance computing systems because of its huge scalability in comparison with snoop-based method. Though directory capacity miss and long directory access latency are the major problems of this method, the relaxed scalability requirement of enterprise servers is advantageous to us to solve these problems along with an advanced LSI technology. Our proposed directory solves both problems by implementing a full bit vector level map of the coherence directory on an LSI chip. Our experimental results validate that a system controlled by our proposed directory can surpass a snoop-based system in performance even without applying data localization optimization to an online transaction processing (OLTP) workload.
Archive | 2012
Hideyuki Koinuma; Go Sugizaki; Toshikazu Ueki
Archive | 2004
Go Sugizaki
Archive | 2012
Toshikazu Ueki; Seishi Okada; Hideyuki Koinuma; Go Sugizaki
Archive | 2012
Toshikazu Ueki; Seishi Okada; Hideyuki Koinuma; Go Sugizaki
Archive | 2010
Go Sugizaki
Archive | 2009
Go Sugizaki
Archive | 2012
Hideyuki Koinuma; Seishi Okada; Go Sugizaki
Archive | 2012
Toshikazu Ueki; Seishi Okada; Hideyuki Koinuma; Go Sugizaki