Tyan's GT75-BP012

Getting to the meat of today's article, we have the Tyan GT75-BP012. Anton has already described the Tyan GT75 servers in great detail here, so we will recap and add a few details.

The Tyan GT75 machines (just like the Tyan TN71-BP012 servers launched a year ago) are based on one IBM POWER8 Turismo Single Chip Module (SCM) processor, offering either eight or ten cores. This CPU finds itself paired with Tyan's Habanero motherboard, the same as in IBM's most affordable OpenPOWER server, the S812LC.

The board has 32 DIMM slots using four IBM Centaur memory buffer chips (MBCs). Since the operational voltage of the Centaur chip PHY maxes out at 1.43V, only low power DDR3 DIMMs are supported. The largest supported DIMMs are the quad ranked 32 GB DIMMs with 4 Gbit chips, allowing the server to have up to 1 TB of RAM. Unfortunately, the latest 8 Gbit based DIMMs are not supported. Tyan ships the server with eight 16 GB DIMMs – for a total of 128GB – if you take the standard configuration.

Tyan GT75: IBM POWER8 Turismo CPU Options
  POWER8 8-Core POWER8 10-Core
Core Count 8 10
Threads 64 80
Nominal Freq.
2.33 GHz
3.025 GHz
2.095 GHz
2.926 GHz
L2 Cache 512 KB per core 512 KB per core
L3 Cache 8 MB eDRAM per core
64 MB per CPU
8 MB eDRAM per core
80 MB per CPU
DRAM Interface DDR3L-1600 (Low Power Only)
PCI Express 3 × PCIe controllers, 32 lanes
TDP 130W

As the OpenPOWER POWER8 has to fit and operate within a 1U home, the clockspeed is limited to 2.328 GHz nominal. However, that is just a paper spec just like the clockspeed of the Xeon E5. In reality, the power governor defaults to "on demand". In that case, the CPU runs at 2.06 GHz at low load, and boost up to 3.025 GHz when the CPU is fully loaded. The speedsteps are very small, only +/- 30 MHz, so the second highest speedstep is 2.99 GHz. Below you find the configuration table of all Tyan GT75 servers.

Comparison of Tyan GT75 Servers
  BSP012G75V4H-B4C BSP012G75V4H-Q4T BSP012G75V4H-Q4F
2.328 GHz
130 W/169 W TDP
2.095 GHz
130 W/169 W TDP
2.095 GHz
130 W/169 W TDP
Installed RAM 8 × 16 GB R-DDR3L 16 × 16 GB R-DDR3L 32 × 16 GB R-DDR3L
RAM (subsystem) Up to 1 TB of DDR3L-1333 DRAM, 32 RDIMM modules, four IBM Centaur MBCs
Storage 2 × 512 GB SSDs 2 × 1 TB SSDs 4 × 1 TB SSDs
Tyan Storage Mezzanine MP012-9235-4I
(4-port SATA 6Gb/s IOC w/o RAID stack)
LAN 4 × GbE ports 4 × 10 GbE ports 4 × 10 GbE ports
Tyan LAN Mezzanine MP012-5719-4C
Broadcom 1GbE LAN Mezz Card
Qlogic+Broadcom 10GbE LAN Mezz Card-
MP012-Q840-4F Qlogic 10GbE LAN Mezz Card

In today's article we're review the basic model, the BSP012G75V4H-B4C. Notice the twelve (!) fans.

The Tyan GT75-BP012 makes use of Tyan's mezzanine cards for networking and for the storage controller. As a result, it can equipped with up to four 3.5” hot-swappable SATA 6G HDD/SSDs and four network controllers (1 GbE or 10 GbE) without using the 8-lane PCIe riser.

Now if you've been counting the CPIe lanes required for all of this, it seems like we should be a bit short, and indeed that's the case. Digging a bit deeper, we'll find that the server is using a PLX PEX8748 PCIe switch to take a PCIe 3.0 x8 root port from the CPU and switch it among the LAN riser, SATA riser, and the two black PCIe x8 slots.

The OpenPOWER Saga Continues: Can You Get POWER Inside 1U? Benchmark Configuration and Methodology
Comments Locked


View All Comments

  • Zzzoom - Friday, February 24, 2017 - link

    "As important as performance per watt is, several markets – HPC, Analytics, and AI chief among them – consider performance the most important metric. Wattage has to be kept under control, but that is it."

    What a load of garbage.
  • JohanAnandtech - Saturday, February 25, 2017 - link

    And now maybe some arguments that substantiate your opinion?
  • SarahKerrigan - Sunday, February 26, 2017 - link

    In HPC specifically, power consumption is a major issue. This was the entire root of the success of the Blue Gene line back in the day, and why NEC is shifting its supercomputing CPUs to progressively more efficient cores instead of higher-performance cores now (SX-9: 102.4GF/core; SX-ACE: 64GF/core.) . HPC is sensitive to running cost, and power dissipation is a critical factor in that.
  • Zzzoom - Monday, February 27, 2017 - link

    Go read the 7+ years worth of materials from the EE HPC Working Group.
  • JohanAnandtech - Wednesday, March 1, 2017 - link

    In a system with 2-4 GPUs, 512 GB of RAM, the TDP of the CPU is not a dealbreaker. I can agree that some HPC markets are more sensitive to perf/watt; but I have seen a lot of examples where raw performance per dollar was just as important.
  • Zzzoom - Wednesday, March 1, 2017 - link

    POWER8 TDP is 45W-102W higher per socket than the highest spec Xeon E5. That's 90W-204W higher per node where each node consumes 1500W-2000W, or 6-10% total on a site with a multi-million dollar power bill that went to great lengths to bring down the PUE by a similar amount. So for anyone to pick POWER8 it has to do better on energy to solution through its unique features, or be considerably cheaper (ha!). POWER8's advantage is NVLink, but TSUBAME3 going with Intel+PLX switches on top of NVLink shows that it's not that big of a deal.
    Anyway, the efficiency requirements on the CORAL procurements are pretty strict so scale-out POWER9+Volta will have to shed a lot of weight.
  • Zzzoom - Wednesday, March 1, 2017 - link

    I forgot about the memory buffers. It's even worse.
  • mystic-pokemon - Sunday, March 5, 2017 - link

    Guys, I know shit ton of stuff about a server Johan listed above. He has a point when he says Power consumption is only so much important.
    In short, when you combine all aspects to TCO model: POWER8 server delivers most optimal TCO value
    We consider all the following into our TCO model
    a) Cost of ownership of the server
    b) Warranty (Lesser than conventional server, different model of operations)
    c) What it delivers (How many independent threads (SMT8 on POWER8 remember ? 192 hardware threads), how much Memory Bandwidth (230 GBPs), how much total memory capacity in 1 server ( 1 TB with 32 GB)
    d) For a public cloud use-case, how many VMs (with x HW threads and x memory cap / bw ) can you deliver on 1 POWER8 server compared to other servers in fleet today ? Based on above stats, a lot .
    e) Data center floor lease cost in DC ( 24 of these servers in 1 Rack, much denser. Average the lease over age of server: 3 years ). This includes all DC services like aggers, connectivity and such.
    f) Cost per KWH in the specific DC ( 1 Rack has nominal power 750W)

    All this combined POWER has good TCO. Its a massively parallel server, what where major advantage comes from. Choose your workload wisely. That's why companies continue to work on it.

    I am talking about all this without actually combining with CAPI over PCIe and openCAPI. Get it ? POWER is going no where.
  • Michael Bay - Friday, February 24, 2017 - link

    I think at this point in time intel has more to fear from goddamn ARM than IBM in server space.
    Okay, maybe AMD as well.
  • JohanAnandtech - Friday, February 24, 2017 - link

    Personally I think OpenPOWER is a viable competitor, but in the right niches (In memory databases, GPU accelerated + NVlink HPC). Just don't put that MHz beast in a far too small 1U cage. :-)

Log in

Don't have an account? Sign up now