A Principled Technologies report: Hands-on testing. Real-world results.

Improving energy efficiency in the data center: Endure higher temperatures with confidence with Dell PowerEdge HS5620 servers

In high-temperature test scenarios, a Dell PowerEdge HS5620 server continued running an intensive workload without component warnings or failures, while a Supermicro SYS-621C-TN12R server failed

The Dell PowerEdge HS5620 maintained OS SSD temperatures 33°C cooler in 25°C ambient conditions (when comparing average temperatures over the course of the two-hour workload to those of the Supermicro SYS‑621C-TN12R server). It also continued without failure in 35°C conditions where the Supermicro SYS‑621C-TN12R server failed, and it kept OS SSD temperatures 34°C cooler during an HVAC malfunction scenario (when comparing average temperatures over the course of the two-hour workload to those of the Supermicro SYS‑621C-TN12R server).

Running a data center at higher temperatures can help organizations save on cooling costs and energy consumption. ENERGY STAR® notes that such measures can be financially beneficial: “Every 1°F increase in temperature can save 4% to 5% in energy costs.”1 But a data center can only remain as warm as its hardware will allow. Servers that can withstand higher day-to-day temperatures can help organizations meet sustainability goals. With the right thermal design, a server can also continue functioning at even hotter temperatures in the event of unexpected circumstances, such as internal fan failures or external environment malfunctions.

At PT, we tested two 2U cloud-optimized servers: the Dell PowerEdge HS5620 and the Supermicro® SYS-621C-TN12R. To create an environment where we could fully control and measure the temperature, we constructed an enclosure around a fully populated server rack. All the servers in the rack ran an intensive synthetic floating-point workload similar to a machine learning inference workload, which stressed the systems’ processors and generated heat in the environment.

We monitored the Dell and Supermicro servers’ internal temperatures during three types of scenario-based tests: ambient temperatures of 25°C, a fan failure, and a data center HVAC malfunction. The Dell system performed without component heat warnings or failures in these tests. In contrast, the Supermicro system yielded warnings in all scenarios and component failures in the second and third scenarios, resulting in system downtime that required manual intervention. Equipped with compelling cooling design advantages, the Dell PowerEdge HS5620 handled each challenge without fail.

How we tested

Table 1 shows key configuration information for the two cloud-optimized servers we tested. See the science behind the report for more details.

Table 1: Key configuration details of the servers we tested.
Dell PowerEdge HS5620 Supermicro SYS-621C-TN12R
Processors 2x Intel® Xeon® Gold 6444Y processors 2x Intel Xeon Gold 6444Y processors
Memory 1,024 GB of DDR5 RAM 1,024 GB of DDR5 RAM
Network interface card (NIC) Intel E810-XXV with 2x 25GbE ports Intel E810-XXVAM2 (AOC-S25GC-i2S) with 2x 25GbE ports
Storage 2x M.2 NVMe® SSDs 2x M.2 NVMe SSDs
Storage controller Dell BOSS N1 Direct-attached PCIe® storage
Power supply unit (PSU) 2x 1,800W Dell 05222NA00 2x 1,200W Supermicro HMCG94MEBRA123N
Fans 5x Dell HPR Silver 1x Dell HPR Gold 3x Supermicro Middle Fan FAN-0206L4

We configured the Dell server with a storage controller and two M.2 NVMe drives to match the storage configuration of the Supermicro server. The Dell server configuration featured five Dell HPR Silver fans and one Dell HPR Gold fan. The Supermicro server supported three 8cm fans, the maximum number it could accommodate at the time of testing. To enable performance monitoring, we adjusted the BIOS system profile settings on the Dell server to “Performance Per Watt (OS).” We kept the Supermicro server’s default BIOS configuration, “OS Controls EPB,” as this setting allowed us monitor the data our tests required.

To create an environment for which we could control and measure the temperature, we built a custom enclosure around a fully loaded 42U server rack. We placed the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12R in middle of the rack at the same position when we tested them. We configured the rest of the 42U rack with a top-of-rack switch and a variety of 2U and 1U servers, as well as blade servers and chassis, which generated heat while running their workloads. We captured baseband management and OS-level telemetry for component monitoring with third-party tools Telegraf and Prometheus.

We tested the servers in three types of scenarios: typical operations in 25°C ambient conditions, an internal fan failure (twice, with different fans disabled each time), and an HVAC malfunction where ambient temperatures rose to 35°C. In each of these scenarios, we used the stress-ng tool to stress the processors’ floating-point capabilities. This type of workload is critical for use cases including AI training and high-performance computing (HPC); for more information, see below. The cabinet of servers launched the workload in four waves, with the Dell and Supermicro systems we tested launching the workload in the fourth wave, 3 minutes and 30 seconds after the first servers began. We monitored temperatures and hardware statistics for 15 minutes prior to starting a workload, throughout the two-hour workload, and for 15 minutes after the workload completed.

For more details on our tests, results, and configurations, see the science behind the report.

Overview of our findings

In Tables 2 and 3, we show an overview of how various components of the servers fared during each test. If at least one of the listed component types showed a warning or failure, we note that below. As Table 2 shows, the components of the Dell PowerEdge HS5620 remained operational in each scenario type without displaying any warnings. On the other hand, the Supermicro SYS-621C-TN12R experienced at least one warning in each scenario type—including in 25°C ambient temperatures—as well as component failure in the HVAC malfunction scenario and in both fan failure scenarios (Table 3). The OS SSD failures we observed in testing resulted in system failure, rendering the Supermicro system unusable and requiring manual intervention. We analyze the servers’ thermal designs and examine these results more closely below.

Table 2: An overview of how key components of the Dell PowerEdge HS5620 server fared in the test scenarios.
Dell PowerEdge HS5620
Component category 25°C ambient temperatures Fan 2 failure Fan 3 failure HVAC malfunction
CPU No warnings or failures No warnings or failures No warnings or failures No warnings or failures
RAM No warnings or failures No warnings or failures No warnings or failures No warnings or failures
NIC No warnings or failures No warnings or failures No warnings or failures No warnings or failures
M.2 SSD No warnings or failures No warnings or failures No warnings or failures No warnings or failures
PSU No warnings or failures No warnings or failures No warnings or failures No warnings or failures
Table 3: An overview of how key components of the Supermicro SYS-621C-TN12R server fared in the test scenarios.
Supermicro SYS-621C-TN12R
Component category 25°C ambient temperatures Fan 1 failure Fan 3 failure HVAC malfunction
CPU No warnings or failures Warning Warning Warning
RAM No warnings or failures Warning No warnings or failures No warnings or failures
NIC No warnings or failures Warning No warnings or failures Warning
M.2 SSD Warning Failure Warning Failure
No warnings or failures No warnings or failures Failure No warnings or failures

System cooling design analysis: A closer look at the advantages of the Dell PowerEdge HS5620

Analyzing the systems’ thermal designs is key to understanding how they performed in each test scenario. Servers utilize several design elements to keep systems cool, such as motherboard design. The placement of sensitive components on the motherboard can help protect these components from causing one another to overheat. Additionally, fans keep air flowing, while chassis design should also help protect components from hot air. Below, we examine these design elements in the Dell PowerEdge HS5620 and Supermicro SYS-621C-TN12R servers.

Motherboard design

The Supermicro system’s motherboard layout was particularly problematic when it came to the placement of the M.2 NVMe modules. For example, in the second and third test scenarios, even the idle SSD’s temperature increased, as it was directly downstream from a processor under load. In addition, on the right side of the chassis, there was no dedicated fan feeding air into the power distribution module (PDU) that linked the twin PSUs to the rest of the system. Instead, the Supermicro system relied on airflow from the fans built into the PSUs at the back of the chassis. While we did not observe a failure of this PDU, the BMC reported a PSU failure during the second fan failure scenario, demonstrating the drawback of this design (see the science behind the report for more on this test). See Figure 1.

In contrast, the motherboard in the Dell PowerEdge HS5620 featured a more intricate design. Processor cooling modules used heat pipes on the heat sinks to allow for more effective cooling. The PDU was integrated onto the motherboard, allowing for smoother airflow over its components. In the configuration we tested, the PDU had both a Dell HPR Gold and Dell HPR Silver fan providing cooling to the components. As Figure 2 shows, openings in the Dell system’s air shroud allowed cool air to pass over the components, mitigating heat transfer from one component to another.

A diagram of Supermicro SYS‑621C-TN12R server’s motherboard. Toward the front of the system, fans 1 through 3 are aligned horizontally, with fan 1 on the left and fan 3 on the right. They push cold air over the processors and memory modules, which are directly behind them and aligned in a row in the middle of the system. The air becomes warmer as it flows past these components, over the two SSDs directly behind the processors, and out the back of the system. Hot air also flows in from behind the server, enters a vent on the left side of the system next to fan 1, and enters the airflow of fan 1. Fan 1 pushes this hot air over a processor and close to an SSD before it exits the system.
The motherboard layout of the Supermicro SYS-621C-TN12 we tested. We added component labels and arrows that show airflow direction from the fans, with colder air indicated by blues and purples, and hotter air indicated by reds, oranges, and yellows. Source: Principled Technologies.
A diagram of the Dell PowerEdge HS5620 server’s motherboard. The six fans are aligned horizontally toward the front of the system, with fan 1 on the left and fan 6 on the right. In the middle of the system, the processors and memory modules are aligned in a row on right side. In the back of the system, the SSDs are behind and to the left of the row of processors and memory modules. An air shroud divides the bottom thermal layer from the top thermal layer. In the bottom thermal layer, fan 1 pushes cool air toward an SSD. Fan 2 pushes cool air next to this SSD, as well as to toward a memory module. Fans 3 through 6 push cool air toward the row of processors and memory modules behind them. Each fan’s airflow is cool when the fan first blows it, hot when it passes through the middle of the system near the components, and warm as it exits the back of the system. In the top thermal layer, the air shroud covers the row of processors and memory modules. All six fans blow air, which remains cold as it passes over the shroud, and becomes hot as it passes the SSDs and exits the back of the system.
The motherboard layout of the Dell PowerEdge HS5620 we tested, showing the bottom thermal layer (left) and the top thermal layer (right) that the shroud divides. We added component labels and arrows that show airflow direction from the fans, with colder air indicated by blues and purples, and hotter air indicated by reds, oranges, and yellows. Source: Principled Technologies.

Fans

A system’s fans are a critical part of cooling design. The Dell PowerEdge HS5620 configuration we tested featured five 60mm Dell HPR Silver fans and one 60mm Dell HPR Gold fan. The Supermicro SYS-621C-TN12R we tested used three primary 80mm fans for cooling. In both servers, each of the power supplies had an additional dedicated fan built in.

Cubic feet per minute (CFM) is a rating that indicates how much air a fan can move. According to their labels, the six fans in the Dell system were 57.26 CFM each (343.56 CFM total), while the three fans in the Supermicro system were 104.7 CFM each2 (314.10 CFM total). Although these totals were close, CFM is only part of the story in a server. In our tests, we also found that the Supermicro SYS-621C-TN12R server’s fans ran at about 13,500 RPM at peak load. The Dell PowerEdge HS5620 server’s dual-rotor fans spun at around 20,000 RPM in our tests. These differences in speed and design allowed the Dell system’s fans to generate higher static pressure, meaning they pushed air through the system with greater force. This also means they counteracted backpressure in the hot aisle, which is critical to effective cooling,3 since equipment in the data center with stronger fans can overwhelm those with underpowered fans, leading to insufficient cooling.

Chassis design

The Supermicro SYS-621C-TN12R chassis included vents on both sides, between the fans and the storage backplane, which were not present in the Dell server. In theory, these vents might work to allow additional airflow into the chassis in an open-air server environment without cabinets or hot aisle containment. However, in practice, the design of data center cabinets means that the sides of a server chassis are within the same thermal zone as the hot aisle: Because server racks and cabinets allow for free airflow around the sides of the servers, those vents are not isolated. So, instead of expelling hot air or pulling in cold air, the vents could allow preheated air from behind the server stack to enter the chassis and circulate through the components, creating a heating loop. See Figure 3.

The left side of the Supermicro SYS‑621C-TN12R server’s chassis. Toward the front, near the row of fans, are seven small rectangular holes (vents) in the side of the chassis.
Vents in the chassis of the Supermicro SYS-621C-TN12 we tested, which could allow preheated air from the hot aisle to circulate through the server’s components. Source: Principled Technologies.

The Dell PowerEdge HS5620 continued without even a component warning; the Supermicro SYS-621C-TN12R failed in two scenarios

Scenario 1: 25°C ambient temperatures

To see how the servers performed under typical data center temperatures, we ran our first scenario. The American Society of Heating, Refrigerating and Air-Conditioning Engineers (ASHRAE) recommends keeping a typical data center at temperatures between 18°C and 27°C, though temperatures up to 45°C are acceptable for certain classes of equipment.4 In this scenario, we targeted an ambient temperature of 25°C in the test environment as we ran the workloads on the servers.

The Dell PowerEdge HS5620 did not display any component warnings or failures in this scenario. Its airflow design kept the thermal zones isolated, maintaining safe operating temperatures for all components. In contrast, 22 minutes into the test, the Supermicro system’s baseboard management controller (BMC) warned that the OS SSD reached a critical temperature. Then, 10 minutes later, it asserted the drive had reached a non-recoverable state, though the SSD did not fail in this scenario. This is because this BMC alert did not poll the component for failure—it simply indicated that the drive had crossed a threshold where failure may be imminent.

Over the course of the two-hour workload, the Dell server’s OS SSD averaged a temperature of 43.9°C, while the idle SSD averaged 45.5°C. The Supermicro system’s OS SSD averaged 77.5°C and the idle SSD averaged 61.7°C during the test—temperatures up to 33.6°C hotter than the Dell system. While the Dell server’s processors averaged 73.7°C and 70.7°C during the workload, the Supermicro server’s processors averaged 77.9°C and 71.1°C.

Figures 4 and 5 show the SSD and processor temperatures for the two systems over the course of the two-hour test. Figure 6 shows the servers’ power consumption, where increases in power correspond to the workload’s effect on the systems, including fans working to cool the servers.

A line graph comparing each system’s SSD temperatures over the course of the two-and-a-half-hour scenario 1 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 server’s two SSDs remained around 40°C before the workload began, stayed between 40°C and 50°C over the course of the two-hour workload, and dropped to between 35°C and 40°C after the workload ended. The Supermicro SYS‑621C-TN12R server’s idle SSD remained between 50°C and 65°C over the two and a half hours, staying between 60°C and 65°C for the majority of the workload. Its OS SSD started just below 60°C before rising to 70°C at 22 minutes and 30 seconds, when its status changed from OK to Critical. It continued to rise in temperature, hitting 75°C at 32 minutes and 30 seconds into the test, where the status changed from Critical to Unresponsive. Its temperature rose to about 80°C around 1 hour and 10 minutes in the test, and it stayed around that range until the test ended. Its temperature decreased when the test ended. When the OS SSD reached 72°C at 2 hours and 19 minutes, its status changed from Unresponsive to Critical. Its temperature continued to cool to just under 60°C toward the end of the idle 15 minutes.
SSD temperatures in the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the first scenario, where the servers ran a floating-point workload in 25°C ambient temperatures. The workload began at 0:15 and ended at 2:15. SSD 1 ran the OS, while SSD 2 was idle. Lower temperatures are better. Source: Principled Technologies.
A line graph comparing each system’s processor temperatures over the course of the two-and-a-half-hour scenario 1 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 server’s two processors remained below 35°C before the workload, rose to just under 80°C at the start of the workload, and dropped down to stay between 68°C and 80°C over the course of the two-hour workload. They dropped below 35°C after the test ended. The Supermicro SYS‑621C-TN12R server’s processors started just under 40°C, then rose to just over 70°C toward the start of the workload. One of the processors stayed under 75°C throughout the two-and-a-half-hour period, while the other rose to just over 80°C and remained in that range for about an hour. When the workload ended, the processors’ temperatures dropped to around 40°C.
Processor temperatures in the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the first scenario, where the servers ran a floating-point workload in 25°C ambient temperatures. The workload began at 0:15 and ended at 2:15. Lower temperatures are better. Source: Principled Technologies.
A line graph comparing each system’s power consumption over the course of the two-and-a-half-hour scenario 1 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 fluctuated between around 300 and 400 watts before the workload, rose to around 900 watts at the start of the workload, and fluctuated between just over 900 watts and about 800 watts while running the two-hour workload. Its power consumption dropped when the workload ended, from about 850 watts to about 300 watts. The Supermicro SYS‑621C-TN12R server’s power consumption remained steady at just over 400 watts before the test, then rose to around 850 watts at the start of the workload. It steadily consumed watts in the mid-800s range during the two-hour workload. When the workload ended, its power consumption dropped to about 450 watts.
Power consumption, in watts, of the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the first scenario, where the servers ran a floating-point workload in 25°C ambient temperatures. The workload began at 0:15 and ended at 2:15. Source: Principled Technologies.

In this scenario reflecting everyday operations in a 25°C data center, the Dell PowerEdge HS5620 performed without raising concerns, while the Supermicro SYS-621C-TN12R issued a critical warning for its OS drive. Table 4 summarizes the results of this test scenario.

Table 4: A summary of our findings over the course of the first two-hour test, where the servers ran a floating-point workload in 25°C ambient temperatures.
Scenario 1: 25°C ambient temperatures
Dell PowerEdge HS5620 Supermicro SYS-621C-TN12
Outcome No system failure No system failure
Component failures None None
Component warnings None OS SSD
OS SSD average temperature 43.9°C 77.5°C
Idle SSD average temperature 45.5°C 61.7°C

Scenario 2: Fan failure

Even though careful monitoring and regular servicing can extend a server’s life, internal components can fail unexpectedly. In our second test scenario, we aimed to see how each server would perform with a fan failure.

In terms of equivalent airflow coverage of components, we determined fan 1 on the Supermicro server aligned best with fan 2 on the Dell system—we manually disabled these fans for our testing. We again started the test at an ambient temperature of 25°C and initiated the workload after 15 minutes of running the servers idle. Over the course of the two-hour test and for 15 minutes afterward, we monitored the systems for warnings and failures.

The Dell PowerEdge HS5620 did not experience any component failures or issue any component warnings, while the Supermicro SYS-621C-TN12R—with just two healthy fans—alerted us to high temperatures for the processors, RAM, and two NICs. The Supermicro system’s OS drive failed 1 hour and 49 minutes into the test, resulting in system failure; once the SSD cooled down and the OS could resume functioning, we had to restart the server via the BMC. The system’s air channels and fans were not able to compensate for cooling design shortcomings, such as airflow patterns that directed hot air from the processors and memory over the SSDs. In contrast, the Dell system’s greater number of fans—each spinning at a higher rate of rotations per minute (RPM) than those of the Supermicro system—and airflow design helped keep the system components cooler and functional.

We again found that the Dell system maintained lower temperatures, on average, over the course of the two-hour workload. Its OS SSD averaged 54.2°C, which was 28.0°C cooler than the Supermicro server’s OS SSD 82.2°C average. The Dell system’s idle SSD averaged 47.0°C, or 21.5°C cooler than the 68.5°C average of the Supermicro server’s idle SSD. When it came to average processor temperatures, those in the Dell server were 56.9°C and 44.3°C, while those in the Supermicro server ran much hotter, at 98.6°C and 72.8°C—a temperature difference of up to 54.3°C. We saw the Dell server’s management systems adjust performance when temperatures crossed safety thresholds based on whether the system detected cooling hardware failures or abnormal environmental conditions.

Figures 7 and 8 show the SSD and processor temperature measurements we collected. Figure 9 compares the systems’ power draw as the systems ran the workload, generating internal heat under load and compensating for the loss of the fan.

A line graph comparing each system’s processor temperatures over the course of the two-and-a-half-hour scenario 2 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 server’s two processors remained around 30°C before the workload and peaked at 64°C 22 minutes into the test, when the system’s inlet temperature reached 33°C. When the inlet temperature reached 39°C at 40 minutes and 30 seconds into the test—when the CPUs measured at around 51°C—the Dell system adjusted the core speed of one of the processors. This processor remained around 60°C throughout the workload and dropped down to around 30°C after completing the two-hour workload. The other processor remained under 45°C for the majority of the workload, and also dropped to around 30°C after completing the workload. In the Supermicro SYS‑621C-TN12R server, one of the processors started at 60°C, then rose to 100°C throughout the entirety of the workload. After it completed the workload, its temperature dropped to about 70°C. Its other processor started at just over 35°C, rose to about 65°C when the test began, and climbed to about 80°C at 50 minutes into the test. It stayed around the 80°C mark throughout the majority of the workload before dropping to around 40°C after completing the workload.
Processor temperatures in the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the second scenario, where the servers ran a floating-point workload with one fan disabled in each server. The workload began at 0:15 and ended at 2:15. Lower temperatures are better. Source: Principled Technologies.
A line graph comparing each system’s power consumption over the course of the two-and-a-half-hour scenario 2 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 fluctuated between about 350 and 470 watts before the workload, rose to around 750 watts at the start of the workload, and remained around the 600-watt mark for the majority of the two-hour workload. Its power consumption dropped to around 375 watts when the workload ended, with a couple of short spikes to about 475 watts. The Supermicro SYS‑621C-TN12R server’s power consumption remained steady at around 450 watts before the test, then rose to just over 800 watts at the start of the workload. Its power consumption dropped to around 650 to 675 watts within 20 minutes of the initial spike, and it stayed in that range for the majority of the test. When the workload ended, its power consumption dropped to just under 400 watts.
Power consumption, in watts, of the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the second scenario, where the servers ran a floating-point workload with one fan disabled in each server. The workload began at 0:15 and ended at 2:15. Source: Principled Technologies.

A server that can remain operational when a fan fails opens up time for an organization to implement a contingency procedure while IT admins service the system. But if a server’s OS drive or other critical component fails shortly after its fan does, important applications can go offline unexpectedly, interrupting users. Waiting on a replacement fan might keep the server down for even longer stretches of time. Table 5 summarizes our findings from this test scenario.

Table 5: A summary of our findings over the course of the second two-hour test, where the servers ran a floating-point workload with one fan disabled in each server.
Scenario 2: Fan failure
Dell PowerEdge HS5620 Supermicro SYS-621C-TN12
Outcome No system failure System failure
Component failures None OS SSD
Component warnings None 1 SSD, 1 CPU, 1 memory module, 2 NICs
OS SSD average temperature 54.2°C 82.2°C
Idle SSD average temperature 47.0°C 68.5°C

We also ran a second fan failure scenario, where the fan we disabled in both servers was in a different position. For this scenario, we determined fan 3 in the Supermicro system was comparable to fan 3 in the Dell system. The Dell PowerEdge HS5620 once again did not experience any component warnings or failures, but the Supermicro SYS-621C-TN12R delivered warnings for the processors and SSDs, and one of its two PSUs failed. (For more details on this test, see the science behind the report.)

Scenario 3: HVAC malfunction

Unexpected failures aren’t limited to internal server components—overheating can also occur when something goes wrong with a facility. Our third scenario mirrors a data center whose cooling system malfunctions.

For 15 minutes, we verified each of the servers’ components were online and healthy in 25°C ambient temperatures. Then, we ran the workload for 15 minutes before shutting off all the air handlers in the test environment. When the ambient temperature in the environment reached 35°C—around one hour later—we turned the air handlers back on to reflect a situation where a facilities team fixes the HVAC system. We followed the servers’ cooling progress until the ambient temperature returned to 25°C.

According to documentation, the Dell PowerEdge HS5620 can run in 30°C conditions in the configuration we tested.5 In this scenario where temperatures rose to 35°C, the system worked past its limit and did not give any component-level warnings or experience any failures. We saw it adjust processor core speed and power consumption to avoid overheating in response to inlet sensor signals (see the science behind the report for more information). Although Supermicro SYS-621C-TN12R documentation states the system can operate in 35°C environments,6 it experienced OS SSD failure in this scenario, resulting in system failure. OS application telemetry stopped nearly an hour into the test. The system stopped responding to SSH and KVM commands, so we shut it down manually with the BMC. Notably, even during this downtime, the system continued to consume more power than the Dell PowerEdge HS5620 (Figure 12). The Supermicro system also gave high-temperature warnings on a NIC and a processor in this scenario.

Throughout the two-hour workload, the Dell system’s OS SSD averaged 48.0°C, and its idle SSD averaged 49.2°C. Compared to the Supermicro system’s average SSD temperatures—82.4°C for the OS drive and 66.4°C for the idle drive—the Dell system kept its SSDs up to 34.4°C cooler.

Figures 10 and 11 show the SSD and processor temperatures for the two systems during this scenario. Figure 12 compares the systems’ power consumption increases as they ran the workload.

A line graph comparing each system’s SSD temperatures over the course of the two-and-a-half-hour scenario 3 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 server’s two SSDs remained around 40°C before the workload, stayed between about 45°C and 60°C over the course of the two-hour workload, and dropped to just over 40°C after completing the workload. We saw processor core throttling at 1 hour and 7 minutes into the test—when the SSD was 50.9°C—and at 1 hour and 55 minutes into the test—when the SSD was 46.9°C. The Supermicro SYS‑621C-TN12R server’s idle SSD was just over 50°C before the workload began, and steadily rose over the course of the two-hour workload, peaking around 75°C before dropping to just over 50°C when the workload completed. Its OS SSD started around 55°C, and temperatures peaked at 93°C 1 hour and 33 minutes into the test. At 25 minutes into the test, when the OS SSD reached 70°C, it reached an upper-critical threshold. Then, at 36 minutes into the test when the SSD reached 75°C, we received a warning that the OS SSD had reached an upper non-recoverable threshold. 1 hour and 10 minutes into the test, when the OS SSD measured 86°C, telemetry stopped, and the drive entered read-only mode. After the workload competed, OS SSD temperatures dropped to just under 60°C.
SSD temperatures in the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the third scenario, where the servers ran a floating-point workload while ambient temperatures rose from 25°C to 35°C to simulate HVAC failure. The workload began at 0:15 and ended at 2:15. SD 1 ran the OS, while SSD 2 was idle. Lower temperatures are better. Source: Principled Technologies.
A line graph comparing each system’s processor temperatures over the course of the two-and-a-half-hour scenario 3 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 server’s two processors remained just over 30°C before the workload began, then rose when the workload started. At 30 minutes into the test, when the first processor was 74°C, we turned the chillers off. The processor temperatures peaked at about 80°C 40 minutes into the test. For the majority of the workload, they remained between 50°C and 60°C, through one of the processors dropped down to about 45°C at two points. We turned the chillers back on at 1 hour and 31 minutes into the test, when the first processor was 60°C. After the workload ended, the processors’ temperatures climbed from around 50°C to just over 70°C. In the Supermicro SYS‑621C-TN12R server, the processors started at between 35°C and 40°C. They rose to between 65°C and 70°C when the workload began, and climbed steadily until about 1 hour and 10 minutes for the first processor, which peaked at 100°C, and about 1 hour and 30 minutes for the second processor, which peaked around 95°C. After the system completed the workload, both processors’ temperatures dropped to around 40°C.
Processor temperatures in the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the third scenario, where the servers ran a floating-point workload while ambient temperatures rose from 25°C to 35°C to simulate HVAC failure. The workload at 0:15 and ended at 2:15. Lower temperatures are better. Source: Principled Technologies.
A line graph comparing each system’s power consumption over the course of the two-and-a-half-hour scenario 3 test, which includes an idle period 15 minutes before the workload began, the two-hour workload, and 15 minutes after the workload completed. The Dell PowerEdge HS5620 fluctuated between about 275 and 400 watts before the workload began, rose to just over 900 watts toward the start of the workload, then began dropping about 40 minutes into the test. Starting at about 55 minutes, it remained around 600 watts. When the workload ended, its power consumption rose again to around 900 watts. The Supermicro SYS‑621C-TN12R server’s power consumption remained steady at just over 400 watts before the workload began, then rose to around 850 watts at the start of the workload. Around 35 minutes into the test, its consumption dropped to 800 watts and just below, but rose to around 925 watts when the system became unresponsive around 1 hour and 10 minutes into the test. It continued to consume around 925 watts throughout the workload, even as it was unresponsive. When the workload ended, the system’s power consumption dropped to 500 watts, and then to around 425 watts.
Power consumption, in watts, of the Dell PowerEdge HS5620 and the Supermicro SYS-621C-TN12 during the third scenario, where the servers ran a floating-point workload while ambient temperatures rose from 25°C to 35°C to simulate HVAC failure. The workload began at 0:15 and ended at 2:15. Source: Principled Technologies.

In a situation where a data center’s HVAC system underperforms or fails, a server that continues to operate can minimize interruption to users and critical operations. Such a solution can also save time and effort for IT administrators: If a server gives warnings or fails due to overheating, an admin must spend extra time checking it, possibly even rebooting it manually. Table 6 shows a summary of our findings from this test scenario.

Table 6: A summary of our findings over the course of the third two-hour test, where the servers ran a floating-point workload while ambient temperatures rose from 25°C to 35°C to simulate HVAC malfunction.
Scenario 3: HVAC malfunction
Dell PowerEdge HS5620 Supermicro SYS-621C-TN12
Outcome No system failure System failure
Component failures None OS SSD
Component warnings None 2 SSDs, 1 CPU, 1 NIC
OS SSD average temperature 48.0°C 82.4°C
Idle SSD average temperature 49.2°C 66.4°C

Conclusion: Remain resilient in high temperatures with the Dell PowerEdge HS5620 to help increase efficiency

Increasing your data center’s temperature can help your organization make strides in energy efficiency and cooling cost savings. With servers that can hold up to these higher everyday temperatures—as well as high temperatures due to unforeseen circumstances—your business can continue to deliver the performance your apps and clients require.

When we ran an intensive floating-point workload on a Dell PowerEdge HS5620 and a Supermicro SYS-621C-TN12R in three scenario types simulating typical operations at 25°C, a fan failure, and an HVAC malfunction, the Dell server experienced no component warnings or failures. In contrast, the Supermicro server experienced warnings in all three scenario types and experienced component failures in the latter two tests, rendering the system unusable. When we inspected and analyzed each system, we found that the Dell PowerEdge HS5620 server’s motherboard layout, fans, and chassis offered cooling design advantages.

For businesses aiming to meet sustainability goals by running hotter data centers, as well as those concerned with server cooling design, the Dell PowerEdge HS5620 is a strong contender to take on higher temperatures during day-to-day operations and unexpected malfunctions.

  1. ENERGY STAR, “5 Simple Ways to Avoid Energy Waste in Your Data Center,” accessed April 8, 2024, https://www.energystar.gov/products/data_center_equipment/5-simple-ways-avoid-energy-waste-your-data-center.
  2. Supermicro, “Supermicro 80mm Hot-Swappable Middle Fan (FAN-0206L4),” accessed April 9, 2024, https://store.supermicro.com/us_en/80mm-fan-0206l4.html.
  3. Electronics Cooling, “The Hidden Risk of Invisible Airflow Imbalance in an Efficient Contained Data Center,” accessed April 4, 2024, https://www.electronics-cooling.com/2016/07/the-hidden-risk-of-invisible-airflow-imbalance-in-an-efficient-contained-data-center/.
  4. ASHRAE TC9.9, “Data Center Power Equipment Thermal Guidelines and Best Practices,” accessed April 24, 2024, https://www.ashrae.org/file%20library/technical%20resources/bookstore/ashrae_tc0909_power_white_paper_22_june_2016_revised.pdf.
  5. Dell, “Dell PowerEdge HS5620 Technical Guide,” accessed May 8, 2024, https://www.delltechnologies.com/asset/en-us/products/servers/technical-support/poweredge-hs5620-technical-guide.pdf.
  6. Supermicro, “CloudDC SuperServer SYS-621C-TN12R,” accessed April 26, 2024, https://www.supermicro.com/en/products/system/datasheet/sys-621c-tn12r.
  7. Dell, “PowerEdge HS5620 Specification Sheet,” accessed April 1, 2024, https://www.delltechnologies.com/asset/en-us/products/servers/technical-support/poweredge-hs5620-spec-sheet.pdf.
  8. Dell, “PowerEdge HS5620 Specification Sheet.”

This project was commissioned by Dell Technologies.

May 2024

Principled Technologies is a registered trademark of Principled Technologies, Inc.

All other product names are the trademarks of their respective owners.

Forgot your password?