With the proliferation of connected devices comes a plethora of new challenges, perhaps none more challenging than the strain this puts on the server space.
Indeed, already today, thousands of kilo watts of power are guzzled by memory components in servers alone, not to mention the added demands of storage technology.
Samsung, as a huge player in the entire connected ecosystem, has decided it needs to tackle the problem and at the Consumer Electronics Show (CES) in Las Vegas, was touting its “green memory” to clean up the data center.
For a start, the firm is promoting the use of lower voltage DRAMs and solid-state drives (SSDs) to significantly lower power consumption and energy bills, while reducing the overall carbon footprint of the data center.
Why bother? Samsung’s DRAM marketing director Sylvie Kadivar told EE Times the potential savings could be enormous.
“Data centers today worldwide consume some 30 billion watts,” she explained, noting that this was the approximate equivalent to some 30 nuclear power plants.
Kadivar said that with Samsung’s 64GB load reduced DIMMs (LRDIMM), enterprises could significantly slash costs by reducing that power. That can translate into hundreds of thousand of dollars saved per year.
Additionally, said Kadivar, even more savings can be achieved by switching out hard disk drives (HDDs) with SSDs, as the lack of moving parts makes the systems inherently more reliable, cooler and more power efficient.
Kadivar pointed out that the world now boasts over one billion smartphones, a phenomenon that is having a huge effect. “All this content and all these applications are leveraging the enterprise space,” she said.
You could have a point there, however given that 26% of the power budget appears to come from memory changing things to reduce power would definitely give them a competitive advantage and be worth pursuing.
How much memory is actually embedded in the 50% networking %?
26% is for local memory on a computer.
how about all the memory up stream in the routers, wireless devices, local and remote servers?
assuming 25% of these devices power is memory then 25% of 50% is 12.5%
SO the total is really 37.5% for memory cradle to grave?
And with compounding it get slightly larger ~ 40%.
So iff you convert cradle to grave and use this low power memory combo (dram or other + flash) these savings will look even better.
If all devices from Google to laptop all are converted to new memory what is the total %?