IBM says these back-office, network edge Power 10 servers would be sweet for - yes, you guessed it - AI

Not to be left out of the AI infrastructure game, on Tuesday IBM unveiled a pair of tiny Power 10 servers designed to preprocess data at the network edge.

The Power S1012 systems are available in both a PC-style tower configuration and a more traditional 2U half-width rack mount chassis. Both can be equipped with IBM's homegrown Power 10 processor with one, four, or eight cores enabled and up to 256GB of onboard memory.

While that might not sound like a lot of cores next to Intel and AMD's edge-centric chips, which can be had with up to 64 cores, it's worth noting that IBM's Power platform is based on a RISC architecture that prioritizes highly-threaded workloads with support for both SMT4 or SMT8.

That means the Power 10 eSCM modules used in these systems can support up to eight threads per core, which on the top-specced configuration works out to a more respectable 64 threads.

IBM boasts its new servers are up to three times more powerful than the outgoing Power S814, which may sound impressive until you consider that system is based on the 10-year-old Power 8 platform. Having said that, the Power 10 family isn't exactly all that fresh anymore either, with it due to celebrate its third birthday in September.

IBM envisions these systems being deployed in a number of scenarios including for AI inferencing in space or power constrained edge deployments, or for running more traditional workloads in remote or back office scenarios.

The chief argument appears to be that by processing all the data streaming in from the edge in place rather than shuttling it all back to a central datacenter, customers can reduce latencies and curb bandwidth consumption.

By all appearances IBM seems to be targeting existing Power customers familiar with the particular hardware and software nuances associated with the SMT-heavy architecture. One those customers is analytics wrangler Equitus, which IBM says is already using the systems to run its AI models at the edge.

As for how IBM is going about processing those AI workloads differs considerably from what you might expect. From what we can tell, these systems aren't equipped with GPUs - IBM's announcement makes no reference to them. Instead, IBM appears to be leaning on the processors' matrix math accelerators (MMAs), four of which are baked into each core, to do the heavy lifting.

In many respects, these MMAs are reminiscent of the AMX engines which appeared in Intel's 4th and 5th-gen Xeon Scalable platforms from 2023. And as we've recently explored, those engines are more than capable of running small large language models ranging between 7 to 13 billion parameters in size.

Alongside its MMAs, IBM also highlighted support for transparent memory encryption to safeguard data moving in and out of AI models on the device. Considering the fact these systems are likely to deployed in remote locations with limited security or supervision, this is likely a welcome feature - particularly for those in highly regulated industries.

The S1012 systems will be available for purchase beginning June 14. ®

Search
About Us
Website HardCracked provides softwares, patches, cracks and keygens. If you have software or keygens to share, feel free to submit it to us here. Also you may contact us if you have software that needs to be removed from our website. Thanks for use our service!
IT News
May 19
Among AI infrastructure hopefuls, Qualcomm has become an unlikely ally

Analysis The enemy of my enemy is my best friend

May 18
Gentoo and NetBSD ban 'AI' code, but Debian doesn't - yet

Comment The problem isn't just that LLM-bot generated code is bad - it's where it came from

May 18
Reddit goes AI agnostic, signs data training deal with OpenAI

Now Google and OpenAI can slurp up your precious memes and priceless comments

May 17
Graph database shows Biden outspends Trump in social media ad war

But incumbent is mentioned a lot more in attack material

May 17
CoreWeave debt deal with investment firms raises $7.5B for AI datacenter startup

Funds to be used for purchasing servers and networking kit

May 17
Hugging Face to make $10M worth of old Nvidia GPUs freely available to AI devs

You get a GPU, you get a GPU, everyone gets a ZeroGPU!

May 17
Discovering the promise of platform engineering

Webinar And what it can mean for your software development team