Categories: NewsTechnology

IBM Zurich Speeds up Big Data-Based AI Training by 10x or More

Artificial intelligence is a booming and competitive industry right now. Various companies and institutions are focusing their attention on this particular technology. Training an AI is still somewhat of a demanding task, although IBM’s Zurich research lab may have come up with a solution. Its generic AI preprocessing block can speed up learning algorithms by an order of magnitude.

Faster AI Training is a Good Thing

Most consumers are legitimately afraid of the rapid developments in the world of artificial intelligence. That is a rather natural reaction, as we have seen some astonishing breakthroughs in this area in the past few months. Even though the industry is booming, there is always a demand for better and faster AI training methods. Coming up with such solutions is not easy by any means. Thanks to IBM Zurich, however, a major breakthrough may be just around the corner.

More specifically, the research institution unveiled a new generic AI preprocessing building block. The main purpose of this building block is to improve the rate at which machine learning algorithms absorb new information. With a strong focus on big data machine learning, IBM Zurich claims its project can speed up the learning process by at least ten times. It uses mathematical duality to filter important pieces of information in a data stream and ignore everything else.

One of the downsides of big data is how there is simply too much information to go through. Even regular AIs can struggle to process all of the necessary information. There is a real information overflow when it comes to big data, and solving this problem has proven quite challenging so far. However, the new breakthrough by IBM Zurich may herald an entirely new era of big data machine learning.

Related Post

According to IBM Zurich mathematician Thomas Parnell, this is the first generic solution with a 10x speedup. Data sets relying on linear machine learning models will be among the first to benefit from this new solution. Do keep in mind the 10x speedup is only a minimum threshold, and the end results may be even better than originally projected. It all depends on which data is introduced and how the machine learning algorithm was designed.

Under the hood, this concept employs hardware accelerators for big data machine learning. Using GPUs and FPGAs in the world of AI training is nothing new, but they sometimes run out of memory to hold all data points. They are less valuable to the learning process in those instances, but they can still be utilized for big data machine learning in other ways. It is an interesting way of repurposing hardware already in use, but in a slightly different capacity from what one would normally expect.

Preprocessing every data point to see if there is a mathematical dual of a point already processed is the key to success here. If such a match is discovered, the algorithm can then skip the data point, as it is already known. This process becomes more and more frequent as more data is processed. Prior to being processed, every data point is assigned an “importance score” by measuring the size of the duality. As this value drops, it becomes less important to be re-processed, and the AI will eventually ignore it altogether.

It is evident this algorithm shows a lot of initial promise. However, there is still a lot of tweaking to be done before it can be commercialized. For now, development will continue in IBM’s Cloud, where it is known as Duality-Gap-Based Heterogeneous Learning. This is a watershed development in the world of big data processing, to say the very least. It will be interesting to see how this technology is made use of in the real world over the next few years.

JP Buntinx

JP Buntinx is a FinTech and Bitcoin enthusiast living in Belgium. His passion for finance and technology made him one of the world's leading freelance Bitcoin writers, and he aims to achieve the same level of respect in the FinTech sector.

Share
Published by
JP Buntinx

Recent Posts

Aptos (APT) and Tron (TRX) Prices Slide, As Volume Soars For Rollblock Suggesting Parabolic Rally

As Aptos and Tron prices take a recent downturn, the spotlight shifts to Rollblock, whose…

3 hours ago

Altcoins to Watch in November: Binance Coin (BNB), Rollblock (RBLK), and Neiro (NEIRO)

As the crypto markets roll into their most bullish time of year, we present three…

3 hours ago

Analysts Forecast $1 for Cardano and Lunex Network As Dogwifhat Plunges To Former Lows

As the crypto market prepares for a major rally, experts believe that two top altcoins,…

3 hours ago

Retail Traders Panic Sell During ‘Fake Dip’; Whales Hold Tight to SOL, DTX, and SHIB for a Millionaire-Maker Bull Run

Solana (SOL): A Strong Ecosystem Despite Volatility Solana (SOL) has been all over the place…

4 hours ago

Llama 3.2 Predicts Price For Dogecoin: $2 Peak By 2025 And $5 Rally For DTX Exchange This Winter

Cryptocurrency trends are keen on the forecast that was recently released by Llama 3.2 model…

4 hours ago

Crypto Whale Sparks 8x Surge In $OPK Price with Massive Buy-in

A mysterious crypto whale, who previously invested 9,600 SOL into tokens $Pnut and $FRED, has…

6 hours ago