Let's Talk AI Trends
- Authors
- Name
- Andy Cao
Table of content
A Break from the Usual
This month has been absolutely hectic for me, so I figured I'd switch things up a bit. Instead of diving into my usual routine, I thought it might be fun to explore some of the latest trends in AI that have caught my eye recently. I've been soaking up a lot of information, reading articles, watching videos, and getting insights from AI itself, which even helped proofread this post!
Building on the Past, Not Replacing It
It's amazing to think about how far AI has come and where it's headed. We've journeyed from the early days of rule-based systems and decision trees, moved through the era of support vector machines (SVMs), and landed squarely in the deep learning and Transformer models age. But here's the kicker—even with all the hype around Transformer and LLMs these days, it's not like we've thrown everything else out the window.
Looking back, AI/ML's evolution hasn't been about discarding the old for the new; it's been more about layering on top of what already exists. Take SVMs, for instance. They were all the rage in the early 2000s, and even though deep learning has taken the spotlight since, SVMs are still in play. They're particularly useful in situations where data is limited or where you really need to understand what's going on under the hood of the model. Deep learning didn't kill off SVMs—it just added more tools to the AI toolkit.
The real game-changer came around 2012 when deep learning took off, thanks to neural networks blowing past previous methods in many applications. And then, in 2017, Transformer models arrived on the scene, shaking up everything from natural language processing (NLP) to other AI domains. But, like every other AI trend, the current dominance of Transformers might just be another chapter in the story, with something new likely waiting in the wings.
Modular vs. End-to-End Models
However, things get tricky when the problems are too interconnected for such a modular division. This approach can struggle with fluid situations where tasks can't be neatly separated. It's not a flaw in the modular approach itself, but it does show its limitations, especially with today's massive AI models running on traditional computing architectures, where data storage and computation are separate.
On the flip side, there are end-to-end models, which are a whole different beast. These models don't bother with breaking things down. Instead, they learn everything from the data in one go, creating a single, cohesive system. This has led to some incredible advancements, particularly in areas like computer vision and NLP. But there's a downside—these models can be pretty opaque, often functioning like a ‘black box' where it's tough to see how decisions are being made.
What's interesting is how both approaches—modular and end-to-end—depend heavily on the underlying hardware. The efficiency and capability of the hardware can either amplify the strengths of these models or exacerbate their weaknesses. This brings us to another crucial aspect of AI's evolution: the role of hardware in driving forward these algorithmic innovations.
Hardware: The Unsung Hero of AI's Evolution
But as AI models get more complex, the cracks in our current hardware start to show. The traditional GPU setup, based on the Von Neumann architecture, is beginning to look outdated. It's power-hungry, and much of that energy is wasted moving data around rather than processing it.
Contrast this with the human brain, which operates on around 20 watts to handle 80 billion neurons. Meanwhile, today's AI models running on GPUs guzzle thousands of watts to manage a similar number of parameters. Clearly, we need more efficient hardware designs.
Enter compute-in-memory architecture—a potential game-changer. In this system, data doesn't have to travel between separate memory and processing units. Instead, everything happens in one place, which could dramatically cut energy use and boost efficiency. This kind of architecture is particularly suited to the large-scale AI models we're seeing today, offering the infrastructure needed for the next wave of AI advancements.
The Future: Exciting, but with Caveats
Current Transformer models, while powerful, are constrained by their architecture. Once trained, these models are largely static, with fine-tuning or retraining required for any updates—a process that is not continuous or automatic during operation. This static nature, while limiting, also serves as a safeguard against unpredictable evolution during deployment. However, as advancements like compute-in-memory architectures emerge, enabling potentially more adaptive and efficient systems, these limitations may be reduced. This shift could necessitate new approaches to AI governance and safety, particularly if models gain the ability to learn and adapt in real-time.