Aussie AI
Commercialized SOTA Research
-
Book Excerpt from "Generative AI in C++"
-
by David Spuler, Ph.D.
Commercialized SOTA Research
The whole of the AI industry is based on commercialization of State-of-the-Art (SOTA) research, from the basic Transformer architecture to the often-used optimizations of quantization and pruning. The areas where recent research from the last year or two is starting to appear in industry models and open source frameworks, with kudos paid to the innovative researchers, includes:
- Flash Attention (linearized attention algorithm)
- Flash Decoding
- RoPE (positional encoding for long contexts)
- Long context window research
Just my opinion, but the areas that seem ripe for greater inclusion in commercial and open source industry AI work include:
- Early exit (dynamic layer pruning)
- Integer-only arithmetic quantization models (end-to-end integers)
• Next: • Up: Table of Contents |
The new AI programming book by Aussie AI co-founders:
Get your copy from Amazon: Generative AI in C++ |