Generative AI Research Literature Survey
Aussie AI
-
edited by David Spuler, Ph.D.
Overview of our active research interests.
Making AI models run faster (down the street).
Will AI spawn a revolution or evolution?
Change your AI engine into a faster Transformer.
Do big models really belong on a small phone?
Multiplication has a bad reputation for running slow.
Measure once, cut a billion times.
AI is coming to a desktop near you.
Take the clippers to cut the links down to size.
Who knew that AI inference was just coding.
Your AI's band should be called The Logarithmics.
A token of our appreciation for tokenizer research.
A quick way to make a real hash of AI models.
AI's are regressing back to their early model-hood.
Getting stuck in the quagmire of embeddings research.
Abandon all hope, ye who enter here.
Are the AI engines working together?
Pruning of the input text tokens.
Like cutting a piece of AI cake, except layer-wise.
When are two heads worse than one?
The third type of model pruning is length-wise
Should models percolate or distill their knowledge?
Making your AI models approximately intelligent.
Your Transformer without any Feeding Forward.
A good AI is a shallow AI.
Apparently it is normal to prune.
AI models are slimmable; who knew?
Peas and carrots go together like matrices and vectors.
Everyone knows that zeroes should be skipped.
A useful innovative inference ideas index.
Your AI might escape by early exiting.
Weights just sit around, so let's precompute.
Going round in circles, inferencing all the way.
Your AI engine will probably be smart.
How to make an AI model in full bloom.
Pruning left-to-right and up-down, from two dimensions.
Padding with zeros works, but fastest is zero zero-padding.
The kindest AI engines always share parameters.
With so many layers, AI models are very fusing.