Aussie AI

Inference Frameworks

  • Last Updated 18 November, 2024
  • by David Spuler, Ph.D.

Inference frameworks are software platforms that take a model and execute it against requests from users. Many inference frameworks also provide training and fine-tuning capabilities, but not all do. Many frameworks have been open-sourced, but there are also many that remain proprietary, and there is much competition occurring in the space.

There is much overlap between the concept of a framework and a "deep learning compiler". And there is also overlap with companies that are offering "AI cloud hosting" services, including both new startups and the major cloud hosts (e.g. Amazon AWS, Microsoft Azure, and Google GCP), which typically include both training and inference features.

Software frameworks are only one part of the AI tech stack. Read more about inference optimization, training optimization, hardware accelerators, ML compilers, and our list of common and obscure AI optimization techniques.

List of Machine Learning Frameworks

Some of the many frameworks include:

  • TensorFlow, open-sourced by Google.
  • PyTorch
  • Torch
  • MXNet
  • HuggingFace Transformers
  • LangChain
  • GGML
  • Llama.cpp
  • Llvm
  • Caffe and Caffe2
  • Theano
  • RNN
  • Keras
  • Microsoft CNTK (Cognitive Toolkit)
  • Amazon ML
  • Google Cloud AutoML
  • Microsoft Azure (various)
  • SciKit-learn

Features of ML Frameworks

Some of the desirable features include:

Survey Papers on ML Software Frameworks

Papers that review or survey software frameworks:

General Research on ML Software Frameworks

Research papers about general issues or specific frameworks:

More AI Research

Read more about: