Realtime AI Machine Learning; Computational Infrastructure
noun phrase
Definition: AI processing designed to generate predictions, decisions, or other outputs under strict latency constraints, typically while data are being received or requests are being served, rather than through delayed batch processing. In adjacent production-ML terminology, this aligns closely with online inference or real-time inference, where predictions are generated on demand at serving time [Google ML Glossary].
Example in context: “Edge computing is a very promising technology for provisioning realtime AI services due to low response time and network bandwidth requirements.” [Rahman et al. 2020]
Synonyms: real-time AI; online AI (context-dependent)
Related terms: online inference; low-latency inference; streaming analytics; edge AI; real-time inference