Inference Time
The moment when a trained AI model receives new input and produces an output or prediction.
In Plain English
Inference time is when an AI model that has already been trained and is ready to use receives new data and makes a prediction or generates text. It's the opposite of training time, when the model is learning from historical data. Inference time happens when you ask ChatGPT a question, upload a photo to facial recognition, or use an AI email filter—the model is 'inferring' or working out an answer based on what it learned during training.
💡Real-World Example
You ask a trained medical imaging AI to review your X-ray. That moment when the model receives your image and delivers its analysis is inference time. The model isn't learning anything new during this moment; it's applying what it learned during months of training to your specific X-ray.
Related Terms
What did you think of our explanation?
