Expand description
AWS Lambda Inference Deployment
Deploy and manage ML inference on AWS Lambda for serverless, pay-per-use inference.
§Features
- Model packaging with Docker/OCI containers
- Cold start optimization with provisioned concurrency
- Automatic scaling with Lambda’s built-in capabilities
- Integration with Pacha registry for model artifacts
§Toyota Way Principles
- Muda Elimination: Pay only for actual inference compute
- Heijunka: Automatic scaling levels inference load
- Jidoka: Built-in error handling and retry logic
Structs§
- Deployment
Estimate - Deployment estimate
- Inference
Request - Lambda inference request
- Inference
Response - Lambda inference response
- Lambda
Client - Lambda inference client
- Lambda
Config - Lambda function configuration for inference
- Lambda
Deployer - Lambda deployment manager
- VpcConfig
- VPC configuration for Lambda
Enums§
- Config
Error - Configuration error
- Deployment
Error - Deployment error
- Deployment
Status - Deployment status
- Lambda
Architecture - Lambda architecture
- Lambda
Runtime - Lambda runtime environment