⬆ ️ Bump torch-model-archiver from 0.3.1 to 0.4.2
Bumps torch-model-archiver from 0.3.1 to 0.4.2.
Release notes
Sourced from torch-model-archiver's releases.
TorchServe v0.4.2 Release Notes
This is a hotfix release of TorchServe v0.4.2.
Improvements
- Fixed the issue of port sharing between management and inference API
- Fixed the issue of cleaning up tmp dir in model archiver
- Fixed KFServing dockerfile
TorchServe v0.4.1 Release Notes
This is the release of TorchServe v0.4.1.
New Features
- PyTorch 1.9.0 support - TorchServe is now certified working with torch 1.9.0 torchvision 0.10.0, torchtext 0.10.0 and torchaudio 0.9.0
- Model configuration support - Added support for model performance tuning on SageMaker via model configuration in config.properties.
- Serialize config snapshots to DynamoDB - Added support for serializing config snapshots to DDB.
- Prometheus metrics plugin support - Added support for Prometheus metrics plugin.
- Kubeflow Pipelines support - Added support for Kubeflow pipelines and Google Vertex AI Manages pipelines, see examples here
- KFServing docker support - Added production docker for KFServing.
- Python 3.9 support - TorchServe is now certified working with Python 3.9.
Improvements
- HF BERT models multiple GPU support - Added multi-gpu support for HuggingFace BERT models.
- Error log for customer python package installation - Added support to log error of customer python package installation.
- Workflow documentation optimization - Optimized workflow documentation.
Tooling improvements
- Mar file automation integration - Integrated mar file generation automation into pytest and postman test.
- Benchmark automation for AWS neuron support - Added support for AWS neuron benchmark automation.
- Staging binary build support - Added support for staging binary build.
Platform Support
Ubuntu 16.04, Ubuntu 18.04, MacOS 10.14+, Windows 10 Pro, Windows Server 2019, Windows subsystem for Linux (Windows Server 2019, WSLv1, Ubuntu 18.0.4)
GPU Support
Torch 1.9.0 + Cuda 10.2, 11.1 Torch 1.8.1 + Cuda 9.2, 10.1
TorchServe v0.4.0 Release Notes
This is the release of TorchServe v0.4.0.
New Features
- Workflow support - Added support for sequential and parallel ensemble models with Language Translation and Computer Vision classification examples.
- S3 Model Store SSE support - Added support for S3 server side model encryption via KMS.
- MMF-activity-recognition model example - Added example MMF-activity-recognition model
- PyTorch 1.8.1 support - TorchServe is now certified working with torch 1.8.1, torchvision 0.9.1, torchtext 0.9.1, and torchaudio 0.8.1
Improvements
- Fixed GPU memory high usage issue and updated model zoo - Fixed duplicate process on GPU device .
- gRPC max_request_size support- Added support for gRPC max_request_size configuration in config.properties.
- Non SSL request support - Added support for non SSL request.
... (truncated)
Commits
-
3afb15fMerge branch 'master' into release_0.4.2 -
cdca64cMerge pull request #1173 from jagadeeshi2i/add-build-script -
f83473eMerge branch 'master' into add-build-script -
cad4b35update version -
cc9ed29Merge pull request #1016 from pytorch/issue_1012 -
0742fb4Merge branch 'master' into issue_1012 -
092ecf0Merge pull request #1175 from Silas606/dev -
5779e75fix pytest error -
926216dMerge branch 'master' into dev -
6ad05b6Merge branch 'master' into issue_1012 - Additional commits viewable in compare view
Dependabot commands
You can trigger Dependabot actions by commenting on this MR
-
$dependabot rebasewill rebase this MR -
$dependabot recreatewill recreate this MR rewriting all the manual changes and resolving conflicts