Releases: oracle/accelerated-data-science
Releases · oracle/accelerated-data-science
ADS 2.8.10
- Improved the
LargeArtifactUploaderclass to understand OCI paths to upload model artifacts to the model catalog by reference. - Removed
ADSDatasetruntime dependency ongeopandas. - Fixed a bug in the progress bar during model registration.
- Fixed a bug where session variable could be referenced before assignment.
- Fixed a bug with model artifact save.
- Fixed a bug with pipelines step.
ADS 2.8.9
- Upgraded the
scikit-learndependency to>=1.0. - Upgraded the
pandasdependency to>1.2.1,<2.1to allow you to use ADS with pandas 2.0. - Implemented multi-part upload in the
ArtifactUploaderto upload model artifacts to the model catalog. - Fixed the "Attribute not found" error, when
deploy()called twice inGenericModel. - Fixed the fetch of the security token, when the relative path for the
security_token_fileis provided (used in session token-bases authentication).
ADS 2.8.8
- Added
PyTorchDistributedruntime option for Data Science jobs to add support for training large language models with PyTorch. - Added options to configure flexible shape in
opctl. - Refactored
deploy()inGenericModelto prioritize the parameters. - Fixed the
opctlcommands delete/cancel/watch/activate/deactivate commands to add missing parameter options. - Fixed the
opctlcommands to call run to start an ML job when no YAML is specified. - Deprecated the
DatasetFactoryclass, and refactored the code.
ADS 2.8.7
- Added support for leveraging pools in the Data Flow applications.
- Added support for token-based authentication.
- Revised help information for
opctlcommands.
ADS 2.8.6
- Resolved an issue in
ads opctl build-image job-localwhen the build ofjob-localwould get stuck. Updated the Python version to 3.8 in the base environment of thejob-localimage. - Fixed a bug that prevented the support of defined tags for Data Science job runs.
- Fixed a bug in the
entryscript.shofads opctlthat attempted to create a temporary folder in the/var/foldersdirectory. - Added support for defined tags in the Data Flow application and application run.
- Deprecated the old
ModelDeploymentPropertiesandModelDeployerclasses, and their corresponding APIs. - Enabled the uploading of large size model artifacts for the
ModelDeploymentclass. - Implemented validation for shape name and shape configuration details in Data Science jobs and Data Flow applications.
- Added the capability to create
ADSDatasetusing the Pandas accessor. - Provided a prebuilt watch command for monitoring Data Science jobs with
ads opctl. - Eliminated the legacy
ads.dataflowpackage from ADS.
2.8.5
ADS
- Added support for
key_contentattribute inads.set_auth()for the API KEY authentication. - Fixed bug in
ModelEvaluatorwhen it returned incorrect ROC AUC characteristics. - Fixed bug in
ADSDataset.suggest_recommendations()API, when it returned an error if the target wasn't specified. - Fixed bug in
ADSDataset.auto_transform()API, when an incorrect sampling was suggested for imbalanced data.
2.8.4
ADS
- Added support for creating ADSDataset from pandas dataframe.
- Added support for multi-model deployment using Triton.
- Added support for model deployment local testing in
ads opctlCLI. - Added support in
ads opctlCLI to generate starter YAML specification for the Data Science Job, Data Flow Application, Data Science Model Deployment and ML Pipeline services. - Added support for invoking model prediction locally with
predict(local=True). - Added support for attaching customized score.py when preparing model.
- Added status check for model deployment delete/activate/deactivate APIs.
- Added support for training and verifying SparkPipelineModel in Dataflow.
- Added support for generating score.py for GPU model deployment.
- Added support for setting defined tags in Data Science jobs.
- Improved model deployment progress bar.
- Fixed bug when using
ads opctlCLI to run jobs locally. - Fixed bug in Dataflow magic when using archive_uri in dataflow config.
2.8.3
ADS
- Added support for custom containers (Bring Your Own Container or BYOC) and environment variables for
ads.model.GenericModel. - Added default values for configuring parameters in
ads.model.ModelDeployment, such as default flex shape, ocpus, memory in gbs, bandwidth, and instance count. - Added support for
ads.jobs.NotebookRuntimeto use directory as job artifact. - Added support for
ads.jobs.PythonRuntimeandads.jobs.GitPythonRuntimeto use shell script as entrypoint.
2.8.2
ADS
- Remove support for Python 3.7.
- Improved the DataScienceMode.create() to support timeout argument and auto extract region from the signer and signer config.
- Support Jupyter Notebook as
entrypointwhen defining Data Science jobs withPythonRuntimeandGitPythonRuntime. - Support environment variable substitution in Data Science job names and output URI.
- Support JSON serialization of list/dictionary when assigning them as Data Science jobs environment variables.
- Support saving the notebook to output URI even if the job run failed when running a Data Science job using
NotebookRuntime. - Added
job.build()method to Data Science job to load default values from environment. - Added
DataScienceJob.fast_launch_shapes()method to list fast launch shapes available for Data Science job. - Added :doc:`HuggingFacePipelineModel class to support prepare, save, deploy and predict for HuggingFace pipelines.
- Updated Data Science job run YAML representation to include configurations inherited from the job.
- Fixed custom conda environment not showing in Data Science Job YAML specification.
- Fixed an issue where model saving was failing in notebook session without ipywidgets installed.
- Fixed "Unknown archive format" error in ads.jobs.PythonRuntime, when the source code folder name ends with "zip". List of supported archive files are: "zip", "tar.gz", "tar" and "tgz".
2.8.1
ADS
- Fixed a bug for
ads opctl runwhen--authflag is passed and image is built by ADS. - Fixed a bug in
GenericModel.save()when the work requests are not successfully populated. - Fixed a bug in
DataScienceModel.create()to when the provenance metadata is not provided.