Accenture Ai Scaling ML and Deep Learning Models
Accenture Ai Scaling ML and Deep Learning Models
Accenture Ai Scaling ML and Deep Learning Models
Copyright © 2024 Amazon Web Services, Inc. and/or its affiliates. All rights reserved.
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Amazon's trademarks and trade dress may not be used in connection with any product or service
that is not Amazon's, in any manner that is likely to cause confusion among customers, or in any
manner that disparages or discredits Amazon. All other trademarks not owned by Amazon are
the property of their respective owners, who may or may not be affiliated with, connected to, or
sponsored by Amazon.
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Table of Contents
Abstract and introduction ................................................................................................................ i
Abstract ........................................................................................................................................................... 1
Are you Well-Architected? .......................................................................................................................... 1
Introduction ................................................................................................................................................... 2
Frictionless ideation to production ........................................................................................................... 2
Workforce analytics use cases ........................................................................................................ 3
An intelligent platform approach .................................................................................................. 4
ML architecture on AWS ................................................................................................................. 5
Feature engineering ........................................................................................................................ 6
The feature store .......................................................................................................................................... 6
The algorithms ................................................................................................................................. 8
Data engineering and data quality ........................................................................................................... 8
Hyper-parameter tuning (HPT) .................................................................................................................. 8
Model registry ............................................................................................................................................. 10
Optimization .................................................................................................................................. 11
Optimization drivers .................................................................................................................................. 11
Fine-tuning and reuse of models ........................................................................................................... 11
Scaling with distributed training ............................................................................................................ 13
Avoiding common missteps to reduce rework ..................................................................................... 13
Machine learning pipelines ........................................................................................................... 15
Going from POC to large-scale deployments ...................................................................................... 15
Applying software engineering principles to data science ........................................................... 16
Machine learning automation through pipelines ................................................................................ 17
Tracking lineage .......................................................................................................................................... 18
Monitoring for performance and bias .......................................................................................... 19
Post-training bias metrics ........................................................................................................................ 20
Monitoring performance ........................................................................................................................... 20
Data quality monitoring ...................................................................................................................... 20
Dealing with drifts ..................................................................................................................................... 21
Augmented AI ................................................................................................................................ 22
Human-in-the-loop workflows ................................................................................................................ 22
Updating model versions .............................................................................................................. 24
Conclusion ...................................................................................................................................... 25
Contributors ................................................................................................................................... 26
iii
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
iv
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Abstract
Today, there is a real-time, global, tectonic shift in the workplace caused by digital transformation.
Accelerated by the Covid pandemic, this digital transformation has created never-seen-before
opportunities and significant workplace disruption. Fully realizing the new market opportunities
demands a modernized workforce. A skills gap contributed to by several factors exist in today's
labor market. Some of these factors are the increase in the number of people entering the
workforce each year, lack of relevant education, and the rise in technology which needs workers to
be equipped with new skills to help them keep up with advancements. Addressing this widening
gap between the current workforce skills and those needed for tomorrow is front and center in the
minds of every C-suite.
This whitepaper outlines an innovative, scalable and automated solution using deep learning (DL)
and machine learning (ML) on Amazon Web Services (AWS), to help solve the problem of bridging
the existing talent and skills gap for both workers and organizations. Combining advanced data
science, ML engineering, deep learning, ethical artificial intelligence (AI), and MLOps on AWS,
this whitepaper provides a roadmap to enterprises and teams to help build production-ready ML
solutions, and derive business value out of the same.
In the Machine Learning Lens, we focus on how to design, deploy, and architect your machine
learning workloads in the AWS Cloud. This lens adds to the best practices described in the Well-
Architected Framework.
Abstract 1
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
For more expert guidance and best practices for your cloud architecture—reference architecture
deployments, diagrams, and whitepapers—refer to the AWS Architecture Center.
Introduction
Today’s rapidly changing environment demands the ability for organizations to adapt to change
to create a sustainable and productive workforce. Thriving in this environment requires rapid
adaptation and readiness for upskilling the workforce for tomorrow.
According to VentureBeat, about 87% of ML models never make it to production. Even though
9 out of 10 business executives believe that AI will be at the center of the next technological
revolution, completion, and successful production deployment is seen as a big challenge as it
requires specific engineering expertise and collaboration between several teams (ML engineering,
IT, Data Science, DevOps, and so on).
Accenture has built a scalable, industrialized, AI-powered solution that is a key component in
helping solve the talent and skilling problem of today and tomorrow to create a productive
workforce. It describes an innovative, cloud-native AWS approach that can be taken to industrialize
the ML solution, and help organizations bridge the skills gap.
This whitepaper describes a technical solution (also referred to as industry solution) for building
and scaling ML, and specifically, DL models for these use cases, and how Accenture is industrializing
the end-to-end process to achieve the technical goals previously detailed. The technical thought
process explained here can be expanded and applied to most problems in other industries. You can
also use it to create a stable and sustainable Enterprise AI system.
Introduction 2
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
To bring about a balance between responsible AI and computing at scale, Accenture created
an advanced, reusable, AWS MLOps architecture for multiple workforce productivity use cases,
including:
• Solutions.AI for Talent and Skilling — An AI-powered solution that delivers intelligent
insights to help close the skills gap in any organization. Through Solutions.AI, Accenture created
enterprise-wide AI solutions that deliver game-changing results, fast.
• FutureofU: Skills. Jobs. Growth — An AI-enabled platform solution created in collaboration
with Accenture partners, who are committed to accelerating a smooth transition to help get
diverse talent skilled and integrated into the AI workforce.
• Intelligent Workforce Insights – Uses the power of cutting-edge AI/ML technologies to develop
role maps that identify declining, stable, and emerging skills, both internally and in the market,
and assess roles for upskilling.
The challenge with these use cases is that the backend needs massive amount of computing
resources to infer and produce the ML score results. For optimal user experience, the scoring needs
to happen fast enough, with near real-time intermediate scores. All three of these use cases have
three common technical traits:
• ML engineering at scale
• AI/ML automation with responsible AI
• End-to-end productionized large-scale resilient AI/ML systems
3
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
For businesses to derive value, ML and DL models need to be productionized, run at scale, and
reused across organizations. Lack of scalability, repeatability, and manual processes diminish any
value that would be otherwise realized from these powerful models.
Complete solution for scaling and productionizing DL models with automated pipelines
This proposed architecture in the previous diagram is designed to help achieve three goals:
4
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
ML architecture on AWS
Expanding on the previous architecture, the following architecture is a drill-down of how a
complete solution can be built with various AWS components and connected for a seamless,
resilient, production-grade solution that is driven by performance and easy to maintain.
5
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Feature engineering
Many DL and ML models are used for the workforce productivity solution; however, text
classification and sentence prediction are inherently the main classifiers you need. Given the
superior performance of neural language models, and because it enables machines to understand
qualitative information, it fits the need of building neural network-based DL models for assessing
peoples’ skills proficiency, and for recommending new career pathways.
Bidirectional Encoder Representations from Transformers (BERT) is the first Natural Language
Processing (NLP) technique to rely solely on a self-attention mechanism, which is made possible by
the bidirectional transformers at the center of BERT's design. This is significant because a word may
change meaning as a sentence develops. Each word added augments the overall meaning of the
sentence, and the context may completely alter the meaning of a specific word.
After Amazon SageMaker Processing applies the transformations defined in the SageMaker Data
Wrangler, the normalized features are stored in an offline feature store so the features can be
shared and reused consistently across the organization among collaborating data scientists. This
means SageMaker Processing and Data Wrangler can be used to generate features, and then store
them in a feature store. This standardization is often key to creating a normalized, reusable set of
features that can be created, shared, and managed as input into training ML models. You can use
this feature consistency across the maturity spectrum, whether you are a startup or an advanced
organization with an ML center of excellence.
The Amazon SageMaker Feature Store is accessible across the organization for different teams to
collaborate, promoting reuse, reducing overall cost, and avoiding silos with duplicate work efforts.
The following query is a sample of the central Feature Store created with BERT embeddings. A
SageMaker Feature Group and a Feature Store are created. Multiple downstream teams can retrieve
and use features from this central store instead of redoing feature engineering repeatedly, adding
to the organization’s operational costs and non-standardization issues.
Feature Store with BERT embeddings ready for reuse across the organization
The algorithms
There are three key pillars to building successful ML applications. If not done correctly, in spite of
all the state-of-the-art continuous integration/continuous delivery (CI/CD), feature store, feature
engineering, and graphics processing unit (GPU)-accelerated DL or automated pipelines, the end-
to-end Enterprise AI platform is bound to fail:
For the talent and skilling industry use cases defined earlier, the DL models need to classify
millions of jobs and skills to predict a good match and user learning sequence. The following
details are some of the things that we found useful and were key in our thought leadership for
creating AI solutions. We define the objective metric that the HPT job will try to optimize, which is
validation accuracy for the talent and skilling use cases. The following is an example code snippet
for the metrics definition:
objective_metric_name = "validation:accuracy"
metrics_definitions = [
{"Name": "train:loss", "Regex": "loss: ([0-9\\.]+)"},
{"Name": "train:accuracy", "Regex": "accuracy: ([0-9\\.]+)"},
{"Name": "validation:loss", "Regex": "val_loss: ([0-9\\.]+)"},
{"Name": "validation:accuracy", "Regex": "val_accuracy: ([0-9\\.]+)"},
]
A crucial setting is the early_stopping_type, which you set so that SageMaker can stop the
tuning job when it starts to overfit, and can help save cost of the overall tuning job. Inaccurate
hyperparameter tuning can not only result in excessive costs, but also an ineffective model even
after hours of training.
objective_metric_name = "validation:accuracy"
tuner = HyperparameterTuner(
estimator=estimator,
objective_type="Maximize",
objective_metric_name=objective_metric_name,
hyperparameter_ranges=hyperparameter_ranges,
metric_definitions=metrics_definitions,
max_jobs=2,
max_parallel_jobs=10,
strategy="Bayesian",
early_stopping_type="Auto",
)
Combining all of this together, you have the following build and training process taking BERT
as an example. Other DL models built with PyTorch, MXNet, or TensorFlow follow the same
process. It is essential to get the following three stages (within the box under MACHINE LEARNING
ENGINEERING) correct to move on to productionizing the system with large scale model
deployments.
Model registry
It is important to catalog models to explain the model predictions and insights. It is also important
that all models promoted to production are cataloged, all model versions managed, metadata such
as training metrics are associated with a model, and the approval status of a model is managed.
This is especially needed when organizations want to move from ad-hoc one-off proof-of-concepts
to embedding AI in their enterprise systems with multiple teams, doing daily DL experiments. This
is implemented in the solution using SageMaker Model Registry.
Model registry 10
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Optimization
DL is simple in essence. In the last few years, AWS has achieved astonishing results on machine-
perception problems with the help of simple parametric models trained with gradient descent (GD).
Extending that, all that is needed at the core is sufficiently large parametric models trained with
GD on a large dataset.
Creating a DL algorithm or identifying the algorithm to use and fine-tune is the first step. The
next step for an enterprise is to derive business value out of the algorithm. That can be achieved
only when the models are appropriately industrialized, scaled, and continuously improved. Ill-
performing models negatively impact a business or organization’s bottom line. In Accenture’s
talent and skilling solution, there are over 50 models running in Production, making a large-scale,
smooth, operationalization process a necessity.
Optimization drivers
DL has positioned itself as an AI revolution and is here to stay. Some of the benefits of using DL
models are:
• Reusability
• Scalability
Optimizing and scaling ML and DL models in production is a crucial set of tasks, and one that
must be done with finesse. To maximize the benefits listed previously, a proper implementation
approach must be taken.
Following are details on how it should be implemented for the industry use cases, taking the
example of a few of the models. The same approach can be used for scaling many other DL models
for new problems.
Optimization drivers 11
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
helps reuse the previous HPT job results, but along with new hyperparameters. This speeds up
converging on the best model faster.
This is particularly important in Enterprise AI systems, as multiple teams may want to reuse the
models created. Training DL models from scratch requires a lot of GPU, compute, and storage
resources. Model reuse across the organization helps in reducing costs. Therefore, a useful
technique for model reuse is fine-tuning. Fine-tuning methodology involves unfreezing a few of
the top layers of a frozen model base for feature extraction, and then jointly training both the
newly added part of the model, which is the fully connected classifier and top layers. With this, a
model can be reused for a different problem, and does not have to be re-trained, saving costs for
the company.
In the following sections, you will see how you can implement and scale the model fine-tuning
strategies previously discussed, while maintaining a laser focus on the business metrics we need to
attain.
WarmStartConfig uses one or more of the previous hyper-parameters tuning job runs called the
parent jobs, and needs a WarmStartType.
estimator = TensorFlow(
entry_point="tf_bert_reviews.py",
source_dir="src",
role=role,
instance_count=train_instance_count,
instance_type=train_instance_type,
volume_size=train_volume_size,
py_version="py37",
framework_version="2.3.1",
hyperparameters={
"epochs": epochs,
"epsilon": epsilon,
"validation_batch_size": validation_batch_size,
"test_batch_size": test_batch_size,
"train_steps_per_epoch": train_steps_per_epoch,
"validation_steps": validation_steps,
"test_steps": test_steps,
"use_xla": use_xla,
"use_amp": use_amp,
"max_seq_length": max_seq_length,
"enable_sagemaker_debugger": enable_sagemaker_debugger,
"enable_checkpointing": enable_checkpointing,
"enable_tensorboard": enable_tensorboard,
"run_validation": run_validation,
"run_test": run_test,
"run_sample_predictions": run_sample_predictions,
},
input_mode=input_mode,
metric_definitions=metrics_definitions,
objective_metric_name = "train:accuracy"
tuner = HyperparameterTuner(
estimator=estimator,
objective_type="Maximize",
objective_metric_name=objective_metric_name,
hyperparameter_ranges=hyperparameter_ranges,
metric_definitions=metrics_definitions,
max_jobs=2,
max_parallel_jobs=1,
strategy="Bayesian",
early_stopping_type="Auto",
warm_start_config=warm_start_config,
)
Secondly, transforming data science prototype scripts from experimentation phase to modular
performant code for production is a deeply involved task, and if not done right, will not produce a
stable production system.
Finally, the ecosystem of ML engineering and MLOps is a culmination of multiple processes and
standards from within DevOps, adding in ML-specific tooling and domain-specific elements,
thereby building repeatable, resilient, production-capable data science solutions on the cloud.
These three tenets alone distinguish a matured AI/ML enterprise from one that has just started in
the journey of using ML for deriving business value.
For industry solutions, as mentioned in the Workforce analytics use cases section of this document,
following are some optimizations that have proved useful to having an efficient, enterprise-grade,
end-to-end, industrialized, ML solution:
In a scaled, industrialized, production version, the full end-to-end automated data engineering
and ML engineering pipeline is the product built on the data science scripts in the experimentation
phase.
Moving from a single laptop setup toward a scalable, production-grade, data science platform is
a completely different challenge from the proof-of-concept stage, and arguably, one of the most
difficult, as it involves collaborating with different teams across an organization. Accenture has
devised a scalable and unique approach for the Accenture talent and skilling AI solutions discussed
in this whitepaper, to go from prototype to full-scale productionized systems in a very short period
of time; enhancing “speed to market” and generating value.
This and the following sections address these problems, and provide a solution.
• Repeatability
• Scalability
• Transparency/Explainability
ML workflow and process with multiple teams that need to collaborate to create a complete ML
solution in production
The single biggest driving factor in making a successful project that has the fewest amount of
rework is that of collaborative involvement between the ML team and the business unit. The
second biggest factor to ensure success is communication within the ML teams.
Decades ago, software engineers refined their processes from large-scale waterfall
implementations to a more flexible agile process. That was a game-changer in terms of bringing
ideas to market faster and more efficiently. Similarly, now ML engineering seeks to define a new set
of practices and tools that will optimize the whole unique realm of software development for data
analysts and scientists.
The idea of ML engineering and MLOps as a paradigm is rooted in the application of similar
principles that DevOps has to software development.
Automated ML pipelines ensure one vital feature and requirement of a stable live enterprise AI
platform: repeatability and reproducibility. Instead of manual ad-hoc Python scripts that may
propagate data quality issues from sources, pipelines help ensure the issues are caught, handled,
retried, or logged at every step of the pipeline.
Implementing these design principles, the following diagram summarizes a modern and effective
way to implement pipelines on AWS. This section describes the key ML processes, and the stages
for creating a modern and performant ML pipeline on AWS.
The ML engineering process and steps to create modern and effective ML pipelines on AWS
Tracking lineage
Model artifact lineage should be recorded and tracked at every stage. This is the only way you can
tie back the model to the dataset it was trained on, the features used, the validation dataset used,
and the approval process and deployment history. For tracking and looking at the lineage artifacts
for our processing job, you can do the following in SageMaker, and the same should be set up as
part of the pipeline artifact lineage.
SageMaker Lineage Tracking API has lineage traversal, contexts, and associations among other
concepts that help in creating a lineage history for all stages of SageMaker Pipelines, training jobs,
endpoints, SageMaker processing jobs, and models, as shown in the following image. Lineage
of feature engineering jobs can and should be tracked in the same way. Amazon SageMaker ML
Lineage Tracking helps in storing information about all the stages of a ML workflow from data
preparation to model deployment.
Tracking lineage 18
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Bias detection with SageMaker Clarify and nearly continuous monitoring with SageMaker Model
Monitor
DL models can be heavily impacted by data bias. Model and data bias detection and rectification
should be constant underlying themes in an Enterprise AI system. You can use SageMaker Clarify
extensively to perform evaluation for detecting data bias, during featuring engineering for
evaluating feature importance, for assessing model bias during training/hyper-parameter tuning
and finally with SageMaker Model Monitor to take actions on live models.
An important need is to generate reports throughout all these stages to maintain transparency
of the process. Using SageMaker Data Wrangler along with SageMaker Clarify, you can generate
reports that explain the features that are considered important, the choices made by the model,
and the reasoning behind the predictions.
Data drift, covariate, label drifts, and concept shift mandate nearly continuous monitoring
and updates to models. Nearly continuous monitoring of model insights, measuring them, and
testing their production effectiveness are critical steps to achieve a successful Enterprise AI.
Baseline metrics first need to be calculated during model training and thresholds should be set for
Kullback–Leibler (KL), Kolmogorov-Smirnov (KS), and Linear Programming (LP) Key Performance
Indicators (KPIs).
19
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
SageMaker Model Monitor detects drift on live model output which can be integrated with
SageMaker Clarify for various detections and takes corrective actions, such as re-training models,
introducing new production variants, and retiring older non-performing models.
Because our focus is always on the end-state, the models and all supporting infrastructure and
data feeds that are employed to solve a problem need to have a measurable characteristic tied to
business outcomes.
For all post-training data and model bias metrics, you can use SageMaker Clarify. Taking the
trained model, you choose the feature to analyze for bias and determine the conditional rejection
(DCR) and other metrics listed previously. Because SageMaker Clarify supports SHAP, you can
determine the contribution of each feature to the model prediction. This helps in performing
feature attribution and generating the model explainability report with SageMaker Clarify. This
provides you with all the information you need for detecting drifts in feature attribution and model
explainability on live model endpoints – a key business goal in ensuring you have responsible AI
baked in into all your solutions.
Monitoring performance
Earlier, this whitepaper described a process of creating data-quality baseline using Deequ, which
helps detect drift in the statistical characteristics of input data being sent to the live model. If
this step is not accurately performed in the ML and DL pipelines, the rest of the components
downstream are deeply impacted, resulting in incorrect or sub-optimal results.
• Feature drift, Label drift and Concept drift, which can be handled by either revisiting feature
engineering, model re-training, or training on new data.
• Prediction drift and Feedback drift, which require more involved approach of new versions of
the models as otherwise it may impact the business objectives.
Regardless of the types and causes of drifts, it is vital to monitor model metrics associated with
training for passive retraining, and the model attribution data for active retraining. Monitoring
these shifts in model efficacy can help with early intervention (Model Monitor), explainable
analytics reports (SageMaker Clarify), and the ability to resolve the issue in a way that will not
cause disruption to the project.
Augmented AI
While you have multiple continually running automated ML pipelines that save the final batch
predictions and insights in a curated zone, it is essential to insert human supervision and guidance
in the automated AI/ML workflows. Humans can provide the necessary critical quality assurances
before pushing sensitive models into production to help the models learn better. Use Amazon
Augmented AI (A2I) along with SageMaker Ground Truth to bring together ML and humans, to
provide automation while continuously improving your models.
With Amazon SageMaker, A2I, Ground Truth, Amazon CloudFront, and Amazon Cognito, you have
a fully functional web application validation system that lets your functional domain experts and
technical leads validate samples of model predictions, which in turn helps quickly optimize and
fine-tune your predictive models.
Human-in-the-loop workflows
In the workforce productivity use cases, one important business requirement is an unbiased,
efficient, and effective model for assessing a person’s current skills based on CV, professional
profile, so on. While you can use custom entity recognition and customize BERT layers (not just the
classifier top layers) for roles, skills, titles, and so on, it is essential to integrate human oversight
in the entire workflow involving many models in the pipeline (custom models, Amazon Textract,
Amazon Comprehend, fine-tuned BERT) and decision-making process, so that your models learn
better.
It is important to know how to make this workflow function well, as it will impact the business
outcomes if an appropriate automated workflow is not put in place to fix low-confidence
predictions and improve the models. For example, in the industry use case, there has been a regular
need to parse a document with a person’s work history and professional skills, which can be in
virtually any format and predicting skill proficiency scores.
Using Amazon Textract (which offers AI powered extraction of text and structured data from
documents) and a series of models in an inference pipeline, AWS was able to provide the
recommended insights. This also allowed AWS to integrate human judgement into the workflow
wherever needed with A2I, to help the models learn better and improve over time.
A front-end web application integrated with augmented AI instills confidence in the predictions
and recommendations being made by the models. This is especially important in creating and
Human-in-the-loop workflows 22
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
maintaining a matured, productionized, enterprise AI system such as the ones in our use case. The
Ground Truth labels that were acquired using A2I and human-in-the-loop workflow are actively
used with SageMaker Model Monitor to continuously evaluate concept drift. For all other drifts
in model bias, feature importance, and explainability, use Model Monitor to compare against the
baseline provided to each SageMaker deployed endpoint.
Human-in-the-loop workflows 23
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
When new versions of the models are available, it is important to see which variant performs
better before routing all traffic to the new model. Models tested during the engineering process
and the same models running “in the wild” may produce different results, and you don’t want
to switch your use base to the new models without first testing them sufficiently with live,
unseen, real traffic. Use Canary testing, A/B testing, and Multi-Armed Bandit (MAB) strategies in
a champion-challenger framework, to choose the best new model among the various SageMaker
hosted production variants.
24
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Conclusion
Taking data science algorithms from experimentation to large-scale production requires an end-
to-end ML engineering process in place to maintain and manage an Enterprise AI system. This
becomes a more involved process for scaling DL models, as in the use cases described here. This
whitepaper describes an approach to scale ML and DL models in production and reduce the time
from “ideation to production”, while minimizing cost and maximizing gains for the organization.
25
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Contributors
Contributors to this document include:
26
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Further reading
• Data Science on AWS by Chris Fregly and Antje Barth - Released April 2021, Publisher(s): O'Reilly
Media, Inc.
• Machine Learning Engineering in Action by Ben Wilson, Manning Publications, Published March
2022, ISBN 9781617298714
• Deep Learning with Python by François Chollet, Manning Publications; Published November
2017, ISBN 9781617294433
• AWS Architecture Center
27
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
About Accenture
Accenture is a global professional services company with leading capabilities in digital, cloud and
security. Combining unmatched experience and specialized skills across more than 40 industries,
we offer Strategy and Consulting, Technology and Operations services and Accenture Song —
all powered by the world’s largest network of Advanced Technology and Intelligent Operations
centers. Our 699,000 people deliver on the promise of technology and human ingenuity every
day, serving clients in more than 120 countries. We embrace the power of change to create value
and shared success for our clients, people, shareholders, partners, and communities. Visit us at
accenture.com.
28
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Document revisions
To be notified about updates to this whitepaper, subscribe to the RSS feed.
29
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
Notices
Customers are responsible for making their own independent assessment of the information in
this document. This document: (a) is for informational purposes only, (b) represents current AWS
product offerings and practices, which are subject to change without notice, and (c) does not create
any commitments or assurances from AWS and its affiliates, suppliers or licensors. AWS products or
services are provided “as is” without warranties, representations, or conditions of any kind, whether
express or implied. The responsibilities and liabilities of AWS to its customers are controlled by
AWS agreements, and this document is not part of, nor does it modify, any agreement between
AWS and its customers.
© 2022 Amazon Web Services, Inc. or its affiliates. All rights reserved.
30
Accenture Enterprise AI – Scaling Machine Learning and Deep Learning Models AWS Whitepaper
AWS Glossary
For the latest AWS terminology, see the AWS glossary in the AWS Glossary Reference.
31