Free Quiz
Write for Us
Learn Artificial Intelligence and Machine Learning
  • Artificial Intelligence
  • Data Science
    • Language R
    • Deep Learning
    • Tableau
  • Machine Learning
  • Python
  • Blockchain
  • Crypto
  • Big Data
  • NFT
  • Technology
  • Interview Questions
  • Others
    • News
    • Startups
    • Books
  • Artificial Intelligence
  • Data Science
    • Language R
    • Deep Learning
    • Tableau
  • Machine Learning
  • Python
  • Blockchain
  • Crypto
  • Big Data
  • NFT
  • Technology
  • Interview Questions
  • Others
    • News
    • Startups
    • Books
Learn Artificial Intelligence and Machine Learning
No Result
View All Result

Home » AWS announces nine major updates for its ML platform SageMaker

AWS announces nine major updates for its ML platform SageMaker

SageMaker aims to provide a machine learning service that can be used to build, train, and deploy ML models for virtually any use case.

Tarun Khanna by Tarun Khanna
January 15, 2021
in Machine Learning, News
Reading Time: 5 mins read
0
aws-nine-major-updates-ml-platform-sagemaker
Share on FacebookShare on TwitterShare on LinkedInShare on WhatsApp

SageMaker aims to provide a machine learning service that can be used to build, train, and deploy ML models for virtually any use case.

Amazon Sagemaker Pipelines

Amazon Web Services (AWS) has announced nine major new updates for its cloud-based machine learning platform, SageMaker. SageMaker aims to provide a machine learning service that can be used to build, train, and deploy ML models for virtually any use case. During this year’s re Invent conference, AWS made several announcements to further improve SageMaker’s capabilities.

Also Read:

How machine learning can spark many discoveries in science and medicine

“Periodic table of machine studying” could fuel AI discovery

Making AI-generated code more correct in any language

The Rise of AI: Leading Computer Scientists anticipate a Star Trek-Like Future

Swami Sivasubramanian, VP of Amazon Machine Learning at AWS, said:

Hundreds of thousands of everyday developers and data scientists have used our industry-leading machine learning service, Amazon SageMaker, to remove barriers to building, training, and deploying custom machine learning models. One of the best parts about having such a widely-adopted service like SageMaker is that we get lots of customer suggestions which fuel our next set of deliverables.

Today, we are announcing a set of tools for Amazon SageMaker that makes it much easier for developers to build end-to-end machine learning pipelines to prepare, build, train, explain, inspect, monitor, debug, and run custom machine learning models with greater visibility, explainability, and automation at scale.

The first announcement is Data Wrangler, a feature that aims to automate the preparation of data for machine learning. Data Wrangler enables customers to choose the data they want from their various data stores and import it with a single click. Over 300 built-in data transformers are included to help customers normalize, transform, and combine features without having to write any code.

Frank Farrall, Principal of AI Ecosystems and Platforms Leader at Deloitte, comments:

SageMaker Data Wrangler enables us to hit the ground running to address our data preparation needs with a rich collection of transformation tools that accelerate the process of machine learning data preparation needed to take new products to market.

In turn, our clients benefit from the rate at which we scale deployments, enabling us to deliver measurable, sustainable results that meet the needs of our clients in a matter of days rather than months.

The second announcement is Feature Store. Amazon SageMaker Feature Store provides a new repository that makes it easy to store, update, retrieve, and share machine learning features for training and inference.

Feature Store aims to overcome the problem of storing features that are mapped to multiple models. A purpose-built feature store helps developers to access and share features that make it much easier to name, organize, find, and share sets of features among teams of developers and data scientists. Because it resides in SageMaker Studio – close to where ML models are run – AWS claims it provides single-digit millisecond inference latency.

Mammad Zadeh, VP of Engineering, Data Platform at Intuit, says:

We have worked closely with AWS in the lead up to the release of Amazon SageMaker Feature Store, and we are excited by the prospect of a fully managed feature store so that we no longer have to maintain multiple feature repositories across our organization.

Our data scientists will be able to use existing features from a central store and drive both standardisation and reuse of features across teams and models.

Next up, we have SageMaker Pipelines—which claims to be the first purpose-built, easy-to-use continuous integration and continuous delivery (CI/CD) service for machine learning.

Developers can define each step of an end-to-end machine learning workflow including the data-load steps, transformations from Amazon SageMaker Data Wrangler, features stored in Amazon SageMaker Feature Store, training configuration and algorithm set up, debugging steps, and optimization steps.

SageMaker Clarify may be one of the most important features being debuted by AWS this week considering ongoing events.

Clarify aims to provide bias detection across the machine learning workflow, enabling developers to build greater fairness and transparency into their ML models. Rather than turn to often time-consuming open-source tools, developers can use the integrated solution to quickly try and counter any bias in models.

Andreas Heyden, Executive VP of Digital Innovations for the DFL Group, says:

Amazon SageMaker Clarify seamlessly integrates with the rest of the Bundesliga Match Facts digital platform and is a key part of our long-term strategy of standardising our machine learning workflows on Amazon SageMaker.

By using AWS’s innovative technologies, such as machine learning, to deliver more in-depth insights and provide fans with a better understanding of the split-second decisions made on the pitch, Bundesliga Match Facts enables viewers to gain deeper insights into the key decisions in each match.

Deep Profiling for Amazon SageMaker automatically monitors system resource utilization and provides alerts were required for any detected training bottlenecks. The feature works across frameworks (PyTorch, Apache MXNet, and TensorFlow) and collects system and training metrics automatically without requiring any code changes in training scripts.

Next up, we have Distributed Training on SageMaker which AWS claims make it possible to train large, complex deep learning models up to two times faster than current approaches.

Kristóf Szalay, CTO at Turbine, comments:

We use machine learning to train our in silico human cell model, called Simulated Cell, based on a proprietary network architecture. By accurately predicting various interventions on the molecular level, Simulated Cell helps us to discover new cancer drugs and find combination partners for existing therapies.

Training of our simulation is something we continuously iterate on, but on a single machine each training takes days, hindering our ability to iterate on new ideas quickly.

We are very excited about Distributed Training on Amazon SageMaker, which we are expecting to decrease our training times by 90% and to help us focus on our main task: to write a best-of-the-breed codebase for the cell model training.

Amazon SageMaker ultimately allows us to become more effective in our primary mission: to identify and develop novel cancer drugs for patients.

SageMaker’s Data Parallelism engine scales training jobs from a single GPU to hundreds or thousands by automatically splitting data across multiple GPUs, improving training time by up to 40 percent.

With edge computing advancements increasing rapidly, AWS is keeping pace with SageMaker Edge Manager.

Edge Manager helps developers to optimize, secure, monitor, and maintain ML models deployed on fleets of edge devices. In addition to helping optimize ML models and manage edge devices, Edge Manager also provides the ability to cryptographically sign models, upload prediction data from devices to SageMaker for monitoring and analysis, and view a dashboard that tracks and provided a visual report on the operation of the deployed models within the SageMaker console.

Igor Bergman, VP of Cloud and Software of PCs and Smart Devices at Lenovo, comments:

SageMaker Edge Manager will help eliminate the manual effort required to optimise, monitor, and continuously improve the models after deployment. With it, we expect our models will run faster and consume less memory than with other comparable machine-learning platforms.

As we extend AI to new applications across the Lenovo services portfolio, we will continue to require a high-performance pipeline that is flexible and scalable both in the cloud and on millions of edge devices. That’s why we selected the Amazon SageMaker platform. With its rich edge-to-cloud and CI/CD workflow capabilities, we can effectively bring our machine learning models to any device workflow for much higher productivity.

Finally, SageMaker JumpStart aims to make it easier for developers which have little experience with machine learning deployments to get started.

JumpStart provides developers with an easy-to-use, searchable interface to find best-in-class solutions, algorithms, and sample notebooks. Developers can select from several end-to-end machine learning templates(e.g. fraud detection, customer churn prediction, or forecasting) and deploy them directly into their SageMaker Studio environments.

AWS has been on a roll with SageMaker improvements—delivering more than 50 new capabilities over the past year. After this bumper feature drop, we probably shouldn’t expect any more until we’ve put 2020 behind us.

Source: Artificial Intelligence News
Tags: amazonaws sagemakerdata wrangleredge computingMachine LearningML
ShareTweetShareSend
Previous Post

Bitcoin Mining Softwares

Next Post

WhatsApp Vs. Signal: Where should you be? Explained.

Tarun Khanna

Tarun Khanna

Founder DeepTech Bytes - Data Scientist | Author | IT Consultant
Tarun Khanna is a versatile and accomplished Data Scientist, with expertise in IT Consultancy as well as Specialization in Software Development and Digital Marketing Solutions.

Related Posts

Researchers teach LLMs to solve complex planning challenges
Machine Learning

Researchers teach LLMs to solve complex planning challenges

April 3, 2025
Like human brains, large language models reason about diverse data in a standard way
Machine Learning

Like human brains, large language models reason about diverse data in a standard way

March 21, 2025
Exclusive: Meta starts testing its first in-house AI training chip
Artificial Intelligence

Exclusive: Meta starts testing its first in-house AI training chip

March 12, 2025
Karnataka’s Budget 2025-26: ₹1,000 crore Accelerator Fund, Quantum Park and EV Cluster
Artificial Intelligence

Karnataka’s Budget 2025-26: ₹1,000 crore Accelerator Fund, Quantum Park and EV Cluster

March 11, 2025
Next Post
whatsapp-vs-signal

WhatsApp Vs. Signal: Where should you be? Explained.

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

37 − 35 =

TRENDING

Data Analytics vs Data Science

Data Analytics vs Data Science
by Tarun Khanna
March 22, 2021
0
ShareTweetShareSend

Huawei readies new AI chip for mass shipment as China seeks Nvidia options, sources stated

Huawei readies new AI chip for mass shipment as China seeks Nvidia options, sources stated

Photo Credit: https://www.reuters.com/

by Tarun Khanna
April 23, 2025
0
ShareTweetShareSend

Natural Language Processing In Finance- Acing Digitization Game

Natural Language Processing
by Vatsal Ghiya
March 31, 2022
0
ShareTweetShareSend

AI memory need propels SK Hynix to historic DRAM market leadership

AI memory need propels SK Hynix to historic DRAM market leadership

Photo Credit: https://www.artificialintelligence-news.com/

by Tarun Khanna
April 28, 2025
0
ShareTweetShareSend

Most Popular Data Science Python Libraries for ML & DL

by Sarah Gomes
January 24, 2021
0
ShareTweetShareSend

How This Uncertain Pandemic Influenced The Time Series Model In Production ?

time series forcasting models
by Manika Sharma
February 6, 2021
0
ShareTweetShareSend

DeepTech Bytes

Deep Tech Bytes is a global standard digital zine that brings multiple facets of deep technology including Artificial Intelligence (AI), Machine Learning (ML), Data Science, Blockchain, Robotics,Python, Big Data, Deep Learning and more.
Deep Tech Bytes on Google News

Quick Links

  • Home
  • Affiliate Programs
  • About Us
  • Write For Us
  • Submit Startup Story
  • Advertise With Us
  • Terms of Service
  • Disclaimer
  • Cookies Policy
  • Privacy Policy
  • DMCA
  • Contact Us

Topics

  • Artificial Intelligence
  • Data Science
  • Python
  • Machine Learning
  • Deep Learning
  • Big Data
  • Blockchain
  • Tableau
  • Cryptocurrency
  • NFT
  • Technology
  • News
  • Startups
  • Books
  • Interview Questions

Connect

For PR Agencies & Content Writers:

connect@deeptechbytes.com

Facebook Twitter Linkedin Instagram
Listen on Apple Podcasts
Listen on Google Podcasts
Listen on Google Podcasts
Listen on Google Podcasts
DMCA.com Protection Status

© 2024 Designed by AK Network Solutions

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • Artificial Intelligence
  • Data Science
    • Language R
    • Deep Learning
    • Tableau
  • Machine Learning
  • Python
  • Blockchain
  • Crypto
  • Big Data
  • NFT
  • Technology
  • Interview Questions
  • Others
    • News
    • Startups
    • Books

© 2023. Designed by AK Network Solutions