Indianext
No Result
View All Result
Subscribe
  • News
    • Project Watch
    • Policy
  • AI Next
  • People
    • Interviews
    • Profiles
  • Companies
  • Make In India
    • Solutions
    • State News
  • About Us
    • Editors Corner
    • Mission
    • Contact Us
    • Work Culture
  • Events
  • Guest post
  • News
    • Project Watch
    • Policy
  • AI Next
  • People
    • Interviews
    • Profiles
  • Companies
  • Make In India
    • Solutions
    • State News
  • About Us
    • Editors Corner
    • Mission
    • Contact Us
    • Work Culture
  • Events
  • Guest post
No Result
View All Result
Latest News on AI, Healthcare & Energy updates in India
No Result
View All Result
Home AI Next

AI Insights: Examining The Free Python Module DeepSpeed-MII

October 29, 2022
AI

A new open-source Python module from DeepSpeed called DeepSpeed-MII accelerates over 20,000 popular deep learning models.

Even though open-source software has increased accessibility for AI, inference time and cost remain two major barriers to its widespread adoption.

Although they are not yet generally accessible, system innovations have the potential to reduce the latency and cost of DL model inference. Low latency and low-cost inference are therefore mostly unattainable as many data scientists require the knowledge to correctly identify and carry out the set of system improvements relevant to a given model. The complexity of the DL model inference landscape, which involves large changes in model size, architecture, system performance characteristics, and hardware requirements, is mostly to blame for this lack of availability.

DeepSpeed-MII

A new open-source Python library called DeepSpeed-MII was created by Microsoft Research to encourage the wider adoption of low-latency, economical inference of high-performance models. Numerous commonly used DL models with highly effective implementations are accessible through MII.

MII employs DeepSpeed-Inference optimizations, such as deep fusion for transformers, automated tensor-slicing for multi-GPU inference, and ZeroQuant quantization, for low latency/cost inference. As a result, it makes it possible to deploy these models quickly, simply, and affordably using Azure and AML on-premises.

MII is powered by DeepSpeed-Inference under the hood. MII optimises DeepSpeed-system Inference based on the model type, batch size, and hardware resources to decrease latency and boost throughput. In order to identify the underlying PyTorch model architecture and replace it with an optimised implementation, MII and DeepSpeed-Inference use a number of pre-specified model injection criteria. As a result, the DeepSpeed-comprehensive Inference’s set of optimizations are instantly available to the tens of thousands of widely-used models offered by MII.

open-source databases

There are dozens such transformer models available in numerous open-source model repositories, including Hugging Face, FairSeq, EluetherAI, and others. A few of the applications that MII provides are text creation, question answering, and classification. It supports models with hundreds of millions of parameters such as BERT, RoBERTa, GPT, OPT, and BLOOM. Additionally, stable diffusion and other modern image-production methods are made possible. Additionally, inference workloads may be latency-critical or cost-sensitive, with the main goal being to minimise both.

One of two DeepSpeed-Inference iterations can be used by MII. The majority of the above improvements are found in the first, ds-public, which is a component of the public DeepSpeed library. Users of Microsoft Azure can also connect to ds-azure using MII for a more reliable connection. Both the MII-Public and MII-Azure DeepSpeed-Inference variants can be used to access MII instances.

Conclusion

When compared to the open-source PyTorch implementation, MII-Public and MII-Azure provide considerable latency and cost benefits (Baseline). However, their performances may differ depending on the type of producing task. MII is perfect for latency-critical applications with batch sizes of one since it can cut latency by up to 6x for open-source models and workloads. To obtain the lowest cost, the team increased baseline and MII throughput with a big batch size. Costly language models like Bloom, OPT, etc. can drastically lower inference costs when combined with MII.

Both locally and on any cloud service, MII-Public is capable of running. For deployment-related enquiries, MII implements a straightforward GRPC server and offers a GRPC inference endpoint. Azure and AML Inference can be used to support MII. The researchers also anticipate that a number of models will be made easier by their discoveries. More advanced AI capabilities in applications and products are made possible by MII’s quick reduction of inferencing latency and cost.

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Editors Corner

How can Artificial Intelligence tools be a blessing for recruiters?

Will Artificial Intelligence ever match human intelligence?

Artificial Intelligence: Features of peer-to-peer networking

What not to share or ask on Chatgpt?

How can Machine Learning help in detecting and eliminating poverty?

How can Artificial Intelligence help in treating Autism?

Speech Recognition and its Wonders in your corporate life

Most groundbreaking Artificial Intelligence-based gadgets to vouch for in 2023

Recommended News

AI Next

Google: AI From All Perspectives

Alphabet subsidiary Google may have been slower than OpenAI to make its AI capabilities publicly available in the past, but...

by India Next
May 31, 2024
AI Next

US And UK Doctors Think Pfizer Is Setting The Standard For AI And Machine Learning In Drug Discovery

New research from Bryter, which involved over 200 doctors from the US and the UK, including neurologists, hematologists, and oncologists,...

by India Next
May 31, 2024
Solutions

An Agreement Is Signed By MEA, MeitY, And CSC To Offer E-Migration Services Via Shared Service Centers

Three government agencies joined forces to form a synergy in order to deliver eMigrate services through Common Services Centers (CSCs)...

by India Next
May 31, 2024
AI Next

PR Handbook For AI Startups: How To Avoid Traps And Succeed In A Crowded Field

The advent of artificial intelligence has significantly changed the landscape of entrepreneurship. The figures say it all. Global AI startups...

by India Next
May 31, 2024

Related Posts

Google
AI Next

Google: AI From All Perspectives

May 31, 2024
Pfizer
AI Next

US And UK Doctors Think Pfizer Is Setting The Standard For AI And Machine Learning In Drug Discovery

May 31, 2024
Artificial-Intelligence
AI Next

PR Handbook For AI Startups: How To Avoid Traps And Succeed In A Crowded Field

May 31, 2024
openai
AI Next

OpenAI Creates An AI Safety Committee Following Significant Departures

May 31, 2024
Load More
Next Post
AI

The Creative Cloud From Adobe Now Includes Generative AI

IndiaNext Logo
IndiaNext Brings you latest news on artificial intelligence, Healthcare & Energy sector from all top sources in India and across the world.

Recent Posts

Google: AI From All Perspectives

US And UK Doctors Think Pfizer Is Setting The Standard For AI And Machine Learning In Drug Discovery

An Agreement Is Signed By MEA, MeitY, And CSC To Offer E-Migration Services Via Shared Service Centers

PR Handbook For AI Startups: How To Avoid Traps And Succeed In A Crowded Field

OpenAI Creates An AI Safety Committee Following Significant Departures

Tags

  • AI
  • EV
  • Mental WellBeing
  • Clean Energy
  • TeleMedicine
  • Healthcare
  • Electric Vehicles
  • Artificial Intelligence
  • Chatbots
  • Data Science
  • Electric Vehicles
  • Energy Storage
  • Machine Learning
  • Renewable Energy
  • Green Energy
  • Solar Energy
  • Solar Power

Follow us

  • Facebook
  • Linkedin
  • Twitter
© India Next. All Rights Reserved.     |     Privacy Policy      |      Web Design & Digital Marketing by Heeren Tanna
No Result
View All Result
  • About Us
  • Activate
  • Activity
  • Advisory Council
  • Archive
  • Career Page
  • Companies
  • Contact Us
  • cryptodemo
  • Energy next
  • Energy Next Archive
  • Home
  • Interviews
  • Make in India
  • Market
  • Members
  • Mission
  • News
  • News Update
  • People
  • Policy
  • Privacy Policy
  • Register
  • Reports
  • Subscription Page
  • Technology
  • Top 10
  • Videos
  • White Papers
  • Work Culture
  • Write For Us

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

IndiaNext Logo

Join Our Newsletter

Get daily access to news updates

no spam, we hate it more than you!