Current directory: /home3/bjinbymy/public_html/indianext/wp-content/mu-plugins Nvidia Participates In The GTC - AI Next
Indianext
No Result
View All Result
Subscribe
  • News
    • Project Watch
    • Policy
  • AI Next
  • People
    • Interviews
    • Profiles
  • Companies
  • Make In India
    • Solutions
    • State News
  • About Us
    • Editors Corner
    • Mission
    • Contact Us
    • Work Culture
  • Events
  • Guest post
  • News
    • Project Watch
    • Policy
  • AI Next
  • People
    • Interviews
    • Profiles
  • Companies
  • Make In India
    • Solutions
    • State News
  • About Us
    • Editors Corner
    • Mission
    • Contact Us
    • Work Culture
  • Events
  • Guest post
No Result
View All Result
Latest News on AI, Healthcare & Energy updates in India
No Result
View All Result
Home AI Next

Nvidia Participates In The GTC

April 5, 2023
Artificial-Intelligence

The NVIDIA GPU Technologies Conference (GTC) this year could not have come at a better time for the business. The Artificial Intelligence (AI) powering ChatGPT, other connected Large Language Models (LLMs), and their applications for generative AI applications are now the trendiest topics in technology. NVIDIA GPUs are the brains behind all of this new AI technologies. Jensen Huang, CEO of NVIDIA, reiterated his company’s commitment to LLMs and generative AI in the future. That is “the iPhone moment for AI,” he says. AI systems can learn the languages of people, software, photos, or chemistry using LLMs. Generative AI may produce fresh, original works based on a query and using a vast knowledge base.

This potential is being advanced by jumbo-sized LLMs, particularly the most recent GPT 4.0, which was unveiled right before GTC. It takes hundreds of GPUs to train these sophisticated models, and additional GPUs are needed for inference when these models are used to solve particular issues. The H100, the most recent Hopper GPU from Nvidia, is known for training, but it can also be partitioned into multiple instances (up to seven), or what Nvidia refers to as MIG (Multi-Instance GPU), which enables the GPU to run different inference models. The GPU converts queries into fresh outputs while in this inference mode by employing learned LLMs.

Nvidia, a supplier of the entire AI stack, including chips, software, accelerator cards, systems, and even services, is utilising its position as market leader to create new revenue prospects. The corporation is expanding its services division into industries like biology, for instance. The cost of the company’s services may be dependent on the amount of time used or the cost of the final product created using those services.

The Prospects for LLMs

These extensive training sets help AI systems comprehend natural language better. A taught AI computer, for instance, may receive a request to produce a programme and utilise its understanding of programming constructs to create the desired programme. Even yet, there is still a large gap between properly designed, tested, and documented programmes and AI. Today is that. Nonetheless, this method is already democratising programming by creating better “no-code” applications and making it simpler for users to create their own programmes using human language without the requirement for a thorough understanding of programming techniques.

The AI’s understanding of ground facts does have certain limitations, though. Only the data needed to train them will allow AI models to function. The ability to retrain on this constant flood of fresh information on the fly while streaming real-time information from the world is currently not possible.

By scaling LLMs to higher levels, you may grant them more authority because their “brains” are bigger and can do more things. Nonetheless, there will still be a need to implement ethical frameworks, minimise biases, and provide safeguards; much more work in AI ethics is required.

According to Nvidia, education is the process of producing intelligence and building a “AI factory.” With the help of inference, this knowledge can be put to use and deployed, and it is scalable from the cloud to end devices. The pre-trained basic models that Nvidia has developed can be used by businesses and customised. Businesses can use internal or secret data and add guardrails to the inference results. Nvidia provides its DGX computer cloud services directly or through hyperscaler clouds for training. Businesses without these internal resources can employ this training-as-a-service.

Deployments of LLM

The capabilities of today are already changing things. New businesses are joining the bandwagon. The major hyperscalers are all creating strategies centred around LLMs in order to not be left out. Microsoft constructed the datacenter for OpenAI, which used the A100 GPU—the H100’s predecessor—to build ChatGPT. Bing for search and conversation, Microsoft 365 as its Copilot assistant for work activities, and Bing all make use of AI technology. These services are all powered by Nvidia GPUs. The new Nvidia H100 GPUs are already being used in private preview by Microsoft Azure cloud services.

In addition to Microsoft’s Azure, NVIDIA has lined up a number of other cloud service providers for its H100 cloud. Limited availability Oracle Cloud Infrastructure is typically offered by Cirrascale and CoreWeave. According to Nvidia, a limited preview of AWS’s H100 cloud will be made accessible in the upcoming weeks. Moreover, Google Cloud plans to offer H100 cloud services with NVIDIA’s cloud partners Lambda, Paperspace, and Vultr.

Three new LLM modality/services are available from Nvidia:

  1. NeMo is used for human language.
  2. Picasso is available for photos (including video).
  3. BioNeMo is a service for biology, the language of proteins.

For the purpose of biology and drug development, BioNeMo can be used to teach AI the language of proteins. According to Huang from Nvidia, this will pave the path for the widespread application of AI in healthcare and drug development. Data filtering, drug trial forecasts, and other drug discovery are only a few of the many uses of AI. The use cases include molecular docking and the prediction of 3D protein structures. Businesses can create unique models using the BioNeMo service and proprietary data. With the use of the technology, model training times can be slashed from six months to just four. Amgen was a founding associate of Nvidia. Moreover, Nvidia and Medtronic are collaborating on intelligent medical equipment. AI will be extremely useful in automating repetitive jobs or simplifying difficult ones, which is appealing to medical practitioners.

Creating a Universe of Digital Twins

Nvidia also made a big announcement about its Omniverse platform. There are numerous uses for this real-world digitalization. Particularly, it appears that the industrial manufacturing sectors of the automobile and other industries are embracing these digital representations of physical machines (so called digital twins).

Huang highlighted the application of Omniverse in factories and consumer experiences while highlighting auto makers in his presentation. Omniverse speeds up the entire workflow, including client involvement and design and manufacturing. GM models aerodynamics with the use of digital twins of automobile designs. Omniverse is used by Toyota, Mercedes, and BMW to build factories. LUCID lures clients with a 3D VR automobile. Also, Microsoft Azure would provide Omniverse services, it was revealed during GTC.

Moreover, Nvidia unveiled its Isaac Sim platform, which enables international teams to work remotely together to develop, train, simulate, validate, and deploy robots.

Hardware for a new data centre

Nvidia also unveiled four inference platforms to support advanced generative AI. Nvidia L4 for creating AI video; Nvidia L40 for creating 2D and 3D images; and Nvidia H100 NVL for deploying big language models are some examples of these. The Grace Hopper “superchip,” which links the Arm-based Grace CPU and Hopper GPU over a fast coherent interface, was also made public by Nvidia. Systems that use very large datasets for recommendations are the focus of Grace Hopper.

The low-profile L4 PCIe card is up to 120 times more energy-efficient and capable of producing AI-powered visual performance than CPUs. The first cloud services provider to provide NVIDIA’s L4 Tensor Core GPU is Google Cloud, which stated G2 virtual machines are currently available in private preview.

The heart of Omniverse is the L40 PCIe card. According to the company, compared to the previous generation Nvidia product, it offers 12x Omniverse performance and 7x Stable Diffusion inference performance. The Ada Lovelace GPU architecture serves as the foundation for the L4 and L40 products.

The H100 has a built-in Transformer Engine and is based on the NVIDIA Hopper GPU computing architecture. It was designed with the development, training, and deployment of recommender systems, big language models, and generative AI in mind. The usage of the H100’s FP8 precision math, which can accelerate AI training by 9x and up to 30x quicker AI inference on LLMs than the previous-generation A100, is one of the performance advantages of the H100 over the A100.

NVIDIA H100 NVL for scalably installing large LLMs like ChatGPT. When compared to the previous version A100 at data centre scale, the new H100 NVL with 94GB of memory and Transformer Engine acceleration offers up to 12x quicker inference performance at GPT-3. Two H100 GPUs are connected through the NVLink coherent bus on the H100 NVL PCIe card. In the second half of the year, the H100 NVL GPU is anticipated to launch.

Making Better Chips with GPUs

Computational lithography is another use for Nvidia’s GPU technology. The GPU corrects for optical diffraction (blur) when creating optical masks used to project photolithographic layers onto silicon in one application, chip manufacturing. By leveraging the new Hopper H100 GPU, Nvidia’s CuLitho service speeds up the calculation for makers of optical and EUV masks by around 40 times compared to standard CPU computations. As a result, the entire manufacturing process for creating new chips is sped up. Nvidia’s longtime foundry partner TSMC is one of the first partners. ASML and Synopsys have partnered as well.

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Editors Corner

How can Artificial Intelligence tools be a blessing for recruiters?

Will Artificial Intelligence ever match human intelligence?

Artificial Intelligence: Features of peer-to-peer networking

What not to share or ask on Chatgpt?

How can Machine Learning help in detecting and eliminating poverty?

How can Artificial Intelligence help in treating Autism?

Speech Recognition and its Wonders in your corporate life

Most groundbreaking Artificial Intelligence-based gadgets to vouch for in 2023

Recommended News

AI Next

Google: AI From All Perspectives

Alphabet subsidiary Google may have been slower than OpenAI to make its AI capabilities publicly available in the past, but...

by India Next
May 31, 2024
AI Next

US And UK Doctors Think Pfizer Is Setting The Standard For AI And Machine Learning In Drug Discovery

New research from Bryter, which involved over 200 doctors from the US and the UK, including neurologists, hematologists, and oncologists,...

by India Next
May 31, 2024
Solutions

An Agreement Is Signed By MEA, MeitY, And CSC To Offer E-Migration Services Via Shared Service Centers

Three government agencies joined forces to form a synergy in order to deliver eMigrate services through Common Services Centers (CSCs)...

by India Next
May 31, 2024
AI Next

PR Handbook For AI Startups: How To Avoid Traps And Succeed In A Crowded Field

The advent of artificial intelligence has significantly changed the landscape of entrepreneurship. The figures say it all. Global AI startups...

by India Next
May 31, 2024

Related Posts

Google
AI Next

Google: AI From All Perspectives

May 31, 2024
Pfizer
AI Next

US And UK Doctors Think Pfizer Is Setting The Standard For AI And Machine Learning In Drug Discovery

May 31, 2024
Artificial-Intelligence
AI Next

PR Handbook For AI Startups: How To Avoid Traps And Succeed In A Crowded Field

May 31, 2024
openai
AI Next

OpenAI Creates An AI Safety Committee Following Significant Departures

May 31, 2024
Load More
Next Post
Speech Recognition and its Wonders in your corporate life

Speech Recognition and its Wonders in your corporate life

IndiaNext Logo
IndiaNext Brings you latest news on artificial intelligence, Healthcare & Energy sector from all top sources in India and across the world.

Recent Posts

Google: AI From All Perspectives

US And UK Doctors Think Pfizer Is Setting The Standard For AI And Machine Learning In Drug Discovery

An Agreement Is Signed By MEA, MeitY, And CSC To Offer E-Migration Services Via Shared Service Centers

PR Handbook For AI Startups: How To Avoid Traps And Succeed In A Crowded Field

OpenAI Creates An AI Safety Committee Following Significant Departures

Tags

  • AI
  • EV
  • Mental WellBeing
  • Clean Energy
  • TeleMedicine
  • Healthcare
  • Electric Vehicles
  • Artificial Intelligence
  • Chatbots
  • Data Science
  • Electric Vehicles
  • Energy Storage
  • Machine Learning
  • Renewable Energy
  • Green Energy
  • Solar Energy
  • Solar Power

Follow us

  • Facebook
  • Linkedin
  • Twitter
© India Next. All Rights Reserved.     |     Privacy Policy      |      Web Design & Digital Marketing by Heeren Tanna
No Result
View All Result
  • About Us
  • Activate
  • Activity
  • Advisory Council
  • Archive
  • Career Page
  • Companies
  • Contact Us
  • cryptodemo
  • Energy next
  • Energy Next Archive
  • Home
  • Interviews
  • Make in India
  • Market
  • Members
  • Mission
  • News
  • News Update
  • People
  • Policy
  • Privacy Policy
  • Register
  • Reports
  • Subscription Page
  • Technology
  • Top 10
  • Videos
  • White Papers
  • Work Culture
  • Write For Us

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

IndiaNext Logo

Join Our Newsletter

Get daily access to news updates

no spam, we hate it more than you!