Skip to main content

Vertex AI – One AI platform, every ML tool you need

 

This year on Google I/O (Google’s Developer conference) Google presented a new platform that unites all ML tools. Vertex AI brings together the Google Cloud services for building ML under one, unified UI and API.

There are many benefits to using Vertex AI. You can train models without code, with minimal expertise required, and take advantage of AutoML to build models in less time. Also, Vertex AI’s custom model tooling supports advanced ML coding, with nearly 80% fewer lines of code required to train a model with custom libraries than competitive platforms.



Google Vertex AI logo

You can use Vertex AI to manage the following stages in the ML workflow:

  • Define and upload a dataset.
  • Train an ML model on your data:
    • Train model
    • Evaluate model accuracy
    • Tune hyperparameters (custom training only)
  • Upload and store your model in Vertex AI.
  • Deploy your trained model and get an endpoint for serving predictions.
  • Send prediction requests to your endpoint.
  • Specify a prediction traffic split in your endpoint.
  • Manage your models and endpoints.


“We had two guiding lights while building Vertex AI: get data scientists and engineers out of the orchestration weeds, and create an industry-wide shift that would make everyone get serious about moving AI out of pilot purgatory and into full-scale production,” said Andrew Moore, vice president and general manager of Cloud AI and Industry Solutions at Google Cloud. “We are very proud of what we came up with within this platform, as it enables serious deployments for a new generation of AI that will empower data scientists and engineers to do fulfilling and creative work.”

“Data science practitioners hoping to put AI to work across the enterprise aren’t looking to wrangle tooling. Rather, they want tooling that can tame the ML lifecycle. Unfortunately, that is no small order,” said Bradley Shimmin, chief analyst for AI Platforms, Analytics, and Data Management at Media. “It takes a supportive infrastructure capable of unifying the user experience, plying AI itself as a supportive guide, and putting data at the very heart of the process — all while encouraging the flexible adoption of diverse technologies.”

To learn more about how to get started on the platform, check out the ML on GCP best practices, this practitioner’s guide to MLOps whitepaper, and sign up to attend the Applied ML Summit for data scientists and ML engineers on June 10th.

If you want to try it and get to know it better you can watch a Codelab (It will cost you about $2 for running it on Google Cloud, but you can register for a $300 trial on GCP 😁 )

Comments

Popular posts from this blog

How to build FAQ Chatbot on Dialogflow?

  After Google I/O I’m inspired and excited to try new APIs and learn new stuff from Google. This time I decided to try Dialogflow and build a Flutter Chatbot app that will answer some frequently asked questions about Dialogflow. This time I want to be focused more on Dialogflow rather than Flutter. Firstly, go to  Dialogflow ES console , create a new Agent, specify the agent’s name, choose English as a language and click “Create”. As you created a new agent go to setting and enable beta features and APIs and Save. Now let’s model our Dialogflow agent When you create a new Dialogflow agent, two default intents will be created automatically. The  Default Welcome Intent  is the first flow you get to when you start a conversation with the agent. The  Default Fallback Intent  is the flow you’ll get once the agent can’t understand you or can not match intent with what you just said. Click  Intents > Default Welcome Intent Scroll down to  Responses . Clear all Text Responses. In the defau

What the Flutter? ExpansionPanel

  A common pattern in apps is to have a list of items that you can expand to show more details. Sometimes these details don’t justify an entirely separate view, and you just need them to show up inline in the list. For that, check out ExpansionPanel, a widget that when tapped on will expand a panel. Start with the  headerBuilder , which returns what the first line of this panel will be. It takes a context and a boolean, so you can change what it looks like when the panel is open vs closed and returns a widget. Next up is the body, which contains the contents of the opened panel. And finally is the boolean  isExpanded  to indicate whether or not this panel is currently open. But what to do with this flag? Well, ExpansionPanels almost exclusively appear as children of ExpansionPanelLists. Here, we can maintain a list of which panels are open and use  ExpansionPanelList’s  expansionCallback parameter to update them. This callback takes an index of the panel that’s just been tapped and whe