What are the pros and cons of working with a containerized model instead of working with the traditional model?
When considering deployment strategies for machine learning (ML) models on Google Cloud, particularly within the context of serverless predictions at scale, practitioners frequently encounter a choice between containerized model deployment and traditional (often framework-native) model deployment. Both approaches are supported in Google Cloud's AI Platform (now Vertex AI) and other managed services. Each method presents
- Published in Artificial Intelligence, EITC/AI/GCML Google Cloud Machine Learning, First steps in Machine Learning, Serverless predictions at scale
What happens when you upload a trained model into Google’s Cloud Machine Learning Engine? What processes does Google’s Cloud Machine Learning Engine perform in the background that facilitate our life?
When you upload a trained machine learning model to Google Cloud Machine Learning Engine (now known as Vertex AI), a series of intricate and automated backend processes are activated, streamlining the transition from model development to large-scale production deployment. This managed infrastructure is designed to abstract operational complexity, providing a seamless environment for deploying, serving,
Can we use streaming data to train and use a model continuously and improve it at the same time?
The ability to use streaming data for both continuous model training and real-time inference is a significant topic in machine learning, particularly within modern data-driven applications. The traditional approach to building machine learning models typically involves collecting a batch of data, cleaning and preparing it, training a model, evaluating it, deploying it, and then periodically
Do I need to install TensorFlow?
The inquiry regarding whether one needs to install TensorFlow when working with plain and simple estimators, particularly within the context of Google Cloud Machine Learning and introductory machine learning tasks, is one that touches on both the technical requirements of certain tools and the practical workflow considerations in applied machine learning. TensorFlow is an open-source
How do Vertex AI and AI Platform API differ?
Vertex AI and AI Platform API are both services provided by Google Cloud that aim to facilitate the development, deployment, and management of machine learning (ML) workflows. While they share a similar objective of supporting ML practitioners and data scientists in leveraging Google Cloud for their projects, these platforms differ significantly in their architecture, feature
How can soft systems analysis and satisficing approaches be used in evaluating the potential of Google Cloud AI machine learning?
Soft systems analysis and satisficing are methodologies with distinct heritages in systems thinking and decision theory, respectively, both offering nuanced alternatives to purely quantitative, optimization-centric evaluation paradigms. Their application to the assessment of Google Cloud AI machine learning—specifically in the context of serverless, scalable prediction—provides valuable frameworks for grappling with the complex, multifaceted, and often
How do I get access to Google Cloud AI?
Accessing Google Cloud AI involves several procedural and conceptual steps, each grounded in the broader context of cloud-based machine learning and artificial intelligence services. Google Cloud Platform (GCP) offers a wide array of tools and services designed to facilitate the development, deployment, and management of AI and machine learning models. The process to gain access
How do I deploy a custom container on Google Cloud AI Platform?
Deploying a custom container on Google Cloud AI Platform (now part of Vertex AI) is a process that allows practitioners to leverage their own software environments, dependencies, and frameworks for training and prediction tasks. This approach is particularly beneficial when default environments do not meet the requirements of a project, such as when custom libraries,
- Published in Artificial Intelligence, EITC/AI/GCML Google Cloud Machine Learning, Google Cloud AI Platform, Training models with custom containers on Cloud AI Platform
What is the difference between using CREATE MODEL with LINEAR_REG in BigQuery ML versus training a custom model with TensorFlow in Vertex AI for time series prediction?
The distinction between using the `CREATE MODEL` statement with `LINEAR_REG` in BigQuery ML and training a custom model with TensorFlow in Vertex AI for time series prediction lies in multiple dimensions, including model complexity, configurability, scalability, operational workflow, integration into data pipelines, and typical use cases. Both approaches offer unique advantages and trade-offs, and the
How to create model and version on GCP after uploading model.joblib on bucket?
To create a model and version on Google Cloud Platform (GCP) after uploading a Scikit-learn model artifact (e.g., `model.joblib`) to a Cloud Storage bucket, you need to use Google Cloud’s Vertex AI (previously AI Platform) for model management and deployment. The process involves several structured steps: preparing your model and artifacts, setting up the environment,
- Published in Artificial Intelligence, EITC/AI/GCML Google Cloud Machine Learning, Advancing in Machine Learning, Scikit-learn models at scale
- 1
- 2

