
GCP
335+ preguntas de práctica gratuitas con respuestas verificadas por IA
Impulsado por IA
Cada respuesta de Google Professional Machine Learning Engineer es verificada de forma cruzada por 3 modelos de IA líderes para garantizar la máxima precisión. Obtén explicaciones detalladas por opción y análisis profundo de cada pregunta.
You deployed a TensorFlow recommendation model to a Vertex AI Prediction endpoint in us-central1 with autoscaling enabled. Over the last week, you observed sustained traffic of ~1,200 requests per hour (about 20 RPS) during business hours, which is 2x higher than your original estimate, and you need to keep P95 latency under 150 ms during future surges. You want the endpoint to scale efficiently to handle this higher baseline and upcoming spikes without causing user-visible latency. What should you do?
You plan to fine-tune a video-frame classifier via transfer learning using a pre-trained ResNet-50 backbone. Your labeled dataset contains 18,000 1080p frames, and you will retrain the model once per day; each training run completes in under 60 minutes on 4 V100 GPUs, and you must minimize infrastructure cost and operational overhead. Which platform components and configuration should you choose?
Your team is preparing to train a fraud detection model using data in BigQuery that includes several fields containing PII (for example, card_number, customer_email, and phone_number). The dataset has approximately 250 million rows and every column is required as a feature. Security requires that you reduce the sensitivity of PII before training while preserving each column’s format and length so downstream SQL joins and validations continue to work. The transformation must be deterministic so the same input always maps to the same protected value, and authorized teams must be able to decrypt values for audits. How should you proceed?
Your team deployed a regression model that predicts hourly water usage for industrial chillers. Four months after launch, a vendor firmware update changed sensor sampling and units for three input features, and the live feature distributions diverged: 5 of 18 features now have a population stability index > 0.25, 27% of temperature readings fall outside the training range, and production RMSE increased from 0.62 to 1.45. How should you address the input differences in production?
You are a data scientist at a city transportation agency tasked with forecasting hourly bike-share demand per station to optimize rebalancing. Your historical trips table in BigQuery contains 24 months of data (~22 million rows) with columns: timestamp, station_id, neighborhood, weather_condition (sunny/rainy/snow), special_event (boolean), and surge_pricing_flag (boolean). You need to choose the most effective combination of a BigQuery ML model and feature engineering to minimize RMSE while capturing weekly/seasonal patterns and handling multiple categorical variables; what should you do?
¿Quieres practicar todas las preguntas en cualquier lugar?
Descarga Cloud Pass gratis — incluye exámenes de práctica, seguimiento de progreso y más.
You work for a vacation rental marketplace with 1.8 million property listings stored across BigQuery and Cloud Storage; the current search relies on keyword matching and filter chips, but you are seeing more complex semantic queries that reference amenities and metadata (for example, "quiet pet-friendly cabin near a lake with a fireplace, sleeps 6, under $200/night, host rating > 4.7"). You must deliver a revamped semantic search proof of concept within 2 weeks with minimal custom modeling and integration effort that can quickly index both structured listing attributes and unstructured descriptions; what should you choose as the search backend?
You are a data scientist at a national power utility analyzing 850 million smart-meter readings from 3,000 substations collected over 5 years; for exploratory analysis, you must compute descriptive statistics (mean, median, mode) by device and region, perform complex hypothesis tests (e.g., differences between peak vs off-peak and seasonal periods with multiple comparisons), and plot feature variations at hourly and daily granularity over time, while using as much of the telemetry as possible and minimizing computational resources—what should you do?
You are launching a grocery delivery mobile app across 3 cities and will use Google Cloud's Recommendations AI to build, test, and deploy product suggestions; you currently capture about 2.5 million user events per day, maintain a catalog of 120,000 SKUs with accurate price and availability, and your business objective is to raise average order value (AOV) by at least 6% within the next quarter while adhering to best practices. Which approach should you take to develop recommendations that most directly increase revenue under these constraints?
You are training a LightGBM model to forecast daily inventory for 120 stores using a small dataset (~60 MB) on Vertex AI; your training script needs a system library (libgomp) and several custom Python packages, and each run takes about 10 minutes, so you want job startup time to be under 2 minutes to minimize overhead. How should you configure the Vertex AI custom training job to minimize startup time while keeping the dataset easy to update?
You are building an end-to-end scikit-learn MLOps workflow in Vertex AI Pipelines (Kubeflow Pipelines) that ingests 50 GB of CSV data from Cloud Storage, performs data cleaning, feature selection, model training, and model evaluation, then writes a .pkl model artifact to a versioned path in a GCS bucket. You are iterating on multiple versions of the feature selection and training components, submitting each version as a new pipeline run in us-central1 on n1-standard-4 CPU-only executors; each end-to-end run currently takes about 80 minutes. You want to reduce iteration time during development without increasing your GCP costs; what should you do?
Your team must deliver an ML solution on Google Cloud to triage warranty claim emails for a global appliance manufacturer into 8 categories within 4 weeks. You are required to use TensorFlow to maintain full control over the model's code, serving, and deployment, and you will orchestrate the workflow with Kubeflow Pipelines. You have 30,000 labeled examples and want to accelerate delivery by leveraging existing resources and managed services instead of training a brand-new model from scratch. How should you build the classifier?
You are building an anomaly detection model for an industrial IoT platform using Keras and TensorFlow. The last 24 months of sensor events (~900 million rows, ~2.6 TB) are stored in a single partitioned table in BigQuery, and you need to apply feature scaling, categorical encoding, and time-window aggregations in a cost-effective and efficient way before training. The trained model will be used to run weekly batch inference directly in BigQuery against newly ingested partitions. How should you implement the preprocessing workflow?
Your edtech company operates a live Q&A chat in virtual classrooms, where an automated text moderation model flags toxic messages. After recent complaints, you discover that benign messages referencing certain indigenous festivals are being misclassified as abusive; an audit on a 10,000-message holdout shows a 12–15% false positive rate for messages containing those festival names versus 3% overall, and those references make up <1% of your training set. With a tight budget and an overextended team this quarter, a major overhaul or full replacement is not feasible; what should you do?
A fintech analytics team has migrated 12 time-series forecasting and anomaly-detection models to Google Cloud over the last 90 days and is now standardizing new training on Vertex AI. You must implement a system that automatically tracks model artifacts (datasets, feature snapshots, checkpoints, and model binaries) and end-to-end lineage across pipeline steps for dev, staging, and prod; the solution must be simple to adopt via reusable templates, require minimal custom code, retain lineage for at least 180 days, and scale to future models without re-architecting; what should you do?
You manage the ML engineering team for a regional logistics network; most training runs are multi-node PyTorch Lightning jobs on managed training with NVIDIA T4 GPUs where a single experiment consumes ~3,000 GPU-hours, new model versions are released every 6–10 weeks, and finance requires at least a 40% reduction in training compute spend without degrading model quality or materially increasing wall-clock time; your pipeline already writes restartable checkpoints to Cloud Storage every 10 minutes with <2% overhead and can tolerate node interruptions. What should you do to reduce Google Cloud compute costs without impacting the model’s performance?
Your robotics team is deploying a quality inspection system for ceramic floor tiles on a high-speed conveyor line; each tile is captured as a 3840x3840 RGB image under controlled lighting, and you have 60,000 labeled images (defective vs. non-defective); operations managers require pixel-level attribution heatmaps overlaid on each image so they can pinpoint hairline cracks and decide whether to discard a tile within the same shift. How should you build the model?
A digital payments startup trained a binary classification model on Vertex AI to flag potentially fraudulent card transactions using 24 months of historical data (validation AUC = 0.93) and deployed it to a Vertex AI online endpoint that processes ~60,000 requests per day; after 4 weeks, the production AUC computed from feedback labels has dropped to 0.76, while autoscaling shows sufficient replicas and Cloud Monitoring reports P95 latency around 110 ms and error rate < 0.1%. What should you do first to troubleshoot the drop in predictive performance?
You are part of a data science team at a ride‑sharing platform and need to train and compare multiple TensorFlow models on Vertex AI using 850 million labeled trip records (≈2.3 TB) stored in a BigQuery table; training will run on 4–8 workers and you want to minimize data‑ingestion bottlenecks while ensuring the pipeline remains scalable and repeatable. What should you do?
Your e-commerce price-optimization model serves about 30,000 predictions per hour on a Vertex AI endpoint, and a Vertex AI Model Monitoring job is configured to detect training-serving skew using a 24-hour sliding window with a 0.3 sampling rate and a baseline dataset at gs://retail-ml/training/2025-06/data.parquet; after three consecutive windows reporting skew on features inventory_days and competitor_price, you retrained the model using the last 45 days of data at gs://retail-ml/training/2025-08/data.parquet and deployed version v2 to the same endpoint, but the monitoring job still raises the same skew alert—what should you do?
You trained an automated scholarship eligibility classifier for a national education nonprofit using Vertex AI on 1.2 million labeled applications, reaching an offline ROC AUC of 0.95; the review board is concerned that predictions may be biased by applicant demographics (e.g., gender, ZIP-code–derived income bracket, first-generation college status) and asks you to deliver transparent insight into how the model makes decisions for 500 sampled approvals and denials and to identify any fairness issues across these cohorts. What should you do?
Periodo de estudio: 1 month
Just want to say a massive thank you to the entire Cloud pass, for helping me pass my exam first time. I wont lie, it wasn't easy, especially the way the real exam is worded, however the way practice questions teaches you why your option was wrong, really helps to frame your mind and helps you to understand what the question is asking for and the solutions your mind should be focusing on. Thanks once again.
Periodo de estudio: 1 month
Good questions banks and explanations that help me practise and pass the exam.
Periodo de estudio: 1 month
강의 듣고 바로 문제 풀었는데 정답률 80% 가량 나왔고, 높은 점수로 시험 합격했어요. 앱 잘 이용했습니다
Periodo de estudio: 1 month
Good mix of theory and practical scenarios
Periodo de estudio: 1 month
I used the app mainly to review the fundamentals—data preparation, model tuning, and deployment options on GCP. The explanations were simple and to the point, which really helped before the exam.

Professional

Associate

Professional

Associate

Foundational

Professional

Professional

Professional

Professional

Professional
Obtén la app gratis