Aporia Documentation
Get StartedBook a Demo🚀 Cool StuffBlog
V2
V2
  • 📖Aporia Docs
  • 🤗Introduction
    • Quickstart
    • Support
  • 💡Core Concepts
    • Why Monitor ML Models?
    • Understanding Data Drift
    • Analyzing Performance
    • Tracking Data Segments
    • Models & Versions
  • 🚀Deployment
    • AWS
    • Google Cloud
    • Azure
    • Databricks
    • Offline / On-Prem
    • Platform Architecture
  • 🏠Storing your Predictions
    • Overview
    • Real-time Models (Postgres)
    • Real-time Models (Kafka)
    • Batch Models
    • Kubeflow / KServe
  • 🧠Model Types
    • Regression
    • Binary Classification
    • Multiclass Classification
    • Multi-Label Classification
    • Ranking
  • 🌈Explainability
    • SHAP values
  • 📜NLP
    • Intro to NLP Monitoring
    • Example: Text Classification
    • Example: Token Classification
    • Example: Question Answering
  • 🍪Data Sources
    • Overview
    • Amazon S3
    • Athena
    • BigQuery
    • Databricks
    • Glue Data Catalog
    • Google Cloud Storage
    • PostgreSQL
    • Redshift
    • Snowflake
    • Microsoft SQL Server
    • Oracle
  • ⚡Monitors & Alerts
    • Overview
    • Data Drift
    • Metric Change
    • Missing Values
    • Model Activity
    • Model Staleness
    • Performance Degradation
    • Prediction Drift
    • Value Range
    • Custom Metric
    • New Values
    • Alerts Consolidation
  • 🎨Dashboards
    • Overview
  • 🤖ML Monitoring as Code
    • Getting started
    • Adding new models
    • Data Segments
    • Custom metrics
    • Querying metrics
    • Monitors
    • Dashboards
  • 📡Integrations
    • Slack
    • Webhook
    • Teams
    • Single Sign On (SAML)
    • Cisco
  • 🔐Administration
    • Role Based Access Control (RBAC)
  • 🔑API Reference
    • REST API
    • API Extended Reference
    • Custom Segment Syntax
    • Custom Metric Syntax
    • Code-Based Metrics
    • Metrics Glossary
  • ⏩Release Notes
    • Release Notes 2024
    • Release Notes 2023
Powered by GitBook
On this page
  • Create a materialization dataset for Aporia queries
  • Update the Aporia Service Account for BigQuery access
  • Create a BigQuery data source in Aporia
  1. Data Sources

BigQuery

PreviousAthenaNextDatabricks

Last updated 1 year ago

This guide describes how to connect Aporia to a BigQuery data source in order to monitor your ML Model in production.

We will assume that your model inputs, outputs and optionally delayed actuals can be queried with SQL. This data source may also be used to connect to your model's training set to be used as a baseline for model monitoring.

Create a materialization dataset for Aporia queries

Create a materialization dataset for Aporia to use to perform queries, see instructions .

A separate materialization dataset location, to which query results will be written, must be designated for each project from which you want to query.

Update the Aporia Service Account for BigQuery access

In order to provide access to BigQuery, you'll need to update your Aporia service account with the necessary API permissions.

Step 1: Obtain your aporia service account

Use the same service account used for the Aporia deployment. If someone else on your team has deployed Aporia, please reach out to them to obtain it.

Step 2: Grant read access to the relevant project

  1. Go to the and login.

  2. Find the Aporia service account you obtain in the previous step and click on 🖋 Edit Principle

  3. In the "Edit access" window click on ADD ANOTHER ROLE

  4. Add the BigQuery Data Viewer and BigQuery Job User roles and click Save

Step 3: Grant access to the materialization dataset

  1. In the left-hand panel, expand the relevant project and find the materialization dataset you created in the previous steps.

  2. Click on "..." by the dataset name, then click on Share

  3. In the "Share permissions" window click on Add Principal

  4. In the "New principal" box, enter the email of the Aporia service account you have obtained. Choose the BigQuery Data Editor role and click Save.

Now Aporia has the permission it needs to connect to the BigQuery datasets and tables you have specified in the policy.

Create a BigQuery data source in Aporia

  1. Go to Integrations page and click on the Data Connectors tab

  2. Scroll to Connect New Data Source section

  3. Click Connect on the BigQuery card and follow the instructions

Go to the and login.

Go to and login to your account.

Bravo! now you can use the data source you've created across all your models in Aporia.

🍪
👏
BigQuery console
Aporia platform
here
IAM console