Yes, BigQuery can technically handle 700B+ rows, however DBT should not handle that in one shot during a full_refresh. The best approach is partitioned, batched processing and that means breaking it down by day. Consider using DBT's microbatch strategy if your DBT version supports it, or implement a daily processing loop in your DAG orchestration.