![rw-book-cover](https://readwise-assets.s3.amazonaws.com/static/images/article2.74d541386bbf.png) ## Metadata - Author: [[services.google.com|Services.google]] - Full Title:: Firat Tekiner - Data Lakehouse Whitepaper - Category:: #🗞️Articles - URL:: https://services.google.com/fh/files/misc/building-a-data-lakehouse.pdf - Finished date:: [[2023-04-11]] ## Highlights > a Dataproc environment to > connect to either Google Cloud Storage or the BigQuery storage > subsystem and read/write data at storage speeds ([View Highlight](https://read.readwise.io/read/01gxqa9pqp92j9f5vjj0szchpk)) > BigQuery as a > data lakehouse ([View Highlight](https://read.readwise.io/read/01gxqan57y8e1d699dr2wfdw51)) The problem is compression... but they seem to be testing charging by "physical storage"? > A modern data warehouse like BigQuery can handle > massive data volumes and has cost parity with other > data storage mechanisms such as Cloud Storage. ([View Highlight](https://read.readwise.io/read/01gxqap16ekn0d1ees57brhfvy))