Cloud Storage Overview and gsutil for the PDE Exam

GCP Study Hub
619c7c8da6d7b95cf26f6f70
November 8, 2025

Cloud Storage shows up everywhere on the Professional Data Engineer exam. It is the staging area in front of BigQuery, the landing zone for Dataflow jobs, the source for Dataproc clusters, and the archive tier for anything that has aged out of hot analytics. If you cannot reason about buckets, objects, storage classes, and the gsutil command line, a surprising chunk of the exam becomes guesswork. Here is the mental model I want every Professional Data Engineer candidate to walk in with.

What Cloud Storage actually is

Cloud Storage is blob storage. It holds every data type you can throw at it, from CSV exports and Parquet files to images, videos, raw application logs, and database backups. It is Google Cloud's equivalent of AWS S3, and the terminology lines up almost one to one. In both services, the container is called a bucket, and an individual file inside that container is called an object.

That object model is important for the exam. Cloud Storage does not care about folders the way a filesystem does. Forward slashes in object names just create the appearance of a folder hierarchy in the console. Under the hood, every object is keyed by its full path within the bucket. When you sync a directory tree or list a prefix, you are really asking Cloud Storage to filter by that key.

Why it ends up in so many pipelines

A few properties make Cloud Storage the default home for data on GCP:

  • It accepts every transfer type. Bulk loads from on prem, streaming writes from applications, scheduled exports from Cloud SQL or BigQuery, raw files dropped by partners, it all lands in a bucket.
  • It is cheap relative to other storage tiers. Backups, archives, and infrequently read logs do not need to sit in a database. Cloud Storage is usually the right tier.
  • Access control is granular. You can set IAM at the bucket level for the common case and reach down to ACLs on individual objects when you genuinely need to share or restrict a single file.
  • Versioning and redundancy are built in. Versioning lets you recover from an accidental overwrite or delete by keeping noncurrent versions of an object. Redundancy ensures your data is replicated across locations so a single failure does not take it offline.
  • Storage classes and regions are configurable. You can put hot analytics data in Standard, cooler data in Nearline or Coldline, and rarely accessed archives in Archive. You can also pick the region or multi region the bucket lives in to satisfy compliance or latency requirements.

For the Professional Data Engineer exam, the lifecycle angle matters a lot. If a question describes data that gets queried daily for thirty days, then quarterly for a year, then almost never, the right answer is almost always a lifecycle rule that transitions objects between storage classes on a schedule.

Access control and the data engineer scenarios that come up

Most exam questions about Cloud Storage permissions are not asking you to memorize IAM role names. They are asking you to pick the right scope. Two patterns show up repeatedly:

  • A pipeline service account needs to read raw files and write transformed output. The right answer is a role bound at the bucket level, not project wide.
  • An external partner needs to drop files into one prefix without seeing anything else. The right answer is bucket level IAM with conditions, or signed URLs if it is a one off transfer.

Knowing that bucket scope exists, and that object level ACLs are still available for the edge cases, is usually enough.

gsutil, the command line you have to know

gsutil is the command line tool built specifically for Cloud Storage. The exam can ask about either gcloud or gsutil for storage tasks, so you should be comfortable reading both. In practice, the storage specific verbs live under gsutil.

The handful of commands worth burning into memory:

gsutil cp local-file.csv gs://my-bucket/path/
gsutil cp gs://my-bucket/path/file.csv ./
gsutil rsync -r ./local-dir gs://my-bucket/mirror
gsutil ls gs://my-bucket/path/
gsutil rm gs://my-bucket/path/old-file.csv

A few specifics that have helped Professional Data Engineer candidates pick the right answer on the exam:

  • cp is the workhorse for one off uploads and downloads. It handles single files and recursive directory copies with the -r flag.
  • rsync is the right verb when you want a directory or bucket to mirror another location. It only transfers what has changed, which matters when the source is large.
  • ls lists bucket contents and is the fastest way to confirm an object actually landed where you expected.
  • Parallel composite uploads let gsutil split a large file into chunks and upload them in parallel, then reassemble them server side. When an exam scenario mentions a multi gigabyte upload that is bottlenecked on bandwidth, that is the feature being hinted at.

How I study this for the exam

I treat Cloud Storage as the connective tissue between everything else on the Professional Data Engineer blueprint. When you read a scenario, ask three questions. Where is the data landing. What storage class fits the access pattern. Which command or tool moves it where it needs to go next. If you can answer those three, you can untangle almost any storage question on the exam.

My Professional Data Engineer course covers Cloud Storage buckets, storage classes, lifecycle rules, and the gsutil commands you need to recognize on exam day.

Get tips and updates from GCP Study Hub

arrow