1
0
Files
hudi/docs/gcs_filesystem.md
2017-05-18 10:30:34 -07:00

63 lines
1.8 KiB
Markdown

---
title: GCS Filesystem (experimental)
keywords: sql hive gcs spark presto
sidebar: mydoc_sidebar
permalink: gcs_hoodie.html
toc: false
summary: In this page, we go over how to configure hoodie with Google Cloud Storage.
---
Hoodie works with HDFS by default and GCS **regional** buckets provide an HDFS API with strong consistency.
## GCS Configs
There are two configurations required for Hoodie GCS compatibility:
- Adding GCS Credentials for Hoodie
- Adding required jars to classpath
### GCS Credentials
Add the required configs in your core-site.xml from where Hoodie can fetch them. Replace the `fs.defaultFS` with your GCS bucket name and Hoodie should be able to read/write from the bucket.
```xml
<property>
<name>fs.defaultFS</name>
<value>gs://hoodie-bucket</value>
</property>
<property>
<name>fs.gs.impl</name>
<value>com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystem</value>
<description>The FileSystem for gs: (GCS) uris.</description>
</property>
<property>
<name>fs.AbstractFileSystem.gs.impl</name>
<value>com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS</value>
<description>The AbstractFileSystem for gs: (GCS) uris.</description>
</property>
<property>
<name>fs.gs.project.id</name>
<value>GCS_PROJECT_ID</value>
</property>
<property>
<name>google.cloud.auth.service.account.enable</name>
<value>true</value>
</property>
<property>
<name>google.cloud.auth.service.account.email</name>
<value>GCS_SERVICE_ACCOUNT_EMAIL</value>
</property>
<property>
<name>google.cloud.auth.service.account.keyfile</name>
<value>GCS_SERVICE_ACCOUNT_KEYFILE</value>
</property>
```
### GCS Libs
GCS hadoop libraries to add to our classpath
- com.google.cloud.bigdataoss:gcs-connector:1.6.0-hadoop2