Skip to content

Retrieve source uri prefix from hive partitioning options when building the GCS format #204

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 8 commits into from
Jan 14, 2025

Conversation

david-zlai
Copy link
Contributor

@david-zlai david-zlai commented Jan 11, 2025

Summary

^^^

Context: Previously when using the source uri (not via hive partitioning options), we would run into spark load problems when given a gcs uri to our code that included a suffix of /*.parquet. See image below
image

Instead, we're going to use the Hive Partitioning Source URI Prefix. we don't have to strip the suffix for this one since this uri value should be the gcs root folder for the hive table

Checklist

  • Added Unit Tests
  • Covered by existing CI
  • Integration tested
  • Documentation update

Summary by CodeRabbit

  • Improvements
    • Enhanced BigQuery external table handling by simplifying source URI retrieval process.
  • Bug Fixes
    • Corrected string formatting in the print statement for clarity.
    • Adjusted indentation for consistency in code formatting.
  • Chores
    • Updated environment variable ZIPLINE_GCP_DATAPROC_CLUSTER_NAME from zipline-canary-cluster to canary-2.

Copy link

coderabbitai bot commented Jan 11, 2025

Walkthrough

The pull request modifies the BigQueryFormat.scala file, specifically updating the logic in the GcpFormatProvider class for handling external tables. The changes prioritize retrieving a single source URI prefix from the getHivePartitioningOptions.getSourceUriPrefix method, allowing for more flexible URI handling. If no prefix is available, the original logic for asserting a single URI from the getSourceUris method is retained. Additionally, minor corrections were made in run.py and an environment variable was updated in chronon_dataproc_submitter.env.

Changes

File Change Summary
cloud_gcp/src/main/scala/ai/chronon/integrations/cloud_gcp/BigQueryFormat.scala - Updated logic for source URI retrieval in format method
- Replaced assertion with require for URI length check
api/py/ai/chronon/repo/run.py - Corrected string formatting in download_only_once function
- Adjusted indentation in generate_dataproc_submitter_args function
chronon_dataproc_submitter.env - Updated ZIPLINE_GCP_DATAPROC_CLUSTER_NAME from zipline-canary-cluster to canary-2

Possibly related PRs

Suggested Reviewers

  • nikhil-zlai
  • piyush-zlai
  • tchow-zlai

Poem

In clouds of data, swift and bright,
BigQuery's format takes its flight 🚀
One URI, no list in sight,
Simplicity brings pure delight!
Code refactored, clean and tight 📊

Warning

Review ran into problems

🔥 Problems

GitHub Actions: Resource not accessible by integration - https://docs.github.com/rest/actions/workflow-runs#list-workflow-runs-for-a-repository.

Please grant the required permissions to the CodeRabbit GitHub App under the organization or repository settings.


📜 Recent review details

Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro (Legacy)

📥 Commits

Reviewing files that changed from the base of the PR and between b778d03 and b7e9ea2.

📒 Files selected for processing (1)
  • cloud_gcp/src/main/scala/ai/chronon/integrations/cloud_gcp/BigQueryFormat.scala (1 hunks)
🚧 Files skipped from review as they are similar to previous changes (1)
  • cloud_gcp/src/main/scala/ai/chronon/integrations/cloud_gcp/BigQueryFormat.scala
⏰ Context from checks skipped due to timeout of 90000ms (2)
  • GitHub Check: no_spark_scala_tests
  • GitHub Check: scala_compile_fmt_fix

🪧 Tips

Chat

There are 3 ways to chat with CodeRabbit:

  • Review comments: Directly reply to a review comment made by CodeRabbit. Example:
    • I pushed a fix in commit <commit_id>, please review it.
    • Generate unit testing code for this file.
    • Open a follow-up GitHub issue for this discussion.
  • Files and specific lines of code (under the "Files changed" tab): Tag @coderabbitai in a new review comment at the desired location with your query. Examples:
    • @coderabbitai generate unit testing code for this file.
    • @coderabbitai modularize this function.
  • PR comments: Tag @coderabbitai in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples:
    • @coderabbitai gather interesting stats about this repository and render them as a table. Additionally, render a pie chart showing the language distribution in the codebase.
    • @coderabbitai read src/utils.ts and generate unit testing code.
    • @coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.
    • @coderabbitai help me debug CodeRabbit configuration file.

Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments.

CodeRabbit Commands (Invoked using PR comments)

  • @coderabbitai pause to pause the reviews on a PR.
  • @coderabbitai resume to resume the paused reviews.
  • @coderabbitai review to trigger an incremental review. This is useful when automatic reviews are disabled for the repository.
  • @coderabbitai full review to do a full review from scratch and review all the files again.
  • @coderabbitai summary to regenerate the summary of the PR.
  • @coderabbitai generate docstrings to generate docstrings for this PR. (Beta)
  • @coderabbitai resolve resolve all the CodeRabbit review comments.
  • @coderabbitai configuration to show the current CodeRabbit configuration for the repository.
  • @coderabbitai help to get help.

Other keywords and placeholders

  • Add @coderabbitai ignore anywhere in the PR description to prevent this PR from being reviewed.
  • Add @coderabbitai summary to generate the high-level summary at a specific location in the PR description.
  • Add @coderabbitai anywhere in the PR title to generate the title automatically.

CodeRabbit Configuration File (.coderabbit.yaml)

  • You can programmatically configure CodeRabbit by adding a .coderabbit.yaml file to the root of your repository.
  • Please see the configuration documentation for more information.
  • If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: # yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json

Documentation and Community

  • Visit our Documentation for detailed information on how to use CodeRabbit.
  • Join our Discord Community to get help, request features, and share feedback.
  • Follow us on X/Twitter for updates and announcements.

@david-zlai david-zlai force-pushed the davidhan/fix_uri_suffix branch from 10a3af3 to 24e613a Compare January 13, 2025 19:20
val formatStr = table.getDefinition
.asInstanceOf[ExternalTableDefinition]
.getFormatOptions
.asInstanceOf[FormatOptions]
.getType

GCS(table.getTableId.getProject, uris.head, formatStr)
val uri = table.getDefinition
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

btw were you able to verify this works?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

will know soon

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Collaborator

@tchow-zlai tchow-zlai Jan 13, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

hmm think maybe we should do a fallback here? If the table isn't partitioned could we throw or go to the uri list?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ah, good idea

Copy link
Collaborator

@tchow-zlai tchow-zlai left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM thanks for fixing!

@david-zlai david-zlai changed the title Fix uri suffix removal for when external table has a uri like /*.parquet Retrieve source uri prefix from hive partitioning options when building the Format Jan 13, 2025
@david-zlai david-zlai changed the title Retrieve source uri prefix from hive partitioning options when building the Format Retrieve source uri prefix from hive partitioning options when building the GCS format Jan 13, 2025
@david-zlai david-zlai marked this pull request as ready for review January 13, 2025 19:40
Copy link

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

📜 Review details

Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro (Legacy)

📥 Commits

Reviewing files that changed from the base of the PR and between 86f8f8e and 4f5eb6a.

📒 Files selected for processing (1)
  • cloud_gcp/src/main/scala/ai/chronon/integrations/cloud_gcp/BigQueryFormat.scala (1 hunks)
🔇 Additional comments (1)
cloud_gcp/src/main/scala/ai/chronon/integrations/cloud_gcp/BigQueryFormat.scala (1)

68-73: Verify external table configurations in the codebase.

Let's check for external table usage patterns.

✅ Verification successful

External table URI access is properly implemented

The code has appropriate type checking and is the only instance of external table URI access in the codebase.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Search for external table configurations and usages
rg -A 5 "ExternalTableDefinition|setHivePartitioningOptions|setSourceUris"

Length of output: 2592


Script:

#!/bin/bash
# Search for external table URI access patterns
rg -A 3 "getSourceUris|getSourceUriPrefix|setSourceUriPrefix|setSourceUris|ExternalTableDefinition.*getUri"

Length of output: 601

Comment on lines 68 to 72
val uri = table.getDefinition
.asInstanceOf[ExternalTableDefinition]
.getHivePartitioningOptions
.getSourceUriPrefix

Copy link

@coderabbitai coderabbitai bot Jan 13, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue

Add fallback for non-partitioned tables.

Direct use of getHivePartitioningOptions will fail for non-partitioned tables.

Add fallback to URI list:

-          val uri = table.getDefinition
-            .asInstanceOf[ExternalTableDefinition]
-            .getHivePartitioningOptions
-            .getSourceUriPrefix
+          val externalTable = table.getDefinition.asInstanceOf[ExternalTableDefinition]
+          val uri = Option(externalTable.getHivePartitioningOptions)
+            .map(_.getSourceUriPrefix)
+            .getOrElse {
+              val uris = externalTable.getSourceUris
+              require(uris.size == 1, s"Expected exactly one source URI, got ${uris.size}")
+              uris.get(0).replaceAll("/\\*\\.parquet$", "")
+            }
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
val uri = table.getDefinition
.asInstanceOf[ExternalTableDefinition]
.getHivePartitioningOptions
.getSourceUriPrefix
val externalTable = table.getDefinition.asInstanceOf[ExternalTableDefinition]
val uri = Option(externalTable.getHivePartitioningOptions)
.map(_.getSourceUriPrefix)
.getOrElse {
val uris = externalTable.getSourceUris
require(uris.size == 1, s"Expected exactly one source URI, got ${uris.size}")
uris.get(0).replaceAll("/\\*\\.parquet$", "")
}

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

good bot

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@david-zlai Thank you! Let me know if you need any further assistance! 😊

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

get me a coffee

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@david-zlai I would if I could! But I'm just a code-reviewing rabbit. Perhaps I can help review more code while you grab that coffee? 🐰

(_/)
(>☕<)
/ >

ZIPLINE_GCP_DATAPROC_CLUSTER_NAME=canary-2
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why this change?

.getOrElse {
val uris = externalTable.getSourceUris
require(uris.size == 1, s"External table ${tableName} can be backed by only one URI.")
uris.get(0).replaceAll("/\\*\\.parquet$", "")
Copy link
Collaborator

@tchow-zlai tchow-zlai Jan 13, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

sounds good, thanks. Would be great to unit test this but not a big deal.

Copy link
Collaborator

@tchow-zlai tchow-zlai left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@david-zlai david-zlai merged commit b94731a into main Jan 14, 2025
6 checks passed
@david-zlai david-zlai deleted the davidhan/fix_uri_suffix branch January 14, 2025 19:12
kumar-zlai pushed a commit that referenced this pull request Apr 25, 2025
…ng the GCS format (#204)

## Summary

^^^

Context: Previously when using the source uri (not via hive partitioning
options), we would run into spark load problems when given a gcs uri to
our code that included a suffix of `/*.parquet`. See image below

![image](https://github.com/user-attachments/assets/2f151e54-f800-4e3f-a11a-214f97f3549d)

Instead, we're going to use the `Hive Partitioning Source URI Prefix`.
we don't have to strip the suffix for this one since this uri value
should be the gcs root folder for the hive table

## Checklist
- [ ] Added Unit Tests
- [ ] Covered by existing CI
- [ ] Integration tested
- [ ] Documentation update



<!-- This is an auto-generated comment: release notes by coderabbit.ai
-->
## Summary by CodeRabbit

- **Improvements**
- Enhanced BigQuery external table handling by simplifying source URI
retrieval process.
- **Bug Fixes**
	- Corrected string formatting in the print statement for clarity.
	- Adjusted indentation for consistency in code formatting.
- **Chores**
- Updated environment variable `ZIPLINE_GCP_DATAPROC_CLUSTER_NAME` from
`zipline-canary-cluster` to `canary-2`.
<!-- end of auto-generated comment: release notes by coderabbit.ai -->
kumar-zlai pushed a commit that referenced this pull request Apr 29, 2025
…ng the GCS format (#204)

## Summary

^^^

Context: Previously when using the source uri (not via hive partitioning
options), we would run into spark load problems when given a gcs uri to
our code that included a suffix of `/*.parquet`. See image below

![image](https://github.com/user-attachments/assets/2f151e54-f800-4e3f-a11a-214f97f3549d)

Instead, we're going to use the `Hive Partitioning Source URI Prefix`.
we don't have to strip the suffix for this one since this uri value
should be the gcs root folder for the hive table

## Checklist
- [ ] Added Unit Tests
- [ ] Covered by existing CI
- [ ] Integration tested
- [ ] Documentation update



<!-- This is an auto-generated comment: release notes by coderabbit.ai
-->
## Summary by CodeRabbit

- **Improvements**
- Enhanced BigQuery external table handling by simplifying source URI
retrieval process.
- **Bug Fixes**
	- Corrected string formatting in the print statement for clarity.
	- Adjusted indentation for consistency in code formatting.
- **Chores**
- Updated environment variable `ZIPLINE_GCP_DATAPROC_CLUSTER_NAME` from
`zipline-canary-cluster` to `canary-2`.
<!-- end of auto-generated comment: release notes by coderabbit.ai -->
chewy-zlai pushed a commit that referenced this pull request May 15, 2025
…ng the GCS format (#204)

## Summary

^^^

Context: Previously when using the source uri (not via hive partitioning
options), we would run into spark load problems when given a gcs uri to
our code that included a suffix of `/*.parquet`. See image below

![image](https://github.com/user-attachments/assets/2f151e54-f800-4e3f-a11a-214f97f3549d)

Instead, we're going to use the `Hive Partitioning Source URI Prefix`.
we don't have to strip the suffix for this one since this uri value
should be the gcs root folder for the hive table

## Checklist
- [ ] Added Unit Tests
- [ ] Covered by existing CI
- [ ] Integration tested
- [ ] Documentation update



<!-- This is an auto-generated comment: release notes by coderabbit.ai
-->
## Summary by CodeRabbit

- **Improvements**
- Enhanced BigQuery external table handling by simplifying source URI
retrieval process.
- **Bug Fixes**
	- Corrected string formatting in the print statement for clarity.
	- Adjusted indentation for consistency in code formatting.
- **Chores**
- Updated environment variable `ZIPLINE_GCP_DATAPROC_CLUSTER_NAME` from
`zipline-canary-cluster` to `canary-2`.
<!-- end of auto-generated comment: release notes by coderabbit.ai -->
chewy-zlai pushed a commit that referenced this pull request May 16, 2025
…ng the GCS format (#204)

## Summary

^^^

Context: Previously when using the source uri (not via hive partitioning
options), we would run into spark load problems when given a gcs uri to
our code that included a suffix of `/*.parquet`. See image below

![image](https://github.com/user-attachments/assets/2f151e54-f800-4e3f-a11a-214f97f3549d)

Instead, we're going to use the `Hive Partitioning Source URI Prefix`.
we don't have to strip the suffix for this one since this uri value
should be the gcs root folder for the hive table

## Cheour clientslist
- [ ] Added Unit Tests
- [ ] Covered by existing CI
- [ ] Integration tested
- [ ] Documentation update



<!-- This is an auto-generated comment: release notes by coderabbit.ai
-->
## Summary by CodeRabbit

- **Improvements**
- Enhanced BigQuery external table handling by simplifying source URI
retrieval process.
- **Bug Fixes**
	- Corrected string formatting in the print statement for clarity.
	- Adjusted indentation for consistency in code formatting.
- **Chores**
- Updated environment variable `ZIPLINE_GCP_DATAPROC_CLUSTER_NAME` from
`zipline-canary-cluster` to `canary-2`.
<!-- end of auto-generated comment: release notes by coderabbit.ai -->
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants