-
Notifications
You must be signed in to change notification settings - Fork 209
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: allow hive sql to be provided as config #312
Conversation
Codecov Report
@@ Coverage Diff @@
## master #312 +/- ##
==========================================
+ Coverage 74.30% 75.10% +0.79%
==========================================
Files 105 105
Lines 4492 4997 +505
Branches 419 518 +99
==========================================
+ Hits 3338 3753 +415
- Misses 1049 1127 +78
- Partials 105 117 +12
Continue to review full report at Codecov.
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Left one comment.
where_clause_suffix=conf.get_string(HiveTableMetadataExtractor.WHERE_CLAUSE_SUFFIX_KEY)) | ||
|
||
self.sql_stmt = conf.get_string(HiveTableMetadataExtractor.EXTRACT_SQL.format( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think it's missing closing bracket?
conf.get_string(HiveTableMetadataExtractor.EXTRACT_SQL.format
--> conf.get_string(HiveTableMetadataExtractor.EXTRACT_SQL).format
By the way, we may not need to add where clause if they provide SQL statement. WDYT?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
actually yeah, we should just let them provide the sql.
* commit 'e14b33e776929f8b020f1c6fec75d0fb83687693': (23 commits) Fix Athena sample DAG (amundsen-io#341) fix: Update postgres_sample_dag to set table extract job as upstream for elastic search publisher (amundsen-io#340) chore: mypy cleanup (convert last comment types, remove noqa imports) (amundsen-io#338) chore: Convert typings to mypy (amundsen-io#311) chore: replace all references of Lyft repo with Amundsen (amundsen-io#323) feat: add github actions for databuilder (amundsen-io#336) build: fix broken tests in Python 3.7, test in CI (amundsen-io#334) fix(deps): Unpin attrs (amundsen-io#332) ci: add dependabot config (amundsen-io#330) Change repo name in travis file (amundsen-io#324) tests: add mock for bigquery auth (amundsen-io#313) feat: allow hive sql to be provided as config (amundsen-io#312) chore: remove python2 (amundsen-io#310) chore: update deps for databuilder (amundsen-io#309) fix: cypher statement param issue in Neo4jStalenessRemovalTask (amundsen-io#307) fix: Added missing job tag key in hive_sample_dag.py (amundsen-io#308) feat: enhance glue extractor (amundsen-io#306) fix: Fix sql for missing columns and mysql based dialects (#550) (amundsen-io#305) docs: Fix broken doc link to dashboard_execution model (amundsen-io#296) chore: apply license headers to all the source files (amundsen-io#304) ... # Conflicts: # README.md # databuilder/extractor/kafka_source_extractor.py # databuilder/publisher/neo4j_csv_publisher.py # docs/models.md # example/scripts/sample_data_loader.py # setup.py
Summary of Changes
This pr is to fix amundsen-io/amundsen#552 which allows user to provide hive metastore sql.
Tests
yes. add a unit test to test the new config.
Documentation
What documentation did you add or modify and why? Add any relevant links then remove this line
CheckList
Make sure you have checked all steps below to ensure a timely review.
make test