-
Notifications
You must be signed in to change notification settings - Fork 13.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Added support for Vertica Grains #515
Merged
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
For the record, which one is the working lib/dialect? |
Thanks for the PR btw! |
The original dialect is not working (It is using an old version of SQLAlchemy). My version works fine for the SQLAlchemy version used in Caravel. |
mistercrunch
pushed a commit
that referenced
this pull request
Jun 2, 2016
* Created migration to fix the bug * Working also on MySQL * Added support for Vertica Grains (#515) * Fix #529 1 "This Session's transaction has been rolled back" (#530) * Fixing the specific issue * Added an additional fix for a similar error in #529 Background: - When an object is modified by SQLAlchemy, it is invalidated so need to be fetched again from the DB - If there's an exception during a transaction, SQLAlchemy performs a rollback and mark the connection as dirty. Bug: - When handling exceptions, the exception handler tries to access the name of the cluster in the main object. Since the name has been invalidated due to a write, SQLAlchemy tries to fetch it on a 'dirty' connection and spits out an error. Solution: - Fetch the information for handling the exception before starting the process. * Modified the migration function to to automatically detect the the foreign keys based on the signature. It supports also sqlite using batch migrations * i18n: Fix typo in Druid cluster broker port label (#512) * Update models.py (#541) removing duplicated `user_id` def
zhaoyongjie
pushed a commit
to zhaoyongjie/incubator-superset
that referenced
this pull request
Nov 17, 2021
* add controls to chart * update package
zhaoyongjie
pushed a commit
to zhaoyongjie/incubator-superset
that referenced
this pull request
Nov 24, 2021
* add controls to chart * update package
zhaoyongjie
pushed a commit
to zhaoyongjie/incubator-superset
that referenced
this pull request
Nov 25, 2021
* add controls to chart * update package
zhaoyongjie
pushed a commit
to zhaoyongjie/incubator-superset
that referenced
this pull request
Nov 26, 2021
* add controls to chart * update package
mistercrunch
added
🏷️ bot
A label used by `supersetbot` to keep track of which PR where auto-tagged with release labels
🚢 0.9.1
labels
Feb 19, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Added support for HP Vertica
Vertica works with Caravel using the proper SQLAlchemy dialects (The original dialects has a problem with reflection https://github.com/LAlbertalli/vertica-sqlalchemy).
This patch adds the time grains specifications (from PgSQL) needed to do time aggregate.