Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Mapping multiple source streams into a single destination table on Postgres #2224

Open
varunbpatil opened this issue Feb 26, 2021 · 4 comments
Labels
connectors/destination/postgres connectors/destinations-database frozen Not being actively worked on lang/java normalization team/destinations Destinations team's backlog transformation Issues for future transformation backlog type/enhancement New feature or request

Comments

@varunbpatil
Copy link
Contributor

varunbpatil commented Feb 26, 2021

Tell us about the problem you're trying to solve

I have a 3rd party API from which I read industrial sensor data (time-series data) of the form (timestamp, value) for each of the close to 1000 sensors. I'm planning to write a custom HTTP source for this. The source's discover method returns one stream per sensor so that I can filter which sensors I want to sync in the Airbyte UI. So, I have close to 1000 source streams. On the Postgres destination (using TimescaleDB), I obviously don't want to create 1000 tables corresponding to 1000 streams. Instead, I want all streams to be merged into a single table with the format (stream_name/sensor name, timestamp, value).

Describe the solution you’d like

Ability to merge multiple source streams into a single destination table by adding the stream name as an additional column without having to write a custom destination in Airbyte, especially when all the streams have the same schema.

Describe the alternative you’ve considered or used

Custom pipelines.

┆Issue is synchronized with this Asana task by Unito

@jonniedarko
Copy link

@varunbpatil did you find a work around for this, I too would like to include the source as a colum in my tables

@albertlieyingadrian
Copy link

Interested in hearing the answers for this as well! Have you found a work around? @varunbpatil

@grishick grishick added the team/destinations Destinations team's backlog label Sep 27, 2022
@grishick
Copy link
Contributor

grishick commented Oct 1, 2022

@misteryeo this is potentially a feature for the future of normalization/transformation

@grishick grishick added normalization transformation Issues for future transformation backlog and removed team/databases labels Oct 1, 2022
@walker-philips
Copy link

I'd like to add another situation that would benefit from this:
Multiple source systems handling similar datasets, but varying primary keys among them. A separate/new source system is created that maps PK to PK of all source systems rather than trying to save/track PKs within each source system.

The PK mapping source can be merged with any of the other sources to allow correct importing into the desired destination (which is also a source). This way, systems can be agnostic of the unique identifiers of other systems, but still be able to send/receive data.

@bleonard bleonard added the frozen Not being actively worked on label Mar 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
connectors/destination/postgres connectors/destinations-database frozen Not being actively worked on lang/java normalization team/destinations Destinations team's backlog transformation Issues for future transformation backlog type/enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

8 participants