Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Kafka source #22

Closed
tims opened this issue Dec 31, 2018 · 12 comments
Closed

Kafka source #22

tims opened this issue Dec 31, 2018 · 12 comments
Assignees
Labels
kind/feature New feature or request

Comments

@tims
Copy link
Contributor

tims commented Dec 31, 2018

The kafka IO PR #19 needs some follow ups.

  • It appears to have broken the build? Or another merge removed an import. IOException symbol not found in ImportJob.
  • Import job gets google application default creds and uses options.setGcpCredential? Why? I've not needed to do that before if I set them in my environment. It would also seems odd to get an error logged about it if you're using the direct or flink runner and not using GCP at all.
  • ImportJobOptions should not inherit FlinkPipelineOptions and GcpOptions, as that suggests it should always use Flink and GCP. These should already be registered by default, so you can just make use of them using options.as(GcpOptions.class) if needed.
@tims
Copy link
Contributor Author

tims commented Dec 31, 2018

I also feel that the deserializer classes should be either inner classes or in a separate package with the feature row kafka IO class. As they are very kafka specific. But we can just move them when we get around to modularizing all the source classes.

@zhilingc
Copy link
Collaborator

zhilingc commented Dec 31, 2018

I think i accidentally removed the IOException import when I did the merge for PR #12, sorry :(

@tims
Copy link
Contributor Author

tims commented Dec 31, 2018

I'm fixing it, pull request incoming

@tims tims mentioned this issue Dec 31, 2018
@tims
Copy link
Contributor Author

tims commented Dec 31, 2018

@baskaranz can you take a look at this.

Also, can you please use the google-java-format plugin if you're using intellij, we standardised on that.

@tims
Copy link
Contributor Author

tims commented Dec 31, 2018

I think we should also write an integration test before we close this issue.

@woop woop mentioned this issue Jan 2, 2019
@woop
Copy link
Member

woop commented Jan 3, 2019

I think we should also write an integration test before we close this issue.

I agree. Not having an integration test is part of the challenge. We merge based on trust that all relevant testing was done. Integration tests would remove that problem.

@tims
Copy link
Contributor Author

tims commented Jan 3, 2019

@baskaranz can you take on making unit test for the FeatureRowKafkaIO?

@baskaranz
Copy link
Contributor

baskaranz commented Jan 3, 2019 via email

@tims
Copy link
Contributor Author

tims commented Jan 10, 2019

@baskaranz any updates on this issue?

As discussed in #13. Can you also remove the FeatureRowKey proto message? We don't actually care what the key is, as everything is in the message, so we don't need it when reading. That's a concern for the publisher.

@pradithya pradithya added the kind/feature New feature or request label Jan 15, 2019
@zhilingc
Copy link
Collaborator

One more bug that I'd like to tag to this issue:

On the core side, the import spec validation prevents the submission of kafka jobs.

@woop
Copy link
Member

woop commented Jan 26, 2019

@tims @baskaranz @zhilingc, what is the current status of this issue, are we still pending the following?

  • Creation of unit tests for FeatureRowKafkaIO
  • Creation of integraton tests for testing KafkaIO
  • Removal of FeatureRowKey proto message
  • Fixing Import spec validation that prevents submission of Kafka jobs

@baskaranz
Copy link
Contributor

baskaranz commented Feb 14, 2019

@woop apart from Creation of unit tests for FeatureRowKafkaIO, the rest of the items were taken care since there's an issue running the tests using FeatureRowKafkaIO. I'm working with @pradithya on this. It works anyway, we use it to read from feature stream. We can close this issue for now and open a new issue for Creation of unit tests for FeatureRowKafkaIO.

Yanson pushed a commit to Yanson/feast that referenced this issue Jul 29, 2020
Closes KE-654

Added third data science scenario, ingestion of data for fraud detection.
dmartinol pushed a commit to dmartinol/feast that referenced this issue Jun 27, 2024
docstrings for permissions package
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/feature New feature or request
Projects
None yet
Development

No branches or pull requests

5 participants