Snowplow
This code sample demonstrates how to read data from Snowplow and publish it to a Kafka topic.
Note that this connector works with Snowplow instances on AWS Kinesis, so you will need connection details for Kinesis.
How to run
Create a Quix account or log-in and visit the Samples to use this project.
Clicking Deploy
on the Sample, deploys a pre-built container in Quix. Complete the environment variables to configure the container.
Clicking Edit code
on the Sample, forks the project to your own Git repo so you can customize it before deploying.
Environment variables
The code sample uses the following environment variables:
- output: This is the Quix Topic that will receive the stream.
- aws_access_key_id: AWS Access Key Id.
- aws_secret_access_key: AWS Secret Access Key.
- aws_region_name: AWS Region Name.
- aws_stream_name: The name of the AWS stream you want to use.
Requirements/prerequisites
You will need Snowplow deployed to AWS to use this project.
Contribute
Submit forked projects to the Quix GitHub repo. Any new project that we accept will be attributed to you and you'll receive $200 in Quix credit.
Open source
This project is open source under the Apache 2.0 license and available in our GitHub repo.
Please star us and mention us on social to show your appreciation.