Add this line to your application's Gemfile:
gem "streamy"
Streamy includes support for two different types of event encoding (JSON and Avro).
Add this to config/initializer/streamy.rb
require "streamy/message_buses/kafka_message_bus"
Streamy.message_bus = Streamy::MessageBuses::KafkaMessageBus.new(
client_id: "streamy",
seed_brokers: "broker.remote:9092",
ssl_ca_certs_from_system: true
)
Create an event:
module Events
class ReceivedPayment < Streamy::JsonEvent
def topic
"payments.transactions"
end
def body
{
amount: 200
}
end
def event_time
Time.now
end
end
end
Publish it:
Events::ReceivedPayment.publish
Add this to config/initializer/streamy.rb
require "streamy/message_buses/kafka_message_bus"
Streamy.message_bus = Streamy::MessageBuses::KafkaMessageBus.new(
client_id: "streamy",
seed_brokers: "broker.remote:9092",
ssl_ca_certs_from_system: true,
)
Streamy.configure do |config|
config.avro_schema_registry_url = "http://registry.example.com",
config.avro_schemas_path = "app/schemas"
end
Default schemas path is "app/schemas" Schema Registry Url is required for encoding with Avro
Create an event:
module Events
class ReceivedPayment < Streamy::AvroEvent
def topic
"payments.transactions"
end
def body
{
amount: 200
}
end
def event_time
Time.now
end
end
end
Create Avro schema (received_payment.asvc
) for event in schema path above:
{
"type": "record",
"name": "received_payment",
"fields": [
{
"name": "type",
"type": "string"
},
{
"name": "event_time",
"type": {
"type": "long",
"logicalType": "timestamp-micros"
}
},
{
"name": "body",
"type": {
"type": "record",
"name": "body",
"fields": [
{
"name": "amount",
"type": ["null", "int"],
"default": null
}
]
}
}
]
}
Publish event:
Events::ReceivedPayment.publish
We use karafka to handle the bulk of the consumer logic. You can also use karafka/avro to consume Avro encoded events.
Configure karafka consumer:
class ApplicationConsumer < Karafka::BaseConsumer
def consume
params_batch.each do |message|
Streamy::MessageProcessor.new(message).run
end
end
end
Add event handler(s):
# app/handlers/received_payment_handler.rb
class ReceivedPaymentHandler
def initialize(body)
@body = message
end
def process
PaymentCounter.increment(body[:amount])
end
private
attr_reader :body
end
Streamy provides an avro deserializer that is used with the schema registry as detailed above. More information on serialization can be found here
Put the following line in your karafka routes file:
deserializer Streamy::Deserializers::AvroDeserializer.new
You can choose a priority for your events. This is done by overriding the priority
method on your event:
:low
- The event will be sent to Kafka by a background thread, events are buffered untildelivery_threshold
messages are waiting or untildelivery_interval
seconds have passed since the last delivery. Calling publish on a low priority event is non blocking, and no errors should be thrown, unless the buffer is full.:standard
(default) - The event will be sent to Kafka by a background thread, but the thread is signaled to send any buffered events as soon as possible. The call to publish is non blocking, and should not throw errors, unless the buffer is full.:essential
- The event will be sent to Kafka immediately. The call to publish is blocking, and may throw errors.:batched
- The event will be queued to send to Kafka using a synchronous producer, but no events are sent untilbatched_message_limit
is reached (which is set tomax_buffer_size - 1
), or the synchronous producer in the specific thread hasdeliver_messages
called by another service. This allows efficient event batching, when creating many events, e.g. in batch jobs. When a batch of events is being delivered the call to publish will block, and may throw errors.
Please read the ruby-kafka
notes here on buffering and error handling
To ensure that all :low
:batched
or :standard
priority events are published Streamy.shutdown
should be called before your process exits to avoid losing any events.
Streamy automatically adds an at_exit
hook to initiate this, but if you are doing something unusual you might need to be aware of this.
Streamy provides a few helpers to make testing a breeze:
it "does publish an received payment" do
ReceivedPayment.publish
expect_event(
type: "received_payment",
topic: "payments.transactions",
body: {
amount: 200
}
)
end
it "does not publish an received payment" do
ReceivedPayment.publish
expect_no_event(type: "received_payment")
end
def test_publish_received_payment
ReceivedPayment.publish
assert_event(
type: "received_payment",
topic: "payments.transactions",
body: {
amount: 200
}
)
end
Streamy will test your Avro messages against your Avro schemas located in your host application for type errors and schema composition errors. To do this you will need to set up your schema path and registry_url in your specs, and stub any requests to FakeConfluentSchemaRegistryServer
. Again for example with Webmock and RSpec:
RSpec.configure do |config|
config.before(:each) do
require "avro_turf/test/fake_confluent_schema_registry_server"
Streamy.configuration.avro_schema_registry_url = "http://registry.example.com"
Streamy.configuration.avro_schemas_path = "app/events/schemas"
stub_request(:any, /^#{Streamy.configuration.avro_schema_registry_url}/).to_rack(FakeConfluentSchemaRegistryServer)
end
end
The gem is available as open source under the terms of the MIT License.