|
4 | 4 | require "digest"
|
5 | 5 | require "rspec/wait"
|
6 | 6 |
|
| 7 | +def thread_it(kafka_input, queue) |
| 8 | + Thread.new do |
| 9 | + begin |
| 10 | + kafka_input.run(queue) |
| 11 | + end |
| 12 | + end |
| 13 | +end |
| 14 | + |
| 15 | +def run_with_kafka(&block) |
| 16 | + queue = Queue.new |
| 17 | + t = thread_it(kafka_input, queue) |
| 18 | + begin |
| 19 | + wait(timeout_seconds).for {queue.length}.to eq(expected_num_events) |
| 20 | + yield(queue) |
| 21 | + ensure |
| 22 | + t.kill |
| 23 | + t.join(30_000) |
| 24 | + end |
| 25 | +end |
| 26 | + |
| 27 | +shared_examples 'consumes all expected messages' do |
| 28 | + it 'should consume all expected messages' do |
| 29 | + run_with_kafka do |queue| |
| 30 | + expect(queue.length).to eq(expected_num_events) |
| 31 | + end |
| 32 | + end |
| 33 | +end |
| 34 | + |
7 | 35 | # Please run kafka_test_setup.sh prior to executing this integration test.
|
8 | 36 | describe "inputs/kafka", :integration => true do
|
| 37 | + subject(:kafka_input) { LogStash::Inputs::Kafka.new(config) } |
| 38 | + let(:execution_context) { double("execution_context")} |
| 39 | + |
| 40 | + before :each do |
| 41 | + allow(kafka_input).to receive(:execution_context).and_return(execution_context) |
| 42 | + allow(execution_context).to receive(:pipeline_id).and_return(pipeline_id) |
| 43 | + end |
| 44 | + |
9 | 45 | # Group ids to make sure that the consumers get all the logs.
|
10 | 46 | let(:group_id_1) {rand(36**8).to_s(36)}
|
11 | 47 | let(:group_id_2) {rand(36**8).to_s(36)}
|
12 | 48 | let(:group_id_3) {rand(36**8).to_s(36)}
|
13 | 49 | let(:group_id_4) {rand(36**8).to_s(36)}
|
14 |
| - let(:group_id_5) {rand(36**8).to_s(36)} |
15 |
| - let(:plain_config) { { 'topics' => ['logstash_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_1, 'auto_offset_reset' => 'earliest'} } |
16 |
| - let(:multi_consumer_config) { plain_config.merge({"group_id" => group_id_4, "client_id" => "spec", "consumer_threads" => 3}) } |
17 |
| - let(:snappy_config) { { 'topics' => ['logstash_topic_snappy'], 'codec' => 'plain', 'group_id' => group_id_1, 'auto_offset_reset' => 'earliest'} } |
18 |
| - let(:lz4_config) { { 'topics' => ['logstash_topic_lz4'], 'codec' => 'plain', 'group_id' => group_id_1, 'auto_offset_reset' => 'earliest'} } |
19 |
| - let(:pattern_config) { { 'topics_pattern' => 'logstash_topic_.*', 'group_id' => group_id_2, 'codec' => 'plain', 'auto_offset_reset' => 'earliest'} } |
20 |
| - let(:decorate_config) { { 'topics' => ['logstash_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_3, 'auto_offset_reset' => 'earliest', 'decorate_events' => true} } |
21 |
| - let(:manual_commit_config) { { 'topics' => ['logstash_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_5, 'auto_offset_reset' => 'earliest', 'enable_auto_commit' => 'false'} } |
| 50 | + let(:pipeline_id) {rand(36**8).to_s(36)} |
| 51 | + let(:config) { { 'codec' => 'plain', 'auto_offset_reset' => 'earliest'}} |
22 | 52 | let(:timeout_seconds) { 30 }
|
23 | 53 | let(:num_events) { 103 }
|
| 54 | + let(:expected_num_events) { num_events } |
24 | 55 |
|
25 |
| - describe "#kafka-topics" do |
26 |
| - def thread_it(kafka_input, queue) |
27 |
| - Thread.new do |
28 |
| - begin |
29 |
| - kafka_input.run(queue) |
30 |
| - end |
31 |
| - end |
32 |
| - end |
33 |
| - |
34 |
| - it "should consume all messages from plain 3-partition topic" do |
35 |
| - kafka_input = LogStash::Inputs::Kafka.new(plain_config) |
36 |
| - queue = Queue.new |
37 |
| - t = thread_it(kafka_input, queue) |
38 |
| - begin |
39 |
| - t.run |
40 |
| - wait(timeout_seconds).for {queue.length}.to eq(num_events) |
41 |
| - expect(queue.length).to eq(num_events) |
42 |
| - ensure |
43 |
| - t.kill |
44 |
| - t.join(30_000) |
45 |
| - end |
46 |
| - end |
47 |
| - |
48 |
| - it "should consume all messages from snappy 3-partition topic" do |
49 |
| - kafka_input = LogStash::Inputs::Kafka.new(snappy_config) |
50 |
| - queue = Queue.new |
51 |
| - t = thread_it(kafka_input, queue) |
52 |
| - begin |
53 |
| - t.run |
54 |
| - wait(timeout_seconds).for {queue.length}.to eq(num_events) |
55 |
| - expect(queue.length).to eq(num_events) |
56 |
| - ensure |
57 |
| - t.kill |
58 |
| - t.join(30_000) |
59 |
| - end |
60 |
| - end |
| 56 | + context 'from a plain 3 partition topic' do |
| 57 | + let(:config) { super.merge({ 'topics' => ['logstash_topic_plain'], 'group_id' => group_id_1}) } |
| 58 | + it_behaves_like 'consumes all expected messages' |
| 59 | + end |
61 | 60 |
|
62 |
| - it "should consume all messages from lz4 3-partition topic" do |
63 |
| - kafka_input = LogStash::Inputs::Kafka.new(lz4_config) |
64 |
| - queue = Queue.new |
65 |
| - t = thread_it(kafka_input, queue) |
66 |
| - begin |
67 |
| - t.run |
68 |
| - wait(timeout_seconds).for {queue.length}.to eq(num_events) |
69 |
| - expect(queue.length).to eq(num_events) |
70 |
| - ensure |
71 |
| - t.kill |
72 |
| - t.join(30_000) |
73 |
| - end |
74 |
| - end |
| 61 | + context 'from snappy 3 partition topic' do |
| 62 | + let(:config) { { 'topics' => ['logstash_topic_snappy'], 'codec' => 'plain', 'group_id' => group_id_1, 'auto_offset_reset' => 'earliest'} } |
| 63 | + it_behaves_like 'consumes all expected messages' |
| 64 | + end |
75 | 65 |
|
76 |
| - it "should consumer all messages with multiple consumers" do |
77 |
| - kafka_input = LogStash::Inputs::Kafka.new(multi_consumer_config) |
78 |
| - queue = Queue.new |
79 |
| - t = thread_it(kafka_input, queue) |
80 |
| - begin |
81 |
| - t.run |
82 |
| - wait(timeout_seconds).for {queue.length}.to eq(num_events) |
83 |
| - expect(queue.length).to eq(num_events) |
84 |
| - kafka_input.kafka_consumers.each_with_index do |consumer, i| |
85 |
| - expect(consumer.metrics.keys.first.tags["client-id"]).to eq("spec-#{i}-main") |
86 |
| - end |
87 |
| - ensure |
88 |
| - t.kill |
89 |
| - t.join(30_000) |
90 |
| - end |
91 |
| - end |
| 66 | + context 'from lz4 3 partition topic' do |
| 67 | + let(:config) { { 'topics' => ['logstash_topic_lz4'], 'codec' => 'plain', 'group_id' => group_id_1, 'auto_offset_reset' => 'earliest'} } |
| 68 | + it_behaves_like 'consumes all expected messages' |
92 | 69 | end
|
93 | 70 |
|
94 |
| - describe "#kafka-topics-pattern" do |
95 |
| - def thread_it(kafka_input, queue) |
96 |
| - Thread.new do |
97 |
| - begin |
98 |
| - kafka_input.run(queue) |
99 |
| - end |
100 |
| - end |
101 |
| - end |
| 71 | + context 'manually committing' do |
| 72 | + let(:config) { { 'topics' => ['logstash_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_2, 'auto_offset_reset' => 'earliest', 'enable_auto_commit' => 'false'} } |
| 73 | + it_behaves_like 'consumes all expected messages' |
| 74 | + end |
102 | 75 |
|
103 |
| - it "should consume all messages from all 3 topics" do |
104 |
| - kafka_input = LogStash::Inputs::Kafka.new(pattern_config) |
105 |
| - queue = Queue.new |
106 |
| - t = thread_it(kafka_input, queue) |
107 |
| - begin |
108 |
| - t.run |
109 |
| - wait(timeout_seconds).for {queue.length}.to eq(3*num_events) |
110 |
| - expect(queue.length).to eq(3*num_events) |
111 |
| - ensure |
112 |
| - t.kill |
113 |
| - t.join(30_000) |
114 |
| - end |
115 |
| - end |
| 76 | + context 'using a pattern to consume from all 3 topics' do |
| 77 | + let(:config) { { 'topics_pattern' => 'logstash_topic_.*', 'group_id' => group_id_3, 'codec' => 'plain', 'auto_offset_reset' => 'earliest'} } |
| 78 | + let(:expected_num_events) { 3*num_events } |
| 79 | + it_behaves_like 'consumes all expected messages' |
116 | 80 | end
|
117 | 81 |
|
118 |
| - describe "#kafka-decorate" do |
119 |
| - def thread_it(kafka_input, queue) |
120 |
| - Thread.new do |
121 |
| - begin |
122 |
| - kafka_input.run(queue) |
| 82 | + context "with multiple consumers" do |
| 83 | + let(:config) { super.merge({'topics' => ['logstash_topic_plain'], "group_id" => group_id_4, "client_id" => "spec", "consumer_threads" => 3}) } |
| 84 | + it 'should should consume all messages' do |
| 85 | + run_with_kafka do |queue| |
| 86 | + expect(queue.length).to eq(num_events) |
| 87 | + kafka_input.kafka_consumers.each_with_index do |consumer, i| |
| 88 | + expect(consumer.metrics.keys.first.tags["client-id"]).to eq("spec-#{i}-#{pipeline_id}") |
123 | 89 | end
|
124 | 90 | end
|
125 | 91 | end
|
| 92 | + end |
126 | 93 |
|
| 94 | + context 'with decorate events set to true' do |
| 95 | + let(:config) { { 'topics' => ['logstash_topic_plain'], 'codec' => 'plain', 'group_id' => group_id_3, 'auto_offset_reset' => 'earliest', 'decorate_events' => true} } |
127 | 96 | it "should show the right topic and group name in decorated kafka section" do
|
128 | 97 | start = LogStash::Timestamp.now.time.to_i
|
129 |
| - kafka_input = LogStash::Inputs::Kafka.new(decorate_config) |
130 |
| - queue = Queue.new |
131 |
| - t = thread_it(kafka_input, queue) |
132 |
| - begin |
133 |
| - t.run |
134 |
| - wait(timeout_seconds).for {queue.length}.to eq(num_events) |
| 98 | + run_with_kafka do |queue| |
135 | 99 | expect(queue.length).to eq(num_events)
|
136 | 100 | event = queue.shift
|
137 | 101 | expect(event.get("[@metadata][kafka][topic]")).to eq("logstash_topic_plain")
|
138 | 102 | expect(event.get("[@metadata][kafka][consumer_group]")).to eq(group_id_3)
|
139 | 103 | expect(event.get("[@metadata][kafka][timestamp]")).to be >= start
|
140 |
| - ensure |
141 |
| - t.kill |
142 |
| - t.join(30_000) |
143 |
| - end |
144 |
| - end |
145 |
| - end |
146 |
| - |
147 |
| - describe "#kafka-offset-commit" do |
148 |
| - def thread_it(kafka_input, queue) |
149 |
| - Thread.new do |
150 |
| - begin |
151 |
| - kafka_input.run(queue) |
152 |
| - end |
153 |
| - end |
154 |
| - end |
155 |
| - |
156 |
| - it "should manually commit offsets" do |
157 |
| - kafka_input = LogStash::Inputs::Kafka.new(manual_commit_config) |
158 |
| - queue = Queue.new |
159 |
| - t = thread_it(kafka_input, queue) |
160 |
| - begin |
161 |
| - t.run |
162 |
| - wait(timeout_seconds).for {queue.length}.to eq(num_events) |
163 |
| - expect(queue.length).to eq(num_events) |
164 |
| - ensure |
165 |
| - t.kill |
166 |
| - t.join(30_000) |
167 | 104 | end
|
168 | 105 | end
|
169 | 106 | end
|
|
0 commit comments