Skip to content

Commit c29c130

Browse files
MINOR: Cleaning ignored streams tests (#20585)
- Test has been disabled for years + it tests ZK setup, but with KRaft there is no more a "controller", so cleaning it up. - Test has also been disabled for years + we did not get any incident about KS breaking during broker upgrade/downgrade
1 parent e37f31e commit c29c130

File tree

2 files changed

+1
-116
lines changed

2 files changed

+1
-116
lines changed

tests/kafkatest/tests/streams/streams_broker_bounce_test.py

Lines changed: 0 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,6 @@
1717
from ducktape.tests.test import Test
1818
from ducktape.mark.resource import cluster
1919
from ducktape.mark import matrix
20-
from ducktape.mark import ignore
2120
from kafkatest.services.kafka import KafkaService, quorum
2221
from kafkatest.services.zookeeper import ZookeeperService
2322
from kafkatest.services.streams import StreamsSmokeTestDriverService, StreamsSmokeTestJobRunnerService
@@ -239,29 +238,6 @@ def test_broker_type_bounce(self, failure_mode, broker_type, sleep_time_secs, nu
239238

240239
return self.collect_results(sleep_time_secs)
241240

242-
@ignore
243-
@cluster(num_nodes=7)
244-
@matrix(failure_mode=["clean_shutdown"],
245-
broker_type=["controller"],
246-
sleep_time_secs=[0])
247-
def test_broker_type_bounce_at_start(self, failure_mode, broker_type, sleep_time_secs):
248-
"""
249-
Start a smoke test client, then kill one particular broker immediately before streams stats
250-
Streams should throw an exception since it cannot create topics with the desired
251-
replication factor of 3
252-
"""
253-
self.setup_system(start_processor=False)
254-
255-
# Sleep to allow test to run for a bit
256-
time.sleep(sleep_time_secs)
257-
258-
# Fail brokers
259-
self.fail_broker_type(failure_mode, broker_type)
260-
261-
self.processor1.start()
262-
263-
return self.collect_results(sleep_time_secs)
264-
265241
@cluster(num_nodes=7)
266242
@matrix(failure_mode=["clean_shutdown", "hard_shutdown", "clean_bounce", "hard_bounce"],
267243
num_failures=[2],

tests/kafkatest/tests/streams/streams_upgrade_test.py

Lines changed: 1 addition & 92 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515

1616
import random
1717
import time
18-
from ducktape.mark import matrix, ignore
18+
from ducktape.mark import matrix
1919
from ducktape.mark.resource import cluster
2020
from ducktape.tests.test import Test
2121
from ducktape.utils.util import wait_until
@@ -111,97 +111,6 @@ def perform_broker_upgrade(self, to_version):
111111
node.version = KafkaVersion(to_version)
112112
self.kafka.start_node(node)
113113

114-
@ignore
115-
@cluster(num_nodes=6)
116-
@matrix(from_version=broker_upgrade_versions, to_version=broker_upgrade_versions)
117-
def test_upgrade_downgrade_brokers(self, from_version, to_version):
118-
"""
119-
Start a smoke test client then perform rolling upgrades on the broker.
120-
"""
121-
122-
if from_version == to_version:
123-
return
124-
125-
self.replication = 3
126-
self.num_kafka_nodes = 3
127-
self.partitions = 1
128-
self.isr = 2
129-
self.topics = {
130-
'echo' : { 'partitions': self.partitions, 'replication-factor': self.replication,
131-
'configs': {"min.insync.replicas": self.isr}},
132-
'data' : { 'partitions': self.partitions, 'replication-factor': self.replication,
133-
'configs': {"min.insync.replicas": self.isr} },
134-
'min' : { 'partitions': self.partitions, 'replication-factor': self.replication,
135-
'configs': {"min.insync.replicas": self.isr} },
136-
'max' : { 'partitions': self.partitions, 'replication-factor': self.replication,
137-
'configs': {"min.insync.replicas": self.isr} },
138-
'sum' : { 'partitions': self.partitions, 'replication-factor': self.replication,
139-
'configs': {"min.insync.replicas": self.isr} },
140-
'dif' : { 'partitions': self.partitions, 'replication-factor': self.replication,
141-
'configs': {"min.insync.replicas": self.isr} },
142-
'cnt' : { 'partitions': self.partitions, 'replication-factor': self.replication,
143-
'configs': {"min.insync.replicas": self.isr} },
144-
'avg' : { 'partitions': self.partitions, 'replication-factor': self.replication,
145-
'configs': {"min.insync.replicas": self.isr} },
146-
'wcnt' : { 'partitions': self.partitions, 'replication-factor': self.replication,
147-
'configs': {"min.insync.replicas": self.isr} },
148-
'tagg' : { 'partitions': self.partitions, 'replication-factor': self.replication,
149-
'configs': {"min.insync.replicas": self.isr} }
150-
}
151-
152-
# Setup phase
153-
self.zk = ZookeeperService(self.test_context, num_nodes=1)
154-
self.zk.start()
155-
156-
# number of nodes needs to be >= 3 for the smoke test
157-
self.kafka = KafkaService(self.test_context, num_nodes=self.num_kafka_nodes,
158-
zk=self.zk, version=KafkaVersion(from_version), topics=self.topics)
159-
self.kafka.start()
160-
161-
# allow some time for topics to be created
162-
wait_until(lambda: self.confirm_topics_on_all_brokers(set(self.topics.keys())),
163-
timeout_sec=60,
164-
err_msg="Broker did not create all topics in 60 seconds ")
165-
166-
self.driver = StreamsSmokeTestDriverService(self.test_context, self.kafka)
167-
168-
processor = StreamsSmokeTestJobRunnerService(self.test_context, self.kafka, "at_least_once")
169-
170-
with self.driver.node.account.monitor_log(self.driver.STDOUT_FILE) as driver_monitor:
171-
self.driver.start()
172-
173-
with processor.node.account.monitor_log(processor.STDOUT_FILE) as monitor:
174-
processor.start()
175-
monitor.wait_until(self.processed_data_msg,
176-
timeout_sec=60,
177-
err_msg="Never saw output '%s' on " % self.processed_data_msg + str(processor.node))
178-
179-
connected_message = "Discovered group coordinator"
180-
with processor.node.account.monitor_log(processor.LOG_FILE) as log_monitor:
181-
with processor.node.account.monitor_log(processor.STDOUT_FILE) as stdout_monitor:
182-
self.perform_broker_upgrade(to_version)
183-
184-
log_monitor.wait_until(connected_message,
185-
timeout_sec=120,
186-
err_msg=("Never saw output '%s' on " % connected_message) + str(processor.node.account))
187-
188-
stdout_monitor.wait_until(self.processed_data_msg,
189-
timeout_sec=60,
190-
err_msg="Never saw output '%s' on" % self.processed_data_msg + str(processor.node.account))
191-
192-
# SmokeTestDriver allows up to 6 minutes to consume all
193-
# records for the verification step so this timeout is set to
194-
# 6 minutes (360 seconds) for consuming of verification records
195-
# and a very conservative additional 2 minutes (120 seconds) to process
196-
# the records in the verification step
197-
driver_monitor.wait_until('ALL-RECORDS-DELIVERED\|PROCESSED-MORE-THAN-GENERATED',
198-
timeout_sec=480,
199-
err_msg="Never saw output '%s' on" % 'ALL-RECORDS-DELIVERED|PROCESSED-MORE-THAN-GENERATED' + str(self.driver.node.account))
200-
201-
self.driver.stop()
202-
processor.stop()
203-
processor.node.account.ssh_capture("grep SMOKE-TEST-CLIENT-CLOSED %s" % processor.STDOUT_FILE, allow_fail=False)
204-
205114
@cluster(num_nodes=6)
206115
@matrix(from_version=metadata_1_versions)
207116
@matrix(from_version=metadata_2_versions)

0 commit comments

Comments
 (0)