From 6264cc155757216c878bcce48e02099a9f5a7d6f Mon Sep 17 00:00:00 2001 From: Ewen Cheslack-Postava Date: Sat, 28 Jan 2017 01:40:10 +0000 Subject: [PATCH] KAFKA-4450; Add upgrade tests for 0.10.1 and rename TRUNK to DEV_BRANCH to reduce confusion Author: Ewen Cheslack-Postava Reviewers: Ismael Juma Closes #2457 from ewencp/kafka-4450-upgrade-tests --- tests/docker/run_tests.sh | 4 ++-- tests/kafkatest/__init__.py | 2 +- .../benchmarks/core/benchmark_test.py | 18 +++++++++--------- tests/kafkatest/directory_layout/kafka_path.py | 14 +++++++------- .../sanity_checks/test_kafka_version.py | 6 +++--- .../sanity_checks/test_performance_services.py | 6 +++--- .../sanity_checks/test_verifiable_producer.py | 16 ++++++++-------- tests/kafkatest/services/console_consumer.py | 4 ++-- tests/kafkatest/services/kafka/kafka.py | 4 ++-- .../performance/consumer_performance.py | 4 ++-- .../services/performance/end_to_end_latency.py | 4 ++-- .../performance/producer_performance.py | 12 ++++++------ tests/kafkatest/services/security/minikdc.py | 6 +++--- .../kafkatest/services/verifiable_consumer.py | 4 ++-- .../kafkatest/services/verifiable_producer.py | 10 +++++----- tests/kafkatest/services/zookeeper.py | 4 ++-- .../client_compatibility_features_test.py | 4 ++-- ...lient_compatibility_produce_consume_test.py | 4 ++-- .../tests/client/message_format_change_test.py | 8 ++++---- .../core/compatibility_test_new_broker_test.py | 17 +++++++++-------- tests/kafkatest/tests/core/upgrade_test.py | 11 +++++++---- tests/kafkatest/version.py | 17 +++++++---------- .../directory_layout/check_project_paths.py | 14 +++++++------- tests/unit/version/check_version.py | 6 +++--- .../kafka/tools/VerifiableLog4jAppender.java | 2 +- .../apache/kafka/tools/VerifiableProducer.java | 2 +- vagrant/base.sh | 6 +++--- vagrant/broker.sh | 2 +- vagrant/zk.sh | 2 +- 29 files changed, 107 insertions(+), 106 deletions(-) diff --git a/tests/docker/run_tests.sh b/tests/docker/run_tests.sh index dadb031c251f1..0e5378a8f26b8 100755 --- a/tests/docker/run_tests.sh +++ b/tests/docker/run_tests.sh @@ -55,7 +55,7 @@ docker network inspect knw for i in $(seq -w 1 ${KAFKA_NUM_CONTAINERS}); do echo knode${i} - docker exec knode${i} bash -c "(tar xfz /kafka_src/core/build/distributions/kafka_*SNAPSHOT.tgz -C /opt || echo missing kafka tgz did you build kafka tarball) && mv /opt/kafka*SNAPSHOT /opt/kafka-trunk && ls -l /opt" + docker exec knode${i} bash -c "(tar xfz /kafka_src/core/build/distributions/kafka_*SNAPSHOT.tgz -C /opt || echo missing kafka tgz did you build kafka tarball) && mv /opt/kafka*SNAPSHOT /opt/kafka-dev && ls -l /opt" docker exec knode01 bash -c "ssh knode$i hostname" done @@ -64,7 +64,7 @@ done (cd ${KAFKA_SRC} && ./gradlew copyDependantTestLibs) for i in $(seq -w 1 ${KAFKA_NUM_CONTAINERS}); do echo knode${i} - docker exec knode${i} bash -c "cp /kafka_src/core/build/dependant-testlibs/* /opt/kafka-trunk/libs/" + docker exec knode${i} bash -c "cp /kafka_src/core/build/dependant-testlibs/* /opt/kafka-dev/libs/" docker exec knode01 bash -c "ssh knode$i hostname" done diff --git a/tests/kafkatest/__init__.py b/tests/kafkatest/__init__.py index b4b0562a44682..bc60fc8c980d1 100644 --- a/tests/kafkatest/__init__.py +++ b/tests/kafkatest/__init__.py @@ -19,7 +19,7 @@ # due to python version naming restrictions, which are enforced by python packaging tools # (see https://www.python.org/dev/peps/pep-0440/) # -# Instead, in trunk, the version should have a suffix of the form ".devN" +# Instead, in development branches, the version should have a suffix of the form ".devN" # # For example, when Kafka is at version 0.9.0.0-SNAPSHOT, this should be something like "0.9.0.0.dev0" __version__ = '0.10.3.0.dev0' diff --git a/tests/kafkatest/benchmarks/core/benchmark_test.py b/tests/kafkatest/benchmarks/core/benchmark_test.py index 14fab2f45ec65..b068fff63caf4 100644 --- a/tests/kafkatest/benchmarks/core/benchmark_test.py +++ b/tests/kafkatest/benchmarks/core/benchmark_test.py @@ -22,7 +22,7 @@ from kafkatest.services.kafka import KafkaService from kafkatest.services.performance import ProducerPerformanceService, EndToEndLatencyService, ConsumerPerformanceService, throughput, latency, compute_aggregate_throughput from kafkatest.services.zookeeper import ZookeeperService -from kafkatest.version import TRUNK, KafkaVersion +from kafkatest.version import DEV_BRANCH, KafkaVersion TOPIC_REP_ONE = "topic-replication-factor-one" TOPIC_REP_THREE = "topic-replication-factor-three" @@ -72,8 +72,8 @@ def start_kafka(self, security_protocol, interbroker_security_protocol, version) @cluster(num_nodes=7) @parametrize(acks=1, topic=TOPIC_REP_THREE, num_producers=3) def test_producer_throughput(self, acks, topic, num_producers=1, message_size=DEFAULT_RECORD_SIZE, - compression_type="none", security_protocol='PLAINTEXT', client_version=str(TRUNK), - broker_version=str(TRUNK)): + compression_type="none", security_protocol='PLAINTEXT', client_version=str(DEV_BRANCH), + broker_version=str(DEV_BRANCH)): """ Setup: 1 node zk + 3 node kafka cluster Produce ~128MB worth of messages to a topic with 6 partitions. Required acks, topic replication factor, @@ -104,8 +104,8 @@ def test_producer_throughput(self, acks, topic, num_producers=1, message_size=DE @parametrize(security_protocol='SSL', interbroker_security_protocol='PLAINTEXT') @matrix(security_protocol=['PLAINTEXT', 'SSL'], compression_type=["none", "snappy"]) def test_long_term_producer_throughput(self, compression_type="none", security_protocol='PLAINTEXT', - interbroker_security_protocol=None, client_version=str(TRUNK), - broker_version=str(TRUNK)): + interbroker_security_protocol=None, client_version=str(DEV_BRANCH), + broker_version=str(DEV_BRANCH)): """ Setup: 1 node zk + 3 node kafka cluster Produce 10e6 100 byte messages to a topic with 6 partitions, replication-factor 3, and acks=1. @@ -162,8 +162,8 @@ def test_long_term_producer_throughput(self, compression_type="none", security_p @cluster(num_nodes=6) @matrix(security_protocol=['SASL_PLAINTEXT', 'SASL_SSL'], compression_type=["none", "snappy"]) def test_end_to_end_latency(self, compression_type="none", security_protocol="PLAINTEXT", - interbroker_security_protocol=None, client_version=str(TRUNK), - broker_version=str(TRUNK)): + interbroker_security_protocol=None, client_version=str(DEV_BRANCH), + broker_version=str(DEV_BRANCH)): """ Setup: 1 node zk + 3 node kafka cluster Produce (acks = 1) and consume 10e3 messages to a topic with 6 partitions and replication-factor 3, @@ -194,7 +194,7 @@ def test_end_to_end_latency(self, compression_type="none", security_protocol="PL @matrix(security_protocol=['PLAINTEXT', 'SSL'], compression_type=["none", "snappy"]) def test_producer_and_consumer(self, compression_type="none", security_protocol="PLAINTEXT", interbroker_security_protocol=None, new_consumer=True, - client_version=str(TRUNK), broker_version=str(TRUNK)): + client_version=str(DEV_BRANCH), broker_version=str(DEV_BRANCH)): """ Setup: 1 node zk + 3 node kafka cluster Concurrently produce and consume 10e6 messages with a single producer and a single consumer, @@ -243,7 +243,7 @@ def test_producer_and_consumer(self, compression_type="none", security_protocol= @matrix(security_protocol=['PLAINTEXT', 'SSL'], compression_type=["none", "snappy"]) def test_consumer_throughput(self, compression_type="none", security_protocol="PLAINTEXT", interbroker_security_protocol=None, new_consumer=True, num_consumers=1, - client_version=str(TRUNK), broker_version=str(TRUNK)): + client_version=str(DEV_BRANCH), broker_version=str(DEV_BRANCH)): """ Consume 10e6 100-byte messages with 1 or more consumers from a topic with 6 partitions (using new consumer iff new_consumer == True), and report throughput. diff --git a/tests/kafkatest/directory_layout/kafka_path.py b/tests/kafkatest/directory_layout/kafka_path.py index 0e60affa2426a..ece8be58214e0 100644 --- a/tests/kafkatest/directory_layout/kafka_path.py +++ b/tests/kafkatest/directory_layout/kafka_path.py @@ -16,7 +16,7 @@ import importlib import os -from kafkatest.version import get_version, KafkaVersion, TRUNK +from kafkatest.version import get_version, KafkaVersion, DEV_BRANCH """This module serves a few purposes: @@ -43,7 +43,7 @@ TOOLS_DEPENDANT_TEST_LIBS_JAR_NAME = "tools-dependant-libs" JARS = { - "trunk": { + "dev": { CORE_JAR_NAME: "core/build/*/*.jar", CORE_LIBS_JAR_NAME: "core/build/libs/*.jar", CORE_DEPENDANT_TEST_LIBS_JAR_NAME: "core/build/dependant-testlibs/*.jar", @@ -97,7 +97,7 @@ def path(self): class KafkaSystemTestPathResolver(object): """Path resolver for Kafka system tests which assumes the following layout: - /opt/kafka-trunk # Current version of kafka under test + /opt/kafka-dev # Current version of kafka under test /opt/kafka-0.9.0.1 # Example of an older version of kafka installed from tarball /opt/kafka- # Other previous versions of kafka ... @@ -106,7 +106,7 @@ def __init__(self, context, project="kafka"): self.context = context self.project = project - def home(self, node_or_version=TRUNK): + def home(self, node_or_version=DEV_BRANCH): version = self._version(node_or_version) home_dir = self.project if version is not None: @@ -114,15 +114,15 @@ def home(self, node_or_version=TRUNK): return os.path.join(KAFKA_INSTALL_ROOT, home_dir) - def bin(self, node_or_version=TRUNK): + def bin(self, node_or_version=DEV_BRANCH): version = self._version(node_or_version) return os.path.join(self.home(version), "bin") - def script(self, script_name, node_or_version=TRUNK): + def script(self, script_name, node_or_version=DEV_BRANCH): version = self._version(node_or_version) return os.path.join(self.bin(version), script_name) - def jar(self, jar_name, node_or_version=TRUNK): + def jar(self, jar_name, node_or_version=DEV_BRANCH): version = self._version(node_or_version) return os.path.join(self.home(version), JARS[str(version)][jar_name]) diff --git a/tests/kafkatest/sanity_checks/test_kafka_version.py b/tests/kafkatest/sanity_checks/test_kafka_version.py index 35500935c69f4..7e65807270e5f 100644 --- a/tests/kafkatest/sanity_checks/test_kafka_version.py +++ b/tests/kafkatest/sanity_checks/test_kafka_version.py @@ -19,7 +19,7 @@ from kafkatest.services.kafka import KafkaService, config_property from kafkatest.services.zookeeper import ZookeeperService from kafkatest.utils import is_version -from kafkatest.version import LATEST_0_8_2, TRUNK +from kafkatest.version import LATEST_0_8_2, DEV_BRANCH class KafkaVersionTest(Test): @@ -47,12 +47,12 @@ def test_0_8_2(self): @cluster(num_nodes=3) def test_multi_version(self): """Test kafka service node-versioning api - ensure we can bring up a 2-node cluster, one on version 0.8.2.X, - the other on trunk.""" + the other on the current development branch.""" self.kafka = KafkaService(self.test_context, num_nodes=2, zk=self.zk, topics={self.topic: {"partitions": 1, "replication-factor": 2}}) self.kafka.nodes[1].version = LATEST_0_8_2 self.kafka.nodes[1].config[config_property.INTER_BROKER_PROTOCOL_VERSION] = "0.8.2.X" self.kafka.start() - assert is_version(self.kafka.nodes[0], [TRUNK.vstring]) + assert is_version(self.kafka.nodes[0], [DEV_BRANCH.vstring]) assert is_version(self.kafka.nodes[1], [LATEST_0_8_2]) diff --git a/tests/kafkatest/sanity_checks/test_performance_services.py b/tests/kafkatest/sanity_checks/test_performance_services.py index b939f2b61ce98..6ff34b8606bee 100644 --- a/tests/kafkatest/sanity_checks/test_performance_services.py +++ b/tests/kafkatest/sanity_checks/test_performance_services.py @@ -21,7 +21,7 @@ from kafkatest.services.performance import ProducerPerformanceService, ConsumerPerformanceService, EndToEndLatencyService from kafkatest.services.performance import latency, compute_aggregate_throughput from kafkatest.services.zookeeper import ZookeeperService -from kafkatest.version import TRUNK, LATEST_0_8_2, LATEST_0_9, KafkaVersion +from kafkatest.version import DEV_BRANCH, LATEST_0_8_2, LATEST_0_9, KafkaVersion class PerformanceServiceTest(Test): @@ -42,8 +42,8 @@ def setUp(self): @parametrize(version=str(LATEST_0_8_2), new_consumer=False) @parametrize(version=str(LATEST_0_9), new_consumer=False) @parametrize(version=str(LATEST_0_9)) - @parametrize(version=str(TRUNK), new_consumer=False) - @parametrize(version=str(TRUNK)) + @parametrize(version=str(DEV_BRANCH), new_consumer=False) + @parametrize(version=str(DEV_BRANCH)) def test_version(self, version=str(LATEST_0_9), new_consumer=True): """ Sanity check out producer performance service - verify that we can run the service with a small diff --git a/tests/kafkatest/sanity_checks/test_verifiable_producer.py b/tests/kafkatest/sanity_checks/test_verifiable_producer.py index 544d7b989193f..b4ca4e94f4b56 100644 --- a/tests/kafkatest/sanity_checks/test_verifiable_producer.py +++ b/tests/kafkatest/sanity_checks/test_verifiable_producer.py @@ -23,7 +23,7 @@ from kafkatest.services.verifiable_producer import VerifiableProducer from kafkatest.services.zookeeper import ZookeeperService from kafkatest.utils import is_version -from kafkatest.version import LATEST_0_8_2, LATEST_0_9, TRUNK, KafkaVersion +from kafkatest.version import LATEST_0_8_2, LATEST_0_9, DEV_BRANCH, KafkaVersion class TestVerifiableProducer(Test): @@ -48,10 +48,10 @@ def setUp(self): @cluster(num_nodes=3) @parametrize(producer_version=str(LATEST_0_8_2)) @parametrize(producer_version=str(LATEST_0_9)) - @parametrize(producer_version=str(TRUNK)) - def test_simple_run(self, producer_version=TRUNK): + @parametrize(producer_version=str(DEV_BRANCH)) + def test_simple_run(self, producer_version=DEV_BRANCH): """ - Test that we can start VerifiableProducer on trunk or against the 0.8.2 jar, and + Test that we can start VerifiableProducer on the current branch snapshot version or against the 0.8.2 jar, and verify that we can produce a small number of messages. """ node = self.producer.nodes[0] @@ -61,11 +61,11 @@ def test_simple_run(self, producer_version=TRUNK): err_msg="Producer failed to start in a reasonable amount of time.") # using version.vstring (distutils.version.LooseVersion) is a tricky way of ensuring - # that this check works with TRUNK - # When running VerifiableProducer 0.8.X, both trunk version and 0.8.X should show up because of the way - # verifiable producer pulls in some trunk directories into its classpath + # that this check works with DEV_BRANCH + # When running VerifiableProducer 0.8.X, both the current branch version and 0.8.X should show up because of the + # way verifiable producer pulls in some development directories into its classpath if node.version <= LATEST_0_8_2: - assert is_version(node, [node.version.vstring, TRUNK.vstring]) + assert is_version(node, [node.version.vstring, DEV_BRANCH.vstring]) else: assert is_version(node, [node.version.vstring]) diff --git a/tests/kafkatest/services/console_consumer.py b/tests/kafkatest/services/console_consumer.py index cdc46cdca985d..94acb65bd0b8b 100644 --- a/tests/kafkatest/services/console_consumer.py +++ b/tests/kafkatest/services/console_consumer.py @@ -21,7 +21,7 @@ from kafkatest.directory_layout.kafka_path import KafkaPathResolverMixin from kafkatest.services.monitor.jmx import JmxMixin -from kafkatest.version import TRUNK, LATEST_0_8_2, LATEST_0_9, LATEST_0_10_0, V_0_10_0_0 +from kafkatest.version import DEV_BRANCH, LATEST_0_8_2, LATEST_0_9, LATEST_0_10_0, V_0_10_0_0 """ 0.8.2.1 ConsoleConsumer options @@ -87,7 +87,7 @@ class ConsoleConsumer(KafkaPathResolverMixin, JmxMixin, BackgroundThreadService) } def __init__(self, context, num_nodes, kafka, topic, group_id="test-consumer-group", new_consumer=True, - message_validator=None, from_beginning=True, consumer_timeout_ms=None, version=TRUNK, + message_validator=None, from_beginning=True, consumer_timeout_ms=None, version=DEV_BRANCH, client_id="console-consumer", print_key=False, jmx_object_names=None, jmx_attributes=None, enable_systest_events=False, stop_timeout_sec=15, print_timestamp=False): """ diff --git a/tests/kafkatest/services/kafka/kafka.py b/tests/kafkatest/services/kafka/kafka.py index 8ef0f356f43cc..2e65e099ac437 100644 --- a/tests/kafkatest/services/kafka/kafka.py +++ b/tests/kafkatest/services/kafka/kafka.py @@ -30,7 +30,7 @@ from kafkatest.services.monitor.jmx import JmxMixin from kafkatest.services.security.minikdc import MiniKdc from kafkatest.services.security.security_config import SecurityConfig -from kafkatest.version import TRUNK +from kafkatest.version import DEV_BRANCH Port = collections.namedtuple('Port', ['name', 'number', 'open']) @@ -67,7 +67,7 @@ class KafkaService(KafkaPathResolverMixin, JmxMixin, Service): def __init__(self, context, num_nodes, zk, security_protocol=SecurityConfig.PLAINTEXT, interbroker_security_protocol=SecurityConfig.PLAINTEXT, client_sasl_mechanism=SecurityConfig.SASL_MECHANISM_GSSAPI, interbroker_sasl_mechanism=SecurityConfig.SASL_MECHANISM_GSSAPI, - authorizer_class_name=None, topics=None, version=TRUNK, jmx_object_names=None, + authorizer_class_name=None, topics=None, version=DEV_BRANCH, jmx_object_names=None, jmx_attributes=None, zk_connect_timeout=5000, zk_session_timeout=6000, server_prop_overides=[]): """ :type context diff --git a/tests/kafkatest/services/performance/consumer_performance.py b/tests/kafkatest/services/performance/consumer_performance.py index 3e02a5ba2fbac..e1cd3a0e1b834 100644 --- a/tests/kafkatest/services/performance/consumer_performance.py +++ b/tests/kafkatest/services/performance/consumer_performance.py @@ -18,7 +18,7 @@ from kafkatest.services.performance import PerformanceService from kafkatest.services.security.security_config import SecurityConfig -from kafkatest.version import TRUNK, V_0_9_0_0, LATEST_0_10_0 +from kafkatest.version import DEV_BRANCH, V_0_9_0_0, LATEST_0_10_0 class ConsumerPerformanceService(PerformanceService): @@ -70,7 +70,7 @@ class ConsumerPerformanceService(PerformanceService): "collect_default": True} } - def __init__(self, context, num_nodes, kafka, topic, messages, version=TRUNK, new_consumer=True, settings={}): + def __init__(self, context, num_nodes, kafka, topic, messages, version=DEV_BRANCH, new_consumer=True, settings={}): super(ConsumerPerformanceService, self).__init__(context, num_nodes) self.kafka = kafka self.security_config = kafka.security_config.client_config() diff --git a/tests/kafkatest/services/performance/end_to_end_latency.py b/tests/kafkatest/services/performance/end_to_end_latency.py index 917ac85bc5dce..bdaf5502c751e 100644 --- a/tests/kafkatest/services/performance/end_to_end_latency.py +++ b/tests/kafkatest/services/performance/end_to_end_latency.py @@ -17,7 +17,7 @@ from kafkatest.services.performance import PerformanceService from kafkatest.services.security.security_config import SecurityConfig -from kafkatest.version import TRUNK, V_0_9_0_0 +from kafkatest.version import DEV_BRANCH, V_0_9_0_0 @@ -45,7 +45,7 @@ class EndToEndLatencyService(PerformanceService): "collect_default": True} } - def __init__(self, context, num_nodes, kafka, topic, num_records, compression_type="none", version=TRUNK, acks=1): + def __init__(self, context, num_nodes, kafka, topic, num_records, compression_type="none", version=DEV_BRANCH, acks=1): super(EndToEndLatencyService, self).__init__(context, num_nodes) self.kafka = kafka self.security_config = kafka.security_config.client_config() diff --git a/tests/kafkatest/services/performance/producer_performance.py b/tests/kafkatest/services/performance/producer_performance.py index 1113e0db65ca0..ff92da86bbf06 100644 --- a/tests/kafkatest/services/performance/producer_performance.py +++ b/tests/kafkatest/services/performance/producer_performance.py @@ -22,7 +22,7 @@ from kafkatest.services.monitor.jmx import JmxMixin from kafkatest.services.performance import PerformanceService from kafkatest.services.security.security_config import SecurityConfig -from kafkatest.version import TRUNK, V_0_9_0_0 +from kafkatest.version import DEV_BRANCH, V_0_9_0_0 class ProducerPerformanceService(JmxMixin, PerformanceService): @@ -34,7 +34,7 @@ class ProducerPerformanceService(JmxMixin, PerformanceService): LOG_FILE = os.path.join(LOG_DIR, "producer_performance.log") LOG4J_CONFIG = os.path.join(PERSISTENT_ROOT, "tools-log4j.properties") - def __init__(self, context, num_nodes, kafka, topic, num_records, record_size, throughput, version=TRUNK, settings=None, + def __init__(self, context, num_nodes, kafka, topic, num_records, record_size, throughput, version=DEV_BRANCH, settings=None, intermediate_stats=False, client_id="producer-performance", jmx_object_names=None, jmx_attributes=None): JmxMixin.__init__(self, num_nodes, jmx_object_names, jmx_attributes or []) @@ -89,11 +89,11 @@ def start_cmd(self, node): cmd = "" - if node.version < TRUNK: + if node.version < DEV_BRANCH: # In order to ensure more consistent configuration between versions, always use the ProducerPerformance - # tool from trunk - tools_jar = self.path.jar(TOOLS_JAR_NAME, TRUNK) - tools_dependant_libs_jar = self.path.jar(TOOLS_DEPENDANT_TEST_LIBS_JAR_NAME, TRUNK) + # tool from the development branch + tools_jar = self.path.jar(TOOLS_JAR_NAME, DEV_BRANCH) + tools_dependant_libs_jar = self.path.jar(TOOLS_DEPENDANT_TEST_LIBS_JAR_NAME, DEV_BRANCH) cmd += "for file in %s; do CLASSPATH=$CLASSPATH:$file; done; " % tools_jar cmd += "for file in %s; do CLASSPATH=$CLASSPATH:$file; done; " % tools_dependant_libs_jar diff --git a/tests/kafkatest/services/security/minikdc.py b/tests/kafkatest/services/security/minikdc.py index b3cbeaef53060..86a713882b010 100644 --- a/tests/kafkatest/services/security/minikdc.py +++ b/tests/kafkatest/services/security/minikdc.py @@ -24,7 +24,7 @@ from ducktape.services.service import Service from kafkatest.directory_layout.kafka_path import KafkaPathResolverMixin, CORE_LIBS_JAR_NAME, CORE_DEPENDANT_TEST_LIBS_JAR_NAME -from kafkatest.version import TRUNK +from kafkatest.version import DEV_BRANCH class MiniKdc(KafkaPathResolverMixin, Service): @@ -103,8 +103,8 @@ def start_node(self, node): principals = 'client ' + kafka_principals + ' ' + self.extra_principals self.logger.info("Starting MiniKdc with principals " + principals) - core_libs_jar = self.path.jar(CORE_LIBS_JAR_NAME, TRUNK) - core_dependant_test_libs_jar = self.path.jar(CORE_DEPENDANT_TEST_LIBS_JAR_NAME, TRUNK) + core_libs_jar = self.path.jar(CORE_LIBS_JAR_NAME, DEV_BRANCH) + core_dependant_test_libs_jar = self.path.jar(CORE_DEPENDANT_TEST_LIBS_JAR_NAME, DEV_BRANCH) cmd = "for file in %s; do CLASSPATH=$CLASSPATH:$file; done;" % core_libs_jar cmd += " for file in %s; do CLASSPATH=$CLASSPATH:$file; done;" % core_dependant_test_libs_jar diff --git a/tests/kafkatest/services/verifiable_consumer.py b/tests/kafkatest/services/verifiable_consumer.py index 711554c07ce99..d7fca876e6a85 100644 --- a/tests/kafkatest/services/verifiable_consumer.py +++ b/tests/kafkatest/services/verifiable_consumer.py @@ -22,7 +22,7 @@ from kafkatest.directory_layout.kafka_path import KafkaPathResolverMixin from kafkatest.services.kafka import TopicPartition -from kafkatest.version import TRUNK +from kafkatest.version import DEV_BRANCH class ConsumerState: @@ -136,7 +136,7 @@ class VerifiableConsumer(KafkaPathResolverMixin, BackgroundThreadService): def __init__(self, context, num_nodes, kafka, topic, group_id, max_messages=-1, session_timeout_sec=30, enable_autocommit=False, assignment_strategy="org.apache.kafka.clients.consumer.RangeAssignor", - version=TRUNK, stop_timeout_sec=30): + version=DEV_BRANCH, stop_timeout_sec=30): super(VerifiableConsumer, self).__init__(context, num_nodes) self.log_level = "TRACE" diff --git a/tests/kafkatest/services/verifiable_producer.py b/tests/kafkatest/services/verifiable_producer.py index d873e1f82b525..f545634aa6419 100644 --- a/tests/kafkatest/services/verifiable_producer.py +++ b/tests/kafkatest/services/verifiable_producer.py @@ -24,7 +24,7 @@ from kafkatest.directory_layout.kafka_path import KafkaPathResolverMixin, TOOLS_JAR_NAME, TOOLS_DEPENDANT_TEST_LIBS_JAR_NAME from kafkatest.utils import is_int, is_int_with_prefix -from kafkatest.version import TRUNK, LATEST_0_8_2 +from kafkatest.version import DEV_BRANCH, LATEST_0_8_2 from kafkatest.utils.remote_account import line_count @@ -50,7 +50,7 @@ class VerifiableProducer(KafkaPathResolverMixin, BackgroundThreadService): } def __init__(self, context, num_nodes, kafka, topic, max_messages=-1, throughput=100000, - message_validator=is_int, compression_types=None, version=TRUNK, acks=None, + message_validator=is_int, compression_types=None, version=DEV_BRANCH, acks=None, stop_timeout_sec=150): """ :param max_messages is a number of messages to be produced per producer @@ -181,9 +181,9 @@ def start_cmd(self, node, idx): cmd = "" if node.version <= LATEST_0_8_2: # 0.8.2.X releases do not have VerifiableProducer.java, so cheat and add - # the tools jar from trunk to the classpath - tools_jar = self.path.jar(TOOLS_JAR_NAME, TRUNK) - tools_dependant_libs_jar = self.path.jar(TOOLS_DEPENDANT_TEST_LIBS_JAR_NAME, TRUNK) + # the tools jar from the development branch to the classpath + tools_jar = self.path.jar(TOOLS_JAR_NAME, DEV_BRANCH) + tools_dependant_libs_jar = self.path.jar(TOOLS_DEPENDANT_TEST_LIBS_JAR_NAME, DEV_BRANCH) cmd += "for file in %s; do CLASSPATH=$CLASSPATH:$file; done; " % tools_jar cmd += "for file in %s; do CLASSPATH=$CLASSPATH:$file; done; " % tools_dependant_libs_jar diff --git a/tests/kafkatest/services/zookeeper.py b/tests/kafkatest/services/zookeeper.py index 8d38d48535143..571956b1a6a45 100644 --- a/tests/kafkatest/services/zookeeper.py +++ b/tests/kafkatest/services/zookeeper.py @@ -23,7 +23,7 @@ from kafkatest.directory_layout.kafka_path import KafkaPathResolverMixin from kafkatest.services.security.security_config import SecurityConfig -from kafkatest.version import TRUNK +from kafkatest.version import DEV_BRANCH class ZookeeperService(KafkaPathResolverMixin, Service): @@ -122,7 +122,7 @@ def query(self, path): """ Queries zookeeper for data associated with 'path' and returns all fields in the schema """ - kafka_run_class = self.path.script("kafka-run-class.sh", TRUNK) + kafka_run_class = self.path.script("kafka-run-class.sh", DEV_BRANCH) cmd = "%s kafka.tools.ZooKeeperMainWrapper -server %s get %s" % \ (kafka_run_class, self.connect_setting(), path) self.logger.debug(cmd) diff --git a/tests/kafkatest/tests/client/client_compatibility_features_test.py b/tests/kafkatest/tests/client/client_compatibility_features_test.py index 36d76b4b06726..d478590dcfd36 100644 --- a/tests/kafkatest/tests/client/client_compatibility_features_test.py +++ b/tests/kafkatest/tests/client/client_compatibility_features_test.py @@ -23,7 +23,7 @@ from kafkatest.services.zookeeper import ZookeeperService from kafkatest.services.kafka import KafkaService from ducktape.tests.test import Test -from kafkatest.version import TRUNK, LATEST_0_10_0, LATEST_0_10_1, V_0_10_1_0, KafkaVersion +from kafkatest.version import DEV_BRANCH, LATEST_0_10_0, LATEST_0_10_1, V_0_10_1_0, KafkaVersion def get_broker_features(broker_version): features = {} @@ -93,7 +93,7 @@ def invoke_compatibility_program(self, features): self.logger.info("** Command failed. See %s for log messages." % ssh_log_file) raise e - @parametrize(broker_version=str(TRUNK)) + @parametrize(broker_version=str(DEV_BRANCH)) @parametrize(broker_version=str(LATEST_0_10_0)) @parametrize(broker_version=str(LATEST_0_10_1)) def run_compatibility_test(self, broker_version): diff --git a/tests/kafkatest/tests/client/client_compatibility_produce_consume_test.py b/tests/kafkatest/tests/client/client_compatibility_produce_consume_test.py index 35b904b77ae01..c5e80cbec4d0a 100644 --- a/tests/kafkatest/tests/client/client_compatibility_produce_consume_test.py +++ b/tests/kafkatest/tests/client/client_compatibility_produce_consume_test.py @@ -22,7 +22,7 @@ from kafkatest.services.console_consumer import ConsoleConsumer from kafkatest.tests.produce_consume_validate import ProduceConsumeValidateTest from kafkatest.utils import is_int_with_prefix -from kafkatest.version import TRUNK, LATEST_0_10_0, LATEST_0_10_1, KafkaVersion +from kafkatest.version import DEV_BRANCH, LATEST_0_10_0, LATEST_0_10_1, KafkaVersion class ClientCompatibilityProduceConsumeTest(ProduceConsumeValidateTest): """ @@ -52,7 +52,7 @@ def min_cluster_size(self): # Override this since we're adding services outside of the constructor return super(ClientCompatibilityProduceConsumeTest, self).min_cluster_size() + self.num_producers + self.num_consumers - @parametrize(broker_version=str(TRUNK)) + @parametrize(broker_version=str(DEV_BRANCH)) @parametrize(broker_version=str(LATEST_0_10_0)) @parametrize(broker_version=str(LATEST_0_10_1)) def test_produce_consume(self, broker_version): diff --git a/tests/kafkatest/tests/client/message_format_change_test.py b/tests/kafkatest/tests/client/message_format_change_test.py index edcead2abf080..e58f58b06b156 100644 --- a/tests/kafkatest/tests/client/message_format_change_test.py +++ b/tests/kafkatest/tests/client/message_format_change_test.py @@ -22,7 +22,7 @@ from kafkatest.services.zookeeper import ZookeeperService from kafkatest.tests.produce_consume_validate import ProduceConsumeValidateTest from kafkatest.utils import is_int -from kafkatest.version import LATEST_0_9, LATEST_0_10, TRUNK, KafkaVersion +from kafkatest.version import LATEST_0_9, LATEST_0_10, DEV_BRANCH, KafkaVersion class MessageFormatChangeTest(ProduceConsumeValidateTest): @@ -58,7 +58,7 @@ def produce_and_consume(self, producer_version, consumer_version, group): err_msg="Producer did not produce all messages in reasonable amount of time")) @cluster(num_nodes=10) - @parametrize(producer_version=str(TRUNK), consumer_version=str(TRUNK)) + @parametrize(producer_version=str(DEV_BRANCH), consumer_version=str(DEV_BRANCH)) @parametrize(producer_version=str(LATEST_0_9), consumer_version=str(LATEST_0_9)) def test_compatibility(self, producer_version, consumer_version): """ This tests performs the following checks: @@ -70,7 +70,7 @@ def test_compatibility(self, producer_version, consumer_version): - The producers and consumers should not have any issue. - Note that for 0.9.x consumers/producers we only do steps 1 and 2 """ - self.kafka = KafkaService(self.test_context, num_nodes=3, zk=self.zk, version=TRUNK, topics={self.topic: { + self.kafka = KafkaService(self.test_context, num_nodes=3, zk=self.zk, version=DEV_BRANCH, topics={self.topic: { "partitions": 3, "replication-factor": 3, 'configs': {"min.insync.replicas": 2}}}) @@ -84,7 +84,7 @@ def test_compatibility(self, producer_version, consumer_version): self.kafka.alter_message_format(self.topic, str(LATEST_0_10)) self.produce_and_consume(producer_version, consumer_version, "group2") - if producer_version == str(TRUNK) and consumer_version == str(TRUNK): + if producer_version == str(DEV_BRANCH) and consumer_version == str(DEV_BRANCH): self.logger.info("Third format change back to 0.9.0") self.kafka.alter_message_format(self.topic, str(LATEST_0_9)) self.produce_and_consume(producer_version, consumer_version, "group3") diff --git a/tests/kafkatest/tests/core/compatibility_test_new_broker_test.py b/tests/kafkatest/tests/core/compatibility_test_new_broker_test.py index f3931ec9864f3..da6483d9023b3 100644 --- a/tests/kafkatest/tests/core/compatibility_test_new_broker_test.py +++ b/tests/kafkatest/tests/core/compatibility_test_new_broker_test.py @@ -23,7 +23,7 @@ from kafkatest.services.zookeeper import ZookeeperService from kafkatest.tests.produce_consume_validate import ProduceConsumeValidateTest from kafkatest.utils import is_int -from kafkatest.version import LATEST_0_10_0, LATEST_0_9, LATEST_0_8_2, TRUNK, KafkaVersion +from kafkatest.version import LATEST_0_10_1, LATEST_0_10_0, LATEST_0_9, LATEST_0_8_2, DEV_BRANCH, KafkaVersion # Compatibility tests for moving to a new broker (e.g., 0.10.x) and using a mix of old and new clients (e.g., 0.9.x) @@ -47,17 +47,18 @@ def setUp(self): @cluster(num_nodes=6) @parametrize(producer_version=str(LATEST_0_8_2), consumer_version=str(LATEST_0_8_2), compression_types=["none"], new_consumer=False, timestamp_type=None) @parametrize(producer_version=str(LATEST_0_8_2), consumer_version=str(LATEST_0_9), compression_types=["none"], new_consumer=False, timestamp_type=None) - @parametrize(producer_version=str(LATEST_0_9), consumer_version=str(TRUNK), compression_types=["none"], new_consumer=False, timestamp_type=None) - @parametrize(producer_version=str(TRUNK), consumer_version=str(LATEST_0_9), compression_types=["none"], new_consumer=False, timestamp_type=None) - @parametrize(producer_version=str(LATEST_0_9), consumer_version=str(TRUNK), compression_types=["snappy"], timestamp_type=None) - @parametrize(producer_version=str(TRUNK), consumer_version=str(LATEST_0_9), compression_types=["snappy"], timestamp_type=str("CreateTime")) - @parametrize(producer_version=str(TRUNK), consumer_version=str(TRUNK), compression_types=["snappy"], timestamp_type=str("LogAppendTime")) + @parametrize(producer_version=str(LATEST_0_9), consumer_version=str(DEV_BRANCH), compression_types=["none"], new_consumer=False, timestamp_type=None) + @parametrize(producer_version=str(DEV_BRANCH), consumer_version=str(LATEST_0_9), compression_types=["none"], new_consumer=False, timestamp_type=None) + @parametrize(producer_version=str(LATEST_0_9), consumer_version=str(DEV_BRANCH), compression_types=["snappy"], timestamp_type=None) + @parametrize(producer_version=str(DEV_BRANCH), consumer_version=str(LATEST_0_9), compression_types=["snappy"], timestamp_type=str("CreateTime")) + @parametrize(producer_version=str(DEV_BRANCH), consumer_version=str(DEV_BRANCH), compression_types=["snappy"], timestamp_type=str("LogAppendTime")) + @parametrize(producer_version=str(LATEST_0_10_1), consumer_version=str(LATEST_0_10_1), compression_types=["snappy"], timestamp_type=str("LogAppendTime")) @parametrize(producer_version=str(LATEST_0_10_0), consumer_version=str(LATEST_0_10_0), compression_types=["snappy"], timestamp_type=str("LogAppendTime")) @parametrize(producer_version=str(LATEST_0_9), consumer_version=str(LATEST_0_9), compression_types=["snappy"], timestamp_type=str("LogAppendTime")) - @parametrize(producer_version=str(TRUNK), consumer_version=str(TRUNK), compression_types=["none"], new_consumer=False, timestamp_type=str("LogAppendTime")) + @parametrize(producer_version=str(DEV_BRANCH), consumer_version=str(DEV_BRANCH), compression_types=["none"], new_consumer=False, timestamp_type=str("LogAppendTime")) def test_compatibility(self, producer_version, consumer_version, compression_types, new_consumer=True, timestamp_type=None): - self.kafka = KafkaService(self.test_context, num_nodes=3, zk=self.zk, version=TRUNK, topics={self.topic: { + self.kafka = KafkaService(self.test_context, num_nodes=3, zk=self.zk, version=DEV_BRANCH, topics={self.topic: { "partitions": 3, "replication-factor": 3, 'configs': {"min.insync.replicas": 2}}}) diff --git a/tests/kafkatest/tests/core/upgrade_test.py b/tests/kafkatest/tests/core/upgrade_test.py index 34af4eb56c7e1..8d89b7c13c36d 100644 --- a/tests/kafkatest/tests/core/upgrade_test.py +++ b/tests/kafkatest/tests/core/upgrade_test.py @@ -25,7 +25,7 @@ from kafkatest.services.zookeeper import ZookeeperService from kafkatest.tests.produce_consume_validate import ProduceConsumeValidateTest from kafkatest.utils import is_int -from kafkatest.version import LATEST_0_8_2, LATEST_0_9, LATEST_0_10_0, TRUNK, KafkaVersion +from kafkatest.version import LATEST_0_8_2, LATEST_0_9, LATEST_0_10_0, LATEST_0_10_1, DEV_BRANCH, KafkaVersion class TestUpgrade(ProduceConsumeValidateTest): @@ -46,7 +46,7 @@ def perform_upgrade(self, from_kafka_version, to_message_format_version=None): self.logger.info("First pass bounce - rolling upgrade") for node in self.kafka.nodes: self.kafka.stop_node(node) - node.version = TRUNK + node.version = DEV_BRANCH node.config[config_property.INTER_BROKER_PROTOCOL_VERSION] = from_kafka_version node.config[config_property.MESSAGE_FORMAT_VERSION] = from_kafka_version self.kafka.start_node(node) @@ -62,6 +62,8 @@ def perform_upgrade(self, from_kafka_version, to_message_format_version=None): self.kafka.start_node(node) @cluster(num_nodes=6) + @parametrize(from_kafka_version=str(LATEST_0_10_1), to_message_format_version=None, compression_types=["lz4"]) + @parametrize(from_kafka_version=str(LATEST_0_10_1), to_message_format_version=None, compression_types=["snappy"], new_consumer=False) @parametrize(from_kafka_version=str(LATEST_0_10_0), to_message_format_version=None, compression_types=["snappy"], new_consumer=False) @parametrize(from_kafka_version=str(LATEST_0_10_0), to_message_format_version=None, compression_types=["snappy"]) @parametrize(from_kafka_version=str(LATEST_0_9), to_message_format_version=None, compression_types=["none"], new_consumer=False) @@ -80,7 +82,7 @@ def perform_upgrade(self, from_kafka_version, to_message_format_version=None): @parametrize(from_kafka_version=str(LATEST_0_8_2), to_message_format_version=None, compression_types=["snappy"], new_consumer=False) def test_upgrade(self, from_kafka_version, to_message_format_version, compression_types, new_consumer=True, security_protocol="PLAINTEXT"): - """Test upgrade of Kafka broker cluster from 0.8.2, 0.9.0 or 0.10.0 to the current version + """Test upgrade of Kafka broker cluster from 0.8.2, 0.9.0, 0.10.0, 0.10.1 to the current version from_kafka_version is a Kafka version to upgrade from: either 0.8.2.X, 0.9.0.x or 0.10.0.x @@ -112,7 +114,8 @@ def test_upgrade(self, from_kafka_version, to_message_format_version, compressio compression_types=compression_types, version=KafkaVersion(from_kafka_version)) - assert self.zk.query("/cluster/id") is None + if from_kafka_version <= LATEST_0_10_0: + assert self.zk.query("/cluster/id") is None # TODO - reduce the timeout self.consumer = ConsoleConsumer(self.test_context, self.num_consumers, self.kafka, diff --git a/tests/kafkatest/version.py b/tests/kafkatest/version.py index cd361e836b49d..7cd489d87ac64 100644 --- a/tests/kafkatest/version.py +++ b/tests/kafkatest/version.py @@ -31,11 +31,8 @@ class KafkaVersion(LooseVersion): assert v10 > v9 # assertion passes! """ def __init__(self, version_string): - self.is_trunk = (version_string.lower() == "trunk") - if self.is_trunk: - # Since "trunk" may actually be a branch that is not trunk, - # use kafkatest_version() for comparison purposes, - # and track whether we're in "trunk" with a flag + self.is_dev = (version_string.lower() == "dev") + if self.is_dev: version_string = kafkatest_version() # Drop dev suffix if present @@ -48,22 +45,22 @@ def __init__(self, version_string): LooseVersion.__init__(self, version_string) def __str__(self): - if self.is_trunk: - return "trunk" + if self.is_dev: + return "dev" else: return LooseVersion.__str__(self) def get_version(node=None): """Return the version attached to the given node. - Default to trunk if node or node.version is undefined (aka None) + Default to DEV_BRANCH if node or node.version is undefined (aka None) """ if node is not None and hasattr(node, "version") and node.version is not None: return node.version else: - return TRUNK + return DEV_BRANCH -TRUNK = KafkaVersion("trunk") +DEV_BRANCH = KafkaVersion("dev") # 0.8.2.X versions V_0_8_2_1 = KafkaVersion("0.8.2.1") diff --git a/tests/unit/directory_layout/check_project_paths.py b/tests/unit/directory_layout/check_project_paths.py index 284e4d5150c41..b9b76f13276be 100644 --- a/tests/unit/directory_layout/check_project_paths.py +++ b/tests/unit/directory_layout/check_project_paths.py @@ -16,7 +16,7 @@ from kafkatest.directory_layout.kafka_path import create_path_resolver, KafkaSystemTestPathResolver, \ KAFKA_PATH_RESOLVER_KEY -from kafkatest.version import V_0_9_0_1, TRUNK, KafkaVersion +from kafkatest.version import V_0_9_0_1, DEV_BRANCH, KafkaVersion class DummyContext(object): @@ -56,9 +56,9 @@ def check_paths(self): """Check expected path resolution without any version specified.""" resolver = create_path_resolver(DummyContext()) - assert resolver.home() == "/opt/kafka-trunk" - assert resolver.bin() == "/opt/kafka-trunk/bin" - assert resolver.script("kafka-run-class.sh") == "/opt/kafka-trunk/bin/kafka-run-class.sh" + assert resolver.home() == "/opt/kafka-dev" + assert resolver.bin() == "/opt/kafka-dev/bin" + assert resolver.script("kafka-run-class.sh") == "/opt/kafka-dev/bin/kafka-run-class.sh" def check_versioned_source_paths(self): """Check expected paths when using versions.""" @@ -74,16 +74,16 @@ def check_node_or_version_helper(self): """ resolver = create_path_resolver(DummyContext()) - # Node with no version attribute should resolve to TRUNK + # Node with no version attribute should resolve to DEV_BRANCH node = DummyNode() - assert resolver._version(node) == TRUNK + assert resolver._version(node) == DEV_BRANCH # Node with version attribute should resolve to the version attribute node.version = V_0_9_0_1 assert resolver._version(node) == V_0_9_0_1 # A KafkaVersion object should resolve to itself - assert resolver._version(TRUNK) == TRUNK + assert resolver._version(DEV_BRANCH) == DEV_BRANCH version = KafkaVersion("999.999.999") assert resolver._version(version) == version diff --git a/tests/unit/version/check_version.py b/tests/unit/version/check_version.py index 7b855518207da..8cf8e9a06e7b9 100644 --- a/tests/unit/version/check_version.py +++ b/tests/unit/version/check_version.py @@ -15,18 +15,18 @@ from mock import Mock -from kafkatest.version import TRUNK, V_0_8_2_2, get_version +from kafkatest.version import DEV_BRANCH, V_0_8_2_2, get_version class CheckVersion(object): def check_get_version(self): """Check default and override behavior of get_version""" node = None - assert get_version(node) == TRUNK + assert get_version(node) == DEV_BRANCH node = Mock() node.version = None - assert get_version(node) == TRUNK + assert get_version(node) == DEV_BRANCH node = Mock() node.version = V_0_8_2_2 diff --git a/tools/src/main/java/org/apache/kafka/tools/VerifiableLog4jAppender.java b/tools/src/main/java/org/apache/kafka/tools/VerifiableLog4jAppender.java index daf569cd77173..faa9698e5a48d 100644 --- a/tools/src/main/java/org/apache/kafka/tools/VerifiableLog4jAppender.java +++ b/tools/src/main/java/org/apache/kafka/tools/VerifiableLog4jAppender.java @@ -158,7 +158,7 @@ private static ArgumentParser argParser() { * * Note: this duplication of org.apache.kafka.common.utils.Utils.loadProps is unfortunate * but *intentional*. In order to use VerifiableProducer in compatibility and upgrade tests, - * we use VerifiableProducer from trunk tools package, and run it against 0.8.X.X kafka jars. + * we use VerifiableProducer from the development tools package, and run it against 0.8.X.X kafka jars. * Since this method is not in Utils in the 0.8.X.X jars, we have to cheat a bit and duplicate. */ public static Properties loadProps(String filename) throws IOException, FileNotFoundException { diff --git a/tools/src/main/java/org/apache/kafka/tools/VerifiableProducer.java b/tools/src/main/java/org/apache/kafka/tools/VerifiableProducer.java index 54d53f10bea13..e81eb8f3ae01b 100644 --- a/tools/src/main/java/org/apache/kafka/tools/VerifiableProducer.java +++ b/tools/src/main/java/org/apache/kafka/tools/VerifiableProducer.java @@ -160,7 +160,7 @@ private static ArgumentParser argParser() { * * Note: this duplication of org.apache.kafka.common.utils.Utils.loadProps is unfortunate * but *intentional*. In order to use VerifiableProducer in compatibility and upgrade tests, - * we use VerifiableProducer from trunk tools package, and run it against 0.8.X.X kafka jars. + * we use VerifiableProducer from the development tools package, and run it against 0.8.X.X kafka jars. * Since this method is not in Utils in the 0.8.X.X jars, we have to cheat a bit and duplicate. */ public static Properties loadProps(String filename) throws IOException, FileNotFoundException { diff --git a/vagrant/base.sh b/vagrant/base.sh index 5f07cc3658b36..448c640d2bfc9 100755 --- a/vagrant/base.sh +++ b/vagrant/base.sh @@ -37,11 +37,11 @@ if [ -z `which javac` ]; then fi chmod a+rw /opt -if [ -h /opt/kafka-trunk ]; then +if [ -h /opt/kafka-dev ]; then # reset symlink - rm /opt/kafka-trunk + rm /opt/kafka-dev fi -ln -s /vagrant /opt/kafka-trunk +ln -s /vagrant /opt/kafka-dev get_kafka() { version=$1 diff --git a/vagrant/broker.sh b/vagrant/broker.sh index 06dee6b97abec..986f0fae43650 100755 --- a/vagrant/broker.sh +++ b/vagrant/broker.sh @@ -23,7 +23,7 @@ PUBLIC_ADDRESS=$2 PUBLIC_ZOOKEEPER_ADDRESSES=$3 JMX_PORT=$4 -kafka_dir=/opt/kafka-trunk +kafka_dir=/opt/kafka-dev cd $kafka_dir sed \ diff --git a/vagrant/zk.sh b/vagrant/zk.sh index 67924d129d98e..e8c690a060e32 100755 --- a/vagrant/zk.sh +++ b/vagrant/zk.sh @@ -22,7 +22,7 @@ ZKID=$1 NUM_ZK=$2 JMX_PORT=$3 -kafka_dir=/opt/kafka-trunk +kafka_dir=/opt/kafka-dev cd $kafka_dir cp $kafka_dir/config/zookeeper.properties $kafka_dir/config/zookeeper-$ZKID.properties