summaryrefslogtreecommitdiff
path: root/test/resources
diff options
context:
space:
mode:
authorDavid Arthur <mumrah@gmail.com>2013-02-22 23:09:25 -0500
committerDavid Arthur <mumrah@gmail.com>2013-04-02 20:19:30 -0400
commit2a3d231aa61642c57537bc2128dd4f2bd30f35dd (patch)
tree6bfdfa13b228481df9c79bcb926c2036b476b891 /test/resources
parente87c561723be25fcfa2564030367196231aa366e (diff)
downloadkafka-python-2a3d231aa61642c57537bc2128dd4f2bd30f35dd.tar.gz
Protocol and low-level client done, adding tests
Diffstat (limited to 'test/resources')
-rw-r--r--test/resources/log4j.properties2
-rw-r--r--test/resources/server.properties63
2 files changed, 33 insertions, 32 deletions
diff --git a/test/resources/log4j.properties b/test/resources/log4j.properties
index c4ecd2c..47a817a 100644
--- a/test/resources/log4j.properties
+++ b/test/resources/log4j.properties
@@ -12,7 +12,7 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
-log4j.rootLogger=DEBUG, stdout
+log4j.rootLogger=TRACE, stdout
log4j.appender.stdout=org.apache.log4j.ConsoleAppender
log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
diff --git a/test/resources/server.properties b/test/resources/server.properties
index 2eefe3b..0d01fca 100644
--- a/test/resources/server.properties
+++ b/test/resources/server.properties
@@ -17,31 +17,32 @@
############################# Server Basics #############################
# The id of the broker. This must be set to a unique integer for each broker.
-brokerid=0
-
-# Hostname the broker will advertise to consumers. If not set, kafka will use the value returned
-# from InetAddress.getLocalHost(). If there are multiple interfaces getLocalHost
-# may not be what you want.
-#hostname=
-
+broker.id=0
############################# Socket Server Settings #############################
# The port the socket server listens on
port=%(kafka.port)d
-# The number of processor threads the socket server uses for receiving and answering requests.
-# Defaults to the number of cores on the machine
-num.threads=2
+# Hostname the broker will bind to and advertise to producers and consumers.
+# If not set, the server will bind to all interfaces and advertise the value returned from
+# from java.net.InetAddress.getCanonicalHostName().
+#host.name=localhost
+
+# The number of threads handling network requests
+num.network.threads=2
+
+# The number of threads doing disk I/O
+num.io.threads=2
# The send buffer (SO_SNDBUF) used by the socket server
-socket.send.buffer=1048576
+socket.send.buffer.bytes=1048576
# The receive buffer (SO_RCVBUF) used by the socket server
-socket.receive.buffer=1048576
+socket.receive.buffer.bytes=1048576
# The maximum size of a request that the socket server will accept (protection against OOM)
-max.socket.request.bytes=104857600
+socket.request.max.bytes=104857600
############################# Log Basics #############################
@@ -53,9 +54,6 @@ log.dir=%(kafka.tmp.dir)s
# for consumption, but also mean more files.
num.partitions=%(kafka.partitions)d
-# Overrides for for the default given by num.partitions on a per-topic basis
-#topic.partition.count.map=topic1:3, topic2:4
-
############################# Log Flush Policy #############################
# The following configurations control the flush of data to disk. This is the most
@@ -68,16 +66,13 @@ num.partitions=%(kafka.partitions)d
# every N messages (or both). This can be done globally and overridden on a per-topic basis.
# The number of messages to accept before forcing a flush of data to disk
-log.flush.interval=1
+log.flush.interval.messages=10000
# The maximum amount of time a message can sit in a log before we force a flush
-log.default.flush.interval.ms=10000
+log.flush.interval.ms=1000
-# Per-topic overrides for log.default.flush.interval.ms
-#topic.flush.intervals.ms=topic1:1000, topic2:3000
-
-# The interval (in ms) at which logs are checked to see if they need to be flushed to disk.
-log.default.flush.scheduler.interval.ms=10000
+# Per-topic overrides for log.flush.interval.ms
+#log.flush.intervals.ms.per.topic=topic1:1000, topic2:3000
############################# Log Retention Policy #############################
@@ -90,11 +85,11 @@ log.default.flush.scheduler.interval.ms=10000
log.retention.hours=168
# A size-based retention policy for logs. Segments are pruned from the log as long as the remaining
-# segments don't drop below log.retention.size.
-#log.retention.size=1073741824
+# segments don't drop below log.retention.bytes.
+#log.retention.bytes=1073741824
# The maximum size of a log segment file. When this size is reached a new log segment will be created.
-log.file.size=536870912
+log.segment.bytes=536870912
# The interval at which log segments are checked to see if they can be deleted according
# to the retention policies
@@ -102,15 +97,21 @@ log.cleanup.interval.mins=1
############################# Zookeeper #############################
-# Enable connecting to zookeeper
-enable.zookeeper=false
-
# Zk connection string (see zk docs for details).
# This is a comma separated host:port pairs, each corresponding to a zk
# server. e.g. "127.0.0.1:3000,127.0.0.1:3001,127.0.0.1:3002".
# You can also append an optional chroot string to the urls to specify the
# root directory for all kafka znodes.
-zk.connect=localhost:2181
+zk.connect=localhost:2181/kafka-python
# Timeout in ms for connecting to zookeeper
-zk.connectiontimeout.ms=1000000
+zk.connection.timeout.ms=1000000
+
+# metrics reporter properties
+kafka.metrics.polling.interval.secs=5
+kafka.metrics.reporters=kafka.metrics.KafkaCSVMetricsReporter
+kafka.csv.metrics.dir=/tmp/kafka_metrics
+# Disable csv reporting by default.
+kafka.csv.metrics.reporter.enabled=false
+
+log.cleanup.policy=delete