diff options
author | David Arthur <mumrah@gmail.com> | 2013-02-22 23:09:25 -0500 |
---|---|---|
committer | David Arthur <mumrah@gmail.com> | 2013-04-02 20:19:30 -0400 |
commit | 2a3d231aa61642c57537bc2128dd4f2bd30f35dd (patch) | |
tree | 6bfdfa13b228481df9c79bcb926c2036b476b891 /test/resources | |
parent | e87c561723be25fcfa2564030367196231aa366e (diff) | |
download | kafka-python-2a3d231aa61642c57537bc2128dd4f2bd30f35dd.tar.gz |
Protocol and low-level client done, adding tests
Diffstat (limited to 'test/resources')
-rw-r--r-- | test/resources/log4j.properties | 2 | ||||
-rw-r--r-- | test/resources/server.properties | 63 |
2 files changed, 33 insertions, 32 deletions
diff --git a/test/resources/log4j.properties b/test/resources/log4j.properties index c4ecd2c..47a817a 100644 --- a/test/resources/log4j.properties +++ b/test/resources/log4j.properties @@ -12,7 +12,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -log4j.rootLogger=DEBUG, stdout +log4j.rootLogger=TRACE, stdout log4j.appender.stdout=org.apache.log4j.ConsoleAppender log4j.appender.stdout.layout=org.apache.log4j.PatternLayout diff --git a/test/resources/server.properties b/test/resources/server.properties index 2eefe3b..0d01fca 100644 --- a/test/resources/server.properties +++ b/test/resources/server.properties @@ -17,31 +17,32 @@ ############################# Server Basics ############################# # The id of the broker. This must be set to a unique integer for each broker. -brokerid=0 - -# Hostname the broker will advertise to consumers. If not set, kafka will use the value returned -# from InetAddress.getLocalHost(). If there are multiple interfaces getLocalHost -# may not be what you want. -#hostname= - +broker.id=0 ############################# Socket Server Settings ############################# # The port the socket server listens on port=%(kafka.port)d -# The number of processor threads the socket server uses for receiving and answering requests. -# Defaults to the number of cores on the machine -num.threads=2 +# Hostname the broker will bind to and advertise to producers and consumers. +# If not set, the server will bind to all interfaces and advertise the value returned from +# from java.net.InetAddress.getCanonicalHostName(). +#host.name=localhost + +# The number of threads handling network requests +num.network.threads=2 + +# The number of threads doing disk I/O +num.io.threads=2 # The send buffer (SO_SNDBUF) used by the socket server -socket.send.buffer=1048576 +socket.send.buffer.bytes=1048576 # The receive buffer (SO_RCVBUF) used by the socket server -socket.receive.buffer=1048576 +socket.receive.buffer.bytes=1048576 # The maximum size of a request that the socket server will accept (protection against OOM) -max.socket.request.bytes=104857600 +socket.request.max.bytes=104857600 ############################# Log Basics ############################# @@ -53,9 +54,6 @@ log.dir=%(kafka.tmp.dir)s # for consumption, but also mean more files. num.partitions=%(kafka.partitions)d -# Overrides for for the default given by num.partitions on a per-topic basis -#topic.partition.count.map=topic1:3, topic2:4 - ############################# Log Flush Policy ############################# # The following configurations control the flush of data to disk. This is the most @@ -68,16 +66,13 @@ num.partitions=%(kafka.partitions)d # every N messages (or both). This can be done globally and overridden on a per-topic basis. # The number of messages to accept before forcing a flush of data to disk -log.flush.interval=1 +log.flush.interval.messages=10000 # The maximum amount of time a message can sit in a log before we force a flush -log.default.flush.interval.ms=10000 +log.flush.interval.ms=1000 -# Per-topic overrides for log.default.flush.interval.ms -#topic.flush.intervals.ms=topic1:1000, topic2:3000 - -# The interval (in ms) at which logs are checked to see if they need to be flushed to disk. -log.default.flush.scheduler.interval.ms=10000 +# Per-topic overrides for log.flush.interval.ms +#log.flush.intervals.ms.per.topic=topic1:1000, topic2:3000 ############################# Log Retention Policy ############################# @@ -90,11 +85,11 @@ log.default.flush.scheduler.interval.ms=10000 log.retention.hours=168 # A size-based retention policy for logs. Segments are pruned from the log as long as the remaining -# segments don't drop below log.retention.size. -#log.retention.size=1073741824 +# segments don't drop below log.retention.bytes. +#log.retention.bytes=1073741824 # The maximum size of a log segment file. When this size is reached a new log segment will be created. -log.file.size=536870912 +log.segment.bytes=536870912 # The interval at which log segments are checked to see if they can be deleted according # to the retention policies @@ -102,15 +97,21 @@ log.cleanup.interval.mins=1 ############################# Zookeeper ############################# -# Enable connecting to zookeeper -enable.zookeeper=false - # Zk connection string (see zk docs for details). # This is a comma separated host:port pairs, each corresponding to a zk # server. e.g. "127.0.0.1:3000,127.0.0.1:3001,127.0.0.1:3002". # You can also append an optional chroot string to the urls to specify the # root directory for all kafka znodes. -zk.connect=localhost:2181 +zk.connect=localhost:2181/kafka-python # Timeout in ms for connecting to zookeeper -zk.connectiontimeout.ms=1000000 +zk.connection.timeout.ms=1000000 + +# metrics reporter properties +kafka.metrics.polling.interval.secs=5 +kafka.metrics.reporters=kafka.metrics.KafkaCSVMetricsReporter +kafka.csv.metrics.dir=/tmp/kafka_metrics +# Disable csv reporting by default. +kafka.csv.metrics.reporter.enabled=false + +log.cleanup.policy=delete |