Coding Guidelines
These guidelines are meant to encourage consistency and best practices amongst people working on the Kafka code base. They should be observed unless there is a compelling reason to ignore them.
Basic Stuff
- Avoid cryptic abbreviations. Single letter variable names are fine in very short methods with few variables, otherwise make them informative.
- Clear code is preferable to comments. When possible make your naming so good you don't need comments. When that isn't possible comments should be thought of as mandatory, write them to be read.
- Logging, configuration, and public APIs are our "UI". Make them pretty, consistent, and usable.
- There is not a maximum line length (certainly not 80 characters, we don't work on punch cards any more), but be reasonable.
- Don't be sloppy. Don't check in commented out code: we use version control, it is still there in the history. Don't leave TODOs in the code or FIXMEs if you can help it. Don't leave println statements in the code. Hopefully this is all obvious.
- We want people to use our stuff, which means we need clear, correct documentation. User documentation should be considered a part of any user-facing the feature, just like unit tests or performance results.
- Don't duplicate code (duh).
- Kafka is system software, and certain things are appropriate in system software that are not appropriate elsewhere. Sockets, bytes, concurrency, and distribution are our core competency which means we will have a more "from scratch" implementation of some of these things then would be appropriate for software elsewhere in the stack. This is because we need to be exceptionally good at these things. This does not excuse fiddly low-level code, but it does excuse spending a little extra time to make sure that our filesystem structures, networking code, threading model, are all done perfectly right for our application rather than just trying to glue together ill-fitting off-the-shelf pieces (well-fitting off-the-shelf pieces are great though).
Scala
We are following the style guide given here (though not perfectly). Below are some specifics worth noting:
- Scala is a very flexible language. Use restraint. Magic cryptic one-liners do not impress us, readability impresses us.
- Use
val
s when possible.
- Use private when possible for member variables.
- Method and member variable names should be in camel case with an initial lower case character like
aMethodName
.
- Constants should be camel case with an initial capital
LikeThis
not LIKE_THIS
.
- Prefer a single top-level class per file for ease of finding things.
- Do not use semi-colons unless required.
- Avoid getters and setters - stick to plain
val
s or var
s instead. If (later on) you require a custom setter (or getter) for a var
named myVar
then add a shadow var myVar_underlying
and override the setter (def myVar_=
) and the getter (def myVar = myVar_underlying
).
- Perfer Option to null in scala APIs.
- Use named arguments when passing in literal values if the meaning is at all unclear, for example instead of
Utils.delete(true)
prefer Utils.delete(recursive=true)
.
- Indentation is 2 spaces and never tabs. One could argue the right amount of indentation, but 2 seems to be standard for scala and consistency is best here since there is clearly no "right" way.
- Include the optional paranthesis on a no-arg method only if the method has a side-effect, otherwise omit them. For example
fileChannel.force()
and fileChannel.size
. This helps emphasize that you are calling the method for the side effect, which is changing some state, not just getting the return value.
- Prefer case classes to tuples in important APIs to make it clear what the intended contents are.
Logging
- Logging is one third of our "UI" and it should be taken seriously. Please take the time to assess the logs when making a change to ensure that the important things are getting logged and there is no junk there.
- Logging statements should be complete sentences with proper capitalization that are written to be read by a person not necessarily familiar with the source code. It is fine to put in hacky little logging statements when debugging, but either clean them up or remove them before checking in. So logging something like "INFO: entering SyncProducer send()" is not appropriate.
- Logging should not mention class names or internal variables.
- There are six levels of logging
TRACE
, DEBUG
, INFO
, WARN
, ERROR
, and FATAL
, they should be used as follows.
INFO
is the level you should assume the software will be run in. INFO messages are things which are not bad but which the user will definitely want to know about every time they occur.
TRACE
and DEBUG
are both things you turn on when something is wrong and you want to figure out what is going on. DEBUG
should not be so fine grained that it will seriously effect the performance of the server. TRACE
can be anything. Both DEBUG
and TRACE
statements should be wrapped in an if(logger.isDebugEnabled)
check to avoid pasting together big strings all the time.
WARN
and ERROR
indicate something that is bad. Use WARN
if you aren't totally sure it is bad, and ERROR
if you are.
- Use
FATAL
only right before calling System.exit()
.
Monitoring
- Monitoring is the second third of our "UI" and it should also be taken seriously.
- We use JMX for monitoring.
- Any new features should come with appropriate monitoring to know the feature is working correctly. This is at least as important as unit tests as it verifies production.
Unit Tests
- New patches should come with unit tests that verify the functionality being added.
- Unit tests are first rate code, and should be treated like it. They should not contain code duplication, cryptic hackery, or anything like that.
- Be aware of the methods in
kafka.utils.TestUtils
, they make a lot of the below things easier to get right.
- Unit tests should test the least amount of code possible, don't start the whole server unless there is no other way to test a single class or small group of classes in isolation.
- Tests should not depend on any external resources, they need to set up and tear down their own stuff. This means if you want zookeeper it needs to be started and stopped, you can't depend on it already being there. Likewise if you need a file with some data in it, you need to write it in the beginning of the test and delete it (pass or fail).
- It is okay to use the filesystem and network in tests since that is our business but you need to clean up after yourself. There are helpers for this in
TestUtils
.
- Do not use sleep or other timing assumptions in tests, it is always, always, always wrong and will fail intermittently on any test server with other things going on that causes delays. Write tests in such a way that they are not timing dependent. Seriously. One thing that will help this is to never directly use the system clock in code (i.e.
System.currentTimeMillis
) but instead to use the kafka.utils.Time
. This is a trait that has a mock implementation allowing you to programmatically and deterministically cause the passage of time when you inject this mock clock instead of the system clock.
- It must be possible to run the tests in parallel, without having them collide. This is a practical thing to allow multiple branches to CI on a single CI server. This means you can't hard code directories or ports or things like that in tests because two instances will step on each other. Again
TestUtils
has helpers for this stuff (e.g. TestUtils.choosePort
will find a free port for you).
Configuration
- Configuration is the final third of our "UI".
- Names should be thought through from the point of view of the person using the config, but often programmers choose configuration names that make sense for someone reading the code.
- Often the value that makes most sense in configuration is not the one most useful to program with. For example, let's say you want to throttle I/O to avoid using up all the I/O bandwidth. The easiest thing to implement is to give a "sleep time" configuration that let's the program sleep after doing I/O to throttle down its rate. But notice how hard it is to correctly use this configuration parameter, the user has to figure out the rate of I/O on the machine, and then do a bunch of arithmetic to calculate the right sleep time to give the desired rate of I/O on the system. It is much, much, much better to just have the user configure the maximum I/O rate they want to allow (say 5MB/sec) and then calculate the appropriate sleep time from that and the actual I/O rate. Another way to say this is that configuration should always be in terms of the quantity that the user knows, not the quantity you want to use.
- Configuration is the answer to problems we can't solve up front for some reason--if there is a way to just choose a best value do that instead.
- Configuration should come from the instance-level properties file. No additional sources of config (environment variables, system properties, etc) should be added as these usually inhibit running multiple instances of a broker on one machine.
Concurrency
- Encapsulate synchronization. That is, locks should be private member variables within a class and only one class or method should need to be examined to verify the correctness of the synchronization strategy.
- Annotate things as
@threadsafe
when they are supposed to be and @notthreadsafe
when they aren't to help track this stuff.
- There are a number of gotchas with threads and threadpools: is the daemon flag set appropriately for your threads? are your threads being named in a way that will distinguish their purpose in a thread dump? What happens when the number of queued tasks hits the limit (do you drop stuff? do you block?).
- Prefer the java.util.concurrent packages to either low-level wait-notify, custom locking/synchronization, or higher level scala-specific primitives. The util.concurrent stuff is well thought out and actually works correctly. There is a generally feeling that threads and locking are not going to be the concurrency primitives of the future because of a variety of well-known weaknesses they have. This is probably true, but they have the advantage of actually being mature enough to use for high-performance software right now; their well-known deficiencies are easily worked around by equally well known best-practices. So avoid actors, software transactional memory, tuple spaces, or anything else not written by Doug Lea and used by at least a million other productions systems. :-)
Backwards Compatibility
- Our policy is that the Kafka protocols and data formats should support backwards compatibility for one release to enable no-downtime upgrades (unless there is a very compelling counter-argument). This means the server MUST be able to support requests from both old and new clients simultaneously. This compatability need only be retained for one release (e.g. 0.7 must accept requests from 0.6 clients, but this need not be maintained indefinitely). The code handling the "old" path can be removed in the next release or whenever we can get all the clients up-to-date. A typical upgrade sequence for binary format changes would be (1) upgrade consumers to handle new message format, (2) upgrade server, (3) clients.
- There are three things which require this binary compatibility: request objects, persistent data structure (messages and message sets), and zookeeper structures and protocols. The message binary structure has a "magic" byte to allow it to be evolved, this number should be incremented when the format is changed and the number can be checked to apply the right logic and fill in defaults appropriately. Network requests have a request id which serve a similar purpose, any change to a request object must be accompanied by a change in the request id. Any change here should be accompanied by compatibility tests that save requests or messages in the old format as a binary file which is tested for compatibility with the new code.
Client Code
There are a few things that need to be considered in client code that are not a major concern on the server side.
- Libraries needed by the client should be avoided whenever possible. Clients are run in someone else's code and it is very possible that they may have the same library we have, but a different and incompatible version. This will mean they can't use our client. For this reason the client should not use any libraries that are not strictly necessary.
- We should attempt to maintain API compatibility when possible, though at this point in the project's lifecycle it is more important to make things good rather avoid breakage.