continuous integration - Computer Definition
A software development system that ensures code changes are compiled and tested routinely. Continuous integration (CI) is used for large applications developed by several programmers, each writing a part of the program. After they make changes to their modules and check in the new code, the CI system typically compiles and tests the application immediately. See continuous delivery.