This file contains documentation about the build process and unittests - put short - if you want to hack parts of arangod this could be interesting for you. Documentation readme can be found in directory Documentation.
Setting up git for automatically merging certain automatically generated files in the arangodb source tree:
git config --global merge.ours.driver true
Essentially, you can compile ArangoDB from source by issueing the following commands from a clone of the source repository:
mkdir build
cd build
cmake ..
make
cd ..
After that, the binaries will reside in build/bin
. To quickly start
up your compiled ArangoDB, simply do
build/bin/arangod -c etc/relative/arangod.conf data
This will use a configuration file that is included in the source repository.
- -DUSE_MAINTAINER_MODE=1 - generate lex/yacc and errors files
- -DUSE_BACKTRACE=1 - add backtraces to native code asserts & exceptions
- -DUSE_FAILURE_TESTS=1 - adds javascript hook to crash the server for data integrity tests
- *-DUSE_CATCH_TESTS=On (default is On so this is set unless you explicitly disable it)
If one changes any error in the ArangoDB system, then one has to:
- Only touch
lib/Basics/errors.dat
and not the files which are automatically generated from it (lib/Basics/voc-errors.h
,lib/Basics/voc-errors.cpp
andjs/common/bootstrap/errors.js
) - Always do a full build with
USE_MAINTAINER_MODE
switched ON afterwards, before you commit the change. - A
make arangod
is not enough! Since it will not recreate these files!
Reason: These files are only built in maintainer mode, we want that a build in non-maintainer-mode works from every commit. If you only change the generated files, the next build with maintainer mode will delete your changes.
Add backtraces to cluster requests so you can easily track their origin:
-DDEBUG_CLUSTER_COMM
V8 Special flags:
-DENABLE_GDB_JIT_INTERFACE
(enable (broken) GDB intergation of JIT) At runtime arangod needs to be started with these options:
--javascript.v8-options="--gdbjit_dump"
--javascript.v8-options="--gdbjit_full"
If the compile goes wrong for no particular reason, appending 'verbose=' adds more output. For some reason V8 has VERBOSE=1 for the same effect.
-DUSE_JEMALLOC=Off -DBASE_LD_FLAGS="-fsanitize=address" -DBASE_CXX_FLAGS="-fsanitize=address -fno-omit-frame-pointer"
Depending on the native way ArangoDB tries to locate the temporary directory.
- Linux/Mac: the environment variable
TMPDIR
is evaluated. - Windows: the W32 API function GetTempPath() is called
- all platforms:
--temp.path
overrules the above system provided settings.
- start arangod with
--console
to get a debug console - Cheapen startup for valgrind:
--server.rest-server false --javascript.gc-frequency 1000000 --javascript.gc-interval 65536 --scheduler.threads=1 --javascript.v8-contexts=1
- to have backtraces output set this on the prompt:
ENABLE_NATIVE_BACKTRACES(true)
We now have a startup rc file: ~/.arangod.rc
. It's evaled as javascript.
A sample version to help working with the arangod rescue console may look like that:
ENABLE_NATIVE_BACKTRACES(true);
internal = require("internal");
fs = require("fs");
db = internal.db;
time = internal.time;
timed = function (cb) {
var s = time();
cb();
return time() - s;
};
print = internal.print;
HINT: You shouldn't lean on these variables in your Foxx services.
(we switched to eslint a while back - this is still named jslint for historical reasons)
use
./utils/gitjslint.sh
to lint your modified files.
./utils/jslint.sh
to find out whether all of your files comply to jslint. This is required to make continuous integration work smoothly.
if you want to add new files / patterns to this make target, edit the respective shell scripts.
to be safe from committing non-linted stuff add .git/hooks/pre-commit with:
./utils/jslint.sh
If you want to search errors in your js file, jslint is very handy - like a compiler is for C/C++. You can invoke it like this:
bin/arangosh --jslint js/client/modules/@arangodb/testing.js
- Ruby, rspec, httparty to install the required dependencies run:
cd UnitTests/HttpInterface; bundler
- catch (compile time, shipped in the 3rdParty directory)
Since several testing technoligies are utilized, and different arangodb startup options may be required (even different compilation options may be required) the framework is split into testsuites. Get a list of the available testsuites and options by invoking:
./scripts/unittest
To locate the suite(s) associated with a specific test file use:
./scripts/unittest find --test js/common/tests/shell/shell-aqlfunctions.js
or to run all of them:
./scripts/unittest auto --test js/common/tests/shell/shell-aqlfunctions.js
Special patterns in the test filenames are used to select tests to be executed or skipped depending on parameters:
Make use of existing external server. (example scripts/unittest http_server --server tcp://127.0.0.1:8529/ )
These tests will only run if clustering is tested. (option 'cluster' needs to be true).
These tests will only run if no cluster is used. (option 'cluster' needs to be false)
These tests are critical to execution time - and thus may fail if arangod is to slow. This may happen i.e. if you run the tests in valgrind, so you want to avoid them since they will fail anyways. To skip them, set the option skipTimeCritical to true.
These tests are disabled. You may however want to run them by hand.
(only applies to ruby tests) These tests aren't run automatically since they require a manual set up environment.
These tests are ran using the jasmine framework instead of jsunity.
These tests produce a certain thread on infrastructure or the test system, and therefore should only be executed once per day.
There are several major places where unittests live:
- UnitTests/HttpInterface - rspec tests
- tests/* - catch unittests
- js/server/tests - runneable on the server
- js/common/tests - runneable on the server & via arangosh
- js/common/test-data
- js/client/tests - runneable via arangosh
- js/apps/system/aardvark/test
runnuing single rspec test
./scripts/unittest http_server --test api-import-spec.rb
debugging rspec with gdb
server> ./scripts/unittest http_server --test api-import-spec.rb --server tcp://127.0.0.1:7777
- or -
server> ARANGO_SERVER="127.0.0.1:6666" rspec -IUnitTests/HttpInterface --format d --color UnitTests/HttpInterface/api-import-spec.rb
client> gdb --args ./build/bin/arangod --server.endpoint http+tcp://127.0.0.1:6666 --server.authentication false --log.level communication=trace ../arangodb-data-test-mmfiles
debugging a storage engine
host> rm -fr ../arangodb-data-rocksdb/; gdb --args ./build/bin/arangod --console --server.storage-engine rocksdb --foxx.queues false --server.statistics false --server.endpoint http+tcp://0.0.0.0:7777 ../arangodb-data-rocksdb
(gdb) catch throw
(gdb) r
arangod> require("jsunity").runTest("js/client/tests/shell/shell-client.js");
These tests work on the plain RESTfull interface of arangodb, and thus also test invalid HTTP-requests and so forth, plus check error handling in the server.
Assume that you have a test file containing
function exampleTestSuite () {
return {
testSizeOfTestCollection : function () {
assertEqual(5, 5);
};
}
jsUnity.run(aqlTestSuite);
return jsunity.done();
Then you can run the test suite using jsunity.runTest
arangosh> require("jsunity").runTest("test.js");
2012-01-28T19:10:23Z [10671] INFO Running aqlTestSuite
2012-01-28T19:10:23Z [10671] INFO 1 test found
2012-01-28T19:10:23Z [10671] INFO [PASSED] testSizeOfTestCollection
2012-01-28T19:10:23Z [10671] INFO 1 test passed
2012-01-28T19:10:23Z [10671] INFO 0 tests failed
2012-01-28T19:10:23Z [10671] INFO 1 millisecond elapsed
you can engage single tests when running arangod with console like this:
require("jsunity").runTest("js/server/tests/aql-queries-simple.js");
arangosh is similar, however, you can only run tests which are intended to be ran via arangosh:
require("jsunity").runTest("js/client/tests/shell/shell-client.js");
All tests with -spec in their names are using the mochajs.org framework.
To run those tests, e.g. in the arangosh, use this:
const runTest = require("@arangodb/mocha-runner");
runTest(<filepath>, true)
(used in our local Jenkins and TravisCI integration; required for running cluster tests)
Invoked like that:
scripts/unittest all
calling it without parameters like this:
scripts/unittest
will give you a extensive usage help which we won't duplicate here.
Choosing facility
The first parameter chooses the facility to execute. Available choices include:
- all: (calls multiple) This target is utilized by most of the jenkins builds invoking unit tests.
- single_client: (see Running a single unittestsuite)
- single_server: (see Running a single unittestsuite)
- many more - call without arguments for more details.
Different facilities may take different options. The above mentioned usage output contains the full detail.
A commandline for running a single test (-> with the facility 'single_server') using valgrind could look like this. Options are passed as regular long values in the syntax --option value --sub:option value. Using Valgrind could look like this:
./scripts/unittest single_server --test js/server/tests/aql/aql-escaping.js \
--extraArgs:server.threads 1 \
--extraArgs:scheduler.threads 1 \
--extraArgs:javascript.gc-frequency 1000000 \
--extraArgs:javascript.gc-interval 65536 \
--extraArgs:log.level debug \
--extraArgs:log.force-direct true \
--javascript.v8-contexts 2 \
--valgrind /usr/bin/valgrind \
--valgrindargs:log-file /tmp/valgrindlog.%p
- we specify the test to execute
- we specify some arangod arguments via --extraArgs which increase the server performance
- we specify to run using valgrind (this is supported by all facilities)
- we specify some valgrind commandline arguments
- we set the log level to debug
- we force the logging not to happen asynchronous
- eventually you may still add temporary
console.log()
statements to tests you debug.
To debug AQL execution blocks, two steps are required:
- turn on logging for queries using
--extraArgs:log.level queries=info
- send queries enabling block debugging:
db._query('RETURN 1', {}, { profile: 4 })
you now will get log-entries with the contents being passed between the blocks.
Testing a single test with the framework directly on a server:
scripts/unittest single_server --test js/server/tests/aql/aql-escaping.js
You can also only execute a single test case in a jsunity testsuite (in this case testTokens
:
scripts/unittest single_server --test js/server/tests/aql/aql-escaping.js --testCase testTokens
Testing a single test with the framework via arangosh:
scripts/unittest single_client --test js/server/tests/aql/aql-escaping.js
Testing a single rspec test:
scripts/unittest http_server --test api-users-spec.rb
Running a test against a ready started server (in contrast to starting one by itself):
scripts/unittest http_server --test api-batch-spec.rb --server tcp://127.0.0.1:8529 --serverRoot /tmp/123
scripts/unittest is mostly only a wrapper; The backend functionality lives in: js/client/modules/@arangodb/testing.js
Since downloading fox apps from github can be cumbersome with shaky DSL and DOS'ed github, we can fake it like this:
export FOXX_BASE_URL="http://germany/fakegit/"
./scripts/unittest single_server --test 'js/server/tests/shell/shell-foxx-manager-spec.js'
require("jsunity").runTest("js/server/tests/aql/aql-escaping.js");
Filtering for one test case (in this case testTokens
):
require("jsunity").runTest("js/server/tests/aql/aql-escaping.js", false, "testTokens");
require("jsunity").runTest("js/server/tests/aql/aql-escaping.js");
bin/arangod /tmp/dataUT --javascript.unit-tests="js/server/tests/aql/aql-escaping.js" --no-server
js/common/modules/loadtestrunner.js
Generally coredumps have to be enabled using:
ulimit -c unlimited
You should then see:
ulimit -a
core file size (blocks, -c) unlimited
for each shell and its subsequent processes.
Hint: on Ubuntu the apport
package may interfere with this; however you may use the systemd-coredump
package
which automates much of the following:
So that the unit testing framework can autorun gdb it needs to reliably find the corefiles.
In Linux this is configured via the /proc
filesystem, you can make this reboot permanent by
creating the file /etc/sysctl.d/corepattern.conf
(or add the following lines to /etc/sysctl.conf
)
# We want core files to be located in a central location
# and know the PID plus the process name for later use.
kernel.core_uses_pid = 1
kernel.core_pattern = /var/tmp/core-%e-%p-%t
to reload the above settings most systems support:
sudo sysctl -p
Note that the proc
paths translate sub-directories to dots.
The non permanent way of doing this in a running system is:
echo 1 > /proc/sys/kernel/core_uses_pid
echo '/var/tmp/core-%e-%p-%t' > /proc/sys/kernel/core_pattern
(you may also inspect these files to validate the current settings)
More modern systems facilitate systemd-coredump
(via a similar named package) to controll coredumps.
On most systems it will put compressed coredumps to /var/lib/systemd/coredump
.
In order to use automatic coredump analysis with the unittests you need to configure
/etc/systemd/coredump.conf
and set Compress=no
- so instant analysis may take place.
Please note that we can't support Ubuntu Apport.
Please use apport-unpack
to send us the bare coredumps.
In order to get coredumps from binaries changing their UID the system needs to be told that its allowed to write cores from them. Default ArangoDB installations will do exactly that, so the following is neccessary to make the system produce coredumps from production arangodb instances:
Edit /etc/security/limits.conf
to contain:
arangodb - core infinity
Edit the systemd unit file /lib/systemd/system/arangodb3.service
(USE infinity!!!):
## setting for core files
# Any dir that is writable by the user running arangod
WorkingDirectory=/var/lib/arangodb3
# core limit - set this to infinity to enable cores
LimitCORE=0
enable new systemd settings:
systemctl daemon-reload && systemctl restart arangodb3.service
enable suid process dumping:
echo 1 >/proc/sys/fs/suid_dumpable
make the above change permanent:
echo "sys.fs.suid_dumpable = 1" >> /etc/sysctl.d/99-suid-coredump.conf
We offer debug packages containing the debug symbols for your binaries. Please install them if you didn't compile yourselves.
Given you saw in the log of the arangod with the PID 25216
that it died, you should then find
/var/tmp/core-V8 WorkerThread-25216-1490887259
with this information. We may now start GDB and inspect whats going on:
gdb /usr/sbin/arangod /var/tmp/*25216*
These commands give usefull information about the incident:
backtrace full
thread apply all bt
The first gives the full stacktrace including variables of the last active thread, the later one the stacktraces of all threads.
For the average *nix user windows debugging has some awkward methods.
Coredumps can be created using the task manager; switch it to detail view, the context menu offers to create dump file; the generated file ends in a directory that explorer hides from you - AppData - you have to type that in the location bar. This however only for running processes which is not as useful as having dumps of crashing processes. While its a common feature to turn on coredumps with the system facilities on *nix systems, its not as easy in windows. You need an external program from the Sysinternals package: ProcDump. First look up the PID of arangod, you can finde it in the brackets in the arangodb logfile. Then invoke procdump like this:
procdump -accepteula -e -ma < PID of arangod >
It will keep on running and monitor arangod until eventually a crash happens. You will then get a core dump if an incident occurs or Dump count not reached. if nothing happened, Dump count reached. if a dump was written - the filename will be printed above.
Releases are supported by a public symbol server so you will be able to debug cores. Releases starting with 2.5.6, 2.6.3 onwards are supported; Note that you should run the latest version of a release series before reporting bugs. Either WinDbg or Visual studio support setting the symbol path via the environment variable or in the menu. Given we want to store the symbols on e:\symbol_cach we add the arangodb symbolserver like this:
set _NT_SYMBOL_PATH=cache*e:\symbol_cache\cache;srv*e:\symbol_cache\arango*https://www.arangodb.com/repositories/symsrv/;SRV*e:\symbol_cache\ms*http://msdl.microsoft.com/download/symbols
You then will be able to see stack traces in the debugger.
You may also try to download the symbols manually using:
symchk.exe arangod.exe /s SRV*e:/symbol_cache/cache*https://www.arangodb.com/repositories/symsrv/
The symbolserver over at https://www.arangodb.com/repositories/symsrv/ is browseable; thus you can easily download the files you need by hand. It contains of a list of directories corosponding to the components of arangodb:
- arango - the basic arangodb library needed by all components
- arango_v8 - the basic V8 wrappers needed by all components
- arangod - the server process
- the client utilities:
- arangob
- arangobench
- arangoexport
- arangoimp
- arangorestore
- arangosh
- arangovpack
In these directories you will find subdirectories with the hash corosponding to the id of the binaries. Their date should corrospond to the release date of their respective arango release.
This means i.e. for ArangoDB 3.1.11:
This file is a microsoft cabinet file, which is a little bit compressed. You can dismantle it so the windows explorer offers you its proper handler by renaming it to .cab; click on the now named arangod.cab
, copy the contained arangod.pdb into your symbol path.
While Visual studio may cary a nice shiny gui, the concept of GUI fails miserably i.e. in testautomation. Getting an overview over all running threads is a tedious task with it. Here the commandline version of WinDBG cdb comes to the aid. testing.js
utilizes it to obtain automatical stack traces for crashes.
We run it like that:
cdb -z <dump file> -c 'kp; ~*kb; dv; !analyze -v; q'
These commands for -c
mean:
kp print curren threads backtrace with arguments
~*kb print all threads stack traces
dv analyze local variables (if)
!analyze -v print verbose analysis
q quit the debugger
If you don't specify them via -c you can also use them in an interactive manner.
There are two scripts scripts/startLocalCluster
and
scripts/stopLocalCluster
which help you to quickly fire up a testing
cluster on your local machine. scripts/startLocalCluster
takes 0, 2 or
three arguments. In the 0 argument version, it simply starts 2 DBservers
and one coordinator in the background, running on ports 8629, 8630 and
8530 respectively. The agency runs on port 4001. With 2 arguments the
first is the number of DBServers and the second is the number of
coordinators.
If there is a third argument and it is "C", then the first coordinator
will be started with --console`` in a separate window (using an
xterm`).
If there is a third argument and it is "D", then all servers are started
up in the GNU debugger in separate windows (using xterm
s). In that
case one has to hit ENTER in the original terminal where the script runs
to continue, once all processes have been start up in the debugger.
This will spawn a temporary local mesos cluster.
Requirements:
- Somewhat recent linux
- docker 1.10+
- curl
- sed (for file editing)
- jq (for json parsing)
- git
- at least 8GB RAM
- fully open firewall inside the docker network
To startup a local mesos cluster:
git clone https://github.com/m0ppers/mesos-cluster
cd mesos-cluster
mkdir /tmp/mesos-cluster
./start-cluster.sh /tmp/mesos-cluster/ --num-slaves=5 --rm --name mesos-cluster
Then save the following configuration to a local file and name it arangodb3.json
:
{
"id": "arangodb",
"cpus": 0.25,
"mem": 256.0,
"ports": [0, 0, 0],
"instances": 1,
"args": [
"framework",
"--framework_name=arangodb",
"--master=zk://172.17.0.2:2181/mesos",
"--zk=zk://172.17.0.2:2181/arangodb",
"--user=",
"--principal=pri",
"--role=arangodb",
"--mode=cluster",
"--async_replication=false",
"--minimal_resources_agent=mem(*):512;cpus(*):0.25;disk(*):512",
"--minimal_resources_dbserver=mem(*):1024;cpus(*):0.25;disk(*):1024",
"--minimal_resources_secondary=mem(*):1024;cpus(*):0.25;disk(*):1024",
"--minimal_resources_coordinator=mem(*):1024;cpus(*):0.25;disk(*):1024",
"--nr_agents=3",
"--nr_dbservers=2",
"--nr_coordinators=2",
"--failover_timeout=86400",
"--arangodb_privileged_image=false",
"--arangodb_force_pull_image=true",
"--arangodb_image=arangodb/arangodb-mesos:3.0",
"--secondaries_with_dbservers=false",
"--coordinators_with_dbservers=false"
],
"container": {
"type": "DOCKER",
"docker": {
"image": "arangodb/arangodb-mesos-framework:3.0",
"network": "HOST"
}
},
"healthChecks": [
{
"protocol": "HTTP",
"path": "/framework/v1/health.json",
"gracePeriodSeconds": 3,
"intervalSeconds": 10,
"portIndex": 0,
"timeoutSeconds": 10,
"maxConsecutiveFailures": 0
}
]
}
Adjust the lines --master
and --zk
to match the IP of your mesos-cluster:
MESOS_IP=`docker inspect mesos-cluster | \
jq '.[0].NetworkSettings.Networks.bridge.IPAddress' | \
sed 's;";;g'`
sed -i -e "s;172.17.0.2;${MESOS_IP};g" arangodb3.json
And deploy the modified file to your local mesos cluster:
MESOS_IP=`docker inspect mesos-cluster | \
jq '.[0].NetworkSettings.Networks.bridge.IPAddress' | \
sed 's;";;g'`
curl -X POST ${MESOS_IP}:8080/v2/apps \
-d @arangodb3.json \
-H "Content-Type: application/json" | \
jq .
Point your webbrowser to the IP of your echo "http://${MESOS_IP}:8080"
.
Wait until arangodb is healthy.
Then click on arangodb
.
On the following screen click on the first port next to the IP Address of your cluster.
Create local docker images using the following repositories:
- https://github.com/arangodb/arangodb-docker
- https://github.com/arangodb/arangodb-mesos-docker
- https://github.com/arangodb/arangodb-mesos-framework
Then adjust the docker images in the config (arangodb3.json
) and redeploy it using the curl command above.
To build the current frontend build please use the command: make frontend
.
The command make frontend_clean
will remove all available node modules and
start a clean installation.
To add new NPM dependencies switch into the js/node
folder and install them
with npm using the following options:
npm install [<@scope>/]<name> --global-style --save --save-exact
or simply
npm install [<@scope>/]<name> --global-style -s -E
The save
and save-exact
options are necessary to make sure the package.json
file is updated correctly.
The global-style
option prevents newer versions of npm from unrolling nested
dependencies inside the node_modules
folder. Omitting this option results in
exposing all dependencies of all modules to ArangoDB users.
Finally add the module's licensing information to LICENSES-OTHER-COMPONENTS.md
.
When updating dependencies make sure that any mocked dependencies (like glob
for mocha
) match the versions required by the updated module and delete any
duplicated nested dependencies if necessary (e.g. mocha/node_modules/glob
)
to make sure the global (mocked) version is used instead.