This merges the version-3.45.1 tag from upstream SQLite git repository to libSQL with the following conflicts resolved: Conflicts: README.md ext/jni/src/org/sqlite/jni/capi/ConfigSqllogCallback.java libsql-sqlite3/configure libsql-sqlite3/doc/jsonb.md libsql-sqlite3/ext/fts5/test/fts5faultH.test libsql-sqlite3/ext/fts5/test/fts5origintext.test libsql-sqlite3/ext/fts5/test/fts5origintext2.test libsql-sqlite3/ext/fts5/test/fts5origintext3.test libsql-sqlite3/ext/fts5/test/fts5origintext4.test libsql-sqlite3/ext/fts5/test/fts5origintext5.test libsql-sqlite3/ext/fts5/test/fts5secure8.test libsql-sqlite3/ext/fts5/test/fts5tokenizer2.test libsql-sqlite3/ext/fts5/test/fts5trigram2.test libsql-sqlite3/ext/jni/src/org/sqlite/jni/annotation/Experimental.java libsql-sqlite3/ext/jni/src/org/sqlite/jni/capi/ConfigSqlLogCallback.java libsql-sqlite3/ext/jni/src/org/sqlite/jni/capi/ConfigSqllogCallback.java libsql-sqlite3/ext/jni/src/org/sqlite/jni/wrapper1/WindowFunction.java libsql-sqlite3/ext/wasm/GNUmakefile libsql-sqlite3/ext/wasm/batch-runner-sahpool.html libsql-sqlite3/ext/wasm/batch-runner-sahpool.js libsql-sqlite3/src/pager.c libsql-sqlite3/src/shell.c.in libsql-sqlite3/src/sqliteInt.h libsql-sqlite3/src/wal.c libsql-sqlite3/test/fts3integrity.test libsql-sqlite3/test/json/jsonb-q1.txt libsql-sqlite3/test/json106.test libsql-sqlite3/test/json107.test libsql-sqlite3/test/jsonb01.test libsql-sqlite3/test/mmapcorrupt.test libsql-sqlite3/test/releasetest_data.tcl libsql-sqlite3/test/shell9.test libsql-sqlite3/test/wapp.tcl libsql-sqlite3/test/wapptest.tcl
The files in this subdirectory are used to help measure the performance of the SQLite JSON functions, especially in relation to handling large JSON inputs.
1.0 Prerequisites
-
Standard SQLite build environment (SQLite source tree, compiler, make, etc.)
-
Valgrind
-
Fossil (only the "fossil xdiff" command is used by this procedure)
-
tclsh
2.0 Setup
-
Run: "
tclsh json-generator.tcl | sqlite3 json100mb.db
" to create the 100 megabyte test database. Do this so that the "json100mb.db" file lands in the directory from which you will run tests, not in the test/json subdirectory of the source tree. -
Make a copy of "json100mb.db" into "jsonb100mb.db" - change the prefix from "json" to "jsonb".
-
Bring up jsonb100mb.db in the sqlite3 command-line shell. Convert all of the content into JSONB using a commands like this:
UPDATE data1 SET x=jsonb(x); VACUUM;
- Build the baseline sqlite3.c file with sqlite3.h and shell.c.
make clean sqlite3.c
-
Run "
sh json-speed-check.sh trunk
". This creates the baseline profile in "jout-trunk.txt" for the preformance test using text JSON. -
Run "
sh json-speed-check.sh trunk --jsonb
". This creates the baseline profile in "joutb-trunk.txt" for the performance test for processing JSONB -
(Optional) Verify that the json100mb.db database really does contain approximately 100MB of JSON content by running:
SELECT sum(length(x)) FROM data1; SELECT * FROM data1 WHERE NOT json_valid(x);
3.0 Testing
-
Build the sqlite3.c (with sqlite3.h and shell.c) to be tested.
-
Run "
sh json-speed-check.sh x1
". The profile output will appear in jout-x1.txt. Substitute any label you want in place of "x1". -
Run "
sh json-speed-check.sh x1 --jsonb
". The profile output will appear in joutb-x1.txt. Substitute any label you want in place of "x1". -
Run the script shown below in the CLI. Divide 2500 by the real elapse time from this test to get an estimate for number of MB/s that the JSON parser is able to process.
.open json100mb.db .timer on WITH RECURSIVE c(n) AS (VALUES(1) UNION ALL SELECT n+1 FROM c WHERE n<25) SELECT sum(json_valid(x)) FROM c, data1;