Popularity
5.2
Growing
Activity
9.6
Declining
253
18
36

Programming language: Scala
Tags: JSON    
Latest version: v0.55.2

jsoniter-scala alternatives and similar packages

Based on the "JSON" category

Do you think we are missing an alternative of jsoniter-scala or a related project?

Add another 'JSON' Package

README

Jsoniter Scala

AppVeyor build TravisCI build Code coverage Gitter chat Maven Central

Scala macros that generate codecs for case classes, standard types, and collections to get maximum performance of JSON parsing and serialization.

Latest results of benchmarks which compare parsing and serialization performance of Jsoniter Scala with Borer, Circe, Play-JSON, Spray-JSON, Jackson, uPickle, AVSystem's scala-commons, DSL-JSON and Jsoniter Java libraries using different JDK and GraalVM versions on the following environment: Intel® Core™ i7-7700 CPU @ 3.6GHz (max 4.2GHz), RAM 16Gb DDR4-2400, Ubuntu 18.04, and latest versions of OpenJDK 8/11/13, OpenJDK 13 + Graal, and GraalVM CE/EE

Acknowledgments

This library started from macros that reused Jsoniter Java reader & writer and generated codecs for them but then evolved to have own core of mechanics for parsing and serialization.

The idea to generate codecs by Scala macros and main details was borrowed from Kryo Macros and adapted for needs of the JSON domain.

Other Scala macros features were peeped in AVSystem Commons Library for Scala

Goals

  1. Safety: validate JSON format, UTF-8 encoding of strings, and mapped values safely with the fail-fast approach and clear reporting, provide configurable limits for suboptimal data structures with safe defaults to be resilient for DoS attacks, generate codecs that create instances of a fixed set of classes during parsing to avoid RCE attacks
  2. Correctness: parse and serialize numbers without loosing of precision doing half even rounding for too long JSON numbers when they bounded to floats or doubles, do not replace illegally encoded characters of string values by placeholder characters
  3. Speed: do parsing and serialization of JSON directly from UTF-8 bytes to your data structures and back, do it crazily fast without using of run-time reflection, intermediate ASTs, strings or hash maps, with minimum allocations and copying
  4. Productivity: derive codecs recursively for complex types using one line macro, do it in compile-time to minimize the probability of run-time issues, optionally print generated sources as compiler output to be inspected for proving safety and correctness or to be reused as a starting point for the implementation of custom codecs, prohibit serializing of null Scala values and parsing immediately to them in generated codecs
  5. Ergonomics: have preconfigured defaults for the safest and common usage that can be easily altered by compile- and run-time configuration instances, combined with compile-time annotations and implicits, embrace the textual representation of JSON providing a pretty printing option, provide a hex dump in the parse error message to speed up the view of an error context

The library targets JDK 8+ and GraalVM 19+ (including compilation to native images) without any platform restrictions.

Support of Scala.js and Scala Native is not a goal for the moment.

Features and limitations

  • JSON parsing from Array[Byte], java.io.InputStream or java.nio.ByteBuffer
  • JSON serialization to Array[Byte], java.io.OutputStream or java.nio.ByteBuffer
  • Support of parsing from or writing to part of Array[Byte] or java.nio.ByteBuffer by specifying of position and limit
  • Parsing of streaming JSON values and JSON arrays from java.io.InputStream without the need of holding all parsed values in the memory
  • Only UTF-8 encoding is supported when working with buffered bytes directly but there is a fallback to parse and serialize JSON from/to String (while this is much less efficient)
  • Parsing of strings with escaped characters for JSON keys and string values
  • Codecs can be generated for primitives, boxed primitives, enums, tuples, String, BigInt, BigDecimal, Option, Either, java.util.UUID, java.time.* (to/from ISO-8601 representation only), Scala collections, arrays, module classes, literal types, value classes, and case classes with values/fields having any of types listed here
  • Classes should be defined with a primary constructor that has one list of arguments for all non-transient fields
  • Non-case Scala classes also supported but they should have getter accessors for all arguments of a primary constructor
  • Types that supported as map keys are primitives, boxed primitives, enums, String, BigInt, BigDecimal, java.util.UUID, java.time.*, literal types, and value classes for any of them
  • Codecs for sorted maps and sets can be customized by implicit Ordering[K] instances for keys that are available at the scope of the make macro call
  • Parsing of escaped characters is not supported for strings which are mapped to numeric and java.time.* types
  • Support of first-order and higher-kind types
  • Support of 2 representations of ADTs with a sealed trait or a Scala class as a base type and non-abstract Scala classes or objects as leaf classes: 1st representation uses discriminator field with string type of value, 2nd one uses string values for objects and a wrapper JSON object with a discriminator key for case class instances
  • Implicitly resolvable value codecs for JSON values and key codecs for JSON object keys that are mapped to maps allows to inject your custom codecs for adding support of other types or for altering representation in JSON for already supported classes
  • Type aliases are also supported for all types mentioned above
  • Only acyclic graphs of class instances are supported by generated codecs
  • Order of instance fields is preserved during serialization for generated codecs
  • Throws a parsing exception if duplicated keys were detected for a class instance (except maps)
  • Serialization of null values is prohibited by throwing of NullPointerException errors
  • Parsing of null values allowed only for optional of collection types (that means the None value or an empty collection accordingly) and for fields which have defined non-null default values
  • Fields with default values that defined in the constructor are optional, other fields are required (no special annotation required)
  • Fields with values that are equals to default values, or are empty options/collections/arrays are not serialized to provide a sparse output
  • Any values that used directly or as part of default values of the constructor parameters should have right implementations of the equals method (it mostly concerns non-case classes or other types that have custom codecs)
  • Fields can be annotated as transient or just not defined in the constructor to avoid parsing and serializing at all
  • Field names can be overridden for serialization/parsing by field annotation in the primary constructor of classes
  • Reading and writing of any arbitrary bytes or raw values are possible by using custom codecs
  • Parsing exception always reports a hexadecimal offset of Array[Byte], java.io.InputStream or java.nio.ByteBuffer where it occurs and an optional hex dump affected by error part of an internal byte buffer
  • Configurable by field annotation ability to read/write numeric fields from/to string values
  • Both key and value codecs are specialized to be work with primitives efficiently without boxing/unboxing
  • No extra buffering is required when parsing from java.io.InputStream or serializing to java.io.OutputStream
  • Using black box macros only for codec generation ensures that your types will never be changed
  • Ability to print all generated code for codecs using a custom scala compiler option: -Xmacro-settings:print-codecs
  • No dependencies on extra libraries in runtime excluding Scala's scala-library
  • Releases for different Scala versions: 2.11, 2.12, 2.13
  • Support of shading to another package for locking on a particular released version
  • Patch versions are backward and forward compatible
  • Support of compilation to a native image by GraalVM

There are configurable options that can be set in compile-time:

  • Ability to read/write numbers from/to string values
  • Skipping of unexpected fields or throwing of parse exceptions
  • Skipping of serialization of fields that have empty collection values can be turned off to force serialization of them
  • Skipping of serialization of fields that have empty optional values can be turned off to force serialization of them
  • Skipping of serialization of fields which values are matched with defaults that are defined in the primary constructor can be turned off to force serialization of that values
  • Mapping function for property names between classes and JSON, including predefined functions which enforce snake_case, kebab-case, camelCase or PascalCase names for all fields in the generated codec
  • An optional name of the discriminator field for ADTs
  • Mapping function for values of a discriminator field that is used for distinguishing classes of ADTs
  • Ability to set precision, scale limit, and the max number of significant digits when parsing BigDecimal values
  • Ability to set the max number of significant digits when parsing BigInt values
  • Ability to set max allowed value when parsing bit sets
  • Ability to set a limit for the number of inserts when parsing sets or maps
  • Throwing of compilation error for recursive data structures can be turned off

List of options that change parsing and serialization in runtime:

  • Serialization of strings with escaped Unicode characters to be ASCII compatible
  • Indenting of output and its step
  • Throwing of stack-less parsing exceptions by default to greatly reduce the impact on performance, while stack traces can be turned on in development for debugging
  • Turning off hex dumping affected by error part of an internal byte buffer to reduce the impact on performance
  • Preferred size of internal in buffers when parsing from java.io.InputStream or java.nio.DirectByteBuffer
  • Preferred size of internal out buffers when serializing to java.io.OutputStream or java.nio.DirectByteBuffer
  • Preferred size of char buffers when parsing string values

For upcoming features and fixes see Commits and Issues page.

How to use

Add the core library with a "compile" scope and the macros library with a "provided" scope to your dependencies list:

libraryDependencies ++= Seq(
  "com.github.plokhotnyuk.jsoniter-scala" %% "jsoniter-scala-core"   % "0.55.2" % Compile,
  "com.github.plokhotnyuk.jsoniter-scala" %% "jsoniter-scala-macros" % "0.55.2" % Provided // required only in compile-time
)

Generate codecs for your Scala classes and collections:

import com.github.plokhotnyuk.jsoniter_scala.macros._
import com.github.plokhotnyuk.jsoniter_scala.core._

case class Device(id: Int, model: String)

case class User(name: String, devices: Seq[Device])

implicit val codec: JsonValueCodec[User] = JsonCodecMaker.make[User](CodecMakerConfig())

That's it! You have generated an instance of com.github.plokhotnyuk.jsoniter_scala.core.JsonValueCodec.

Now you can use it for parsing and serialization:

val user = readFromArray("""{"name":"John","devices":[{"id":1,"model":"HTC One X"}]}""".getBytes("UTF-8"))
val json = writeToArray(User(name = "John", devices = Seq(Device(id = 2, model = "iPhone X"))))

If you don't know-how to make your data structures from scratch but have a JSON sample then use on-line services 1 2 to generate an initial version of them.

To see generated code for codecs add the following line to your sbt build file

scalacOptions ++= Seq("-Xmacro-settings:print-codecs")

Full code see in the examples directory

For more use cases, please, check out tests:

Samples for integration with different web frameworks:

For all dependent projects it is recommended to use sbt-updates plugin or Scala steward service to keep up with using of latest releases.

Known issues

  1. Scalac has a bug that affects case classes which have 2 fields where the name of one is a prefix for another name that contains a character which should be encoded immediately after the prefix (like o and o-o). You will get compilation or runtime error, depending on the version of the compiler, see details here.

The workaround is to move a definition of the field with encoded chars (o-o in our case) to be after the field that is affected by the exception (after the o field)

  1. A configuration parameter for the make macro is evaluated in compile-time only that require no dependency on other code that uses a result of the macro's call. In that case the following compilation error will be reported:
[error] Cannot evaluate a parameter of the 'make' macro call for type 'full.name.of.YourType'. It should not depend on
        code from the same compilation module where the 'make' macro is called. Use a separated submodule of the project
        to compile all such dependencies before their usage for generation of codecs.

But sometime scalac (or zinc) can fail to compile the make macro call with the same error message for configuration that has not clear dependencies on other code. For those cases workarounds can be simpler than recommended usage of separated submodule:

  • isolate the make macro call(s) in the separated object, like in this PR
  • move jsoniter-scala imports to be local, like here and here
  • use sbt clean compile stage or sbt clean test stage instead of just sbt clean stage, like in this repo
  1. Scalac can throw the following stack overflow exception on make call for ADTs with objects if the call and the ADT definition are enclosed in the definition of some outer class (for more details see: https://github.com/scala/bug/issues/11157):
java.lang.StackOverflowError
    ...
    at scala.tools.nsc.transform.ExplicitOuter$OuterPathTransformer.outerPath(ExplicitOuter.scala:267)
    at scala.tools.nsc.transform.ExplicitOuter$OuterPathTransformer.outerPath(ExplicitOuter.scala:267)
    at scala.tools.nsc.transform.ExplicitOuter$OuterPathTransformer.outerPath(ExplicitOuter.scala:267)
    at scala.tools.nsc.transform.ExplicitOuter$OuterPathTransformer.outerPath(ExplicitOuter.scala:267)
    ...

Workarounds are:

  • don't enclose ADTs with an object into outer classes
  • use the outer object (not a class) instead

How to develop

Feel free to ask questions in chat, open issues, or contribute by creating pull requests (fixes and improvements to docs, code, and tests are highly appreciated)

Run tests, check coverage and binary compatibility

sbt clean coverage test coverageReport
sbt clean +test +mimaReportBinaryIssues

BEWARE: jsoniter-scala is included into Scala Community Build for 2.11.x, 2.12.x, and 2.13.x versions of Scala.

Printing of code generated by macros

To see and check code generated by the make macro add the -Dmacro.settings=print-codecs option like here:

sbt -Dmacro.settings=print-codecs clean test

Also, to print code generated by the eval macro use the -Dmacro.settings=print-expr-results option.

Both options can be combined: -Dmacro.settings=print-codecs,print-expr-results

Run benchmarks

Before benchmark running check if your CPU works in performance mode (not a powersave one). On Linux use following commands to print current and set the performance mode:

cat /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance | sudo tee /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor

Sbt plugin for JMH tool is used for benchmarking, to see all their features and options please check Sbt-JMH docs and [JMH tool docs](https://openjdk.java.net/projects/code-tools/jmh/

Learn how to write benchmarks in JMH samples and JMH articles posted in Aleksey Shipilёv’s and Nitsan Wakart’s blogs.

List of available option can be printed by:

sbt 'jsoniter-scala-benchmark/jmh:run -h'

Results of benchmark can be stored in different formats: *.csv, *.json, etc. All supported formats can be listed by:

sbt 'jsoniter-scala-benchmark/jmh:run -lrf'

JMH allows to run benchmarks with different profilers, to get a list of supported use (can require entering of user password):

sbt 'jsoniter-scala-benchmark/jmh:run -lprof'

Help for profiler options can be printed by following command (<profiler_name> should be replaced by name of the supported profiler from the command above):

sbt 'jsoniter-scala-benchmark/jmh:run -prof <profiler_name>:help'

For parametrized benchmarks the constant value(s) for parameter(s) can be set by -p option:

sbt clean 'jsoniter-scala-benchmark/jmh:run -p size=1,10,100,1000 ArrayOf.*'

To see throughput with allocation rate of generated codecs run benchmarks with GC profiler using the following command:

sbt clean 'jsoniter-scala-benchmark/jmh:run -prof gc -rf json -rff jdk8.json .*Reading.*'

Results that are stored in JSON can be easy plotted in JMH Visualizer by drugging & dropping of your file to the drop zone or using the source parameter with an HTTP link to your file in the URL like here.

On Linux the perf profiler can be used to see CPU event statistics normalized per ops:

sbt clean 'jsoniter-scala-benchmark/jmh:run -prof perfnorm TwitterAPIReading.jsoniterScala'

To get a result for some benchmarks with an in-flight recording file from JFR profiler use command like this:

sbt clean 'jsoniter-scala-benchmark/jmh:run -jvmArgsAppend "-XX:+UnlockDiagnosticVMOptions -XX:+DebugNonSafepoints" -prof "jmh.extras.JFR:dir=/tmp/profile-jfr;flameGraphDir=/home/andriy/Projects/com/github/brendangregg/FlameGraph;jfrFlameGraphDir=/home/andriy/Projects/com/github/chrishantha/jfr-flame-graph;verbose=true" -wi 10 -i 60 TwitterAPIReading.jsoniterScala'

Now you can open files from the /tmp/profile-jfr directory:

profile.jfr                             # JFR profile, open and analyze it using JMC
jfr-collapsed-cpu.txt                   # Data from JFR profile that are extracted for Flame Graph tool
flame-graph-cpu.svg                     # Flame graph of CPU usage
flame-graph-cpu-reverse.svg             # Reversed flame graph of CPU usage
flame-graph-allocation-tlab.svg         # Flame graph of heap allocations in TLAB
flame-graph-allocation-tlab-reverse.svg # Reversed flame graph of heap allocations in TLAB

To run benchmarks with recordings by Async profiler, clone its repository and use command like this:

sbt -no-colors 'jsoniter-scala-benchmark/jmh:run -jvmArgsAppend "-XX:+UnlockDiagnosticVMOptions -XX:+DebugNonSafepoints" -prof "jmh.extras.Async:event=cpu;dir=/tmp/profile-async;asyncProfilerDir=/home/andriy/Projects/com/github/jvm-profiling-tools/async-profiler;flameGraphDir=/home/andriy/Projects/com/github/brendangregg/FlameGraph;flameGraphOpts=--color,java;verbose=true" -wi 10 -i 60 TwitterAPIReading.jsoniterScala'

To see list of available events need to start your app or benchmark, and run jps command. I will show list of PIDs and names for currently running Java processes. While your Java process still running launch the Async Profiler with the list option and ID of your process like here:

$ ~/Projects/com/github/jvm-profiling-tools/async-profiler/profiler.sh list 6924
Basic events:
  cpu
  alloc
  lock
  wall
  itimer
Perf events:
  page-faults
  context-switches
  cycles
  instructions
  cache-references
  cache-misses
  branches
  branch-misses
  bus-cycles
  L1-dcache-load-misses
  LLC-load-misses
  dTLB-load-misses
  mem:breakpoint
  trace:tracepoint

Following command can be used to profile and print assembly code of hottest methods, but it requires a setup of an additional library to make PrintAssembly feature enabled:

sbt clean 'jsoniter-scala-benchmark/jmh:run -prof perfasm -wi 10 -i 10 -p size=128 BigIntReading.jsoniterScala'

More info about extras, options and ability to generate flame graphs see in Sbt-JMH docs

Other benchmarks with results for jsoniter-scala:

  • comparison with other JSON parser for Scala mostly on samples from real APIs
  • comparison with best binary parsers and serializers for Scala
  • comparison with different binary and text serializers for Scala
  • comparison with JSON serializers for Scala on synthetic samples
  • comparison with a state of the art filter that by "building structural indices converts control flow into data flow, thereby largely eliminating inherently unpredictable branches in the program and exploiting the parallelism available in modern processors"

Publish locally

Publish to local Ivy repo:

sbt clean +publishLocal

Publish to local Maven repo:

sbt clean +publishM2

Release

For version numbering use Recommended Versioning Scheme that is used in the Scala ecosystem.

Double-check binary and source compatibility, including behavior, and release using the following command (credentials are required):

sbt release

Do not push changes to github until promoted artifacts for the new version are not available for download on Maven Central Repository to avoid binary compatibility check failures in triggered Travis CI builds.

The last step is updating of the tag info in a release list.