Skip to main content

Taking a Spark Profile

Bloom logo

Taking a Spark Profile

What does the plugin do?

Spark is a performance profiling plugin (with limited availability for Forge and Fabric) that can display server information, such as TPS, memory, tick durations, CPU usage, and disk usage. But more importantly it can create a performance profile that can be viewed on a fully feature web interface. It does not require any configuration and is incredibly easy to install.

Setup

note

You can ignore this step if you are running Purpur 1.19.1 or above.

Paper 1.21+ includes Spark by default and you do not need to download Spark separately.

Download the latest build from Spark and drop it into your plugins (or mod folder if you are running Forge/Fabric) folder. Turn on or restart the server, and you're done! If you need help installing plugins, check How to install plugins.

Using The Profiler

note

Some versions of spark use different command roots, /sparkb for BungeeCord, /sparkv for Velocity, and /sparkc for Fabric/Forge. If you are having trouble running the command try using the corresponding subfix.

The most useful tool in terms of diagnostics is the profiler feature which can be run with /spark profiler start.

Usually you'll want to run the profile during a time of stress for your server, to get the most out of the report you should try to leave it running for 10 minutes or more. You can also add additional parameters to the report, for more information check the Spark Documentation

Once you are ready to check the results of the profiler you should run /spark profiler stop and copy the link produeced.

When you open the link you'll find a useful web interface that displays all the data collected.

img

Other Useful Commands

Running /tps will display an output that looks like this:

img

Running /spark healthreport will display an output that looks like this:

img

The values will depend on your server, newer versions of Spark will contain more useful information, as the utility is in active development.

For more parameters to use, and other useful commands you can check the Official Spark Documentation

Need Help Reading the Profile?

If you require help with reading your Spark profile, create a ticket in Bloom.host Discord!.

Info

Spark's Official Website Github