The Nuxeo Platform can be used for a very broad range of application, each with different use cases and each involving differing expectations around performance.
Depending on your particular use cases and the throughput you expect, you will likely need to adapt the Nuxeo Platform configuration as well as the configuration of the underlying systems, in order to obtain the performance you seek.
This page provides guidance on how to optimize your application performance, as well as how to troubleshoot performance-related issues through monitoring. This information will also come in very handy should you seek assistance through our Support Department.
Context is key – when gathering information about performance (whether you are setting your internal objectives or troubleshooting problems), it's important to clarify what you are trying to achieve and how you are going about it.
In short, one must be able to quantify or measure performance issues, in order to be able to solve them.
Describe Your Test Scenario
What you are testing, and how you are testing it? Consider the following elements:
- Which of the Nuxeo entry points are you testing?
- JSF UI,
- Rest API,
- Java API,
Your usage scenario:
- Bulk import? Browse? Search? Convert? ...
- How many concurrent users? What is the ramp up?
Your test data:
- How many documents do you have in the repository (approximately)?
- What content types are you using? Where to check for this?
- How many users do you have?
- How many ACEs (access control entries) do you have? Where to check for this?
The testing tools are you using, if any:
- Manual load testing?
- FunkLoad, JMeter, LoadRunner ?
Having this information will help identify the part of the platform you are putting pressure on, and provide insight into how to reproduce the issue, which is an important step in identifying the underlying cause(s), setting up targeted monitoring, and qualifying potential resolutions.
Metrics &Amp; Measurements
Performance is relative – if something appears slow, we need to define what "slow" means for your application.
The following metrics should be considered when qualifying a performance issue.
- Number of request/sec,
- Number of document injected/sec,
- Number of queries executed/sec.
- Information about your system:
- JVM monitoring (Heap, GC, Threads) need links to how to install & monitor this;
- Nuxeo Host system (CPU, Memory, I/O) PerfMon, fs_usage vm_stat, top, netstat;
- DB Server (CPU, Memory, I/O, Locks, Slow queries ...) PerfMon for SQLServer, Enterprise Manager for Oracle, ...?);
- Are any systems running on a VM (virtual machine)?
- Is the DB on the same host/VM?
- Where is the binary store (local disk, network mount, same disk as Nuxeo host server or database, AmazonS3, ...)?
- Is this on an SSD (solid state disk)?
- Is the DB storage using SSD (solid state disks)?
You can find more information about monitoring Nuxeo in the Nuxeo Metrics page.
For simple and automatic GC (garbage collection in the JVM) monitoring, see the Monitoring Page.
- What version of Nuxeo are you using?
- What distribution are you using (CAP, DM, DAM, ...)?
- Any specific add-ons that may be relevant (Quotas, Multi-Tenant, custom code)?
- What database server are you using? MSSQL, Oracle, PostgreSQL, other?
- What is the hardware you are testing on?
Key Points For Consideration
Monitoring is Key
Performing benchmarking and performance tests is only effective when good monitoring is in place – the metrics you capture from monitoring the various systems and components is crucial when trying to pinpoint bottlenecks, and crucial in determining if a particular modification yields any improvement.
If you use FunkLoad, it includes system monitoring.
For more information on monitoring the Nuxeo Platform, please see Nuxeo Metrics page.
On most modern hardware, CPUs are multi-core. This means you will need to use multi-threading in order to take full advantage of the parallelism benefits of multi-core CPUs.
To control how you handle threads on Nuxeo, you can adjust its pool settings.
There are three settings in Nuxeo that are directly related and that should be adjusted in a consistent manner:
- Thread pool size (
maxThreadsin server.xml or common-base/conf/server.xml.nxftl). This is the number of concurrent in-bound HTTP requests that Tomcat will handle in parallel;
- Waiting queue: This is the size of the in-bound HTTP requests that are accepted and held while waiting for a free thread.
- Thread pool size (
- AJP connector
- Nuxeo: DB connection pool
- VCS pool (
- Other DB pool (
- VCS pool (
- Database server: Maximum concurrent connection and transactions.
In general the rule-of-thumb is that for each running Tomcat HTTP thread:
- You will need one connection from VCS to access the repository;
- You may need one connection from the generic pool to access an other datasource.
This means that for a typical configuration, you will have:
maxThreads = nuxeo.vcs.max-pool-size.
If you are not using Nuxeo in cluster mode you must ensure that:
nuxeo.vcs.max-pool-size + 1 (lock management) < nuxeo.db.max-pool-size.
If you are using Nuxeo in cluster mode you must ensure that:
nuxeo.vcs.max-pool-size + 1 (cluster connection) + 1 (lock management) < nuxeo.db.max-pool-size.
nuxeo.db.max-pool-size is set to
nuxeo.vcs.max-pool-size + 10% to handle any thread requesting a DB connection but not a VCS connection (VCS is used to manipulate the repository).
Failure to set the DB server's in-bound connections to a large enough number to accommodate all possible connections from all running Nuxeo nodes, can create a bottleneck and severely impact performance.
It's important to understand that using more threads won't always speed up the application: the more threads you have, the more they will share the resources (CPU, DB ...), and the longer the transactions will be. This may create locking and conflicts issues at database level.
Adjusting the correct number of threads is dependent on many factors. This is where benchmarking can help, as it will allow you to adjust various settings and compare results in a consistent manner.
You may very well have at some point more requests than available threads: that's where the Tomcat queuing will be useful.
Common Problems & Hints
DB Connections / Transaction Starvation
When the database connection pool size is too high, you can end up in a situation where the database cannot provide anymore connections or transactions. In this case you will start having JDBC errors coming from the pools saying that transaction or connection was refused by the server.
When the HTTP thread pool size is too large in proportion to the database connection pool, you can end up in a situation where HTTP requests:
- will be waiting for next available connection in the pool;
- may eventually exit with an error such as "can not obtains a connection from the pool within the blocking timeout".
Deadlocks & Transaction Rollbacks
Deadlocks and rollbacks will occur when there are several transactions affective the same records in the database. Inside the Nuxeo Platform, this implies that more than one thread is attempting to modify the same document. For interactive users this is very unlikely to happen as the application provides thread-safe locking protection.
This kind of problem would occur if several asynchronous processes affect the same document. Typically, this could occur when several asynchronous listeners are triggered when a document is created or updated: for example virus scan + OCR + quota. However, most databases will correctly handle this via built-in row-locking. Therefore deadlocks are often the sign of additional problems being present, such as:
- Long running transactions:
- Creates more concurrent transactions;
- Creates more overlapping between transactions;
- Too many concurrent transactions:
- Lock escalation moves from row to page;
- Interleaving creates long running transactions.
- Reduce as much as possible long running transactions:
- Speed up transactions by speeding up I/O and reducing disk accesses;
- Split long running transactions in sub transactions.
- Configure the pool to limit concurrency to what the database can actually handle.
The DBA is Your Friend
By default the Nuxeo Platform will create database tables with very few indexes – this is by design as we cannot foresee what queries you will require and therefore which indexes should be created to make them more efficient.
In the default configuration, we create a set of default indexes that will be useful in most of the cases.
But to optimize your database, you should:
- Run benchmarks of typical user activity;
- Use the database monitoring tools to highlight the slowest queries;
- create indexes in the database that will speed these queries.
Please refer to the VCS documentation in order to have a better understanding about how the Nuxeo Platform uses the database.
For more information about database-specific configurations and how to report problems, please see the Database section.
CPU Is Not 100%: Don't Panic
When running performance benchmarks, you may very well reach the maximum system throughput without seeing CPU utilization 100%. This is normal – it merely implied that the bottleneck is not the processing power. Other system bottleneck can include:
- Disk I/O,
- Database speed,
- Memory allocation and GC,
- Network latency and I/O.
If your benchmarks show that you use 100% of the CPU, this is usually a good sign as CPU is typically the easiest resource to add more of, and can add more Nuxeo nodes in order to increase performance.