In this article, I will establish how the SQL language and its implementations distinguish between changed values and modified values, where a changed value is a value that has been “touched”, but not necessarily modified, i.e. the value might be the same before and after the change.
Many ORMs, unfortunately, either update all of a record’s values, or only the modified ones. The first can be inefficient, and the latter can be wrong. Updating the changed values would be correct.
Note that you may have a different definition of changed and modified. For this article, let’s just assume that the above definition is as valid as it is useful.
Introduction
A very interesting discussion was triggered recently by Vlad Mihalcea who was looking for an answer to this interesting question:
What’s the overhead of updating all columns, even the ones that haven’t changed?
Apart from the question being very interesting from a performance perspective, the tweet also inspired functional aspects of a distinction between updating all columns vs. updating some columns, which I’ll summarise in this article.
What’s the Problem?
The problem is one that all ORM vendors need to solve: ORMs have a client side representation of the relational model, and that representation is cached (or “out of sync”) for a user to change and then persist again. The problem is now how to re-synchronise the client side representation with the server side representation in a consistent and correct way.
Sidenote: By ORM I understand any tool that maps from a client side representation of your database schema to the database schema itself, regardless if the product supports full-fledged JPA-style object graph persistence, or “merely” implements an “active record” pattern, such as jOOQ 3.x (I find that distinction a bit academic).
All such ORMs have a client side representation of a database record, for instance given the following table (I’m going to be using PostgreSQL syntax):
CREATE TABLE customer ( customer_id SERIAL8 NOT NULL PRIMARY KEY, first_name VARCHAR(50) NOT NULL, last_name VARCHAR(50) NOT NULL )
You’re going to have a client side representation as the following (using Java, e.g. jOOQ or JPA):
// jOOQ generated UpdatableRecord public class CustomerRecord extends UpdatableRecordImpl<CustomerRecord> { public CustomerRecord setCustomerId(Long customerId) { ... } public Long getCustomerId() { ... } public CustomerRecord setFirstName(String firstName) { ... } public String getFirstName() { ... } ... } // JPA annotated entity @Entity public class Customer { @Id @GeneratedValue(strategy = IDENITITY) public long customerId; @Column public String firstName; ... }
In principle, these two approaches are the same thing with the distinction that jOOQ explicitly governs all UpdatableRecord
interactions through type inheritance, whereas JPA makes this dependency more implicit through annotations:
- jOOQ – explicit behavioural dependency between entity and jOOQ logic
- JPA – implicit behavioural dependency between entity and JPA entity manager
In principle, the distinction is just a matter of taste, a programming style: Explicit vs. declarative.
But from a practical perspective, the JPA implementation lacks an important feature when it comes to synching the state back to the database. It cannot reflect change, only modification.
How to synch the state back to the database?
Let’s assume we have a customer called John Doe:
INSERT INTO customer (first_name, last_name) VALUES ('John', 'Doe');
And that customer now changes their names to John Smith. We have several options of sending that update to the database, through “PATCH” or “PUT” semantics – terminology used by Morgan Tocker in another tweet in that discussion:
-- PATCH UPDATE customer SET last_name = 'Smith' WHERE id = ? -- PUT UPDATE customer SET first_name = 'John', last_name = 'Smith' WHERE customer_id = ?
A “PATCH” operation sends only the changed values back to the server, whereas a “PUT” operation sends the entire entity back to the server.
Discussion – Semantics.
In favour of PUT
The two operations are semantically very different. If another session attempts to rename this customer to Jane Doe concurrently (and without optimistic locking being in place), then the PATCH operation might result in an inconsistent outcome (Jane Smith), whereas the PUT operation would still produce one of the expected results, depending on what write is executed first:
-- PATCH result: Jane Smith -- PATCH 1 UPDATE customer SET last_name = 'Smith' WHERE customer_id = ? -- PATCH 2 UPDATE customer SET first_name = 'Jane' WHERE customer_id = ? -- PUT result: Jane Doe -- PUT 1 UPDATE customer SET first_name = 'John', last_name = 'Smith' WHERE customer_id = ? -- PUT 2 UPDATE customer SET first_name = 'Jane', last_name = 'Doe' WHERE customer_id = ?
This is one of the reasons why Hibernate, as a JPA implementation, always implements PUT semantics by default, sending all the columns at once. You can opt out of this by using the @DynamicUpdate
, which will only update modified values (not “changed” values, I’ll explain this distinction later).
This makes perfect sense in such a trivial setup, but it is a short-sighted solution, when the table has many more columns. We’ll see right away why:
In favour of PATCH
One size doesn’t fit all. Sometimes, you do want concurrent updates to happen, and you do want to implement PATCH semantics, because sometimes, two concurrent updates do not work against each other. Take the following example using an enhancement of the customer table.
Business is asking us to collect some aggregate metrics for each customer. The number of clicks they made on our website, as well as the number of purchases they made:
CREATE TABLE customer ( customer_id SERIAL8 NOT NULL PRIMARY KEY, first_name VARCHAR(50) NOT NULL, last_name VARCHAR(50) NOT NULL, clicks BIGINT NOT NULL DEFAULT 0, purchases BIGINT NOT NULL DEFAULT 0 )
And, of course, once you agree that the above design is a suitable one, you’ll immediately agree that here, PATCH semantics is more desirable than PUT semantics:
-- Updating clicks UPDATE customer SET clicks = clicks+1 WHERE customer_id = ? -- Updating purchases UPDATE customer SET purchases = purchases+1 WHERE customer_id = ?
Not only do we update only an individual column, we’re doing it entirely in SQL, including the calculation. With this approach, we do not even need optimistic locking to guarantee update correctness, as we’re not using any client side cached version of the customer record, which could be out of date and would need optimistic (or worse: pessimistic) locking.
If we implemented this differently, using client side calculation of the updated clicks / purchases counters…
-- Updating clicks UPDATE customer SET clicks = ? WHERE customer_id = ? -- Updating purchases UPDATE customer SET purchases = ? WHERE customer_id = ?
… then we’d need one of these techniques:
- Pessimistic locking: Nope, won’t work. We could still get incorrect updates
- Optimistic locking: Indeed, any update would need to be done on a versioned customer record, so if there are two concurrent updates, one of them will fail and could try again. This guarantees data integrity, but will probably make this functionality very painful, because a lot of click updates are probably done in a short amount of time, and they would need to be repeated until they work!
- Client side synchronisation: Of course, we could prevent concurrency for these updates on the client side, making sure that only one concurrent process ever updates click counts (for a given customer). We could implement a click count update queue for this.
All of the above options have significant drawbacks, the easiest solution is really to just increment the counter directly in the database.
And don’t forget, if you choose a bind-variable based solution, and opt for updating ALL the columns, rather than just the changed one, your first_name / last_name updates might conflict with these counter updates as well, making things even more complicated.
Partial PUT (or compound PATCH)
In fact, from a semantics perspective, if you do want to use an ORM to update an entity, you should think about a “partial PUT” semantics, which separates the different entity elements in “sub entities”. From a relational perspective, of course, no such thing as a subentity exists. The above example should be normalised into this, and we would have much less concurrency issues:
CREATE TABLE customer ( customer_id SERIAL8 NOT NULL PRIMARY KEY, first_name VARCHAR(50) NOT NULL, last_name VARCHAR(50) NOT NULL ); CREATE TABLE customer_clicks customer_id BIGINT NOT NULL PRIMARY KEY REFERENCES customer, clicks BIGINT NOT NULL DEFAULT 0 ); CREATE TABLE customer_purchases customer_id BIGINT NOT NULL PRIMARY KEY REFERENCES customer, purchases BIGINT NOT NULL DEFAULT 0 );
This way, the previously mentioned PUT semantics would not create situations where individual, semantically unrelated updates (updates to names, updates to clicks) would interfere with each other. We would only need to make sure that e.g. two competing updates to clicks are correctly serialised.
Practically, we often don’t design our databases this way, either for convenience reasons, for optimised storage, for optimised querying (see also our article when normalisation and surrogate keys hurt performance).
jOOQ’s “changed” value semantics
So that “sub entity” is really just a logical thing, which can be represented either as a logically separate entity in JPA, or we can use jOOQ, which works a bit differently here. In jOOQ, we can change an UpdatableRecord
only partially, and that partial change is sent to the server:
CustomerRecord customer = ctx .selectFrom(CUSTOMER) .where(CUSTOMER.CUSTOMER_ID.eq(customerId)) .fetchOne(); customer.setFirstName("John"); customer.setLastName("Smith"); assertTrue(customer.changed(CUSTOMER.FIRST_NAME)); assertTrue(customer.changed(CUSTOMER.LAST_NAME)); assertFalse(customer.changed(CUSTOMER.CLICKS)); assertFalse(customer.changed(CUSTOMER.PURCHASES)); customer.store(); assertFalse(customer.changed(CUSTOMER.FIRST_NAME)); assertFalse(customer.changed(CUSTOMER.LAST_NAME)); assertFalse(customer.changed(CUSTOMER.CLICKS)); assertFalse(customer.changed(CUSTOMER.PURCHASES));
This will send the following statement to the server:
UPDATE customer SET first_name = ?, last_name = ? WHERE customer_id = ?
Optionally, just as with JPA, you can turn on optimistic locking on this statement. The important thing here is that the clicks
and purchases
columns are left untouched, because they were not changed by the client code. This is different from JPA, which either sends all the values by default, or if you specify @DynamicUpdate
in Hibernate, it would send only the last_name
column, because while first_name
was changed it was not modified.
My definition:
- changed: The value is “touched”, its state is “dirty” and the state needs to be synched to the database, regardless of modification.
- modified: The value is different from its previously known value. By necessity, a modified value is always changed.
As you can see, these are different things, and it is quite hard for a JPA-based API like Hibernate to implement changed semantics because of the annotation-based declarative nature of how entities are defined. We’d need some sophisticated instrumentation to intercept all data changes even when the values have not been modified (I didn’t make those attributes public by accident).
Without this distinction, however, it is unreasonable to use @DynamicUpdate
in Hibernate, as we might run into that situation we didn’t want to run into, where we get a customer called “Jane Smith” – or we use optimistic locking, in case of which there’s not much point in using @DynamicUpdate
.
The database perspective
From a database perspective, it is also important to distinguish between change and modification semantics. In the answer I gave on Stack Exchange, I’ve illustrated two situations:
INSERTs and DEFAULT values
Thus far, we’ve discussed only UPDATE
statements, but similar reasoning may be made for INSERT
as well. These two statements are the same:
INSERT INTO t (a, b) VALUES (?, ?); INSERT INTO t (a, b, c) VALUES (?, ?, DEFAULT);
This one, however, is different:
INSERT INTO t (a, b, c) VALUES (?, ?, ?);
In the first case, a DEFAULT
clause (e.g. timestamp generation, identity generation, trigger value generation, etc.) may apply to the column c
. In the second case, the value c
is provided explicitly by the client.
Languages like Java do not have any way to represent this distinction between
NULL
(which is usually, but not always, theDEFAULT
) in SQL- an actual
DEFAULT
This can only be achieved when an ORM implements changed semantics, like jOOQ does. When you create a customer with jOOQ, then clicks
and purchases
will have their DEFAULT
applied:
CustomerRecord c1 = ctx.newRecord(CUSTOMER); c1.setFirstName("John"); c1.setLastName("Doe"); c1.store(); CustomerRecord c2 = ctx.newRecord(CUSTOMER); c2.setFirstName("Jane"); c2.setLastName("Smith"); c2.setClicks(1); c2.setPurchases(1); c2.store();
Resulting SQL:
-- c1.store(); INSERT INTO customer (first_name, last_name) VALUES (?, ?); -- c2.store(); INSERT INTO customer (first_name, last_name, clicks, purchases) VALUES (?, ?, ?, ?);
In both cases, that’s what the user tells jOOQ to do, so jOOQ will generate a query accordingly.
Back to UPDATE statements
Consider the following example using Oracle triggers:
CREATE TABLE x (a INT PRIMARY KEY, b INT, c INT, d INT); INSERT INTO x VALUES (1, 1, 1, 1); CREATE OR REPLACE TRIGGER t BEFORE UPDATE OF c, d -- Doesn't fire on UPDATE OF b! ON x BEGIN IF updating('c') THEN dbms_output.put_line('Updating c'); END IF; IF updating('d') THEN dbms_output.put_line('Updating d'); END IF; END; / SET SERVEROUTPUT ON UPDATE x SET b = 1 WHERE a = 1; UPDATE x SET c = 1 WHERE a = 1; UPDATE x SET d = 1 WHERE a = 1; UPDATE x SET b = 1, c = 1, d = 1 WHERE a = 1;
It results in the following output:
table X created. 1 rows inserted. TRIGGER T compiled 1 rows updated. 1 rows updated. Updating c 1 rows updated. Updating d 1 rows updated. Updating c Updating d
As you can see, the trigger doesn’t fire when we update only column b
, which it is not interested in. Again, this goes in the direction of distinguishing between changed and modified values, where a trigger fires only when a value is changed (but not necessarily modified).
Now, if an ORM will always update all the columns, this trigger will not work correctly. Sure, we can compare :OLD.b
and :NEW.b
, but that would check for modification, not change, and it might be costly to do so for large strings!
Speaking of costs…
Performance
Statement caching: Weakly in favour of PUT
While one of the reasons the Hibernate team mentioned in favour of updating all the columns is improved cursor cache performance (fewer distinct SQL statements need to be parsed by the database as there are fewer distinct update configurations), I suggest that this “premature optimisation” is negligible. If a client application runs dynamic updates (in the jOOQ sense, where changed values are updated, not just modified values), then chances that the possible SQL statements that need to be parsed will explode are slim to non-existent.
I would definitely like to see real-world benchmarks on this topic!
Batching: Weakly in favour of PUT
When you want to batch tons of update statements from JDBC, then indeed, you will need to ensure that they all have the exact same SQL string. However, this is not a good argument in favour of using PUT semantics and updating all columns.
I’m saying “not good”, because such a batched update should still only consider a subset of the columns for update, not all the columns. And that subset should be determined on aggregated changed flags, not data modification.
Index updates: In favour of PATCH (depending on the database)
Most databases optimise index updates to ignore indexes whose columns have not been changed. Oracle also doesn’t update indexes whose columns have not been modified, in case of which PUT and PATCH semantics both work the same way from an indexing perspective. Other databases may not work this way, where PATCH semantics is favourable.
But even if the optimisation is in place, the old and the new values have to be compared for equality (i.e. to see if a modification took place). You don’t want to compare millions of strings per second if there’s no need to do so! Check out Morgan Tocker’s interesting answer on Stack Exchange, from a MySQL perspective
So, why not just prevent expensive modification checks by telling the database what has changed, instead?
UNDO overhead: In favour of PATCH
Every statement has a footprint on the UNDO / REDO logs. As I’ve shown above, the statements are semantically different in many ways, so if your statement is bigger (more columns are updated), then the impact on the UNDO / REDO log is bigger as well. This can have drastic effects depending on the size of your table / columns:
Don’t forget that this can also affect backup performance!
More performance related information in this blog post:
https://jonathanlewis.wordpress.com/2007/01/02/superfluous-updates
Note: While these bits of information were mostly Oracle-specific, common sense dictates that other RDBMS will behave in similar ways.
Conclusion
With all these negative aspects to including unnecessary columns for update through an ORM compared to the almost negligible benefits, I’d say that users should move forward and completely avoid this mess. Here’s how:
- jOOQ optimises this out of the box, if users set the changed values explicitly. Beware that when you “load” a POJO into a Record, it will set all the columns to changed, which may or may not be the desired effect!
- Hibernate allows for
@DynamicUpdate
, which may work incorrectly as we have minimal “PATCH” semantics based on modified values, not on changed values. However, JPA allows for declaring more than one entity per table, which might certainly be a valid option for this kind of problem - Normalisation is always an option, with its own trade offs. The
clicks
andpurchases
columns could be externalised in separate tables, if this benefits the overall design. - More often than not, writing an UPDATE with SQL directly is the best choice. As we’ve seen in this article, the counters should be updated with expressions of the form
clicks = clicks + 1
, which circumvents most problems exposed in this article.
In short, as Michael Simons said:
And we all do feel very dirty when we write SELECT *
, right? So we should at least be wary of updating all the columns as well.
Filed under: java, sql Tagged: Active Record, hibernate, java, jooq, jpa, Optimistic Locking, ORMs, pessimistic locking, sql, Updating Image may be NSFW.
Clik here to view.
Clik here to view.
Clik here to view.
Clik here to view.
Clik here to view.
Clik here to view.
Clik here to view.
Clik here to view.
