=================================
PostGreSQL/JSONB Data Persistence
=================================
This document outlines the general capabilities of the ``pjpersist``
package. ``pjpersist`` is a PostGreSQL/JSONB storage implementation for
persistent Python objects. It is *not* a storage for the ZODB.
The goal of ``pjpersist`` is to provide a data manager that serializes
objects to JSONB blobs at transaction boundaries. The PJ data manager is a
persistent data manager, which handles events at transaction boundaries (see
``transaction.interfaces.IDataManager``) as well as events from the
persistency framework (see ``persistent.interfaces.IPersistentDataManager``).
An instance of a data manager is supposed to have the same life time as the
transaction, meaning that it is assumed that you create a new data manager
when creating a new transaction:
>>> import transaction
Note: The ``conn`` object is a ``psycopg.Connection`` instance. In this case
our tests use the ``pjpersist_test`` database.
Let's now define a simple persistent object:
>>> import datetime
>>> import persistent
>>> class Person(persistent.Persistent):
...
... def __init__(self, name, phone=None, address=None, friends=None,
... visited=(), birthday=None):
... self.name = name
... self.address = address
... self.friends = friends or {}
... self.visited = visited
... self.phone = phone
... self.birthday = birthday
... self.today = datetime.datetime(2014, 5, 14, 12, 30)
...
... def __str__(self):
... return self.name
...
... def __repr__(self):
... return '<%s %s>' %(self.__class__.__name__, self)
We will fill out the other objects later. But for now, let's create a new
person and store it in PJ:
>>> stephan = Person('Stephan')
>>> stephan
<Person Stephan>
The datamanager provides a ``root`` attribute in which the object tree roots
can be stored. It is special in the sense that it immediately writes the data
to the DB:
>>> dm.root['stephan'] = stephan
>>> dm.root['stephan']
<Person Stephan>
Custom Persistence Tables
-------------------------
By default, persistent objects are stored in a table having the escaped
Python path of the class:
>>> from pjpersist import serialize
>>> person_cn = serialize.get_dotted_name(Person, True)
>>> person_cn
'u__main___dot_Person'
>>> transaction.commit()
>>> dumpTable(person_cn)
[{'data': {'_py_persistent_type': '__main__.Person',
'address': None,
'birthday': None,
'friends': {},
'name': 'Stephan',
'phone': None,
'today': {'_py_type': 'datetime.datetime',
'value': '2014-05-14T12:30:00.000000'},
'visited': []},
'id': '0001020304050607080a0b0c0'}]
As you can see, the stored document for the person looks very much like a
natural JSON document. But oh no, I forgot to specify the full name for
Stephan. Let's do that:
>>> dm.root['stephan'].name = 'Stephan Richter'
>>> dm.root['stephan']._p_changed
True
This time, the data is not automatically saved:
>>> fetchone(person_cn)['data']['name']
'Stephan'
So we have to commit the transaction first:
>>> dm.root['stephan']._p_changed
True
>>> transaction.commit()
>>> dm.root['stephan']._p_changed
>>> fetchone(person_cn)['data']['name']
'Stephan Richter'
Let's now add an address for Stephan. Addresses are also persistent objects:
>>> class Address(persistent.Persistent):
... _p_pj_table = 'address'
...
... def __init__(self, city, zip):
... self.city = city
... self.zip = zip
...
... def __str__(self):
... return '%s (%s)' %(self.city, self.zip)
...
... def __repr__(self):
... return '<%s %s>' %(self.__class__.__name__, self)
pjpersist supports a special attribute called ``_p_pj_table``,
which allows you to specify a custom table to use.
>>> stephan = dm.root['stephan']
>>> stephan.address = Address('Maynard', '01754')
>>> stephan.address
<Address Maynard (01754)>
Note that the address is not immediately saved in the database:
>>> dumpTable('address', isolate=True)
relation "address" does not exist
...
But once we commit the transaction, everything is available:
>>> transaction.commit()
>>> dumpTable('address')
[{'data': {'_py_persistent_type': '__main__.Address',
'city': 'Maynard',
'zip': '01754'},
'id': '0001020304050607080a0b0c0'}]
>>> dumpTable(person_cn)
[{'data': {'_py_persistent_type': '__main__.Person',
'address': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'address'},
'birthday': None,
'friends': {},
'name': 'Stephan Richter',
'phone': None,
'today': {'_py_type': 'datetime.datetime',
'value': '2014-05-14T12:30:00.000000'},
'visited': []},
'id': '0001020304050607080a0b0c0'}]
>>> dm.root['stephan'].address
<Address Maynard (01754)>
Non-Persistent Objects
----------------------
As you can see, even the reference looks nice and all components are easily
visible. But what about arbitrary non-persistent, but picklable,
objects? Well, let's create a phone number object for that:
>>> class Phone(object):
...
... def __init__(self, country, area, number):
... self.country = country
... self.area = area
... self.number = number
...
... def __str__(self):
... return '%s-%s-%s' %(self.country, self.area, self.number)
...
... def __repr__(self):
... return '<%s %s>' %(self.__class__.__name__, self)
>>> dm.root['stephan'].phone = Phone('+1', '978', '394-5124')
>>> dm.root['stephan'].phone
<Phone +1-978-394-5124>
Let's now commit the transaction and look at the JSONB document again:
>>> transaction.commit()
>>> dm.root['stephan'].phone
<Phone +1-978-394-5124>
>>> dumpTable(person_cn)
[{'data': {'_py_persistent_type': '__main__.Person',
'address': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'address'},
'birthday': None,
'friends': {},
'name': 'Stephan Richter',
'phone': {'_py_type': '__main__.Phone',
'area': '978',
'country': '+1',
'number': '394-5124'},
'today': {'_py_type': 'datetime.datetime',
'value': '2014-05-14T12:30:00.000000'},
'visited': []},
'id': '0001020304050607080a0b0c0'}]
As you can see, for arbitrary non-persistent objects we need a small hint in
the sub-document, but it is very minimal. If the ``__reduce__`` method returns
a more complex construct, more meta-data is written. We will see that next
when storing a date and other arbitrary data:
>>> dm.root['stephan'].friends = {'roy': Person('Roy Mathew')}
>>> dm.root['stephan'].visited = ('Germany', 'USA')
>>> dm.root['stephan'].birthday = datetime.date(1980, 1, 25)
>>> transaction.commit()
>>> dm.root['stephan'].friends
{'roy': <Person Roy Mathew>}
>>> dm.root['stephan'].visited
['Germany', 'USA']
>>> dm.root['stephan'].birthday
datetime.date(1980, 1, 25)
As you can see, a dictionary key is always converted to unicode and tuples are
always maintained as lists, since JSON does not have two sequence types.
>>> import pprint
>>> pprint.pprint(dict(
... fetchone(person_cn, """data @> '{"name": "Stephan Richter"}'""")))
{'data': {'_py_persistent_type': '__main__.Person',
'address': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'address'},
'birthday': {'_py_type': 'datetime.date',
'value': '1980-01-25'},
'friends': {'roy': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'u__main___dot_Person'}},
'name': 'Stephan Richter',
'phone': {'_py_type': '__main__.Phone',
'area': '978',
'country': '+1',
'number': '394-5124'},
'today': {'_py_type': 'datetime.datetime',
'value': '2014-05-14T12:30:00.000000'},
'visited': ['Germany', 'USA']},
'id': '0001020304050607080a0b0c0'}
Custom Serializers
------------------
(A patch to demonstrate)
>>> dm.root['stephan'].birthday = datetime.date(1981, 1, 25)
>>> transaction.commit()
>>> pprint.pprint(
... fetchone(person_cn,
... """data @> '{"name": "Stephan Richter"}'""")['data']['birthday'])
{'_py_type': 'datetime.date', 'value': '1981-01-25'}
As you can see, the serialization of the birthay is an ISO string. We can,
however, provide a custom serializer that uses the ordinal to store the data.
>>> class DateSerializer(serialize.ObjectSerializer):
...
... def can_read(self, state):
... return isinstance(state, dict) and \
... state.get('_py_type') == 'custom_date'
...
... def read(self, state):
... return datetime.date.fromordinal(state['ordinal'])
...
... def can_write(self, obj):
... return isinstance(obj, datetime.date)
...
... def write(self, obj):
... return {'_py_type': 'custom_date',
... 'ordinal': obj.toordinal()}
>>> serialize.SERIALIZERS.append(DateSerializer())
>>> dm.root['stephan']._p_changed = True
>>> transaction.commit()
Let's have a look again:
>>> dm.root['stephan'].birthday
datetime.date(1981, 1, 25)
>>> pprint.pprint(dict(
... fetchone(person_cn, """data @> '{"name": "Stephan Richter"}'""")))
{'data': {'_py_persistent_type': '__main__.Person',
'address': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'address'},
'birthday': {'_py_type': 'custom_date', 'ordinal': 723205},
'friends': {'roy': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'u__main___dot_Person'}},
'name': 'Stephan Richter',
'phone': {'_py_type': '__main__.Phone',
'area': '978',
'country': '+1',
'number': '394-5124'},
'today': {'_py_type': 'custom_date', 'ordinal': 735367},
'visited': ['Germany', 'USA']},
'id': '0001020304050607080a0b0c0'}
Much better!
>>> del serialize.SERIALIZERS[:]
Persistent Objects as Sub-Documents
-----------------------------------
In order to give more control over which objects receive their own tables
and which do not, the developer can provide a special flag marking a
persistent class so that it becomes part of its parent object's document:
>>> class Car(persistent.Persistent):
... _p_pj_sub_object = True
...
... def __init__(self, year, make, model):
... self.year = year
... self.make = make
... self.model = model
...
... def __str__(self):
... return '%s %s %s' %(self.year, self.make, self.model)
...
... def __repr__(self):
... return '<%s %s>' %(self.__class__.__name__, self)
The ``_p_pj_sub_object`` is used to mark a type of object to be just part
of another document:
>>> dm.root['stephan'].car = car = Car('2005', 'Ford', 'Explorer')
>>> transaction.commit()
>>> dm.root['stephan'].car
<Car 2005 Ford Explorer>
>>> pprint.pprint(dict(
... fetchone(person_cn, """data @> '{"name": "Stephan Richter"}'""")))
{'data': {'_py_persistent_type': '__main__.Person',
'address': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'address'},
'birthday': {'_py_type': 'datetime.date',
'value': '1981-01-25'},
'car': {'_py_persistent_type': '__main__.Car',
'make': 'Ford',
'model': 'Explorer',
'year': '2005'},
'friends': {'roy': {'_py_type': 'DBREF',
'database': 'pjpersist_test',
'id': '0001020304050607080a0b0c0',
'table': 'u__main___dot_Person'}},
'name': 'Stephan Richter',
'phone': {'_py_type': '__main__.Phone',
'area': '978',
'country': '+1',
'number': '394-5124'},
'today': {'_py_type': 'datetime.date', 'value': '2014-05-14'},
'visited': ['Germany', 'USA']},
'id': '0001020304050607080a0b0c0'}
The reason we want objects to be persistent is so that they pick up changes
automatically:
>>> dm.root['stephan'].car.year = '2004'
>>> transaction.commit()
>>> dm.root['stephan'].car
<Car 2004 Ford Explorer>
Table Sharing
-------------
Since PostGreSQL/JSONB is so flexible, it sometimes makes sense to store
multiple types of (similar) objects in the same table. In those cases you
instruct the object type to store its Python path as part of the document.
Warning: Please note though that this method is less efficient, since the
document must be loaded in order to create a ghost causing more database
access.
>>> class ExtendedAddress(Address):
...
... def __init__(self, city, zip, country):
... super(ExtendedAddress, self).__init__(city, zip)
... self.country = country
...
... def __str__(self):
... return '%s (%s) in %s' %(self.city, self.zip, self.country)
In order to accomplish table sharing, you simply create another class
that has the same ``_p_pj_table`` string as another (sub-classing will
ensure that).
So let's give Stephan two extended addresses now.
>>> dm.root['stephan'].address2 = ExtendedAddress(
... 'Tettau', '01945', 'Germany')
>>> dm.root['stephan'].address2
<ExtendedAddress Tettau (01945) in Germany>
>>> dm.root['stephan'].address3 = ExtendedAddress(
... 'Arnsdorf', '01945', 'Germany')
>>> dm.root['stephan'].address3
<ExtendedAddress Arnsdorf (01945) in Germany>
>>> transaction.commit()
When loading the addresses, they should be of the right type:
>>> dm.root['stephan'].address
<Address Maynard (01754)>
>>> dm.root['stephan'].address2
<ExtendedAddress Tettau (01945) in Germany>
>>> dm.root['stephan'].address3
<ExtendedAddress Arnsdorf (01945) in Germany>
Persistent Serialization Hooks
------------------------------
When persistent components implement the ``IPersistentSerializationHooks``, it
is possible for the object to conduct some custom storage function.
>>> from pjpersist.persistent import PersistentSerializationHooks
>>> class Usernames(PersistentSerializationHooks):
... _p_pj_table = 'usernames'
... format = 'email'
...
... def _pj_after_store_hook(self, conn):
... print('After Store Hook')
...
... def _pj_after_load_hook(self, conn):
... print('After Load Hook')
When we store the object, the hook is called:
(actually twice, because this is a new object)
>>> dm.root['stephan'].usernames = Usernames()
>>> transaction.commit()
After Store Hook
After Store Hook
When loading, the same happens:
>>> dm.root['stephan'].usernames.format
After Load Hook
'email'
The store hook fires just once if the object is not new:
>>> dm.root['stephan'].usernames.format = 'snailmail'
>>> transaction.commit()
After Store Hook
Column Serialization
--------------------
pjpersist also allows for the object to specify values, usually attributes or
properties, to be stored as columns on the object's storage table.
Note that we support only a one-way transformation, because object state
will be always deserialized from the ``data`` jsonb field.
>>> import zope.schema
>>> class IPerson(zope.interface.Interface):
...
... name = zope.schema.TextLine(title='Name')
... address = zope.schema.TextLine(title='Address')
... visited = zope.schema.Datetime(title='Visited')
... phone = zope.schema.TextLine(title='Phone')
Initially, we are storing only the name in a column:
>>> from pjpersist.persistent import SimpleColumnSerialization, select_fields
>>> @zope.interface.implementer(IPerson)
... class ColumnPerson(SimpleColumnSerialization, Person):
... _p_pj_table = 'cperson'
... _pj_column_fields = select_fields(IPerson, 'name')
So once I create such a person and commit the transaction, the person table is
extended to store the attribute and the person is added to the table:
>>> dm.root['anton'] = anton = ColumnPerson('Anton')
>>> transaction.commit()
>>> dumpTable('cperson')
[{'data': {'_py_persistent_type': '__main__.ColumnPerson',
'address': None,
'birthday': None,
'friends': {},
'name': 'Anton',
'phone': None,
'today': {'_py_type': 'datetime.datetime',
'value': '2014-05-14T12:30:00.000000'},
'visited': []},
'id': '0001020304050607080a0b0c0',
'name': 'Anton'}]
Tricky Cases
------------
Changes in Basic Mutable Type
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Tricky, tricky. How do we make the framework detect changes in mutable
objects, such as lists and dictionaries? Answer: We keep track of which
persistent object they belong to and provide persistent implementations.
>>> type(dm.root['stephan'].friends)
<class 'pjpersist.serialize.PersistentDict'>
>>> dm.root['stephan'].friends['roger'] = Person('Roger')
>>> transaction.commit()
>>> sorted(dm.root['stephan'].friends.keys())
['roger', 'roy']
The same is true for lists:
>>> type(dm.root['stephan'].visited)
<class 'pjpersist.serialize.PersistentList'>
>>> dm.root['stephan'].visited.append('France')
>>> transaction.commit()
>>> dm.root['stephan'].visited
['Germany', 'USA', 'France']
Circular Non-Persistent References
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Any mutable object that is stored in a sub-document, cannot have multiple
references in the object tree, since there is no global referencing. These
circular references are detected and reported:
>>> class Top(persistent.Persistent):
... foo = None
>>> class Foo(object):
... bar = None
>>> class Bar(object):
... foo = None
>>> top = Top()
>>> foo = Foo()
>>> bar = Bar()
>>> top.foo = foo
>>> foo.bar = bar
>>> bar.foo = foo
>>> dm.root['top'] = top
Traceback (most recent call last):
...
CircularReferenceError: <...>
Circular Persistent References
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
In general, circular references among persistent objects are not a problem,
since we always only store a link to the object. However, there is a case when
the circular dependencies become a problem.
If you set up an object tree with circular references and then add the tree to
the storage at once, it must insert objects during serialization, so that
references can be created. However, care needs to be taken to only create a
minimal reference object, so that the system does not try to recursively
reduce the state.
>>> class PFoo(persistent.Persistent):
... bar = None
>>> class PBar(persistent.Persistent):
... foo = None
>>> top = Top()
>>> foo = PFoo()
>>> bar = PBar()
>>> top.foo = foo
>>> foo.bar = bar
>>> bar.foo = foo
>>> dm.root['ptop'] = top
Containers and Tables
---------------------
Now that we have talked so much about the gory details on storing one object,
what about mappings that reflect an entire table, for example a
table of people.
There are many approaches that can be taken. The following implementation
defines an attribute in the document as the mapping key and names a
table:
>>> from pjpersist import mapping
>>> class People(mapping.PJTableMapping):
... __pj_table__ = person_cn
... __pj_mapping_key__ = 'short_name'
The mapping takes the data manager as an argument. One can easily create a
sub-class that assigns the data manager automatically. Let's have a look:
>>> People(dm).keys()
[]
The reason no person is in the list yet, is because no document has the key
yet or the key is null. Let's change that:
>>> People(dm)['stephan'] = dm.root['stephan']
>>> transaction.commit()
>>> People(dm).keys()
['stephan']
>>> People(dm)['stephan']
<Person Stephan Richter>
Also note that setting the "short-name" attribute on any other person will add
it to the mapping:
>>> dm.root['stephan'].friends['roy'].short_name = 'roy'
>>> transaction.commit()
>>> sorted(People(dm).keys())
['roy', 'stephan']
=======
CHANGES
=======
3.1.4 (2024-03-27)
------------------
- Moving CI to github actions.
- Declare Python 3.11 compatibility
- Fix serialization for dates where year is shorter than 4 digits long
3.1.3 (2022-11-23)
------------------
- Faster PJDataManager flush() when it contains a lot of objects.
3.1.2 (2022-07-15)
------------------
- Replace collections with collections.abc for python 3.10 compatibility
3.1.1 (2022-06-06)
------------------
- Fix `MappingView` fix, do not emit `hint` in the log message, hint object is the one
being loaded, this just causes pain.
3.1.0 (2022-06-03)
------------------
- Fix `collections.abc.MappingView` subclass persistence. It was very bad, did not
store the underlying mapping at all, **caused silent data loss**.
Urgent update advised!
Also failed on loading the state of such.
3.0.2 (2022-05-03)
------------------
- Fix `DBRef` comparisons to return valid results instead
of failing hard when comparing to `None` and not `DBRef` instances.
(`__neq__` was unused because `__ne__` is the right method and `__ne__` anyway
delegates to `__eq__`)
3.0.1 (2022-02-03)
------------------
- Fix imports in `testing.py`
- Declare Python 3.9 compatibility
- Small improvement in `pjpersist.zope.container.PJContainer._load_one`
Get the local cache just once, since the `_cache` property became
a bit more expensive.
- Added `sqlbuilder.ILIKE` -- case-insensitive pair of LIKE
3.0.0 (2021-02-22)
------------------
- Backwards incompatible change: PJDataManager now accepts a pool instead
of connection object. PJDataManager will get the connection from the pool
when joining the transaction, and return it back when transaction
completes (aborts or commits). This allows for more flexible connection
management. The connection pool must implement IPJConnectionPool interface
(it is compatible with psycopg2.pool).
- `IPJDataManager.begin()` is renamed to `setTransactionOptions()`
- Errors executing SQL statements now doom the entire transaction,
causing `transaction.interfaces.DoomedTransaction` exception on
any attempts to commit it. A failed transaction must be aborted.
2.0.1 (2020-10-13)
------------------
- Fixed persisting tuple keyed dicts. Persisting such objects worked,
but reading failed.
2.0.0 (2020-06-02)
------------------
- Drop Python 2.7 and 3.6 support, add 3.8.
- Remove buildout support.
- Support for nested flushing. In complex use cases it can happen that during
serialization of an object, a query is made to look up another object. That
in turn causes a flush, resulting in a flush inside a flush. The `flush()`
method did not expect that behavior and failed if the inner flush would
flush objects that the outer flush had already handled.
1.7.2 (2020-02-10)
------------------
- Optimization: do not dig in `data` when we have a native field for
`_pj_mapping_key`, should allow creating indexes for lookup
1.7.1 (2019-06-19)
------------------
- Fixed an edge case when the serializer gets a mapping with a key `dict_data`.
Reading such object failed.
- Fixed an edge case with the serializer, when an object's state living
in a persistent object became 'empty'. Basically the state was just
`{'_py_persistent_type': 'SomeClass'}`
`SomeClass.__setstate__` was not called, thus the object could miss
attributes. Like a subclass of `UserDict` would miss the `data` attribute.
- Removed checking for 0x00 chars in dict keys. Turns out PostGreSQL just
can not store 0x00.
1.7.0 (2019-05-29)
------------------
- Support for sub-second datetime and time resolution during serialization.
- Add `use_cache` argument to `PJContainer._load_one()` to support ignoring
the cache. (This became handy if a container keeps track of multiple
versions of an item and you try to load all old revisions.)
1.6.0 (2019-05-29)
------------------
- Make `id` and `data` column name configurable via `_pj_id_column` and
`_pj_data_column` attributes in `PJContainer`, respectively.
- Auto-assign a name to objects when using `PJContainer`, not just
`IdNamesPJContainer`.
1.5.0 (2018-10-10)
------------------
- Support for Python 3.7. Removed Python 3.5 testing from tox.
1.4.1 (2018-09-13)
------------------
- No need to log in tpc_finish.
1.4.0 (2018-09-13)
------------------
- Implemented skipping tpc_prepare when DM has no writes.
We found out that AWS Aurora is dog slow at the moment on tpc_prepare.
When the DataManager has no writes, there's no need to call tpc_prepare.
See `CALL_TPC_PREPARE_ON_NO_WRITE_TRANSACTION`, by default True for backwards
compatibility.
- Added ability to log whether the transaction had writes.
See `LOG_READ_WRITE_TRANSACTION`, by default False
1.3.2 (2018-04-19)
------------------
- More precise flushing of datamanager to avoid unnecessary database
writes.
1.3.1 (2018-04-11)
------------------
- Enabled concurrent adds to IdNamesPJContainer by eliminating a query
that was causing transaction conflicts.
1.3.0 (2018-03-22)
------------------
- Python 3 compatibility fixes
- More efficient PJContainer.values() implementation
1.2.2 (2017-12-12)
------------------
- Need to protect all DB calls against `DatabaseDisconnected`
1.2.1 (2017-12-12)
------------------
- `psycopg2.OperationalError` and `psycopg2.InterfaceError` will be caught
on SQL command execution and reraised as `DatabaseDisconnected`
1.2.0 (2017-10-24)
------------------
- Added a new helper function to link subobject to main doc object. This is
needed when a custom `__getstate__()` and `__setstate__()` is implemented. A
detailed example is provided.
- Implemented `flush_hint` argument for `IDataManager.execute()` to allow
flushing only some objects during query. `flush_hints` is a list table names
that need to be flushed for the query to return a correct result.
- The Zope-specific containers use the `flush_hint` to only flush objects they
manage when a query is run on the container.
- While flushing objects, every main document object is now only flushed
once. Before that fix, any subobject would cause its doc object to be dumped
again.
Note: These optimizations provide a 15% performance improvements in real-world
applications.
1.1.2 (2017-09-14)
------------------
- Make sure changed objects aren't `_p_changed` anymore after commit.
1.1.1 (2017-07-03)
------------------
- Nothing changed yet.
1.0.0 (2017-03-18)
------------------
- Initial Public Release
- Project forked from mongopersist to work with PostGreSQL and JSONB data
type. The main motiviation is the ability to utilize PostGreSQL's great
transactional support.
Raw data
{
"_id": null,
"home_page": "https://github.com/Shoobx/pjpersist",
"name": "pjpersist",
"maintainer": null,
"docs_url": null,
"requires_python": null,
"maintainer_email": null,
"keywords": "postgres jsonb persistent",
"author": "Shoobx Team",
"author_email": "dev@shoobx.com",
"download_url": "https://files.pythonhosted.org/packages/f0/f0/6c93fa9d415bdb296903e40cdedf6ff20e3986dbbfd2f567ab74a7037edb/pjpersist-3.1.4.tar.gz",
"platform": null,
"description": "=================================\nPostGreSQL/JSONB Data Persistence\n=================================\n\nThis document outlines the general capabilities of the ``pjpersist``\npackage. ``pjpersist`` is a PostGreSQL/JSONB storage implementation for\npersistent Python objects. It is *not* a storage for the ZODB.\n\nThe goal of ``pjpersist`` is to provide a data manager that serializes\nobjects to JSONB blobs at transaction boundaries. The PJ data manager is a\npersistent data manager, which handles events at transaction boundaries (see\n``transaction.interfaces.IDataManager``) as well as events from the\npersistency framework (see ``persistent.interfaces.IPersistentDataManager``).\n\nAn instance of a data manager is supposed to have the same life time as the\ntransaction, meaning that it is assumed that you create a new data manager\nwhen creating a new transaction:\n\n >>> import transaction\n\nNote: The ``conn`` object is a ``psycopg.Connection`` instance. In this case\nour tests use the ``pjpersist_test`` database.\n\nLet's now define a simple persistent object:\n\n >>> import datetime\n >>> import persistent\n\n >>> class Person(persistent.Persistent):\n ...\n ... def __init__(self, name, phone=None, address=None, friends=None,\n ... visited=(), birthday=None):\n ... self.name = name\n ... self.address = address\n ... self.friends = friends or {}\n ... self.visited = visited\n ... self.phone = phone\n ... self.birthday = birthday\n ... self.today = datetime.datetime(2014, 5, 14, 12, 30)\n ...\n ... def __str__(self):\n ... return self.name\n ...\n ... def __repr__(self):\n ... return '<%s %s>' %(self.__class__.__name__, self)\n\nWe will fill out the other objects later. But for now, let's create a new\nperson and store it in PJ:\n\n >>> stephan = Person('Stephan')\n >>> stephan\n <Person Stephan>\n\nThe datamanager provides a ``root`` attribute in which the object tree roots\ncan be stored. It is special in the sense that it immediately writes the data\nto the DB:\n\n >>> dm.root['stephan'] = stephan\n >>> dm.root['stephan']\n <Person Stephan>\n\nCustom Persistence Tables\n-------------------------\n\nBy default, persistent objects are stored in a table having the escaped\nPython path of the class:\n\n >>> from pjpersist import serialize\n >>> person_cn = serialize.get_dotted_name(Person, True)\n >>> person_cn\n 'u__main___dot_Person'\n\n >>> transaction.commit()\n >>> dumpTable(person_cn)\n [{'data': {'_py_persistent_type': '__main__.Person',\n 'address': None,\n 'birthday': None,\n 'friends': {},\n 'name': 'Stephan',\n 'phone': None,\n 'today': {'_py_type': 'datetime.datetime',\n 'value': '2014-05-14T12:30:00.000000'},\n 'visited': []},\n 'id': '0001020304050607080a0b0c0'}]\n\n\nAs you can see, the stored document for the person looks very much like a\nnatural JSON document. But oh no, I forgot to specify the full name for\nStephan. Let's do that:\n\n >>> dm.root['stephan'].name = 'Stephan Richter'\n >>> dm.root['stephan']._p_changed\n True\n\nThis time, the data is not automatically saved:\n\n >>> fetchone(person_cn)['data']['name']\n 'Stephan'\n\nSo we have to commit the transaction first:\n\n >>> dm.root['stephan']._p_changed\n True\n >>> transaction.commit()\n >>> dm.root['stephan']._p_changed\n >>> fetchone(person_cn)['data']['name']\n 'Stephan Richter'\n\nLet's now add an address for Stephan. Addresses are also persistent objects:\n\n >>> class Address(persistent.Persistent):\n ... _p_pj_table = 'address'\n ...\n ... def __init__(self, city, zip):\n ... self.city = city\n ... self.zip = zip\n ...\n ... def __str__(self):\n ... return '%s (%s)' %(self.city, self.zip)\n ...\n ... def __repr__(self):\n ... return '<%s %s>' %(self.__class__.__name__, self)\n\npjpersist supports a special attribute called ``_p_pj_table``,\nwhich allows you to specify a custom table to use.\n\n >>> stephan = dm.root['stephan']\n >>> stephan.address = Address('Maynard', '01754')\n >>> stephan.address\n <Address Maynard (01754)>\n\nNote that the address is not immediately saved in the database:\n\n >>> dumpTable('address', isolate=True)\n relation \"address\" does not exist\n ...\n\nBut once we commit the transaction, everything is available:\n\n >>> transaction.commit()\n >>> dumpTable('address')\n [{'data': {'_py_persistent_type': '__main__.Address',\n 'city': 'Maynard',\n 'zip': '01754'},\n 'id': '0001020304050607080a0b0c0'}]\n\n >>> dumpTable(person_cn)\n [{'data': {'_py_persistent_type': '__main__.Person',\n 'address': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'address'},\n 'birthday': None,\n 'friends': {},\n 'name': 'Stephan Richter',\n 'phone': None,\n 'today': {'_py_type': 'datetime.datetime',\n 'value': '2014-05-14T12:30:00.000000'},\n 'visited': []},\n 'id': '0001020304050607080a0b0c0'}]\n\n >>> dm.root['stephan'].address\n <Address Maynard (01754)>\n\n\nNon-Persistent Objects\n----------------------\n\nAs you can see, even the reference looks nice and all components are easily\nvisible. But what about arbitrary non-persistent, but picklable,\nobjects? Well, let's create a phone number object for that:\n\n >>> class Phone(object):\n ...\n ... def __init__(self, country, area, number):\n ... self.country = country\n ... self.area = area\n ... self.number = number\n ...\n ... def __str__(self):\n ... return '%s-%s-%s' %(self.country, self.area, self.number)\n ...\n ... def __repr__(self):\n ... return '<%s %s>' %(self.__class__.__name__, self)\n\n >>> dm.root['stephan'].phone = Phone('+1', '978', '394-5124')\n >>> dm.root['stephan'].phone\n <Phone +1-978-394-5124>\n\nLet's now commit the transaction and look at the JSONB document again:\n\n >>> transaction.commit()\n >>> dm.root['stephan'].phone\n <Phone +1-978-394-5124>\n\n >>> dumpTable(person_cn)\n [{'data': {'_py_persistent_type': '__main__.Person',\n 'address': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'address'},\n 'birthday': None,\n 'friends': {},\n 'name': 'Stephan Richter',\n 'phone': {'_py_type': '__main__.Phone',\n 'area': '978',\n 'country': '+1',\n 'number': '394-5124'},\n 'today': {'_py_type': 'datetime.datetime',\n 'value': '2014-05-14T12:30:00.000000'},\n 'visited': []},\n 'id': '0001020304050607080a0b0c0'}]\n\nAs you can see, for arbitrary non-persistent objects we need a small hint in\nthe sub-document, but it is very minimal. If the ``__reduce__`` method returns\na more complex construct, more meta-data is written. We will see that next\nwhen storing a date and other arbitrary data:\n\n >>> dm.root['stephan'].friends = {'roy': Person('Roy Mathew')}\n >>> dm.root['stephan'].visited = ('Germany', 'USA')\n >>> dm.root['stephan'].birthday = datetime.date(1980, 1, 25)\n\n >>> transaction.commit()\n >>> dm.root['stephan'].friends\n {'roy': <Person Roy Mathew>}\n >>> dm.root['stephan'].visited\n ['Germany', 'USA']\n >>> dm.root['stephan'].birthday\n datetime.date(1980, 1, 25)\n\nAs you can see, a dictionary key is always converted to unicode and tuples are\nalways maintained as lists, since JSON does not have two sequence types.\n\n >>> import pprint\n >>> pprint.pprint(dict(\n ... fetchone(person_cn, \"\"\"data @> '{\"name\": \"Stephan Richter\"}'\"\"\")))\n {'data': {'_py_persistent_type': '__main__.Person',\n 'address': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'address'},\n 'birthday': {'_py_type': 'datetime.date',\n 'value': '1980-01-25'},\n 'friends': {'roy': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'u__main___dot_Person'}},\n 'name': 'Stephan Richter',\n 'phone': {'_py_type': '__main__.Phone',\n 'area': '978',\n 'country': '+1',\n 'number': '394-5124'},\n 'today': {'_py_type': 'datetime.datetime',\n 'value': '2014-05-14T12:30:00.000000'},\n 'visited': ['Germany', 'USA']},\n 'id': '0001020304050607080a0b0c0'}\n\n\nCustom Serializers\n------------------\n\n(A patch to demonstrate)\n\n >>> dm.root['stephan'].birthday = datetime.date(1981, 1, 25)\n >>> transaction.commit()\n\n >>> pprint.pprint(\n ... fetchone(person_cn,\n ... \"\"\"data @> '{\"name\": \"Stephan Richter\"}'\"\"\")['data']['birthday'])\n {'_py_type': 'datetime.date', 'value': '1981-01-25'}\n\nAs you can see, the serialization of the birthay is an ISO string. We can,\nhowever, provide a custom serializer that uses the ordinal to store the data.\n\n >>> class DateSerializer(serialize.ObjectSerializer):\n ...\n ... def can_read(self, state):\n ... return isinstance(state, dict) and \\\n ... state.get('_py_type') == 'custom_date'\n ...\n ... def read(self, state):\n ... return datetime.date.fromordinal(state['ordinal'])\n ...\n ... def can_write(self, obj):\n ... return isinstance(obj, datetime.date)\n ...\n ... def write(self, obj):\n ... return {'_py_type': 'custom_date',\n ... 'ordinal': obj.toordinal()}\n\n >>> serialize.SERIALIZERS.append(DateSerializer())\n >>> dm.root['stephan']._p_changed = True\n >>> transaction.commit()\n\nLet's have a look again:\n\n >>> dm.root['stephan'].birthday\n datetime.date(1981, 1, 25)\n\n >>> pprint.pprint(dict(\n ... fetchone(person_cn, \"\"\"data @> '{\"name\": \"Stephan Richter\"}'\"\"\")))\n {'data': {'_py_persistent_type': '__main__.Person',\n 'address': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'address'},\n 'birthday': {'_py_type': 'custom_date', 'ordinal': 723205},\n 'friends': {'roy': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'u__main___dot_Person'}},\n 'name': 'Stephan Richter',\n 'phone': {'_py_type': '__main__.Phone',\n 'area': '978',\n 'country': '+1',\n 'number': '394-5124'},\n 'today': {'_py_type': 'custom_date', 'ordinal': 735367},\n 'visited': ['Germany', 'USA']},\n 'id': '0001020304050607080a0b0c0'}\n\nMuch better!\n\n >>> del serialize.SERIALIZERS[:]\n\n\nPersistent Objects as Sub-Documents\n-----------------------------------\n\nIn order to give more control over which objects receive their own tables\nand which do not, the developer can provide a special flag marking a\npersistent class so that it becomes part of its parent object's document:\n\n >>> class Car(persistent.Persistent):\n ... _p_pj_sub_object = True\n ...\n ... def __init__(self, year, make, model):\n ... self.year = year\n ... self.make = make\n ... self.model = model\n ...\n ... def __str__(self):\n ... return '%s %s %s' %(self.year, self.make, self.model)\n ...\n ... def __repr__(self):\n ... return '<%s %s>' %(self.__class__.__name__, self)\n\nThe ``_p_pj_sub_object`` is used to mark a type of object to be just part\nof another document:\n\n >>> dm.root['stephan'].car = car = Car('2005', 'Ford', 'Explorer')\n >>> transaction.commit()\n\n >>> dm.root['stephan'].car\n <Car 2005 Ford Explorer>\n\n >>> pprint.pprint(dict(\n ... fetchone(person_cn, \"\"\"data @> '{\"name\": \"Stephan Richter\"}'\"\"\")))\n {'data': {'_py_persistent_type': '__main__.Person',\n 'address': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'address'},\n 'birthday': {'_py_type': 'datetime.date',\n 'value': '1981-01-25'},\n 'car': {'_py_persistent_type': '__main__.Car',\n 'make': 'Ford',\n 'model': 'Explorer',\n 'year': '2005'},\n 'friends': {'roy': {'_py_type': 'DBREF',\n 'database': 'pjpersist_test',\n 'id': '0001020304050607080a0b0c0',\n 'table': 'u__main___dot_Person'}},\n 'name': 'Stephan Richter',\n 'phone': {'_py_type': '__main__.Phone',\n 'area': '978',\n 'country': '+1',\n 'number': '394-5124'},\n 'today': {'_py_type': 'datetime.date', 'value': '2014-05-14'},\n 'visited': ['Germany', 'USA']},\n 'id': '0001020304050607080a0b0c0'}\n\n\nThe reason we want objects to be persistent is so that they pick up changes\nautomatically:\n\n >>> dm.root['stephan'].car.year = '2004'\n >>> transaction.commit()\n >>> dm.root['stephan'].car\n <Car 2004 Ford Explorer>\n\n\nTable Sharing\n-------------\n\nSince PostGreSQL/JSONB is so flexible, it sometimes makes sense to store\nmultiple types of (similar) objects in the same table. In those cases you\ninstruct the object type to store its Python path as part of the document.\n\nWarning: Please note though that this method is less efficient, since the\ndocument must be loaded in order to create a ghost causing more database\naccess.\n\n >>> class ExtendedAddress(Address):\n ...\n ... def __init__(self, city, zip, country):\n ... super(ExtendedAddress, self).__init__(city, zip)\n ... self.country = country\n ...\n ... def __str__(self):\n ... return '%s (%s) in %s' %(self.city, self.zip, self.country)\n\nIn order to accomplish table sharing, you simply create another class\nthat has the same ``_p_pj_table`` string as another (sub-classing will\nensure that).\n\nSo let's give Stephan two extended addresses now.\n\n >>> dm.root['stephan'].address2 = ExtendedAddress(\n ... 'Tettau', '01945', 'Germany')\n >>> dm.root['stephan'].address2\n <ExtendedAddress Tettau (01945) in Germany>\n\n >>> dm.root['stephan'].address3 = ExtendedAddress(\n ... 'Arnsdorf', '01945', 'Germany')\n >>> dm.root['stephan'].address3\n <ExtendedAddress Arnsdorf (01945) in Germany>\n\n >>> transaction.commit()\n\nWhen loading the addresses, they should be of the right type:\n\n >>> dm.root['stephan'].address\n <Address Maynard (01754)>\n >>> dm.root['stephan'].address2\n <ExtendedAddress Tettau (01945) in Germany>\n >>> dm.root['stephan'].address3\n <ExtendedAddress Arnsdorf (01945) in Germany>\n\n\nPersistent Serialization Hooks\n------------------------------\n\nWhen persistent components implement the ``IPersistentSerializationHooks``, it\nis possible for the object to conduct some custom storage function.\n\n\n >>> from pjpersist.persistent import PersistentSerializationHooks\n >>> class Usernames(PersistentSerializationHooks):\n ... _p_pj_table = 'usernames'\n ... format = 'email'\n ...\n ... def _pj_after_store_hook(self, conn):\n ... print('After Store Hook')\n ...\n ... def _pj_after_load_hook(self, conn):\n ... print('After Load Hook')\n\nWhen we store the object, the hook is called:\n(actually twice, because this is a new object)\n\n >>> dm.root['stephan'].usernames = Usernames()\n >>> transaction.commit()\n After Store Hook\n After Store Hook\n\nWhen loading, the same happens:\n\n >>> dm.root['stephan'].usernames.format\n After Load Hook\n 'email'\n\nThe store hook fires just once if the object is not new:\n\n >>> dm.root['stephan'].usernames.format = 'snailmail'\n >>> transaction.commit()\n After Store Hook\n\n\nColumn Serialization\n--------------------\n\npjpersist also allows for the object to specify values, usually attributes or\nproperties, to be stored as columns on the object's storage table.\n\nNote that we support only a one-way transformation, because object state\nwill be always deserialized from the ``data`` jsonb field.\n\n >>> import zope.schema\n >>> class IPerson(zope.interface.Interface):\n ...\n ... name = zope.schema.TextLine(title='Name')\n ... address = zope.schema.TextLine(title='Address')\n ... visited = zope.schema.Datetime(title='Visited')\n ... phone = zope.schema.TextLine(title='Phone')\n\nInitially, we are storing only the name in a column:\n\n >>> from pjpersist.persistent import SimpleColumnSerialization, select_fields\n >>> @zope.interface.implementer(IPerson)\n ... class ColumnPerson(SimpleColumnSerialization, Person):\n ... _p_pj_table = 'cperson'\n ... _pj_column_fields = select_fields(IPerson, 'name')\n\nSo once I create such a person and commit the transaction, the person table is\nextended to store the attribute and the person is added to the table:\n\n >>> dm.root['anton'] = anton = ColumnPerson('Anton')\n >>> transaction.commit()\n\n >>> dumpTable('cperson')\n [{'data': {'_py_persistent_type': '__main__.ColumnPerson',\n 'address': None,\n 'birthday': None,\n 'friends': {},\n 'name': 'Anton',\n 'phone': None,\n 'today': {'_py_type': 'datetime.datetime',\n 'value': '2014-05-14T12:30:00.000000'},\n 'visited': []},\n 'id': '0001020304050607080a0b0c0',\n 'name': 'Anton'}]\n\n\nTricky Cases\n------------\n\nChanges in Basic Mutable Type\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\nTricky, tricky. How do we make the framework detect changes in mutable\nobjects, such as lists and dictionaries? Answer: We keep track of which\npersistent object they belong to and provide persistent implementations.\n\n >>> type(dm.root['stephan'].friends)\n <class 'pjpersist.serialize.PersistentDict'>\n\n >>> dm.root['stephan'].friends['roger'] = Person('Roger')\n >>> transaction.commit()\n >>> sorted(dm.root['stephan'].friends.keys())\n ['roger', 'roy']\n\nThe same is true for lists:\n\n >>> type(dm.root['stephan'].visited)\n <class 'pjpersist.serialize.PersistentList'>\n\n >>> dm.root['stephan'].visited.append('France')\n >>> transaction.commit()\n >>> dm.root['stephan'].visited\n ['Germany', 'USA', 'France']\n\n\nCircular Non-Persistent References\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\nAny mutable object that is stored in a sub-document, cannot have multiple\nreferences in the object tree, since there is no global referencing. These\ncircular references are detected and reported:\n\n >>> class Top(persistent.Persistent):\n ... foo = None\n\n >>> class Foo(object):\n ... bar = None\n\n >>> class Bar(object):\n ... foo = None\n\n >>> top = Top()\n >>> foo = Foo()\n >>> bar = Bar()\n >>> top.foo = foo\n >>> foo.bar = bar\n >>> bar.foo = foo\n\n >>> dm.root['top'] = top\n Traceback (most recent call last):\n ...\n CircularReferenceError: <...>\n\n\nCircular Persistent References\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\nIn general, circular references among persistent objects are not a problem,\nsince we always only store a link to the object. However, there is a case when\nthe circular dependencies become a problem.\n\nIf you set up an object tree with circular references and then add the tree to\nthe storage at once, it must insert objects during serialization, so that\nreferences can be created. However, care needs to be taken to only create a\nminimal reference object, so that the system does not try to recursively\nreduce the state.\n\n >>> class PFoo(persistent.Persistent):\n ... bar = None\n\n >>> class PBar(persistent.Persistent):\n ... foo = None\n\n >>> top = Top()\n >>> foo = PFoo()\n >>> bar = PBar()\n >>> top.foo = foo\n >>> foo.bar = bar\n >>> bar.foo = foo\n\n >>> dm.root['ptop'] = top\n\n\nContainers and Tables\n---------------------\n\nNow that we have talked so much about the gory details on storing one object,\nwhat about mappings that reflect an entire table, for example a\ntable of people.\n\nThere are many approaches that can be taken. The following implementation\ndefines an attribute in the document as the mapping key and names a\ntable:\n\n >>> from pjpersist import mapping\n >>> class People(mapping.PJTableMapping):\n ... __pj_table__ = person_cn\n ... __pj_mapping_key__ = 'short_name'\n\nThe mapping takes the data manager as an argument. One can easily create a\nsub-class that assigns the data manager automatically. Let's have a look:\n\n >>> People(dm).keys()\n []\n\nThe reason no person is in the list yet, is because no document has the key\nyet or the key is null. Let's change that:\n\n >>> People(dm)['stephan'] = dm.root['stephan']\n >>> transaction.commit()\n\n >>> People(dm).keys()\n ['stephan']\n >>> People(dm)['stephan']\n <Person Stephan Richter>\n\nAlso note that setting the \"short-name\" attribute on any other person will add\nit to the mapping:\n\n >>> dm.root['stephan'].friends['roy'].short_name = 'roy'\n >>> transaction.commit()\n >>> sorted(People(dm).keys())\n ['roy', 'stephan']\n\n\n=======\nCHANGES\n=======\n\n\n3.1.4 (2024-03-27)\n------------------\n\n- Moving CI to github actions.\n- Declare Python 3.11 compatibility\n- Fix serialization for dates where year is shorter than 4 digits long\n\n\n3.1.3 (2022-11-23)\n------------------\n\n- Faster PJDataManager flush() when it contains a lot of objects.\n\n\n3.1.2 (2022-07-15)\n------------------\n\n- Replace collections with collections.abc for python 3.10 compatibility\n\n\n3.1.1 (2022-06-06)\n------------------\n\n- Fix `MappingView` fix, do not emit `hint` in the log message, hint object is the one\n being loaded, this just causes pain.\n\n\n3.1.0 (2022-06-03)\n------------------\n\n- Fix `collections.abc.MappingView` subclass persistence. It was very bad, did not\n store the underlying mapping at all, **caused silent data loss**.\n Urgent update advised!\n Also failed on loading the state of such.\n\n\n3.0.2 (2022-05-03)\n------------------\n\n- Fix `DBRef` comparisons to return valid results instead\n of failing hard when comparing to `None` and not `DBRef` instances.\n (`__neq__` was unused because `__ne__` is the right method and `__ne__` anyway\n delegates to `__eq__`)\n\n3.0.1 (2022-02-03)\n------------------\n\n- Fix imports in `testing.py`\n\n- Declare Python 3.9 compatibility\n\n- Small improvement in `pjpersist.zope.container.PJContainer._load_one`\n Get the local cache just once, since the `_cache` property became\n a bit more expensive.\n\n- Added `sqlbuilder.ILIKE` -- case-insensitive pair of LIKE\n\n\n3.0.0 (2021-02-22)\n------------------\n\n- Backwards incompatible change: PJDataManager now accepts a pool instead\n of connection object. PJDataManager will get the connection from the pool\n when joining the transaction, and return it back when transaction\n completes (aborts or commits). This allows for more flexible connection\n management. The connection pool must implement IPJConnectionPool interface\n (it is compatible with psycopg2.pool).\n\n- `IPJDataManager.begin()` is renamed to `setTransactionOptions()`\n\n- Errors executing SQL statements now doom the entire transaction,\n causing `transaction.interfaces.DoomedTransaction` exception on\n any attempts to commit it. A failed transaction must be aborted.\n\n\n2.0.1 (2020-10-13)\n------------------\n\n- Fixed persisting tuple keyed dicts. Persisting such objects worked,\n but reading failed.\n\n\n2.0.0 (2020-06-02)\n------------------\n\n- Drop Python 2.7 and 3.6 support, add 3.8.\n\n- Remove buildout support.\n\n- Support for nested flushing. In complex use cases it can happen that during\n serialization of an object, a query is made to look up another object. That\n in turn causes a flush, resulting in a flush inside a flush. The `flush()`\n method did not expect that behavior and failed if the inner flush would\n flush objects that the outer flush had already handled.\n\n\n1.7.2 (2020-02-10)\n------------------\n\n- Optimization: do not dig in `data` when we have a native field for\n `_pj_mapping_key`, should allow creating indexes for lookup\n\n\n1.7.1 (2019-06-19)\n------------------\n\n- Fixed an edge case when the serializer gets a mapping with a key `dict_data`.\n Reading such object failed.\n\n- Fixed an edge case with the serializer, when an object's state living\n in a persistent object became 'empty'. Basically the state was just\n `{'_py_persistent_type': 'SomeClass'}`\n `SomeClass.__setstate__` was not called, thus the object could miss\n attributes. Like a subclass of `UserDict` would miss the `data` attribute.\n\n- Removed checking for 0x00 chars in dict keys. Turns out PostGreSQL just\n can not store 0x00.\n\n1.7.0 (2019-05-29)\n------------------\n\n- Support for sub-second datetime and time resolution during serialization.\n\n- Add `use_cache` argument to `PJContainer._load_one()` to support ignoring\n the cache. (This became handy if a container keeps track of multiple\n versions of an item and you try to load all old revisions.)\n\n\n1.6.0 (2019-05-29)\n------------------\n\n- Make `id` and `data` column name configurable via `_pj_id_column` and\n `_pj_data_column` attributes in `PJContainer`, respectively.\n\n- Auto-assign a name to objects when using `PJContainer`, not just\n `IdNamesPJContainer`.\n\n\n1.5.0 (2018-10-10)\n------------------\n\n- Support for Python 3.7. Removed Python 3.5 testing from tox.\n\n\n1.4.1 (2018-09-13)\n------------------\n\n- No need to log in tpc_finish.\n\n\n1.4.0 (2018-09-13)\n------------------\n\n- Implemented skipping tpc_prepare when DM has no writes.\n We found out that AWS Aurora is dog slow at the moment on tpc_prepare.\n When the DataManager has no writes, there's no need to call tpc_prepare.\n See `CALL_TPC_PREPARE_ON_NO_WRITE_TRANSACTION`, by default True for backwards\n compatibility.\n\n- Added ability to log whether the transaction had writes.\n See `LOG_READ_WRITE_TRANSACTION`, by default False\n\n\n1.3.2 (2018-04-19)\n------------------\n\n- More precise flushing of datamanager to avoid unnecessary database\n writes.\n\n\n1.3.1 (2018-04-11)\n------------------\n\n- Enabled concurrent adds to IdNamesPJContainer by eliminating a query\n that was causing transaction conflicts.\n\n1.3.0 (2018-03-22)\n------------------\n\n- Python 3 compatibility fixes\n- More efficient PJContainer.values() implementation\n\n\n1.2.2 (2017-12-12)\n------------------\n\n- Need to protect all DB calls against `DatabaseDisconnected`\n\n\n1.2.1 (2017-12-12)\n------------------\n\n- `psycopg2.OperationalError` and `psycopg2.InterfaceError` will be caught\n on SQL command execution and reraised as `DatabaseDisconnected`\n\n\n1.2.0 (2017-10-24)\n------------------\n\n- Added a new helper function to link subobject to main doc object. This is\n needed when a custom `__getstate__()` and `__setstate__()` is implemented. A\n detailed example is provided.\n\n- Implemented `flush_hint` argument for `IDataManager.execute()` to allow\n flushing only some objects during query. `flush_hints` is a list table names\n that need to be flushed for the query to return a correct result.\n\n- The Zope-specific containers use the `flush_hint` to only flush objects they\n manage when a query is run on the container.\n\n- While flushing objects, every main document object is now only flushed\n once. Before that fix, any subobject would cause its doc object to be dumped\n again.\n\nNote: These optimizations provide a 15% performance improvements in real-world\napplications.\n\n\n1.1.2 (2017-09-14)\n------------------\n\n- Make sure changed objects aren't `_p_changed` anymore after commit.\n\n\n1.1.1 (2017-07-03)\n------------------\n\n- Nothing changed yet.\n\n\n1.0.0 (2017-03-18)\n------------------\n\n- Initial Public Release\n\n- Project forked from mongopersist to work with PostGreSQL and JSONB data\n type. The main motiviation is the ability to utilize PostGreSQL's great\n transactional support.\n\n",
"bugtrack_url": null,
"license": "ZPL 2.1",
"summary": "PostgreSQL/JSONB Persistence Backend",
"version": "3.1.4",
"project_urls": {
"Homepage": "https://github.com/Shoobx/pjpersist"
},
"split_keywords": [
"postgres",
"jsonb",
"persistent"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "f0f06c93fa9d415bdb296903e40cdedf6ff20e3986dbbfd2f567ab74a7037edb",
"md5": "b00b75e97de64a359a58d7e4fe3d9124",
"sha256": "0cda09009de89d90bf437d7a5df710f9b8d71869b26fa0e6f339b2db9ac067d8"
},
"downloads": -1,
"filename": "pjpersist-3.1.4.tar.gz",
"has_sig": false,
"md5_digest": "b00b75e97de64a359a58d7e4fe3d9124",
"packagetype": "sdist",
"python_version": "source",
"requires_python": null,
"size": 122982,
"upload_time": "2024-03-27T15:59:28",
"upload_time_iso_8601": "2024-03-27T15:59:28.429498Z",
"url": "https://files.pythonhosted.org/packages/f0/f0/6c93fa9d415bdb296903e40cdedf6ff20e3986dbbfd2f567ab74a7037edb/pjpersist-3.1.4.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-03-27 15:59:28",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "Shoobx",
"github_project": "pjpersist",
"travis_ci": false,
"coveralls": true,
"github_actions": true,
"tox": true,
"lcname": "pjpersist"
}