| #include "Python.h" |
| |
| #include "pycore_dict.h" // _PyDict_UniqueId() |
| #include "pycore_lock.h" // PyMutex_LockFlags() |
| #include "pycore_pystate.h" // _PyThreadState_GET() |
| #include "pycore_object.h" // _Py_IncRefTotal |
| #include "pycore_uniqueid.h" |
| |
| // This contains code for allocating unique ids for per-thread reference |
| // counting and re-using those ids when an object is deallocated. |
| // |
| // Currently, per-thread reference counting is only used for heap types. |
| // |
| // See Include/internal/pycore_uniqueid.h for more details. |
| |
| #ifdef Py_GIL_DISABLED |
| |
| #define POOL_MIN_SIZE 8 |
| |
| #define LOCK_POOL(pool) PyMutex_LockFlags(&pool->mutex, _Py_LOCK_DONT_DETACH) |
| #define UNLOCK_POOL(pool) PyMutex_Unlock(&pool->mutex) |
| |
| static int |
| resize_interp_type_id_pool(struct _Py_unique_id_pool *pool) |
| { |
| if ((size_t)pool->size > PY_SSIZE_T_MAX / (2 * sizeof(*pool->table))) { |
| return -1; |
| } |
| |
| Py_ssize_t new_size = pool->size * 2; |
| if (new_size < POOL_MIN_SIZE) { |
| new_size = POOL_MIN_SIZE; |
| } |
| |
| _Py_unique_id_entry *table = PyMem_Realloc(pool->table, |
| new_size * sizeof(*pool->table)); |
| if (table == NULL) { |
| return -1; |
| } |
| |
| Py_ssize_t start = pool->size; |
| for (Py_ssize_t i = start; i < new_size - 1; i++) { |
| table[i].next = &table[i + 1]; |
| } |
| table[new_size - 1].next = NULL; |
| |
| pool->table = table; |
| pool->freelist = &table[start]; |
| _Py_atomic_store_ssize(&pool->size, new_size); |
| return 0; |
| } |
| |
| static int |
| resize_local_refcounts(_PyThreadStateImpl *tstate) |
| { |
| if (tstate->refcounts.is_finalized) { |
| return -1; |
| } |
| |
| struct _Py_unique_id_pool *pool = &tstate->base.interp->unique_ids; |
| Py_ssize_t size = _Py_atomic_load_ssize(&pool->size); |
| |
| Py_ssize_t *refcnts = PyMem_Realloc(tstate->refcounts.values, |
| size * sizeof(Py_ssize_t)); |
| if (refcnts == NULL) { |
| return -1; |
| } |
| |
| Py_ssize_t old_size = tstate->refcounts.size; |
| if (old_size < size) { |
| memset(refcnts + old_size, 0, (size - old_size) * sizeof(Py_ssize_t)); |
| } |
| |
| tstate->refcounts.values = refcnts; |
| tstate->refcounts.size = size; |
| return 0; |
| } |
| |
| Py_ssize_t |
| _PyObject_AssignUniqueId(PyObject *obj) |
| { |
| PyInterpreterState *interp = _PyInterpreterState_GET(); |
| struct _Py_unique_id_pool *pool = &interp->unique_ids; |
| |
| LOCK_POOL(pool); |
| if (pool->freelist == NULL) { |
| if (resize_interp_type_id_pool(pool) < 0) { |
| UNLOCK_POOL(pool); |
| return _Py_INVALID_UNIQUE_ID; |
| } |
| } |
| |
| _Py_unique_id_entry *entry = pool->freelist; |
| pool->freelist = entry->next; |
| entry->obj = obj; |
| _PyObject_SetDeferredRefcount(obj); |
| // The unique id is one plus the index of the entry in the table. |
| Py_ssize_t unique_id = (entry - pool->table) + 1; |
| assert(unique_id > 0); |
| UNLOCK_POOL(pool); |
| return unique_id; |
| } |
| |
| void |
| _PyObject_ReleaseUniqueId(Py_ssize_t unique_id) |
| { |
| PyInterpreterState *interp = _PyInterpreterState_GET(); |
| struct _Py_unique_id_pool *pool = &interp->unique_ids; |
| |
| LOCK_POOL(pool); |
| assert(unique_id > 0 && unique_id <= pool->size); |
| Py_ssize_t idx = unique_id - 1; |
| _Py_unique_id_entry *entry = &pool->table[idx]; |
| entry->next = pool->freelist; |
| pool->freelist = entry; |
| UNLOCK_POOL(pool); |
| } |
| |
| static Py_ssize_t |
| clear_unique_id(PyObject *obj) |
| { |
| Py_ssize_t id = _Py_INVALID_UNIQUE_ID; |
| if (PyType_Check(obj)) { |
| if (PyType_HasFeature((PyTypeObject *)obj, Py_TPFLAGS_HEAPTYPE)) { |
| PyHeapTypeObject *ht = (PyHeapTypeObject *)obj; |
| id = ht->unique_id; |
| ht->unique_id = _Py_INVALID_UNIQUE_ID; |
| } |
| } |
| else if (PyCode_Check(obj)) { |
| PyCodeObject *co = (PyCodeObject *)obj; |
| id = co->_co_unique_id; |
| co->_co_unique_id = _Py_INVALID_UNIQUE_ID; |
| } |
| else if (PyDict_Check(obj)) { |
| PyDictObject *mp = (PyDictObject *)obj; |
| id = _PyDict_UniqueId(mp); |
| mp->_ma_watcher_tag &= ~(UINT64_MAX << DICT_UNIQUE_ID_SHIFT); |
| } |
| return id; |
| } |
| |
| void |
| _PyObject_DisablePerThreadRefcounting(PyObject *obj) |
| { |
| Py_ssize_t id = clear_unique_id(obj); |
| if (id != _Py_INVALID_UNIQUE_ID) { |
| _PyObject_ReleaseUniqueId(id); |
| } |
| } |
| |
| void |
| _PyObject_ThreadIncrefSlow(PyObject *obj, size_t idx) |
| { |
| _PyThreadStateImpl *tstate = (_PyThreadStateImpl *)_PyThreadState_GET(); |
| if (((Py_ssize_t)idx) < 0 || resize_local_refcounts(tstate) < 0) { |
| // just incref the object directly. |
| Py_INCREF(obj); |
| return; |
| } |
| |
| assert(idx < (size_t)tstate->refcounts.size); |
| tstate->refcounts.values[idx]++; |
| #ifdef Py_REF_DEBUG |
| _Py_IncRefTotal((PyThreadState *)tstate); |
| #endif |
| _Py_INCREF_STAT_INC(); |
| } |
| |
| void |
| _PyObject_MergePerThreadRefcounts(_PyThreadStateImpl *tstate) |
| { |
| if (tstate->refcounts.values == NULL) { |
| return; |
| } |
| |
| struct _Py_unique_id_pool *pool = &tstate->base.interp->unique_ids; |
| |
| LOCK_POOL(pool); |
| for (Py_ssize_t i = 0, n = tstate->refcounts.size; i < n; i++) { |
| Py_ssize_t refcnt = tstate->refcounts.values[i]; |
| if (refcnt != 0) { |
| PyObject *obj = pool->table[i].obj; |
| _Py_atomic_add_ssize(&obj->ob_ref_shared, |
| refcnt << _Py_REF_SHARED_SHIFT); |
| tstate->refcounts.values[i] = 0; |
| } |
| } |
| UNLOCK_POOL(pool); |
| } |
| |
| void |
| _PyObject_FinalizePerThreadRefcounts(_PyThreadStateImpl *tstate) |
| { |
| _PyObject_MergePerThreadRefcounts(tstate); |
| |
| PyMem_Free(tstate->refcounts.values); |
| tstate->refcounts.values = NULL; |
| tstate->refcounts.size = 0; |
| tstate->refcounts.is_finalized = 1; |
| } |
| |
| void |
| _PyObject_FinalizeUniqueIdPool(PyInterpreterState *interp) |
| { |
| struct _Py_unique_id_pool *pool = &interp->unique_ids; |
| |
| // First, set the free-list to NULL values |
| while (pool->freelist) { |
| _Py_unique_id_entry *next = pool->freelist->next; |
| pool->freelist->obj = NULL; |
| pool->freelist = next; |
| } |
| |
| // Now everything non-NULL is a object. Clear their unique ids as the |
| // object outlives the interpreter. |
| for (Py_ssize_t i = 0; i < pool->size; i++) { |
| PyObject *obj = pool->table[i].obj; |
| pool->table[i].obj = NULL; |
| if (obj != NULL) { |
| Py_ssize_t id = clear_unique_id(obj); |
| (void)id; |
| assert(id == i + 1); |
| } |
| } |
| PyMem_Free(pool->table); |
| pool->table = NULL; |
| pool->freelist = NULL; |
| pool->size = 0; |
| } |
| |
| #endif /* Py_GIL_DISABLED */ |