2014-01-13 19:52:22 +01:00
|
|
|
Database interactions
|
|
|
|
*********************
|
|
|
|
|
|
|
|
Database object
|
|
|
|
===============
|
|
|
|
|
|
|
|
.. py:class:: rocksdb.DB
|
|
|
|
|
|
|
|
.. py:method:: __init__(db_name, Options opts, read_only=False)
|
|
|
|
|
2014-01-16 08:51:01 +01:00
|
|
|
:param unicode db_name: Name of the database to open
|
2014-01-13 19:52:22 +01:00
|
|
|
:param opts: Options for this specific database
|
|
|
|
:type opts: :py:class:`rocksdb.Options`
|
|
|
|
:param bool read_only: If ``True`` the database is opened read-only.
|
|
|
|
All DB calls which modify data will raise an
|
|
|
|
Exception.
|
|
|
|
|
|
|
|
|
|
|
|
.. py:method:: put(key, value, sync=False, disable_wal=False)
|
|
|
|
|
|
|
|
Set the database entry for "key" to "value".
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Name for this entry
|
|
|
|
:param bytes value: Data for this entry
|
2014-01-13 19:52:22 +01:00
|
|
|
:param bool sync:
|
|
|
|
If ``True``, the write will be flushed from the operating system
|
|
|
|
buffer cache (by calling WritableFile::Sync()) before the write
|
|
|
|
is considered complete. If this flag is true, writes will be
|
|
|
|
slower.
|
|
|
|
|
|
|
|
If this flag is ``False``, and the machine crashes, some recent
|
|
|
|
writes may be lost. Note that if it is just the process that
|
|
|
|
crashes (i.e., the machine does not reboot), no writes will be
|
|
|
|
lost even if ``sync == False``.
|
|
|
|
|
|
|
|
In other words, a DB write with ``sync == False`` has similar
|
|
|
|
crash semantics as the "write()" system call. A DB write
|
|
|
|
with ``sync == True`` has similar crash semantics to a "write()"
|
|
|
|
system call followed by "fdatasync()".
|
|
|
|
|
|
|
|
:param bool disable_wal:
|
|
|
|
If ``True``, writes will not first go to the write ahead log,
|
|
|
|
and the write may got lost after a crash.
|
|
|
|
|
|
|
|
.. py:method:: delete(key, sync=False, disable_wal=False)
|
|
|
|
|
|
|
|
Remove the database entry for "key".
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Name to delete
|
2014-01-13 19:52:22 +01:00
|
|
|
:param sync: See :py:meth:`rocksdb.DB.put`
|
|
|
|
:param disable_wal: See :py:meth:`rocksdb.DB.put`
|
|
|
|
:raises rocksdb.errors.NotFound: If the key did not exists
|
|
|
|
|
|
|
|
.. py:method:: merge(key, value, sync=False, disable_wal=False)
|
|
|
|
|
|
|
|
Merge the database entry for "key" with "value".
|
|
|
|
The semantics of this operation is determined by the user provided
|
|
|
|
merge_operator when opening DB.
|
|
|
|
|
|
|
|
See :py:meth:`rocksdb.DB.put` for the parameters
|
|
|
|
|
|
|
|
:raises:
|
|
|
|
:py:exc:`rocksdb.errors.NotSupported` if this is called and
|
|
|
|
no :py:attr:`rocksdb.Options.merge_operator` was set at creation
|
|
|
|
|
|
|
|
|
|
|
|
.. py:method:: write(batch, sync=False, disable_wal=False)
|
|
|
|
|
|
|
|
Apply the specified updates to the database.
|
|
|
|
|
|
|
|
:param rocksdb.WriteBatch batch: Batch to apply
|
|
|
|
:param sync: See :py:meth:`rocksdb.DB.put`
|
|
|
|
:param disable_wal: See :py:meth:`rocksdb.DB.put`
|
|
|
|
|
2014-05-31 20:21:42 +02:00
|
|
|
.. py:method:: get(key, verify_checksums=False, fill_cache=True, snapshot=None, read_tier="all")
|
2014-01-13 19:52:22 +01:00
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Name to get
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
:param bool verify_checksums:
|
|
|
|
If ``True``, all data read from underlying storage will be
|
|
|
|
verified against corresponding checksums.
|
|
|
|
|
|
|
|
:param bool fill_cache:
|
|
|
|
Should the "data block", "index block" or "filter block"
|
|
|
|
read for this iteration be cached in memory?
|
|
|
|
Callers may wish to set this field to ``False`` for bulk scans.
|
|
|
|
|
|
|
|
:param snapshot:
|
|
|
|
If not ``None``, read as of the supplied snapshot
|
|
|
|
(which must belong to the DB that is being read and which must
|
|
|
|
not have been released). Is it ``None`` a implicit snapshot of the
|
|
|
|
state at the beginning of this read operation is used
|
|
|
|
:type snapshot: :py:class:`rocksdb.Snapshot`
|
|
|
|
|
|
|
|
:param string read_tier:
|
|
|
|
Specify if this read request should process data that ALREADY
|
|
|
|
resides on a particular cache. If the required data is not
|
|
|
|
found at the specified cache,
|
|
|
|
then :py:exc:`rocksdb.errors.Incomplete` is raised.
|
|
|
|
|
|
|
|
| Use ``all`` if a fetch from disk is allowed.
|
|
|
|
| Use ``cache`` if only data from cache is allowed.
|
|
|
|
|
|
|
|
:returns: ``None`` if not found, else the value for this key
|
|
|
|
|
2014-05-31 20:21:42 +02:00
|
|
|
.. py:method:: multi_get(keys, verify_checksums=False, fill_cache=True, snapshot=None, read_tier="all")
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
:param keys: Keys to fetch
|
2014-01-17 07:42:48 +01:00
|
|
|
:type keys: list of bytes
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
For the other params see :py:meth:`rocksdb.DB.get`
|
|
|
|
|
|
|
|
:returns:
|
2014-01-17 07:42:48 +01:00
|
|
|
A ``dict`` where the value is either ``bytes`` or ``None`` if not found
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
:raises: If the fetch for a single key fails
|
|
|
|
|
|
|
|
.. note::
|
|
|
|
keys will not be "de-duplicated".
|
|
|
|
Duplicate keys will return duplicate values in order.
|
|
|
|
|
2014-05-31 20:21:42 +02:00
|
|
|
.. py:method:: key_may_exist(key, fetch=False, verify_checksums=False, fill_cache=True, snapshot=None, read_tier="all")
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
If the key definitely does not exist in the database, then this method
|
|
|
|
returns ``False``, else ``True``. If the caller wants to obtain value
|
|
|
|
when the key is found in memory, fetch should be set to ``True``.
|
|
|
|
This check is potentially lighter-weight than invoking DB::get().
|
|
|
|
One way to make this lighter weight is to avoid doing any IOs.
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Key to check
|
2014-01-13 19:52:22 +01:00
|
|
|
:param bool fetch: Obtain also the value if found
|
|
|
|
|
|
|
|
For the other params see :py:meth:`rocksdb.DB.get`
|
|
|
|
|
|
|
|
:returns:
|
|
|
|
* ``(True, None)`` if key is found but value not in memory
|
|
|
|
* ``(True, None)`` if key is found and ``fetch=False``
|
|
|
|
* ``(True, <data>)`` if key is found and value in memory and ``fetch=True``
|
|
|
|
* ``(False, None)`` if key is not found
|
|
|
|
|
2018-03-24 19:03:20 +01:00
|
|
|
.. py:method:: iterkeys(verify_checksums=False, fill_cache=True, snapshot=None, read_tier="all")
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
Iterate over the keys
|
|
|
|
|
|
|
|
For other params see :py:meth:`rocksdb.DB.get`
|
|
|
|
|
|
|
|
:returns:
|
|
|
|
A iterator object which is not valid yet.
|
|
|
|
Call first one of the seek methods of the iterator to position it
|
|
|
|
|
|
|
|
:rtype: :py:class:`rocksdb.BaseIterator`
|
|
|
|
|
2018-03-24 19:03:20 +01:00
|
|
|
.. py:method:: itervalues(verify_checksums=False, fill_cache=True, snapshot=None, read_tier="all")
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
Iterate over the values
|
|
|
|
|
|
|
|
For other params see :py:meth:`rocksdb.DB.get`
|
|
|
|
|
|
|
|
:returns:
|
|
|
|
A iterator object which is not valid yet.
|
|
|
|
Call first one of the seek methods of the iterator to position it
|
|
|
|
|
|
|
|
:rtype: :py:class:`rocksdb.BaseIterator`
|
|
|
|
|
2018-03-24 19:03:20 +01:00
|
|
|
.. py:method:: iteritems(verify_checksums=False, fill_cache=True, snapshot=None, read_tier="all")
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
Iterate over the items
|
|
|
|
|
|
|
|
For other params see :py:meth:`rocksdb.DB.get`
|
|
|
|
|
|
|
|
:returns:
|
|
|
|
A iterator object which is not valid yet.
|
|
|
|
Call first one of the seek methods of the iterator to position it
|
|
|
|
|
|
|
|
:rtype: :py:class:`rocksdb.BaseIterator`
|
|
|
|
|
2022-01-18 01:55:20 +01:00
|
|
|
.. py:method:: iterator(start=None, column_family=None, iterate_lower_bound=None, iterate_upper_bound=None, reverse=False, include_key=True, include_value=True, fill_cache=True, prefix_same_as_start=False, auto_prefix_mode=False)
|
|
|
|
|
|
|
|
:param start: prefix to seek to
|
|
|
|
:type start: bytes
|
|
|
|
|
|
|
|
:param column_family: column family handle
|
|
|
|
:type column_family: :py:class:`rocksdb.ColumnFamilyHandle`
|
|
|
|
|
|
|
|
:param iterate_lower_bound:
|
|
|
|
defines the smallest key at which the backward iterator can return an entry.
|
|
|
|
Once the bound is passed, Valid() will be false. `iterate_lower_bound` is
|
|
|
|
inclusive ie the bound value is a valid entry.
|
|
|
|
If prefix_extractor is not null, the Seek target and `iterate_lower_bound`
|
|
|
|
need to have the same prefix. This is because ordering is not guaranteed
|
|
|
|
outside of prefix domain.
|
|
|
|
:type iterate_lower_bound: bytes
|
|
|
|
|
|
|
|
:param iterate_upper_bound:
|
|
|
|
defines the extent up to which the forward iterator
|
|
|
|
can returns entries. Once the bound is reached, Valid() will be false.
|
|
|
|
"iterate_upper_bound" is exclusive ie the bound value is
|
|
|
|
not a valid entry. If prefix_extractor is not null:
|
|
|
|
1. If auto_prefix_mode = true, iterate_upper_bound will be used
|
|
|
|
to infer whether prefix iterating (e.g. applying prefix bloom filter)
|
|
|
|
can be used within RocksDB. This is done by comparing
|
|
|
|
iterate_upper_bound with the seek key.
|
|
|
|
2. If auto_prefix_mode = false, iterate_upper_bound only takes
|
|
|
|
effect if it shares the same prefix as the seek key. If
|
|
|
|
iterate_upper_bound is outside the prefix of the seek key, then keys
|
|
|
|
returned outside the prefix range will be undefined, just as if
|
|
|
|
iterate_upper_bound = null.
|
|
|
|
If iterate_upper_bound is not null, SeekToLast() will position the iterator
|
|
|
|
at the first key smaller than iterate_upper_bound.
|
|
|
|
:type iterate_upper_bound: bytes
|
|
|
|
|
|
|
|
:param reverse: run the iteration in reverse - using `reversed` is also supported
|
|
|
|
:type reverse: bool
|
|
|
|
|
|
|
|
:param include_key: the iterator should include the key in each iteration
|
|
|
|
:type include_key: bool
|
|
|
|
|
|
|
|
:param include_value: the iterator should include the value in each iteration
|
|
|
|
:type include_value: bool
|
|
|
|
|
|
|
|
:param fill_cache: Should the "data block"/"index block" read for this iteration be placed in
|
|
|
|
block cache? Callers may wish to set this field to false for bulk scans.
|
|
|
|
This would help not to the change eviction order of existing items in the
|
|
|
|
block cache. Default: true
|
|
|
|
:type fill_cache: bool
|
|
|
|
|
|
|
|
:param bool prefix_same_as_start:
|
|
|
|
Enforce that the iterator only iterates over the same prefix as the seek.
|
|
|
|
This option is effective only for prefix seeks, i.e. prefix_extractor is
|
|
|
|
non-null for the column family and total_order_seek is false. Unlike
|
|
|
|
iterate_upper_bound, prefix_same_as_start only works within a prefix
|
|
|
|
but in both directions. Default: false
|
|
|
|
:type prefix_same_as_start: bool
|
|
|
|
|
|
|
|
:param bool auto_prefix_mode: When true, by default use total_order_seek = true, and RocksDB can
|
2022-01-18 00:37:23 +01:00
|
|
|
selectively enable prefix seek mode if won't generate a different result
|
|
|
|
from total_order_seek, based on seek key, and iterator upper bound.
|
|
|
|
Not supported in ROCKSDB_LITE mode, in the way that even with value true
|
|
|
|
prefix mode is not used. Default: false
|
2022-01-18 01:55:20 +01:00
|
|
|
:type auto_prefix_mode: bool
|
2022-01-18 00:37:23 +01:00
|
|
|
|
|
|
|
:returns:
|
|
|
|
A iterator object which is valid and ready to begin using. It will be either a key, item or value
|
|
|
|
iterator depending on the arguments provided.
|
|
|
|
:rtype: :py:class:`rocksdb.BaseIterator`
|
|
|
|
|
2014-01-13 19:52:22 +01:00
|
|
|
.. py:method:: snapshot()
|
|
|
|
|
|
|
|
Return a handle to the current DB state.
|
|
|
|
Iterators created with this handle will all observe a stable snapshot
|
|
|
|
of the current DB state.
|
|
|
|
|
|
|
|
:rtype: :py:class:`rocksdb.Snapshot`
|
|
|
|
|
|
|
|
|
|
|
|
.. py:method:: get_property(prop)
|
|
|
|
|
|
|
|
DB implementations can export properties about their state
|
|
|
|
via this method. If "property" is a valid property understood by this
|
2014-01-17 07:42:48 +01:00
|
|
|
DB implementation, a byte string with its value is returned.
|
2014-01-13 19:52:22 +01:00
|
|
|
Otherwise ``None``
|
|
|
|
|
|
|
|
Valid property names include:
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
* ``b"rocksdb.num-files-at-level<N>"``: return the number of files at level <N>,
|
2014-01-13 19:52:22 +01:00
|
|
|
where <N> is an ASCII representation of a level number (e.g. "0").
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
* ``b"rocksdb.stats"``: returns a multi-line byte string that describes statistics
|
2014-01-13 19:52:22 +01:00
|
|
|
about the internal operation of the DB.
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
* ``b"rocksdb.sstables"``: returns a multi-line byte string that describes all
|
2014-01-13 19:52:22 +01:00
|
|
|
of the sstables that make up the db contents.
|
|
|
|
|
2014-04-27 16:36:27 +02:00
|
|
|
* ``b"rocksdb.num-immutable-mem-table"``: Number of immutable mem tables.
|
|
|
|
|
|
|
|
* ``b"rocksdb.mem-table-flush-pending"``: Returns ``1`` if mem table flush is pending, otherwise ``0``.
|
|
|
|
|
|
|
|
* ``b"rocksdb.compaction-pending"``: Returns ``1`` if a compaction is pending, otherweise ``0``.
|
|
|
|
|
|
|
|
* ``b"rocksdb.background-errors"``: Returns accumulated background errors encountered.
|
|
|
|
|
|
|
|
* ``b"rocksdb.cur-size-active-mem-table"``: Returns current size of the active memtable.
|
|
|
|
|
2014-01-13 19:52:22 +01:00
|
|
|
.. py:method:: get_live_files_metadata()
|
|
|
|
|
|
|
|
Returns a list of all table files.
|
|
|
|
|
|
|
|
It returns a list of dict's were each dict has the following keys.
|
|
|
|
|
|
|
|
``name``
|
|
|
|
Name of the file
|
|
|
|
|
|
|
|
``level``
|
|
|
|
Level at which this file resides
|
|
|
|
|
|
|
|
``size``
|
|
|
|
File size in bytes
|
|
|
|
|
|
|
|
``smallestkey``
|
|
|
|
Smallest user defined key in the file
|
|
|
|
|
|
|
|
``largestkey``
|
|
|
|
Largest user defined key in the file
|
|
|
|
|
|
|
|
``smallest_seqno``
|
|
|
|
smallest seqno in file
|
|
|
|
|
|
|
|
``largest_seqno``
|
|
|
|
largest seqno in file
|
2014-05-31 17:59:29 +02:00
|
|
|
|
2015-08-27 21:45:09 +02:00
|
|
|
.. py:method:: compact_range(begin=None, end=None, ** options)
|
2014-05-31 17:59:29 +02:00
|
|
|
|
|
|
|
Compact the underlying storage for the key range [begin,end].
|
|
|
|
The actual compaction interval might be superset of [begin, end].
|
|
|
|
In particular, deleted and overwritten versions are discarded,
|
|
|
|
and the data is rearranged to reduce the cost of operations
|
|
|
|
needed to access the data.
|
|
|
|
|
|
|
|
This operation should typically only be invoked by users who understand
|
|
|
|
the underlying implementation.
|
|
|
|
|
|
|
|
``begin == None`` is treated as a key before all keys in the database.
|
|
|
|
``end == None`` is treated as a key after all keys in the database.
|
|
|
|
Therefore the following call will compact the entire database: ``db.compact_range()``.
|
|
|
|
|
|
|
|
Note that after the entire database is compacted, all data are pushed
|
|
|
|
down to the last level containing any data. If the total data size
|
|
|
|
after compaction is reduced, that level might not be appropriate for
|
2015-08-27 21:45:09 +02:00
|
|
|
hosting all the files. In this case, client could set change_level
|
2014-05-31 17:59:29 +02:00
|
|
|
to ``True``, to move the files back to the minimum level capable of holding
|
|
|
|
the data set or a given level (specified by non-negative target_level).
|
|
|
|
|
|
|
|
:param bytes begin: Key where to start compaction.
|
|
|
|
If ``None`` start at the beginning of the database.
|
|
|
|
:param bytes end: Key where to end compaction.
|
|
|
|
If ``None`` end at the last key of the database.
|
2015-08-27 21:45:09 +02:00
|
|
|
:param bool change_level: If ``True``, compacted files will be moved to
|
|
|
|
the minimum level capable of holding the data
|
|
|
|
or given level (specified by non-negative target_level).
|
2014-05-31 17:59:29 +02:00
|
|
|
If ``False`` you may end with a bigger level
|
2015-08-27 21:45:09 +02:00
|
|
|
than configured. Default is ``False``.
|
|
|
|
:param int target_level: If change_level is true and target_level have non-negative
|
|
|
|
value, compacted files will be moved to target_level.
|
|
|
|
Default is ``-1``.
|
|
|
|
:param string bottommost_level_compaction:
|
|
|
|
For level based compaction, we can configure if we want to
|
|
|
|
skip/force bottommost level compaction. By default level based
|
|
|
|
compaction will only compact the bottommost level if there is a
|
|
|
|
compaction filter. It can be set to the following values.
|
|
|
|
|
|
|
|
``skip``
|
|
|
|
Skip bottommost level compaction
|
|
|
|
|
|
|
|
``if_compaction_filter``
|
|
|
|
Only compact bottommost level if there is a compaction filter.
|
|
|
|
This is the default.
|
|
|
|
|
|
|
|
``force``
|
|
|
|
Always compact bottommost level
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
.. py:attribute:: options
|
|
|
|
|
|
|
|
Returns the associated :py:class:`rocksdb.Options` instance.
|
|
|
|
|
|
|
|
.. note::
|
|
|
|
|
|
|
|
Changes to this object have no effect anymore.
|
|
|
|
Consider this as read-only
|
|
|
|
|
|
|
|
Iterator
|
|
|
|
========
|
|
|
|
|
|
|
|
.. py:class:: rocksdb.BaseIterator
|
|
|
|
|
|
|
|
Base class for all iterators in this module. After creation a iterator is
|
|
|
|
invalid. Call one of the seek methods first before starting iteration
|
|
|
|
|
|
|
|
.. py:method:: seek_to_first()
|
|
|
|
|
|
|
|
Position at the first key in the source
|
|
|
|
|
|
|
|
.. py:method:: seek_to_last()
|
|
|
|
|
|
|
|
Position at the last key in the source
|
|
|
|
|
|
|
|
.. py:method:: seek(key)
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Position at the first key in the source that at or past
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
Methods to support the python iterator protocol
|
|
|
|
|
|
|
|
.. py:method:: __iter__()
|
|
|
|
.. py:method:: __next__()
|
|
|
|
.. py:method:: __reversed__()
|
|
|
|
|
|
|
|
Snapshot
|
|
|
|
========
|
|
|
|
|
|
|
|
.. py:class:: rocksdb.Snapshot
|
|
|
|
|
|
|
|
Opaque handler for a single Snapshot.
|
|
|
|
Snapshot is released if nobody holds a reference on it.
|
|
|
|
Retrieved via :py:meth:`rocksdb.DB.snapshot`
|
|
|
|
|
|
|
|
WriteBatch
|
|
|
|
==========
|
|
|
|
|
|
|
|
.. py:class:: rocksdb.WriteBatch
|
|
|
|
|
|
|
|
WriteBatch holds a collection of updates to apply atomically to a DB.
|
|
|
|
|
|
|
|
The updates are applied in the order in which they are added
|
|
|
|
to the WriteBatch. For example, the value of "key" will be "v3"
|
|
|
|
after the following batch is written::
|
|
|
|
|
|
|
|
batch = rocksdb.WriteBatch()
|
2014-01-17 07:42:48 +01:00
|
|
|
batch.put(b"key", b"v1")
|
|
|
|
batch.delete(b"key")
|
|
|
|
batch.put(b"key", b"v2")
|
|
|
|
batch.put(b"key", b"v3")
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
.. py:method:: __init__(data=None)
|
|
|
|
|
|
|
|
Creates a WriteBatch.
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes data:
|
2014-01-13 19:52:22 +01:00
|
|
|
A serialized version of a previous WriteBatch. As retrieved
|
|
|
|
from a previous .data() call. If ``None`` a empty WriteBatch is
|
|
|
|
generated
|
|
|
|
|
|
|
|
.. py:method:: put(key, value)
|
|
|
|
|
|
|
|
Store the mapping "key->value" in the database.
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Name of the entry to store
|
|
|
|
:param bytes value: Data of this entry
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
.. py:method:: merge(key, value)
|
|
|
|
|
|
|
|
Merge "value" with the existing value of "key" in the database.
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Name of the entry to merge
|
|
|
|
:param bytes value: Data to merge
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
.. py:method:: delete(key)
|
|
|
|
|
|
|
|
If the database contains a mapping for "key", erase it. Else do nothing.
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:param bytes key: Key to erase
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
.. py:method:: clear()
|
|
|
|
|
|
|
|
Clear all updates buffered in this batch.
|
|
|
|
|
2015-07-05 13:35:15 +02:00
|
|
|
.. note::
|
|
|
|
Don't call this method if there is an outstanding iterator.
|
|
|
|
Calling :py:meth:`rocksdb.WriteBatch.clear()` with outstanding
|
|
|
|
iterator, leads to SEGFAULT.
|
|
|
|
|
2014-01-13 19:52:22 +01:00
|
|
|
.. py:method:: data()
|
|
|
|
|
|
|
|
Retrieve the serialized version of this batch.
|
|
|
|
|
2014-01-17 07:42:48 +01:00
|
|
|
:rtype: ``bytes``
|
2014-01-13 19:52:22 +01:00
|
|
|
|
|
|
|
.. py:method:: count()
|
|
|
|
|
|
|
|
Returns the number of updates in the batch
|
|
|
|
|
|
|
|
:rtype: int
|
|
|
|
|
2015-07-05 13:35:15 +02:00
|
|
|
.. py:method:: __iter__()
|
|
|
|
|
|
|
|
Returns an iterator over the current contents of the write batch.
|
|
|
|
|
|
|
|
If you add new items to the batch, they are not visible for this
|
|
|
|
iterator. Create a new one if you need to see them.
|
|
|
|
|
|
|
|
.. note::
|
|
|
|
Calling :py:meth:`rocksdb.WriteBatch.clear()` on the write batch
|
|
|
|
invalidates the iterator. Using a iterator where its corresponding
|
|
|
|
write batch has been cleared, leads to SEGFAULT.
|
|
|
|
|
|
|
|
:rtype: :py:class:`rocksdb.WriteBatchIterator`
|
|
|
|
|
|
|
|
WriteBatchIterator
|
|
|
|
==================
|
|
|
|
|
|
|
|
.. py:class:: rocksdb.WriteBatchIterator
|
|
|
|
|
|
|
|
.. py:method:: __iter__()
|
|
|
|
|
|
|
|
Returns self.
|
|
|
|
|
|
|
|
.. py:method:: __next__()
|
|
|
|
|
|
|
|
Returns the next item inside the corresponding write batch.
|
|
|
|
The return value is a tuple of always size three.
|
|
|
|
|
|
|
|
First item (Name of the operation):
|
|
|
|
|
|
|
|
* ``"Put"``
|
|
|
|
* ``"Merge"``
|
|
|
|
* ``"Delete"``
|
|
|
|
|
|
|
|
Second item (key):
|
|
|
|
Key for this operation.
|
|
|
|
|
|
|
|
Third item (value):
|
|
|
|
The value for this operation. Empty for ``"Delete"``.
|
|
|
|
|
2015-08-15 16:34:35 +02:00
|
|
|
Repair DB
|
|
|
|
=========
|
|
|
|
|
|
|
|
.. py:function:: repair_db(db_name, opts)
|
|
|
|
|
|
|
|
:param unicode db_name: Name of the database to open
|
|
|
|
:param opts: Options for this specific database
|
|
|
|
:type opts: :py:class:`rocksdb.Options`
|
|
|
|
|
|
|
|
If a DB cannot be opened, you may attempt to call this method to
|
|
|
|
resurrect as much of the contents of the database as possible.
|
|
|
|
Some data may be lost, so be careful when calling this function
|
|
|
|
on a database that contains important information.
|
|
|
|
|
2015-07-05 13:35:15 +02:00
|
|
|
|
2014-01-13 19:52:22 +01:00
|
|
|
Errors
|
|
|
|
======
|
|
|
|
|
|
|
|
.. py:exception:: rocksdb.errors.NotFound
|
|
|
|
.. py:exception:: rocksdb.errors.Corruption
|
|
|
|
.. py:exception:: rocksdb.errors.NotSupported
|
|
|
|
.. py:exception:: rocksdb.errors.InvalidArgument
|
|
|
|
.. py:exception:: rocksdb.errors.RocksIOError
|
|
|
|
.. py:exception:: rocksdb.errors.MergeInProgress
|
|
|
|
.. py:exception:: rocksdb.errors.Incomplete
|
|
|
|
|
|
|
|
|