Getting Started

First, install the BedquiltDB extension on your PostgreSQL instance, then create enable the extension on a database:

-- in psql
CREATE DATABASE bedquilt_test;

-- connect to bedquilt_test and then run:
CREATE EXTENSION pgcrypto;
CREATE EXTENSION bedquilt;

Then, install the pybedquilt libary:

# python2
$ pip install pybedquilt

# python3
$ pip3 install pybedquilt

Inserting Data

Now, let's open a python interpreter and create a BedquiltClient:

$ python
>>> import pybedquilt
>>> client = pybedquilt.BedquiltClient("dbname=bedquilt_test")

Once we have a client, connected to the PostgreSQL server, we can get a BedquiltCollection object:

>>> people = client['people']

Lets put some people documents in our collection:

>>> sarah = {
...    '_id': "sarah@example.com",
...    'name': "Sarah Jones",
...    'age': 27,
...    'city': "Edinburgh",
...    'likes': ["icecream", "code", "tigers"]
... }
>>> dave = {
...    'name': "Dave",
...    'age': 40,
...    'city': "London",
...    'likes': ["code", "puppies", "knitting"]
... }
>>> elaine = {
...    'name': "Elaine",
...    'age': 36,
...    'city': "Edinburgh",
...    'likes': ["board games", "cider", "owls"]
... }
>>> people.insert(sarah)
'sarah@example.com'
>>> people.insert(dave)
'99a5cb8648566901436e2967'
>>> people.insert(elaine)
'cabb5c9f6a55ecdebbc42c69'

As you can see, we can insert any python dictionary which can be serialized to JSON. When we insert a document (or dict, if you prefer), the insert method returns the _id field of the document. If the document does not have an _id key already, a random one will be generated on the server. The _id field must be unique for all documents in the collection.

Adding Constraints

Right now our people collection is completely un-structured. The insert function will accept a document of any structure and happily add it to the collection, even if it doesn't make sense to do so. For example, we could add an blog-post to the people collection like this:

>>> people.insert({
...    '_id': 'a-cool-article',
...    'title': 'A Cool Article',
...    'content': '...',
...    'tags': ['coolness', 'awesome']
... })

Or, we could do something silly and insert a document which looks like a person, but is still wrong:

>>> people.insert({
...    'name': None,
...    'age': 'Elaine',
...    'city': ["board games", "cider", "owls"],
...    'lol': 'wtf',
...    'likes': {
...        'street': 'Mill Lane',
...        'city': None
...    }
... })

Many NoSQL databases behave in the same way. On the one hand, this makes it easy to work with loosely structured data. On the other hand, it makes it too easy to add bad data to the collection, which can be very difficult to deal with later on.

With BedquiltDB we can add constraints to collections, which allow us to specify a set of checks to be performed on incoming data. If any of the checks fail then the document will be rejected and not saved to the collection.

Lets add some constraints to our people collection:

>>> people.add_constraints({
...     'name': {'$required': 1,
...              '$notnull': 1,
...              '$type': 'string'},
...     'age': {'$required': 1,
...             '$type': 'number'},
...     'city': {'$type': 'string'},
...     'likes': {'$required': 1,
...               '$type': 'array'}
... })

Now, when we try to insert garbage data into the people collection, the bad data will be rejected, and an exception raised to tell us about how silly we've been.

Finding Data

With a few documents in our people collection, we can then query the collection and see what comes back:

>>> people.find({'city': "Edinburgh"})
<pybedquilt.core.BedquiltCursor at 0x1038b7690>

The find method returns a BedquiltCursor, an object which represents a stream of results coming from the server. The cursor is lazy, so no results are fetched from the server until we ask for them, by consuming the cursor. Let's use the python list function on the cursor and see what happens:

>>> list( people.find({'city': "Edinburgh"}) )
[{'_id': 'sarah@example.com', ....}, {'_id': 'cabb5c9f6a55ecdebbc42c69', ...}]
>>> list( people.find({'city': {'$noteq': 'Glasgow'}}))

In this case we get a list of two documents back, corresponding to the two people who live in Edinburgh, sarah and elaine.

We could also have written a for loop to consume the cursor:

>>> for doc in people.find({'city': "Edinburgh"}):
...     print doc['name']

Or any other way one would usually consume a generator in python.


The first argument to the find method is a query dictionary. A document matches the query if its contents are a superset of the query data. In our example, we're basically saying "find documents where the city field is 'Edinburgh".

We can also query on the contents of nested JSON data structures, for example if we wanted to get all documents where the person likes "code":

>>> list( people.find({'likes': ['code']}) )
[ {...}, {...} ]

That query returns the two documents for sarah and dave.

If we omit the query document, then no filter is applied to the query, we just get back all the documents in the collection:

>>> list( people.find() )

We can also apply a sort to the stream of documents we receive, say, by the age field, in ascending order:

>>> people.find({'likes' :['code']}, sort=[{'age': 1}])

Nah, let's sort by name, but in descending order:

>>> people.find({'likes' :['code']}, sort=[{'name': -1}])

Even better, let's sort by age descending, then by name ascending:

>>> people.find({'likes' :['code']}, sort=[{'age': -1}, {'name': 1}])

We can limit the results to an arbitrary number of documents:

>>> people.find({'likes' :['code']}, limit=4)

And we can skip over the first n results:

>>> people.find({'likes' :['code']}, skip=2)

In fact, we can do all of these at once, or in any combination:

>>> people.find(
...  {'likes' :['code']},
...  skip=4,
...  limit=5,
...  sort=[{'age': -1}, {'name': 1}])

The skip, limit and sort options are pretty powerful.


If we know we are only interested in a single document, we can limit the result set to a single document matching our query by using the find_one method:

>>> people.find_one({'likes': ['code']})
{'_id': 'sarah@example.com' ...}

If we know the _id key of a document we want, we can query for that document directly, using find_one_by_id:

>>> people.find_one_by_id('sarah@example.com')
{'_id': 'sarah@example.com', 'name': 'Sarah Jones'...}

Updating Existing Documents

We modify and save an existing document too, using the save method:

doc = people.find_one_by_id('sarah@example.com')
doc['job'] = "Web Developer"
doc['likes'].append("Javascript")
people.save(doc)

If the saved document has an _id which matches a document in the collection, it will be overwritten with the new data. If not, then save behaves just like the insert method.

Removing Documents

We can delete data from a collection using the remove, remove_one and remove_one_by_id methods. They work exactly as you'd expect:

people.remove({'cool': False})
people.remove_one({'likes': ['cabbage']})
people.remove_one_by_id('magic_man@example.com')

Using Raw SQL

The BedquiltClient object has a connection property, which is a psycopg2 connection (spoiler warning: pybedquilt uses psycopg2 internally). You can use the connection to execute arbitrary SQL commands:

cur = client.connection.cursor()
cur.execute("select 1;")
cur.fetchall()  # => [(1,)]

cur.execute("select bq_find_one_by_id('users', 'joe@example.com')")
cur.fetchall()  # => [('{"_id": "joe@example.com", ...}')]

Alternatively, you can import psycopg2 yourself and use the same connection string as you used to build the BedquiltClient instance.