Attention

This is no longer maintained and has been superseded by datajoint-company/datajoint-docs. Please file new issues there (or help contribute!). We are currently migrating and generating new content until December 2022 after which we’ll be decomissioning https://docs.datajoint.org and https://tutorials.datajoint.org in favor of https://datajoint.com/docs/.

Creating Schemas

Schemas

On the database server, related tables are grouped into a named collection called a schema. This grouping organizes the data and allows control of user access. A database server may contain multiple schemas each containing a subset of the tables. A single pipeline may comprise multiple schemas. Tables are defined within a schema, so a schema must be created before the creation of any tables.

Note

By convention, the datajoint package is imported as dj. The documentation refers to the package as dj throughout.

Create a new schema using the dj.schema function:

import datajoint as dj
schema = dj.schema('alice_experiment')

This statement creates the database schema alice_experiment on the server.

The returned object schema will then serve as a decorator for DataJoint classes, as described in Creating Tables.

It is a common practice to have a separate Python module for each schema. Therefore, each such module has only one dj.schema object defined and is usually named schema.

The dj.schema constructor can take a number of optional parameters after the schema name.

  • context - Dictionary for looking up foreign key references. Defaults to None to use local context.

  • connection - Specifies the DataJoint connection object. Defaults to dj.conn().

  • create_schema - When False, the schema object will not create a schema on the database and will raise an error if one does not already exist. Defaults to True.

  • create_tables - When False, the schema object will not create tables on the database and will raise errors when accessing missing tables. Defaults to True.

Working with existing data

See the chapter Work with Existing Pipelines for how to work with data in existing pipelines, including accessing a pipeline from one language when the pipeline was developed using another.

Talk to the Community