The python data pipeline defined for BrodyLab.
- Install conda on your system: https://conda.io/projects/conda/en/latest/user-guide/install/index.html
- If running in Windows get git
- (Optional for ERDs) Install graphviz
- Open a new terminal
- Clone this repository:
git@github.com:Brody-Lab/bl_pipeline_python.git
- If you cannot clone repositories with ssh, set keys
- Create a conda environment:
conda create -n bl_pipeline_python_env python==3.7
. - Activate environment:
conda activate bl_pipeline_python_env
. (Activate environment each time you use the project) - Change directory to this repository
cd bl_pipeline_python
. - Install all required libraries
pip install -e .
- Datajoint Configuration:
jupyter notebook notebooks/00-datajoint-configuration.ipynb
We have created some tutorial notebooks to help you start working with datajoint
- Querying data (Strongly recommended)
jupyter notebook notebooks/tutorials/Explore_Sessions_Data.ipynb
jupyter notebook notebooks/tutorials/1-Explore U19 data pipeline with DataJoint.ipynb
-
This portion of the workflow usings DataJoint's standard python package element-array-ephys.
-
This workflow uses the
ephys_chronic
module fromelement-array-ephys
.
bl_pipeline_python/process/process.py
- Copy data from source tables (to shadow tables) to new tables
- Shadow table allows for renaming of primary key
- Shadow table has same definition of the new table, except that the primary key is the same as the source table
- For each shadow table set the keys as a secondary field when not used as primary key
bl_pipeline_python/notebooks/debugging_data_integrity.ipynb