Port is a research tool that enables individuals to donate their digital trace data for academic research in a secure, transparent, and privacy-preserving way.
Data donation allows researchers to invite participants to share their data download packages (DDPs). A major challenge is however that DDPs potentially contain very sensitive data, and often not all data is needed to answer the specific research question under investigation. To circumvent these challenges, an alternative framework was developed:
- The research participant requests their personal DDP at the platform of interest.
- They download it onto their own personal device.
- By means of local processing, only the features of interest to the researcher are extracted from that DDP.
- The participant inspects the extracted features after which they can consent (or decline) to donate.
To allow for the local processing (step 3) to take place, we developed the software Port. Port creates a frontend that guides participants through the data donation steps.
Port is open-source and allows for researchers to fully configure their own data donation study.
Note: Port is only a frontend. In order for it to be used in a live study, it needs to be hosted with a server and integrate with a solution to store and retrieve the donated data. This repository will discuss a some readily available options.
In order to start a local instance of Port go through the following steps:
-
Pre-requisites
-
Install dependencies:
cd ./port npm install
-
Start the local web server (with hotloading enabled):
npm run watch
-
You can now go to the browser:
http://localhost:3000
.
If the installation went correctly you should be greeted with a mock data donation study.
A researcher can implement their own data donation flow by writing a Python script. The Python scripts has 2 main functions:
- It determines the data donation flow. i.e. what screens (for example a file prompt) does the participant gets to see and when. You can use the Port API (
props.py
) for this. - It determines what data gets extract from the participants submission. Here is were Python really shines, you can use most data extraction methods you are familiar with! (As long as it's available in Pyodide)
A typical script includes the following steps:
- Prompt the participant to submit a file
- Handling the submission from step 1. This is the step where you can extract the data you are interested in.
- The extracted data is presented on screen accompanied with a consent button. After consent is given, the data is sent to a storage location of the researcher.
A example such a script is included in this repo: script.py
.
We recommend you use that script as starting point for your own data donation study.
You can find another example of such a script in this repository.
Below you can find examples on how to use the Port API in your script.py
Main function
Every `script.py` should have this function:def process(sessionId):
This function is a generator of commands by using yield
statements. No return
statements should be used.
def process(sessionId):
result1 = yield CommandUIRender(page1)
result2 = yield CommandUIRender(page2)
# last yield should not expect a result
yield CommandUIRender(page3)
ScriptWrapper
and py_worker using send
to iterate over the commands one by one. For more information on yield and Generators: https://realpython.com/introduction-to-python-generators
API imports
from port.api.props as props
from port.api.commands import (CommandUIRender, CommandUIDonate)
Create file input
platform = "Twitter"
progress = 25
file_input_description = props.Translatable({
"en": f"Please follow the download instructions and choose the file that you stored on your device. Click “Skip” at the right bottom, if you do not have a {platform} file. ",
"nl": f"Volg de download instructies en kies het bestand dat u opgeslagen heeft op uw apparaat. Als u geen {platform} bestand heeft klik dan op “Overslaan” rechts onder."
})
allowed_extensions = "application/zip, text/plain"
file_input = props.PropsUIPromptFileInput(file_input_description, allowed_extensions)
Create a consent page
import pandas as pd
table1_title = props.Translatable({
"en": "Title 1",
"nl": "Titel 1"
})
table1_data = pd.DataFrame(data, columns=["columnX", "columnY", "columnZ"])
table1 = props.PropsUIPromptConsentFormTable("table_1", table1_title, table1_data)
table2_title = props.Translatable({
"en": "Title 2",
"nl": "Titel 2"
})
table2_data = pd.DataFrame(data, columns=["columnA", "columnB", "columnC", "columnD"])
table2 = props.PropsUIPromptConsentFormTable("table_2", table2_title, table2_data)
tables = [table1, table1]
# Meta tables currently not supported
meta_tables = []
consent_form = props.PropsUIPromptConsentForm(tables, meta_tables)
Create donation page
header = props.PropsUIHeader(title)
footer = props.PropsUIFooter(progress)
body = props.PropsUIPromptFileInput(file_input_description, allowed_extensions)
page = props.PropsUIPageDonation(platform, header, body, footer)
Create page with radio buttons
header = props.PropsUIHeader(title)
footer = props.PropsUIFooter(progress)
body = props.PropsUIPromptRadioInput(title, description, [{"id": 0, "value": "Selection 1"}, {"id": 1, "value": "Selection 2"}])
page = props.PropsUIPageDonation(platform, header, body, footer)
Handling the result from a file input
page = props.PropsUIPageDonation(platform, header, file_input, footer)
result = yield CommandUIRender(page)
# Result is a dictionary (Payload)
if result.__type__ == "PayloadString":
# File selected
filename = result.value
zipfile = zipfile.ZipFile(filename)
# Extract the data you are interested contained in zipfile
# Typically you will use your own written functions here
...
else:
# No file selected
Handling consent result
platform = "Twitter"
donation_key = f"{sessionId}-{platform}"
page = props.PropsUIPageDonation(platform, header, consent_form, footer)
result = yield CommandUIRender(page)
# Response is a dictionary (Payload)
if result.__type__ == "PayloadJSON":
# User gave consent
yield CommandSystemDonate(donation_key, result.value)
else:
# User declined or skipped
Track user behaviour
tracking_key = f"{sessionId}-tracking"
data = "any json string"
# Use the donate command to store tracking data
yield CommandSystemDonate(tracking_key, data)
Port is a frontend it creates the app that participants will end up interacting with: it determines the data donation flow and extraction logic. If you want to use Port in a data dontion study it needs to be hosted by a server an needs a way of storing donated data.
In order to do a data donation study with Port you have several options:
-
Create production code
npm run prod:clean && npm run prod:build
The production code can now be found in the
dist
folder. -
Run Port on top of Next on the Azure cloud.
For instructions on running Next + Port see: https://github.com/eyra/mono/blob/d3i/latest/PORT.md
-
Create production code
npm run prod:clean && npm run prod:build
The production code can now be found in the
dist
folder. -
Run Port with Portserver.
This is a fairly minimal and cloud agnostic implementation of a server that can serve Port (just like Next it uses Elixir and Phoenix). It uses a database for storing donated data.
Data donation very interesting but sounds like way too much work? Feel free to contact Eyra.
You can also implement your own server for Port. If you want to use your own server solution, please check out these instuctions below:
Add Port app as dependency
Make sure to link your forked Port app as a dependency in your web server code. If `npm` is your dependency manager, add a dependency in `package.json` as follows:
"dependencies": {
"port": "github:<your_organisation>/port",
},
And resolve dependencies:
npm install
Import code from Port dist folder
import Assembly from "port/dist/framework/assembly";
import Worker from "port/dist/framework/processing/py_worker.js";
import "port/dist/port-0.0.0-py3-none-any.whl";
import "port/dist/styles.css";
In this example above imports can be resolved by Webpack (worker-loader, css-loader, and file-loader). Make sure to include the correct config in your webpack.config.js
:
module: {
rules: [
{
test: /worker\.js$/,
use: { loader: "worker-loader" },
},
{
test: /\.css$/,
use: [MiniCssExtractPlugin.loader, "css-loader", "postcss-loader"],
},
{
test: /\.whl$/,
use: [
{
loader: "file-loader",
options: { name: "[name].[ext]", outputPath: "../" },
},
],
},
]
}
Implement System callback interface
// Typescript example
import { CommandSystem, CommandSystemDonate, isCommandSystemDonate } from 'port/dist/framework/types/commands'
import { System } from 'port/dist/framework/types/modules'
export default class MySystem implements System {
send (command: CommandSystem): void {
if (isCommandSystemDonate(command)) {
this.handleDonation(command)
} else {
console.log('[LocalSystem] received unknown command: ' + JSON.stringify(command))
}
}
handleDonation (command: CommandSystemDonate): void {
console.log(`[LocalSystem] received donation: ${command.key}=${command.json_string}`)
}
}
Implement start donation session
// Typescript example
import MySystem from "my_system.ts";
import Assembly from "port/dist/framework/assembly";
import Worker from "port/dist/framework/processing/py_worker.js";
import "port/dist/port-0.0.0-py3-none-any.whl";
import "port/dist/styles.css";
const locale = 'en' // preferred locale
const parent = 'parent_id' // parent DOM element id
const worker = new Worker();
const system = new MySystem();
const container = document.getElementById(parent);
this.assembly = new Assembly(worker, system);
this.assembly.visualisationEngine.start(container, locale);
this.assembly.processingEngine.start();
Workers can not be reused over sessions. Re-run code above for every new donation session.
You can use Port as is, but if your study has demand for something you are free to change the way Port works. Luckily Port is customizable, is this something you want? Use these technical details to your advantage.
Port uses the following data model (also see: src/framework/types)
-
Module Description ProcessingEngine Responsible for processing donation flows VisualizationEngine Responsible for presenting the UI and accepting user input CommandHandler Decoupling of ProcessingEngine and VisualizationEngine System Callback interface for System Commands (e.g. Donation) -
Page Description SplashScreen First page that is rendered before the Python script is loaded with GDPR consent logic Donation Page that uses several prompts to get a file from the user and consent to donate the extracted data End Final page with instructions on how to continue -
Prompt Description FileInput File selection RadioInput Multiple choice question ConsentForm Displays extracted data in tables and asks for user consent Confirm General dialog to ask for extra confirmation -
Command Description Render Render the page Donate Save the extracted data Commands can be send from the Python script using the
yield
keyword. -
Payload Description Void Command without user input as a response True Positive user input (e.g. Ok button in confirm prompt) False Negative user input (e.g. Cancel button in confirm prompt) Error Unexpected problem when handling command String String result File Only used in Javascript. This is intercepted in py_worker.js and translated into a String (filename), while the bytes of the file are written to the Pyodide file system JSON User input structured as JSON, used to return the consent data from the consent form Payloads are part of a Response back to the Python script after sending commands:
export interface Response { __type__: 'Response' command: Command payload: Payload }
Responses are intercepted in py_worker.js and only the payload is returned to the Python script. Payloads don't have a Python representation in the API yet. They are translated into a dictionary (default Pyodide behaviour).
See: src/framework/processing/py/port
-
This object is used in main to wrap the
process
generator function in your script. It translates incoming Javascript and outgoing Python commands. -
- commands.py: Defines commands, pages and prompts that are used to communicate from the Python script to the
VisualisationEngine
andSystem
. - props.py: Defines property objects for pages and prompts
- commands.py: Defines commands, pages and prompts that are used to communicate from the Python script to the
These instructions give you some pointers on things you might like to change or add to Port.
Change copy (texts shown on the web pages)
The app has two types of copy:
- Dynamic copy: part of the Python script
- Static copy: part of React components
Currently two languages are supported (Dutch and English). The Translatable object plays a central role and has a Python and a Typescript implementation
From Python code copy can be used as follows:
from port.api.props import Translatable
copy = Translatable({
"en": "English text",
"nl": "Nederlandse tekst"
})
In React components copy is handled as follows:
import TextBundle from '../../../../text_bundle'
import { Translator } from '../../../../translator'
import { Translatable } from '../../../../types/elements'
interface Props {
dynamicCopy: Translatable // from Python script
locale: string
}
export const MyComponent = ({ dynamicCopy, locale }: Props): JSX.Element => {
const dynamicText = Translator.translate(dynamicCopy, locale)
const staticText = Translator.translate(staticCopy(), locale)
return (
<>
<div>{dynamicText}</div>
<div>{staticText}</div>
</>
)
}
const staticCopy = (): Translatable => {
return new TextBundle()
.add('en', 'English')
.add('nl', 'Nederlands')
}
Add new prompt
Add the properties of the prompt in [src/framework/types/prompts.ts](src/framework/types/prompts.ts) with the following template:
export type PropsUIPrompt =
PropsUIPromptNew |
...
export interface PropsUIPromptNew {
__type__: 'PropsUIPromptNew'
title: Text
description: Text
...
}
export function isPropsUIPromptNew (arg: any): arg is PropsUIPromptNew {
return isInstanceOf<PropsUIPromptNew>(arg, 'PropsUIPromptNew', ['title', 'description', ... ])
}
Add the prompt component to src/framework/visualisation/react/ui/prompts with the following template:
import { Weak } from '../../../../helpers'
import { ReactFactoryContext } from '../../factory'
import { PropsUIPromptNew } from '../../../../types/prompts'
import { Translator } from '../../../../translator'
import { Title2, BodyLarge } from '../elements/text'
import { PrimaryButton } from '../elements/button'
type Props = Weak<PropsUIPromptNew> & ReactFactoryContext
export const New = (props: Props): JSX.Element => {
const { resolve } = props
const { title, description, continueButton } = prepareCopy(props)
function handleContinue (): void {
// Send payload back to script
resolve?.({ __type__: 'PayloadTrue', value: true })
}
return (
<>
<Title2 text={title} />
<BodyLarge text={description} />
<PrimaryButton label={continueButton} onClick={handleContinue} />
</>
)
}
interface Copy {
title: string
description: string
continueButton: string
}
function prepareCopy ({ title, locale }: Props): Copy {
return {
title: Translator.translate(title, locale),
description: Translator.translate(description, locale),
continueButton: Translator.translate(continueButtonLabel(), locale),
}
}
const continueButtonLabel = (): Translatable => {
return new TextBundle()
.add('en', 'Continue')
.add('nl', 'Verder')
}
Use external Python libraries
Python packages are loaded using micropip:
await micropip.install("https://domain.com/path/to/python.whl", deps=False)
Add the above statement to the py_worker.js file as follows:
function installPortPackage() {
console.log('[ProcessingWorker] load port package')
return self.pyodide.runPythonAsync(`
import micropip
await micropip.install("https://domain.com/path/to/python.whl", deps=False)
await micropip.install("/port-0.0.0-py3-none-any.whl", deps=False)
import port
`);
}
The standard library is available by default. Please check The Pyodide docs for other packages you can use.
Implement support for alternative web framework
Create a new folder in [src/framework/visualisation](src/framework/visualisation) with a file inside called `engine.ts` to add support for your web framework of choice.
import { VisualisationEngine } from '../../types/modules'
import { Response, CommandUIRender } from '../../types/commands'
export default class MyEngine implements VisualisationEngine {
locale!: string
root!: HTMLElement
start (root: HTMLElement, locale: string): void {
this.root = root
this.locale = locale
}
async render (command: CommandUIRender): Promise<Response> {
// Render page and return user input as a response
...
}
terminate (): void {
...
}
}
Change implementation of assembly.ts to support your new VisualisationEngine:
import MyEngine from './visualisation/my/engine'
import WorkerProcessingEngine from './processing/worker_engine'
import { VisualisationEngine, ProcessingEngine, System } from './types/modules'
import CommandRouter from './command_router'
export default class Assembly {
visualisationEngine: VisualisationEngine
processingEngine: ProcessingEngine
router: CommandRouter
constructor (worker: Worker, system: System) {
const sessionId = String(Date.now())
this.visualisationEngine = new MyEngine()
this.router = new CommandRouter(system, this.visualisationEngine)
this.processingEngine = new WorkerProcessingEngine(sessionId, worker, this.router)
}
}
Implement support for alternative script language
To support an alternative for Python scripts, create a Javascript file (eg: `my_worker.js`) in [src/framework/processing](src/framework/processing) with the following template:
onmessage = (event) => {
const { eventType } = event.data
switch (eventType) {
case 'initialise':
// Insert initialisation code here
self.postMessage({ eventType: 'initialiseDone' })
break
case 'firstRunCycle':
runCycle(null)
break
case 'nextRunCycle':
const { response } = event.data
runCycle(response.payload)
break
default:
console.log('[ProcessingWorker] Received unsupported event: ', eventType)
}
}
function runCycle (payload) {
console.log('[ProcessingWorker] runCycle ' + JSON.stringify(payload))
// Insert script code here:
// 1. Handle the payload
// 2. Create next command, eg:
nextCommand = new CommandUIRender(new PropsUIPageDonation(...))
self.postMessage({
eventType: 'runCycleDone',
scriptEvent: nextCommand
})
}
Change the implementation of index.tsx to support your new worker file:
const workerFile = new URL('./framework/processing/my_worker.js', import.meta.url)
Make sure to add the worker to the ts-standard
ignore list in package.json:
"ts-standard": {
"ignore": [
"src/framework/processing/my_worker.js"
]
}
Note: don't forget to import this new worker file in your server code
-
Automatic
Jest is used as a testing framework. Tests can be found here: src/test.
Run all unit tests:
npm run dev:test
-
Manual
Start the local web server (with hotloading enabled):
npm run dev:start
-
Integration with Next
To run the Port app on top of Next locally see: https://github.com/eyra/mono/blob/d3i/latest/PORT.md
Code in Javascript types and Python api are currently created by hand. Since both of them are implementions of the same model we will seek the opportunity in the future to define this model in a more abstract way and generate the needed Javascript and Python code accordingly.
The project is a react app created by create-react-app. This is not set in stone for the future but it was a nice way to speed up the development process in the beginning. Using this strongly opinionated setup hides most of the configuration. It uses webpack to bundle and serve the app.
The project uses ts-standard for managing the code style. This is a TypeScript Style Guide, with linter and automatic code fixer based on StandardJS.
Before committing to github Husky runs all the necessary scripts to make sure the code conforms to ts-standard
, all the tests run green, and the dist
folder is up-to-date.
Port is funded by the PDI-SSH and is a collaboration between six Dutch universities and Eyra.
The consortium is composed of researchers from:
- University of Amsterdam
- Radboud University Nijmegen
- VU Amsterdam
- Utrecht University
- Tilburg University
- Erasmus University Rotterdam
The first phase of the project ended in December 2022 and resulted in an MVP solution to run one Port app on top of a Next bundle. This Next + Port combi can be released as a Docker image and deployed on Azure Web App Service.