# Getting started

The simplest way to get started with trawler is using docker.

# Trawler platform

The trawler platform can be run as a single docker image. All it needs is a running postgres database. The simplest way to get started is using docker-compose.

curl https://raw.githubusercontent.com/scalar-dev/trawler/master/docker-compose.example.yml
docker-compose -f docker-compose.example.yml up

# Agent

The trawler agent connects to your databases and other systems in order to extract metadata and upload it to the trawler platform. This is also most easily run with docker.

First, you'll need to create a user to upload data. This is currently a bit of a faff will be easier in the near future. Firstly create a user:

curl -g \
-H "Content-Type: application/json" \
-d '{"query":"mutation { createUser(email: \"me@example.com\", password: \"password\") }"}' \

This should give you a <userid>.

You'll then need to create yourself a project and a role in the postgres (e.g. using psql):

insert into project values('63255f7a-e383-457a-9c30-4c7f95308749', 'test')
insert into account_role(account_id, project_id, role) values('<userid>', '63255f7a-e383-457a-9c30-4c7f95308749', 'admin');

Then login with this user's credentials get a graphql JWT:

curl -g \
-H "Content-Type: application/json" \
-d '{"query":"mutation { login(email: \"me@example.com\", password: \"password\") { jwt } }"}' \

Finally take this token and create an API key for yourself

curl -g \
-H "Content-Type: application/json" \
-H "Authorization: Bearer <token>" \
-d '{"query":"mutation { createApiKey { secret } }"}' \

The key from this step will be used as TRAWLER_TOKEN below.

Create an agent configuration file, example.yml:

- type: sql
    every: minutes
    interval: 60
    on_startup: true
  override_dbname: "my-db-name"
  # Replace this with your actual database URI
  uri: postgresql://postgres:postgres@localhost/postgres%       

Then run:

docker run --net=host \
    -v $(pwd)/example.yml:/app/config/example.yml \
    -e TRAWLER_TOKEN=<token> \
    -e TRAWLER_ENDPOINT=http://localhost:9090 \
    scalardev/trawler-agent \
    run --now /app/config/example.yml 

This should scan the database and upload it to your local trawler instance.