102 lines
3.1 KiB
Markdown
102 lines
3.1 KiB
Markdown
Run your own inventaire in a docker environment
|
|
|
|
## Requirements
|
|
|
|
- [docker-compose](https://docs.docker.com/compose/gettingstarted/) up and ready
|
|
- git
|
|
|
|
## Install
|
|
|
|
```
|
|
git clone https://github.com/inventaire/inventaire-docker.git
|
|
```
|
|
|
|
got to `cd inventaire-docker`
|
|
|
|
clone the two repos inventaire needs to run :
|
|
|
|
- `inventaire` core application server -> [setup](https://github.com/inventaire/inventaire#installation)
|
|
- `entities-search-engine` for querying entities -> [go to repo](https://github.com/inventaire/entities-search-engine)
|
|
|
|
```
|
|
git clone https://github.com/inventaire/inventaire.git
|
|
git clone https://github.com/inventaire/entities-search-engine.git
|
|
```
|
|
|
|
Create empty folders for docker volume to dump backup. In accordance with docker-compose volumes, example: `mkdir data couch-test couch es`
|
|
|
|
Start the magic, build everything at once !
|
|
|
|
```
|
|
docker-compose up --build
|
|
```
|
|
|
|
## Useful commands
|
|
|
|
`docker-compose up` : start containers if already built
|
|
|
|
`docker-compose down` : kill active containers
|
|
|
|
`docker rm $(docker ps -a -q)` : delete stopped containers
|
|
|
|
`docker rmi $(docker images -q -f dangling=true)` : delete untagged images
|
|
|
|
Check out [official doc](https://docs.docker.com/compose/)
|
|
|
|
## Load wikidata into elasticsearch
|
|
|
|
Make sure ES import limit is above entities-search-engige import rate, by [closing the index](https://www.elastic.co/guide/en/elasticsearch/reference/current/indices-open-close.html) first, raising the limit & reopen the index
|
|
|
|
```
|
|
curl -XPOST http://localhost:9200/wikidata/_close
|
|
curl -XPUT http://localhost:9200/wikidata/_settings -d '{"index.mapping.total_fields.limit": 20000}'
|
|
curl -XPOST http://localhost:9200/wikidata/_open
|
|
```
|
|
|
|
start the containers `docker-compose up`
|
|
|
|
```
|
|
claim=P31:Q5
|
|
type=humans
|
|
docker-compose exec entities-search-engine ./bin/dump_wikidata_subset $claim $type
|
|
```
|
|
|
|
[more info on importing some wikidata items](https://github.com/inventaire/inventaire-deploy/install_entities_search_engine)
|
|
|
|
more docs [wikidata filtered dump import](https://github.com/inventaire/entities-search-engine/blob/master/docs/wikidata_filtered_dump_import.mdFv)
|
|
|
|
## Fixtures
|
|
|
|
In case you would like to play with out-of-the-box data.
|
|
|
|
Run api tests to populate tests dbs (see Tests section)
|
|
```
|
|
docker-compose -f docker-compose.yml -f docker-compose.test.yml up exec inventaire npm run test-api
|
|
```
|
|
|
|
- Replicate `*-tests` dbs documents into `*` dbs
|
|
|
|
```
|
|
`docker-compose exec inventaire npm run replicate-tests-db`
|
|
```
|
|
|
|
## Tests
|
|
|
|
Start services with test environnement with [multiple compose files](https://docs.docker.com/compose/extends/#understanding-multiple-compose-files)
|
|
|
|
```
|
|
docker-compose -f docker-compose.yml -f docker-compose.test.yml up
|
|
```
|
|
|
|
Execute tests script
|
|
|
|
`docker-compose exec inventaire npm run test-api`
|
|
|
|
or execute directly the test command
|
|
|
|
`docker-compose exec inventaire ./node_modules/.bin/mocha --compilers coffee:coffee-script/register --timeout 20000 /opt/inventaire/path/to/test/file`
|
|
|
|
Tip : create a symbolic link on your machine between the inventaire folder and docker working directory on your machine at `/opt/`, in order to autocomplete path to test file to execute
|
|
|
|
`sudo ln ~/path/to/inventaire-docker/inventaire /opt -s`
|