siulkilulki
c99c218436
Add comments to makefile. Fix get_utterances condition. Adjust craweler settings. Change split-data script
62 lines
2.4 KiB
Makefile
62 lines
2.4 KiB
Makefile
SHELL := /bin/bash
|
|
PREPARE_ENVIRONMENT := $(shell ./prepare-environment.sh > /tmp/makeenv)
|
|
include /tmp/makeenv
|
|
JOBS := 100
|
|
|
|
.PHONY: all update data clean clean-data clean-cache clean-ml
|
|
|
|
all: data
|
|
|
|
# annotator_console.py needs running redis instance
|
|
# move belo
|
|
score.txt: predicted.txt test.txt
|
|
paste $< <(egrep -o "__label__(yes|no)" $(word 2,$^)) | ./evaluate.py > $@
|
|
|
|
predicted.txt: fs-model.bin test.txt
|
|
./fasttext predict $< $(word 2,$^) > $@
|
|
|
|
fs-model.bin: train.txt
|
|
./fasttext supervised -input $< -output `basename $@ .bin` #-dim 300 -ws 10 -wordNgrams 2 -loss ns
|
|
|
|
train.txt test.txt dev.txt: ./annotator_console.py tsv2fasttext.py split-data.sh
|
|
./$< 2tsv | ./$(word 2,$^) > all.txt
|
|
# paste -d ' ' <(cut -f1 -d' ' all.txt) <(cut -d' ' -f2- all.txt | ./tokenizer.perl -threads 12 -l pl) | sponge all.txt
|
|
./split-data.sh all.txt
|
|
rm all.txt
|
|
|
|
clean-ml:
|
|
rm -f train.txt test.txt dev.txt fs-model* predicted.txt score.txt
|
|
|
|
parish2text: parishwebsites/parish2text.py parishwebsites/parish2text-commands.sh
|
|
mkdir -p parishwebsites/{text-data,text-data-logs}
|
|
cd parishwebsites && ./parish2text-commands.sh data > p2t-commands.txt && parallel --jobs -2 < p2t-commands.txt
|
|
|
|
data-add: parishwebsites/spider-commands-add.txt parishwebsites/domain-blacklist.txt parishwebsites/deal-with-not-completed.sh
|
|
cd parishwebsites && ./deal-with-not-completed.sh
|
|
cd parishwebsites && parallel --jobs $(JOBS) < spider-commands-add.txt
|
|
|
|
data: parishwebsites/spider-commands.txt parishwebsites/domain-blacklist.txt
|
|
rm -f parishwebsites/*processed.txt
|
|
cd parishwebsites && parallel --jobs $(JOBS) < spider-commands.txt
|
|
|
|
parishwebsites/spider-commands.txt: parishes-with-urls.tsv parishwebsites/domain-blacklist.txt
|
|
cut -f3 $< | tail -n +2 | grep http | parishwebsites/generate_spider_commands.sh | sort -u | parishwebsites/remove_blacklisted.py $(word 2,$^) | parishwebsites/remove_duplicate_commands.py > $@
|
|
|
|
parishes-with-urls.tsv: apikey.txt parishes-deon.tsv scraper/get_parishes_urls.py
|
|
scraper/get_parishes_urls.py -a $< -p $(word 2,$^) >> $@ 2> get-parishes-urls.log
|
|
|
|
parishes-deon.tsv: scraper/crawl_deon.py
|
|
scraper/crawl_deon.py > $@ 2> crawl-deon.log
|
|
|
|
update: environment.yml
|
|
conda env update -f $<
|
|
|
|
clean:
|
|
rm -rf parishes-deon.tsv parishes-with-urls.tsv spider-commands.txt
|
|
|
|
clean-data:
|
|
rm -rf parishwebsites/{data,processed.txt,crawler-log.txt}
|
|
|
|
clean-cache:
|
|
rm -rf parishwebsites/.scrapy/httpcache
|