Commit Graph

115 Commits

Author SHA1 Message Date
61631c52a3 lexicon search 2017-10-10 15:39:47 +02:00
5e809efcce corrected tokenizer 2017-05-05 12:58:32 +02:00
96a5bc3108 original sentence in tokenized sentence 2017-04-28 13:48:32 +02:00
4faae4e91a slight change 2017-04-27 13:52:03 +02:00
dceb0d9f47 date recognition 2017-04-27 10:37:29 +02:00
bd73749388 new tokenizer 2017-04-26 17:02:18 +02:00
a0673df75a cpplint corrections 2017-04-22 23:47:48 +02:00
970dda5dc2 option of white space tokenization while searching 2017-04-22 23:45:51 +02:00
31e4f091ad mutliple results 2017-04-21 14:51:58 +02:00
c3826919ba changes in CMakeLists.txt 2017-03-03 11:28:54 +01:00
cf7b1592f7 updated todo 2016-11-01 22:23:30 +01:00
7e005bfca7 changed significance factor to 2 2016-10-22 18:02:04 +02:00
8bc739ff20 added boundary on simple search results 2016-01-25 22:42:42 +01:00
b3d7c993aa tokenize only option - no word map 2016-01-01 20:45:07 +01:00
bbf3853d2a added lowercasing when tokenizing by space 2015-12-29 21:44:46 +01:00
0a8d2fdd39 tokenize by whitespace option 2015-12-27 20:54:40 +01:00
873d7c300c added parameterless constructor for concordia 2015-10-19 15:38:10 +02:00
1adabf4833 add index path as required argument to concordia constructor 2015-10-16 22:14:11 +02:00
f585ff9e01 corpus figures creator 2015-10-06 13:34:03 +02:00
96c74c47ac corpus analyzer 2015-10-04 16:24:58 +02:00
2601dc83bf test corpus for corpus analyzer 2015-10-03 16:19:10 +02:00
4e17e28f7f working corpus analyzer 2015-10-03 16:18:49 +02:00
fa3138df29 count occurences feature 2015-10-01 13:36:54 +02:00
fd32ff7e12 todo 2015-09-07 08:15:46 +02:00
cdeb57ccfa todo 2015-08-26 20:14:43 +02:00
bd62420cd5 updated tutorial 2015-08-24 14:30:20 +02:00
0a3fd8a04e added an extremely important improvement to the concordia search algorithm - gapped overlays cut-off 2015-08-24 13:10:06 +02:00
209e374226 repaired concordia test 2015-08-19 20:53:40 +02:00
68fecaddf8 adding all tokenized examples 2015-08-19 20:49:26 +02:00
a765443a01 simple search returns matched pattern fragments 2015-08-07 12:54:57 +02:00
28704c2f43 separated tokenization and adding to index 2015-08-01 17:03:39 +02:00
5a57406875 finished original word positions 2015-06-27 12:40:24 +02:00
a8c5fa0c75 original word positions 2015-06-27 10:09:49 +02:00
dba70b4e24 done word positions 2015-06-26 22:50:53 +02:00
724bf0d080 new responsibilities of tokenized sentence 2015-06-26 15:38:24 +02:00
9b1735516c working sentence tokenizer 2015-06-25 20:49:22 +02:00
8432dd321f tokenizer in progress 2015-06-25 10:12:51 +02:00
0baf3e4ef2 character intervals in progress 2015-06-22 13:52:56 +02:00
4c0f2fd08d modified todo 2015-06-12 12:25:02 +02:00
dff52abff7 modified todo 2015-06-11 11:17:45 +02:00
680eb54ae5 modified todo, removed concordia-server 2015-06-09 13:09:10 +02:00
07d5d4438b clear index, examples 2015-05-04 20:40:44 +02:00
abbd5b1ae8 finished documentation 2015-05-01 14:52:53 +02:00
9e550ca1cf more doc 2015-04-30 22:22:54 +02:00
87a26bfa3b cleaned configuration, doc 2015-04-30 21:15:18 +02:00
b790c6898f stable release 2015-04-30 09:29:10 +02:00
db63cf776e doc 2015-04-28 21:34:07 +02:00
952b94971f only configure prod resources if available 2015-04-27 16:20:54 +02:00
bb7608d05e anubis searcher -> concordia searcher
Former-commit-id: 8afe194adf3163ee62caa30732d9c9dd095df66b
2015-04-24 11:48:32 +02:00
23aa113747 order in scripts
Former-commit-id: 945423dec6f4007f780b27fd590fb09578117b54
2015-04-24 11:10:17 +02:00