-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathfullrefs.bib
2389 lines (2212 loc) · 240 KB
/
fullrefs.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@misc{wandb,
title = {Experiment Tracking with Weights and Biases},
year = {2020},
note = {Software available from wandb.com},
url={https://www.wandb.com/},
author = {Biewald, Lukas},
}
@ARTICLE{2020SciPy-NMeth,
author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and
Haberland, Matt and Reddy, Tyler and Cournapeau, David and
Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and
Bright, Jonathan and {van der Walt}, St{\'e}fan J. and
Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and
Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and
Kern, Robert and Larson, Eric and Carey, C J and
Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and
{VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and
Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and
Harris, Charles R. and Archibald, Anne M. and
Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and
{van Mulbregt}, Paul and {SciPy 1.0 Contributors}},
title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific
Computing in Python}},
journal = {Nature Methods},
year = {2020},
volume = {17},
pages = {261--272},
adsurl = {https://rdcu.be/b08Wh},
doi = {10.1038/s41592-019-0686-2},
}
@inproceedings{hoffman2016elbo,
title={Elbo surgery: yet another way to carve up the variational evidence lower bound},
author={Hoffman, Matthew D and Johnson, Matthew J},
booktitle={Workshop in Advances in Approximate Bayesian Inference, NIPS},
volume={1},
pages={2},
year={2016}
}
@article{ebner_facesdatabase_2010,
title = {{FACES}—{A} database of facial expressions in young, middle-aged, and older women and men: {Development} and validation},
volume = {42},
issn = {1554-351X, 1554-3528},
shorttitle = {{FACES}—{A} database of facial expressions in young, middle-aged, and older women and men},
url = {http://link.springer.com/10.3758/BRM.42.1.351},
doi = {10.3758/BRM.42.1.351},
language = {en},
number = {1},
urldate = {2021-01-07},
journal = {Behavior Research Methods},
author = {Ebner, Natalie C. and Riediger, Michaela and Lindenberger, Ulman},
month = feb,
year = {2010},
pages = {351--362},
file = {Ebner et al. - 2010 - FACES—A database of facial expressions in young, m.pdf:/home/wolf/Zotero/storage/ZCEE2PG2/Ebner et al. - 2010 - FACES—A database of facial expressions in young, m.pdf:application/pdf},
}
@article{belli_image-conditioned_2019,
title = {Image-{Conditioned} {Graph} {Generation} for {Road} {Network} {Extraction}},
url = {http://arxiv.org/abs/1910.14388},
abstract = {Deep generative models for graphs have shown great promise in the area of drug design, but have so far found little application beyond generating graph-structured molecules. In this work, we demonstrate a proof of concept for the challenging task of road network extraction from image data. This task can be framed as image-conditioned graph generation, for which we develop the Generative Graph Transformer (GGT), a deep autoregressive model that makes use of attention mechanisms for image conditioning and the recurrent generation of graphs. We benchmark GGT on the application of road network extraction from semantic segmentation data. For this, we introduce the Toulouse Road Network dataset, based on real-world publicly-available data. We further propose the StreetMover distance: a metric based on the Sinkhorn distance for effectively evaluating the quality of road network generation. The code and dataset are publicly available.},
urldate = {2020-05-08},
journal = {arXiv:1910.14388 [cs, stat]},
author = {Belli, Davide and Kipf, Thomas},
month = oct,
year = {2019},
note = {arXiv: 1910.14388},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning},
annote = {Comment: Presented at NeurIPS 2019 Workshop on Graph Representation Learning},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/BA9GESIJ/1910.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/J62CRRMB/Belli and Kipf - 2019 - Image-Conditioned Graph Generation for Road Networ.pdf:application/pdf},
}
@article{kingma_introduction_2019,
title = {An {Introduction} to {Variational} {Autoencoders}},
volume = {12},
issn = {1935-8237, 1935-8245},
url = {https://www.nowpublishers.com/article/Details/MAL-056},
doi = {10.1561/2200000056},
abstract = {An Introduction to Variational Autoencoders},
language = {English},
number = {4},
urldate = {2020-05-07},
journal = {Foundations and Trends® in Machine Learning},
author = {Kingma, Diederik P. and Welling, Max},
month = nov,
year = {2019},
note = {Publisher: Now Publishers, Inc.},
pages = {307--392},
file = {Snapshot:/home/wolf/Zotero/storage/K2AB4SFQ/MAL-056.html:text/html;Full Text PDF:/home/wolf/Zotero/storage/8I69DBEQ/Kingma and Welling - 2019 - An Introduction to Variational Autoencoders.pdf:application/pdf},
}
@article{yang_feedback_2020,
title = {Feedback {Recurrent} {AutoEncoder}},
url = {http://arxiv.org/abs/1911.04018},
abstract = {In this work, we propose a new recurrent autoencoder architecture, termed Feedback Recurrent AutoEncoder (FRAE), for online compression of sequential data with temporal dependency. The recurrent structure of FRAE is designed to efficiently extract the redundancy along the time dimension and allows a compact discrete representation of the data to be learned. We demonstrate its effectiveness in speech spectrogram compression. Specifically, we show that the FRAE, paired with a powerful neural vocoder, can produce high-quality speech waveforms at a low, fixed bitrate. We further show that by adding a learned prior for the latent space and using an entropy coder, we can achieve an even lower variable bitrate.},
urldate = {2020-05-07},
journal = {arXiv:1911.04018 [cs, eess, stat]},
author = {Yang, Yang and Sautière, Guillaume and Ryu, J. Jon and Cohen, Taco S.},
month = feb,
year = {2020},
note = {arXiv: 1911.04018},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning, Computer Science - Sound, Electrical Engineering and Systems Science - Audio and Speech Processing},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/UJVMX3D7/1911.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/8X92ZTZG/Yang et al. - 2020 - Feedback Recurrent AutoEncoder.pdf:application/pdf},
}
@article{rezende_variational_2016,
title = {Variational {Inference} with {Normalizing} {Flows}},
url = {http://arxiv.org/abs/1505.05770},
abstract = {The choice of approximate posterior distribution is one of the core problems in variational inference. Most applications of variational inference employ simple families of posterior approximations in order to allow for efficient inference, focusing on mean-field or other simple structured approximations. This restriction has a significant impact on the quality of inferences made using variational methods. We introduce a new approach for specifying flexible, arbitrarily complex and scalable approximate posterior distributions. Our approximations are distributions constructed through a normalizing flow, whereby a simple initial density is transformed into a more complex one by applying a sequence of invertible transformations until a desired level of complexity is attained. We use this view of normalizing flows to develop categories of finite and infinitesimal flows and provide a unified view of approaches for constructing rich posterior approximations. We demonstrate that the theoretical advantages of having posteriors that better match the true posterior, combined with the scalability of amortized variational approaches, provides a clear improvement in performance and applicability of variational inference.},
urldate = {2020-05-05},
journal = {arXiv:1505.05770 [cs, stat]},
author = {Rezende, Danilo Jimenez and Mohamed, Shakir},
month = jun,
year = {2016},
note = {arXiv: 1505.05770},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning, Computer Science - Artificial Intelligence, Statistics - Computation, Statistics - Methodology},
annote = {Comment: Proceedings of the 32nd International Conference on Machine Learning},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/RK8HUUL2/1505.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/ZYV38DAF/Rezende and Mohamed - 2016 - Variational Inference with Normalizing Flows.pdf:application/pdf},
}
@inproceedings{jin_unsupervised_2019,
address = {Florence, Italy},
title = {Unsupervised {Learning} of {PCFGs} with {Normalizing} {Flow}},
url = {https://www.aclweb.org/anthology/P19-1234},
doi = {10.18653/v1/P19-1234},
abstract = {Unsupervised PCFG inducers hypothesize sets of compact context-free rules as explanations for sentences. PCFG induction not only provides tools for low-resource languages, but also plays an important role in modeling language acquisition (Bannard et al., 2009; Abend et al. 2017). However, current PCFG induction models, using word tokens as input, are unable to incorporate semantics and morphology into induction, and may encounter issues of sparse vocabulary when facing morphologically rich languages. This paper describes a neural PCFG inducer which employs context embeddings (Peters et al., 2018) in a normalizing flow model (Dinh et al., 2015) to extend PCFG induction to use semantic and morphological information. Linguistically motivated sparsity and categorical distance constraints are imposed on the inducer as regularization. Experiments show that the PCFG induction model with normalizing flow produces grammars with state-of-the-art accuracy on a variety of different languages. Ablation further shows a positive effect of normalizing flow, context embeddings and proposed regularizers.},
urldate = {2020-05-05},
booktitle = {Proceedings of the 57th {Annual} {Meeting} of the {Association} for {Computational} {Linguistics}},
publisher = {Association for Computational Linguistics},
author = {Jin, Lifeng and Doshi-Velez, Finale and Miller, Timothy and Schwartz, Lane and Schuler, William},
month = jul,
year = {2019},
pages = {2442--2452},
file = {Full Text PDF:/home/wolf/Zotero/storage/IYM2VFG9/Jin et al. - 2019 - Unsupervised Learning of PCFGs with Normalizing Fl.pdf:application/pdf},
}
@article{che_towards_2018,
title = {Towards {Better} {UD} {Parsing}: {Deep} {Contextualized} {Word} {Embeddings}, {Ensemble}, and {Treebank} {Concatenation}},
shorttitle = {Towards {Better} {UD} {Parsing}},
url = {http://arxiv.org/abs/1807.03121},
abstract = {This paper describes our system (HIT-SCIR) submitted to the CoNLL 2018 shared task on Multilingual Parsing from Raw Text to Universal Dependencies. We base our submission on Stanford's winning system for the CoNLL 2017 shared task and make two effective extensions: 1) incorporating deep contextualized word embeddings into both the part of speech tagger and parser; 2) ensembling parsers trained with different initialization. We also explore different ways of concatenating treebanks for further improvements. Experimental results on the development data show the effectiveness of our methods. In the final evaluation, our system was ranked first according to LAS (75.84\%) and outperformed the other systems by a large margin.},
urldate = {2020-05-05},
journal = {arXiv:1807.03121 [cs]},
author = {Che, Wanxiang and Liu, Yijia and Wang, Yuxuan and Zheng, Bo and Liu, Ting},
month = jul,
year = {2018},
note = {arXiv: 1807.03121},
keywords = {Computer Science - Computation and Language},
annote = {Comment: System description paper of our system (HIT-SCIR) for the CoNLL 2018 shared task on Universal Dependency parsing, which was ranked first in the LAS evaluation. Fix typos and grammar errors. Add the results of parser without ensemble},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/KL6352GX/1807.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/IMFY9XFP/Che et al. - 2018 - Towards Better UD Parsing Deep Contextualized Wor.pdf:application/pdf},
}
@article{shazeer_swivel_2016,
title = {Swivel: {Improving} {Embeddings} by {Noticing} {What}'s {Missing}},
shorttitle = {Swivel},
url = {http://arxiv.org/abs/1602.02215},
abstract = {We present Submatrix-wise Vector Embedding Learner (Swivel), a method for generating low-dimensional feature embeddings from a feature co-occurrence matrix. Swivel performs approximate factorization of the point-wise mutual information matrix via stochastic gradient descent. It uses a piecewise loss with special handling for unobserved co-occurrences, and thus makes use of all the information in the matrix. While this requires computation proportional to the size of the entire matrix, we make use of vectorized multiplication to process thousands of rows and columns at once to compute millions of predicted values. Furthermore, we partition the matrix into shards in order to parallelize the computation across many nodes. This approach results in more accurate embeddings than can be achieved with methods that consider only observed co-occurrences, and can scale to much larger corpora than can be handled with sampling methods.},
urldate = {2020-05-04},
journal = {arXiv:1602.02215 [cs]},
author = {Shazeer, Noam and Doherty, Ryan and Evans, Colin and Waterson, Chris},
month = feb,
year = {2016},
note = {arXiv: 1602.02215},
keywords = {Computer Science - Computation and Language},
annote = {Comment: 9 pages, 4 figures},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/8IXFYWWP/1602.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/D2WWGN24/Shazeer et al. - 2016 - Swivel Improving Embeddings by Noticing What's Mi.pdf:application/pdf},
}
@article{vrandevcic2014wikidata,
title={Wikidata: a free collaborative knowledgebase},
author={Vrande{\v{c}}i{\'c}, Denny and Kr{\"o}tzsch, Markus},
journal={Communications of the ACM},
volume={57},
number={10},
pages={78--85},
year={2014},
publisher={ACM New York, NY, USA}
}
@inproceedings{perozzi2014deepwalk,
title={Deepwalk: Online learning of social representations},
author={Perozzi, Bryan and Al-Rfou, Rami and Skiena, Steven},
booktitle={Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining},
pages={701--710},
year={2014}
}
@article{goodfellow2014generative,
title={Generative adversarial nets},
author={Goodfellow, Ian and Pouget-Abadie, Jean and Mirza, Mehdi and Xu, Bing and Warde-Farley, David and Ozair, Sherjil and Courville, Aaron and Bengio, Yoshua},
journal={Advances in neural information processing systems},
volume={27},
pages={2672--2680},
year={2014}
}
@inproceedings{glorot2010understanding,
title={Understanding the difficulty of training deep feedforward neural networks},
author={Glorot, Xavier and Bengio, Yoshua},
booktitle={Proceedings of the thirteenth international conference on artificial intelligence and statistics},
pages={249--256},
year={2010}
}
@article{fengvaleriu,
title={Valeriu Codreanu, SURFsara, Netherlands Ian Foster, UChicago \& ANL, USA Zhao Zhang, TACC, USA},
author={Feng, Song and Torsten Hoefler, ETH and Li, Switzerland Jessy and Podareanu, Damian and Pu, Qifan and Qiu, Judy and Saletore, Vikram and Smorkalov, Mikhail E and Torres, Jordi}
}
@article{tang2011leveraging,
title={Leveraging social media networks for classification},
author={Tang, Lei and Liu, Huan},
journal={Data Mining and Knowledge Discovery},
volume={23},
number={3},
pages={447--478},
year={2011},
publisher={Springer}
}
@article{lerer_pytorch-biggraph_nodate,
title = {{PyTorch}-{BigGraph}: {A} {Large}-scale {Graph} {Embedding} {System}},
abstract = {Graph embedding methods produce unsupervised node features from graphs that can then be used for a variety of machine learning tasks. Modern graphs, particularly in industrial applications, contain billions of nodes and trillions of edges, which exceeds the capability of existing embedding systems. We present PyTorch-BigGraph (PBG), an embedding system that incorporates several modifications to traditional multi-relation embedding systems that allow it to scale to graphs with billions of nodes and trillions of edges. PBG uses graph partitioning to train arbitrarily large embeddings on either a single machine or in a distributed environment. We demonstrate comparable performance with existing embedding systems on common benchmarks, while allowing for scaling to arbitrarily large graphs and parallelization on multiple machines. We train and evaluate embeddings on several large social network graphs as well as the full Freebase dataset, which contains over 100 million nodes and 2 billion edges.},
language = {en},
author = {Lerer, Adam and Wu, Ledell and Shen, Jiajun and Lacroix, Timothee and Wehrstedt, Luca and Bose, Abhijit and Peysakhovich, Alex},
pages = {12},
file = {Lerer et al. - PyTorch-BigGraph A Large-scale Graph Embedding Sy.pdf:/home/wolf/Zotero/storage/3K2HXJB3/Lerer et al. - PyTorch-BigGraph A Large-scale Graph Embedding Sy.pdf:application/pdf},
}
@inproceedings{karim_drug-drug_2019,
title = {Drug-drug interaction prediction based on knowledge graph embeddings and convolutional-{LSTM} network},
url = {https://research.vu.nl/en/publications/drug-drug-interaction-prediction-based-on-knowledge-graph-embeddi},
doi = {10.1145/3307339.3342161},
language = {English},
urldate = {2020-05-01},
booktitle = {{ACM}-{BCB} 2019 - {Proceedings} of the 10th {ACM} {International} {Conference} on {Bioinformatics}, {Computational} {Biology} and {Health} {Informatics}},
publisher = {Association for Computing Machinery, Inc},
author = {Karim, Md Rezaul and Cochez, Michael and Jares, Joao Bosco and Uddin, Mamtaz and Beyan, Oya and Decker, Stefan},
month = sep,
year = {2019},
pages = {113--123},
file = {Snapshot:/home/wolf/Zotero/storage/TBI6IG9Z/drug-drug-interaction-prediction-based-on-knowledge-graph-embeddi.html:text/html},
}
@incollection{groth_rdf2vec_2016,
address = {Cham},
title = {{RDF2Vec}: {RDF} {Graph} {Embeddings} for {Data} {Mining}},
volume = {9981},
isbn = {978-3-319-46522-7 978-3-319-46523-4},
shorttitle = {{RDF2Vec}},
url = {http://link.springer.com/10.1007/978-3-319-46523-4_30},
abstract = {Linked Open Data has been recognized as a valuable source for background information in data mining. However, most data mining tools require features in propositional form, i.e., a vector of nominal or numerical features associated with an instance, while Linked Open Data sources are graphs by nature. In this paper, we present RDF2Vec, an approach that uses language modeling approaches for unsupervised feature extraction from sequences of words, and adapts them to RDF graphs. We generate sequences by leveraging local information from graph substructures, harvested by Weisfeiler-Lehman Subtree RDF Graph Kernels and graph walks, and learn latent numerical representations of entities in RDF graphs. Our evaluation shows that such vector representations outperform existing techniques for the propositionalization of RDF graphs on a variety of different predictive machine learning tasks, and that feature vector representations of general knowledge graphs such as DBpedia and Wikidata can be easily reused for different tasks.},
language = {en},
urldate = {2020-05-01},
booktitle = {The {Semantic} {Web} – {ISWC} 2016},
publisher = {Springer International Publishing},
author = {Ristoski, Petar and Paulheim, Heiko},
editor = {Groth, Paul and Simperl, Elena and Gray, Alasdair and Sabou, Marta and Krötzsch, Markus and Lecue, Freddy and Flöck, Fabian and Gil, Yolanda},
year = {2016},
doi = {10.1007/978-3-319-46523-4_30},
note = {Series Title: Lecture Notes in Computer Science},
pages = {498--514},
file = {Ristoski and Paulheim - 2016 - RDF2Vec RDF Graph Embeddings for Data Mining.pdf:/home/wolf/Zotero/storage/SNHAH8XG/Ristoski and Paulheim - 2016 - RDF2Vec RDF Graph Embeddings for Data Mining.pdf:application/pdf},
}
@article{wilcke_knowledge_2017,
title = {The knowledge graph as the default data model for learning on heterogeneous knowledge},
volume = {1},
issn = {2451-8484},
url = {https://research.vu.nl/en/publications/the-knowledge-graph-as-the-default-data-model-for-machine-learnin},
doi = {10.3233/DS-170007},
language = {English},
number = {1-2},
urldate = {2020-05-01},
journal = {Data Science},
author = {Wilcke, Xander and Bloem, Peter and Boer, Victor De},
month = dec,
year = {2017},
note = {Publisher: IOS Press},
pages = {39--57},
file = {Full Text:/home/wolf/Zotero/storage/WXXSCZZ7/Wilcke et al. - 2017 - The knowledge graph as the default data model for .pdf:application/pdf;Snapshot:/home/wolf/Zotero/storage/GDBR87DG/the-knowledge-graph-as-the-default-data-model-for-machine-learnin.html:text/html},
}
@inproceedings{wilcke_knowledge_2018,
title = {The {Knowledge} {Graph} for {End}-to-{End} {Learning} on {Heterogeneous} {Knowledge}},
url = {https://research.vu.nl/en/publications/the-knowledge-graph-for-end-to-end-learning-on-heterogeneous-know-2},
language = {English},
urldate = {2020-05-01},
author = {Wilcke, W. X. and Bloem, P. and Boer, Viktor de},
month = mar,
year = {2018},
file = {Snapshot:/home/wolf/Zotero/storage/XMFFDMBU/the-knowledge-graph-for-end-to-end-learning-on-heterogeneous-know-2.html:text/html;Full Text PDF:/home/wolf/Zotero/storage/W9K67T7R/Wilcke et al. - 2018 - The Knowledge Graph for End-to-End Learning on Het.pdf:application/pdf},
}
@incollection{socher_reasoning_2013,
title = {Reasoning {With} {Neural} {Tensor} {Networks} for {Knowledge} {Base} {Completion}},
url = {http://papers.nips.cc/paper/5028-reasoning-with-neural-tensor-networks-for-knowledge-base-completion.pdf},
urldate = {2020-04-20},
booktitle = {Advances in {Neural} {Information} {Processing} {Systems} 26},
publisher = {Curran Associates, Inc.},
author = {Socher, Richard and Chen, Danqi and Manning, Christopher D and Ng, Andrew},
editor = {Burges, C. J. C. and Bottou, L. and Welling, M. and Ghahramani, Z. and Weinberger, K. Q.},
year = {2013},
pages = {926--934},
file = {NIPS Snapshot:/home/wolf/Zotero/storage/CRWB7IY5/5028-reasoning-with-neural-tensor-networks-for-knowledge-base-completion.html:text/html;NIPS Full Text PDF:/home/wolf/Zotero/storage/WINTUXFA/Socher et al. - 2013 - Reasoning With Neural Tensor Networks for Knowledg.pdf:application/pdf},
}
@article{serafini_logic_2016,
title = {Logic {Tensor} {Networks}: {Deep} {Learning} and {Logical} {Reasoning} from {Data} and {Knowledge}},
shorttitle = {Logic {Tensor} {Networks}},
url = {http://arxiv.org/abs/1606.04422},
abstract = {We propose Logic Tensor Networks: a uniform framework for integrating automatic learning and reasoning. A logic formalism called Real Logic is defined on a first-order language whereby formulas have truth-value in the interval [0,1] and semantics defined concretely on the domain of real numbers. Logical constants are interpreted as feature vectors of real numbers. Real Logic promotes a well-founded integration of deductive reasoning on a knowledge-base and efficient data-driven relational machine learning. We show how Real Logic can be implemented in deep Tensor Neural Networks with the use of Google's tensorflow primitives. The paper concludes with experiments applying Logic Tensor Networks on a simple but representative example of knowledge completion.},
urldate = {2020-04-17},
journal = {arXiv:1606.04422 [cs]},
author = {Serafini, Luciano and Garcez, Artur d'Avila},
month = jul,
year = {2016},
note = {arXiv: 1606.04422},
keywords = {Computer Science - Machine Learning, Computer Science - Artificial Intelligence, Computer Science - Logic in Computer Science, Computer Science - Neural and Evolutionary Computing},
annote = {Comment: 12 pages, 2 figs, 1 table, 27 references},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/UB6KLTXA/1606.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/52JYXFJI/Serafini and Garcez - 2016 - Logic Tensor Networks Deep Learning and Logical R.pdf:application/pdf},
}
@article{van_krieken_integrated_nodate,
title = {Integrated {Learning} and {Reasoning} using {Gradient} {Descent}},
language = {en},
author = {van Krieken, Emile},
pages = {85},
file = {f106624807.pdf:/home/wolf/Zotero/storage/Q74UGEP2/van Krieken - Integrated Learning and Reasoning using Gradient D.pdf:application/pdf},
}
@article{wilcke_end--end_2020,
title = {End-to-{End} {Entity} {Classification} on {Multimodal} {Knowledge} {Graphs}},
url = {http://arxiv.org/abs/2003.12383},
abstract = {End-to-end multimodal learning on knowledge graphs has been left largely unaddressed. Instead, most end-to-end models such as message passing networks learn solely from the relational information encoded in graphs' structure: raw values, or literals, are either omitted completely or are stripped from their values and treated as regular nodes. In either case we lose potentially relevant information which could have otherwise been exploited by our learning methods. To avoid this, we must treat literals and non-literals as separate cases. We must also address each modality separately and accordingly: numbers, texts, images, geometries, et cetera. We propose a multimodal message passing network which not only learns end-to-end from the structure of graphs, but also from their possibly divers set of multimodal node features. Our model uses dedicated (neural) encoders to naturally learn embeddings for node features belonging to five different types of modalities, including images and geometries, which are projected into a joint representation space together with their relational information. We demonstrate our model on a node classification task, and evaluate the effect that each modality has on the overall performance. Our result supports our hypothesis that including information from multiple modalities can help our models obtain a better overall performance.},
urldate = {2020-04-15},
journal = {arXiv:2003.12383 [cs]},
author = {Wilcke, W. X. and Bloem, P. and de Boer, V. and Veer, R. H. van t and van Harmelen, F. A. H.},
month = mar,
year = {2020},
note = {arXiv: 2003.12383},
keywords = {Computer Science - Machine Learning, Computer Science - Artificial Intelligence, Computer Science - Computation and Language, Computer Science - Computer Vision and Pattern Recognition},
annote = {Comment: Submitted to the 17th International Conference on Principles of Knowledge Representation and Reasoning (2020)},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/DAZZK22X/2003.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/DGIQXPX7/Wilcke et al. - 2020 - End-to-End Entity Classification on Multimodal Kno.pdf:application/pdf},
}
@article{wang_submassive_nodate,
title = {{SUBMASSIVE}: {Resolving} {Subclass} {Cycles} in {Very} {Large} {Knowledge} {Graphs}},
abstract = {Large knowledge graphs capture information of a large number of entities and their relations. Among the many relations they capture, class subsumption assertions are usually present and expressed using the rdfs:subClassOf construct. From our examination, publicly available knowledge graphs contain many potentially erroneous cyclic subclass relations, a problem that can be exacerbated when different knowledge graphs are integrated as Linked Open Data. This paper presents an automatic approach for resolving such cycles at scale using automated reasoning by encoding the problem of cycle-resolving to a MAXSAT solver. The approach is tested on the LOD-a-lot dataset, and compared against a semi-automatic version of our algorithm. We show how the number of removed triples is a trade-off against the efficiency of the algorithm. The code and the resulting cycle-free class hierarchy of the LOD-a-lot are published at www.submassive.cc.},
language = {en},
author = {Wang, Shuai and Bloem, Peter and Raad, Joe and van Harmelen, Frank},
pages = {10},
file = {Wang et al. - SUBMASSIVE Resolving Subclass Cycles in Very Larg.pdf:/home/wolf/Zotero/storage/7RUKLRUJ/Wang et al. - SUBMASSIVE Resolving Subclass Cycles in Very Larg.pdf:application/pdf},
}
@inproceedings{grainger_semantic_2016,
title = {The {Semantic} {Knowledge} {Graph}: {A} {Compact}, {Auto}-{Generated} {Model} for {Real}-{Time} {Traversal} and {Ranking} of any {Relationship} within a {Domain}},
shorttitle = {The {Semantic} {Knowledge} {Graph}},
doi = {10.1109/DSAA.2016.51},
abstract = {This paper describes a new kind of knowledge representation and mining system which we are calling the Semantic Knowledge Graph. At its heart, the Semantic Knowledge Graph leverages an inverted index, along with a complementary uninverted index, to represent nodes (terms) and edges (the documents within intersecting postings lists for multiple terms/nodes). This provides a layer of indirection between each pair of nodes and their corresponding edge, enabling edges to materialize dynamically from underlying corpus statistics. As a result, any combination of nodes can have edges to any other nodes materialize and be scored to reveal latent relationships between the nodes. This provides numerous benefits: the knowledge graph can be built automatically from a real-world corpus of data, new nodes - along with their combined edges - can be instantly materialized from any arbitrary combination of preexisting nodes (using set operations), and a full model of the semantic relationships between all entities within a domain can be represented and dynamically traversed using a highly compact representation of the graph. Such a system has widespread applications in areas as diverse as knowledge modeling and reasoning, natural language processing, anomaly detection, data cleansing, semantic search, analytics, data classification, root cause analysis, and recommendations systems. The main contribution of this paper is the introduction of a novel system - the Semantic Knowledge Graph - which is able to dynamically discover and score interesting relationships between any arbitrary combination of entities (words, phrases, or extracted concepts) through dynamically materializing nodes and edges from a compact graphical representation built automatically from a corpus of data representative of a knowledge domain. The source code for our Semantic Knowledge Graph implementation is being published along with this paper to facilitate further research and extensions of this work.},
booktitle = {2016 {IEEE} {International} {Conference} on {Data} {Science} and {Advanced} {Analytics} ({DSAA})},
author = {Grainger, Trey and Aljadda, Khalifeh and Korayem, Mohammed and Smith, Andries},
month = oct,
year = {2016},
keywords = {anomaly detection, Anomaly Detection, auto-generated model, compact graph representation, compact graphical representation, Context, corpus statistics, data classification, data cleansing, data mining, Data models, Graph Compression, graph theory, indexing, Information Retrieval, inverted index, Knowledge Graph, knowledge modeling, knowledge representation, Learning systems, mining system, natural language processing, Natural Language Processing, Natural languages, Ontologies, Ontology Learning, Pragmatics, real-time relationship traversal, recommendations systems, Relationship Extraction, relationship ranking, root cause analysis, semantic knowledge graph, semantic search, Semantic Search, semantic Web, Semantics, Text Analytics},
pages = {420--429},
file = {IEEE Xplore Abstract Record:/home/wolf/Zotero/storage/7JDYA2HY/references.html:text/html;IEEE Xplore Full Text PDF:/home/wolf/Zotero/storage/TTT5Z28P/Grainger et al. - 2016 - The Semantic Knowledge Graph A Compact, Auto-Gene.pdf:application/pdf},
}
@inproceedings{grainger_semantic_2016-1,
title = {The {Semantic} {Knowledge} {Graph}: {A} {Compact}, {Auto}-{Generated} {Model} for {Real}-{Time} {Traversal} and {Ranking} of any {Relationship} within a {Domain}},
shorttitle = {The {Semantic} {Knowledge} {Graph}},
doi = {10.1109/DSAA.2016.51},
abstract = {This paper describes a new kind of knowledge representation and mining system which we are calling the Semantic Knowledge Graph. At its heart, the Semantic Knowledge Graph leverages an inverted index, along with a complementary uninverted index, to represent nodes (terms) and edges (the documents within intersecting postings lists for multiple terms/nodes). This provides a layer of indirection between each pair of nodes and their corresponding edge, enabling edges to materialize dynamically from underlying corpus statistics. As a result, any combination of nodes can have edges to any other nodes materialize and be scored to reveal latent relationships between the nodes. This provides numerous benefits: the knowledge graph can be built automatically from a real-world corpus of data, new nodes - along with their combined edges - can be instantly materialized from any arbitrary combination of preexisting nodes (using set operations), and a full model of the semantic relationships between all entities within a domain can be represented and dynamically traversed using a highly compact representation of the graph. Such a system has widespread applications in areas as diverse as knowledge modeling and reasoning, natural language processing, anomaly detection, data cleansing, semantic search, analytics, data classification, root cause analysis, and recommendations systems. The main contribution of this paper is the introduction of a novel system - the Semantic Knowledge Graph - which is able to dynamically discover and score interesting relationships between any arbitrary combination of entities (words, phrases, or extracted concepts) through dynamically materializing nodes and edges from a compact graphical representation built automatically from a corpus of data representative of a knowledge domain. The source code for our Semantic Knowledge Graph implementation is being published along with this paper to facilitate further research and extensions of this work.},
booktitle = {2016 {IEEE} {International} {Conference} on {Data} {Science} and {Advanced} {Analytics} ({DSAA})},
author = {Grainger, Trey and Aljadda, Khalifeh and Korayem, Mohammed and Smith, Andries},
month = oct,
year = {2016},
keywords = {anomaly detection, Anomaly Detection, auto-generated model, compact graph representation, compact graphical representation, Context, corpus statistics, data classification, data cleansing, data mining, Data models, Graph Compression, graph theory, indexing, Information Retrieval, inverted index, Knowledge Graph, knowledge modeling, knowledge representation, Learning systems, mining system, natural language processing, Natural Language Processing, Natural languages, Ontologies, Ontology Learning, Pragmatics, real-time relationship traversal, recommendations systems, Relationship Extraction, relationship ranking, root cause analysis, semantic knowledge graph, semantic search, Semantic Search, semantic Web, Semantics, Text Analytics},
pages = {420--429},
file = {IEEE Xplore Abstract Record:/home/wolf/Zotero/storage/CXYHJUKJ/7796928.html:text/html;IEEE Xplore Full Text PDF:/home/wolf/Zotero/storage/BNULC8WD/Grainger et al. - 2016 - The Semantic Knowledge Graph A Compact, Auto-Gene.pdf:application/pdf},
}
@article{byun_chronograph_2020,
title = {{ChronoGraph}: {Enabling} {Temporal} {Graph} {Traversals} for {Efficient} {Information} {Diffusion} {Analysis} over {Time}},
volume = {32},
issn = {1558-2191},
shorttitle = {{ChronoGraph}},
doi = {10.1109/TKDE.2019.2891565},
abstract = {ChronoGraph is a novel system enabling temporal graph traversals. Compared to snapshot-oriented systems, this traversal-oriented system is suitable for analyzing information diffusion over time without violating a time constraint on temporal paths. The cornerstone of ChronoGraph aims at bridging the chasm between point-based semantics and period-based semantics and the gap between temporal graph traversals and static graph traversals. Therefore, our graph model and traversal language provide the temporal syntax for both semantics, and we present a method converting point-based semantics to period-based ones. Also, ChronoGraph exploits the temporal support and parallelism to handle the temporal degree, which explosively increases compared to static graphs. We demonstrate how three traversal recipes can be implemented on top of our system: temporal breadth-first search (tBFS), temporal depth-first search (tDFS), and temporal single source shortest path (tSSSP). According to our evaluation, our temporal support and parallelism enhance temporal graph traversals in terms of convenience and efficiency. Also, ChronoGraph outperforms existing property graph databases in terms of temporal graph traversals. We prototype ChronoGraph by extending Tinkerpop, a de facto standard for property graphs. Therefore, we expect that our system would be readily accessible to existing property graph users.},
number = {3},
journal = {IEEE Transactions on Knowledge and Data Engineering},
author = {Byun, Jaewook and Woo, Sungpil and Kim, Daeyoung},
month = mar,
year = {2020},
note = {Conference Name: IEEE Transactions on Knowledge and Data Engineering},
keywords = {graph theory, Semantics, ChronoGraph, Databases, graph model, graph traversal language, Parallel processing, parallelism, point-based semantics, programming language semantics, Prototypes, Standards, static graph traversals, Syntactics, temporal aggregation, temporal breadth-first search, temporal degree, temporal depth-first search, temporal graph, temporal graph traversals, temporal networks, temporal paths, temporal single source shortest path, temporal support, temporal syntax, Time factors, Tinkerpop, traversal language, traversal-oriented system, tree searching},
pages = {424--437},
file = {IEEE Xplore Abstract Record:/home/wolf/Zotero/storage/PCGCTDJX/8606161.html:text/html;IEEE Xplore Full Text PDF:/home/wolf/Zotero/storage/6T2NPKPT/Byun et al. - 2020 - ChronoGraph Enabling Temporal Graph Traversals fo.pdf:application/pdf},
}
@article{chen_review_2020,
title = {A review: {Knowledge} reasoning over knowledge graph},
volume = {141},
shorttitle = {A review},
doi = {10.1016/j.eswa.2019.112948},
abstract = {Mining valuable hidden knowledge from large-scale data relies on the support of reasoning technology. Knowledge graphs, as a new type of knowledge representation, have gained much attention in natural language processing. Knowledge graphs can effectively organize and represent knowledge so that it can be efficiently utilized in advanced applications. Recently, reasoning over knowledge graphs has become a hot research topic, since it can obtain new knowledge and conclusions from existing data. Herein we review the basic concept and definitions of knowledge reasoning and the methods for reasoning over knowledge graphs. Specifically, we dissect the reasoning methods into three categories: rule-based reasoning, distributed representation-based reasoning and neural network-based reasoning. We also review the related applications of knowledge graph reasoning, such as knowledge graph completion, question answering, and recommender systems. Finally, we discuss the remaining challenges and research opportunities for knowledge graph reasoning. © 2019 Elsevier Ltd},
journal = {Expert Systems with Applications},
author = {Chen, X. and Jia, S. and Xiang, Y.},
year = {2020},
keywords = {Distributed representation-based reasoning, Knowledge graph, Neural network-based reasoning, Reasoning, Rule-based reasoning},
annote = {Cited By :2},
file = {SCOPUS Snapshot:/home/wolf/Zotero/storage/IHHPGCZP/display.html:text/html;Chen et al. - 2020 - A review Knowledge reasoning over knowledge graph.pdf:/home/wolf/Zotero/storage/T4CDVCYP/Chen et al. - 2020 - A review Knowledge reasoning over knowledge graph.pdf:application/pdf},
}
@inproceedings{kertkeidkachorn_t2kg_2018,
title = {{T2KG}: {A} demonstration of knowledge graph population from text and its challenges},
volume = {2293},
shorttitle = {{T2KG}},
abstract = {Knowledge Graphs play an important role in many AI applications as prior knowledge. In recent years, there are many existing Knowledge Graphs such as DBpedia, Freebase, YAGO. Nevertheless, massive amounts of knowledge are being produced every day. Consequently, Knowledge Graphs become more obsolete over time. It is therefore necessary to populate new knowledge into Knowledge Graphs in order to keep them useable. In this study, we present our end-to-end system for populating knowledge graph from natural language text, namely T2KG. Also, we demonstrate use-cases, achievements, challenges, and lessons learned of the system in practice. © 2018 CEUR-WS. All Rights Reserved.},
author = {Kertkeidkachorn, N. and Ichise, R.},
year = {2018},
pages = {110--113},
file = {SCOPUS Snapshot:/home/wolf/Zotero/storage/SJUD6MD3/display.html:text/html;Kertkeidkachorn and Ichise - T2KG A Demonstration of Knowledge Graph Populati.pdf:/home/wolf/Zotero/storage/VMMCIQF4/Kertkeidkachorn and Ichise - T2KG A Demonstration of Knowledge Graph Populati.pdf:application/pdf},
}
@article{kertkeidkachorn_gtranse_2020,
title = {{GTransE}: {Generalizing} {Translation}-{Based} {Model} on {Uncertain} {Knowledge} {Graph} {Embedding}},
volume = {1128 AISC},
shorttitle = {{GTransE}},
doi = {10.1007/978-3-030-39878-1_16},
abstract = {This is an extension from a selected paper from JSAI2019. Knowledge graphs are useful for many AI applications. Many recent studies have been focused on learning numerical representations of a knowledge graph in a low-dimensional vector space. Learning representations benefits the deep learning framework for encoding real-world knowledge. However, most of the studies do not consider uncertain knowledge graphs. Uncertain knowledge graphs, e.g., NELL, are valuable because they can express the likelihood of triples. In this study, we proposed a novel loss function for translation-based models, GTransE, to deal with uncertainty on knowledge graphs. Experimental results show that GTransE can robustly learn representations on uncertain knowledge graphs. © 2020, Springer Nature Switzerland AG.},
journal = {Advances in Intelligent Systems and Computing},
author = {Kertkeidkachorn, N. and Liu, X. and Ichise, R.},
year = {2020},
keywords = {Knowledge graph, Knowledge representation, Uncertainty},
pages = {170--178},
file = {SCOPUS Snapshot:/home/wolf/Zotero/storage/Q2M9X3E4/display.html:text/html;Kertkeidkachorn et al. - 2020 - GTransE Generalizing Translation-Based Model on U.pdf:/home/wolf/Zotero/storage/JXC44RB7/Kertkeidkachorn et al. - 2020 - GTransE Generalizing Translation-Based Model on U.pdf:application/pdf},
}
@article{kertkeidkachorn_t2kg_nodate,
title = {{T2KG} : {A} {Demonstration} of {Knowledge} {Graph} {Population} from {Text} and {Its} {Challenges}},
abstract = {Knowledge Graphs play an important role in many AI applications as prior knowledge. In recent years, there are many existing Knowledge Graphs such as DBpedia, Freebase, YAGO. Nevertheless, massive amounts of knowledge are being produced every day. Consequently, Knowledge Graphs become more obsolete over time. It is therefore necessary to populate new knowledge into Knowledge Graphs in order to keep them useable. In this study, we present our end-to-end system for populating knowledge graph from natural language text, namely T2KG. Also, we demonstrate use-cases, achievements, challenges, and lessons learned of the system in practice.},
language = {en},
author = {Kertkeidkachorn, Natthawut and Ichise, Ryutaro},
pages = {4},
}
@inproceedings{mehta_scalable_2019,
title = {Scalable knowledge graph construction over text using deep learning based predicate mapping},
doi = {10.1145/3308560.3317708},
abstract = {Automatic extraction of information from text and its transformation into a structured format is an important goal in both Semantic Web Research and computational linguistics. Knowledge Graphs (KG) serve as an intuitive way to provide structure to unstructured text. A fact in a KG is expressed in the form of a triple which captures entities and their interrelationships (predicates). Multiple triples extracted from text can be semantically identical but they may have a vocabulary gap which could lead to an explosion in the number of redundant triples. Hence, to get rid of this vocabulary gap, there is a need to map triples to a homogeneous namespace. In this work, we present an end-to-end KG construction system, which identifies and extracts entities and relationships from text and maps them to the homogenous DBpedia namespace. For Predicate Mapping, we propose a Deep Learning architecture to model semantic similarity. This mapping step is computation heavy, owing to the large number of triples in DBpedia. We identify and prune unnecessary comparisons to make this step scalable. Our experiments show that the proposed approach is able to construct a richer KG at a significantly lower computation cost with respect to previous work. © 2019 IW3C2 (International World Wide Web Conference Committee), published under Creative Commons CC-BY 4.0 License.},
author = {Mehta, A. and Singhal, A. and Karlapalem, K.},
year = {2019},
keywords = {Knowledge Graph, Deep Learning, Predicate Mapping, Scalability, Sentence Simplification},
pages = {705--713},
annote = {Cited By :1},
file = {SCOPUS Snapshot:/home/wolf/Zotero/storage/64T8MVY3/display.html:text/html},
}
@article{trotzek_utilizing_2020,
title = {Utilizing {Neural} {Networks} and {Linguistic} {Metadata} for {Early} {Detection} of {Depression} {Indications} in {Text} {Sequences}},
volume = {32},
issn = {1558-2191},
doi = {10.1109/TKDE.2018.2885515},
abstract = {Depression is ranked as the largest contributor to global disability and is also a major reason for suicide. Still, many individuals suffering from forms of depression are not treated for various reasons. Previous studies have shown that depression also has an effect on language usage and that many depressed individuals use social media platforms or the internet in general to get information or discuss their problems. This paper addresses the early detection of depression using machine learning models based on messages on a social platform. In particular, a convolutional neural network based on different word embeddings is evaluated and compared to a classification based on user-level linguistic metadata. An ensemble of both approaches is shown to achieve state-of-the-art results in a current early detection task. Furthermore, the currently popular ERDE score as metric for early detection systems is examined in detail and its drawbacks in the context of shared tasks are illustrated. A slightly modified metric is proposed and compared to the original score. Finally, a new word embedding was trained on a large corpus of the same domain as the described task and is evaluated as well.},
number = {3},
journal = {IEEE Transactions on Knowledge and Data Engineering},
author = {Trotzek, Marcel and Koitka, Sven and Friedrich, Christoph M.},
month = mar,
year = {2020},
note = {Conference Name: IEEE Transactions on Knowledge and Data Engineering},
keywords = {behavioural sciences computing, convolutional neural nets, convolutional neural network, depressed individuals, Depression, depression indication detection, early detection, early detection systems, early detection task, ERDE score, Europe, Internet, learning (artificial intelligence), linguistic metadata, Linguistics, Machine learning, machine learning models, meta data, Metadata, Natural language processing, shared tasks, social media platforms, Social network services, social networking (online), Task analysis, text sequences, user-level linguistic metadata, word embeddings},
pages = {588--601},
file = {IEEE Xplore Abstract Record:/home/wolf/Zotero/storage/GFBJYSYW/8580405.html:text/html;IEEE Xplore Full Text PDF:/home/wolf/Zotero/storage/DJUBBMLG/Trotzek et al. - 2020 - Utilizing Neural Networks and Linguistic Metadata .pdf:application/pdf},
}
@inproceedings{noauthor_struc2vec_nodate,
title = {struc2vec},
url = {https://arxiv.org/pdf/1704.03165.pdf},
}
@inproceedings{grover_node2vec_2016,
address = {San Francisco, California, USA},
series = {{KDD} '16},
title = {node2vec: {Scalable} {Feature} {Learning} for {Networks}},
isbn = {978-1-4503-4232-2},
shorttitle = {node2vec},
url = {https://doi.org/10.1145/2939672.2939754},
doi = {10.1145/2939672.2939754},
abstract = {Prediction tasks over nodes and edges in networks require careful effort in engineering features used by learning algorithms. Recent research in the broader field of representation learning has led to significant progress in automating prediction by learning the features themselves. However, present feature learning approaches are not expressive enough to capture the diversity of connectivity patterns observed in networks. Here we propose node2vec, an algorithmic framework for learning continuous feature representations for nodes in networks. In node2vec, we learn a mapping of nodes to a low-dimensional space of features that maximizes the likelihood of preserving network neighborhoods of nodes. We define a flexible notion of a node's network neighborhood and design a biased random walk procedure, which efficiently explores diverse neighborhoods. Our algorithm generalizes prior work which is based on rigid notions of network neighborhoods, and we argue that the added flexibility in exploring neighborhoods is the key to learning richer representations. We demonstrate the efficacy of node2vec over existing state-of-the-art techniques on multi-label classification and link prediction in several real-world networks from diverse domains. Taken together, our work represents a new way for efficiently learning state-of-the-art task-independent representations in complex networks.},
urldate = {2020-04-08},
booktitle = {Proceedings of the 22nd {ACM} {SIGKDD} {International} {Conference} on {Knowledge} {Discovery} and {Data} {Mining}},
publisher = {Association for Computing Machinery},
author = {Grover, Aditya and Leskovec, Jure},
month = aug,
year = {2016},
keywords = {feature learning, graph representations, information networks, node embeddings},
pages = {855--864},
file = {Full Text PDF:/home/wolf/Zotero/storage/X8MYYMZP/Grover and Leskovec - 2016 - node2vec Scalable Feature Learning for Networks.pdf:application/pdf},
}
@inproceedings{perozzi_deepwalk_2014,
address = {New York, New York, USA},
series = {{KDD} '14},
title = {{DeepWalk}: online learning of social representations},
isbn = {978-1-4503-2956-9},
shorttitle = {{DeepWalk}},
url = {https://doi.org/10.1145/2623330.2623732},
doi = {10.1145/2623330.2623732},
abstract = {We present DeepWalk, a novel approach for learning latent representations of vertices in a network. These latent representations encode social relations in a continuous vector space, which is easily exploited by statistical models. DeepWalk generalizes recent advancements in language modeling and unsupervised feature learning (or deep learning) from sequences of words to graphs. DeepWalk uses local information obtained from truncated random walks to learn latent representations by treating walks as the equivalent of sentences. We demonstrate DeepWalk's latent representations on several multi-label network classification tasks for social networks such as BlogCatalog, Flickr, and YouTube. Our results show that DeepWalk outperforms challenging baselines which are allowed a global view of the network, especially in the presence of missing information. DeepWalk's representations can provide F1 scores up to 10\% higher than competing methods when labeled data is sparse. In some experiments, DeepWalk's representations are able to outperform all baseline methods while using 60\% less training data. DeepWalk is also scalable. It is an online learning algorithm which builds useful incremental results, and is trivially parallelizable. These qualities make it suitable for a broad class of real world applications such as network classification, and anomaly detection.},
urldate = {2020-04-08},
booktitle = {Proceedings of the 20th {ACM} {SIGKDD} international conference on {Knowledge} discovery and data mining},
publisher = {Association for Computing Machinery},
author = {Perozzi, Bryan and Al-Rfou, Rami and Skiena, Steven},
month = aug,
year = {2014},
keywords = {deep learning, latent representations, learning with partial labels, network classification, online learning, social networks},
pages = {701--710},
file = {Full Text PDF:/home/wolf/Zotero/storage/PS3DVJIA/Perozzi et al. - 2014 - DeepWalk online learning of social representation.pdf:application/pdf},
}
@article{rossi_deep_2020,
title = {Deep {Inductive} {Graph} {Representation} {Learning}},
volume = {32},
issn = {1558-2191},
doi = {10.1109/TKDE.2018.2878247},
abstract = {This paper presents a general inductive graph representation learning framework called \${\textbackslash}textDeepGL\$DeepGL for learning deep node and edge features that generalize across-networks. In particular, \${\textbackslash}textDeepGL\$DeepGL begins by deriving a set of base features from the graph (e.g., graphlet features) and automatically learns a multi-layered hierarchical graph representation where each successive layer leverages the output from the previous layer to learn features of a higher-order. Contrary to previous work, \${\textbackslash}textDeepGL\$DeepGL learns relational functions (each representing a feature) that naturally generalize across-networks and are therefore useful for graph-based transfer learning tasks. Moreover, \${\textbackslash}textDeepGL\$DeepGL naturally supports attributed graphs, learns interpretable inductive graph representations, and is space-efficient (by learning sparse feature vectors). In addition, \${\textbackslash}textDeepGL\$DeepGL is expressive, flexible with many interchangeable components, efficient with a time complexity of \${\textbackslash}mathcal O({\textbar}E{\textbar})\$O({\textbar}E{\textbar}), and scalable for large networks via an efficient parallel implementation. Compared with recent methods, \${\textbackslash}textDeepGL\$DeepGL is (1) effective for across-network transfer learning tasks and large (attributed) graphs, (2) space-efficient requiring up to 6x less memory, (3) fast with up to 106x speedup in runtime performance, and (4) accurate with an average improvement in AUC of 20 percent or more on many learning tasks and across a wide variety of networks.},
number = {3},
journal = {IEEE Transactions on Knowledge and Data Engineering},
author = {Rossi, Ryan A. and Zhou, Rong and Ahmed, Nesreen K.},
month = mar,
year = {2020},
note = {Conference Name: IEEE Transactions on Knowledge and Data Engineering},
keywords = {graph theory, learning (artificial intelligence), Natural language processing, Task analysis, attributed graphs, deep inductive graph representation learning, DeepGL, Electronic mail, graph based transfer learning tasks, Graph representation learning, graph-based feature learning, graphlet features, higher-order structures, inductive representation learning, interpretable inductive graph representations, multilayered hierarchical graph representation, Orbits, relational function learning, Runtime, sparse feature vectors, transfer learning},
pages = {438--452},
file = {IEEE Xplore Abstract Record:/home/wolf/Zotero/storage/E2I4CVZ7/8519335.html:text/html;IEEE Xplore Full Text PDF:/home/wolf/Zotero/storage/NRM7T4Q6/Rossi et al. - 2020 - Deep Inductive Graph Representation Learning.pdf:application/pdf},
}
@article{kilgarriff_review_2000,
title = {Review of {WordNet}: {An} {Electronic} {Lexical} {Database}},
volume = {76},
issn = {0097-8507},
shorttitle = {Review of {WordNet}},
url = {https://www.jstor.org/stable/417141},
doi = {10.2307/417141},
number = {3},
urldate = {2020-04-07},
journal = {Language},
author = {Kilgarriff, Adam},
collaborator = {Fellbaum, Christiane},
year = {2000},
note = {Publisher: Linguistic Society of America},
pages = {706--708},
file = {Submitted Version:/home/wolf/Zotero/storage/2C2FSA9P/Kilgarriff - 2000 - Review of WordNet An Electronic Lexical Database.pdf:application/pdf},
}
@article{kertkeidkachorn_automatic_2018,
title = {An {Automatic} {Knowledge} {Graph} {Creation} {Framework} from {Natural} {Language} {Text}},
volume = {E101.D},
issn = {0916-8532, 1745-1361},
url = {https://www.jstage.jst.go.jp/article/transinf/E101.D/1/E101.D_2017SWP0006/_article},
doi = {10.1587/transinf.2017SWP0006},
abstract = {Knowledge graphs (KG) play a crucial role in many modern applications. However, constructing a KG from natural language text is challenging due to the complex structure of the text. Recently, many approaches have been proposed to transform natural language text to triples to obtain KGs. Such approaches have not yet provided efficient results for mapping extracted elements of triples, especially the predicate, to their equivalent elements in a KG. Predicate mapping is essential because it can reduce the heterogeneity of the data and increase the searchability over a KG. In this article, we propose T2KG, an automatic KG creation framework for natural language text, to more effectively map natural language text to predicates. In our framework, a hybrid combination of a rule-based approach and a similarity-based approach is presented for mapping a predicate to its corresponding predicate in a KG. Based on experimental results, the hybrid approach can identify more similar predicate pairs than a baseline method in the predicate mapping task. An experiment on KG creation is also conducted to investigate the performance of the T2KG. The experimental results show that the T2KG also outperforms the baseline in KG creation. Although KG creation is conducted in open domains, in which prior knowledge is not provided, the T2KG still achieves an F1 score of approximately 50\% when generating triples in the KG creation task. In addition, an empirical study on knowledge population using various text sources is conducted, and the results indicate the T2KG could be used to obtain knowledge that is not currently available from DBpedia.},
language = {en},
number = {1},
urldate = {2020-04-06},
journal = {IEICE Transactions on Information and Systems},
author = {Kertkeidkachorn, Natthawut and Ichise, Ryutaro},
year = {2018},
pages = {90--98},
file = {Kertkeidkachorn and Ichise - 2018 - An Automatic Knowledge Graph Creation Framework fr.pdf:/home/wolf/Zotero/storage/WP9AAHMA/Kertkeidkachorn and Ichise - 2018 - An Automatic Knowledge Graph Creation Framework fr.pdf:application/pdf},
}
@article{kertkeidkachorn_automatic_2018-1,
title = {An {Automatic} {Knowledge} {Graph} {Creation} {Framework} from {Natural} {Language} {Text}},
volume = {E101-D},
issn = {1745-1361, 0916-8532},
url = {https://search.ieice.org/bin/summary.php?id=e101-d_1_90&category=D&year=2018&lang=E&abst=},
abstract = {Knowledge graphs (KG) play a crucial role in many modern applications. However, constructing a KG from natural language text is challenging due to the complex structure of the text. Recently, many approaches have been proposed to transform natural language text to triples to obtain KGs. Such approaches have not yet provided efficient results for mapping extracted elements of triples, especially the predicate, to their equivalent elements in a KG. Predicate mapping is essential because it can reduce the heterogeneity of the data and increase the searchability over a KG. In this article, we propose T2KG, an automatic KG creation framework for natural language text, to more effectively map natural language text to predicates. In our framework, a hybrid combination of a rule-based approach and a similarity-based approach is presented for mapping a predicate to its corresponding predicate in a KG. Based on experimental results, the hybrid approach can identify more similar predicate pairs than a baseline method in the predicate mapping task. An experiment on KG creation is also conducted to investigate the performance of the T2KG. The experimental results show that the T2KG also outperforms the baseline in KG creation. Although KG creation is conducted in open domains, in which prior knowledge is not provided, the T2KG still achieves an F1 score of approximately 50\% when generating triples in the KG creation task. In addition, an empirical study on knowledge population using various text sources is conducted, and the results indicate the T2KG could be used to obtain knowledge that is not currently available from DBpedia.},
number = {1},
urldate = {2020-04-06},
journal = {IEICE TRANSACTIONS on Information and Systems},
author = {Kertkeidkachorn, Natthawut and Ichise, Ryutaro},
month = jan,
year = {2018},
note = {Publisher: The Institute of Electronics, Information and Communication Engineers},
pages = {90--98},
file = {Snapshot:/home/wolf/Zotero/storage/VTTW8LL4/summary.html:text/html},
}
@article{kertkeidkachorn_automatic_2018-2,
title = {An {Automatic} {Knowledge} {Graph} {Creation} {Framework} from {Natural} {Language} {Text}},
volume = {E101.D},
issn = {0916-8532, 1745-1361},
url = {https://www.jstage.jst.go.jp/article/transinf/E101.D/1/E101.D_2017SWP0006/_article},
doi = {10.1587/transinf.2017SWP0006},
abstract = {Knowledge graphs (KG) play a crucial role in many modern applications. However, constructing a KG from natural language text is challenging due to the complex structure of the text. Recently, many approaches have been proposed to transform natural language text to triples to obtain KGs. Such approaches have not yet provided efficient results for mapping extracted elements of triples, especially the predicate, to their equivalent elements in a KG. Predicate mapping is essential because it can reduce the heterogeneity of the data and increase the searchability over a KG. In this article, we propose T2KG, an automatic KG creation framework for natural language text, to more effectively map natural language text to predicates. In our framework, a hybrid combination of a rule-based approach and a similarity-based approach is presented for mapping a predicate to its corresponding predicate in a KG. Based on experimental results, the hybrid approach can identify more similar predicate pairs than a baseline method in the predicate mapping task. An experiment on KG creation is also conducted to investigate the performance of the T2KG. The experimental results show that the T2KG also outperforms the baseline in KG creation. Although KG creation is conducted in open domains, in which prior knowledge is not provided, the T2KG still achieves an F1 score of approximately 50\% when generating triples in the KG creation task. In addition, an empirical study on knowledge population using various text sources is conducted, and the results indicate the T2KG could be used to obtain knowledge that is not currently available from DBpedia.},
language = {en},
number = {1},
urldate = {2020-04-06},
journal = {IEICE Transactions on Information and Systems},
author = {Kertkeidkachorn, Natthawut and Ichise, Ryutaro},
year = {2018},
pages = {90--98},
file = {Kertkeidkachorn and Ichise - 2018 - An Automatic Knowledge Graph Creation Framework fr.pdf:/home/wolf/Zotero/storage/96LRXMRC/Kertkeidkachorn and Ichise - 2018 - An Automatic Knowledge Graph Creation Framework fr.pdf:application/pdf},
}
@incollection{you_graph_2018,
title = {Graph {Convolutional} {Policy} {Network} for {Goal}-{Directed} {Molecular} {Graph} {Generation}},
url = {http://papers.nips.cc/paper/7877-graph-convolutional-policy-network-for-goal-directed-molecular-graph-generation.pdf},
urldate = {2020-03-27},
booktitle = {Advances in {Neural} {Information} {Processing} {Systems} 31},
publisher = {Curran Associates, Inc.},
author = {You, Jiaxuan and Liu, Bowen and Ying, Zhitao and Pande, Vijay and Leskovec, Jure},
editor = {Bengio, S. and Wallach, H. and Larochelle, H. and Grauman, K. and Cesa-Bianchi, N. and Garnett, R.},
year = {2018},
pages = {6410--6421},
file = {NIPS Snapshot:/home/wolf/Zotero/storage/IU79DNYN/7877-graph-convolutional-policy-network-for-goal-directed-molecular-graph-generation.html:text/html;NIPS Full Text PDF:/home/wolf/Zotero/storage/GPW9RRLY/You et al. - 2018 - Graph Convolutional Policy Network for Goal-Direct.pdf:application/pdf},
}
@inproceedings{svetlik_automatic_2017,
title = {Automatic {Curriculum} {Graph} {Generation} for {Reinforcement} {Learning} {Agents}},
url = {https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14961},
abstract = {In recent years, research has shown that transfer learning methods can be leveraged to construct curricula that sequence a series of simpler tasks such that performance on a final target task is improved. A major limitation of existing approaches is that such curricula are handcrafted by humans that are typically domain experts. To address this limitation, we introduce a method to generate a curriculum based on task descriptors and a novel metric of transfer potential. Our method automatically generates a curriculum as a directed acyclic graph (as opposed to a linear sequence as done in existing work). Experiments in both discrete and continuous domains show that our method produces curricula that improve the agent's learning performance when compared to the baseline condition of learning on the target task from scratch.},
language = {en},
urldate = {2020-03-27},
booktitle = {Thirty-{First} {AAAI} {Conference} on {Artificial} {Intelligence}},
author = {Svetlik, Maxwell and Leonetti, Matteo and Sinapov, Jivko and Shah, Rishi and Walker, Nick and Stone, Peter},
month = feb,
year = {2017},
file = {Snapshot:/home/wolf/Zotero/storage/WRTKNPRX/14961.html:text/html;Full Text PDF:/home/wolf/Zotero/storage/HNTXER68/Svetlik et al. - 2017 - Automatic Curriculum Graph Generation for Reinforc.pdf:application/pdf},
}
@article{zhang_deep_2020,
title = {Deep {Learning} on {Graphs}: {A} {Survey}},
shorttitle = {Deep {Learning} on {Graphs}},
url = {http://arxiv.org/abs/1812.04202},
abstract = {Deep learning has been shown to be successful in a number of domains, ranging from acoustics, images, to natural language processing. However, applying deep learning to the ubiquitous graph data is non-trivial because of the unique characteristics of graphs. Recently, substantial research efforts have been devoted to applying deep learning methods to graphs, resulting in beneficial advances in graph analysis techniques. In this survey, we comprehensively review the different types of deep learning methods on graphs. We divide the existing methods into five categories based on their model architectures and training strategies: graph recurrent neural networks, graph convolutional networks, graph autoencoders, graph reinforcement learning, and graph adversarial methods. We then provide a comprehensive overview of these methods in a systematic manner mainly by following their development history. We also analyze the differences and compositions of different methods. Finally, we briefly outline the applications in which they have been used and discuss potential future research directions.},
urldate = {2020-03-27},
journal = {arXiv:1812.04202 [cs, stat]},
author = {Zhang, Ziwei and Cui, Peng and Zhu, Wenwu},
month = mar,
year = {2020},
note = {arXiv: 1812.04202},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning, Computer Science - Social and Information Networks},
annote = {Comment: Accepted by Transactions on Knowledge and Data Engineering. 24 pages, 11 figures},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/UVTXS9Z7/1812.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/WQ3NVAPL/Zhang et al. - 2020 - Deep Learning on Graphs A Survey.pdf:application/pdf},
}
@inproceedings{zhuang_dual_2018,
address = {Lyon, France},
series = {{WWW} '18},
title = {Dual {Graph} {Convolutional} {Networks} for {Graph}-{Based} {Semi}-{Supervised} {Classification}},
isbn = {978-1-4503-5639-8},
url = {https://doi.org/10.1145/3178876.3186116},
doi = {10.1145/3178876.3186116},
abstract = {The problem of extracting meaningful data through graph analysis spans a range of different fields, such as the internet, social networks, biological networks, and many others. The importance of being able to effectively mine and learn from such data continues to grow as more and more structured data become available. In this paper, we present a simple and scalable semi-supervised learning method for graph-structured data in which only a very small portion of the training data are labeled. To sufficiently embed the graph knowledge, our method performs graph convolution from different views of the raw data. In particular, a dual graph convolutional neural network method is devised to jointly consider the two essential assumptions of semi-supervised learning: (1) local consistency and (2) global consistency. Accordingly, two convolutional neural networks are devised to embed the local-consistency-based and global-consistency-based knowledge, respectively. Given the different data transformations from the two networks, we then introduce an unsupervised temporal loss function for the ensemble. In experiments using both unsupervised and supervised loss functions, our method outperforms state-of-the-art techniques on different datasets.},
urldate = {2020-03-27},
booktitle = {Proceedings of the 2018 {World} {Wide} {Web} {Conference}},
publisher = {International World Wide Web Conferences Steering Committee},
author = {Zhuang, Chenyi and Ma, Qiang},
month = apr,
year = {2018},
keywords = {adjacency matrix, graph convolutional networks, graph diffusion, pointwise mutual information, semi-supervised learning},
pages = {499--508},
file = {Submitted Version:/home/wolf/Zotero/storage/2YVUNFFD/Zhuang and Ma - 2018 - Dual Graph Convolutional Networks for Graph-Based .pdf:application/pdf},
}
@article{wei_eda_2019,
title = {{EDA}: {Easy} {Data} {Augmentation} {Techniques} for {Boosting} {Performance} on {Text} {Classification} {Tasks}},
shorttitle = {{EDA}},
url = {http://arxiv.org/abs/1901.11196},
abstract = {We present EDA: easy data augmentation techniques for boosting performance on text classification tasks. EDA consists of four simple but powerful operations: synonym replacement, random insertion, random swap, and random deletion. On five text classification tasks, we show that EDA improves performance for both convolutional and recurrent neural networks. EDA demonstrates particularly strong results for smaller datasets; on average, across five datasets, training with EDA while using only 50\% of the available training set achieved the same accuracy as normal training with all available data. We also performed extensive ablation studies and suggest parameters for practical use.},
urldate = {2020-02-19},
journal = {arXiv:1901.11196 [cs]},
author = {Wei, Jason and Zou, Kai},
month = aug,
year = {2019},
note = {arXiv: 1901.11196},
keywords = {Computer Science - Computation and Language},
annote = {Comment: EMNLP-IJCNLP 2019 short paper},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/6827NNJE/1901.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/Z7CGKHNG/Wei and Zou - 2019 - EDA Easy Data Augmentation Techniques for Boostin.pdf:application/pdf},
}
@inproceedings{heinerman_evolution_2015,
title = {Evolution, {Individual} {Learning}, and {Social} {Learning} in a {Swarm} of {Real} {Robots}},
doi = {10.1109/SSCI.2015.152},
abstract = {We investigate a novel adaptive system based on evolution, individual learning, and social learning in a swarm of physical Thymio II robots. The system is based on distinguishing inheritable and learnable features in the robots and defining appropriate operators for both categories. In this study we choose to make the sensory layout of the robots inheritable, thus evolvable, and the robot controllers learnable. We run tests with a basic system that employs only evolution and individual learning and compare this with an extended system where robots can disseminate their learned controllers. Results show that social learning increases the learning speed and leads to better controllers.},
booktitle = {2015 {IEEE} {Symposium} {Series} on {Computational} {Intelligence}},
author = {Heinerman, Jacqueline and Rango, Massimiliano and Eiben, A.E.},
month = dec,
year = {2015},
note = {ISSN: null},
keywords = {learning (artificial intelligence), adaptive system, Bioinformatics, Collision avoidance, evolutionary computation, Genomics, individual learning, Layout, mobile robots, multi-robot systems, physical Thymio II robots, robot controllers, Robot sensing systems, sensory layout, social learning, swarm},
pages = {1055--1062},
file = {IEEE Xplore Full Text PDF:/home/wolf/Zotero/storage/3ADCLDVQ/Heinerman et al. - 2015 - Evolution, Individual Learning, and Social Learnin.pdf:application/pdf},
}
@article{heinerman_importance_2019,
title = {Importance of {Parameter} {Settings} on the {Benefits} of {Robot}-to-{Robot} {Learning} in {Evolutionary} {Robotics}},
volume = {6},
issn = {2296-9144},
url = {https://www.frontiersin.org/articles/10.3389/frobt.2019.00010/full},
doi = {10.3389/frobt.2019.00010},
abstract = {Robot-to-robot learning, a specific case of social learning in robotics, enables multiple robots to share learned skills while completing a task. The literature offers various statements of its benefits. Robots using this type of social learning can reach a higher performance, an increased learning speed, or both, compared to robots using individual learning only. No general explanation has been advanced for the difference in observations, which make the results highly dependent on the particular system and parameter setting. In this paper, we perform a detailed analysis into the effects of robot-to-robot learning. As a result, we show that this type of social learning can reduce the sensitivity of the learning process to the choice of parameters in two ways. First, robot-to-robot learning can reduce the number of bad performing individuals in the population. Second, robot-to-robot learning can increase the chance of having a successful run, where success is defined as the presence of a high performing individual. Additionally, we show that robot-to-robot learning results in an increased learning speed for almost all parameter settings. Our results indicate that robot-to-robot learning is a powerful mechanism which leads to benefits in both performance and learning speed.},
language = {English},
urldate = {2019-12-30},
journal = {Frontiers in Robotics and AI},
author = {Heinerman, Jacqueline and Haasdijk, Evert and Eiben, A. E.},
year = {2019},
keywords = {evolutionary algorithm, Evolutionary Robotics, parameter tuning, Robot-to-robot learning, Social learning},
file = {Full Text PDF:/home/wolf/Zotero/storage/DLW2RUA4/Heinerman et al. - 2019 - Importance of Parameter Settings on the Benefits o.pdf:application/pdf},
}
@misc{noauthor_unsupervised_nodate,
title = {Unsupervised identification and recognition of situations for high-dimensional sensori-motor streams {\textbar} {Elsevier} {Enhanced} {Reader}},
url = {https://reader.elsevier.com/reader/sd/pii/S0925231217309840?token=FA834DAEFF1933B4D5FC4F184CF41CDFB96C90399B57F8F9B3E29C5D738EDBFE655A91B4654AFD5BC5241CFC3188B930},
language = {en},
urldate = {2019-12-30},
doi = {10.1016/j.neucom.2017.02.090},
file = {Snapshot:/home/wolf/Zotero/storage/9VVGVM4H/S0925231217309840.html:text/html},
}
@inproceedings{ratner_snorkel_2018,
address = {Houston, TX, USA},
title = {Snorkel {MeTaL}: {Weak} {Supervision} for {Multi}-{Task} {Learning}},
isbn = {978-1-4503-5828-6},
shorttitle = {Snorkel {MeTaL}},
url = {http://dl.acm.org/citation.cfm?doid=3209889.3209898},
doi = {10.1145/3209889.3209898},
abstract = {Many real-world machine learning problems are challenging to tackle for two reasons: (i) they involve multiple sub-tasks at different levels of granularity; and (ii) they require large volumes of labeled training data. We propose Snorkel MeTaL, an end-toend system for multi-task learning that leverages weak supervision provided at multiple levels of granularity by domain expert users. In MeTaL, a user specifies a problem consisting of multiple, hierarchically-related sub-tasks—for example, classifying a document at multiple levels of granularity—and then provides labeling functions for each sub-task as weak supervision. MeTaL learns a re-weighted model of these labeling functions, and uses the combined signal to train a hierarchical multi-task network which is automatically compiled from the structure of the sub-tasks. Using MeTaL on a radiology report triage task and a fine-grained news classification task, we achieve average gains of 11.2 accuracy points over a baseline supervised approach and 9.5 accuracy points over the predictions of the user-provided labeling functions.},
language = {en},
urldate = {2019-12-27},
booktitle = {Proceedings of the {Second} {Workshop} on {Data} {Management} for {End}-{To}-{End} {Machine} {Learning} - {DEEM}'18},
publisher = {ACM Press},
author = {Ratner, Alex and Hancock, Braden and Dunnmon, Jared and Goldman, Roger and Ré, Christopher},
year = {2018},
pages = {1--4},
file = {Ratner et al. - 2018 - Snorkel MeTaL Weak Supervision for Multi-Task Lea.pdf:/home/wolf/Zotero/storage/DFCMGMK7/Ratner et al. - 2018 - Snorkel MeTaL Weak Supervision for Multi-Task Lea.pdf:application/pdf},
}
@article{joulin_fasttext.zip:_2016,
title = {{FastText}.zip: {Compressing} text classification models},
shorttitle = {{FastText}.zip},
url = {http://arxiv.org/abs/1612.03651},
abstract = {We consider the problem of producing compact architectures for text classification, such that the full model fits in a limited amount of memory. After considering different solutions inspired by the hashing literature, we propose a method built upon product quantization to store word embeddings. While the original technique leads to a loss in accuracy, we adapt this method to circumvent quantization artefacts. Our experiments carried out on several benchmarks show that our approach typically requires two orders of magnitude less memory than fastText while being only slightly inferior with respect to accuracy. As a result, it outperforms the state of the art by a good margin in terms of the compromise between memory usage and accuracy.},
urldate = {2019-12-23},
journal = {arXiv:1612.03651 [cs]},
author = {Joulin, Armand and Grave, Edouard and Bojanowski, Piotr and Douze, Matthijs and Jégou, Hérve and Mikolov, Tomas},
month = dec,
year = {2016},
note = {arXiv: 1612.03651},
keywords = {Computer Science - Machine Learning, Computer Science - Computation and Language},
annote = {Comment: Submitted to ICLR 2017},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/P67INB7F/1612.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/JSHEYQQ9/Joulin et al. - 2016 - FastText.zip Compressing text classification mode.pdf:application/pdf},
}
@article{higgins_beta-vae_2016,
title = {beta-{VAE}: {Learning} {Basic} {Visual} {Concepts} with a {Constrained} {Variational} {Framework}},
shorttitle = {beta-{VAE}},
url = {https://openreview.net/forum?id=Sy2fzU9gl},
abstract = {We introduce beta-VAE, a new state-of-the-art framework for automated discovery of interpretable factorised latent representations from raw image data in a completely unsupervised manner.},
language = {en},
urldate = {2021-01-02},
author = {Higgins, Irina and Matthey, Loic and Pal, Arka and Burgess, Christopher and Glorot, Xavier and Botvinick, Matthew and Mohamed, Shakir and Lerchner, Alexander},
month = nov,
year = {2016},
file = {Snapshot:/home/wolf/Zotero/storage/7X3MUUFS/forum.html:text/html;Full Text PDF:/home/wolf/Zotero/storage/VEBUUZTQ/Higgins et al. - 2016 - beta-VAE Learning Basic Visual Concepts with a Co.pdf:application/pdf},
}
@article{paszke_pytorch_2019,
title = {{PyTorch}: {An} {Imperative} {Style}, {High}-{Performance} {Deep} {Learning} {Library}},
shorttitle = {{PyTorch}},
url = {http://arxiv.org/abs/1912.01703},
abstract = {Deep learning frameworks have often focused on either usability or speed, but not both. PyTorch is a machine learning library that shows that these two goals are in fact compatible: it provides an imperative and Pythonic programming style that supports code as a model, makes debugging easy and is consistent with other popular scientific computing libraries, while remaining efficient and supporting hardware accelerators such as GPUs. In this paper, we detail the principles that drove the implementation of PyTorch and how they are reflected in its architecture. We emphasize that every aspect of PyTorch is a regular Python program under the full control of its user. We also explain how the careful and pragmatic implementation of the key components of its runtime enables them to work together to achieve compelling performance. We demonstrate the efficiency of individual subsystems, as well as the overall speed of PyTorch on several common benchmarks.},
urldate = {2021-01-02},
journal = {arXiv:1912.01703 [cs, stat]},
author = {Paszke, Adam and Gross, Sam and Massa, Francisco and Lerer, Adam and Bradbury, James and Chanan, Gregory and Killeen, Trevor and Lin, Zeming and Gimelshein, Natalia and Antiga, Luca and Desmaison, Alban and Köpf, Andreas and Yang, Edward and DeVito, Zach and Raison, Martin and Tejani, Alykhan and Chilamkurthy, Sasank and Steiner, Benoit and Fang, Lu and Bai, Junjie and Chintala, Soumith},
month = dec,
year = {2019},
note = {arXiv: 1912.01703},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning, Computer Science - Mathematical Software},
annote = {Comment: 12 pages, 3 figures, NeurIPS 2019},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/CLTGJQQA/1912.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/8E2CLU5P/Paszke et al. - 2019 - PyTorch An Imperative Style, High-Performance Dee.pdf:application/pdf},
}
@online{nickel_three-way_nodate,
title = {A {Three}-{Way} {Model} for {Collective} {Learning} on {Multi}-{Relational} {Data}},
abstract = {Relational learning is becoming increasingly important in many areas of application. Here, we present a novel approach to relational learning based on the factorization of a three-way tensor. We show that unlike other tensor approaches, our method is able to perform collective learning via the latent components of the model and provide an efficient algorithm to compute the factorization. We substantiate our theoretical considerations regarding the collective learning capabilities of our model by the means of experiments on both a new dataset and a dataset commonly used in entity resolution. Furthermore, we show on common benchmark datasets that our approach achieves better or on-par results, if compared to current state-of-the-art relational learning solutions, while it is significantly faster to compute.},
language = {en},
author = {Nickel, Maximilian and Tresp, Volker and Kriegel, Hans-Peter},
pages = {8},
file = {Nickel et al. - A Three-Way Model for Collective Learning on Multi.pdf:/home/wolf/Zotero/storage/VKHTK2Y9/Nickel et al. - A Three-Way Model for Collective Learning on Multi.pdf:application/pdf},
}
@article{mills-tettey_dynamic_nodate,
title = {The {Dynamic} {Hungarian} {Algorithm} for the {Assignment} {Problem} with {Changing} {Costs}},
language = {en},
author = {Mills-Tettey, G Ayorkor and Stentz, Anthony and Dias, M Bernardine},
pages = {19},
year = {2007},
file = {Mills-Tettey et al. - The Dynamic Hungarian Algorithm for the Assignment.pdf:/home/wolf/Zotero/storage/EC48KYH8/Mills-Tettey et al. - The Dynamic Hungarian Algorithm for the Assignment.pdf:application/pdf},
}
@misc{mills-tettey_dynamic_2007,
title = {The {Dynamic} {Hungarian} {Algorithm} for the {Assignment} {Problem} with {Changing} {Costs}},
url = {/paper/The-Dynamic-Hungarian-Algorithm-for-the-Assignment-Mills-Tettey-Stentz/23a53ffca21c4d4aebd085fc426a7a68137bcf90},
abstract = {In this paper, we present the dynamic Hungarian algorithm, applicable to optimally solving the assignment problem in situations with changing edge costs or weights. This problem is relevant, for example, in a transportation domain where the unexpected closing of a road translates to changed transportation costs. When such cost changes occur after an initial assignment has been made, the new problem, like the original problem, may be solved from scratch using the well-known Hungarian algorithm. However, the dynamic version of the algorithm which we present solves the new problem more efficiently by repairing the initial solution obtained before the cost changes. We present proofs of the correctness and efficiency of our algorithm and present simulation results illustrating its efficiency.},
language = {en},
urldate = {2020-12-27},
journal = {undefined},
author = {Mills-Tettey, G. A. and Stentz, Anthony and Dias, M.},
year = {2007},
file = {Snapshot:/home/wolf/Zotero/storage/CVVAKYUA/23a53ffca21c4d4aebd085fc426a7a68137bcf90.html:text/html},
}
@book{chung1997spectral,
title={Spectral graph theory},
author={Chung, Fan RK and Graham, Fan Chung},
number={92},
year={1997},
publisher={American Mathematical Soc.}
}
@article{yong_gradient_2020,
title = {Gradient {Centralization}: {A} {New} {Optimization} {Technique} for {Deep} {Neural} {Networks}},
shorttitle = {Gradient {Centralization}},
url = {http://arxiv.org/abs/2004.01461},
abstract = {Optimization techniques are of great importance to effectively and efficiently train a deep neural network (DNN). It has been shown that using the first and second order statistics (e.g., mean and variance) to perform Z-score standardization on network activations or weight vectors, such as batch normalization (BN) and weight standardization (WS), can improve the training performance. Different from these existing methods that mostly operate on activations or weights, we present a new optimization technique, namely gradient centralization (GC), which operates directly on gradients by centralizing the gradient vectors to have zero mean. GC can be viewed as a projected gradient descent method with a constrained loss function. We show that GC can regularize both the weight space and output feature space so that it can boost the generalization performance of DNNs. Moreover, GC improves the Lipschitzness of the loss function and its gradient so that the training process becomes more efficient and stable. GC is very simple to implement and can be easily embedded into existing gradient based DNN optimizers with only one line of code. It can also be directly used to fine-tune the pre-trained DNNs. Our experiments on various applications, including general image classification, fine-grained image classification, detection and segmentation, demonstrate that GC can consistently improve the performance of DNN learning. The code of GC can be found at https://github.com/Yonghongwei/Gradient-Centralization.},
urldate = {2020-12-24},
journal = {arXiv:2004.01461 [cs]},
author = {Yong, Hongwei and Huang, Jianqiang and Hua, Xiansheng and Zhang, Lei},
month = apr,
year = {2020},
note = {arXiv: 2004.01461},
keywords = {Computer Science - Computer Vision and Pattern Recognition},
annote = {Comment: 20 pages, 7 figures, conference},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/FDVVQHVS/2004.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/GLC88XY6/Yong et al. - 2020 - Gradient Centralization A New Optimization Techni.pdf:application/pdf},
}
@article{paulheim_knowledge_2016,
title = {Knowledge graph refinement: {A} survey of approaches and evaluation methods},
volume = {8},
issn = {22104968, 15700844},
shorttitle = {Knowledge graph refinement},
url = {https://www.medra.org/servlet/aliasResolver?alias=iospress&doi=10.3233/SW-160218},
doi = {10.3233/SW-160218},
abstract = {In the recent years, different Web knowledge graphs, both free and commercial, have been created. While Google coined the term “Knowledge Graph” in 2012, there are also a few openly available knowledge graphs, with DBpedia, YAGO, and Freebase being among the most prominent ones. Those graphs are often constructed from semi-structured knowledge, such as Wikipedia, or harvested from the web with a combination of statistical and linguistic methods. The result are large-scale knowledge graphs that try to make a good trade-off between completeness and correctness. In order to further increase the utility of such knowledge graphs, various refinement methods have been proposed, which try to infer and add missing knowledge to the graph, or identify erroneous pieces of information. In this article, we provide a survey of such knowledge graph refinement approaches, with a dual look at both the methods being proposed as well as the evaluation methodologies used.},
language = {en},
number = {3},
urldate = {2020-12-24},
journal = {Semantic Web},
author = {Paulheim, Heiko},
editor = {Cimiano, Philipp},
month = dec,
year = {2016},
pages = {489--508},
file = {Paulheim - 2016 - Knowledge graph refinement A survey of approaches.pdf:/home/wolf/Zotero/storage/WXA6WR5W/Paulheim - 2016 - Knowledge graph refinement A survey of approaches.pdf:application/pdf},
}
@article{zhang_lookahead_2019,
title = {Lookahead {Optimizer}: k steps forward, 1 step back},
shorttitle = {Lookahead {Optimizer}},
url = {http://arxiv.org/abs/1907.08610},
abstract = {The vast majority of successful deep neural networks are trained using variants of stochastic gradient descent (SGD) algorithms. Recent attempts to improve SGD can be broadly categorized into two approaches: (1) adaptive learning rate schemes, such as AdaGrad and Adam, and (2) accelerated schemes, such as heavy-ball and Nesterov momentum. In this paper, we propose a new optimization algorithm, Lookahead, that is orthogonal to these previous approaches and iteratively updates two sets of weights. Intuitively, the algorithm chooses a search direction by looking ahead at the sequence of fast weights generated by another optimizer. We show that Lookahead improves the learning stability and lowers the variance of its inner optimizer with negligible computation and memory cost. We empirically demonstrate Lookahead can significantly improve the performance of SGD and Adam, even with their default hyperparameter settings on ImageNet, CIFAR-10/100, neural machine translation, and Penn Treebank.},
urldate = {2020-12-24},
journal = {arXiv:1907.08610 [cs, stat]},
author = {Zhang, Michael R. and Lucas, James and Hinton, Geoffrey and Ba, Jimmy},
month = jul,
year = {2019},
note = {arXiv: 1907.08610
version: 1},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning, Computer Science - Neural and Evolutionary Computing},
annote = {Comment: Accepted to Neural Information Processing Systems 2019. Code available at: https://github.com/michaelrzhang/lookahead},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/92MZWPIM/1907.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/V2E575NK/Zhang et al. - 2019 - Lookahead Optimizer k steps forward, 1 step back.pdf:application/pdf},
}
@article{liu_variance_2020,
title = {On the {Variance} of the {Adaptive} {Learning} {Rate} and {Beyond}},
url = {http://arxiv.org/abs/1908.03265},
abstract = {The learning rate warmup heuristic achieves remarkable success in stabilizing training, accelerating convergence and improving generalization for adaptive stochastic optimization algorithms like RMSprop and Adam. Here, we study its mechanism in details. Pursuing the theory behind warmup, we identify a problem of the adaptive learning rate (i.e., it has problematically large variance in the early stage), suggest warmup works as a variance reduction technique, and provide both empirical and theoretical evidence to verify our hypothesis. We further propose RAdam, a new variant of Adam, by introducing a term to rectify the variance of the adaptive learning rate. Extensive experimental results on image classification, language modeling, and neural machine translation verify our intuition and demonstrate the effectiveness and robustness of our proposed method. All implementations are available at: https://github.com/LiyuanLucasLiu/RAdam.},
urldate = {2020-12-24},
journal = {arXiv:1908.03265 [cs, stat]},
author = {Liu, Liyuan and Jiang, Haoming and He, Pengcheng and Chen, Weizhu and Liu, Xiaodong and Gao, Jianfeng and Han, Jiawei},
month = apr,
year = {2020},
note = {arXiv: 1908.03265},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning, Computer Science - Computation and Language},
annote = {Comment: ICLR 2020. Fix several typos in the previous version},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/AQUQBQVE/1908.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/PDKA3GF8/Liu et al. - 2020 - On the Variance of the Adaptive Learning Rate and .pdf:application/pdf},
}
@article{nguyen_survey_2020,
title = {A survey of embedding models of entities and relationships for knowledge graph completion},
url = {http://arxiv.org/abs/1703.08098},
abstract = {Knowledge graphs (KGs) of real-world facts about entities and their relationships are useful resources for a variety of natural language processing tasks. However, because knowledge graphs are typically incomplete, it is useful to perform knowledge graph completion or link prediction, i.e. predict whether a relationship not in the knowledge graph is likely to be true. This paper serves as a comprehensive survey of embedding models of entities and relationships for knowledge graph completion, summarizing up-to-date experimental results on standard benchmark datasets and pointing out potential future research directions.},
urldate = {2020-12-24},
journal = {arXiv:1703.08098 [cs]},
author = {Nguyen, Dat Quoc},
month = oct,
year = {2020},
note = {arXiv: 1703.08098},
keywords = {Computer Science - Artificial Intelligence, Computer Science - Computation and Language, Computer Science - Information Retrieval},
annote = {Comment: In Proceedings of the 14th Workshop on Graph-Based Natural Language Processing (TextGraphs 2020); 16 pages, 2 figures, 6 tables},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/3CVD5S4W/1703.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/IDKD8TQJ/Nguyen - 2020 - A survey of embedding models of entities and relat.pdf:application/pdf},
}
@article{nickel_review_2016,
title = {A {Review} of {Relational} {Machine} {Learning} for {Knowledge} {Graphs}},
volume = {104},
issn = {0018-9219, 1558-2256},
url = {http://arxiv.org/abs/1503.00759},
doi = {10.1109/JPROC.2015.2483592},
abstract = {Relational machine learning studies methods for the statistical analysis of relational, or graph-structured, data. In this paper, we provide a review of how such statistical models can be "trained" on large knowledge graphs, and then used to predict new facts about the world (which is equivalent to predicting new edges in the graph). In particular, we discuss two fundamentally different kinds of statistical relational models, both of which can scale to massive datasets. The first is based on latent feature models such as tensor factorization and multiway neural networks. The second is based on mining observable patterns in the graph. We also show how to combine these latent and observable models to get improved modeling power at decreased computational cost. Finally, we discuss how such statistical models of graphs can be combined with text-based information extraction methods for automatically constructing knowledge graphs from the Web. To this end, we also discuss Google's Knowledge Vault project as an example of such combination.},
number = {1},
urldate = {2020-12-24},
journal = {Proceedings of the IEEE},
author = {Nickel, Maximilian and Murphy, Kevin and Tresp, Volker and Gabrilovich, Evgeniy},
month = jan,
year = {2016},
note = {arXiv: 1503.00759},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning},
pages = {11--33},
annote = {Comment: To appear in Proceedings of the IEEE},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/77TE68FB/1503.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/GSGA8N7P/Nickel et al. - 2016 - A Review of Relational Machine Learning for Knowle.pdf:application/pdf},
}
@article{tong_calibrating_2019,
title = {Calibrating the {Adaptive} {Learning} {Rate} to {Improve} {Convergence} of {ADAM}},
url = {http://arxiv.org/abs/1908.00700},
abstract = {Adaptive gradient methods (AGMs) have become popular in optimizing the nonconvex problems in deep learning area. We revisit AGMs and identify that the adaptive learning rate (A-LR) used by AGMs varies significantly across the dimensions of the problem over epochs (i.e., anisotropic scale), which may lead to issues in convergence and generalization. All existing modified AGMs actually represent efforts in revising the A-LR. Theoretically, we provide a new way to analyze the convergence of AGMs and prove that the convergence rate of {\textbackslash}textsc\{Adam\} also depends on its hyper-parameter \${\textbackslash}epsilon\$, which has been overlooked previously. Based on these two facts, we propose a new AGM by calibrating the A-LR with an activation (\{{\textbackslash}em softplus\}) function, resulting in the {\textbackslash}textsc\{Sadam\} and {\textbackslash}textsc\{SAMSGrad\} methods {\textbackslash}footnote\{Code is available at https://github.com/neilliang90/Sadam.git.\}. We further prove that these algorithms enjoy better convergence speed under nonconvex, non-strongly convex, and Polyak-\{{\textbackslash}L\}ojasiewicz conditions compared with {\textbackslash}textsc\{Adam\}. Empirical studies support our observation of the anisotropic A-LR and show that the proposed methods outperform existing AGMs and generalize even better than S-Momentum in multiple deep learning tasks.},
urldate = {2020-11-27},
journal = {arXiv:1908.00700 [cs, math, stat]},
author = {Tong, Qianqian and Liang, Guannan and Bi, Jinbo},
month = sep,
year = {2019},
note = {arXiv: 1908.00700
version: 2},
keywords = {Computer Science - Machine Learning, Statistics - Machine Learning, Mathematics - Optimization and Control},
file = {arXiv.org Snapshot:/home/wolf/Zotero/storage/ZPJI32AH/1908.html:text/html;arXiv Fulltext PDF:/home/wolf/Zotero/storage/EVJQCVCC/Tong et al. - 2019 - Calibrating the Adaptive Learning Rate to Improve .pdf:application/pdf},
}
@inproceedings{toutanova_representing_2015,
address = {Lisbon, Portugal},
title = {Representing {Text} for {Joint} {Embedding} of {Text} and {Knowledge} {Bases}},
url = {https://www.aclweb.org/anthology/D15-1174},
doi = {10.18653/v1/D15-1174},
urldate = {2020-10-22},
booktitle = {Proceedings of the 2015 {Conference} on {Empirical} {Methods} in {Natural} {Language} {Processing}},
publisher = {Association for Computational Linguistics},
author = {Toutanova, Kristina and Chen, Danqi and Pantel, Patrick and Poon, Hoifung and Choudhury, Pallavi and Gamon, Michael},
month = sep,
year = {2015},
pages = {1499--1509},
file = {Full Text PDF:/home/wolf/Zotero/storage/Z3RFNDLV/Toutanova et al. - 2015 - Representing Text for Joint Embedding of Text and .pdf:application/pdf},
}
@incollection{bordes_translating_2013,
title = {Translating {Embeddings} for {Modeling} {Multi}-relational {Data}},
url = {http://papers.nips.cc/paper/5071-translating-embeddings-for-modeling-multi-relational-data.pdf},
urldate = {2020-10-22},
booktitle = {Advances in {Neural} {Information} {Processing} {Systems} 26},
publisher = {Curran Associates, Inc.},
author = {Bordes, Antoine and Usunier, Nicolas and Garcia-Duran, Alberto and Weston, Jason and Yakhnenko, Oksana},
editor = {Burges, C. J. C. and Bottou, L. and Welling, M. and Ghahramani, Z. and Weinberger, K. Q.},
year = {2013},
pages = {2787--2795},
file = {NIPS Snapshot:/home/wolf/Zotero/storage/2UDIEXWE/5071-translating-embeddings-for-modeling-multi-relational-data.html:text/html;NIPS Full Text PDF:/home/wolf/Zotero/storage/DASKLYKA/Bordes et al. - 2013 - Translating Embeddings for Modeling Multi-relation.pdf:application/pdf},
}
@incollection{liu_stein_2016,
title = {Stein {Variational} {Gradient} {Descent}: {A} {General} {Purpose} {Bayesian} {Inference} {Algorithm}},
shorttitle = {Stein {Variational} {Gradient} {Descent}},
url = {http://papers.nips.cc/paper/6338-stein-variational-gradient-descent-a-general-purpose-bayesian-inference-algorithm.pdf},
urldate = {2020-09-30},
booktitle = {Advances in {Neural} {Information} {Processing} {Systems} 29},
publisher = {Curran Associates, Inc.},
author = {Liu, Qiang and Wang, Dilin},
editor = {Lee, D. D. and Sugiyama, M. and Luxburg, U. V. and Guyon, I. and Garnett, R.},
year = {2016},
pages = {2378--2386},
file = {NIPS Snapshot:/home/wolf/Zotero/storage/CTRLGD5A/6338-stein-variational-gradient-descent-a-general-purpose-bayesian-inference-algorithm.html:text/html;NIPS Full Text PDF:/home/wolf/Zotero/storage/JQWBY7N2/Liu and Wang - 2016 - Stein Variational Gradient Descent A General Purp.pdf:application/pdf},
}
@article{date_gpu-accelerated_2016,
title = {{GPU}-accelerated {Hungarian} algorithms for the {Linear} {Assignment} {Problem}},
volume = {57},
issn = {0167-8191},
url = {http://www.sciencedirect.com/science/article/pii/S016781911630045X},
doi = {10.1016/j.parco.2016.05.012},
abstract = {In this paper, we describe parallel versions of two different variants (classical and alternating tree) of the Hungarian algorithm for solving the Linear Assignment Problem (LAP). We have chosen Compute Unified Device Architecture (CUDA) enabled NVIDIA Graphics Processing Units (GPU) as the parallel programming architecture because of its ability to perform intense computations on arrays and matrices. The main contribution of this paper is an efficient parallelization of the augmenting path search phase of the Hungarian algorithm. Computational experiments on problems with up to 25 million variables reveal that the GPU-accelerated versions are extremely efficient in solving large problems, as compared to their CPU counterparts. Tremendous parallel speedups are achieved for problems with up to 400 million variables, which are solved within 13 seconds on average. We also tested multi-GPU versions of the two variants on up to 16 GPUs, which show decent scaling behavior for problems with up to 1.6 billion variables and dense cost matrix structure.},
language = {en},
urldate = {2020-09-27},
journal = {Parallel Computing},
author = {Date, Ketan and Nagi, Rakesh},