-
Notifications
You must be signed in to change notification settings - Fork 5
/
DLC.bib
8986 lines (8626 loc) · 490 KB
/
DLC.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
% Generated by Paperpile. Check out http://paperpile.com for more information.
% BibTeX export options can be customized via Settings -> BibTeX.
@ARTICLE{Pascanu2017-jl,
title = "Learning model-based planning from scratch",
author = "Pascanu, Razvan and Li, Yujia and Vinyals, Oriol and Heess,
Nicolas and Buesing, Lars and Racani{\`e}re, Sebastien and
Reichert, David and Weber, Th{\'e}ophane and Wierstra, Daan
and Battaglia, Peter",
abstract = "Conventional wisdom holds that model-based planning is a
powerful approach to sequential decision-making. It is often
very challenging in practice, however, because while a model
can be used to evaluate a plan, it does not prescribe how to
construct a plan. Here we introduce the ``Imagination-based
Planner'', the first model-based, sequential decision-making
agent that can learn to construct, evaluate, and execute
plans. Before any action, it can perform a variable number
of imagination steps, which involve proposing an imagined
action and evaluating it with its model-based imagination.
All imagined actions and outcomes are aggregated,
iteratively, into a ``plan context'' which conditions future
real and imagined actions. The agent can even decide how to
imagine: testing out alternative imagined actions, chaining
sequences of actions together, or building a more complex
``imagination tree'' by navigating flexibly among the
previously imagined states using a learned policy. And our
agent can learn to plan economically, jointly optimizing for
external rewards and computational costs associated with
using its imagination. We show that our architecture can
learn to solve a challenging continuous control problem, and
also learn elaborate planning strategies in a discrete
maze-solving task. Our work opens a new direction toward
learning the components of a model-based planning system and
how to use them.",
month = "19~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.AI",
eprint = "1707.06170"
}
@ARTICLE{Luo_undated-qj,
title = "Learning Deep Architectures via Generalized Whitened Neural
Networks",
author = "Luo, Ping"
}
@ARTICLE{Cisse2017-rg,
title = "Houdini: Fooling Deep Structured Prediction Models",
author = "Cisse, Moustapha and Adi, Yossi and Neverova, Natalia and
Keshet, Joseph",
abstract = "Generating adversarial examples is a critical step for
evaluating and improving the robustness of learning
machines. So far, most existing methods only work for
classification and are not designed to alter the true
performance measure of the problem at hand. We introduce a
novel flexible approach named Houdini for generating
adversarial examples specifically tailored for the final
performance measure of the task considered, be it
combinatorial and non-decomposable. We successfully apply
Houdini to a range of applications such as speech
recognition, pose estimation and semantic segmentation. In
all cases, the attacks based on Houdini achieve higher
success rate than those based on the traditional surrogates
used to train the models while using a less perceptible
adversarial perturbation.",
month = "17~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "stat.ML",
eprint = "1707.05373"
}
@ARTICLE{Melis2017-jk,
title = "On the State of the Art of Evaluation in Neural Language
Models",
author = "Melis, G{\'a}bor and Dyer, Chris and Blunsom, Phil",
abstract = "Ongoing innovations in recurrent neural network
architectures have provided a steady influx of apparently
state-of-the-art results on language modelling benchmarks.
However, these have been evaluated using differing code
bases and limited computational resources, which represent
uncontrolled sources of experimental variation. We
reevaluate several popular architectures and regularisation
methods with large-scale automatic black-box hyperparameter
tuning and arrive at the somewhat surprising conclusion that
standard LSTM architectures, when properly regularised,
outperform more recent models. We establish a new state of
the art on the Penn Treebank and Wikitext-2 corpora, as well
as strong baselines on the Hutter Prize dataset.",
month = "18~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.CL",
eprint = "1707.05589"
}
@ARTICLE{Tozzi2017-ah,
title = "Topodynamics of metastable brains",
author = "Tozzi, Arturo and Peters, James F and Fingelkurts, Andrew A
and Fingelkurts, Alexander A and Mariju{\'a}n, Pedro C",
affiliation = "Center for Nonlinear Science, University of North Texas, 1155
Union Circle, \#311427, Denton, TX 76203-5017, USA. Electronic
address: [email protected]. Department of Electrical and
Computer Engineering, University of Manitoba, 75A Chancellor's
Circle Winnipeg, MB R3T 5V6 Canada; Department of Mathematics,
Ad{\i}yaman University, 02040 Ad{\i}yaman, Turkey. Electronic
address: [email protected]. BM-Science - Brain and
Mind Technologies Research Centre, Espoo, Finland. Electronic
address: [email protected]. BM-Science - Brain
and Mind Technologies Research Centre, Espoo, Finland.
Electronic address: [email protected].
Bioinformation Group, Aragon Institute of Health Science
(IACS), Aragon Health Research Institute (IIS Aragon),
Zaragoza, 50009 Spain. Electronic address:
abstract = "The brain displays both the anatomical features of a vast
amount of interconnected topological mappings as well as the
functional features of a nonlinear, metastable system at the
edge of chaos, equipped with a phase space where mental random
walks tend towards lower energetic basins. Nevertheless, with
the exception of some advanced neuro-anatomic descriptions and
present-day connectomic research, very few studies have been
addressing the topological path of a brain embedded or
embodied in its external and internal environment. Herein, by
using new formal tools derived from algebraic topology, we
provide an account of the metastable brain, based on the
neuro-scientific model of Operational Architectonics of
brain-mind functioning. We introduce a ``topodynamic''
description that shows how the relationships among the
countless intertwined spatio-temporal levels of brain
functioning can be assessed in terms of projections and
mappings that take place on abstract structures, equipped with
different dimensions, curvatures and energetic constraints.
Such a topodynamical approach, apart from providing a
biologically plausible model of brain function that can be
operationalized, is also able to tackle the issue of a
long-standing dichotomy: it throws indeed a bridge between the
subjective, immediate datum of the na{\"\i}ve complex of
sensations and mentations and the objective, quantitative,
data extracted from experimental neuro-scientific procedures.
Importantly, it opens the door to a series of new predictions
and future directions of advancement for neuroscientific
research.",
journal = "Phys. Life Rev.",
month = "23~" # mar,
year = 2017,
keywords = "Borsuk--Ulam theorem; Central nervous system; Mind; Nonlinear
dynamics; Topology",
language = "en"
}
@ARTICLE{Gomez2017-pn,
title = "The Reversible Residual Network: Backpropagation Without
Storing Activations",
author = "Gomez, Aidan N and Ren, Mengye and Urtasun, Raquel and
Grosse, Roger B",
abstract = "Deep residual networks (ResNets) have significantly pushed
forward the state-of-the-art on image classification,
increasing in performance as networks grow both deeper and
wider. However, memory consumption becomes a bottleneck, as
one needs to store the activations in order to calculate
gradients using backpropagation. We present the Reversible
Residual Network (RevNet), a variant of ResNets where each
layer's activations can be reconstructed exactly from the
next layer's. Therefore, the activations for most layers
need not be stored in memory during backpropagation. We
demonstrate the effectiveness of RevNets on CIFAR-10,
CIFAR-100, and ImageNet, establishing nearly identical
classification accuracy to equally-sized ResNets, even
though the activation storage requirements are independent
of depth.",
month = "14~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.CV",
eprint = "1707.04585"
}
@ARTICLE{Teh2017-bq,
title = "Distral: Robust Multitask Reinforcement Learning",
author = "Teh, Yee Whye and Bapst, Victor and Czarnecki, Wojciech
Marian and Quan, John and Kirkpatrick, James and Hadsell,
Raia and Heess, Nicolas and Pascanu, Razvan",
abstract = "Most deep reinforcement learning algorithms are data
inefficient in complex and rich environments, limiting their
applicability to many scenarios. One direction for improving
data efficiency is multitask learning with shared neural
network parameters, where efficiency may be improved through
transfer across related tasks. In practice, however, this is
not usually observed, because gradients from different tasks
can interfere negatively, making learning unstable and
sometimes even less data efficient. Another issue is the
different reward schemes between tasks, which can easily
lead to one task dominating the learning of a shared model.
We propose a new approach for joint training of multiple
tasks, which we refer to as Distral (Distill \& transfer
learning). Instead of sharing parameters between the
different workers, we propose to share a ``distilled''
policy that captures common behaviour across tasks. Each
worker is trained to solve its own task while constrained to
stay close to the shared policy, while the shared policy is
trained by distillation to be the centroid of all task
policies. Both aspects of the learning process are derived
by optimizing a joint objective function. We show that our
approach supports efficient transfer on complex 3D
environments, outperforming several related methods.
Moreover, the proposed learning process is more robust and
more stable---attributes that are critical in deep
reinforcement learning.",
month = "13~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.LG",
eprint = "1707.04175"
}
@INPROCEEDINGS{Fernandes2015-fn,
title = "A Proactive Intelligent Decision Support System for Predicting
the Popularity of Online News",
booktitle = "Progress in Artificial Intelligence",
author = "Fernandes, Kelwin and Vinagre, Pedro and Cortez, Paulo",
abstract = "Due to the Web expansion, the prediction of online news
popularity is becoming a trendy research topic. In this paper,
we propose a novel and proactive Intelligent Decision Support
System (IDSS) that analyzes articles prior to their
publication. Using a broad set of extracted features (e.g.,
keywords, digital media content, earlier popularity of news
referenced in the article) the IDSS first predicts if an
article will become popular. Then, it optimizes a subset of the
articles features that can more easily be changed by authors,
searching for an enhancement of the predicted popularity
probability. Using a large and recently collected dataset, with
39,000 articles from the Mashable website, we performed a
robust rolling windows evaluation of five state of the art
models. The best result was provided by a Random Forest with a
discrimination power of 73\%. Moreover, several stochastic hill
climbing local searches were explored. When optimizing 1000
articles, the best optimization method obtained a mean gain
improvement of 15 percentage points in terms of the estimated
popularity probability. These results attest the proposed IDSS
as a valuable tool for online news authors.",
publisher = "Springer, Cham",
pages = "535--546",
month = "8~" # sep,
year = 2015,
keywords = "Viralization",
language = "en",
conference = "Portuguese Conference on Artificial Intelligence"
}
@ARTICLE{Lloret2016-ze,
title = "Analysing and evaluating the task of automatic tweet generation:
Knowledge to business",
author = "Lloret, Elena and Palomar, Manuel",
abstract = "In this paper a study concerning the evaluation and analysis of
natural language tweets is presented. Based on our experience in
text summarisation, we carry out a deep analysis on user's
perception through the evaluation of tweets manual and
automatically generated from news. Specifically, we consider two
key issues of a tweet: its informativeness and its
interestingness. Therefore, we analyse: (1) do users equally
perceive manual and automatic tweets?; (2) what linguistic
features a good tweet may have to be interesting, as well as
informative? The main challenge of this proposal is the analysis
of tweets to help companies in their positioning and reputation
on the Web. Our results show that: (1) automatically informative
and interesting natural language tweets can be generated as a
result of summarisation approaches; and (2) we can characterise
good and bad tweets based on specific linguistic features not
present in other types of tweets.",
journal = "Comput. Ind.",
volume = 78,
pages = "3--15",
month = "1~" # may,
year = 2016,
keywords = "Natural language processing; Text summarisation; Natural language
tweet generation; User study; Linguistic analysis; Descriptive
statistics;Viralization"
}
@ARTICLE{Varol2017-el,
title = "Analyzing Social Big Data to Study Online Discourse and Its
Manipulation",
author = "Varol, Onur",
abstract = "The widespread use of social media helps people connect and
share their opinions and experiences with millions of others,
while simultaneously bringing new threats. This dissertation
aims to provide insights into analysis of online conversations
and mechanisms that might be used for their manipulation. The
first part delves into the effect of geography on information
dissemination and user roles in online discourse. I study
trending topics on Twitter to highlight mechanisms governing the
diffusion of local and national trends. My analysis points to
three locally geographic regions and one cluster that contains
trendsetting cities coinciding with major travel hubs. When
factors limiting information spread are considered, censorship
mechanisms mandated by governments are found to be ineffective
and even show a correlation with increasing popularity. I also
present an analysis of spatiotemporal characteristics and
distinct user roles in the Gezi movement. Next, I discuss
different forms of social media manipulation. Malicious entities
can employ promotion campaigns and social bots. We build machine
learning frameworks that exploit features extracted from
network, content, and users to train accurate supervised
learning models. Our system for early detection of promoted
social media trends harnesses multidimensional time series
signals to reveal subtle differences between promoted and
organic trends. In my research on social bots, I carried out the
largest study of the human-bot ecosystem to date. Our estimates
suggest that between 9 and 15\% of active Twitter accounts are
bots. I present distinct behavioral groups and interaction
strategies among human and bot accounts. This body of work
contributes to a more comprehensive understanding of online user
behavior and to the development of systems to detect online
abuse.",
publisher = "[Bloomington, Ind.] : Indiana University",
month = jun,
year = 2017,
keywords = "Network Science; Social Media Analysis; Bot Detection; Doctoral
Dissertation;Viralization",
language = "en"
}
@ARTICLE{He2016-em,
title = "Deep Reinforcement Learning with a Combinatorial Action
Space for Predicting Popular Reddit Threads",
author = "He, Ji and Ostendorf, Mari and He, Xiaodong and Chen,
Jianshu and Gao, Jianfeng and Li, Lihong and Deng, Li",
abstract = "We introduce an online popularity prediction and tracking
task as a benchmark task for reinforcement learning with a
combinatorial, natural language action space. A specified
number of discussion threads predicted to be popular are
recommended, chosen from a fixed window of recent comments
to track. Novel deep reinforcement learning architectures
are studied for effective modeling of the value function
associated with actions comprised of interdependent
sub-actions. The proposed model, which represents dependence
between sub-actions through a bi-directional LSTM, gives the
best performance across different experimental
configurations and domains, and it also generalizes well
with varying numbers of recommendation requests.",
month = "12~" # jun,
year = 2016,
keywords = "Viralization",
archivePrefix = "arXiv",
primaryClass = "cs.CL",
eprint = "1606.03667"
}
@INPROCEEDINGS{Wang2015-la,
title = "{I} Can Has Cheezburger? A Nonparanormal Approach to Combining
Textual and Visual Information for Predicting and Generating
Popular Meme Descriptions",
booktitle = "{HLT-NAACL}",
author = "Wang, William Yang and Wen, Miaomiao",
pages = "355--365",
year = 2015,
keywords = "Viralization"
}
@ARTICLE{Farzindar2015-vx,
title = "Natural Language Processing for Social Media",
author = "Farzindar, Atefeh and Inkpen, Diana",
abstract = "Abstract In recent years, online social networking has
revolutionized interpersonal communication. The newer research
on language analysis in social media has been increasingly
focusing on the latter's impact on our daily lives, both on a
personal and a professional level. Natural language processing
(NLP) is one of the most promising avenues for social media data
processing. It is a scientific challenge to develop powerful
methods and algorithms which extract relevant information from a
large volume of data coming from multiple sources and languages
in various formats or in free form. We discuss the challenges in
analyzing social media texts in contrast with traditional
documents. Research methods in information extraction, automatic
categorization and clustering, automatic summarization and
indexing, and statistical machine translation need to be adapted
to a new kind of data. This book reviews the current research on
Natural Language Processing (NLP) tools and methods for
processing the non-traditional information from social media
data that is available in large amounts (big data), and shows
how innovative NLP approaches can integrate appropriate
linguistic information in various fields such as social media
monitoring, health care, business intelligence, industry,
marketing, and security and defense. We review the existing
evaluation metrics for NLP and social media applications, and
the new efforts in evaluation campaigns or shared tasks on new
datasets collected from social media. Such tasks are organized
by the Association for Computational Linguistics (such as
SemEval tasks) or by the National Institute of Standards and
Technology via the Text REtrieval Conference (TREC) and the Text
Analysis Conference (TAC). In the concluding chapter, we discuss
the importance of this dynamic discipline and its great
potential for NLP in the coming decade, in the context of
changes in mobile technology, cloud computing, and social
networking. Table of Contents: Preface / Acknowledgments /
Introduction to Social Media Analysis / Linguistic
Pre-processing\textbackslash\textbackslash of Social Media Texts
/ Semantic Analysis of Social Media Texts / Applications of
Social Media Text Analysis / Data Collection, Annotation, and
Evaluation / Conclusion and Perspectives / Glossary /
Bibliography / Authors' Biographies",
journal = "Synthesis Lectures on Human Language Technologies",
publisher = "Morgan \& Claypool Publishers",
volume = 8,
number = 2,
pages = "1--166",
month = "28~" # aug,
year = 2015,
keywords = "Viralization"
}
@ARTICLE{Tan2014-eo,
title = "The effect of wording on message propagation: Topic- and
author-controlled natural experiments on Twitter",
author = "Tan, Chenhao and Lee, Lillian and Pang, Bo",
abstract = "Consider a person trying to spread an important message on a
social network. He/she can spend hours trying to craft the
message. Does it actually matter? While there has been
extensive prior work looking into predicting popularity of
social-media content, the effect of wording per se has
rarely been studied since it is often confounded with the
popularity of the author and the topic. To control for these
confounding factors, we take advantage of the surprising
fact that there are many pairs of tweets containing the same
url and written by the same user but employing different
wording. Given such pairs, we ask: which version attracts
more retweets? This turns out to be a more difficult task
than predicting popular topics. Still, humans can answer
this question better than chance (but far from perfectly),
and the computational methods we develop can do better than
both an average human and a strong competing method trained
on non-controlled data.",
month = "6~" # may,
year = 2014,
keywords = "Viralization",
archivePrefix = "arXiv",
primaryClass = "cs.SI",
eprint = "1405.1438"
}
@MISC{noauthor_undated-pl,
title = "Want to be retweeted more? - Home",
abstract = "A demo for the paper The effect of wording on message
propagation: Topic- and author-controlled natural experiments
on Twitter by Chenhao Tan, Lillian Lee and Bo Pang. It can be
used to predict which tweet will be retweeted more among a
pair of tweets on the same topic.",
howpublished = "\url{https://chenhaot.com/retweetedmore/}",
note = "Accessed: 2017-7-14",
keywords = "Viralization"
}
% The entry below contains non-ASCII chars that could not be converted
% to a LaTeX equivalent.
@MISC{noauthor_undated-fa,
title = "Using Deep Learning at Scale in Twitter’s Timelines",
abstract = "Using Deep Learning at Scale in Twitter’s Timelines",
howpublished = "\url{https://blog.twitter.com/engineering/en_us/topics/insights/2017/using-deep-learning-at-scale-in-twitters-timelines.html}",
note = "Accessed: 2017-7-14",
keywords = "Viralization"
}
@MISC{Novet2017-nh,
title = "Twitter is now using a trendy type of {AI} to figure out
which tweets to show you",
booktitle = "{CNBC}",
author = "Novet, Jordan",
abstract = "Twitter has started using artificial intelligence to do a
better job of recommending relevant tweets at the top of
users' timelines.",
publisher = "CNBC",
month = "29~" # mar,
year = 2017,
howpublished = "\url{http://www.cnbc.com/2017/05/09/twitter-using-deep-learning-ai-to-rank-tweets.html}",
note = "Accessed: 2017-7-14",
keywords = "Viralization"
}
% The entry below contains non-ASCII chars that could not be converted
% to a LaTeX equivalent.
@MISC{Brownlee2016-ze,
title = "{MIT’s} {DeepDrumpf} Twitter Bot Uses Neural Networks To
Tweet Like Donald Trump",
booktitle = "{Co.Design}",
author = "Brownlee, John",
abstract = "And don’t worry. DeepLearnTheBern is next!",
publisher = "Co.Design",
month = "4~" # mar,
year = 2016,
howpublished = "\url{https://www.fastcodesign.com/3057501/mits-deepdrumpf-twitter-bot-uses-neural-networks-to-tweet-like-donald-trump}",
note = "Accessed: 2017-7-14",
keywords = "Viralization"
}
@ARTICLE{Stokowiec_undated-cg,
title = "Shallow reading with Deep Learning: Predicting popularity of
online content using only its title",
author = "Stokowiec, Wojciech and Trzci, Tomasz and Lk, Krzysztof Wo and
Marasek, Krzysztof and Rokita, Przemys Law",
keywords = "Viralization"
}
@MISC{noauthor_undated-gz,
title = "Buzzfeed Title Generator - Ravi Parikh's Website",
abstract = "Official website of Ravi Parikh.",
howpublished = "\url{http://www.ravi.io/buzzfeed-title-generator?mode=history}",
note = "Accessed: 2017-7-14",
keywords = "Viralization"
}
@MISC{noauthor_undated-mz,
title = "Buzzfeed \& Upworthy Clickbait Headline Generator",
abstract = "Looking for linkbait \& clickbait headline ideas for your
blog? Get popular Buzzfeed \& Upworthy article headlines
along with their social share count.",
howpublished = "\url{http://www.contentforest.com/copywriting-tools/clickbait-headline-generator}",
note = "Accessed: 2017-7-14",
keywords = "Viralization"
}
% The entry below contains non-ASCII chars that could not be converted
% to a LaTeX equivalent.
@MISC{Larseidnes2015-ti,
title = "{Auto-Generating} Clickbait With Recurrent Neural Networks",
booktitle = "Lars Eidnes' blog",
author = "{larseidnes}",
abstract = "``F.D.R.'s War Plans!'' reads a headline from a 1941 Chicago
Daily Tribune. Had this article been written today, it might
rather have said ``21 War Plans F.D.R. Does Not Want You To
Know About. Number 6 may shock you!''. Modern writers have
become very good at squeezing out the maximum clickability
out of every headline.…",
month = "13~" # oct,
year = 2015,
howpublished = "\url{https://larseidnes.com/2015/10/13/auto-generating-clickbait-with-recurrent-neural-networks/}",
note = "Accessed: 2017-7-14",
keywords = "Viralization"
}
@ARTICLE{Blom2015-tv,
title = "Click bait: Forward-reference as lure in online news headlines",
author = "Blom, Jonas Nygaard and Hansen, Kenneth Reinecke",
abstract = "This is why you should read this article. Although such an
opening statement does not make much sense read in isolation,
journalists often write headlines like this on news websites.
They use the forward-referring technique as a stylistic and
narrative luring device trying to induce anticipation and
curiosity so the readers click (or tap on) the headline and read
on. In this article, we map the use of forward-referring
headlines in online news journalism by conducting an analysis of
100,000 headlines from 10 different Danish news websites. The
results show that commercialization and tabloidization seem to
lead to a recurrent use of forward-reference in Danish online
news headlines. In addition, the article contributes to reference
theory by expanding previous models on phoricity to include
multimodal references on the web.",
journal = "J. Pragmat.",
volume = 76,
pages = "87--100",
month = "1~" # jan,
year = 2015,
keywords = "Online news headlines; Forward-reference; Cataphora; Discourse
deixis; Media commercialization; Tabloidization;Viralization"
}
@ARTICLE{Forbes2017-gw,
title = "Verb Physics: Relative Physical Knowledge of Actions and
Objects",
author = "Forbes, Maxwell and Choi, Yejin",
abstract = "Learning commonsense knowledge from natural language text is
nontrivial due to reporting bias: people rarely state the
obvious, e.g., ``My house is bigger than me.'' However,
while rarely stated explicitly, this trivial everyday
knowledge does influence the way people talk about the
world, which provides indirect clues to reason about the
world. For example, a statement like, ``Tyler entered his
house'' implies that his house is bigger than Tyler. In this
paper, we present an approach to infer relative physical
knowledge of actions and objects along five dimensions
(e.g., size, weight, and strength) from unstructured natural
language text. We frame knowledge acquisition as joint
inference over two closely related problems: learning (1)
relative physical knowledge of object pairs and (2) physical
implications of actions when applied to those object pairs.
Empirical results demonstrate that it is possible to extract
knowledge of actions and objects from language and that
joint inference over different types of knowledge improves
performance.",
month = "12~" # jun,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.CL",
eprint = "1706.03799"
}
@ARTICLE{noauthor_undated-ty,
title = "Multitask Learning for {Fine-Grained} Twitter Sentiment Analysis",
author = "Balikas, Georgios and Moura, Simon and Amini, Massih-Reza",
abstract = "Traditional sentiment analysis approaches tackle problems like
ternary (3-category) and fine-grained (5-category) classification
by learning the tasks separately. We argue that such
classification tasks are correlated and we propose a multitask
approach based on a recurrent neural network that benefits by
jointly learning them. Our study demonstrates the potential of
multitask models on this type of problems and improves the
state-of-the-art results in the fine-grained sentiment
classification problem.",
journal = "arXiv [cs.IR]",
month = "12~" # jul,
year = 2017
}
@ARTICLE{Evert_undated-mh,
title = "Bachelor of Science Thesis in Computer Science and Engineering",
author = "Evert, Alex and Genander, Jacob and Lallo, Nicklas and Lantz,
Rickard and Nilsson, Filip",
keywords = "Viralization"
}
@ARTICLE{Rony2017-sd,
title = "Diving Deep into Clickbaits: Who Use Them to What Extents in
Which Topics with What Effects?",
author = "Rony, Md Main Uddin and Hassan, Naeemul and Yousuf, Mohammad",
abstract = "The use of alluring headlines (clickbait) to tempt the
readers has become a growing practice nowadays. For the sake
of existence in the highly competitive media industry, most
of the on-line media including the mainstream ones, have
started following this practice. Although the wide-spread
practice of clickbait makes the reader's reliability on
media vulnerable, a large scale analysis to reveal this fact
is still absent. In this paper, we analyze 1.67 million
Facebook posts created by 153 media organizations to
understand the extent of clickbait practice, its impact and
user engagement by using our own developed clickbait
detection model. The model uses distributed sub-word
embeddings learned from a large corpus. The accuracy of the
model is 98.3\%. Powered with this model, we further study
the distribution of topics in clickbait and non-clickbait
contents.",
month = "28~" # mar,
year = 2017,
keywords = "Viralization",
archivePrefix = "arXiv",
primaryClass = "cs.SI",
eprint = "1703.09400"
}
@ARTICLE{Wei2017-qt,
title = "Learning to Identify Ambiguous and Misleading News Headlines",
author = "Wei, Wei and Wan, Xiaojun",
abstract = "Accuracy is one of the basic principles of journalism.
However, it is increasingly hard to manage due to the
diversity of news media. Some editors of online news tend to
use catchy headlines which trick readers into clicking.
These headlines are either ambiguous or misleading,
degrading the reading experience of the audience. Thus,
identifying inaccurate news headlines is a task worth
studying. Previous work names these headlines ``clickbaits''
and mainly focus on the features extracted from the
headlines, which limits the performance since the
consistency between headlines and news bodies is
underappreciated. In this paper, we clearly redefine the
problem and identify ambiguous and misleading headlines
separately. We utilize class sequential rules to exploit
structure information when detecting ambiguous headlines.
For the identification of misleading headlines, we extract
features based on the congruence between headlines and
bodies. To make use of the large unlabeled data set, we
apply a co-training method and gain an increase in
performance. The experiment results show the effectiveness
of our methods. Then we use our classifiers to detect
inaccurate headlines crawled from different sources and
conduct a data analysis.",
month = "17~" # may,
year = 2017,
keywords = "Viralization",
archivePrefix = "arXiv",
primaryClass = "cs.CL",
eprint = "1705.06031"
}
% The entry below contains non-ASCII chars that could not be converted
% to a LaTeX equivalent.
@INPROCEEDINGS{Anand2017-nf,
title = "We Used Neural Networks to Detect Clickbaits: You Won’t Believe
What Happened Next!",
booktitle = "Advances in Information Retrieval",
author = "Anand, Ankesh and Chakraborty, Tanmoy and Park, Noseong",
abstract = "Online content publishers often use catchy headlines for their
articles in order to attract users to their websites. These
headlines, popularly known as clickbaits, exploit a user’s
curiosity gap and lure them to click on links that often
disappoint them. Existing methods for automatically detecting
clickbaits rely on heavy feature engineering and domain
knowledge. Here, we introduce a neural network architecture
based on Recurrent Neural Networks for detecting clickbaits.
Our model relies on distributed word representations learned
from a large unannotated corpora, and character embeddings
learned via Convolutional Neural Networks. Experimental results
on a dataset of news headlines show that our model outperforms
existing techniques for clickbait detection with an accuracy of
0.98 with F1-score of 0.98 and ROC-AUC of 0.99.",
publisher = "Springer, Cham",
pages = "541--547",
month = "8~" # apr,
year = 2017,
keywords = "Viralization",
language = "en",
conference = "European Conference on Information Retrieval"
}
@ARTICLE{Sun2017-pr,
title = "Revisiting Unreasonable Effectiveness of Data in Deep
Learning Era",
author = "Sun, Chen and Shrivastava, Abhinav and Singh, Saurabh and
Gupta, Abhinav",
abstract = "The success of deep learning in vision can be attributed to:
(a) models with high capacity; (b) increased computational
power; and (c) availability of large-scale labeled data.
Since 2012, there have been significant advances in
representation capabilities of the models and computational
capabilities of GPUs. But the size of the biggest dataset
has surprisingly remained constant. What will happen if we
increase the dataset size by 10x or 100x? This paper takes a
step towards clearing the clouds of mystery surrounding the
relationship between `enormous data' and deep learning. By
exploiting the JFT-300M dataset which has more than 375M
noisy labels for 300M images, we investigate how the
performance of current vision tasks would change if this
data was used for representation learning. Our paper
delivers some surprising (and some expected) findings.
First, we find that the performance on vision tasks still
increases linearly with orders of magnitude of training data
size. Second, we show that representation learning (or
pre-training) still holds a lot of promise. One can improve
performance on any vision tasks by just training a better
base model. Finally, as expected, we present new
state-of-the-art results for different vision tasks
including image classification, object detection, semantic
segmentation and human pose estimation. Our sincere hope is
that this inspires vision community to not undervalue the
data and develop collective efforts in building larger
datasets.",
month = "10~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.CV",
eprint = "1707.02968"
}
@ARTICLE{Costa2017-va,
title = "Automatic Generation of Natural Language Explanations",
author = "Costa, Felipe and Ouyang, Sixun and Dolog, Peter and Lawlor,
Aonghus",
abstract = "An important task for recommender system is to generate
explanations according to a user's preferences. Most of the
current methods for explainable recommendations use
structured sentences to provide descriptions along with the
recommendations they produce. However, those methods have
neglected the review-oriented way of writing a text, even
though it is known that these reviews have a strong
influence over user's decision. In this paper, we propose a
method for the automatic generation of natural language
explanations, for predicting how a user would write about an
item, based on user ratings from different items' features.
We design a character-level recurrent neural network (RNN)
model, which generates an item's review explanations using
long-short term memories (LSTM). The model generates text
reviews given a combination of the review and ratings score
that express opinions about different factors or aspects of
an item. Our network is trained on a sub-sample from the
large real-world dataset BeerAdvocate. Our empirical
evaluation using natural language processing metrics shows
the generated text's quality is close to a real user written
review, identifying negation, misspellings, and domain
specific vocabulary.",
month = "4~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.CL",
eprint = "1707.01561"
}
@ARTICLE{noauthor_undated-qo,
title = "A causal framework for explaining the predictions of black-box
sequence-to-sequence models",
author = "Alvarez-Melis, David and Jaakkola, Tommi S",
abstract = "We interpret the predictions of any black-box structured
input-structured output model around a specific input-output
pair. Our method returns an ``explanation''' consisting of groups
of input-output tokens that are causally related. Our method
infers these dependencies by querying the model with perturbed
inputs, generating a graph over tokens from the responses, and
solving a partitioning problem to select the most relevant
components. We focus the general approach on sequence-to-sequence
problems, adopting a variational autoencoder to yield meaningful
input perturbations. We test our method across several NLP
sequence generation tasks.",
journal = "arXiv [cs.LG]",
month = "6~" # jul,
year = 2017
}
@ARTICLE{Ruder2016-hu,
title = "An overview of gradient descent optimization algorithms",
author = "Ruder, Sebastian",
abstract = "Gradient descent optimization algorithms, while increasingly
popular, are often used as black-box optimizers, as
practical explanations of their strengths and weaknesses are
hard to come by. This article aims to provide the reader
with intuitions with regard to the behaviour of different
algorithms that will allow her to put them to use. In the
course of this overview, we look at different variants of
gradient descent, summarize challenges, introduce the most
common optimization algorithms, review architectures in a
parallel and distributed setting, and investigate additional
strategies for optimizing gradient descent.",
month = "15~" # sep,
year = 2016,
archivePrefix = "arXiv",
primaryClass = "cs.LG",
eprint = "1609.04747"
}
@ARTICLE{Serra2017-rl,
title = "Getting deep recommenders fit: Bloom embeddings for sparse
binary input/output networks",
author = "Serr{\`a}, Joan and Karatzoglou, Alexandros",
abstract = "Recommendation algorithms that incorporate techniques from
deep learning are becoming increasingly popular. Due to the
structure of the data coming from recommendation domains
(i.e., one-hot-encoded vectors of item preferences), these
algorithms tend to have large input and output
dimensionalities that dominate their overall size. This
makes them difficult to train, due to the limited memory of
graphical processing units, and difficult to deploy on
mobile devices with limited hardware. To address these
difficulties, we propose Bloom embeddings, a compression
technique that can be applied to the input and output of
neural network models dealing with sparse high-dimensional
binary-coded instances. Bloom embeddings are computationally
efficient, and do not seriously compromise the accuracy of
the model up to 1/5 compression ratios. In some cases, they
even improve over the original accuracy, with relative
increases up to 12\%. We evaluate Bloom embeddings on 7 data
sets and compare it against 4 alternative methods, obtaining
favorable results. We also discuss a number of further
advantages of Bloom embeddings, such as 'on-the-fly'
constant-time operation, zero or marginal space
requirements, training time speedups, or the fact that they
do not require any change to the core model architecture or
training configuration.",
month = "13~" # jun,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.LG",
eprint = "1706.03993"
}
@ARTICLE{Shwartz-Ziv2017-jx,
title = "Opening the Black Box of Deep Neural Networks via
Information",
author = "Shwartz-Ziv, Ravid and Tishby, Naftali",
abstract = "Despite their great success, there is still no comprehensive
theoretical understanding of learning with Deep Neural
Networks (DNNs) or their inner organization. Previous work
proposed to analyze DNNs in the
\textbackslashtextit\{Information Plane\}; i.e., the plane
of the Mutual Information values that each layer preserves
on the input and output variables. They suggested that the
goal of the network is to optimize the Information
Bottleneck (IB) tradeoff between compression and prediction,
successively, for each layer. In this work we follow up on
this idea and demonstrate the effectiveness of the
Information-Plane visualization of DNNs. Our main results
are: (i) most of the training epochs in standard DL are
spent on \{\textbackslashemph compression\} of the input to
efficient representation and not on fitting the training
labels. (ii) The representation compression phase begins
when the training errors becomes small and the Stochastic
Gradient Decent (SGD) epochs change from a fast drift to
smaller training error into a stochastic relaxation, or
random diffusion, constrained by the training error value.
(iii) The converged layers lie on or very close to the
Information Bottleneck (IB) theoretical bound, and the maps
from the input to any hidden layer and from this hidden
layer to the output satisfy the IB self-consistent
equations. This generalization through noise mechanism is
unique to Deep Neural Networks and absent in one layer
networks. (iv) The training time is dramatically reduced
when adding more hidden layers. Thus the main advantage of
the hidden layers is computational. This can be explained by
the reduced relaxation time, as this it scales
super-linearly (exponentially for simple diffusion) with the
information compression from the previous layer.",
month = "2~" # mar,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.LG",
eprint = "1703.00810"
}
@ARTICLE{Fortunato2017-ij,
title = "Noisy Networks for Exploration",
author = "Fortunato, Meire and Azar, Mohammad Gheshlaghi and Piot,
Bilal and Menick, Jacob and Osband, Ian and Graves, Alex and
Mnih, Vlad and Munos, Remi and Hassabis, Demis and Pietquin,
Olivier and Blundell, Charles and Legg, Shane",
abstract = "We introduce NoisyNet, a deep reinforcement learning agent
with parametric noise added to its weights, and show that
the induced stochasticity of the agent's policy can be used
to aid efficient exploration. The parameters of the noise
are learned with gradient descent along with the remaining
network weights. NoisyNet is straightforward to implement
and adds little computational overhead. We find that
replacing the conventional exploration heuristics for A3C,
DQN and dueling agents (entropy reward and $\epsilon$-greedy
respectively) with NoisyNet yields substantially higher
scores for a wide range of Atari games, in some cases
advancing the agent from sub to super-human performance.",
month = "30~" # jun,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.LG",
eprint = "1706.10295"
}
@UNPUBLISHED{Beaulieu-Jones2017-qt,
title = "Privacy-preserving generative deep neural networks support
clinical data sharing",
author = "Beaulieu-Jones, Brett K and Wu, Zhiwei Steven and Williams, Chris
and Greene, Casey S",
abstract = "Though it is widely recognized that data sharing enables faster
scientific progress, the sensible need to protect participant
privacy hampers this practice in medicine. We train deep neural
networks that generate synthetic subjects closely resembling
study participants. Using the SPRINT trial as an example, we show
that machine-learning models built from simulated participants
generalize to the original dataset. We incorporate differential
privacy, which offers strong guarantees on the likelihood that a
subject could be identified as a member of the trial.
Investigators who have compiled a dataset can use our method to
provide a freely accessible public version that enables other
scientists to perform discovery-oriented analyses. Generated data
can be released alongside analytical code to enable fully
reproducible workflows, even when privacy is a concern. By
addressing data sharing challenges, deep neural networks can
facilitate the rigorous and reproducible investigation of
clinical datasets.",
journal = "bioRxiv",
pages = "159756",
month = "5~" # jul,
year = 2017,
language = "en"
}
@ARTICLE{Lerer2017-ee,
title = "Maintaining cooperation in complex social dilemmas using
deep reinforcement learning",
author = "Lerer, Adam and Peysakhovich, Alexander",
abstract = "In social dilemmas individuals face a temptation to increase
their payoffs in the short run at a cost to the long run
total welfare. Much is known about how cooperation can be
stabilized in the simplest of such settings: repeated
Prisoner's Dilemma games. However, there is relatively
little work on generalizing these insights to more complex
situations. We start to fill this gap by showing how to use
modern reinforcement learning methods to generalize a highly
successful Prisoner's Dilemma strategy: tit-for-tat. We
construct artificial agents that act in ways that are simple
to understand, nice (begin by cooperating), provokable (try
to avoid being exploited), and forgiving (following a bad
turn try to return to mutual cooperation). We show both
theoretically and experimentally that generalized
tit-for-tat agents can maintain cooperation in more complex
environments. In contrast, we show that employing purely
reactive training techniques can lead to agents whose
behavior results in socially inefficient outcomes.",
month = "4~" # jul,
year = 2017,
archivePrefix = "arXiv",
primaryClass = "cs.AI",
eprint = "1707.01068"
}
@ARTICLE{Carreira-Perpinan2017-if,
title = "Model compression as constrained optimization, with
application to neural nets. Part I: general framework",
author = "Carreira-Perpi{\~n}{\'a}n, Miguel {\'A}",
abstract = "Compressing neural nets is an active research problem, given
the large size of state-of-the-art nets for tasks such as
object recognition, and the computational limits imposed by
mobile devices. We give a general formulation of model
compression as constrained optimization. This includes many
types of compression: quantization, low-rank decomposition,
pruning, lossless compression and others. Then, we give a
general algorithm to optimize this nonconvex problem based
on the augmented Lagrangian and alternating optimization.